The Anna Karenina principle was mentioned in December, 2004, interview with IBM's storage virtualisation architect, Steve Legg. In it he said that the best and architecturally cleanest place for a storage virtualisation function was inside the storage fabric. This was partly because the fabric could better see all the storage devices than a storage array controller such as HDS' TagmaStore.
IBM chose to implement its virtalisation product, the SAN Volume Controller or SVC, in an appliance which links to a SAN fabric switch or director. Since then there has been the development of intelligent directors which can run storage applications. A standard interface is being developed for these, known as the Fabric Application Interface Standard or FAIS.
We talked again to Steve Legg and found out IBM's current thinking on storage virtualisation.
TW:What is the current situation with SVC?
Steve Legg: There are now over 1,800 SVC systems installed. We're probably the storage virtualisation leader based on a units or revenue measure.
TW: Could you remind of us your product platform choices when you developed it?
Steve Legg: We chose our in-band technology very much on the grounds that, then in 1999, it served the purpose with the hardware available.
TW: And the situation now?
Steve Legg: I'm not hung up on the platform where it runs. We need memory mips and MB/sec bandwidth. It's been designed so that we could always move it elsewhere if we want to. Now processing in the fabric switch (director) is starting to mature. Whilst we have no product plans right now, we are having discussions about how we can use that processing power in the switch.
The location of storage virtualisation won't have moved; it would still be in the SAN fabric. But within the SAN fabric you can do it in a variety of places; an appliance or the switch.
TW: As EMC's InVista does. Could you give us your view of InVista here?
Steve Legg: InVista is not out there yet; it's in beta. InVista will provide logical-to-physical translation. It doesn't cover cacheing, point-in-time copy or other copy services that SVC does. It's not trivial to do this.
If you just want to do virtualisation then do it in the switch. If you want to do more then you need more than you can do in the switch.
TW: IBM knows about what you can and can't do in switches because it has experimented with a switch-based version of SVC.
Steve Legg: Our SVC in one implementation is on two appliances that are Intel-based and in an active:active pairing. They link to the switch in a full-duplex way through four fabric ports.
Another SVC implementation has it fit inside a Cisco switch as an embedded product based on a Mips processor. Approximately nine were sold. Technically it was a complete success, commercially a complete flop.
Switch slots represent a very specialised space. The switch has a specific backplane. An embedded design in the switch is limited in scalability. Cisco and others are now putting intelligence into fabric ports themselves. We could exploit these mips, MB and MB/sec for virtualisation. (But) for copy services, etc. we can best do it still in the appliance. We could split off virtualisation functions into the switch and use the (existing) appliance for cacheing and helping out with copy services.
TW: Steve Legg than looked at the appliance's capacity and scalability. (Previously, in-band virtualisation approaches have been criticised for introducing a potential bottleneck into the SAN fabric.)
Steve Legg: Our SVC appliance has active:active nodes; they're paired for availability. We cluster them for scalability. There hasn't been anything said about SVC bottlenecking.
The appliance has to deal with the aggregate workload coming from the hosts, not the theoretical maximum I/O bandwidth of all the switch ports or I/O controllers. A host connected to a SAN don't generate hundreds of I/Os a second. It's three or four I/Os a second because most of the time hosts are running business logic. The fan-in from hosts to the SVC can be huge.
TW: So the picture here is that a SAN fabric director (switch in Legg's phraseology) can be an effective place to run basic virtualisation but the physical space available in the switch limits the scalability and the storage application stack that could be needed. The best place to have applications that are layered onto the virtualisation function is in a connected (in-band) appliance which can scale up in power and availability as required.
We then asked Steve Legg about FAIS, and a couple of questions about Network Appliance and storage products from suppliers such as 3Par, Exanet and BlueArc.
What is your view of FAIS?
Steve Legg: FAIS is a proposed standard. We're totally committed to that being the way to do it.
TW: Can you briefly mention where NetApp fits in with SVC?
Steve Legg: NetApp filers don't connect to SVC, The NETApp gateway, the V-series has backend storage which are Fibre Channel arrays. The SVC fits there. There is overlap here. There always will be.
TW: What do think of the virtualisation approaches from 3PAR, Exanet, BlueArc and so on?
Steve Legg: Here the all-in-one environments are a limiting factor. You can't easily change the architecture if you want to split off functions. The SVC is open. You can buy whatever disk or disk arrays you like. You can move virtualisation to the switch if you like.
TW: And lastly, a quick look at storage controller-based approaches to virtualisation?
Steve Legg: If you put virtualisation and storage applications in the storage controller it's inflexible. You can't move functions easily to the fabric; it's limiting. Also suppliers will inevitably optimise components to improve performance and/or cost and thereby prejudice openness.
TW: As an architect Legg is strongly biased in favour of clean interfaces between functional levels in a stack or interfaces in a network. It's the OSI 7-layer model approach, one that has proved its enduring worth over time.
We might expect to see IBM hiving basic virtualisation off the SVC appliance and running it inside a Brocade, Cisco or McDATA director. But the indications are that it believes it can provide better storage virtualisation-based functions by running them, still, on its SVC appliance.
With customer installations heading towards 2,000 it is an approach that resonates well with a lot of customers. Many of them, if not all of them, will have Brocade, Cisco or McDATA directors. Why should they change their approach?
In one sense it will be an easy choice to make. What will the director-based storage applications be able to do now and in the future that the SVC appliance won't? What are the comparable costs and management tasks? What do the roadmaps look like? Tick the boxes and choose. To throw out an incumbent the director-based products have got to be really good. Expect a flood of white papers and analyst reports trying to prove things one way or the other.