- The 20 Best iPhone/iPad Games of 2013 So Far
- 9 Steps to Build Your Personal Brand (and Your Career)
- 7 Consumer Technologies Coming to an Enterprise Near You
- 11 Signs Your IT Project is Doomed
Network World - Despite vendor pledges to support existing or developing industry standards, users are expected to deploy single-vendor data center and cloud switching fabrics from their primary suppliers.
Standards such as Transparent Interconnect of Lots of Links (TRILL), Shortest Path Bridging and Multi-Chassis Link Aggregation will be embraced by those vendors looking to dent Cisco's dominance in data center switching. The standards support will soothe customers looking to avoid vendor lock-in, but it's unlikely IT shops will mix and match multivendor switches within and between data centers.
"In the data center, customers do look for that path with the least bumps in the road just because of the critical need of the data center," says Zeus Kerravala, principal at ZK Research.
Cisco, meanwhile, will continue to expand and enhance its FabricPath approach to data center fabric switching -- which the company says is a "superset" of TRILL but is not founded on it. Juniper will continue advocating a tagging mechanism in the Broadcom silicon inside its QFabric line to support multiple active paths and one-hop reachability in data centers and cloud environments.
Brocade says the current implementation of its VCS fabric technology does not include all of the features that are found in the TRILL standard. The company says its VDX data center switches can operate in VCS mode or in "classic" mode, which is more adherent to IEEE 802.x standards. The data plane in a VCS implementation uses TRILL; but the control plane does not, it uses Fabric Shortest Path First (FSPF), an ANSI standard used by all Fibre Channel SAN fabrics as the link-state routing protocol.
Alternative link-state routing protocols can be supported in VCS when they are standardized and available, Brocade says. But the link state protocol in TRILL, Intermediate System-to-Intermediate System (IS-IS), has been documented since July 2011 and even tested for interoperability at the University of New Hampshire almost a year before that, says Donald Eastlake, chairman of the TRILL working group in the IETF.
"RFC 6326 is TRILL support of IS-IS," Eastlake says. "That has been out for a bit now. I think this year there will be a number of switches from a number of vendors that will support the TRILL control plane."
Eastlake wouldn't speculate on why vendors have been delaying full compliance of TRILL data and control planes in their fabric switches.
"There's nothing inherently evil about using the TRILL data plane instead of the control plane," he says. "It won't interoperate with a standard TRILL control plane but people manufacture things that don't interoperate all the time. They make the business decision that that's what they want to do. Other people make things that try very hard to be very interoperable because that's the business decision that they make."
HP says it supports TRILL and its own Intelligent Resilient Framework (IRF) multichassis bonding technology for flattening the data center network. HP just announced its new 5900 line of top-of-rack switches that support both.
Arista Networks, Extreme Networks and Alcatel-Lucent all currently advocate MC-LAG as the fabric architecture for their switches. MC-LAG is an IEEE standard that is intended to replace the Ethernet Spanning Tree Protocol to improve resiliency and uptime, and reduce latency, by creating active/active network paths for load balancing and redundancy.
Since the fabrics are based on Ethernet, they should be as easy to deploy as Ethernet, according to Extreme.
"Ethernet is open," says Doug Wills, senior director of marketing at Extreme. "Why not Ethernet fabrics?"
Avaya supports the IEEE's Shortest Path Bridging (SPB) specification, an alternative to TRILL, for its VENA fabric architectures. Shortest Path Bridging is founded on the IEEE 802.1ah Provider Backbone Bridging MAC-in-MAC standard, which in the telecom world is proposed as a Layer 2 alternative to MPLS for Metro Ethernet deployments.
Some vendors, though, hedge their bets by supporting both SPB and TRILL. Huawei is in this camp.
Juniper chose to implement a tagging mechanism in the Broadcom silicon in its QFabric line because it functions like a fabric inside the switch itself; if a switch itself functions as a fabric, a fabric configured with like switches can form a flatter topology, Juniper asserts.
"You don't actually run a protocol," says Andy Ingram, worldwide managing director of data center sales for Juniper. "We use the hardware to transport the bits between the ingress and the egress point in the switch fabric."
Ingram says this has several advantages: It can be managed as a single logical device -- there is no need to manage each node of the spine and leaf topology; it lowers latency because there are fewer ASICs in the data path; and it should be less expensive because of the fewer components.
"The cards inside of a fabric switch interconnect, we don't ask them to be intelligent," Ingram says. "We're at 3kW versus something that might be 13kW. Because we're not asking it to be a switch, we can make it more efficient, we can make it faster, we can make it less expensive."
A packet is tagged and hashed on which of the routes it will take through the interconnect, Ingram says. This capability is built into the Broadcom Titan and Trident ASICs used in the QFabric nodes and interconnects.
Juniper says it has 100 customers for QFabric, more than half of which are in production mode -- for the stand-alone QFX3500 node/top-of-rack switch and/or the QFabric Interconnect, according to Ingram. He also says at least one-third are brand-new Juniper customers who've literally bought into the QFabric architecture.
There's been speculation that QFabric is having a hard time living up to its 6,144 10G port/one-hop billing in customer trials, but Ingram says Juniper's been working closely with customers through a prolonged testing period for what's essentially a new data center networking architecture.
"We have one that will go production in the next 60 days that was a beta account," he says. "It's a big deal, it's a big part of their business, it's a service provider in the cloud space, and they've been testing it extensively for some time. They're waiting for the next release to come out which would have some of the capabilities that are important for them to go to production. We have quite a few that will go into production in the next three to four months."