VMDC 3.0.1 Design Overview

The Virtualized Multiservice Data Center (VMDC) architecture is based on the foundational design principles of modularity, high availability, differentiated service support, secure multi-tenancy, and automated service orchestration(Figure 2-1).

Design Principles

Design principles provide the benefits of streamlined turn-up of new services, maximized service availability, resource optimization, facilitated business compliance and support for self-service IT models, thus maximizing operational efficiency and allowing the private or public cloud provider to focus on their core business objectives.

Figure 2-1 VMDC Design Principles

VMDC 3.0 introduced FabricPath into data center designs because FabricPath allows for the creation of simple, scalable, and efficient Layer 2 domains that apply to many network scenarios. FabricPath brings the stability and scalability of routing to Layer 2. The switched domain does not have to be segmented any more, providing data center-wide workload mobility. Because traffic is no longer forwarded using Spanning Tree Protocol (STP), the bisectional bandwidth of the network is expanded, providing enhanced scalability.

•FabricPath is simple to configure. The only necessary configuration consists of distinguishing the core ports, which link the switches, from the edge ports, where end devices are attached. No parameters need to be tuned to achieve operational status, and switch addresses are assigned automatically.

•A single control protocol is used for unicast forwarding, multicast forwarding, and VLAN pruning. Networks designed using FabricPath protocol require less combined configuration than equivalent networks based on Spanning Tree Protocol, further reducing the overall management needed for the solution.

•Static network designs make assumptions about traffic patterns and the locations of servers and services. If those assumptions are incorrect, which often becomes the case over time, complex redesign may be necessary. A fabric switching system based on Cisco FabricPath can be easily expanded as needed with additional access nodes in a plug and play manner, with minimal operational impact.

•Switches that do not support Cisco FabricPath can still be attached to the FabricPath fabric in a redundant way without resorting to Spanning Tree Protocol.

•The capabilities of Cisco FabricPath Layer 2 troubleshooting tools provide parity with those currently available in the IP community for non-fabric path environments. For example, the Ping and Traceroute features now offered at Layer 2 with FabricPath can measure latency and test a particular path among the multiple equal-cost paths to a destination within the fabric.

Reliability Based on Proven Technology

•Though Cisco FabricPath offers a plug-and-play user interface, its control protocol is built on top of the powerful Intermediate System-to-Intermediate System (IS-IS) routing protocol, an industry standard that provides fast convergence and is proven to scale in the largest service provider environments.

•Loop prevention and mitigation is available in the data plane, helping ensure safe forwarding unmatched by any transparent bridging technology. Cisco FabricPath frames include a time-to-live (TTL) field similar to the one used in IP, and an applied reverse-path forwarding (RPF) check.

Efficiency and High Performance

•With FabricPath, equal-cost multipath (ECMP) protocols can be used in the data plane enabling the network to find optimal paths among all the available links between any two devices. First-generation hardware supporting Cisco FabricPath can perform 16-way ECMP, which, when combined with 16-port 10 Gbps port-channels, represents bandwidth of up to 2.56 terabits per second (Tbps) between switches.

•With FabricPath, frames are forwarded along the shortest path to their destination, reducing the latency of the exchanges between end stations compared to a Spanning Tree based solution.

•Cisco FabricPath needs to learn at the edge of the fabric only a subset of the MAC addresses present in the network, enabling massive scalability of the switched domain.

Terminology

FabricPath comprises two types of nodes: spine nodes and leaf nodes. A spine node is one that connects to other switches in the fabric and a leaf node is one that connects to servers. These terms are useful in greenfield scenarios but may be vague for migration situations, where one has built a hierarchical topology and is accustomed to using traditional terminology to describe functional roles. In this document, we expand our set of terms to correlate fabric path nodes and functional roles to hierarchical network terminology. These terms are:

•Aggregation-Edge—This is a FabricPath node that sits at the "edge" of the fabric, corresponding to the aggregation node in a hierarchical topology.

•Access-Edge—This is a Fabricpath node that sits at the edge of the fabric, corresponding to the access node in a hierarchical topology.

These nodes may perform Layer 2 and/or Layer 3 functions. Hence at times, we also refer to an Layer 3 spine or an Layer 3 edge node to clarify the location of Layer 2/Layer 3 boundaries and distinguish between nodes that are performing Layer 3 functions versus Layer 2-only functions.

FabricPath Topologies

FabricPath can be implemented in a variety of network designs, from full-mesh to ring topologies. The following sections discuss three DC design options based on FabricPath that were considered in VMDC 3.0 design and validation. These design options are:

Note VMDC 3.0.1 is based solely on the Typical Data Center design option.

Typical Data Center

A Typical Data Center design is a 2-tier FabricPath design as depicted in Figure 2-2. All VMDC architectures are built around modular building blocks called PoDs. Each PoD uses a localized Services attachment model. Within a PoD, Virtual Port Channels (vPCs) handle Layer 2 switching provide an active-active environment that does not depend on Spanning Tree Protocol (STP) but converges quickly in the event of failure. Figure 2-2 shows a VMDC PoD with FabricPath as a vPC replacement.

Figure 2-2 Typical Data Center Design

From a resilience perspective, a vPC-based design is sufficient at this scale, although there are other benefits of using FabricPath, including:

•FabricPath is simple to configure and manage. There is no need to identify a pair of peers or configure Port Channels. Nevertheless, Port Channels can still be leveraged in FabricPath topologies if needed.

•FabricPath is flexible, it does not require a particular topology, and will function even if the network is currently cabled for the classic triangle vPC topology. FabricPath can accommodate any future design.

•FabricPath does not use or extend STP. Even a partial introduction of FabricPath benefits the network because it segments the span of STP.

•FabricPath can be extended easily without degrading operations. Adding a switch or a link in a FabricPath-based fabric does not result in lost frames. Therefore, it is possible to start with a small network and extend it gradually, as needed.

•FabricPath increases the pool of servers that are candidates for VM mobility and thereby enables more efficient server utilization.

Note Certain application environments, especially those that generate high levels of broadcast, may not tolerate extremely large Layer 2 environments.

Switched Fabric Data Center

Once FabricPath is introduced into a basic topology, additional options can be used to expand the data center. The first option is to extend the spine, adding more spine nodes for additional Layer 2 or Layer 3 forwarding to form a Switched Fabric Data Center (Figure 2-3), which can be viewed as a single expanded FabricPath-based PoD.

A switched fabric datacenter can take advantage of FabricPath capabilities, including 16-way ECMP, to create a non-blocking fabric. In this design, servers can communicate with each other with little oversubscription.

Figure 2-3 Switched Fabric Data Center or Clos Fabric

The topology shown inFigure 2-4shows a Switched Fabric Data Center using an edge switch pair as the Layer 3 boundary with Localized Services attached. It is drawn unfolded for comparison to the Typical Data Center Design in the previous section. A key feature of this design is the dedicated spine that allows the Layer 2 domain to be extended easily and optimizes the paths for East-West type traffic, giving predictable network characteristics.

Figure 2-4 Switched Fabric Data Center with Localized Services

Extended Switched Fabric Data Center (3 Tier)

The Extended Switched Fabric Data Center (Figure 2-5) blends concepts from the Typical Data Center and Switched Fabric Data Center designs into a single functional model. As previously described, a PoD can be designed around the Typical Data Center or Switched Fabric Data Center model. In this case, a dedicated spine is used to further extend a Layer 2 domain across multiple PoDs which can reside in a single building on a single floor or between floors, or in a data center deployed in a campus type environment where different PoDs reside within different physical buildings. As FabricPath extends the Data Center network horizontally, services may remain localized within a PoD, be centralized between PoDs or distributed. Figure 2-5 shows distributed service node attachment.

VMDC Virtualized Containers

The VMDC architecture is capable of supporting multiple different virtual containers, referred to as consumer models. These consumer models are described in greater detail later in this document, and in earlier release material:

In VMDC 3.0.1, it is possible to define and deploy all previously defined VMDC consumer models. FabricPath does not impose limitations on this aspect of design. However, for validation purposes, VMDC 3.0.1 (like 3.0) focused on the Palladium model, which was introduced in VMDC 2.1 and is further refined in VMDC 3.0.1 to fit the needs of private cloud deployments.Figure 2-6 shows a single tenant modified Palladium-type container. A key difference between VMDC 3.0 and VMDC 2.1 is in the details of the public zone implementation: in VMDC 3.0.1, a single public default VRF is shared across all tenants, as opposed to employing dedicated VRFs per tenant. The default VRF is part of the global routing table.

Figure 2-6 VMDC Palladium Container

In summary, the validated VMDC 3.0 Palladium containers have the following characteristics:

•A single, shared public zone with multiple server VLANs and a single ACE context (or multiple contexts) for Server Load Balancing. This public zone is in the global routing table.

•Multiple, firewall private zones reachable via the public zone - in other words, each firewall's outside interface is in the public zone. These private zones include an ACE SLB and may have one or more VLANs.

•A VSG virtual firewall can be applied in a multi-tenant/shared fashion to the public zone.

•A VSG can be applied in a dedicated fashion to each of the private zones, providing a second tier of policy enforcement, and back-end (E/W) zoning.

Solution Components

The following sections provide the network components utilized in the VMDC 3.0 and 3.0.1 solutions (Table 2-1). Snapshots of the two final network topology models validated in VMDC 3.0, Typical End-to-End (Figure 2-7), and Extended End-to-End (Figure 2-8) are included, plus diagrams specific to the Typical DC topologies employed in the subsequent VMDC 3.0.1 update. However, it should be noted that earlier analysis focusing on FabricPath resiliency explored additional options such as Nexus 5000-based leaf and spine nodes, and the alternative of routing at the access-edge node rather than the aggregation-edge. This analysis yielded good convergence results across the fabric; generally sub-3 second for most aggregation-edge failures and subsecond for access-edge failures but did not include service node attachment. The models included in this document were selected for expanded end-to-end validation as generally more representative of current deployment trends and relevance.

•Release 2.2—Builds on top of releases 2.0 and 2.1 for a common release supporting public, private, and hybrid cloud deployments. Enhancements include "defense in depth" security, multi-media QoS support, and Layer 2 (VPLS) based DCI.

•VMDC 2.3—Further expands on topology models in previous 2.X releases, providing a more collapsed architectural model, offering smaller footprint and entry point option. Enhancements include introduction of a new "copper" tenancy container mode.

Related Documents

The following documents are available for reference and consideration.