Site icon

ACI Multi-Pod Caveats and Considerations

Screen Shot 2018-07-25 at 1.58.51 PM

Originally I was going to create a detailed configuration guide for Multi-Pod, however, after checking out the the Cisco ACI MultiPod Configuration Whitepaper on CCO I realized I would be duplicating efforts at best. The configuration whitepaper on CCO contains detailed configuration examples for the IPN and screenshots from the APIC. If you are looking to deploy Multi-Pod, this is what you will want to use!

https://www.cisco.com/c/en/us/solutions/collateral/data-center-virtualization/application-centric-infrastructure/white-paper-c11-739714.html#_Toc495712857
So – instead of the detailed configuration guide, I have worked to collect important caveats, design guidelines, and requirements that can serve as a one-stop shop for ACI Multi-Pod deployments.

Why Multi-Pod?

Before we dig in, let’s examine some of the benefits that come with deploying ACI Multi-Pod.

  • ACI Multi-Pod allows for fault isolation of Control Plane Protocols inside of each respective Pod (IS-IS and COOP protocols operate intra-Pod). This is important because it ensures that a routing or protocol issue in one Pod will not take down the entire fabric.
  • Centralized Management through one APIC Cluster (Cluster of 3, 5, of 7).
  • Eliminates fate sharing if one DC goes down
  • End-to-end Policy Enforcement
  • Pervasive GW across all leafs in all Pods

Hardware and Software Requirements

APIC Software Requirement

  • 2.0(1) on APIC
  • 12.0(1) on Leafs and Spines
  • Ability to change MTU for Inter-Pod traffic (CPU generated traffic) – available starting with 2.2(2e)
  • 50msec latency across Pods – Supported starting 2.3(1)

APIC Hardware Requirement

  • Any APIC Model
  • All Nexus 9000 Platforms for Fabric nodes

IPN node options

While the list below is not all inclusive, it gives a quick reference of platforms that can act as Inter-Pod Node devices.

  • Nexus 9200, 9300-EX or later (not first generation 9300/9500).
  • Nexus 7000
  • ASR9k
  • ASR1k

Regardless of which you choose, the IPN device must support the following:

  • OSPF
  • PIM Bi-Dir (will handle Unknown unicast, multicast and broadcast traffic)
  • Jumbo MTU (will allow for VXLAN encapsulated traffic between Pods)
  • DHCP Relay (will allow for Spine and Leaf nodes to be discovered across the IPN)
  • QOS (will support the prioritization of APIC-to-APIC communication over the IPN)
  • Support for 40/100G interfaces (or 10Gig with QSA adapters on the Spine)

Multi-Pod Scalability numbers

  • 2.2(x) – 10 pods / 200 leafs per pod /No more than 400 leafs per fabric
  • 3.2(x) – 12 pods / 200 leafs per pod / No more than 400 leafs per fabric
  • 4.2(x) – 12 pods / 400 leafs per pod / No more than 500 leafs per site

Multi-Site vs Multi-Pod from an Availability Zone / Region perspective

First – some definitions (courtesy of AWS). If you’d like to know more about AZs/Regions on AWS, check out this link.

Regions – Geographically separated areas, each of which can be comprised of multiple availability zones.

Availability Zones – Locations within a Region that are designed to be isolated fault domains which provide low latency network connectivity to other Availability zones in the same Region.

* Note – What you read below is not mandate for every use-case of a Multi-Pod or Multi-Site design, rather it is a guideline that you can use when deciding which is more appropriate for your overall design.

When it comes to mapping the Availability Zone / Region concepts to ACI designs, here are some thoughts:

  • Each “Fabric”, which is made up of multiple Pods, is considered a Single Availability Zone.
  • Configuration Zones (an ACI concept) can be used to span certain switches in your fabric, and can be used to map the configuration zone to an availability zone (applies to configuration and policy only).
  • “Multiple Fabrics (whether a single Pod or Multi-Pod) connected by Multi-Site” is considered a Region with multiple Availability Zones.
  • The original use case for Multi-Pod is to have multiple “pods” of ACI at a customer location, linked together via Multi-Pod (i.e., AZ in one Region)
  • Multi-Site allows you to manage and connect multiple sites (Several Fabrics) via the Multi-Site Orchestrator (MSO)*

* NOTE – The ability to connect Multi-Pod Fabrics with Multi-Site became available beginning with APIC 3.2 code.

Multi-Pod vs Multi-Site at a glance

https://www.cisco.com/c/dam/en/us/td/docs/switches/datacenter/aci/aci_multi-site/sw/1x/fundamentals/Multi-Pod_vs_Multi-Site_Infographic.pdf

Guidelines and Limitations

Caveats

Design Considerations

Design Considerations for APICs

When it comes to deploying APICs in an ACI Multi-Pod Design, there are several ways to spread out your APICs across the various Pods, depending on how many Pods you intend to connect together. Cisco has a lot of good information to consider located on their Multi-Pod Whitepaper on CCO.

If you are just going with (2) Pods, which is very prevalent in DR/BC deployments, my rule of thumb is (2) APICs in Pod1, (1) APIC in Pod2, and a Standby APIC in both Pod1/2. My main reason for this is that you will alway have a full copy of the APIC Shards in both Pods.

Design Considerations with Firewalls

Deploying firewalls across pods is a common requirement for most customers. With ACI MultiPod, there are several validated design options for deploying firewalls across your MultiPod environment.
For more details on these options, please check out the ACI MultiPod and Service Node Whitepaper on CCO.

Active/Active Firewalls

As of APIC 3.2(4d), you can use an active/active firewall cluster that is stretched across pods. When deploying Cisco ASA or Cisco Firepower firewall appliances, this deployment model takes the name of “split spanned EtherChannel” and ensures that all the nodes of the cluster “own” the same MAC/IP values so that the stretched firewall cluster appears as a single logical entity to the ACI Multi-Pod fabric. This deployment model removes any concern about the possible creation of asymmetric traffic paths for both east-west and north-south traffic flows, as traffic will be dynamically redirected to the specific firewall node owning the connection state for that specific traffic flow.

Active/Standby Firewalls

Option 1 – Active/Standby FW with NO use of vPC

Option 2 – Active/Standby FW with vPC

  • Active and Standby FWs can be connected to respective Leaf pairs in each Pod.
  • This is supported in the 2.3 release starting with 2.3(1) on EX/FX Leaf switches ONLY. This option is not supported on first generation ACI Leaf switches due to CSCux29124.

Troubleshooting

APIC in Pod1 cannot ssh to spine/leaf in Pod2

APICs are not VRF aware; because of this, if you are trying to SSH from your APIC to Leaf or Spines in other Pods you will have to specify the the interface or use the attach command.
Using the “attach” command will automatically source the correct interface to allow for SSH.
coast-apic1# acidiag fnvread
      ID   Pod ID                 Name    Serial Number         IP Address    Role        State   LastUpdMsgId
————————————————————————————————————–
     101        1               Spine1      FOX2123PLDK      10.0.32.65/32   spine         active   0
     102        2               Spine2      FOX2123PLD2     10.1.152.64/32   spine         active   0
     201        1              Leaf201      FDO21250W9K      10.0.32.67/32    leaf         active   0
     202        1              Leaf202      FDO21260TFC      10.0.32.66/32    leaf         active   0
     203        1              Leaf203      FDO21242YD1      10.0.32.68/32    leaf         active   0
     204        1              Leaf204      FDO21260TBZ      10.0.32.64/32    leaf         active   0
     205        2              Leaf205      FDO21250W1Y     10.1.152.66/32    leaf         active   0
     206        2              Leaf206      FDO21253CGH     10.1.152.65/32    leaf         active   0
 
Total 8 nodes
 
coast-apic1# attach Leaf206
This command is being deprecated on APIC controller, please use NXOS-style equivalent command
# Executing command: ssh Leaf206 -b 10.0.0.1
Warning: Permanently added ‘leaf206,10.1.152.65’ (RSA) to the list of known hosts.
 
Password:
Cisco Nexus Operating System (NX-OS) Software
TAC support: http://www.cisco.com/tac
Specifying the the interface in your SSH command
coast-apic1# ssh admin@10.1.152.65 -b 10.0.0.1
Password:
Last login: Mon Jul 23 11:24:48 2018 from 10.0.0.1
Cisco Nexus Operating System (NX-OS) Software
TAC support: http://www.cisco.com/tac
Exit mobile version