This document provides network configuration guidance for Azure Local clusters using disaggregated storage architecture with Cisco Nexus Top-of-Rack (ToR) switches.
- Azure Local Disaggregated Storage Network Design with Cisco Nexus Switches
This document assists administrators in designing a network architecture that aligns with Azure Local cluster requirements. It provides reference architectures and sample configurations for network devices supporting cluster deployments. Equipment such as switches, firewalls, or routers located on customer premises is considered out of scope, as it is assumed to be part of the existing infrastructure. The focus is on Node-to-ToR, ToR-to-ToR, and ToR uplink configurations.
Definitions
- ToR: Top of Rack network switch. Supports Management, Compute, and Storage intent traffic.
- p-NIC: Physical Network Interface Card attached to an Azure Local node.
- v-Switch: Virtual Switch configured on the Azure Local Cluster.
- VLAN: Virtual Local Area Network.
- SET: Switch Embedded Teaming, supporting switch-independent teaming (Windows Server feature).
- MLAG: Multi-Chassis Link Aggregation, a technique that lets two or more network switches work together as if they were one logical switch. (Cisco Nexus uses vPC for MLAG.)
- Border Router: Uplink device with the ToR switches, providing routing to endpoints external to the Azure Local environment.
- AS: Autonomous System number used to define a BGP neighbor.
- WRED: Weighted Random Early Detection, a congestion avoidance mechanism used in QoS policies.
- ECN: Explicit Congestion Notification, a congestion notification mechanism used to mark packets when congestion is encountered in the communication path. A DSCP bit is modified in the packet to identify congestion.
- Make: Cisco
- Model: Nexus 93180YC-FX
- Firmware: 10.3 or greater recommended
Before implementing this configuration, ensure the following requirements are met:
-
Azure Local cluster with 2-16 nodes
-
Each node equipped with:
- 2x 25Gbps NICs for compute/management (p-NIC A/B)
- 2x RDMA-capable NICs for storage (p-NIC C/D supporting RoCEv2 or iWARP)
-
Switch Embedded Teaming (SET) configured on compute/management interfaces
-
Storage Spaces Direct configured for storage workloads
- 2x Cisco Nexus 93180YC-FX switches (or equivalent meeting Azure Local physical network requirements)
- NX-OS firmware version 10.3 or later
- Sufficient port capacity for node connections and inter-switch links
- Layer 3 connectivity to border routers/upstream infrastructure
Switch Requirements: Ensure your switch hardware meets the specifications outlined in the Azure Local Physical Network Requirements documentation, including:
- Port density and bandwidth requirements
- Buffer sizing for lossless RDMA transport
- QoS and Priority Flow Control (PFC) capabilities
- LLDP and DCBX support for automatic configuration discovery
- Determine RDMA protocol: RoCEv2 or iWARP
- Configure appropriate MTU settings based on RDMA protocol choice
- Ensure PFC and QoS policies align with RDMA requirements
This document discusses a 2-16 node environment where the Management and Compute network intents share a SET team interface. Storage1 and Storage2 network intents utilize isolated network interfaces and connect to separate ToR switches to support storage traffic isolation.
Within this environment, there are two ToR devices (TOR1 and TOR2). The devices are configured with HSRP for gateway redundancy and iBGP for routing, but do NOT use vPC (MLAG) for the storage network interfaces to maintain strict traffic isolation required for RDMA operations.
Key Architecture Points:
- Management and Compute intents: Utilize redundancy across both ToR switches
- Storage intents: Isolated to individual ToR switches (no inter-switch communication)
- Gateway redundancy: Achieved through HSRP configuration
- Routing: iBGP between ToR switches, eBGP to external networks
Each node is equipped with two physical network interface cards, each with two physical interfaces (p-NICs):
- 🟦 p-NICs A and B handle both compute and management intent traffic.
- 🟦 p-NIC interfaces support 25Gbps bandwidth.
- 🟦 p-NICs A and B are configured as part of a Switch Embedded Teaming (SET) team, transmitting compute and management traffic. These NICs are assigned to a virtual switch (v-Switch) to support multiple network intents. Management intent traffic is untagged, and Compute traffic is tagged.
- 🟪 p-NICs C and D are dedicated to storage intent traffic and are RDMA-capable devices.
- 🟪 p-NICs C and D can support RoCEv2 or iWARP.
- 🟪 p-NICs C and D are connected to the ToR devices. VLAN 711 is assigned to p-NIC C and VLAN 712 to p-NIC D. The interfaces operate in trunk mode, and only one storage intent VLAN is assigned per interface.
Important
If your Azure Local environment uses iWARP-based network cards and you enable Jumbo Frames, ensure that Jumbo Frames are also enabled on all switch interfaces and network paths carrying Storage Intent traffic. For RoCEv2-based systems, enabling Jumbo Frames on the switch is not required for Storage Intent traffic.
- TOR1 is in an MLAG (vPC) configuration with TOR2. Three interfaces are assigned to the MLAG peer link between the devices. Peer link bandwidth should be sized based on customer workloads.
- MLAG supports two network intents: Management and Compute.
- Layer 2 traffic is supported between the cluster nodes and the ToR devices. External connectivity is only established via Layer 3 sessions.
- Storage intent traffic is isolated to a specific ToR and is not expected to traverse between the ToR devices.
- TOR1 and TOR2 operate as Layer 2/Layer 3 devices. Layer 3 is supported by BGP as the primary routing protocol.
- TOR1 and TOR2 are configured to support an iBGP session; all external links are configured as eBGP sessions.
The cable map below shows two nodes as an example. For larger environments, extend the pattern accordingly. Maintaining a consistent wiring pattern is essential for successful Azure Local deployments.
Note
Azure Local deployments require consistent wiring patterns for proper operation:
- All p-NIC A interfaces should connect to TOR1
- All p-NIC B interfaces should connect to TOR2
- All p-NIC C interfaces should connect to TOR1
- All p-NIC D interfaces should connect to TOR2
Consistent cabling ensures proper traffic flow, simplifies troubleshooting, and helps avoid deployment challenges. We recommend verifying your wiring pattern matches this guidance before beginning deployment.
| Device | Interface | Device | Interface | Device | Interface | Device | Interface | |||
|---|---|---|---|---|---|---|---|---|---|---|
| Node1 | p-NIC A | <==> | TOR1 | Ethernet1/1 | Node2 | p-NIC A | <==> | TOR1 | Ethernet1/2 | |
| Node1 | p-NIC B | <==> | TOR2 | Ethernet1/1 | Node2 | p-NIC B | <==> | TOR2 | Ethernet1/2 | |
| Node1 | p-NIC C | <==> | TOR1 | Ethernet1/15 | Node2 | p-NIC C | <==> | TOR1 | Ethernet1/16 | |
| Node1 | p-NIC D | <==> | TOR2 | Ethernet1/15 | Node2 | p-NIC D | <==> | TOR2 | Ethernet1/16 |
| Device | Interface | Device | Interface | Device | Interface | Device | Interface | |||
|---|---|---|---|---|---|---|---|---|---|---|
| TOR1 | Ethernet1/1 | <==> | Node1 | p-NIC A | TOR2 | Ethernet1/1 | <==> | Node1 | p-NIC B | |
| TOR1 | Ethernet1/2 | <==> | Node2 | p-NIC A | TOR2 | Ethernet1/2 | <==> | Node2 | p-NIC B | |
| TOR1 | Ethernet1/15 | <==> | Node1 | p-NIC C | TOR2 | Ethernet1/15 | <==> | Node1 | p-NIC D | |
| TOR1 | Ethernet1/16 | <==> | Node2 | p-NIC C | TOR2 | Ethernet1/16 | <==> | Node2 | p-NIC D | |
| TOR1 | Ethernet1/41 | <==> | TOR2 | Ethernet1/41 | TOR2 | Ethernet1/41 | <==> | TOR1 | Ethernet1/41 | |
| TOR1 | Ethernet1/42 | <==> | TOR2 | Ethernet1/42 | TOR2 | Ethernet1/42 | <==> | TOR1 | Ethernet1/42 | |
| TOR1 | Ethernet1/47 | <==> | Border1 | Ethernet1/x | TOR2 | Ethernet1/47 | <==> | Border1 | Ethernet1/x | |
| TOR1 | Ethernet1/48 | <==> | Border2 | Ethernet1/x | TOR2 | Ethernet1/48 | <==> | Border2 | Ethernet1/x | |
| TOR1 | Ethernet1/49 | <==> | TOR2 | Ethernet1/49 | TOR2 | Ethernet1/49 | <==> | TOR1 | Ethernet1/49 | |
| TOR1 | Ethernet1/50 | <==> | TOR2 | Ethernet1/50 | TOR2 | Ethernet1/50 | <==> | TOR1 | Ethernet1/50 | |
| TOR1 | Ethernet1/51 | <==> | TOR2 | Ethernet1/51 | TOR2 | Ethernet1/51 | <==> | TOR1 | Ethernet1/51 |
This section provides a high-level overview of the switch configuration required for Azure Local environments. The configuration focuses on enabling features and policies that support Management, Compute, and Storage network intents, as well as redundancy and high availability.
Quality of Service (QoS) policies are implemented to prioritize critical traffic types, such as storage (RDMA) and cluster heartbeat, and to ensure efficient bandwidth allocation for all network intents. The following sections detail the required class maps, policy maps, and system QoS application.
- Compute and Management: These intents share a SET team interface and are typically carried over tagged and untagged VLANs, respectively. Ensure proper VLAN tagging and trunk configuration on switch ports.
- Storage: Storage intent traffic uses dedicated RDMA-capable interfaces and is isolated to specific VLANs. Lossless transport is required for RDMA (RoCEv2 or iWARP), which is achieved through Priority Flow Control (PFC) and appropriate buffer allocation.
Enable the following features on Cisco Nexus 93180YC-FX to support this Azure Local environment:
feature bgp
feature interface-vlan
feature hsrp
feature lacp
feature lldp- BGP is used as the primary routing protocol.
- interface-vlan enables SVI creation for management and compute networks.
- LACP is used for port-channels between ToR devices.
- HSRP provides gateway redundancy for SVIs.
- LLDP is used to transmit host interface configuration values, which can be leveraged by Azure Local for enhanced support scenarios.
Quality of Service (QoS) Policy
The spanning tree configuration implements Multiple Spanning Tree (MST) protocol to prevent network loops while allowing redundant paths. The spanning-tree port type edge bpduguard default command configures all ports as edge ports by default (typically used for end devices) and enables BPDU guard protection, which will disable ports that unexpectedly receive spanning tree BPDUs. The priority settings establish a hierarchy where MST instances 0 and 1 have higher priority (lower numerical value of 20480) compared to MST instance 2 (28672), making instances 0 and 1 more likely to become the root bridge.
The MST configuration section creates a spanning tree region named "AzureLocal" with revision 1, which helps identify switches that should share the same MST configuration. The VLAN-to-instance mappings distribute different VLANs across multiple spanning tree instances: instance 1 handles VLANs 1-710 and 713-1999, instance 2 manages VLANs 2000-4094, and instance 3 is dedicated to VLANs 711-712. This segmentation allows for better load balancing and convergence times by having different spanning tree calculations for different groups of VLANs.
spanning-tree port type edge bpduguard default
spanning-tree mst 0-1 priority 20480
spanning-tree mst 2 priority 28672
spanning-tree mst configuration
name AzureLocal
revision 1
instance 1 vlan 1-710,713-1999
instance 2 vlan 2000-4094
instance 3 vlan 711-712This section configures the necessary VLANs to support Azure Local network intents and establishes Switched Virtual Interfaces (SVIs) for Layer 3 connectivity. The VLAN structure is designed to provide network segmentation, security, and optimal traffic flow for different Azure Local workloads.
vlan 1-2,7-8,99,711
vlan 2
name UNUSED_INTERFACE
vlan 7
name Management_7
vlan 8
name Compute_8
vlan 99
name NativeVlan
vlan 711
name Storage_711_TOR1
interface Vlan7
description Management_7
no shutdown
mtu 9216
no ip redirects
ip address 10.101.176.2/26
no ipv6 redirects
hsrp version 2
hsrp 7
priority 150 forwarding-threshold lower 1 upper 150
ip 10.101.176.1
interface Vlan8
description Compute_8
no shutdown
mtu 9216
no ip redirects
ip address 10.101.177.2/26
no ipv6 redirects
hsrp version 2
hsrp 8
priority 150 forwarding-threshold lower 1 upper 150
ip 10.101.177.1Management VLAN (VLAN 7): Configured as an SVI with IP address 10.101.176.2/26, this VLAN provides out-of-band management access to Azure Local nodes. This network segment carries Windows Admin Center traffic, PowerShell remoting, and other administrative communications. The SVI enables the ToR switch to act as the default gateway for management traffic, providing Layer 3 routing to external management systems and administrative networks. HSRP is configured with virtual IP 10.101.176.1 and priority 150 to provide gateway redundancy with TOR2.
Compute VLAN (VLAN 8): Configured as an SVI with IP address 10.101.177.2/26, this VLAN handles Azure Local virtual machine traffic and compute workloads. The SVI provides gateway services for tenant VMs and enables north-south traffic flow between Azure Local workloads and external networks. This configuration supports SDN virtual networks and provides connectivity for Azure Arc-enabled services. HSRP is configured with virtual IP 10.101.177.1 and priority 150 to ensure high availability gateway services across both ToR switches.
Storage VLAN (VLAN 711): This VLAN operates exclusively at Layer 2 and intentionally has no SVI configuration. Storage traffic uses RDMA protocols (RoCEv2 or iWARP) that bypass the traditional TCP/IP stack, eliminating the need for Layer 3 routing. VLAN 711 is isolated to TOR1 only and carries storage traffic from Azure Local nodes connected to TOR1 (p-NIC C interfaces). This ensures optimal performance and maintains the lossless characteristics required for Storage Spaces Direct operations.
Native VLAN (VLAN 99): Serves as a security boundary for untagged traffic. Any misconfigured or untagged frames are automatically assigned to this VLAN, preventing unauthorized access to production networks. This follows Cisco security best practices and provides operational visibility into potential configuration issues.
Unused Interface VLAN (VLAN 2): Acts as a parking VLAN for unused switch ports. Assign unused interfaces to this VLAN to prevent unauthorized device connections and maintain network security posture.
-
MTU 9216: Jumbo frames are configured on management and compute SVIs to support Software Defined Networking (SDN) services in Azure Local environments. This MTU setting is specifically required for SDN virtual networks, Network Controller communications, and gateway services. If your Azure Local deployment does not utilize SDN features, standard MTU (1500) can be used instead.
-
IP Redirects Disabled: Both IPv4 and IPv6 redirects are disabled on SVIs as a best practice for Azure Local deployments. IP redirects can conflict with various Azure Local services including Network Controller, SDN gateway operations, and cluster networking components. Disabling redirects ensures proper traffic flow and prevents routing conflicts within the Azure Local environment.
-
Subnet Sizing: /26 subnets provide 62 usable IP addresses per network intent, sufficient for typical Azure Local cluster deployments while maintaining efficient IP space utilization. See the Azure Local Network Patterns guide to determine the correct subnet size.
Note
TOR2 Storage VLAN Configuration: TOR2 requires a separate storage VLAN 712 configuration:
vlan 712
name Storage_712_TOR2VLAN 712 is isolated to TOR2 and carries storage traffic from Azure Local nodes connected to TOR2 (p-NIC D interfaces). Each ToR switch maintains its own dedicated storage VLAN to ensure traffic isolation and optimal RDMA performance. Storage traffic never traverses between ToR switches, even in the vPC configuration.
Important
Ensure HSRP is configured on VLANs 7 and 8 to provide gateway redundancy between TOR1 and TOR2. This configuration will be detailed in the HSRP section of the interface configuration.
These interfaces connect to Azure Local nodes' p-NIC A and p-NIC B interfaces, which are configured as Switch Embedded Teaming (SET) members. This configuration supports both management and compute network intents through VLAN separation while providing redundancy across both ToR switches.
interface Ethernet1/1
description Switched-Compute-Management
no cdp enable
switchport
switchport mode trunk
switchport trunk native vlan 7
switchport trunk allowed vlan 8
spanning-tree port type edge trunk
mtu 9216
logging event port link-status
no shutdown
interface Ethernet1/2
description Switched-Compute-Management
no cdp enable
switchport
switchport mode trunk
switchport trunk native vlan 7
switchport trunk allowed vlan 8
spanning-tree port type edge trunk
mtu 9216
logging event port link-status
no shutdownVLAN Configuration:
-
VLAN 7 (Management Intent): Configured as the native VLAN to carry untagged management traffic from Azure Local nodes. This includes Windows Admin Center communications, PowerShell remoting, cluster management traffic, and health monitoring services.
-
VLAN 8 (Compute Intent): Configured as a tagged VLAN to carry compute workload traffic, including virtual machine communications, tenant network traffic, and Azure Arc-enabled services.
Network Redundancy: Both VLANs 7 and 8 are supported on both TOR1 and TOR2 switches, providing high availability through the vPC (MLAG) configuration. The HSRP configuration on the corresponding SVIs (interface Vlan7 and Vlan8) utilizes the peer link between ToR devices to maintain gateway redundancy and ensure seamless failover for Azure Local nodes.
MTU Configuration: The MTU 9216 setting is optional and specifically configured to support Software Defined Networking (SDN) environments. This jumbo frame configuration is required for:
- Network Controller communications
- SDN gateway operations
- Virtual network overlay protocols
If your Azure Local deployment does not utilize SDN features, the MTU can be set to the standard 1500 bytes.
Interface Optimization:
- CDP Disabled: Cisco Discovery Protocol is disabled as it's not required for Azure Local environments and reduces unnecessary protocol overhead
- Spanning Tree Edge: Configured as edge trunk ports since these connect directly to end devices (Azure Local nodes), enabling faster convergence
- Link Status Logging: Enabled to provide visibility into interface state changes for troubleshooting and monitoring
Note
These interfaces support the SET team configuration on Azure Local nodes, where p-NIC A connects to TOR1 and p-NIC B connects to TOR2. This provides active-active connectivity with automatic failover capabilities, ensuring high availability for both management and compute traffic.
These interfaces connect to Azure Local nodes' dedicated storage NICs (p-NIC C interfaces) and are specifically configured to support RDMA traffic for Storage Spaces Direct operations. The configuration ensures lossless transport and optimal performance for storage workloads.
interface Ethernet1/15
description Switched-Storage-Node1-pNIC-C
no cdp enable
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 711
priority-flow-control mode on send-tlv
spanning-tree port type edge trunk
mtu 9216
logging event port link-status
service-policy type qos input AZS_SERVICES
no shutdown
interface Ethernet1/16
description Switched-Storage-Node2-pNIC-C
no cdp enable
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 711
priority-flow-control mode on send-tlv
spanning-tree port type edge trunk
mtu 9216
logging event port link-status
service-policy type qos input AZS_SERVICES
no shutdownPhysical Connectivity:
- Ethernet1/15: Connects to Node1 p-NIC C (dedicated storage interface)
- Ethernet1/16: Connects to Node2 p-NIC C (dedicated storage interface)
- Pattern: All p-NIC C interfaces from every Azure Local node connect to TOR1 for storage traffic isolation
VLAN Configuration:
- VLAN 99 (Native): Configured as the default native VLAN to capture any untagged traffic. This serves as a security boundary and operational best practice, ensuring that misconfigured or untagged frames are isolated from production storage traffic.
- VLAN 711 (Storage): The dedicated storage intent VLAN that carries all RDMA traffic for Storage Spaces Direct operations. Each storage interface is assigned to only one storage VLAN (711 for TOR1, 712 for TOR2) to maintain traffic isolation and prevent cross-switch storage communication.
LLDP and DCBX Integration:
- priority-flow-control mode on send-tlv: Enables the switch to advertise its Priority Flow Control (PFC) capabilities and Data Center Bridging Exchange (DCBX) configuration to connected Azure Local nodes via LLDP.
The switch transmits QoS policy information including:
- Traffic class mappings
- PFC enabled priorities
- Buffer allocation settings
- Congestion management parameters
Azure Local nodes operate in "non-willing mode" for DCBX, meaning they will not accept configuration changes from the switch. Instead, this information is used for:
- Validation of switch/host configuration alignment
- Troubleshooting QoS and RDMA connectivity issues
- Support diagnostic data collection
MTU Configuration:
- MTU 9216: Configures jumbo frame support, but actual utilization depends on the RDMA transport protocol:
- iWARP-based RDMA: Utilizes the configured jumbo frames (9216 bytes) for improved performance and reduced CPU overhead
- RoCEv2-based RDMA: Operates with standard Ethernet frames (1500 bytes) and ignores the jumbo frame configuration, as RoCEv2 handles frame segmentation differently
Quality of Service:
- service-policy type qos input AZS_SERVICES: Applies the QoS policy that enables traffic classification, bandwidth guarantees, and congestion management for storage and cluster heartbeat traffic. This policy ensures that storage RDMA traffic receives appropriate priority and lossless transport characteristics.
Lossless Transport:
- Priority Flow Control (PFC): Combined with the QoS policy, PFC ensures lossless delivery of storage traffic by providing per-priority pause capabilities, preventing packet drops that would severely impact RDMA performance.
Note
TOR2 Storage Configuration: TOR2 requires corresponding storage interface and VLAN configuration:
vlan 712
name Storage_712_TOR2
interface Ethernet1/15
description Switched-Storage-Node1-pNIC-D
no cdp enable
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 712
priority-flow-control mode on send-tlv
spanning-tree port type edge trunk
mtu 9216
logging event port link-status
service-policy type qos input AZS_SERVICES
no shutdown
interface Ethernet1/16
description Switched-Storage-Node2-pNIC-D
no cdp enable
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 712
priority-flow-control mode on send-tlv
spanning-tree port type edge trunk
mtu 9216
logging event port link-status
service-policy type qos input AZS_SERVICES
no shutdownImportant
RDMA Protocol Considerations: Verify your Azure Local cluster's RDMA configuration (iWARP vs. RoCEv2) to ensure optimal switch configuration. While jumbo frames benefit iWARP deployments, they are not utilized by RoCEv2. Properly configure the MTU and QoS settings based on the deployed RDMA protocol to achieve the best performance and reliability for storage workloads.
The iBGP peer link provides dedicated connectivity between TOR1 and TOR2 for BGP route exchange and high-availability communications. This link operates independently of the vPC peer link and ensures reliable routing convergence and failover capabilities.
interface port-channel50
description iBGP_PEER_LINK
logging event port link-status
mtu 9216
ip address 10.71.55.25/30
interface Ethernet1/41
description iBGP_PEER_LINK
mtu 9216
logging event port link-status
channel-group 50 mode active
no shutdown
interface Ethernet1/42
description iBGP_PEER_LINK
mtu 9216
logging event port link-status
channel-group 50 mode active
no shutdownConfiguration Details
Physical Connectivity:
- Ethernet1/41: First physical interface in the iBGP peer link bundle
- Ethernet1/42: Second physical interface providing redundancy and increased bandwidth
- Port-Channel 50: LACP-based aggregation of the two physical interfaces for high availability
Link Aggregation:
- LACP Active Mode: Both interfaces are configured with
channel-group 50 mode active, ensuring active LACP negotiation with the peer switch. This provides automatic failover if one physical link fails and enables load balancing across both interfaces.
IP Addressing:
- IP Address 10.71.55.25/30: Point-to-point /30 subnet providing efficient IP utilization. TOR1 uses .25/30, while TOR2 would be configured with .26/30. This dedicated subnet isolates BGP traffic from other network segments and provides a reliable communication path for routing protocol exchanges.
MTU Configuration:
- MTU 9216: Jumbo frame configuration supports efficient routing table exchanges and reduces fragmentation for large BGP updates. This is particularly beneficial in environments with extensive route advertisements or frequent topology changes.
Interface Optimization:
- Link Status Logging: Enabled on all interfaces to provide visibility into link state changes, critical for troubleshooting routing convergence issues
BGP Integration:
This port-channel serves as the transport for iBGP sessions between TOR1 and TOR2, enabling:
- Route Synchronization: Ensures both switches maintain consistent routing tables for Azure Local network segments
- Failover Coordination: Provides reliable communication for BGP convergence during link or device failures
- Load Balancing: Supports ECMP (Equal Cost Multi-Path) routing decisions across both ToR switches
Note
TOR2 Configuration: The corresponding configuration on TOR2 would be identical except for the IP address:
interface port-channel50
ip address 10.71.55.26/30This creates the point-to-point link between the two switches for iBGP communication.
Important
Separation from vPC Peer Link: This iBGP peer link operates independently of the HSRP peer link (port-channel 101) to ensure routing protocol stability.
interface port-channel101
description HSRP_PEER
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 7-8
spanning-tree port type network
logging event port link-status
interface Ethernet1/49
description HSRP_PEER
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 7-8
logging event port link-status
channel-group 101 mode active
no shutdown
interface Ethernet1/50
description HSRP_PEER
switchport
switchport mode trunk
switchport trunk native vlan 99
switchport trunk allowed vlan 7-8
logging event port link-status
channel-group 101 mode active
no shutdownThe HSRP peer link provides Layer 2 connectivity between TOR1 and TOR2, enabling HSRP synchronization for management and compute network intents. This link ensures reliable gateway redundancy and enables seamless failover capabilities for Azure Local cluster connectivity.
HSRP Integration:
This peer link enables critical HSRP functionality between TOR1 and TOR2:
- Gateway Redundancy: Supports HSRP hello message exchange between the switches for VLANs 7 and 8, ensuring seamless gateway failover for Azure Local nodes.
- State Synchronization: Enables real-time synchronization of HSRP active/standby states and priority changes between the ToR switches.
- Preemption Control: Facilitates coordinated failover using HSRP forwarding thresholds configured on the SVIs to prevent unnecessary state changes.
Azure Local SDN and Gateway configuration
After deploying this configuration, validate the following:
show interface brief
show interface trunk
show spanning-tree briefshow vlan brief
show interface vlan brief
show hsrp briefshow ip bgp summary
show ip route bgp
show hsrp briefshow policy-map interface ethernet1/15
show interface ethernet1/15 priority-flow-control- Verify SET team formation on cluster nodes
- Confirm RDMA connectivity using Azure Local validation tools
- Test storage performance and verify lossless transport
- Physical network requirements for Azure Local
- Teaming in Azure Stack HCI
- Network considerations for cloud deployments of Azure Local
- Manage Azure Local gateway connections
- Microsoft Azure Local Connectivity to Cisco Nexus 9000 Series Switches in Cisco NX-OS and Cisco® Application Centric Infrastructure (Cisco ACI™) Mode
- RoCE Storage Implementation over NX-OS VXLAN Fabrics
- Cisco Nexus 9000 Series NX-OS Quality of Service Configuration Guide, Release 10.5(x)
- Cisco Nexus Configure Queuing and Scheduling
- Cisco WRED-Explicit Congestion Notification
- RFC 3168 - The Addition of Explicit Congestion Notification (ECN) to IP
- Azure Local network deployment patterns
