This document provides a comprehensive reference for implementing a fully converged physical network design for Azure Local deployments. It includes configuration examples, validation procedures, and best practices for customers and partners deploying Azure Local in production environments.
- Azure Local - Fully Converged Physical Network Design Reference
Azure Local's fully converged network design provides a unified approach to handling management, compute, and storage traffic over the same physical infrastructure. This design pattern offers:
- Simplified Infrastructure: Reduced hardware footprint and cabling complexity
- Cost Optimization: Efficient use of network resources and equipment
- Scalability: Support for 2 to 16 nodes with consistent configuration patterns
The fully converged physical network architecture integrates management, compute, and storage traffic over the same physical Ethernet interfaces. This design minimizes hardware footprint while maximizing scalability and deployment simplicity.
Key Design Principle: The fully converged storage deployment follows the same pattern as switched deployment for consistency: One Storage VLAN per TOR is the baseline configuration. This approach ensures storage traffic always uses RDMA instead of falling back to regular TCP traffic, providing reliable high-performance storage connectivity.
This section describes the essential components required for implementing a fully converged Azure Local network design.
Physical switches that provide redundant Layer 2/Layer 3 connectivity for the Azure Local cluster. Key characteristics:
- Each Azure Local node connects to two separate ToR switches for high availability
- All switch ports must be configured as IEEE 802.1Q trunk ports to support multiple VLANs
- Supports standard data center switch features including MLAG, BGP routing, and QoS policies
Physical hosts running the Azure Local operating system. In a fully converged design:
- Each node requires minimum two high-speed physical NICs (10Gbps or higher)
- NICs must support RDMA (Remote Direct Memory Access) for optimal storage performance
- All traffic types (management, compute, and storage) flow over the same physical interfaces using VLAN segmentation
Azure Local's intent-based networking framework that simplifies network deployment:
- Defines and deploys logical networking configurations through "intents"
- In fully converged deployments, uses a single
Management + Compute + Storageintent - Automatically configures VLANs, IP addressing, and adapter settings across the cluster
Windows-native NIC teaming technology that provides link-level redundancy:
- Creates a single logical interface from multiple physical NICs
- Operates in switch-independent mode for maximum compatibility
- Only supported virtual switch technology on Azure Local
- Provides load balancing and failover capabilities
This section demonstrates a fully converged Azure Local deployment where management, compute, and storage traffic all share the same NICs using VLAN segmentation. The design scales seamlessly from 2 to 16 nodes with consistent configuration patterns.
- Fully Converged: All traffic types (Management, Compute, Storage) utilize the same physical links
- Redundant Infrastructure: Each node connects to both TOR1 and TOR2 for high availability
- Switch Embedded Teaming: Host-level NIC bonding provides fault tolerance and load balancing
- VLAN Segmentation: Traffic isolation using IEEE 802.1Q VLAN tagging
The following tables demonstrate physical connectivity between Azure Local nodes and Top-of-Rack switches in a three-node cluster. This pattern scales consistently for larger deployments.
Cabling Requirements:
- Each node requires two network connections (NIC A and NIC B)
- Each NIC connects to a different ToR switch to ensure redundancy
- Use identical cable types and speeds for consistent performance
| Azure Local Node | Interface | ToR Switch | Interface |
|---|---|---|---|
| Host1 | NIC A | TOR1 | Ethernet1/1 |
| Host1 | NIC B | TOR2 | Ethernet1/1 |
| Azure Local Node | Interface | ToR Switch | Interface |
|---|---|---|---|
| Host2 | NIC A | TOR1 | Ethernet1/2 |
| Host2 | NIC B | TOR2 | Ethernet1/2 |
| Azure Local Node | Interface | ToR Switch | Interface |
|---|---|---|---|
| Host3 | NIC A | TOR1 | Ethernet1/3 |
| Host3 | NIC B | TOR2 | Ethernet1/3 |
The fully converged design uses VLAN segmentation to isolate different traffic types while sharing the same physical infrastructure.
| Traffic Type | Purpose | VLAN ID | Configuration Notes |
|---|---|---|---|
| Management | Cluster and host management traffic | 7 | Native VLAN, L3 routed (SVI) |
| Compute | Virtual machine workload traffic | 201 | Tagged VLAN, L3 routed (SVI) |
| Storage 1 | SMB1 over RDMA | 711 | Tagged VLAN, L2 only (no SVI) |
| Storage 2 | SMB2 over RDMA | 712 | Tagged VLAN, L2 only (no SVI) |
Important
Storage VLAN Design Pattern: The fully converged deployment follows the same pattern as switched deployment: One Storage VLAN per TOR is the baseline configuration. This design ensures storage traffic always uses RDMA instead of falling back to regular TCP traffic.
- Baseline Approach: TOR1 carries Storage VLAN 711, TOR2 carries Storage VLAN 712
- Optional Enhancement: Both storage VLANs can be configured on both switches for additional resilience (not mandatory)
- Primary Goal: Maintain design consistency and guarantee RDMA connectivity
This design utilizes two physical switches as Top-of-Rack (ToR) devices to provide redundant uplink connectivity for all Azure Local nodes.
Switch Requirements:
- Support for IEEE 802.1Q VLAN tagging and trunking
- Layer 3 routing capabilities (SVI interfaces)
- RDMA/RoCE support with Data Center Bridging (DCB)
- Quality of Service (QoS) policy enforcement
- Optional: Multi-Chassis Link Aggregation (MLAG) support
Redundancy Design:
- ToR switches may be configured with MLAG (Multi-Chassis Link Aggregation) for enhanced redundancy
- MLAG configuration is optional, as redundancy is primarily achieved through host-side SET (Switch Embedded Teaming)
- Core network layer connectivity (routers, firewalls) is considered out of scope for this document
This section provides configuration guidance using Cisco Nexus 93180YC-FX3 (NX-OS 10.x) as the reference platform. The principles apply to other vendor switches with appropriate syntax modifications.
VLAN Configuration Summary:
- VLAN 7 (Management): Layer 3 routed VLAN, configured as native VLAN on trunk ports
- VLAN 201 (Compute): Layer 3 routed VLAN, tagged on trunk ports
- VLAN 711 (Storage - TOR1): Layer 2 only VLAN (no SVI), tagged on TOR1 trunk ports for RDMA traffic
- VLAN 712 (Storage - TOR2): Layer 2 only VLAN (no SVI), tagged on TOR2 trunk ports for RDMA traffic
Key Configuration Requirements:
- Enable Priority Flow Control (PFC) for RDMA support
- Configure jumbo frames (MTU 9216) for optimal performance
- Apply QoS service policies for traffic prioritization
- Set appropriate spanning-tree settings for edge ports
Baseline Configuration - TOR1 (One Storage VLAN per TOR):
vlan 7
name Management_7
vlan 201
name Compute_201
vlan 711
name Storage_711
interface Vlan7
description Management
no shutdown
mtu 9216
ip address 10.101.176.2/24
hsrp 7
ip 100.101.176.1
interface Vlan201
description Compute
no shutdown
mtu 9216
ip address 10.101.177.2/24
hsrp 201
ip 100.101.177.1
interface Ethernet1/1-3
description To_Azure_Local_Host_FullyConverged
switchport
switchport mode trunk
switchport trunk native vlan 7
switchport trunk allowed vlan 7,201,711
priority-flow-control mode on send-tlv
spanning-tree port type edge trunk
mtu 9216
service-policy type qos input AZS_SERVICES
no shutdownBaseline Configuration - TOR2 (One Storage VLAN per TOR):
vlan 7
name Management_7
vlan 201
name Compute_201
vlan 712
name Storage_712
interface Vlan7
description Management
no shutdown
mtu 9216
ip address 10.101.176.3/24
hsrp 7
ip 100.101.176.1
interface Vlan201
description Compute
no shutdown
mtu 9216
ip address 10.101.177.3/24
hsrp 201
ip 100.101.177.1
interface Ethernet1/1-3
description To_Azure_Local_Host_FullyConverged
switchport
switchport mode trunk
switchport trunk native vlan 7
switchport trunk allowed vlan 7,201,712
priority-flow-control mode on send-tlv
spanning-tree port type edge trunk
mtu 9216
service-policy type qos input AZS_SERVICES
no shutdownNote
- The above shows the baseline "One Storage VLAN per TOR" configuration
- QoS policies and routing design (e.g., uplinks, BGP/OSPF, default gateway) will be introduced in a separate document
This section provides comprehensive validation procedures to verify the network configuration is functioning correctly in already deployed Azure Local environments. These validation commands are intended for troubleshooting and verifying existing production deployments, not for fresh installations.
Use these PowerShell commands on deployed Azure Local nodes to validate network adapter configuration and RDMA connectivity:
1. Verify Network Adapter Configuration
# Display network adapters and VLAN assignments
Get-NetAdapter | Format-Table InterfaceAlias, VlanID, MacAddress -AutoSize
# Check VLAN isolation settings for virtual adapters
Get-VMNetworkAdapterIsolation -ManagementOS | Format-Table ParentAdapter, IsolationMode, DefaultIsolationID -AutoSize2. Validate SMB Multichannel and RDMA Connectivity
[Host3]: PS C:\Users\Administrator\Documents> Get-NetAdapter | ft InterfaceAlias, VlanID, MacAddress
InterfaceAlias VlanID MacAddress
-------------- ------ ----------
ethernet 0 0C-42-A1-F9-69-4A
vSMB(managementcompute#ethernet 2) 00-15-5D-C8-20-07
ethernet 2 0 0C-42-A1-F9-69-4B
vManagement(managementcompute) 0C-42-A1-F9-69-4A
vSMB(managementcompute#ethernet) 00-15-5D-C8-20-06
[Host3]: PS C:\Users\Administrator\Documents> Get-VMNetworkAdapterIsolation -ManagementOS | ft ParentAdapter, IsolationMode, DefaultIsolationID
ParentAdapter IsolationMode DefaultIsolationID
------------- ------------- ------------------
VMInternalNetworkAdapter, Name = 'vSMB(managementcompute#ethernet)' Vlan 711
VMInternalNetworkAdapter, Name = 'vManagement(managementcompute)' Vlan 0
VMInternalNetworkAdapter, Name = 'vSMB(managementcompute#ethernet 2)' Vlan 712
# Verify SMB Multichannel connections and RDMA usage
[Host3]: PS C:\Users\Administrator\Documents> Get-SmbMultichannelConnection | Format-Table ClientInterfaceFriendlyName, ClientIpAddress, ServerIpAddress, ClientRdmaCapable, RdmaConnectionCount, TcpConnectionCount -AutoSize
ClientInterfaceFriendlyName ClientIpAddress ServerIpAddress ClientRdmaCapable RdmaConnectionCount TcpConnectionCount
--------------------------- --------------- --------------- ----------------- ------------------- ------------------
vSMB(managementcompute#ethernet 2) 10.71.2.10 10.71.2.34 True 2 0
vSMB(managementcompute#ethernet) 10.71.1.249 10.71.1.214 True 2 0
# Verify SMB connections
[Host3]: PS C:\Users\Administrator\Documents> Get-SmbConnection
ServerName ShareName UserName Credential Dialect NumOpens
---------- --------- -------- ---------- ------- --------
Host4 c$ Administrator Administrator 3.1.1 2
Note
SMB Multichannel Validation Key Points:
- Both storage VLANs (711 and 712) are operational with RDMA enabled
RdmaConnectionCount = 2confirms RDMA is being used for storage trafficTcpConnectionCount = 0shows no fallback to regular TCP- SMB 3.1.1 dialect is being used for optimal performance
- This validates that storage traffic is using RDMA as intended
Verify Top-of-Rack switch configuration and MAC address learning to ensure proper connectivity.
Validation Commands for Cisco NX-OS:
# On TOR1 (Storage VLAN 711 only)
TOR1# show mac address-table interface ethernet 1/3
Legend:
* - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC
age - seconds since last seen,+ - primary entry using vPC Peer-Link,
(T) - True, (F) - False, C - ControlPlane MAC, ~ - vsan
VLAN MAC Address Type age Secure NTFY Ports
---------+-----------------+--------+---------+------+----+------------------
* 7 0c42.a1f9.694a dynamic 0 F F Eth1/3
* 711 0015.5dc8.2006 dynamic 0 F F Eth1/3
# On TOR2 (Storage VLAN 712 only)
TOR2# show mac address-table interface ethernet 1/3
Legend:
* - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC
age - seconds since last seen,+ - primary entry using vPC Peer-Link,
(T) - True, (F) - False, C - ControlPlane MAC, ~ - vsan
VLAN MAC Address Type age Secure NTFY Ports
---------+-----------------+--------+---------+------+----+------------------
* 7 0c42.a1f9.694b dynamic 0 F F Eth1/3
* 712 0015.5dc8.2007 dynamic 0 F F Eth1/3
Expected Results:
- TOR1 should show Management VLAN 7 and Storage VLAN 711 learned MAC addresses
- TOR2 should show Management VLAN 7 and Storage VLAN 712 learned MAC addresses
- This validates the baseline "One Storage VLAN per TOR" configuration ensuring RDMA traffic flow
Quality of Service configuration is essential for optimal Azure Local performance. For detailed QoS policy configuration:
Reference: Quality of Service (QoS) Policy
For BGP routing configuration and best practices in Azure Local deployments:
Reference: Azure Local BGP Routing
A: Yes, you can configure both Storage VLANs on both TOR switches, but you won't get too much benefit from it. Here's why:
- We always want storage to stay in RDMA - The baseline configuration already ensures RDMA connectivity is maintained
- We don't want traffic to cross the TOR peer link - Having both VLANs on both switches may introduce unnecessary inter-switch traffic
- It's a balance between resilience vs performance - The added complexity doesn't provide significant resilience improvements
Baseline Configuration (Recommended):
- TOR1: Configure only Storage VLAN 711 (plus Management and Compute VLANs)
- TOR2: Configure only Storage VLAN 712 (plus Management and Compute VLANs)
This approach ensures:
- Design consistency across deployment patterns
- Guaranteed RDMA usage - storage traffic cannot fall back to regular TCP
- Simplified configuration and troubleshooting
- Optimal performance - no unnecessary cross-TOR traffic
Optional Enhanced Configuration: While customers can configure both Storage VLANs (711 and 712) on both TOR switches, this configuration is not recommended as it provides minimal additional resilience. The baseline "One Storage VLAN per TOR" configuration already provides sufficient redundancy through the host-side SET (Switch Embedded Teaming) and ensures reliable RDMA connectivity without the potential performance implications.
Important
The primary goal is to maintain design consistency, force storage traffic to always use RDMA instead of allowing fallback to regular TCP traffic, and avoid unnecessary cross-TOR traffic that could impact performance.
For comprehensive Azure Local networking information, refer to these official Microsoft documentation resources:
-
Network considerations for cloud deployments of Azure Local
Planning guidance for Azure Local network deployments in cloud environments -
Physical network requirements for Azure Local
Detailed hardware and connectivity requirements for Azure Local deployments
- Teaming in Azure Stack HCI
Technical blog covering NIC teaming concepts and implementation
