Skip to content

Latest commit

 

History

History
412 lines (312 loc) · 19.5 KB

File metadata and controls

412 lines (312 loc) · 19.5 KB

Azure Local - Fully Converged Physical Network Design Reference

This document provides a comprehensive reference for implementing a fully converged physical network design for Azure Local deployments. It includes configuration examples, validation procedures, and best practices for customers and partners deploying Azure Local in production environments.

Executive Summary

Azure Local's fully converged network design provides a unified approach to handling management, compute, and storage traffic over the same physical infrastructure. This design pattern offers:

  • Simplified Infrastructure: Reduced hardware footprint and cabling complexity
  • Cost Optimization: Efficient use of network resources and equipment
  • Scalability: Support for 2 to 16 nodes with consistent configuration patterns

Design Overview

The fully converged physical network architecture integrates management, compute, and storage traffic over the same physical Ethernet interfaces. This design minimizes hardware footprint while maximizing scalability and deployment simplicity.

Key Design Principle: The fully converged storage deployment follows the same pattern as switched deployment for consistency: One Storage VLAN per TOR is the baseline configuration. This approach ensures storage traffic always uses RDMA instead of falling back to regular TCP traffic, providing reliable high-performance storage connectivity.

Architecture Components

This section describes the essential components required for implementing a fully converged Azure Local network design.

Top-of-Rack (ToR) Switches

Physical switches that provide redundant Layer 2/Layer 3 connectivity for the Azure Local cluster. Key characteristics:

  • Each Azure Local node connects to two separate ToR switches for high availability
  • All switch ports must be configured as IEEE 802.1Q trunk ports to support multiple VLANs
  • Supports standard data center switch features including MLAG, BGP routing, and QoS policies

Azure Local Nodes

Physical hosts running the Azure Local operating system. In a fully converged design:

  • Each node requires minimum two high-speed physical NICs (10Gbps or higher)
  • NICs must support RDMA (Remote Direct Memory Access) for optimal storage performance
  • All traffic types (management, compute, and storage) flow over the same physical interfaces using VLAN segmentation

Network ATC (Automatic Network Configuration)

Azure Local's intent-based networking framework that simplifies network deployment:

  • Defines and deploys logical networking configurations through "intents"
  • In fully converged deployments, uses a single Management + Compute + Storage intent
  • Automatically configures VLANs, IP addressing, and adapter settings across the cluster

Switch Embedded Teaming (SET)

Windows-native NIC teaming technology that provides link-level redundancy:

  • Creates a single logical interface from multiple physical NICs
  • Operates in switch-independent mode for maximum compatibility
  • Only supported virtual switch technology on Azure Local
  • Provides load balancing and failover capabilities

Physical Network Topology

This section demonstrates a fully converged Azure Local deployment where management, compute, and storage traffic all share the same NICs using VLAN segmentation. The design scales seamlessly from 2 to 16 nodes with consistent configuration patterns.

Design Characteristics

  • Fully Converged: All traffic types (Management, Compute, Storage) utilize the same physical links
  • Redundant Infrastructure: Each node connects to both TOR1 and TOR2 for high availability
  • Switch Embedded Teaming: Host-level NIC bonding provides fault tolerance and load balancing
  • VLAN Segmentation: Traffic isolation using IEEE 802.1Q VLAN tagging

Topology Diagram

AzureLocalPhysicalNetworkDiagram_FullyConverged

Cabling Configuration Example

The following tables demonstrate physical connectivity between Azure Local nodes and Top-of-Rack switches in a three-node cluster. This pattern scales consistently for larger deployments.

Cabling Requirements:

  • Each node requires two network connections (NIC A and NIC B)
  • Each NIC connects to a different ToR switch to ensure redundancy
  • Use identical cable types and speeds for consistent performance

Host 1

Azure Local Node Interface ToR Switch Interface
Host1 NIC A TOR1 Ethernet1/1
Host1 NIC B TOR2 Ethernet1/1

Host 2

Azure Local Node Interface ToR Switch Interface
Host2 NIC A TOR1 Ethernet1/2
Host2 NIC B TOR2 Ethernet1/2

Host 3

Azure Local Node Interface ToR Switch Interface
Host3 NIC A TOR1 Ethernet1/3
Host3 NIC B TOR2 Ethernet1/3

VLAN Architecture

The fully converged design uses VLAN segmentation to isolate different traffic types while sharing the same physical infrastructure.

Traffic Type Purpose VLAN ID Configuration Notes
Management Cluster and host management traffic 7 Native VLAN, L3 routed (SVI)
Compute Virtual machine workload traffic 201 Tagged VLAN, L3 routed (SVI)
Storage 1 SMB1 over RDMA 711 Tagged VLAN, L2 only (no SVI)
Storage 2 SMB2 over RDMA 712 Tagged VLAN, L2 only (no SVI)

Important

Storage VLAN Design Pattern: The fully converged deployment follows the same pattern as switched deployment: One Storage VLAN per TOR is the baseline configuration. This design ensures storage traffic always uses RDMA instead of falling back to regular TCP traffic.

  • Baseline Approach: TOR1 carries Storage VLAN 711, TOR2 carries Storage VLAN 712
  • Optional Enhancement: Both storage VLANs can be configured on both switches for additional resilience (not mandatory)
  • Primary Goal: Maintain design consistency and guarantee RDMA connectivity

Top-of-Rack Switch Configuration

This design utilizes two physical switches as Top-of-Rack (ToR) devices to provide redundant uplink connectivity for all Azure Local nodes.

Switch Requirements:

  • Support for IEEE 802.1Q VLAN tagging and trunking
  • Layer 3 routing capabilities (SVI interfaces)
  • RDMA/RoCE support with Data Center Bridging (DCB)
  • Quality of Service (QoS) policy enforcement
  • Optional: Multi-Chassis Link Aggregation (MLAG) support

Redundancy Design:

  • ToR switches may be configured with MLAG (Multi-Chassis Link Aggregation) for enhanced redundancy
  • MLAG configuration is optional, as redundancy is primarily achieved through host-side SET (Switch Embedded Teaming)
  • Core network layer connectivity (routers, firewalls) is considered out of scope for this document

Interface & VLAN Configuration

This section provides configuration guidance using Cisco Nexus 93180YC-FX3 (NX-OS 10.x) as the reference platform. The principles apply to other vendor switches with appropriate syntax modifications.

VLAN Configuration Summary:

  • VLAN 7 (Management): Layer 3 routed VLAN, configured as native VLAN on trunk ports
  • VLAN 201 (Compute): Layer 3 routed VLAN, tagged on trunk ports
  • VLAN 711 (Storage - TOR1): Layer 2 only VLAN (no SVI), tagged on TOR1 trunk ports for RDMA traffic
  • VLAN 712 (Storage - TOR2): Layer 2 only VLAN (no SVI), tagged on TOR2 trunk ports for RDMA traffic

Key Configuration Requirements:

  • Enable Priority Flow Control (PFC) for RDMA support
  • Configure jumbo frames (MTU 9216) for optimal performance
  • Apply QoS service policies for traffic prioritization
  • Set appropriate spanning-tree settings for edge ports
Sample NX-OS Configuration

Baseline Configuration - TOR1 (One Storage VLAN per TOR):

vlan 7
  name Management_7
vlan 201
  name Compute_201
vlan 711
  name Storage_711

interface Vlan7
  description Management
  no shutdown
  mtu 9216
  ip address 10.101.176.2/24
  hsrp 7
    ip 100.101.176.1

interface Vlan201
  description Compute
  no shutdown
  mtu 9216
  ip address 10.101.177.2/24
  hsrp 201
    ip 100.101.177.1

interface Ethernet1/1-3
  description To_Azure_Local_Host_FullyConverged
  switchport
  switchport mode trunk
  switchport trunk native vlan 7
  switchport trunk allowed vlan 7,201,711
  priority-flow-control mode on send-tlv
  spanning-tree port type edge trunk
  mtu 9216
  service-policy type qos input AZS_SERVICES
  no shutdown

Baseline Configuration - TOR2 (One Storage VLAN per TOR):

vlan 7
  name Management_7
vlan 201
  name Compute_201
vlan 712
  name Storage_712

interface Vlan7
  description Management
  no shutdown
  mtu 9216
  ip address 10.101.176.3/24
  hsrp 7
    ip 100.101.176.1

interface Vlan201
  description Compute
  no shutdown
  mtu 9216
  ip address 10.101.177.3/24
  hsrp 201
    ip 100.101.177.1

interface Ethernet1/1-3
  description To_Azure_Local_Host_FullyConverged
  switchport
  switchport mode trunk
  switchport trunk native vlan 7
  switchport trunk allowed vlan 7,201,712
  priority-flow-control mode on send-tlv
  spanning-tree port type edge trunk
  mtu 9216
  service-policy type qos input AZS_SERVICES
  no shutdown

Note

  • The above shows the baseline "One Storage VLAN per TOR" configuration
  • QoS policies and routing design (e.g., uplinks, BGP/OSPF, default gateway) will be introduced in a separate document

Configuration Validation

This section provides comprehensive validation procedures to verify the network configuration is functioning correctly in already deployed Azure Local environments. These validation commands are intended for troubleshooting and verifying existing production deployments, not for fresh installations.

Azure Local Host Validation

Use these PowerShell commands on deployed Azure Local nodes to validate network adapter configuration and RDMA connectivity:

1. Verify Network Adapter Configuration

# Display network adapters and VLAN assignments
Get-NetAdapter | Format-Table InterfaceAlias, VlanID, MacAddress -AutoSize

# Check VLAN isolation settings for virtual adapters
Get-VMNetworkAdapterIsolation -ManagementOS | Format-Table ParentAdapter, IsolationMode, DefaultIsolationID -AutoSize

2. Validate SMB Multichannel and RDMA Connectivity

[Host3]: PS C:\Users\Administrator\Documents> Get-NetAdapter | ft InterfaceAlias, VlanID, MacAddress

InterfaceAlias                     VlanID MacAddress       
--------------                     ------ ----------       
ethernet                                0 0C-42-A1-F9-69-4A
vSMB(managementcompute#ethernet 2)        00-15-5D-C8-20-07
ethernet 2                              0 0C-42-A1-F9-69-4B
vManagement(managementcompute)            0C-42-A1-F9-69-4A
vSMB(managementcompute#ethernet)          00-15-5D-C8-20-06

[Host3]: PS C:\Users\Administrator\Documents> Get-VMNetworkAdapterIsolation -ManagementOS | ft ParentAdapter, IsolationMode, DefaultIsolationID

ParentAdapter                                                         IsolationMode DefaultIsolationID
-------------                                                         ------------- ------------------
VMInternalNetworkAdapter, Name = 'vSMB(managementcompute#ethernet)'            Vlan                711
VMInternalNetworkAdapter, Name = 'vManagement(managementcompute)'              Vlan                  0
VMInternalNetworkAdapter, Name = 'vSMB(managementcompute#ethernet 2)'          Vlan                712



# Verify SMB Multichannel connections and RDMA usage
[Host3]: PS C:\Users\Administrator\Documents> Get-SmbMultichannelConnection | Format-Table ClientInterfaceFriendlyName, ClientIpAddress, ServerIpAddress, ClientRdmaCapable, RdmaConnectionCount, TcpConnectionCount -AutoSize

ClientInterfaceFriendlyName         ClientIpAddress ServerIpAddress ClientRdmaCapable RdmaConnectionCount TcpConnectionCount
---------------------------         --------------- --------------- ----------------- ------------------- ------------------
vSMB(managementcompute#ethernet 2)  10.71.2.10      10.71.2.34      True              2                   0
vSMB(managementcompute#ethernet)    10.71.1.249     10.71.1.214     True              2                   0

# Verify SMB connections
[Host3]: PS C:\Users\Administrator\Documents> Get-SmbConnection

ServerName ShareName UserName      Credential    Dialect NumOpens
---------- --------- --------      ----------    ------- --------
Host4      c$        Administrator Administrator 3.1.1   2

Note

SMB Multichannel Validation Key Points:

  • Both storage VLANs (711 and 712) are operational with RDMA enabled
  • RdmaConnectionCount = 2 confirms RDMA is being used for storage traffic
  • TcpConnectionCount = 0 shows no fallback to regular TCP
  • SMB 3.1.1 dialect is being used for optimal performance
  • This validates that storage traffic is using RDMA as intended
Switch Validation

Verify Top-of-Rack switch configuration and MAC address learning to ensure proper connectivity.

Validation Commands for Cisco NX-OS:

# On TOR1 (Storage VLAN 711 only)
TOR1# show mac address-table interface ethernet 1/3
Legend:
        * - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC
        age - seconds since last seen,+ - primary entry using vPC Peer-Link,
        (T) - True, (F) - False, C - ControlPlane MAC, ~ - vsan
   VLAN     MAC Address      Type      age     Secure NTFY Ports
---------+-----------------+--------+---------+------+----+------------------
*    7     0c42.a1f9.694a   dynamic  0         F      F    Eth1/3
*  711     0015.5dc8.2006   dynamic  0         F      F    Eth1/3

# On TOR2 (Storage VLAN 712 only)
TOR2# show mac address-table interface ethernet 1/3
Legend:
        * - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC
        age - seconds since last seen,+ - primary entry using vPC Peer-Link,
        (T) - True, (F) - False, C - ControlPlane MAC, ~ - vsan
   VLAN     MAC Address      Type      age     Secure NTFY Ports
---------+-----------------+--------+---------+------+----+------------------
*    7     0c42.a1f9.694b   dynamic  0         F      F    Eth1/3
*  712     0015.5dc8.2007   dynamic  0         F      F    Eth1/3

Expected Results:

  • TOR1 should show Management VLAN 7 and Storage VLAN 711 learned MAC addresses
  • TOR2 should show Management VLAN 7 and Storage VLAN 712 learned MAC addresses
  • This validates the baseline "One Storage VLAN per TOR" configuration ensuring RDMA traffic flow

Quality of Service (QoS)

Quality of Service configuration is essential for optimal Azure Local performance. For detailed QoS policy configuration:

Reference: Quality of Service (QoS) Policy

BGP Routing

For BGP routing configuration and best practices in Azure Local deployments:

Reference: Azure Local BGP Routing

Frequently Asked Questions

Q: Can I configure both Storage VLANs on both TOR switches?

A: Yes, you can configure both Storage VLANs on both TOR switches, but you won't get too much benefit from it. Here's why:

  1. We always want storage to stay in RDMA - The baseline configuration already ensures RDMA connectivity is maintained
  2. We don't want traffic to cross the TOR peer link - Having both VLANs on both switches may introduce unnecessary inter-switch traffic
  3. It's a balance between resilience vs performance - The added complexity doesn't provide significant resilience improvements

Baseline Configuration (Recommended):

  • TOR1: Configure only Storage VLAN 711 (plus Management and Compute VLANs)
  • TOR2: Configure only Storage VLAN 712 (plus Management and Compute VLANs)

This approach ensures:

  • Design consistency across deployment patterns
  • Guaranteed RDMA usage - storage traffic cannot fall back to regular TCP
  • Simplified configuration and troubleshooting
  • Optimal performance - no unnecessary cross-TOR traffic

Optional Enhanced Configuration: While customers can configure both Storage VLANs (711 and 712) on both TOR switches, this configuration is not recommended as it provides minimal additional resilience. The baseline "One Storage VLAN per TOR" configuration already provides sufficient redundancy through the host-side SET (Switch Embedded Teaming) and ensures reliable RDMA connectivity without the potential performance implications.

Important

The primary goal is to maintain design consistency, force storage traffic to always use RDMA instead of allowing fallback to regular TCP traffic, and avoid unnecessary cross-TOR traffic that could impact performance.

Additional Resources

For comprehensive Azure Local networking information, refer to these official Microsoft documentation resources:

Official Documentation

Technical Deep Dives