Juniper NETWORKS EVPN LAG Multihoming in EVPN-VXLAN Cloud Data Center Infrastructures User Guide

June 15, 2024
JUNIPER NETWORKS

Juniper NETWORKS EVPN LAG Multihoming in EVPN-VXLAN Cloud Data Center

Infrastructures

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-product

Product Information

Specifications

  • Product Name: EVPN LAG Multihoming in EVPN-VXLAN Cloud Data Center Infrastructures
  • Published: 2023-10-16
  • Manufacturer: Juniper Networks, Inc.
  • Address: 1133 Innovation Way Sunnyvale, California 94089 USA
  • Contact: 408-745-2000
  • Website: www.juniper.net

About This Guide
This guide provides networking professionals with concepts and tools related to EVPN LAG multihoming in EVPN-VXLAN cloud data center architectures.

Product Usage Instructions

  • Chapter 1: Introduction to EVPN LAG Multihoming
    This chapter provides an introduction to EVPN LAG multihoming.

  • Chapter 2: Ethernet Segment Identifiers, ESI Types, and LACP in EVPN LAGs
    This chapter explains Ethernet Segment Identifiers (ESI), ESI Types, and Link Aggregation Control Protocol (LACP) in EVPN LAGs.

  • Chapter 3: EVPN LAGs in EVPN-VXLAN Reference Architectures
    This chapter discusses the usage of EVPN LAGs in different EVPN-VXLAN reference architectures.

  • Chapter 4: EVPN Features in EVPNs using EVPN LAGs
    This chapter covers the various EVPN features available when using EVPN LAGs.

  • Chapter 5: EVPN LAG Pros and Cons
    This chapter provides an overview of the pros and cons of using EVPN LAGs.

  • Chapter 6: EVPN LAG Configuration Best Practices
    This chapter offers best practices for configuring EVPN LAGs in different architectures.

FAQ

Q: What is the purpose of this document?
A: The purpose of this document is to provide networking professionals with concepts and tools related to EVPN LAG multihoming in EVPN-VXLAN cloud data center architectures.

Juniper Networks, Inc.
1133 Innovation Way Sunnyvale, California 94089 USA
408-745-2000
www.juniper.net

Juniper Networks, the Juniper Networks logo, Juniper, and Junos are registered trademarks of Juniper Networks, Inc. in the United States and other countries. All other trademarks, service marks, registered marks, or registered service marks are the property of their respective owners. Juniper Networks assumes no responsibility for any inaccuracies in this document. Juniper Networks reserves the right to change, modify, transfer, or otherwise revise this publication without notice.

EVPN LAG Multihoming in EVPN-VXLAN Cloud Data Center Infrastructures Copyright © 2023 Juniper Networks, Inc. All rights reserved.

The information in this document is current as of the date.

YEAR 2000 NOTICE
Juniper Networks hardware and software products are Year 2000 compliant. Junos OS has no known time-related limitations through the year 2038. However, the NTP application is known to have some difficulty in the year 2036.

END USER LICENSE AGREEMENT
The Juniper Networks product that is the subject of this technical documentation consists of (or is intended for use with) Juniper Networks software. Use of such software is subject to the terms and conditions of the End User License Agreement (“EULA”) posted at https://support.juniper.net/support/eula/. By downloading, installing or using such software, you agree to the terms and conditions of that EULA.

About This Guide

Use this guide to learn more about EVPN LAG Multihoming implementation options and features in EVPN-VXLAN Cloud Data Center Infrastructures.

  • The purpose of this document is to provide networking professionals with concepts and tools related to EVPN LAG multihoming in EVPN-VXLAN cloud data center architectures.
  • The intended audience for this guide includes system integrators, infrastructure professionals, partners, and customers that want to utilize or optimize the use of EVPN LAG multihoming, also known as ESI-LAG multihoming, in their cloud data centers.

Introduction to EVPN LAG Multihoming

Ethernet link aggregation remains a vital technology in modern data center networking. Link aggregation provides the technology to bundle interfaces and to increase bandwidth in a data center by adding new links to bundled interfaces. Link aggregation can significantly increase a network’s high availability by providing a redundant path or paths that traffic can utilize in the event of a link failure and has a variety of other applications that remain highly relevant in modern data center networks.

New EVPN technology standards—including RFCs 8365, 7432, and 7348—introduce the concept of link aggregation in EVPNs through the use of Ethernet segments. Ethernet segments in an EVPN collect links into a bundle and assign a number—called the Ethernet segment ID (ESI)—to the bundled links. Links from multiple standalone nodes can be assigned into the same ESI, introducing an important link aggregation feature that introduces node level redundancy to devices in an EVPN-VXLAN network. The bundled links that are numbered with an ESI are often referred to as ESI LAGs or EVPN LAGs. The EVPN LAGs term will be used to refer to these link bundles for the remainder of this document.
Layer 2 Multihoming in EVPN networks is dependent on EVPN LAGs. EVPN LAGs, which provide full active-active link support, are also frequently enabled with LACP to ensure multivendor support for devices accessing the data center. Layer 2 Multihoming with LACP is an especially attractive configuration option when deploying servers because the multihoming is transparent from the server point of view; the server thinks it is connected to a single networking device when it is connected to two or more switches.

Figure 1 shows a fabric topology representing a standard 3 stage spine-leaf underlay architecture. An EVPN-VXLAN overlay offers link and node level redundancy and an active-active Layer 2 Ethernet extension between the leaf devices, without the need to implement a Spanning Tree Protocol (STP) to avoid loops. Different workloads in the data center are integrated into the fabric without any prerequisites at the storage system level. This setup would typically use QFX5110 or QFX5120 switches as leaf devices and QFX10002 or QFX10008 switches as spine devices.

Figure 1: Data Center Example Topology with EVPN LAGs

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(1\)

NOTE:
Single-homed end systems are still frequently used in modern data center networks. The iSCSi storage array systems in the figure provide an example of single-homed end systems in an EVPN-VXLAN data center.

See the EVPN Multihoming Implementation section of the EVPN Multihoming Overview document for additional information related to ESIs and ESI numbering. See Multihoming an Ethernet-Connected End System Design and Implementation in the Cloud Data Center Architecture Guide for a step-by-step configuration procedure of an EVPN LAG implementation.

Ethernet Segment Identifiers, ESI Types, and LACP in EVPN LAGs

This section discusses Ethernet Segment Identifiers (ESIs), ESI Types, and LACP in EVPN LAGs.

Ethernet Segment Identifiers and Numbering in EVPN LAGs
An EVPN LAG is identified using an Ethernet segment identifier (ESI). An ESI is a mandatory attribute that is required to enable EVPN LAG server multihoming. ESI values are encoded as 10-byte integers and are used to identify a multihomed segment. The same ESI value enabled on multiple leaf switch interfaces connected to the same server or BladeCenter form an EVPN LAG. This EVPN LAG supports active-active multihoming towards the connected server.

We recommend using an ESI value that uses the same values for the first 8 bytes with changes only in the 9th and 10th bytes on a per-EVPN LAG basis. The four ESIs used later in this document follow these assignment guidelines and use the following ESI values: 00:03:03:03:03:03:03:03:03:01, 00:03:03:03:03:03:03:03:03:02, 00:03:03:03:03:03:03:03:03:03, and 00:03:03:03:03:03:03:03:03:04. Using a structured ESI assignment method simplifies network administration while also ensuring the ESI values can be used across Junos OS releases. The need to be compatible across Junos releases is required in some environments due to changes made to ES import route community handling in Junos OS Release 17.3R3.

See the EVPN Multihoming Implementation section of the EVPN Multihoming Overview document for additional information related to ESIs and ESI numbering.

ESI Types in EVPN LAGs
Junos OS currently supports ESI type 0 (manually hard-coded ESI value), type 1 (auto derived from LACP), and type 5 (IRB-VGA gateway ESI encoded based on autonomous system number) ESI patterns. See the EVPN Multihoming Implementation section of the EVPN Multihoming Overview document for additional information related to ESI types.

EVPN multihoming is managed at the control plane level using EVPN Type 4 routes and the ES-Import Route-Target extended community, which in Junos is an automatically derived 6-byte value taken from the 10-byte ESI value. This extended community is used within the Type-4 ES-Route to signal using EVPN BGP messages when connecting leaf devices to the same multihomed CE device.

LACP in EVPN LAGs
External devices such as servers almost always require an LACP state that matches the ESI-enabled interface on the leaf switches to ensure the server can actively send traffic across all member interfaces in the LAG bundle. Specifically, the same ESI and LACP system identifier should be configured on all links that make up a LAG bundle. To ensure the suggested system ID consistency, its a best-practice to base the system ID on the ESI you assign to each LAG.
For example, if you assign ESI 00:03:03:03:03:03:03:03:03:01 to a LAG, then it is suggested that you derive the corresponding system ID from the last 6 bytes of that ESI, i.e., “03:03:03:03:03:01”. The suggested best practice of deriving the system ID from the LAG’s ESI is demonstrated in the configurations presented later in this document.

EVPN LAGs in EVPN-VXLAN Reference Architectures

This section provides an overview of the Juniper EVPN-VXLAN reference architectures and the role of EVPN LAGs in these architectures. It is intended as a resource to help readers understand EVPN LAG capabilities in different contexts. The standard EVPN-VXLAN architecture consists of a 3-stage spine- leaf architecture. The physical underlay is IP forwarding enabled—all leaf to spine underlay links are typically IPv4 routed—and the logical overlay layer uses MP-BGP with EVPN signaling for control plane-based MAC-IP address learning and to establish VXLAN tunnels between switches.

Juniper Networks has four primary data center architectures:

  • Centrally Routed Bridging (CRB)—inter-VNI routing occurs on the spine switches.
  • Edge Routed Bridging (ERB)—inter-VNI routing occurs on the leaf switches.
  • Bridged Overlay—inter-VLAN and inter-VNI routing occurs outside of the EVPN-VXLAN fabric. Example: Routing occurs at the firewall cluster connected to the EVPN-VXLAN.
  • Centrally-Routed Bridging Mutual (CRB-M)—architecture where the spine switches are also connecting the existing data center infrastructure with the EVPN LAG. CRB-M architectures are often used during data center migrations.

EVPN LAGs in Centrally Routed Bridging Architectures

  • In the CRB architecture, we recommend provisioning the EVPN LAGs at the leaf layer and connecting two or more leaf devices to each server or BladeCenter.
  • Figure 2 illustrates EVPN LAG provisioning in a CRB architecture.

Figure 2: EVPN LAGs in a CRB Architecture

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(2\)Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-
VXLAN-Cloud-Data-Center-Infrastructures-fig- \(3\)

BEST PRACTICE:
The same ESI value and LACP system ID should be used when connecting multiple leaf devices to the same server. Unique ESI values and LACP system IDs should be used per EVPN LAG.

EVPN LAGs in Edge-Routed Bridging Architectures
Figure 3 illustrates the use of EVPN LAGs within an Edge routing bridging (ERB) architecture. The recommended EVPN LAG provisioning in an ERB architecture is similar to the CRB architecture. The major difference between the architectures is that the IP first hop gateway capability is moved to the leaf level using IRB interfaces with anycast addressing.

The ERB architecture offers ARP suppression capability complemented by the advertisement of the most specific host /32 Type-5 EVPN routes from leaf devices toward the spine devices. This technology combination efficiently reduces data center traffic flooding and creates a topology that is often utilized to support Virtual Machine Traffic Optimization (VMTO) capabilities.

Figure 3: EVPN LAGs in ERB Architecture

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(4\) Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-
VXLAN-Cloud-Data-Center-Infrastructures-fig- \(5\)

EVPN LAGs in Bridged Overlay Architectures
In a bridged overlay architecture, VLANs are extended between leaf devices across VXLAN tunnels. EVPN LAGs are used in a bridged overlay to provide multihoming for servers and to connect to first hop gateways outside of the EVPN-VXLAN fabric, which are typically SRX Series Services Gateways or MX Series routers. The bridged overlay architecture helps conserve bandwidth on the gateway devices and increases the bandwidth and resiliency of servers and BladeCenters by delivering active-active forwarding in the same broadcast domain.

  • Figure 4 illustrates EVPN LAGs in a sample bridged overlay architecture.

Figure 4: EVPN LAGs in Bridged Overlay Architecture

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(6\)

EVPN LAGs in Centrally Routed Bridging Migration Architectures
EVPN LAGs might be introduced between the spine and leaf devices during migration to one of the aforementioned EVPN-VXLAN reference architectures. This EVPN LAG is needed in some migration scenarios to integrate the existing legacy ToR-based infrastructure into the EVPN-VXLAN architecture. Figure 5 shows a Virtual Chassis and an MC-LAG architecture connected to spine devices using an EVPN LAG. The EVPN LAG provisioning is done from the spine devices during the migration of these topologies into an EVPN-VXLAN reference architecture.

Figure 5: EVPN LAGs in CRB Migration Architectures

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(7\)

The CRB Migration architecture is often used when migrating an MC-LAG or Virtual Chassis-based data center in phases. In this architecture, the EVPN LAG capability is introduced at the spine level and only one overlay iBGP session is running between the two spine switches. The top-of-rack switches connected to the spine devices are legacy switches configured as Virtual Chassis or MC-LAG clusters with no EVPN iBGP peerings to the spine switches. This architecture helps when deploying EVPN-VXLAN technologies in stages into an existing data center. The first step is building an EVPN LAG-capable spine layer, and then sequentially migrating to an EVPN control plan where MAC addresses for the new leaf switches are learned from the spine layer switches. The new leaf switches, therefore, can benefit from the advanced EVPN features, such as ARP suppression, IGMP suppression, and optimized multicast, supported by the new switches.

The default EVPN core isolation behavior should be disabled in CRB Migration architectures. The default EVPN core-isolation behavior disables local EVPN LAG members if the network loses the last iBGP-EVPN signaled peer. Because this peering between the two spine devices will be lost during the migration, the default behavior-which can be changed by entering the no-core-isolation option in the edit protocols evpn hierarchy—must be changed to prevent core isolation events.

EVPN Features in EVPNs using EVPN LAGs

This section introduces some commonly-used features in EVPNs that are using EVPN LAGs.

EVPN Aliasing
EVPN aliasing is the ability of a remote device to load balance Layer 2 unicast traffic across an Ethernet segment towards an endpoint device. See the Aliasing section of the EVPN Multihoming Overview topic for additional information on aliasing.

Servers and BladeCenter switches connected to an EVPN LAG with multiple links may send ARP requests to one of the leaf devices. The leaf devices respond by advertising the MAC address to the rest of the EVPN topology using the overlay iBGP peerings. The other leaf devices in the EVPN network use the default EVPN aliasing (EVPN Route type 1 per EVI) functionality to learn the EVPN Type 2 MAC routes from other leaf devices. All of the leaf devices are connected to the same ESI, however, for forwarding purposes.

EVPN LAG Multihomed Proxy Advertisement
A leaf device in a network running EVPN responds to a server’s ARP request with a MAC and IP EVPN Type 2 route. Starting in Junos OS Release 18.4, multiple leaf devices that have links in the same EVPN LAG have the capability of also advertising MAC and IP EVPN Type 2 routes to the server that sent the ARP request. This capability is possible using the proxy M-bit in EVPN Type 2 routes. This capability is especially beneficial in failure scenarios. If a switch is multihomed to two leaf devices and the link to one of those devices fails, a Type 2 message can be sent toward the server that initially sent the ARP request and traffic can be sent across the network over the active links.

High Availability in EVPN LAGs
EVPN LAGs include many built-in high availability capabilities.

EVPN LAG link level fast convergence is delivered using the EVPN IBGP route type I massive withdrawal message. Node level fast convergence is handled using the BFD for IBGP in the overlay network. Node-level Layer 2 EVPN LAG redundancy is available in an EVPN-VXLAN fabric using built-in EVPN loop prevention mechanisms like split horizon and designated forwarding. Spanning tree protocol (STP) does not need to run in an EVPN-VXLAN fabric. See Designated Forwarding Election and Split Horizon in the EVPN Multihoming Overview topic for additional information on these features.

The core isolation feature quickly brings down the local EVPN-LAG when all IBGP EVPN overlay peerings are lost. This feature diverts traffic to other paths when a link is down. See When to Use the Core Isolation Feature for additional information on core isolation.

EVPN LAG Interfaces Advanced Capabilities
EVPN LAG multihoming has a series of advanced feature capabilities when compared to other LAG multihoming technologies such as Virtual Chassis or Multichassis link aggregation (MC-LAG). The advanced features include Proxy ARP and ARP suppression, IGMP proxy, MAC mobility, MAC history, and duplicate MAC detection. ARP suppression in an ERB architecture helps reduce Ethernet broadcast traffic flooding across the Layer 2 fabric, thereby freeing up server resources across the fabric. ARP requests, therefore, are not flooded to other leaf devices when the ARP table is already populated with the address from other signaling events, most notably EVPN route sharing. See EVPN Proxy ARP and ARP Suppression, and NDP and NDP Suppression for additional information on Proxy ARP and ARP suppression.

IGMP proxy helps reduce IGMP membership report flooding in a data center by translating the IGMP messages into EVPN type 6 routes to send to all of the leaf devices in the data center. See Overview of IGMP Snooping in an EVPN- VXLAN Environment for additional information on IGMP Proxy. MAC mobility history allows EVPN LAGs to track how often a MAC address moves between ESIs. MAC mobility history allows network administrators to create security actions based on MAC address movements while also simplifying MAC address-related administration in an EVPN. See Overview of MAC Mobility for additional information on this feature.

EVPN LAG Pros and Cons

The following table summarizes the pros and cons of EVPN-VXLAN and EVPN multihomed data centers.

Table 1: EVPN LAG Pros and Cons

Pros Cons

Flexibility.

EVPN-VXLAN is an industry standard based approach to data center fabric creation that appears to be gaining traction and wide adoption across the networking industry.

| More devices to manage than in case of a single aggregated platform like Virtual Chassis.
Full support for active-active multihomed access links.| Perceived as complex.
Integrates with Contrail provisioning.| Use with FCoE/FIP is not supported.
Functional parity with Virtual Chassis for network layer functions in next generation access (NGA) use cases.|
Simplified overall topology.

EVPN, in particular when configured at the leaf layer, reduces the total number of technologies in a data center fabric.

|
Simplified configuration.

The same configuration can be shared per given EVPN LAG between different leaf devices and there is no need to provision back to back links between the leaf devices.

|
Flexible configuration.

EVPN-VXLAN is supported using enterprise-style and service-provider style (ELS and non-ELS) configurations.

|
Improved services delivery.

EVPN-VXLAN introduces the concept of VLAN normalization at the services dedicated border-leafs level and the overlapped vlan-id (unique VNI) capability at the per POD level (pair of leafs).

|
---|---

EVPN LAG Configuration Best Practices

Junos OS on QFX Series switches support Enterprise style configuration and Service Provider style configuration. Both configuration styles can be used to configure EVPN LAGs. We recommend using the Enterprise configuration style because it supports more data center features like storm control profiles and BPDU blocking without having to enable RSTP on the leaf devices. This section covers EVPN LAG configuration using both configuration styles.

EVPN LAG Configuration of a CRB Architecture Using the Enterprise Configuration Style
The Enterprise configuration style is the recommended way to enable EVPN LAG functionality in most data center architectures. Enterprise style provisioning is generally simpler than Service Provider style provisioning and generally is more compatible with other Layer 2 features.

The following configuration provides a sample EVPN LAG configuration done using the Enterprise configuration style on the leaf devices in a Centrally Routed Bridging architecture.

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(8\) Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-
VXLAN-Cloud-Data-Center-Infrastructures-fig- \(9\) Juniper-NETWORKS-EVPN-LAG-
Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-Infrastructures-fig-
\(10\) Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(11\) Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-
VXLAN-Cloud-Data-Center-Infrastructures-fig- \(12\) Juniper-NETWORKS-EVPN-LAG-
Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-Infrastructures-fig-
\(13\) Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(14\)

EVPN LAG Configuration of a CRB Architecture Using the Service Provider Configuration Style
The QFX5110 switches also support the Service Provider style of configuration. When configuring EVPN LAGs in the Service Provider configuration style, multiple units per given EVPN LAG are assigned. These multiple units provide an opportunity to enable more selective filter and rate limiting per unit interface, but these assignments must be done on a per unit basis and are therefore burdensome to configure and maintain. This granularity is generally not needed in data center architectures using spine and leaf topologies, so we do not recommend using Service Provider style configuration when enabling EVPN LAGs in most data center environments.

The following configuration provides a sample EVPN LAG configuration done using the Service Provider configuration style. This configuration is supported when QFX5110 or QFX5120 switches are operating in the leaf device role in an architecture where the IRBs are not on the leaf devices, which are the Centrally Routed Bridging (CRB) and Bridged Overlay (BO) architectures. The Enterprise configuration style must be used to enable EVPN LAGs when a QFX5110 or QFX5120 switch is used as the leaf device in an Edge Routed Bridging (ERB) architecture.

Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-VXLAN-Cloud-Data-Center-
Infrastructures-fig- \(15\) Juniper-NETWORKS-EVPN-LAG-Multihoming-in-EVPN-
VXLAN-Cloud-Data-Center-Infrastructures-fig- \(16\)

RELATED DOCUMENTATION

  • Cloud Data Center Architecture Guide
  • EVPN Feature Guide
  • Multihoming an Ethernet-Connected End System Design and Implementation
  • EVPN Multihoming Overview

Juniper Networks, Inc.

References

Read User Manual Online (PDF format)

Read User Manual Online (PDF format)  >>

Download This Manual (PDF format)

Download this manual  >>

JUNIPER NETWORKS User Manuals

Related Manuals