Data Center Solution

Data Center Bridging Exchange Technology

Coordinate PFC, ETS, and application priority discovery for RDMA-ready networks.

Back to Data Center Solutions

Overview

Data Center Bridging Exchange (DCBX) is the negotiation layer used by adjacent Ethernet devices to discover and align Data Center Bridging behavior. It rides on LLDP and exchanges lossless Ethernet parameters such as Priority Flow Control (PFC), Enhanced Transmission Selection (ETS), and application priority mappings.

For xSONiC data center fabrics, DCBX matters most when RoCEv2, storage, HPC, or AI training traffic needs predictable low-loss forwarding. These workloads can be sensitive to packet loss, retransmission, queue pressure, and inconsistent priority treatment across servers and switches.

Why It Matters

  • Reduces configuration drift between servers, adapters, and switch ports.
  • Keeps RDMA and storage traffic classes mapped consistently across links.
  • Helps validate whether PFC and ETS behavior is aligned before production traffic arrives.
  • Lowers the risk of link-by-link manual QoS errors in large deployments.
  • Gives operators a negotiation mechanism that complements static QoS templates.

Functional Scope

DCBX is not a complete QoS policy by itself. It is a discovery and exchange mechanism that helps devices understand the DCB capabilities and intended configuration on the other side of a link.

FunctionWhat It DoesOperational Value
Peer discoveryLearns DCB capabilities and configuration from the adjacent device.Reduces blind spots during server-to-switch and switch-to-switch bring-up.
Configuration exchangeShares PFC, ETS, and application priority information.Helps align lossless classes and bandwidth treatment across the link.
Change monitoringTracks local and peer-side DCB changes.Helps operators detect drift after maintenance or policy updates.
TechnologyRole in a Lossless FabricTypical xSONiC Use
PFCPauses selected priority classes instead of pausing the entire link.Protects RDMA/storage classes from packet loss under congestion.
ETSAllocates bandwidth and scheduling treatment among traffic classes.Keeps lossless classes from starving normal traffic, and vice versa.
Application PriorityMaps specific applications or protocols to traffic priorities.Aligns RoCEv2, storage, and management traffic with intended queues.
LLDPCarries DCBX information between directly connected neighbors.Provides the transport for link-local DCBX exchange.

DCBX TLV Structure

DCBX uses LLDP extension TLVs to carry DCB information. The LLDP TLV type is vendor-specific, with IEEE DCBX information identified by the IEEE OUI. The payload then identifies which DCBX subtype is being exchanged.

TLVSubtypeTypical LengthPurpose
ETS Configuration TLV0x0925 bytesAdvertises ETS parameters and bandwidth allocation.
ETS Recommendation TLV0x0A25 bytesCarries recommended ETS values for the peer.
PFC Configuration TLV0x0B6 bytesExchanges PFC enablement and capability information.
Application Priority TLV0x0CVariableMaps applications to traffic priorities.

State Machine

DCBX behavior can be understood as a per-port state machine. Each enabled port collects local configuration, advertises it, reads peer information, updates local state when policy allows, then keeps monitoring for changes.

StateProcessingNext Step
Local configuration collectionRead local DCB settings, capability, and willingness.Advertise locally if a peer exists; otherwise monitor for changes.
Local configuration advertisementSend local DCBX information to the adjacent device.Collect peer configuration if negotiation is allowed.
Peer configuration collectionRead peer capability, willingness, and advertised DCB settings.Decide whether local state should be updated.
Local configuration updateReconcile local and peer DCB configuration according to policy.Return to monitoring after convergence.
Configuration change monitoringWatch for local or peer-side changes.Restart collection when state changes are detected.

Workflow View

Local DCB policy
      |
      v
Collect local PFC / ETS / APP priority settings
      |
      v
Advertise settings through LLDP DCBX TLVs
      |
      v
Receive peer DCBX TLVs
      |
      v
Compare willingness, capability, and priority mappings
      |
      v
Apply accepted updates or keep local policy
      |
      v
Monitor for link, peer, or policy changes

PFC Example

Consider a server connected to an xSONiC leaf switch. If the switch has PFC enabled for the RDMA priority but the server NIC does not, congestion can still lead to packet loss. The switch may send PFC pause frames, but the endpoint will continue transmitting if it has not negotiated or enabled the same lossless behavior.

Misaligned configuration

Server NIC          xSONiC Leaf
PFC disabled  --->  PFC enabled on RDMA priority
Traffic keeps       Switch buffer fills under congestion
transmitting        Packet loss or retransmission risk increases

With DCBX enabled on both sides, the server and switch can exchange DCB capability and configuration information before the workload depends on that link. Operators still need to define the intended QoS policy, but DCBX reduces the chance that each link carries a different interpretation of the same traffic class.

Aligned DCBX exchange

Server NIC <--- LLDP DCBX TLVs ---> xSONiC Leaf
     |                                  |
     +-- PFC / ETS / APP priority -----+
     +-- Capability and willingness ---+
     +-- Monitoring after convergence -+

Switch-to-Switch Consistency

DCBX is also useful beyond the server edge. In a leaf/spine or storage fabric, switches may need to keep PFC and ETS behavior consistent along the forwarding path. When two switches advertise incompatible priority treatment, DCBX can surface the difference and, where policy permits, help one side accept the peer’s configuration.

For xSONiC deployments, this is especially relevant across:

  • 100G and 200G storage or frontend links.
  • 400G and 800G AI backend fabrics.
  • Migration stages where some links are newly configured and others are already in service.
  • Multi-pod fabrics where RoCEv2 behavior must remain consistent across a larger path.

Deployment Guidance

Use DCBX as part of a deliberate QoS design, not as a substitute for one.

  1. Define which traffic classes require lossless treatment.
  2. Map those classes to explicit priorities and queues.
  3. Decide which devices are allowed to accept peer recommendations.
  4. Enable DCBX on server-facing and fabric-facing links where negotiation is required.
  5. Validate negotiated PFC, ETS, and application priority state on both sides of each link.
  6. Monitor for configuration drift after NIC, switch, or software updates.

xSONiC Platform Fit

DCBX planning is most important on xSONiC data center switches that participate in RoCEv2, storage, HPC, or AI backend fabrics. It is a natural fit for 100G, 200G, 400G, and 800G leaf/spine designs where small inconsistencies can become expensive once large-scale traffic starts.

Related Products

Products commonly paired with this solution.

Use these related platforms as a starting point for sizing, comparison, and follow-up discussion.

XS-DC-32X100-LS-G1 front panel product image

XS-DC-32X100-LS-G1

Data Center AI

32-port 100G leaf/spine switch for compact data center fabrics, cloud routing, and high-throughput server aggregation.

3.2Tbps
4,760Mpps
XS-DC-64X200-LS-G1 front panel product image

XS-DC-64X200-LS-G1

Data Center AI

64-port 200G leaf/spine switch for high-bandwidth storage, compute, and scale-out data center fabrics.

12.8Tbps
19,040Mpps
XS-DC-64X800-AI-G1 front panel product image

XS-DC-64X800-AI-G1

Data Center AI

64-port 800G AI fabric switch for large-scale GPU clusters, HPC backbones, and ultra-high-throughput data center networks.

51.2Tbps
42,000Mpps
Next Step

Move from Data Center Bridging Exchange Technology into implementation.

Use the related products below to continue comparing platforms, or open a conversation if you need help mapping the solution to your environment.