When you scale AI workloads across a single machine, networking becomes as important as the processors themselves. A cluster of powerful GPUs can be brought to a crawl if they cannot communicate effectively. In distributed systems, where multiple machines work together on a single problem, the network is the fabric that binds them. It's responsible for shipping training data to nodes, synchronizing model parameters, and collecting results. If this fabric is slow or unreliable, your expensive compute resources will spend most of their time waiting.
This section examines the two primary metrics for network performance, bandwidth and latency, and explains why they are significant for common AI training patterns.
When evaluating a network, we focus on two fundamental characteristics:
Bandwidth: Often called throughput, bandwidth is the maximum amount of data that can be transferred over a network connection in a given amount of time. It's typically measured in bits per second, such as gigabits per second (Gbps) or even terabits per second (Tbps). Think of bandwidth as the number of lanes on a highway. A 10-lane highway (high bandwidth) can carry more cars (data) at once than a 2-lane road (low bandwidth). For AI, high bandwidth is essential for moving large objects, like datasets, model checkpoints, or the entire set of model parameters between nodes.
Latency: This is the time it takes for a single piece of data, a packet, to travel from its source to its destination. It's a measure of delay, typically expressed in milliseconds (ms) or microseconds (µs). In our highway analogy, latency is the time it takes for a single car to complete its journey from start to finish, regardless of how many lanes there are. Low latency is critical for operations that involve frequent, small, back-and-forth communications.
In distributed training, particularly with data parallelism, nodes must constantly synchronize model updates (gradients) after each training step. This involves many small, frequent messages. If latency is high, each synchronization step introduces a significant delay. The GPUs finish their work and then wait for the network, leading to poor utilization.
An ideal AI network combines high bandwidth (a wide pipe) to move large volumes of data with low latency (a short pipe) to ensure rapid communication for synchronization tasks.
The effectiveness of a network is determined by how well it handles the specific communication patterns of an application. In distributed deep learning, one of the most demanding patterns is All-to-All. During gradient synchronization, every GPU node needs to send its calculated gradients to every other node.
Imagine a cluster of 8 GPUs. After a training step, each GPU has a piece of the total gradient. To prepare for the next step, every GPU needs the complete, averaged gradient. This requires a complex shuffle of data where every node simultaneously talks to every other node.
High latency in an All-to-All exchange creates a compounding delay, as the slowest connection can hold up the entire cluster. This is why specialized networking hardware and libraries are used in large-scale AI systems.
While standard enterprise Ethernet (1 GbE or 10 GbE) is fine for general IT tasks, it is often insufficient for serious distributed AI workloads. High-performance computing (HPC) environments have long relied on more advanced interconnects.
InfiniBand: This is a high-performance networking standard designed for very low latency and high bandwidth. It is a common choice for building large AI supercomputers. An important feature of InfiniBand is its support for Remote Direct Memory Access (RDMA).
RDMA (Remote Direct Memory Access): In a traditional network stack, moving data from one machine to another requires multiple steps involving the CPU and operating system on both the sending and receiving ends. This process introduces significant latency and consumes CPU cycles. RDMA allows the network interface card (NIC) of one machine to write data directly into the memory (RAM or even VRAM) of another machine, bypassing the CPU and OS. This dramatically reduces latency and frees the CPU to perform other work.
RDMA over Converged Ethernet (RoCE): This is a protocol that allows you to achieve RDMA-like performance over a standard Ethernet network, providing a competitive alternative to InfiniBand if the underlying network is properly configured to be "lossless".
Understanding these networking fundamentals is the first step. As you'll see in later chapters on designing on-premise and cloud infrastructure, the choice of network technology has a direct and significant impact on both system performance and total cost. An investment in a low-latency, high-bandwidth network can pay for itself by ensuring your expensive GPU resources are always busy computing, not waiting.
Was this section helpful?
© 2026 ApX Machine LearningEngineered with