Loading learning content...
Bandwidth tells us how fast data can flow through a pipe, but it reveals nothing about how long that data takes to arrive. A 10 Gbps transcontinental fiber link and a 10 Gbps local data center connection have identical bandwidth—yet the user experience differs dramatically. The difference is delay.
Delay (also called latency) measures the time required for data to travel from source to destination. For real-time applications like voice, video conferencing, online gaming, and financial trading, delay often matters more than raw bandwidth. A video call on a 10 Mbps connection with 20ms delay will feel instantaneous, while the same call on 1 Gbps with 500ms delay becomes an exercise in frustration.
By the end of this page, you will understand the four components of network delay (propagation, transmission, queuing, and processing), how routing protocols incorporate delay into path selection, the distinction between static and dynamic delay metrics, and the trade-offs between bandwidth and delay optimization.
Total end-to-end delay is not a single phenomenon but a composition of four distinct delay types. Understanding each component is essential for network design, troubleshooting, and routing metric configuration.
| Component | Definition | Determinants | Typical Values |
|---|---|---|---|
| Propagation Delay | Time for a signal to travel through the medium | Distance and medium speed (typically 2/3 speed of light) | 5 μs/km (fiber), 500ms+ (satellite) |
| Transmission Delay | Time to push all bits onto the link | Packet size ÷ Link bandwidth | 1.2 μs (1500B @ 10Gbps), 214 ms (1500B @ 56Kbps) |
| Queuing Delay | Time waiting in router output queue | Traffic load and queue length | 0 (empty queue) to seconds (congested) |
| Processing Delay | Time for router to process packet header | Router CPU, forwarding method | Typically < 1 μs in modern hardware |
Total Delay Formula:
Total Delay = Propagation + Transmission + Queuing + Processing
For a path with n hops:
Total Delay = Σ(Propagation[i] + Transmission[i] + Queuing[i] + Processing[i])
where i = 1 to n
Let's examine each component in detail:
In high-bandwidth, uncongested networks, propagation delay typically dominates. In congested networks, queuing delay dominates. In legacy low-bandwidth networks, transmission delay can dominate. Understanding which component dominates in your specific environment is crucial for optimization efforts.
Propagation delay represents a fundamental physical limit that no amount of bandwidth or router upgrades can overcome. Understanding this helps set realistic expectations for network performance, especially for geographically distributed systems.
123456789101112131415161718192021222324252627282930
Propagation Delay Formula:════════════════════════════════════════════════════════ Propagation Delay = Distance / Propagation Speed Where:• Distance = physical path length (km or m)• Propagation Speed = speed of signal in medium Propagation Speeds by Medium:──────────────────────────────────────────────────────• Vacuum: 299,792 km/s (speed of light, c)• Fiber optic: ~200,000 km/s (~0.67c)• Copper cable: ~200,000 km/s (~0.67c)• Wireless (air): ~299,000 km/s (~c)• Geostationary sat: 299,792 km/s but 72,000+ km round trip! Example Calculations:──────────────────────────────────────────────────────New York to London (5,500 km fiber): = 5,500 km ÷ 200,000 km/s = 27.5 ms one-way Round-trip: ~55 ms minimum New York to Tokyo (11,000 km fiber): = 11,000 km ÷ 200,000 km/s = 55 ms one-way Round-trip: ~110 ms minimum Geostationary Satellite (36,000 km altitude): = 72,000 km (up and down) ÷ 299,792 km/s = 240 ms one-way Round-trip: ~480 ms minimum (often 500-600ms with processing)Why Propagation Delay Cannot Be Reduced:
Propagation delay is constrained by the speed of light—a fundamental physical constant. Even with theoretical improvements:
Implications for Global Systems:
| Route | Approximate Distance | Theoretical Min RTT | Realistic RTT |
|---|---|---|---|
| Same datacenter | < 1 km | < 0.01 ms | 0.1-0.5 ms |
| Same city | 10-50 km | 0.1-0.5 ms | 1-5 ms |
| Cross-country (US) | 4,000 km | 40 ms | 60-80 ms |
| Transatlantic | 6,000 km | 60 ms | 70-90 ms |
| Transpacific | 10,000 km | 100 ms | 120-150 ms |
| Global (antipodal) | 20,000 km | 200 ms | 250-350 ms |
| Geostationary satellite | 72,000 km | 480 ms | 500-700 ms |
Since propagation delay cannot be reduced, the solution is to reduce distance. Content Delivery Networks (CDNs) place content closer to users—if you can't make light go faster, bring the destination closer. This is why global services deploy servers on every continent and in every major city.
Transmission delay (also called serialization delay) is the time required to push all bits of a packet onto the transmission medium. Unlike propagation delay, transmission delay can be reduced by increasing bandwidth.
123456789101112131415161718192021
Transmission Delay Formula:════════════════════════════════════════════════════════ Transmission Delay = Packet Size (bits) / Bandwidth (bps) Example: 1,500-byte packet on various links:──────────────────────────────────────────────────────Packet Size = 1,500 bytes = 12,000 bits Link Bandwidth | Calculation | Delay────────────────────────────────────────────────────56 Kbps (dialup) | 12,000 / 56,000 | 214.3 ms1 Mbps (DSL) | 12,000 / 1,000,000 | 12.0 ms10 Mbps (Ethernet) | 12,000 / 10,000,000 | 1.2 ms100 Mbps (FastE) | 12,000 / 100,000,000| 0.12 ms (120 μs)1 Gbps (GigE) | 12,000 / 1×10^9 | 0.012 ms (12 μs)10 Gbps | 12,000 / 10×10^9 | 1.2 μs100 Gbps | 12,000 / 100×10^9 | 0.12 μs Key Insight: Transmission delay becomes negligible at high bandwidthsbut dominates at low bandwidths (legacy WAN links)Store-and-Forward Impact:
At each hop, a router must receive the entire packet before forwarding it (store-and-forward switching). This means transmission delay is incurred at every hop:
Total Transmission Delay = Packet Size × Number of Links / Bandwidth
For a 1,500-byte packet over 10 Gbps links with 5 hops:
= 12,000 bits × 5 / 10,000,000,000 bps
= 6 microseconds total
For the same packet over 1 Mbps links with 5 hops:
= 12,000 bits × 5 / 1,000,000 bps
= 60 milliseconds total
The 10,000× difference in transmission delay explains why bandwidth was historically so important for interactive applications, and why modern high-speed networks largely eliminate transmission delay as a concern.
Some switches use cut-through switching, which begins forwarding a frame as soon as the destination address is read (before the entire frame arrives). This reduces effective transmission delay at Layer 2 but only works for local switching—routers still require store-and-forward behavior for Layer 3 processing.
Of the four delay components, queuing delay is uniquely variable—it can range from zero (empty queues) to unbounded (during severe congestion). This variability makes queuing delay both the most impactful and the hardest to predict or incorporate into routing metrics.
The Queuing Model:
When multiple input interfaces want to send traffic through the same output interface, packets queue up waiting for transmission. The queuing delay depends on:
12345678910111213141516171819202122
M/M/1 Queue Model (Simplified Queuing Theory):════════════════════════════════════════════════════════ Traffic Intensity (ρ) = Arrival Rate (λ) / Service Rate (μ) Average Queuing Delay = ρ / (μ × (1 - ρ)) Example with 1 Gbps output link:──────────────────────────────────────────────────────Service Rate (μ) = 1 Gbps = 1,000,000,000 bps Traffic Load | ρ | Average Queue Delay | Observation────────────────────────────────────────────────────────100 Mbps | 0.1 | 0.11 μs | Nearly empty queue500 Mbps | 0.5 | 1.0 μs | Moderate queuing900 Mbps | 0.9 | 9.0 μs | Significant queuing990 Mbps | 0.99 | 99 μs | Heavy congestion999 Mbps | 0.999| 999 μs | Severe congestion Key Insight: Queuing delay grows EXPONENTIALLY as utilization approaches 100%. This is why networks are designed for 70-80% maximum utilization, not 100%.Why Routing Protocols Avoid Dynamic Delay:
Given queuing delay's variability, one might expect routing protocols to measure and react to actual delays. However, most protocols use STATIC delay values rather than real-time measurements. Here's why:
Routing Instability: If routes changed based on current congestion, traffic would shift to the less-congested path, which would then become congested, causing traffic to shift back—creating oscillations.
Measurement Overhead: Continuously measuring delay consumes bandwidth and router resources.
Convergence Time: By the time measurements propagate and routes converge, the congestion pattern may have changed.
Self-Fulfilling Prophecy: Announcing delay increases can cause traffic to shift away, reducing delay, causing traffic to return—unstable feedback loops.
Early ARPANET experiments with dynamic delay metrics demonstrated these problems dramatically, leading to the adoption of static metrics in production protocols.
Network engineers often target maximum 70% link utilization. At this level, queuing delays remain manageable. As utilization increases toward 100%, queuing delays explode non-linearly. Capacity planning must account for this—a link at 95% utilization is not 'almost full,' it's in crisis mode.
EIGRP (Enhanced Interior Gateway Routing Protocol) is notable for explicitly incorporating delay into its composite metric. Understanding EIGRP's delay handling provides insight into practical delay metric implementation.
1234567891011121314151617181920212223242526
EIGRP Delay Component:════════════════════════════════════════════════════════ Delay Value = Sum of interface delays along path (measured in tens of microseconds, then scaled) In the composite metric formula:Metric = [(K1 × BW) + (K2 × BW)/(256–Load) + (K3 × Delay)] × 256 With defaults (K1=1, K2=0, K3=1, K4=0, K5=0):Metric = (Bandwidth + Delay) × 256 Where:• Bandwidth = 10^7 / minimum BW in Kbps (bottleneck bandwidth)• Delay = Sum of delays / 10 (cumulative delay) Default Interface Delays (Cisco):──────────────────────────────────────────────────────Interface Type | Default Delay | In tens of μs──────────────────────────────────────────────────────10 Gbps Ethernet | 10 μs | 11 Gbps Ethernet | 10 μs | 1FastEthernet (100 Mbps) | 100 μs | 10Ethernet (10 Mbps) | 1,000 μs | 100T1 (1.544 Mbps) | 20,000 μs | 2000Serial (64 Kbps) | 20,000 μs | 2000Key EIGRP Delay Characteristics:
1. Cumulative Summation Unlike bandwidth (which uses the minimum/bottleneck value), EIGRP sums delays across all links in the path. This accurately reflects reality—each link adds to total delay.
2. Static Configuration EIGRP uses statically configured delay values, not measured delays. These defaults approximate typical physical characteristics but can be overridden.
3. Delay Tuning for Traffic Engineering Administrators can modify delay values to influence path selection:
12345678910111213141516
! View current delay valueshow interface GigabitEthernet0/0 | include DLY! Output: DLY 10 usec ! Modify delay to influence EIGRP path selectioninterface GigabitEthernet0/0 delay 100 ! Sets delay to 100 tens of microseconds (1000 μs = 1 ms) ! Make a path less preferred by increasing delayinterface GigabitEthernet0/1 delay 10000 ! Sets delay to 10000 tens of microseconds (100 ms) ! Verificationshow ip eigrp topology all-linksWhy Delay and Bandwidth Together?
EIGRP's combination of bandwidth and delay provides more nuanced path selection:
| Scenario | Bandwidth-Only | Delay-Only | Combined |
|---|---|---|---|
| High-BW, High-Delay (satellite) | Preferred | Not preferred | Balanced consideration |
| Low-BW, Low-Delay (local) | Not preferred | Preferred | Balanced consideration |
| High-BW, Low-Delay (fiber) | Preferred | Preferred | Strongly preferred |
| Low-BW, High-Delay (legacy WAN) | Not preferred | Not preferred | Strongly avoided |
This composite approach captures the trade-offs that network engineers intuitively understand: sometimes you accept lower bandwidth for lower latency (real-time applications), and sometimes you accept higher latency for higher bandwidth (bulk transfers).
EIGRP's K values allow weighting different metric components. Setting K3=0 would ignore delay entirely, making EIGRP behave like OSPF (bandwidth-only). However, Cisco strongly recommends keeping default K values (K1=K3=1, others=0) and ensuring they match across all EIGRP neighbors—mismatched K values prevent neighbor formation.
Unlike EIGRP, OSPF does not natively incorporate delay into its cost calculation. OSPF cost is based purely on bandwidth (Reference BW / Interface BW). However, network engineers can approximate delay-awareness through manual cost configuration.
The OSPF Delay Problem:
Consider two paths from A to B:
OSPF sees both paths as identical (cost 1) because bandwidth is identical. With equal costs, OSPF may even load-balance between them—sending half your traffic through a 500ms delay path!
Manual Delay Compensation:
Administrators can manually increase costs on high-delay links:
123456789101112131415161718192021
! Scenario: Two 10 Gbps links - one local fiber, one satellite ! Reference bandwidth = 100 Gbpsrouter ospf 1 auto-cost reference-bandwidth 100000 ! Local 10 Gbps fiber - auto-calculated cost = 10interface TenGigabitEthernet0/0 description Local Fiber Path ! Cost 10 from auto-calculation (100000/10000) ! No manual override needed ! Satellite 10 Gbps link - override to reflect delayinterface TenGigabitEthernet0/1 description Satellite Path - High Delay ip ospf cost 500 ! Manually set to 500 to make it backup only ! Now local path (cost 10) is strongly preferred ! Result: Traffic uses low-delay fiber path exclusively! Satellite only used during fiber failureDelay-Aware OSPF Extensions:
Recognizing OSPF's limitations, several extensions have been developed:
1. OSPF-TE (Traffic Engineering) OSPF-TE extensions (RFC 3630) allow advertising additional link attributes including delay, but these are primarily used by MPLS-TE for constraint-based routing rather than native OSPF path selection.
2. Performance-Based Routing Some vendor implementations (like Cisco's Performance Routing - PfR) overlay performance measurements on top of IGP routing, allowing delay-aware path selection without modifying OSPF itself.
3. Segment Routing Segment Routing can use Flex-Algo to define computation constraints including delay, enabling delay-optimized paths while maintaining OSPF as the underlying protocol.
4. SD-WAN Software-Defined WAN solutions typically measure real-time latency and select paths accordingly, operating above the IGP layer.
OSPF's simplicity is intentional. By using only bandwidth-derived costs, OSPF remains stable, predictable, and easy to troubleshoot. The trade-off is less sophisticated path selection. In practice, most enterprise networks manually tune costs to account for delay where necessary, accepting this administrative overhead for OSPF's operational simplicity.
While routing protocols use static delay values, network operations require accurate delay measurement for capacity planning, troubleshooting, and SLA verification. Understanding measurement techniques is essential for network engineers.
| Technique | What It Measures | Limitations | Use Case |
|---|---|---|---|
| Ping (ICMP Echo) | Round-trip time (RTT) | ICMP may be deprioritized; only RTT not one-way | Quick connectivity/delay check |
| Traceroute | Per-hop RTT estimates | ICMP deprioritization; load-balanced paths may show inconsistent results | Path analysis, hop-by-hop delay identification |
| IP SLA | Configurable synthetic probes | Adds probe traffic; measures probe path not actual traffic | Continuous monitoring, SLA verification |
| Two-Way Active Measurement (TWAMP) | Bidirectional delay measurement | Requires TWAMP support on both ends | Carrier-grade SLA measurement |
| Netflow/IPFIX with timestamps | Actual traffic timing | Timestamp resolution limits; processing overhead | Real traffic delay analysis |
12345678910111213141516171819202122
! Configure IP SLA to measure delay to remote hostip sla 1 icmp-echo 10.1.1.1 source-ip 10.0.0.1 frequency 60 ! Probe every 60 seconds timeout 5000 ! 5 second timeout threshold 100 ! Alert if RTT > 100ms ip sla schedule 1 start-time now life forever ! Configure tracking based on SLAtrack 1 ip sla 1 reachability ! View resultsshow ip sla statistics 1 ! Example output:! Round Trip Time (RTT) Index 1! Latest RTT: 45 milliseconds! Latest operation start time: 14:32:45 UTC! Number of successes: 1440! Number of failures: 0! Operation time to live: ForeverOne-Way vs Round-Trip Delay:
Most measurement tools report Round-Trip Time (RTT), which includes delay in both directions:
RTT = Forward Delay + Reverse Delay
If paths are symmetric: One-Way Delay ≈ RTT / 2
If paths are asymmetric: One-Way Delay ≠ RTT / 2
Asymmetric routing is common in the Internet, making RTT/2 an approximation at best. For precise one-way delay measurement, synchronized clocks (GPS, PTP) are required at both endpoints.
Delay Variation (Jitter):
Beyond absolute delay, delay variation (jitter) significantly impacts real-time applications:
Jitter = |Delay(packet n) - Delay(packet n-1)|
For voice/video:
- < 30ms jitter: Excellent
- 30-50ms jitter: Good
- > 50ms jitter: May cause quality issues
High jitter requires larger receive buffers, which add to end-to-end delay. This creates a trade-off between jitter tolerance and total delay.
Synthetic probes (ping, IP SLA) measure the network's response to test traffic. Passive measurement observes actual user traffic. Synthetic probes may not reflect real traffic experience if probe packets are treated differently (QoS, routing). Use both approaches for complete visibility.
We've thoroughly explored delay as a fundamental yet often underappreciated routing metric—from its physical components through protocol implementations and measurement techniques. Let's consolidate the essential knowledge:
Looking Ahead:
Delay captures time, but not quality. A link might have low delay but frequent errors requiring retransmissions. In the next page, we'll explore reliability as a routing metric—how protocols account for link stability, error rates, and the probability of successful packet delivery.
Together, bandwidth, delay, and reliability form the foundation for the composite metrics that modern protocols use to make nuanced routing decisions.
You now have comprehensive understanding of delay as a routing metric—its physical components, the distinction between static and dynamic delay values, protocol-specific implementations in EIGRP and OSPF, and practical measurement techniques. This knowledge is essential for designing networks that meet latency-sensitive application requirements.