Loading content...
Why do airlines overbook flights? Why do banks not keep cash equal to all deposits? Why do cellular networks sell more capacity than they have?
The answer in every case is the same fundamental insight: most resources are never used at full capacity simultaneously. By carefully analyzing usage patterns and accepting small probabilities of congestion, shared resources can serve far more users than dedicated alternatives at dramatically lower cost per user.
This principle—statistical multiplexing gain—is the quantitative foundation of modern communications. Understanding it explains why Internet access costs dollars per month instead of thousands, why mobile phones work despite limited spectrum, and why cloud computing is economically viable.
This page develops the mathematics and engineering behind efficiency gains, giving you tools to analyze and optimize shared communication systems.
By the end of this page, you will understand: the mathematical basis of statistical multiplexing gain, how to calculate efficiency for different traffic types, the tradeoff between efficiency and quality of service, real-world examples of efficiency optimization, and how to reason about capacity planning for shared systems.
Utilization is the fundamental efficiency metric: the fraction of channel capacity actually carrying useful data. Let's compare dedicated and shared approaches.
Dedicated Channel Utilization
In a dedicated channel system, each user receives exclusive capacity whether they're using it or not. Consider voice calls:
For bursty data applications, the situation is worse:
Shared Channel Utilization
When multiple sources share a channel, their independent utilization patterns combine favorably:
| Scenario | Dedicated System | Shared System | Improvement |
|---|---|---|---|
| Voice (35% activity factor) | 35% per line | 80%+ aggregate | 2.3x capacity |
| Office workstation (8% active) | 8% per port | 70% aggregate | 8.7x capacity |
| Web browsing (1% active transfer) | 1% per user | 60% aggregate | 60x capacity |
| IoT sensors (0.1% duty cycle) | 0.1% per device | 50% aggregate | 500x capacity |
The activity factor is the fraction of time a source is actively using capacity. Lower activity factors mean greater potential for multiplexing gain. Voice has ~35% activity; interactive data has ~1-5%; IoT sensors often have <1%. Applications with very low activity factors benefit most dramatically from statistical multiplexing.
The Mathematics of Aggregation
Let's formalize why aggregation improves utilization. Consider n independent sources, each with:
Dedicated System:
Shared System:
The shared system doesn't need capacity for all peaks simultaneously—only enough for the statistically likely maximum. As n grows large, this maximum becomes increasingly predictable.
Statistical multiplexing gain is the ratio of users a shared system can support compared to a dedicated system with the same total capacity. This gain derives from probability theory and grows with the number of sources.
Derivation Using Central Limit Theorem
Consider n independent, identically distributed sources, each with mean μ and variance σ². The aggregate load X is the sum:
X = X₁ + X₂ + ... + Xₙ
By the Central Limit Theorem, as n grows:
The coefficient of variation (relative variability) is: CV = StdDev[X] / E[X] = (√n × σ) / (n × μ) = σ / (μ × √n)
Crucially, relative variability decreases with √n. With 100 sources, CV is 10× lower than with 1 source. The aggregate becomes increasingly smooth and predictable.
The coefficient of variation decreasing as 1/√n is sometimes called the 'square root rule.' It means that to reduce relative variability by half, you need to aggregate 4× more sources. This explains why larger networks achieve better statistical multiplexing gains than smaller ones—more sources to average.
Calculating Required Capacity
To support n sources with shared capacity C, we need the probability of aggregate demand exceeding C to be below some acceptable threshold ε:
P(X > C) < ε
Using the normal approximation:
C = n × μ + z_ε × √n × σ
Where z_ε is the z-score corresponding to probability ε (e.g., z = 2.33 for ε = 1%).
Multiplexing Gain Calculation
The multiplexing gain G is the ratio of dedicated to shared capacity:
G = (n × R) / (n × μ + z_ε × √n × σ)
As n → ∞, the z_ε × √n × σ term becomes negligible compared to n × μ, and:
G → R / μ = 1 / a (the inverse of activity factor)
Example: 1000 Voice Sources
Dedicated: 1000 × 64 = 64,000 kbps = 64 Mbps Shared: 1000 × 22.4 + 2.33 × √1000 × 25 = 22,400 + 1,842 = 24,242 kbps ≈ 24.2 Mbps
Gain: 64 / 24.2 ≈ 2.6×
The shared system needs only 38% of the dedicated system's capacity while achieving 99% service probability.
| Number of Sources | Dedicated Capacity | Shared Capacity | Multiplexing Gain |
|---|---|---|---|
| 10 | 640 kbps | 340 kbps | 1.9× |
| 100 | 6.4 Mbps | 2.8 Mbps | 2.3× |
| 1,000 | 64 Mbps | 24.2 Mbps | 2.6× |
| 10,000 | 640 Mbps | 230 Mbps | 2.8× |
| 100,000 | 6.4 Gbps | 2.27 Gbps | 2.8× |
For circuit-switched systems like traditional telephony, the Erlang model provides precise tools for dimensioning shared resources. Developed by Agnar Erlang in the early 1900s, these formulas remain fundamental to telecommunications engineering.
Traffic Load Measurement: The Erlang
The Erlang is the standard unit of telecommunications traffic intensity:
Traffic A (in Erlangs) = λ × h, where:
Example: 600 calls per hour, average duration 3 minutes
Another unit for traffic is CCS (Cent Call Seconds): 100 call-seconds. Since an hour = 3600 seconds, 1 Erlang = 36 CCS. CCS is common in North American telephony; Erlangs are standard internationally. Both measure the same thing: occupancy.
Erlang B Formula: Blocking Probability
For systems where blocked calls are cleared (caller hangs up and may retry later), the Erlang B formula gives the probability of blocking:
P(blocking) = B(A, n) = (Aⁿ/n!) / Σₖ₌₀ⁿ (Aᵏ/k!)
Where:
This formula assumes:
Erlang C Formula: Waiting Probability
For systems where blocked calls wait (call center queues), the Erlang C formula gives the probability of waiting:
P(waiting > 0) = C(A, n) = [n × B(A, n)] / [n - A × (1 - B(A, n))]
This is used for dimensioning systems with queues, like call centers or packet networks.
Using Erlang Tables for Capacity Planning
Erlang tables (or calculators) determine required circuits for given traffic and blocking targets:
Example: Support 50 Erlangs with ≤2% blocking:
The gain increases with traffic volume. Higher Erlang loads allow proportionally fewer circuits per Erlang.
| Offered Traffic (Erlangs) | Circuits Required | Erlangs per Circuit | Gain vs. 1:1 |
|---|---|---|---|
| 10 | 17 | 0.59 | 1.7× |
| 30 | 42 | 0.71 | 2.1× |
| 50 | 62 | 0.81 | 2.5× |
| 100 | 115 | 0.87 | 2.9× |
| 500 | 527 | 0.95 | 3.2× |
| 1,000 | 1,030 | 0.97 | 3.4× |
Packet switching takes statistical multiplexing further than circuit switching by sharing capacity at the packet level rather than the call level. This enables dramatically higher efficiency for bursty traffic.
Circuit vs. Packet Efficiency
In circuit switching:
In packet switching:
The Bursty Data Advantage
Data applications are far more bursty than voice:
Packet switching handles this burstiness naturally. A 100 Mbps link serving 1000 users at 1% average activity can let any user burst to 100 Mbps when idle users aren't transmitting.
| Traffic Type | Activity Factor | Circuit Efficiency | Packet Efficiency | Packet Advantage |
|---|---|---|---|---|
| Voice call | 35% | 35% | 35%* | 1× |
| Video conference | 50% | 50% | 50%* | 1× |
| Web browsing | 2% | 2% | 70% | 35× |
| Email/messaging | 0.5% | 0.5% | 60% | 120× |
| IoT sensor | 0.1% | 0.1% | 50% | 500× |
Voice over IP can achieve higher efficiency than circuit-switched voice by using silence suppression—not sending packets during silence periods. This can push voice efficiency from 35% to 50-60%, at the cost of potential clipping if voice activity detection makes mistakes.
Queuing Theory for Packet Networks
Packet network efficiency analysis uses queuing theory. The key relationship is the utilization factor:
ρ = λ / μ
Where:
For stability, ρ must be < 1 (arrivals must be slower than service on average).
Wait Time and Utilization Tradeoff
For M/M/1 queues (Poisson arrivals, exponential service, single server), average waiting time W is:
W = 1 / (μ - λ) = (1/μ) / (1 - ρ)
As ρ approaches 1:
This nonlinear relationship explains why packet networks can't run at 100% utilization—queuing delays would become unbounded. Practical targets are 70-80% utilization to balance efficiency against latency.
Statistical Multiplexing in Practice
Consider a 1 Gbps link serving 10,000 users:
If all users transmitted simultaneously (all 10% activity aligned), the link would be overwhelmed. But statistically, this never happens—the law of large numbers ensures aggregate demand stays near the 1 Gbps average with high probability.
Efficiency gains translate directly to cost savings. Understanding this relationship helps justify investment in multiplexing infrastructure and informs pricing models.
Capital Cost Savings
Multiplexing reduces infrastructure investment proportionally to multiplexing gain:
Example: Metropolitan Fiber Network
| Configuration | Fiber Strands | Equipment Cost | Total Cost |
|---|---|---|---|
| Dedicated 10G per customer (1000 customers) | 1,000 | $50M | $75M |
| Shared 100G with 10:1 mux (1000 customers) | 10 | $5M | $8M |
| Savings | 99% | 90% | 89% |
The multiplexed approach costs 1/9th as much while serving the same customers equally well (for typical usage patterns).
| Cost Component | Dedicated Approach | Multiplexed Approach | Reduction |
|---|---|---|---|
| Physical infrastructure | $500M | $50M | 90% |
| Network equipment | $200M | $30M | 85% |
| Power/cooling (annual) | $10M | $2M | 80% |
| Maintenance (annual) | $5M | $1M | 80% |
| Total 5-year TCO | $775M | $95M | 88% |
Pricing Implications
Multiplexing efficiency enables consumption-based and shared pricing models:
Committed Information Rate (CIR): Customers purchase guaranteed capacity less than their peak access rate, paying premium only for firm guarantees.
Burst Pricing: Customers pay base rate for normal usage, premium for occasional bursts—matching cost to actual resource consumption.
Best-Effort/Oversubscription: Residential Internet is highly oversubscribed (often 20-50:1 on aggregation links) because most users are inactive most of the time.
Cloud Computing Economics
Cloud providers apply multiplexing principles to compute, storage, and networking:
Aggressive oversubscription increases risk. If usage patterns change (everyone works from home, major event causes simultaneous demand), the statistical assumptions fail. The 2020 pandemic revealed oversubscription limits as residential networks, designed for 5% concurrent use, faced 50% concurrent use. Responsible engineering includes margins for unexpected load.
Higher efficiency comes at the cost of statistical guarantees. Understanding this tradeoff is essential for system design that balances cost against user experience.
The Fundamental Tradeoff
Shared resources achieve efficiency by pooling unused capacity from idle users. But this means:
Quantifying the Tradeoff
For a given multiplexing gain G and number of users n, there's an associated congestion probability P_c:
Service Level Targets
Different applications tolerate different congestion levels:
Quality of Service Mechanisms
To serve multiple service levels on shared infrastructure, QoS mechanisms provide differentiated treatment:
Priority Queuing: Critical traffic (voice, video) gets served before bulk traffic (downloads, backups). During congestion, low-priority traffic waits while high-priority traffic proceeds.
Weighted Fair Queuing: Each class gets proportional share of capacity. Business traffic might get 70% weight, residential 30%—so business users experience less congestion.
Rate Limiting: Each user's traffic is constrained to purchased capacity despite sharing the aggregate link. Prevents any user from monopolizing shared resources.
Admission Control: For circuit-like services (voice), new requests are rejected if accepting them would overload the network. Protects existing sessions at the cost of blocking new ones.
Often, 20% of users generate 80% of traffic. QoS mechanisms can constrain heavy users during congestion, reserving capacity for the majority. This makes high oversubscription ratios practical—the few heavy users are rate-limited, the many light users experience no impact.
Let's examine how these efficiency principles apply in actual systems.
Example 1: Residential Internet Service
A cable ISP serving 100,000 homes:
This works because:
Result: Infrastructure costs 1/250th of dedicated approach, sufficient for 99%+ of situations.
Example 2: Cellular Network
A 5G cell sector serving 1,000 active devices:
Multiplexing gain: 100× This enables mobile broadband at consumer prices despite spectrum scarcity.
Example 3: Cloud Data Center
A cloud provider with 10,000 VMs on 1,000 servers:
Most VMs are idle or low-bandwidth. The few high-bandwidth VMs statistically spread across servers. Total efficiency allows 10,000 × 1 Gbps (10 Tbps) of virtual capacity on ~250 Gbps of physical backbone.
| System | Peak-to-Actual Ratio | Achieved Utilization | Cost Reduction |
|---|---|---|---|
| Residential Cable ISP | 250:1 | ~75% | 99.6% |
| Enterprise WAN | 10:1 | ~60% | 90% |
| 5G Cellular Sector | 100:1 | ~50% | 99% |
| Cloud Data Center Network | 40:1 | ~40% | 97.5% |
| Submarine Cable System | 5:1 | ~85% | 80% |
Submarine cables have modest oversubscription because capacity is extremely expensive ($200-400M per cable) and demand is aggregated from entire continents. Traffic is already smoothed across millions of users before reaching the cable. High utilization (85%+) is essential for ROI, and backup cables handle failures rather than statistical headroom.
We've developed the quantitative framework for understanding multiplexing efficiency. Let's consolidate the key insights:
What's Next:
With the efficiency gains of multiplexing established, the next page explores the types of multiplexing—the specific techniques (FDM, TDM, WDM, CDM, statistical) that implement channel sharing. Each technique offers different tradeoffs suitable for different applications.
You now understand the quantitative framework for multiplexing efficiency—the mathematical and economic foundations that make modern communications affordable. This knowledge lets you reason about capacity planning, pricing, and the tradeoffs between efficiency and quality of service.