Network Information Theory
Overview
Network information theory extends Shannon's point-to-point framework to multi-terminal settings: multiple senders, multiple receivers, or both. Unlike the single-user case, many multi-user capacity problems remain open. The field reveals fundamental tensions between cooperation, interference, and distributed operation.
Multiple Access Channel (MAC)
Multiple senders transmit to a single receiver. For a two-user DM-MAC with inputs and output :
Capacity region (achievable rate pairs):
over all product distributions (no cooperation between encoders). The region is a pentagon (or degenerate cases).
Key properties:
- Corner points achieved by successive decoding: decode one user treating the other as noise, subtract, decode the second
- Time-sharing between corner points achieves the dominant face
- Capacity region is convex (time-sharing achievable)
- Sum capacity
Gaussian MAC: , , power constraints :
Successive interference cancellation (SIC): practical scheme achieving corner points. Decode strongest user first, subtract from received signal, decode next. Foundation of NOMA in 5G.
Broadcast Channel (BC)
One sender transmits to multiple receivers, each seeing a different channel. For a two-user degraded BC (where forms a Markov chain):
Superposition coding achieves capacity:
where is an auxiliary random variable representing the "cloud center" and given represents the "satellite codeword."
Gaussian BC: , stronger user gets higher-rate information layered on top of weaker user's message. Capacity region characterized by power splitting parameter :
MAC-BC duality: the capacity region of the Gaussian BC equals the capacity region of the dual MAC with a sum power constraint. This greatly simplifies computation.
Non-degraded BC: general capacity region unknown. Marton's inner bound (superposition + binning) is the best known achievable region. Converse remains open except for specific cases.
Relay Channel
Three-terminal channel: sender relay receiver, with a direct link from sender to receiver.
Key strategies:
- Decode-and-forward (DF): relay fully decodes, re-encodes, and forwards. Optimal when relay channel is strong. Rate:
- Compress-and-forward (CF): relay quantizes its observation and forwards compressed version. Useful when relay channel is weak. Uses Wyner-Ziv compression
- Amplify-and-forward (AF): relay simply scales and retransmits (Gaussian channels). Simplest but suboptimal
Cut-set bound (upper bound): . DF achieves this for degraded relay channels.
The capacity of the general relay channel remains open. The gap between DF/CF achievable rates and the cut-set bound can be significant.
Multi-hop and diamond networks: extensions with multiple relays in series or parallel. Noisy network coding (Lim, Kim, El Gamal, Chung, 2011) unifies and improves upon CF for general relay networks.
Interference Channel
Two sender-receiver pairs, each causing interference to the other:
Capacity: known only in special cases:
- Strong interference ( and ): each receiver can decode both messages. Capacity = intersection of two MAC regions
- Very strong interference: interference so strong that decoding it is "free." Capacity equals interference-free capacity
- General case: capacity is unknown (one of the major open problems)
Han-Kobayashi scheme: each transmitter splits its message into common and private parts. Common parts are decoded by both receivers; private parts treated as noise by unintended receiver. Best known achievable region. Achieved within 1 bit of capacity for all Gaussian ICs (Etkin, Tse, Wang, 2008).
Interference alignment (Cadambe-Jafar, 2008): in -user interference channel with time/frequency-varying coefficients, each user can achieve DoF. Total degrees of freedom, double what treating interference as noise achieves. Requires infinite symbol extensions or sufficiently varying channels.
Wiretap Channel
Sender communicates to legitimate receiver while keeping message secret from eavesdropper. Models: where is the legitimate channel output and is the eavesdropper's observation.
Secrecy capacity (Wyner, 1975; Csiszar-Korner, 1978):
For degraded channels (): .
Gaussian wiretap: when (main channel better than eavesdropper). If , secrecy capacity is zero.
Information-theoretic security vs. computational security:
- Provides unconditional security (no assumptions about eavesdropper's computational power)
- Secret key agreement via public discussion (Maurer, Ahlswede-Csiszar)
- Connection to quantum key distribution (BB84)
Network Coding
Classical routing: intermediate nodes only store-and-forward packets. Network coding allows intermediate nodes to combine (encode) packets.
Linear Network Coding
Nodes perform linear operations over a finite field GF(). Each outgoing edge carries a linear combination of incoming messages.
Max-flow min-cut theorem for networks (Ahlswede, Cai, Li, Yeung, 2000): for a single multicast session, the capacity equals the minimum of the max-flows to each receiver. Network coding achieves this; routing generally does not.
Butterfly network: canonical example where network coding provides throughput gain. Two multicast receivers each need two source symbols. Routing achieves rate 1.5; network coding achieves rate 2.
Algebraic framework (Koetter-Medard, 2003): encode using transfer matrices over GF(). For multicast, a random linear code over sufficiently large field achieves capacity with high probability.
Random Linear Network Coding
Each node selects random coefficients for linear combinations over GF().
- Achieves multicast capacity with probability where is number of receivers
- Decentralized: nodes need no global knowledge
- Practical protocol: append encoding vector as header. Receiver collects innovative (linearly independent) packets, solves via Gaussian elimination
- Overhead: header size grows logarithmically in field size
- Applications: P2P content distribution, wireless broadcast, disruption-tolerant networks
Practical Considerations
- Batched network coding (BATS): Fountain-like codes adapted for multi-hop networks with recoding
- Fulcrum codes: bridge between random network coding and Raptor codes
- Coded caching (Maddah-Ali, Niesen): information-theoretic formulation of caching with coded multicast delivery. Achieves multiplicative gain in delivery rate
Distributed Source Coding
CEO Problem
agents each observe a noisy version of a source . A central estimator (CEO) must reconstruct from the agents' compressed messages.
- Quadratic Gaussian CEO: rate-distortion function known (Prabhakaran, Tse, Ramchandran). As , achieves centralized distortion
- General case: optimal rate-distortion region open beyond two terminals
Multi-Terminal Source Coding
correlated sources encoded separately, decoded jointly:
- Lossless: Slepian-Wolf region for all
- Lossy: Berger-Tung inner bound (random binning + joint typicality). Tight for some cases (Gaussian, specific distortion measures)
Information-Theoretic Limits for Wireless
MIMO Channel
with transmit and receive antennas.
- Capacity (known at receiver):
- Water-filling over singular values of when CSI at transmitter
- Degrees of freedom: spatial multiplexing gain
- Massive MIMO: as , channel hardens, interference averages out, simple matched-filter processing near-optimal
Fading Channels
- Ergodic capacity: (average over fading realizations)
- Outage capacity:
- Diversity-multiplexing tradeoff (Zheng-Tse, 2003): fundamental tradeoff between reliability (diversity gain ) and rate (multiplexing gain ) in high-SNR regime
Open Problems
Several fundamental problems remain unsolved:
- Capacity of the general interference channel
- Capacity of the general broadcast channel (non-degraded)
- Capacity of the general relay channel
- Distributed lossy source coding (general rate-distortion region)
- Optimality of Gaussian signaling for vector Gaussian networks beyond known special cases
- Separation of source and channel coding in networks (does not hold in general, unlike point-to-point)
These open problems have driven decades of research and continue to motivate new techniques, including connections to theoretical computer science (hardness of approximation) and algebraic geometry.