Two-Way Communication
When Both Parties Want to Talk
In all the channel models we have studied, communication flows in one direction: from encoder to decoder (possibly with feedback assisting the encoder). But in many practical scenarios, two users simultaneously communicate with each other over the same channel. This is the two-way channel, introduced by Shannon in 1961.
The two-way channel is surprisingly tricky: each user's transmission creates interference for the other, but each user also knows its own transmitted signal and can subtract it from the received signal. This self-interference cancellation is automatic and free β the challenge lies in the coding and interaction structure. Can multiple rounds of interaction help? Shannon showed that interaction can indeed improve communication, making this one of the earliest examples of interactive communication protocols.
Definition: Shannon's Two-Way Channel
Shannon's Two-Way Channel
A two-way channel consists of:
- Two users, each with a message to send to the other: user 1 has for user 2, user 2 has for user 1.
- At each time : user 1 sends and receives ; user 2 sends and receives .
- The channel is described by .
The capacity region is the set of achievable rate pairs where is the rate of user 's message.
The two-way channel is inherently interactive: each user's encoding at time depends on its past observations , which in turn depend on the other user's past transmissions. This creates a feedback-like loop where interaction can improve rates.
Two-Way Channel
A communication channel where two users simultaneously send messages to each other. Each user encodes based on its message and past observations, creating an interactive communication protocol. Introduced by Shannon in 1961.
Related: Interactive Communication
Theorem: Shannon's Inner Bound for the Two-Way Channel
For the two-way channel , the following rate region is achievable:
This is the "independent coding" inner bound: each user encodes independently, and each user decodes by subtracting its own known transmission.
User 2 receives and knows its own input . After conditioning on , the effective channel from user 1 to user 2 is with known. The rate is exactly the capacity of this point-to-point channel. Similarly for the reverse direction.
The point is that with independent coding, the two-way channel decomposes into two one-way channels, each with known interference. The open question is whether interaction β adapting each user's encoding based on past observations β can do better.
Independent coding
Each user generates a random codebook independently. User sends without using any feedback.
Decoding with known interference
User 2 receives and knows (its own codeword). It looks for the unique such that is jointly typical. This succeeds if . Similarly for user 1.
Definition: Interactive Communication
Interactive Communication
In interactive communication, users take turns refining their messages over multiple rounds. In each round, each user transmits a signal that depends on:
- Its own message,
- All past observations (from previous rounds),
- (Optionally) a summary of what it has decoded so far.
Interaction allows each user to adapt its transmission based on what the other user has already sent, potentially achieving higher rates than one-shot (non-interactive) coding.
Formally, an -round interactive protocol divides the channel uses into rounds of uses each. In round , user encodes based on .
Interactive Communication
A communication protocol where users adapt their transmissions based on past observations, refining their messages over multiple rounds. Can potentially exceed the rates achievable by non-interactive (one-shot) coding.
Related: Two-Way Channel
Theorem: Outer Bound for the Two-Way Channel
For the two-way channel, the capacity region is contained in: for some joint distribution (not necessarily product).
When the outer bound with joint distributions strictly exceeds the inner bound with product distributions, interaction can potentially help.
The outer bound allows correlated inputs , which interaction can create. The inner bound restricts to independent inputs. If the gap is nonzero, there is room for interaction to improve rates.
Standard converse
By Fano's inequality: . Expanding: . Since can depend on , the effective input distribution can be a joint (non-product) distribution. This is why the outer bound allows joint distributions.
Example: The Binary Multiplying Two-Way Channel
Consider the binary multiplying two-way channel: (the output is the product of the two binary inputs). With independent coding, each user's rate is zero (since depends on both inputs but conditioning on one input reveals nothing if that input is random). Show that interaction achieves positive rates.
Independent coding fails
With independent Bernoulli(1/2) inputs: only when both (probability 1/4). : when , regardless of , so no information. When , , giving 1 bit. Average: bits. So Shannon's inner bound gives bits.
Interactive protocol
Round 1: User 1 sends (its message bit). User 2 sends . Both observe . Now user 2 knows . Round 2: User 2 sends . User 1 sends . Both observe . Now user 1 knows . Rate: each user sends 1 bit in 2 channel uses, giving bits.
Comparison
In this case, the interactive protocol achieves the same rate as the independent coding bound (0.5 bits each). The two-way channel capacity for this example is known to be (achieved by time-sharing). Interaction does not help beyond time-sharing here, but for other channels it can.
Example: The Gaussian Two-Way Channel
For the Gaussian two-way channel , with and power constraints , characterize the capacity region. Does interaction help?
Shannon's inner bound
With independent coding: , . This is a rectangle. Each user gets the full point-to-point capacity because self-interference is perfectly canceled ().
Outer bound
The outer bound with joint distributions: since does not depend on at all, . This matches the inner bound!
Conclusion
For the Gaussian two-way channel with additive noise (and no coupling between directions), the capacity region is: , . Interaction does not help because the channel decomposes into two independent point-to-point channels. Interaction helps only when the two directions are coupled through the channel.
Historical Note: Shannon's Last Great Open Problem
1961-presentShannon introduced the two-way channel in 1961, twenty years before the relay channel gained prominence. He established the inner and outer bounds but could not determine the exact capacity region. More than sixty years later, the capacity of the general two-way channel remains unknown β it is one of the oldest open problems in information theory.
What makes the two-way channel so hard is the interaction: the optimal encoding at each time step depends on the past observations, which depend on the other user's past encoding, which depends on their past observations, creating an infinite regress. This is fundamentally different from the one-way channel where the capacity formula involves a single-letter optimization. Whether the capacity of the two-way channel admits a single-letter characterization is itself an open question.
Historical Note: Interactive Communication in Computer Science
1979-1996The study of interactive communication has deep connections to theoretical computer science. Yao (1979) introduced the concept of communication complexity: how many bits must two parties exchange to compute a joint function of their inputs? This is the two-way channel problem in the noiseless setting.
The noisy version β interactive communication over noisy channels β was studied by Schulman (1996), who showed that interaction can be protected against noise using tree codes. The question of whether interaction helps over noisy channels, and by how much, connects information theory to computational complexity in surprising ways.
Shannon's Two-Way Channel
Summary: Feedback and Interaction Effects
| Channel | Does Feedback/Interaction Help? | Mechanism | Capacity Known? |
|---|---|---|---|
| Point-to-point DMC | No (capacity unchanged) | N/A | Yes (Shannon, 1956) |
| Point-to-point Gaussian | Capacity unchanged, but reliability improves (doubly exponential error) | Schalkwijk-Kailath iterative refinement | Yes |
| MAC (general) | Yes β capacity region enlarged | Input correlation via shared observations | Gaussian: Yes (Ozarow, 1984) |
| BC (degraded) | No | N/A | Yes |
| BC (non-degraded) | Yes β capacity region enlarged | Retransmission / XOR of missed information | Gaussian: partially (Shayevitz-Wigger) |
| Two-way channel | Interaction can help for coupled channels | Adaptive encoding based on past observations | Open in general |
Two-Way Channel: Inner and Outer Bounds
Visualize the inner bound (independent coding) and outer bound for the Gaussian two-way channel with varying power asymmetry.
Parameters
Common Mistake: The Two-Way Channel Does Not Always Decompose
Mistake:
Assuming that the two-way channel always decomposes into two independent one-way channels because each user knows its own input.
Correction:
The Gaussian additive two-way channel does decompose because does not depend on . But for general two-way channels where couples the two directions (e.g., ), the channel does not decompose. In such channels, interaction can create correlation between the inputs that improves rates beyond independent coding. The binary multiplying channel is an example where the outputs are coupled.
Quick Check
Shannon's two-way channel has been open for over 60 years. What makes the two-way channel fundamentally harder than one-way channels?
The encoding functions are more complex
The interaction creates dependencies between time steps that prevent single-letter characterization
There is no known converse technique for two-way channels
Two-way channels require quantum information theory
Correct. In one-way channels, the capacity is characterized by a single-letter formula β a one-dimensional optimization. In the two-way channel, the encoding at time depends on all past observations, which depend on all past encodings by both users. This creates a complex dependency structure that resists reduction to a single-letter formula.
Why This Matters: Two-Way Channels and Full-Duplex Wireless
The two-way channel is the information-theoretic model for full-duplex wireless communication, where two devices transmit and receive simultaneously on the same frequency band. Full-duplex is a major research direction in 5G and 6G, promising to double spectral efficiency compared to half-duplex (TDD/FDD).
The main practical challenge β self-interference cancellation β maps directly to the "known interference" structure of the two-way channel: each device knows its own transmitted signal and can (in principle) subtract it from the received signal. The information-theoretic results show that perfect self-interference cancellation allows both directions to achieve full point-to-point capacity simultaneously.
See Book telecom, Ch. 22 for full-duplex wireless system design.
Key Takeaway
The two-way channel models simultaneous bidirectional communication. With independent coding and known-interference cancellation, each direction achieves point-to-point capacity for channels that decouple (like Gaussian additive). For coupled channels, interaction may help, but the general capacity region remains one of the oldest open problems in information theory.