Two-Way Communication

When Both Parties Want to Talk

In all the channel models we have studied, communication flows in one direction: from encoder to decoder (possibly with feedback assisting the encoder). But in many practical scenarios, two users simultaneously communicate with each other over the same channel. This is the two-way channel, introduced by Shannon in 1961.

The two-way channel is surprisingly tricky: each user's transmission creates interference for the other, but each user also knows its own transmitted signal and can subtract it from the received signal. This self-interference cancellation is automatic and free β€” the challenge lies in the coding and interaction structure. Can multiple rounds of interaction help? Shannon showed that interaction can indeed improve communication, making this one of the earliest examples of interactive communication protocols.

Definition:

Shannon's Two-Way Channel

A two-way channel consists of:

  • Two users, each with a message to send to the other: user 1 has W1W_1 for user 2, user 2 has W2W_2 for user 1.
  • At each time ii: user 1 sends X1,i=f1,i(W1,Y1iβˆ’1)X_{1,i} = f_{1,i}(W_1, Y_1^{i-1}) and receives Y1,iY_{1,i}; user 2 sends X2,i=f2,i(W2,Y2iβˆ’1)X_{2,i} = f_{2,i}(W_2, Y_2^{i-1}) and receives Y2,iY_{2,i}.
  • The channel is described by p(y1,y2∣x1,x2)p(y_1, y_2 | x_1, x_2).

The capacity region is the set of achievable rate pairs (R1,R2)(R_{1}, R_{2}) where RkR_{k} is the rate of user kk's message.

The two-way channel is inherently interactive: each user's encoding at time ii depends on its past observations Ykiβˆ’1Y_k^{i-1}, which in turn depend on the other user's past transmissions. This creates a feedback-like loop where interaction can improve rates.

Two-Way Channel

A communication channel where two users simultaneously send messages to each other. Each user encodes based on its message and past observations, creating an interactive communication protocol. Introduced by Shannon in 1961.

Related: Interactive Communication

Theorem: Shannon's Inner Bound for the Two-Way Channel

For the two-way channel p(y1,y2∣x1,x2)p(y_1, y_2 | x_1, x_2), the following rate region is achievable: Rinner=⋃p(x1)p(x2){(R1,R2):R1≀I(X1;Y2∣X2)R2≀I(X2;Y1∣X1)}.\mathcal{R}_{\text{inner}} = \bigcup_{p(x_1) p(x_2)} \left\{(R_{1}, R_{2}): \begin{array}{l} R_{1} \leq I(X_1; Y_2 | X_2) \\ R_{2} \leq I(X_2; Y_1 | X_1) \end{array}\right\}.

This is the "independent coding" inner bound: each user encodes independently, and each user decodes by subtracting its own known transmission.

User 2 receives Y2Y_2 and knows its own input X2X_2. After conditioning on X2X_2, the effective channel from user 1 to user 2 is p(y2∣x1,x2)p(y_2 | x_1, x_2) with X2X_2 known. The rate I(X1;Y2∣X2)I(X_1; Y_2 | X_2) is exactly the capacity of this point-to-point channel. Similarly for the reverse direction.

The point is that with independent coding, the two-way channel decomposes into two one-way channels, each with known interference. The open question is whether interaction β€” adapting each user's encoding based on past observations β€” can do better.

Definition:

Interactive Communication

In interactive communication, users take turns refining their messages over multiple rounds. In each round, each user transmits a signal that depends on:

  • Its own message,
  • All past observations (from previous rounds),
  • (Optionally) a summary of what it has decoded so far.

Interaction allows each user to adapt its transmission based on what the other user has already sent, potentially achieving higher rates than one-shot (non-interactive) coding.

Formally, an LL-round interactive protocol divides the nn channel uses into LL rounds of n/Ln/L uses each. In round β„“\ell, user kk encodes based on (Wk,Yk(β„“βˆ’1)n/L)(W_k, Y_k^{(\ell-1)n/L}).

Interactive Communication

A communication protocol where users adapt their transmissions based on past observations, refining their messages over multiple rounds. Can potentially exceed the rates achievable by non-interactive (one-shot) coding.

Related: Two-Way Channel

Theorem: Outer Bound for the Two-Way Channel

For the two-way channel, the capacity region is contained in: RouterβŠ‡{(R1,R2):R1≀I(X1;Y2∣X2)R2≀I(X2;Y1∣X1)}\mathcal{R}_{\text{outer}} \supseteq \left\{(R_{1}, R_{2}): \begin{array}{l} R_{1} \leq I(X_1; Y_2 | X_2) \\ R_{2} \leq I(X_2; Y_1 | X_1) \end{array}\right\} for some joint distribution p(x1,x2)p(x_1, x_2) (not necessarily product).

When the outer bound with joint distributions strictly exceeds the inner bound with product distributions, interaction can potentially help.

The outer bound allows correlated inputs (X1,X2)(X_1, X_2), which interaction can create. The inner bound restricts to independent inputs. If the gap is nonzero, there is room for interaction to improve rates.

Example: The Binary Multiplying Two-Way Channel

Consider the binary multiplying two-way channel: Y1=Y2=X1β‹…X2Y_1 = Y_2 = X_1 \cdot X_2 (the output is the product of the two binary inputs). With independent coding, each user's rate is zero (since YkY_k depends on both inputs but conditioning on one input reveals nothing if that input is random). Show that interaction achieves positive rates.

Example: The Gaussian Two-Way Channel

For the Gaussian two-way channel Y1=X2+Z1Y_1 = X_2 + Z_1, Y2=X1+Z2Y_2 = X_1 + Z_2 with Zk∼N(0,N)Z_k \sim \mathcal{N}(0, N) and power constraints P1,P2P_1, P_2, characterize the capacity region. Does interaction help?

Historical Note: Shannon's Last Great Open Problem

1961-present

Shannon introduced the two-way channel in 1961, twenty years before the relay channel gained prominence. He established the inner and outer bounds but could not determine the exact capacity region. More than sixty years later, the capacity of the general two-way channel remains unknown β€” it is one of the oldest open problems in information theory.

What makes the two-way channel so hard is the interaction: the optimal encoding at each time step depends on the past observations, which depend on the other user's past encoding, which depends on their past observations, creating an infinite regress. This is fundamentally different from the one-way channel where the capacity formula involves a single-letter optimization. Whether the capacity of the two-way channel admits a single-letter characterization is itself an open question.

Historical Note: Interactive Communication in Computer Science

1979-1996

The study of interactive communication has deep connections to theoretical computer science. Yao (1979) introduced the concept of communication complexity: how many bits must two parties exchange to compute a joint function of their inputs? This is the two-way channel problem in the noiseless setting.

The noisy version β€” interactive communication over noisy channels β€” was studied by Schulman (1996), who showed that interaction can be protected against noise using tree codes. The question of whether interaction helps over noisy channels, and by how much, connects information theory to computational complexity in surprising ways.

Shannon's Two-Way Channel

Animated overview of the two-way channel: two users simultaneously transmit and receive. Shows how the Gaussian additive two-way channel decomposes into two independent point-to-point channels via self-interference cancellation.

Summary: Feedback and Interaction Effects

ChannelDoes Feedback/Interaction Help?MechanismCapacity Known?
Point-to-point DMCNo (capacity unchanged)N/AYes (Shannon, 1956)
Point-to-point GaussianCapacity unchanged, but reliability improves (doubly exponential error)Schalkwijk-Kailath iterative refinementYes
MAC (general)Yes β€” capacity region enlargedInput correlation via shared observationsGaussian: Yes (Ozarow, 1984)
BC (degraded)NoN/AYes
BC (non-degraded)Yes β€” capacity region enlargedRetransmission / XOR of missed informationGaussian: partially (Shayevitz-Wigger)
Two-way channelInteraction can help for coupled channelsAdaptive encoding based on past observationsOpen in general

Two-Way Channel: Inner and Outer Bounds

Visualize the inner bound (independent coding) and outer bound for the Gaussian two-way channel with varying power asymmetry.

Parameters
10
10
1

Common Mistake: The Two-Way Channel Does Not Always Decompose

Mistake:

Assuming that the two-way channel always decomposes into two independent one-way channels because each user knows its own input.

Correction:

The Gaussian additive two-way channel does decompose because Y1=X2+Z1Y_1 = X_2 + Z_1 does not depend on X1X_1. But for general two-way channels where p(y1,y2∣x1,x2)p(y_1, y_2 | x_1, x_2) couples the two directions (e.g., Y1=f(X1,X2,Z)Y_1 = f(X_1, X_2, Z)), the channel does not decompose. In such channels, interaction can create correlation between the inputs that improves rates beyond independent coding. The binary multiplying channel is an example where the outputs are coupled.

Quick Check

Shannon's two-way channel has been open for over 60 years. What makes the two-way channel fundamentally harder than one-way channels?

The encoding functions are more complex

The interaction creates dependencies between time steps that prevent single-letter characterization

There is no known converse technique for two-way channels

Two-way channels require quantum information theory

Why This Matters: Two-Way Channels and Full-Duplex Wireless

The two-way channel is the information-theoretic model for full-duplex wireless communication, where two devices transmit and receive simultaneously on the same frequency band. Full-duplex is a major research direction in 5G and 6G, promising to double spectral efficiency compared to half-duplex (TDD/FDD).

The main practical challenge β€” self-interference cancellation β€” maps directly to the "known interference" structure of the two-way channel: each device knows its own transmitted signal and can (in principle) subtract it from the received signal. The information-theoretic results show that perfect self-interference cancellation allows both directions to achieve full point-to-point capacity simultaneously.

See Book telecom, Ch. 22 for full-duplex wireless system design.

Key Takeaway

The two-way channel models simultaneous bidirectional communication. With independent coding and known-interference cancellation, each direction achieves point-to-point capacity for channels that decouple (like Gaussian additive). For coupled channels, interaction may help, but the general capacity region remains one of the oldest open problems in information theory.