Noise prediction for channels with side information at the transmitter

Uri Erez*, Ram Zamir

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

16 Scopus citations

Abstract

The computation of channel capacity with side information at the transmitter side (but not at the receiver side) requires, in general, extension of the input alphabet to a space of strategies, and is often hard. We consider the special case of a discrete memoryless modulo-additive noise channel Y = X + Zs, where the encoder observes causally the random state S G S that governs the distribution of the noise ZS. We show that the capacity of this channel is given by C = log |X| - mint:S→X H(ZS - t(S)). This capacity is realized by a state-independent code, followed by a shift by the "noise prediction" tmin(S) that minimizes the entropy of Zs -t(S). If the set of conditional noise distributions {p( z \s), s ∈ S} is such that the optimum predictor tmin(·) is independent of the state weights, then C is also the capacity for a noncausal encoder, that observes the entire state sequence in advance. Furthermore, for this case we also derive a simple formula for the capacity when the state process has memory.

Original languageEnglish
Pages (from-to)1610-1617
Number of pages8
JournalIEEE Transactions on Information Theory
Volume46
Issue number4
DOIs
StatePublished - 2000

Keywords

  • Optimum transmitter
  • Prediction with minimum error entropy
  • Side information
  • Time-varying channels

Fingerprint

Dive into the research topics of 'Noise prediction for channels with side information at the transmitter'. Together they form a unique fingerprint.

Cite this