## Abstract

An additive noise channel is considered, in which the distribution of the noise is nonparametric and unknown. The problem of learning encoders and decoders based on noise samples is considered. For uncoded communication systems, the problem of choosing a codebook and possibly also a generalized minimal distance decoder (which is parameterized by a covariance matrix) is addressed. High probability generalization bounds for the error probability loss function, as well as for a hinge-type surrogate loss function are provided. A stochastic-gradient based alternating-minimization algorithm for the latter loss function is proposed. In addition, a Gibbs-based algorithm that gradually expurgates an initial codebook from codewords in order to obtain a smaller codebook with improved error probability is proposed, and bounds on its average empirical error and generalization error, as well as a high probability generalization bound, are stated. Various experiments demonstrate the performance of the proposed algorithms. For coded systems, the problem of maximizing the mutual information between the input and the output with respect to the input distribution is addressed, and uniform convergence bounds for two different classes of input distributions are obtained.

Original language | English |
---|---|

Pages (from-to) | 1886-1921 |

Number of pages | 36 |

Journal | IEEE Transactions on Information Theory |

Volume | 68 |

Issue number | 3 |

DOIs | |

State | Published - 1 Mar 2022 |

Externally published | Yes |

## Keywords

- Additive noise channels
- Alternating optimization algorithm
- Entropy estimation
- Expurgation
- Generalization bounds
- Gibbs algorithm
- Hinge loss
- Minimal distance decoding
- Mismatch decoding
- Statistical learning
- Stochastic gradient descent