TY - GEN
T1 - Refined Convergence Rates of the Good-Turing Estimator
AU - Painsky, Amichai
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - The Good-Turing (GT) estimator is perhaps the most popular framework for modelling large alphabet distributions. Classical results show that the GT estimator convergences to the occupancy probability, formally defined as the total probability of words that appear exactly k times in the sample. In this work we introduce new convergence guarantees for the GT estimator, based on worst-case MSE analysis. Our results refine and improve upon currently known bounds. Importantly, we introduce a simultaneous convergence rate to the entire collection of occupancy probabilities.
AB - The Good-Turing (GT) estimator is perhaps the most popular framework for modelling large alphabet distributions. Classical results show that the GT estimator convergences to the occupancy probability, formally defined as the total probability of words that appear exactly k times in the sample. In this work we introduce new convergence guarantees for the GT estimator, based on worst-case MSE analysis. Our results refine and improve upon currently known bounds. Importantly, we introduce a simultaneous convergence rate to the entire collection of occupancy probabilities.
UR - http://www.scopus.com/inward/record.url?scp=85123409455&partnerID=8YFLogxK
U2 - 10.1109/ITW48936.2021.9611389
DO - 10.1109/ITW48936.2021.9611389
M3 - ???researchoutput.researchoutputtypes.contributiontobookanthology.conference???
AN - SCOPUS:85123409455
T3 - 2021 IEEE Information Theory Workshop, ITW 2021 - Proceedings
BT - 2021 IEEE Information Theory Workshop, ITW 2021 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 17 October 2021 through 21 October 2021
ER -