TY - JOUR
T1 - On the optimal boolean function for prediction under quadratic loss
AU - Weinberger, Nir
AU - Shayevitz, Ofer
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/7
Y1 - 2017/7
N2 - Suppose Yn is obtained by observing a uniform Bernoulli random vector Xn through a binary symmetric channel. Courtade and Kumar asked how large the mutual information between Yn and a Boolean function b(Xn) could be, and conjectured that the maximum is attained by a dictator function. An equivalent formulation of this conjecture is that dictator minimizes the prediction cost in a sequential prediction of Yn under logarithmic loss, given b(Xn). In this paper, we study the question of minimizing the sequential prediction cost under a different (proper) loss function the quadratic loss. In the noiseless case, we show that majority asymptotically minimizes this prediction cost among all Boolean functions.We further show that for weak noise, majority is better than a dictator, and that for a strong noise dictator outperforms majority. We conjecture that for quadratic loss, there is no single sequence of Boolean functions that is simultaneously (asymptotically) optimal at all noise levels.
AB - Suppose Yn is obtained by observing a uniform Bernoulli random vector Xn through a binary symmetric channel. Courtade and Kumar asked how large the mutual information between Yn and a Boolean function b(Xn) could be, and conjectured that the maximum is attained by a dictator function. An equivalent formulation of this conjecture is that dictator minimizes the prediction cost in a sequential prediction of Yn under logarithmic loss, given b(Xn). In this paper, we study the question of minimizing the sequential prediction cost under a different (proper) loss function the quadratic loss. In the noiseless case, we show that majority asymptotically minimizes this prediction cost among all Boolean functions.We further show that for weak noise, majority is better than a dictator, and that for a strong noise dictator outperforms majority. We conjecture that for quadratic loss, there is no single sequence of Boolean functions that is simultaneously (asymptotically) optimal at all noise levels.
KW - Boolean functions
KW - Logarithmic loss function
KW - Pinsker's inequality
KW - Quadratic loss function
KW - Sequential prediction
UR - http://www.scopus.com/inward/record.url?scp=85025430794&partnerID=8YFLogxK
U2 - 10.1109/TIT.2017.2686437
DO - 10.1109/TIT.2017.2686437
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85025430794
SN - 0018-9448
VL - 63
SP - 4202
EP - 4217
JO - IEEE Transactions on Information Theory
JF - IEEE Transactions on Information Theory
IS - 7
M1 - 7886288
ER -