TY - JOUR
T1 - Agnostically learning halfspaces
AU - Kalai, Adam Tauman
AU - Klivans, Adam R.
AU - Mansour, Yishay
AU - Servedio, Rocco A.
PY - 2007
Y1 - 2007
N2 - We give a computationally efficient algorithm that learns (under distributional assumptions) a halfspace in the difficult agnostic framework of Kearns, Schapire, and Sellie [Mach. Learn., 17 (1994), pp. 115-141], where a learner is given access to a distribution on labelled examples but where the labelling may be arbitrary (similar to malicious noise). It constructs a hypothesis whose error rate on future examples is within an additive ε{lunate} of the optimal halfspace, in time poly(n) for any constant ε{lunate} > 0, for the uniform distribution over {-1, 1}n or unit sphere in ℝn, as well as any log-concave distribution in ℝn. It also agnostically learns Boolean disjunctions in time 2Õ(√n) with respect to any distribution. Our algorithm, which performs L1 polynomial regression, is a natural noise-tolerant arbitrary-distribution generalization of the well-known "low-degree" Fourier algorithm of Linial, Mansour, and Nisan. We observe that significant improvements on the running time of our algorithm would yield the fastest known algorithm for learning parity with noise, a challenging open problem in computational learning theory.
AB - We give a computationally efficient algorithm that learns (under distributional assumptions) a halfspace in the difficult agnostic framework of Kearns, Schapire, and Sellie [Mach. Learn., 17 (1994), pp. 115-141], where a learner is given access to a distribution on labelled examples but where the labelling may be arbitrary (similar to malicious noise). It constructs a hypothesis whose error rate on future examples is within an additive ε{lunate} of the optimal halfspace, in time poly(n) for any constant ε{lunate} > 0, for the uniform distribution over {-1, 1}n or unit sphere in ℝn, as well as any log-concave distribution in ℝn. It also agnostically learns Boolean disjunctions in time 2Õ(√n) with respect to any distribution. Our algorithm, which performs L1 polynomial regression, is a natural noise-tolerant arbitrary-distribution generalization of the well-known "low-degree" Fourier algorithm of Linial, Mansour, and Nisan. We observe that significant improvements on the running time of our algorithm would yield the fastest known algorithm for learning parity with noise, a challenging open problem in computational learning theory.
KW - Agnostic learning
KW - Fourier
KW - Halfspaces
UR - http://www.scopus.com/inward/record.url?scp=55249114173&partnerID=8YFLogxK
U2 - 10.1137/060649057
DO - 10.1137/060649057
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:55249114173
SN - 0097-5397
VL - 37
SP - 1777
EP - 1805
JO - SIAM Journal on Computing
JF - SIAM Journal on Computing
IS - 6
ER -