TY - JOUR
T1 - Using more data to speed-up training time
AU - Shalev-Shwartz, Shai
AU - Shamir, Ohad
AU - Tromer, Eran
N1 - Publisher Copyright:
© Copyright 2012 by the authors.
PY - 2012
Y1 - 2012
N2 - In many recent applications, data is plentiful. By now, we have a rather clear understanding of how more data can be used to improve the accuracy of learning algorithms. Recently, there has been a growing interest in understanding how more data can be leveraged to reduce the required training runtime. In this paper, we study the runtime of learning as a function of the number of available training examples, and underscore the main highlevel techniques. We provide the first formal positive result showing that even in the unrealizable case, the runtime can decrease exponentially while only requiring a polynomial growth of the number of examples. Our construction corresponds to a synthetic learning problem and an interesting open question is whether the tradeoff can be shown for more natural learning problems. We spell out several interesting candidates of natural learning problems for which we conjecture that there is a tradeoff between computational and sample complexity.
AB - In many recent applications, data is plentiful. By now, we have a rather clear understanding of how more data can be used to improve the accuracy of learning algorithms. Recently, there has been a growing interest in understanding how more data can be leveraged to reduce the required training runtime. In this paper, we study the runtime of learning as a function of the number of available training examples, and underscore the main highlevel techniques. We provide the first formal positive result showing that even in the unrealizable case, the runtime can decrease exponentially while only requiring a polynomial growth of the number of examples. Our construction corresponds to a synthetic learning problem and an interesting open question is whether the tradeoff can be shown for more natural learning problems. We spell out several interesting candidates of natural learning problems for which we conjecture that there is a tradeoff between computational and sample complexity.
UR - https://www.scopus.com/pages/publications/105012415466
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.conferencearticle???
AN - SCOPUS:105012415466
SN - 1532-4435
VL - 22
SP - 1019
EP - 1027
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
T2 - 15th International Conference on Artificial Intelligence and Statistics, AISTATS 2012
Y2 - 21 April 2012 through 23 April 2012
ER -