TY - JOUR
T1 - Approximate modularity revisited
AU - Feige, Uriel
AU - Feldman, Michal
AU - Talgam-Cohen, Inbal
N1 - Publisher Copyright:
© 2020 Society for Industrial and Applied Mathematics.
PY - 2020
Y1 - 2020
N2 - Set functions with convenient properties (such as submodularity) appear in application areas of current interest, such as algorithmic game theory, and allow for improved optimization algorithms. It is natural to ask (e.g., in the context of data driven optimization) how robust such properties are, and whether small deviations from them can be tolerated. We consider two such questions in the important special case of linear set functions. One question that we address is whether any set function that approximately satisfies the modularity equation (linear functions satisfy the modularity equation exactly) is close to a linear function. The answer to this is positive (in a precise formal sense) as shown by Kalton and Roberts [Trans. Amer. Math. Soc., 278 (1983), pp. 803-816] (and further improved by Bondarenko, Prymak, and Radchenko [J. Math. Anal. Appl., 402 (2013), pp. 234-241]). We revisit their proof idea that is based on expander graphs and provide significantly stronger upper bounds by combining it with new techniques. Furthermore, we provide improved lower bounds for this problem. Another question that we address is that of how to learn a linear function h that is close to an approximately linear function f, while querying the value of f on only a small number of sets. We present a deterministic algorithm that makes only linearly many (in the number of items) nonadaptive queries, and thus improve upon a previous algorithm of Chierichetti, Das, Dasgupta, and Kumar [Proceedings of the 56th Symposium on Foundations of Computer Science, 2015, pp. 1143-1162] that is randomized and makes more than a quadratic number of queries. Our learning algorithm is based on the Hadamard transform.
AB - Set functions with convenient properties (such as submodularity) appear in application areas of current interest, such as algorithmic game theory, and allow for improved optimization algorithms. It is natural to ask (e.g., in the context of data driven optimization) how robust such properties are, and whether small deviations from them can be tolerated. We consider two such questions in the important special case of linear set functions. One question that we address is whether any set function that approximately satisfies the modularity equation (linear functions satisfy the modularity equation exactly) is close to a linear function. The answer to this is positive (in a precise formal sense) as shown by Kalton and Roberts [Trans. Amer. Math. Soc., 278 (1983), pp. 803-816] (and further improved by Bondarenko, Prymak, and Radchenko [J. Math. Anal. Appl., 402 (2013), pp. 234-241]). We revisit their proof idea that is based on expander graphs and provide significantly stronger upper bounds by combining it with new techniques. Furthermore, we provide improved lower bounds for this problem. Another question that we address is that of how to learn a linear function h that is close to an approximately linear function f, while querying the value of f on only a small number of sets. We present a deterministic algorithm that makes only linearly many (in the number of items) nonadaptive queries, and thus improve upon a previous algorithm of Chierichetti, Das, Dasgupta, and Kumar [Proceedings of the 56th Symposium on Foundations of Computer Science, 2015, pp. 1143-1162] that is randomized and makes more than a quadratic number of queries. Our learning algorithm is based on the Hadamard transform.
KW - Learning
KW - Linear functions
KW - Set functions
UR - http://www.scopus.com/inward/record.url?scp=85079738797&partnerID=8YFLogxK
U2 - 10.1137/18M1173873
DO - 10.1137/18M1173873
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85079738797
SN - 0097-5397
VL - 49
SP - 67
EP - 97
JO - SIAM Journal on Computing
JF - SIAM Journal on Computing
IS - 1
ER -