TY - JOUR
T1 - Beyond trees
T2 - Classification with sparse pairwise dependencies
AU - Tenzer, Yaniv
AU - Moscovich, Amit
AU - Dorn, Mary Frances
AU - Nadler, Boaz
AU - Spiegelman, Clifford
N1 - Publisher Copyright:
© 2020 Yaniv Tenzer, Amit Moscovich, Mary Frances Dorn, Boaz Nadler and Clifford Spiegelman.
PY - 2020/10
Y1 - 2020/10
N2 - Several classification methods assume that the underlying distributions follow tree-structured graphical models. Indeed, trees capture statistical dependencies between pairs of variables, which may be crucial to attaining low classification errors. In this setting, the optimal classifier is linear in the log-transformed univariate and bivariate densities that correspond to the tree edges. In practice, observed data may not be well approximated by trees. Yet, motivated by the importance of pairwise dependencies for accurate classification, here we propose to approximate the optimal decision boundary by a sparse linear combination of the univariate and bivariate log-transformed densities. Our proposed approach is semi-parametric in nature: we non-parametrically estimate the univariate and bivariate densities, remove pairs of variables that are nearly independent using the Hilbert-Schmidt independence criterion, and finally construct a linear SVM using the retained log-transformed densities. We demonstrate on synthetic and real data sets, that our classifier, named SLB (sparse log-bivariate density), is competitive with other popular classification methods.
AB - Several classification methods assume that the underlying distributions follow tree-structured graphical models. Indeed, trees capture statistical dependencies between pairs of variables, which may be crucial to attaining low classification errors. In this setting, the optimal classifier is linear in the log-transformed univariate and bivariate densities that correspond to the tree edges. In practice, observed data may not be well approximated by trees. Yet, motivated by the importance of pairwise dependencies for accurate classification, here we propose to approximate the optimal decision boundary by a sparse linear combination of the univariate and bivariate log-transformed densities. Our proposed approach is semi-parametric in nature: we non-parametrically estimate the univariate and bivariate densities, remove pairs of variables that are nearly independent using the Hilbert-Schmidt independence criterion, and finally construct a linear SVM using the retained log-transformed densities. We demonstrate on synthetic and real data sets, that our classifier, named SLB (sparse log-bivariate density), is competitive with other popular classification methods.
KW - Bayesian network
KW - Binary classification
KW - Graphical model
KW - Semi-parametric
KW - Sparsity
UR - http://www.scopus.com/inward/record.url?scp=85094925124&partnerID=8YFLogxK
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85094925124
SN - 1532-4435
VL - 21
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
ER -