Feature set decomposition for decision trees

Lior Rokach, Oded Maimon

Research output: Contribution to journalArticlepeer-review

35 Scopus citations


This paper presents practical aspects of feature set decomposition in classification problems using decision trees. Feature set decomposition generalizes the task of feature selection which is extensively used in data mining. Feature selection aims to provide a representative set of features from which a classifier is constructed. On the other hand, feature set decomposition decomposes the original set of features into several subsets, and builds a classifier for each subset. The classifiers are then combined for classifying new instances. In order to examine the idea, a general framework that searches for helpful decomposition structures is proposed. This framework nests many algorithms, two of which are tested empirically over a set of benchmark datasets. The first algorithm performs a serial search while using a new Vapnik-Chervonenkis dimension bound for multiple oblivious trees as an evaluating schema. The second algorithm performs a multi-search while using wrapper evaluating schema. This work indicates that feature set decomposition can increase the accuracy of decision trees.

Original languageEnglish
Pages (from-to)131-158
Number of pages28
JournalIntelligent Data Analysis
Issue number2
StatePublished - 2005


  • Vapnik-Chervonenkis
  • decision trees
  • decomposition
  • naïve Bayes


Dive into the research topics of 'Feature set decomposition for decision trees'. Together they form a unique fingerprint.

Cite this