Random subspace method
Random subspace method [1] (or attribute bagging[2]) is an ensemble classifier that consists of several classifiers each operating in a subspace of the original feature space, and outputs the class based on the outputs of these individual classifiers. Random subspace method has been used for decision trees (random decision forests),<ref name="ho1995"'>Ho, Tin Kam (1995). Random Decision Forest (PDF). Proceedings of the 3rd International Conference on Document Analysis and Recognition, Montreal, QC, 14–16 August 1995. pp. 278–282.</ref>[1] linear classifiers,[3] support vector machines,[4] nearest neighbours[5] and other types of classifiers. This method is also applicable to one-class classifiers.[6][7]
The algorithm is an attractive choice for classification problems where the number of features is much larger than the number of training objects, such as fMRI data[8] or gene expression data.[9]
In machine learning and statistics, feature selection, also known as variable selection, attribute selection or variable subset selection, is the process of selecting a subset of relevant features (variables, predictors) for use in model construction. Feature selection techniques are used for three reasons:
- simplification of models to make them easier to interpret by researchers/users,[1]
- shorter training times,
- enhanced generalization by reducing overfitting[2](formally, reduction of variance[1])
The central premise when using a feature selection technique is that the data contains many features that are either redundant or irrelevant, and can thus be removed without incurring much loss of information.[2] Redundant or irrelevant features are two distinct notions, since one relevant feature may be redundant in the presence of another relevant feature with which it is strongly correlated.[3]
References
- 1 2 Ho, Tin Kam (1998). "The Random Subspace Method for Constructing Decision Forests" (PDF). IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (8): 832–844. doi:10.1109/34.709601.
- ↑ Bryll, R. (2003). "Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets". Pattern Recognition 36 (6): 1291–1302. doi:10.1016/s0031-3203(02)00121-8.
- ↑ Skurichina, Marina (2002). "Bagging, boosting and the random subspace method for linear classifiers". Pattern Analysis and Applications 5 (2): 121–135. doi:10.1007/s100440200011.
- ↑ Tao, D. (2006). "Asymmetric bagging and random subspace for support vector machines-based relevance feedback in image retrieval". IEEE Transactions on Pattern Analysis and Machine Intelligence. doi:10.1109/tpami.2006.134.
- ↑ Tremblay, G. (2004). "Optimizing Nearest Neighbour in Random Subspaces using a Multi-Objective Genetic Algorithm" (PDF). 17th International Conference on Pattern Recognition: 208–211.
- ↑ Nanni, L. (2006). "Experimental comparison of one-class classifiers for online signature verification". Neurocomputing 69 (7).
- ↑ Cheplygina, Veronika; David M.J. Tax, "Pruned random subspace method for one-class classifiers", Multiple Classifier Systems 2011 (PDF), pp. 96–105 Cite uses deprecated parameter
|coauthors=
(help) - ↑ Kuncheva, Ludmila; et al. (2010). IEEE Transactions on Medical Imaging (PDF) 29 (2): 531–542 http://pages.bangor.ac.uk/~mas00a/papers/lkjrcpdlsjtmi10.pdf. Missing or empty
|title=
(help);|contribution=
ignored (help) - ↑ Bertoni, Alberto; Folgieri, Raffaella; Valentini, Giorgio (2005). "Bio-molecular cancer prediction with random subspace ensembles of support vector machines". Neurocomputing 63: 535–539. doi:10.1016/j.neucom.2004.07.007.