TY - GEN
T1 - Random feature subset selection for analysis of data with missing features
AU - DePasquale, Joseph
AU - Polikar, Robi
PY - 2007
Y1 - 2007
N2 - We discuss an ensemble-of-classifiers based algorithm for the missing feature problem. The proposed approach is inspired in part by the random subspace method, and in part by the incremental learning algorithm, Learn ++. The premise is to generate an adequately large number of classifiers, each trained on a different and random combination of features, drawn from an iteratively updated distribution. To classify an instance with missing features, only those classifiers whose training data did not include the currently missing feature are used. These classifiers are combined by using a majority voting combination rule to obtain the final classification of the given instance. We had previously presented preliminary results on a similar approach, which could handle up to 10% missing data. In this study, we expand our work to include different types of rules to update the distribution, and also examine the effect of the algorithm's primary free parameter (the number of features used to train the ensemble of classifiers) on the overall classification performance. We show that this algorithm can now accommodate up to 30% of features missing without a significant drop in performance.
AB - We discuss an ensemble-of-classifiers based algorithm for the missing feature problem. The proposed approach is inspired in part by the random subspace method, and in part by the incremental learning algorithm, Learn ++. The premise is to generate an adequately large number of classifiers, each trained on a different and random combination of features, drawn from an iteratively updated distribution. To classify an instance with missing features, only those classifiers whose training data did not include the currently missing feature are used. These classifiers are combined by using a majority voting combination rule to obtain the final classification of the given instance. We had previously presented preliminary results on a similar approach, which could handle up to 10% missing data. In this study, we expand our work to include different types of rules to update the distribution, and also examine the effect of the algorithm's primary free parameter (the number of features used to train the ensemble of classifiers) on the overall classification performance. We show that this algorithm can now accommodate up to 30% of features missing without a significant drop in performance.
UR - http://www.scopus.com/inward/record.url?scp=51749102660&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=51749102660&partnerID=8YFLogxK
U2 - 10.1109/IJCNN.2007.4371330
DO - 10.1109/IJCNN.2007.4371330
M3 - Conference contribution
AN - SCOPUS:51749102660
SN - 142441380X
SN - 9781424413805
T3 - IEEE International Conference on Neural Networks - Conference Proceedings
SP - 2379
EP - 2384
BT - The 2007 International Joint Conference on Neural Networks, IJCNN 2007 Conference Proceedings
T2 - 2007 International Joint Conference on Neural Networks, IJCNN 2007
Y2 - 12 August 2007 through 17 August 2007
ER -