Scaling Boosting by Margin-based Inclusion of Features and Relations

Hoche, Susanne and Wrobel, Stefan (2002) Scaling Boosting by Margin-based Inclusion of Features and Relations. In: Machine Learning: ECML 2002. Lecture Notes in Computer Science, 2430 . Springer Berlin / Heidelberg, FIN, pp. 148-160. ISBN 978-3-540-44036-9

Full text not available from this repository.


Boosting is well known to increase the accuracy of propositional and multi-relational classification learners. However, the base learner’s efficiency vitally determines boosting’s efficiency since the complexity of the underlying learner is amplified by iterated calls of the learner in the boosting framework. The idea of restricting the learner to smaller feature subsets in order to increase efficiency is widely used. Surprisingly, little attention has been paid so far to exploiting characteristics of boosting itself to include features based on the current learning progress. In this paper, we show that the dynamics inherent to boosting offer ideal means to maximize the efficiency of the learning process. We describe how to utilize the training examples’ margins—which are known to be maximized by boosting—to reduce learning times without a deterioration of the learning quality. We suggest to stepwise include features in the learning process in response to a slowdown in the improvement of the margins. Experimental results show that this approach significantly reduces the learning time while maintaining or even improving the predictive accuracy of the underlying fully equipped learner.

Item Type: Book Section
Faculty \ School: Faculty of Science > School of Computing Sciences
Depositing User: Vishal Gautam
Date Deposited: 28 Jul 2011 08:36
Last Modified: 15 Dec 2022 00:52
DOI: 10.1007/3-540-36755-1_13

Actions (login required)

View Item View Item