Boosting Strategy for Classification
Lodhi, H., Karakoulas, G. and Shawe-Taylor, J. (2002) Boosting Strategy for Classification. Intelligent Data Analysis, 6, (2), 149-174.
Full text not available from this repository.
This paper introduces a strategy for training ensemble classifiers by analysing boosting within margin theory. We present a bound on the generalisation error of ensembled classifiers in terms of the 2-norm of the margin slack vector. We develop an effective, adaptive and robust boosting algorithm, DMBoost, by optimising this bound. The soft margin based quadratic loss function is insensitive to points having a large margin. The algorithm improves the generalisation performance of a system by ignoring the examples having small or negative margin. We evaluate the efficacy of the proposed method by applying it to a text categorization task. Experimental results show that DMBoost performs significantly better than AdaBoost, hence validating the effectiveness of the method. Furthermore, experimental results on UCI data sets demonstrate that DMBoost generally outperforms AdaBoost.
|Divisions:||Faculty of Physical and Applied Science > Electronics and Computer Science
|Date Deposited:||17 Aug 2004|
|Last Modified:||02 Mar 2012 12:59|
|Contributors:||Lodhi, H. (Author)
Karakoulas, G. (Author)
Shawe-Taylor, J. (Author)
|Further Information:||Google Scholar|
|RDF:||RDF+N-Triples, RDF+N3, RDF+XML, Browse.|
Actions (login required)