Boosting Strategy for Classification
Boosting Strategy for Classification
This paper introduces a strategy for training ensemble classifiers by analysing boosting within margin theory. We present a bound on the generalisation error of ensembled classifiers in terms of the 2-norm of the margin slack vector. We develop an effective, adaptive and robust boosting algorithm, DMBoost, by optimising this bound. The soft margin based quadratic loss function is insensitive to points having a large margin. The algorithm improves the generalisation performance of a system by ignoring the examples having small or negative margin. We evaluate the efficacy of the proposed method by applying it to a text categorization task. Experimental results show that DMBoost performs significantly better than AdaBoost, hence validating the effectiveness of the method. Furthermore, experimental results on UCI data sets demonstrate that DMBoost generally outperforms AdaBoost.
149-174
Lodhi, H.
80ab75b5-cd7b-4455-a158-aac3c0b4a74d
Karakoulas, G.
10b0760f-a471-4bf0-aacf-53b74e7a8bbd
Shawe-Taylor, J.
c32d0ee4-b422-491f-8c28-78663851d6db
2002
Lodhi, H.
80ab75b5-cd7b-4455-a158-aac3c0b4a74d
Karakoulas, G.
10b0760f-a471-4bf0-aacf-53b74e7a8bbd
Shawe-Taylor, J.
c32d0ee4-b422-491f-8c28-78663851d6db
Lodhi, H., Karakoulas, G. and Shawe-Taylor, J.
(2002)
Boosting Strategy for Classification.
Intelligent Data Analysis, 6 (2), .
Abstract
This paper introduces a strategy for training ensemble classifiers by analysing boosting within margin theory. We present a bound on the generalisation error of ensembled classifiers in terms of the 2-norm of the margin slack vector. We develop an effective, adaptive and robust boosting algorithm, DMBoost, by optimising this bound. The soft margin based quadratic loss function is insensitive to points having a large margin. The algorithm improves the generalisation performance of a system by ignoring the examples having small or negative margin. We evaluate the efficacy of the proposed method by applying it to a text categorization task. Experimental results show that DMBoost performs significantly better than AdaBoost, hence validating the effectiveness of the method. Furthermore, experimental results on UCI data sets demonstrate that DMBoost generally outperforms AdaBoost.
This record has no associated files available for download.
More information
Published date: 2002
Organisations:
Electronics & Computer Science
Identifiers
Local EPrints ID: 259784
URI: http://eprints.soton.ac.uk/id/eprint/259784
ISSN: 1088-467x
PURE UUID: a377dfa5-6705-4fbe-841a-e7510b91ac6f
Catalogue record
Date deposited: 17 Aug 2004
Last modified: 07 Jan 2022 21:12
Export record
Contributors
Author:
H. Lodhi
Author:
G. Karakoulas
Author:
J. Shawe-Taylor
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics