The University of Southampton
University of Southampton Institutional Repository

Boosting Strategy for Classification

Boosting Strategy for Classification
Boosting Strategy for Classification
This paper introduces a strategy for training ensemble classifiers by analysing boosting within margin theory. We present a bound on the generalisation error of ensembled classifiers in terms of the 2-norm of the margin slack vector. We develop an effective, adaptive and robust boosting algorithm, DMBoost, by optimising this bound. The soft margin based quadratic loss function is insensitive to points having a large margin. The algorithm improves the generalisation performance of a system by ignoring the examples having small or negative margin. We evaluate the efficacy of the proposed method by applying it to a text categorization task. Experimental results show that DMBoost performs significantly better than AdaBoost, hence validating the effectiveness of the method. Furthermore, experimental results on UCI data sets demonstrate that DMBoost generally outperforms AdaBoost.
1088-467x
149-174
Lodhi, H.
80ab75b5-cd7b-4455-a158-aac3c0b4a74d
Karakoulas, G.
10b0760f-a471-4bf0-aacf-53b74e7a8bbd
Shawe-Taylor, J.
c32d0ee4-b422-491f-8c28-78663851d6db
Lodhi, H.
80ab75b5-cd7b-4455-a158-aac3c0b4a74d
Karakoulas, G.
10b0760f-a471-4bf0-aacf-53b74e7a8bbd
Shawe-Taylor, J.
c32d0ee4-b422-491f-8c28-78663851d6db

Lodhi, H., Karakoulas, G. and Shawe-Taylor, J. (2002) Boosting Strategy for Classification. Intelligent Data Analysis, 6 (2), 149-174.

Record type: Article

Abstract

This paper introduces a strategy for training ensemble classifiers by analysing boosting within margin theory. We present a bound on the generalisation error of ensembled classifiers in terms of the 2-norm of the margin slack vector. We develop an effective, adaptive and robust boosting algorithm, DMBoost, by optimising this bound. The soft margin based quadratic loss function is insensitive to points having a large margin. The algorithm improves the generalisation performance of a system by ignoring the examples having small or negative margin. We evaluate the efficacy of the proposed method by applying it to a text categorization task. Experimental results show that DMBoost performs significantly better than AdaBoost, hence validating the effectiveness of the method. Furthermore, experimental results on UCI data sets demonstrate that DMBoost generally outperforms AdaBoost.

This record has no associated files available for download.

More information

Published date: 2002
Organisations: Electronics & Computer Science

Identifiers

Local EPrints ID: 259784
URI: http://eprints.soton.ac.uk/id/eprint/259784
ISSN: 1088-467x
PURE UUID: a377dfa5-6705-4fbe-841a-e7510b91ac6f

Catalogue record

Date deposited: 17 Aug 2004
Last modified: 07 Jan 2022 21:12

Export record

Contributors

Author: H. Lodhi
Author: G. Karakoulas
Author: J. Shawe-Taylor

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×