Mutual Information Theory for Adaptive Mixture Models
Mutual Information Theory for Adaptive Mixture Models
Many pattern recognition systems need to estimate an underlying probability density function (pdf). Mixture models are commonly used for this purpose in which an underlying pdf is estimated by a finite mixing of distributions. The basic computational element of a density mixture model is a component with a nonlinear mapping function, which takes part in mixing. Selecting an optimal set of components for mixture models is important to ensure an efficient and accurate estimate of an underlying pdf. Previous work has commonly estimated an underlying pdf based on the information contained in patterns. In this paper, mutual information theory is employed to measure whether two components are statistically dependent. If a component has small mutual information, it is statistically independent of the other components. Hence, that component makes a significant contribution to the system pdf and should not be removed. However, if a particular component has large mutual information, it is unlikely to be statistically independent of the other components and may be removed without significant damage to the estimated pdf. Continuing to remove components with large and positive mutual information will give a density mixture model with an optimal structure, which is very close to the true pdf.
396-403
Zwolinski, M.
adfcb8e7-877f-4bd7-9b55-7553b6cb3ea0
Yang, Z.R.
b71f8582-83df-4c36-9925-136cdfd207f9
April 2001
Zwolinski, M.
adfcb8e7-877f-4bd7-9b55-7553b6cb3ea0
Yang, Z.R.
b71f8582-83df-4c36-9925-136cdfd207f9
Zwolinski, M. and Yang, Z.R.
(2001)
Mutual Information Theory for Adaptive Mixture Models.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 23 (4), .
Abstract
Many pattern recognition systems need to estimate an underlying probability density function (pdf). Mixture models are commonly used for this purpose in which an underlying pdf is estimated by a finite mixing of distributions. The basic computational element of a density mixture model is a component with a nonlinear mapping function, which takes part in mixing. Selecting an optimal set of components for mixture models is important to ensure an efficient and accurate estimate of an underlying pdf. Previous work has commonly estimated an underlying pdf based on the information contained in patterns. In this paper, mutual information theory is employed to measure whether two components are statistically dependent. If a component has small mutual information, it is statistically independent of the other components. Hence, that component makes a significant contribution to the system pdf and should not be removed. However, if a particular component has large mutual information, it is unlikely to be statistically independent of the other components and may be removed without significant damage to the estimated pdf. Continuing to remove components with large and positive mutual information will give a density mixture model with an optimal structure, which is very close to the true pdf.
More information
Published date: April 2001
Identifiers
Local EPrints ID: 255750
URI: http://eprints.soton.ac.uk/id/eprint/255750
PURE UUID: 03a409a5-269f-4b2a-844f-310890d681cb
Catalogue record
Date deposited: 27 Apr 2001
Last modified: 15 Mar 2024 02:39
Export record
Contributors
Author:
M. Zwolinski
Author:
Z.R. Yang
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
Loading...
View more statistics