Training Neurofuzzy Systems
Training Neurofuzzy Systems
A neurofuzzy system combines the positive attributes of a neural network and a fuzzy system by providing a transparent framework for representing linguistic rules with well defined modelling and learning characteristics. Unfortunately, their application is limited to problems involving a small number of input variables by the curse of dimensionality where the size of the rule base and the training set increase as an exponential function of the input dimension. The curse can be alleviated by a number of approaches but one which has recently received much attention is the exploitation of redundancy. Many functions can be adequately approximated by an additive model whose output is a sum over several smaller dimensional submodels. This technique is called global partitioning and the aim of an algorithm designed to construct the approximation is to automatically determine the number of submodels and the subset of input variables for each submodel. The construction algorithm is an iterative process where each iteration must identify a set of candidate refinements and evaluate the associated candidate models. This leads naturally to the problem of how to train the candidate models and the approach taken depends on whether they contain one or multiple submodels.
213--218
Mills, D.J.
bd207c8b-fbf0-41da-bba4-b54d9a29804d
Brown, M.
52cf4f52-6839-4658-8cc5-ec51da626049
Harris, C.J.
c4fd3763-7b3f-4db1-9ca3-5501080f797a
1994
Mills, D.J.
bd207c8b-fbf0-41da-bba4-b54d9a29804d
Brown, M.
52cf4f52-6839-4658-8cc5-ec51da626049
Harris, C.J.
c4fd3763-7b3f-4db1-9ca3-5501080f797a
Mills, D.J., Brown, M. and Harris, C.J.
(1994)
Training Neurofuzzy Systems.
Symp. on Artificial Intelligence in Real-Time Control.
.
Record type:
Conference or Workshop Item
(Other)
Abstract
A neurofuzzy system combines the positive attributes of a neural network and a fuzzy system by providing a transparent framework for representing linguistic rules with well defined modelling and learning characteristics. Unfortunately, their application is limited to problems involving a small number of input variables by the curse of dimensionality where the size of the rule base and the training set increase as an exponential function of the input dimension. The curse can be alleviated by a number of approaches but one which has recently received much attention is the exploitation of redundancy. Many functions can be adequately approximated by an additive model whose output is a sum over several smaller dimensional submodels. This technique is called global partitioning and the aim of an algorithm designed to construct the approximation is to automatically determine the number of submodels and the subset of input variables for each submodel. The construction algorithm is an iterative process where each iteration must identify a set of candidate refinements and evaluate the associated candidate models. This leads naturally to the problem of how to train the candidate models and the approach taken depends on whether they contain one or multiple submodels.
This record has no associated files available for download.
More information
Published date: 1994
Additional Information:
Organisation: IFAC Address: Valencia, Spain
Venue - Dates:
Symp. on Artificial Intelligence in Real-Time Control, 1994-01-01
Organisations:
Southampton Wireless Group
Identifiers
Local EPrints ID: 250268
URI: http://eprints.soton.ac.uk/id/eprint/250268
PURE UUID: e28d6f08-a2c6-4952-ad4d-8255b09e8e2c
Catalogue record
Date deposited: 04 May 1999
Last modified: 10 Dec 2021 20:07
Export record
Contributors
Author:
D.J. Mills
Author:
M. Brown
Author:
C.J. Harris
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics