Extracting moving shapes by evidence gathering
Extracting moving shapes by evidence gathering
Many approaches can track objects moving in sequences of images but can suffer in occlusion and noise, and often require initialisation. These factors can be handled by techniques that extract objects from image sequences, especially when phrased in terms of evidence gathering.
Since the template approach is proven for arbitrary shapes, we re-deploy it for moving arbitrary shapes, but in a way aimed to avoid discretisation problems. In this way, the discrete mapping operation is deferred as far as possible, by using continuous shape descriptions. A further advantage is reduction in computational demand, as seen in use of templates for shape extraction.
This prior specification of motion avoids the need to use an expensive parametric model to capture data that is already known. Furthermore, the complexity of the motion template model remains unchanged with increase in the complexity of motion, whereas a parametric model would require increasingly more parameters leading to an enormous increase in computational requirements.
The new approach combining moving arbitrary shape description with motion templates permits us to achieve the objective of low dimensionality extraction of arbitrarily moving arbitrary shapes with performance advantage as reflected by the results this new technique can achieve.
arbitrary shape extraction, motion estimation, tracking, motion template, hough transform, evidence gathering
1099-1114
Grant, Michael G.
835c848a-f26c-4474-b0f7-8465d44bdacf
Nixon, Mark S.
2b5b9804-5a81-462a-82e6-92ee5fa74e12
Lewis, Paul H.
7aa6c6d9-bc69-4e19-b2ac-a6e20558c020
May 2002
Grant, Michael G.
835c848a-f26c-4474-b0f7-8465d44bdacf
Nixon, Mark S.
2b5b9804-5a81-462a-82e6-92ee5fa74e12
Lewis, Paul H.
7aa6c6d9-bc69-4e19-b2ac-a6e20558c020
Grant, Michael G., Nixon, Mark S. and Lewis, Paul H.
(2002)
Extracting moving shapes by evidence gathering.
[in special issue: Handwriting processing and applications]
Pattern Recognition, 35 (5), .
(doi:10.1016/S0031-3203(01)00078-4).
Abstract
Many approaches can track objects moving in sequences of images but can suffer in occlusion and noise, and often require initialisation. These factors can be handled by techniques that extract objects from image sequences, especially when phrased in terms of evidence gathering.
Since the template approach is proven for arbitrary shapes, we re-deploy it for moving arbitrary shapes, but in a way aimed to avoid discretisation problems. In this way, the discrete mapping operation is deferred as far as possible, by using continuous shape descriptions. A further advantage is reduction in computational demand, as seen in use of templates for shape extraction.
This prior specification of motion avoids the need to use an expensive parametric model to capture data that is already known. Furthermore, the complexity of the motion template model remains unchanged with increase in the complexity of motion, whereas a parametric model would require increasingly more parameters leading to an enormous increase in computational requirements.
The new approach combining moving arbitrary shape description with motion templates permits us to achieve the objective of low dimensionality extraction of arbitrarily moving arbitrary shapes with performance advantage as reflected by the results this new technique can achieve.
This record has no associated files available for download.
More information
Published date: May 2002
Keywords:
arbitrary shape extraction, motion estimation, tracking, motion template, hough transform, evidence gathering
Identifiers
Local EPrints ID: 38777
URI: http://eprints.soton.ac.uk/id/eprint/38777
ISSN: 0031-3203
PURE UUID: 7876b08e-935e-47dd-863f-a156090b9123
Catalogue record
Date deposited: 14 Jun 2006
Last modified: 16 Mar 2024 02:34
Export record
Altmetrics
Contributors
Author:
Michael G. Grant
Author:
Paul H. Lewis
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics