A tongue-movement communication and control concept for hands-free human-machine interfaces
A tongue-movement communication and control concept for hands-free human-machine interfaces
A new communication and control concept using tongue movements is introduced to generate, detect, and classify signals that can be used in novel hands-free human–machine interface applications such as communicating with a computer and controlling devices. The signals that are caused by tongue movements are the changes in the airflow pressure that occur in the ear canal. The goal is to demonstrate that the ear pressure signals that are acquired using a microphone that is inserted into the ear canal, due to specific tongue movements, are distinct and that the signals can be detected and classified very accurately. The strategy that is developed for demonstrating the concept includes energy-based signal detection and segmentation to extract ear pressure signals due to tongue movements, signal normalization to decrease the trial-to-trial variations in the signals, and pairwise cross-correlation signal averaging to obtain accurate estimates from ensembles of pressure signals. A new decision fusion classification algorithm is formulated to assign the pressure signals to their respective tongue-movement classes. The complete strategy of signal detection and segmentation, estimation, and classification is tested on four tongue movements of eight subjects. Through extensive experiments, it is demonstrated that the ear pressure signals due to the tongue movements are distinct and that the four pressure signals can be classified with an accuracy of more than 97% averaged across the eight subjects using the decision fusion classification algorithm. Thus, it is concluded that, through the unique concept that is introduced in this paper, human–computer interfaces that use tongue movements can be designed for hands-free communication and control applications.
533-546
Vaidyanathan, R.
f062a7b1-fc7e-4227-9e1b-ca0b61330237
Chung, B.
bb913e22-b6ae-46d2-ad8c-7ebfa9b7e0ba
Gupta, L.
d2b4415e-e315-431e-8587-fe6d85560485
Kook, H.
ae545691-dbb8-419f-bd42-6c556e5bcb7b
Kota, S.
5cb61382-f881-4f28-8d49-4f4e4bbd24f9
West, J.D.
a5a63c8a-f843-4f6b-ad59-c49a39bec01f
July 2007
Vaidyanathan, R.
f062a7b1-fc7e-4227-9e1b-ca0b61330237
Chung, B.
bb913e22-b6ae-46d2-ad8c-7ebfa9b7e0ba
Gupta, L.
d2b4415e-e315-431e-8587-fe6d85560485
Kook, H.
ae545691-dbb8-419f-bd42-6c556e5bcb7b
Kota, S.
5cb61382-f881-4f28-8d49-4f4e4bbd24f9
West, J.D.
a5a63c8a-f843-4f6b-ad59-c49a39bec01f
Vaidyanathan, R., Chung, B., Gupta, L., Kook, H., Kota, S. and West, J.D.
(2007)
A tongue-movement communication and control concept for hands-free human-machine interfaces.
IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, 37 (4), .
(doi:10.1109/TSMCA.2007.897919).
Abstract
A new communication and control concept using tongue movements is introduced to generate, detect, and classify signals that can be used in novel hands-free human–machine interface applications such as communicating with a computer and controlling devices. The signals that are caused by tongue movements are the changes in the airflow pressure that occur in the ear canal. The goal is to demonstrate that the ear pressure signals that are acquired using a microphone that is inserted into the ear canal, due to specific tongue movements, are distinct and that the signals can be detected and classified very accurately. The strategy that is developed for demonstrating the concept includes energy-based signal detection and segmentation to extract ear pressure signals due to tongue movements, signal normalization to decrease the trial-to-trial variations in the signals, and pairwise cross-correlation signal averaging to obtain accurate estimates from ensembles of pressure signals. A new decision fusion classification algorithm is formulated to assign the pressure signals to their respective tongue-movement classes. The complete strategy of signal detection and segmentation, estimation, and classification is tested on four tongue movements of eight subjects. Through extensive experiments, it is demonstrated that the ear pressure signals due to the tongue movements are distinct and that the four pressure signals can be classified with an accuracy of more than 97% averaged across the eight subjects using the decision fusion classification algorithm. Thus, it is concluded that, through the unique concept that is introduced in this paper, human–computer interfaces that use tongue movements can be designed for hands-free communication and control applications.
This record has no associated files available for download.
More information
Published date: July 2007
Identifiers
Local EPrints ID: 46371
URI: http://eprints.soton.ac.uk/id/eprint/46371
ISSN: 1083-4427
PURE UUID: c1ec4267-e6c1-4b6e-9d70-422a771f2551
Catalogue record
Date deposited: 25 Jun 2007
Last modified: 15 Mar 2024 09:21
Export record
Altmetrics
Contributors
Author:
R. Vaidyanathan
Author:
B. Chung
Author:
L. Gupta
Author:
H. Kook
Author:
S. Kota
Author:
J.D. West
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics