A tongue-movement communication and control concept for hands-free human-machine interfaces
Vaidyanathan, R., Chung, B., Gupta, L., Kook, H., Kota, S. and West, J.D. (2007) A tongue-movement communication and control concept for hands-free human-machine interfaces. IEEE Transactions on Systems, Man and Cybernetics Part A: Systems and Humans, 37, (4), 533-546. (doi:10.1109/TSMCA.2007.897919).
Full text not available from this repository.
A new communication and control concept using tongue movements is introduced to generate, detect, and classify signals that can be used in novel hands-free human–machine interface applications such as communicating with a computer and controlling devices. The signals that are caused by tongue movements are the changes in the airflow pressure that occur in the ear canal. The goal is to demonstrate that the ear pressure signals that are acquired using a microphone that is inserted into the ear canal, due to specific tongue movements, are distinct and that the signals can be detected and classified very accurately. The strategy that is developed for demonstrating the concept includes energy-based signal detection and segmentation to extract ear pressure signals due to tongue movements, signal normalization to decrease the trial-to-trial variations in the signals, and pairwise cross-correlation signal averaging to obtain accurate estimates from ensembles of pressure signals. A new decision fusion classification algorithm is formulated to assign the pressure signals to their respective tongue-movement classes. The complete strategy of signal detection and segmentation, estimation, and classification is tested on four tongue movements of eight subjects. Through extensive experiments, it is demonstrated that the ear pressure signals due to the tongue movements are distinct and that the four pressure signals can be classified with an accuracy of more than 97% averaged across the eight subjects using the decision fusion classification algorithm. Thus, it is concluded that, through the unique concept that is introduced in this paper, human–computer interfaces that use tongue movements can be designed for hands-free communication and control applications.
|Subjects:||T Technology > TK Electrical engineering. Electronics Nuclear engineering
Q Science > QP Physiology
Q Science > QC Physics
|Divisions:||University Structure - Pre August 2011 > Institute of Sound and Vibration Research > Signal Processing and Control
|Date Deposited:||25 Jun 2007|
|Last Modified:||01 Jun 2011 14:28|
|Contributors:||Vaidyanathan, R. (Author)
Chung, B. (Author)
Gupta, L. (Author)
Kook, H. (Author)
Kota, S. (Author)
West, J.D. (Author)
|Contact Email Address:||R.Vaidyanathan@soton.ac.uk|
|RDF:||RDF+N-Triples, RDF+N3, RDF+XML, Browse.|
Actions (login required)