Explore open access research and scholarly works from STORE - University of Staffordshire Online Repository

Advanced Search

Integrated person identification using voice and facial features

Chibelushi, C.C., Mason, J.S.D. and Deravi, F. (1997) Integrated person identification using voice and facial features. In: IEE Colloquium on Image Processing for Security Applications, 10 March 1997, London, UK.

Full text not available from this repository.
Official URL: http://dx.doi.org/10.1049/ic:19970380

Abstract or description

Real-world automatic person recognition requires a consistently high recognition accuracy which is difficult to attain using a single recognition modality. This paper addresses the issue of person identification accuracy resulting from the combination of voice and outer lip-margin features. An assessment of feature fusion - based on audio-visual feature vector concatenation, principal component analysis, and linear discriminant analysis - is conducted. The paper shows that outer lip margins carry speaker identity cues. It is also shown that the joint use of voice and lip-margin features is equivalent to an effective increase in signal-to-noise ratio of the audio signal. Simple audio-visual feature vector concatenation is shown to be an effective method for feature combination, and linear discriminant analysis is shown to possess the capability of packing discriminating audio-visual information into fewer coeficients than principal component analysis.

Item Type: Conference or Workshop Item (Paper)
Faculty: Previous Faculty of Computing, Engineering and Sciences > Computing
Event Title: IEE Colloquium on Image Processing for Security Applications
Event Location: London, UK
Event Dates: 10 March 1997
Depositing User: Claude CHIBELUSHI
Date Deposited: 12 May 2013 20:37
Last Modified: 24 Feb 2023 13:38
URI: https://eprints.staffs.ac.uk/id/eprint/1118

Actions (login required)

View Item
View Item