Share Email Print

Proceedings Paper

Multisensor user authentication
Author(s): John M. Colombi; D. Krepp; Steven K. Rogers; Dennis W. Ruck; Mark E. Oxley
Format Member Price Non-Member Price
PDF $17.00 $21.00

Paper Abstract

User recognition is examined using neural and conventional techniques for processing speech and face images. This article for the first time attempts to overcome this significant problem of distortions inherently captured over multiple sessions (days). Speaker recognition uses both Linear Predictive Coding (LPC) cepstral and auditory neural model representations with speaker dependent codebook designs. For facial imagery, recognition is developed on a neural network that consists of a single hidden layer multilayer perceptron backpropagation network using either the raw data as inputs or principal components of the raw data computed using the Karhunen-Loeve Transform as inputs. The data consists of 10 subjects; each subject recorded utterances and had images collected for 10 days. The utterances collected were 400 rich phonetic sentences (4 sec), 200 subject name recordings (3 sec), and 100 imposter name recordings (3 sec). Face data consists of over 2000, 32 X 32 pixel, 8 bit gray scale images of the 10 subjects. Each subsystem attains over 90% verification accuracy individually using test data gathered on day following the training data.

Paper Details

Date Published: 19 August 1993
PDF: 15 pages
Proc. SPIE 1966, Science of Artificial Neural Networks II, (19 August 1993); doi: 10.1117/12.152612
Show Author Affiliations
John M. Colombi, Air Force Institute of Technology (United States)
D. Krepp, Air Force Institute of Technology (United States)
Steven K. Rogers, Air Force Institute of Technology (United States)
Dennis W. Ruck, Air Force Institute of Technology (United States)
Mark E. Oxley, Air Force Institute of Technology (United States)

Published in SPIE Proceedings Vol. 1966:
Science of Artificial Neural Networks II
Dennis W. Ruck, Editor(s)

© SPIE. Terms of Use
Back to Top