Speaker recognition using neural network and adaptive wavelet transform
Abstract
The same word uttered by different people has different waveforms. It has also been observed that the same word uttered by the same person has different waveform at different times. This difference can be characterized by some time domain dilation effects in the waveform. In our experiment a set of words was selected and each word was uttered eight times by five different speakers. The objective of this work is to extract a wavelet basis function for the speech data generated by each individual speaker. The wavelet filter coefficients are then used as a feature set and fed into a neural network-based speaker recognition system. This is an attempt to cascade a wavelet network (wavenet) and a neural network (neural-net) for feature extraction and classification respectively and applied for speaker recognition. The results show very high promise and good prospects to couple a wavelet network and neural networks.
- Publication:
-
Visual Information Processing II
- Pub Date:
- August 1993
- DOI:
- 10.1117/12.150976
- Bibcode:
- 1993SPIE.1961..391B