Speech coding apparatus having speaker dependent prototypes gene

Patent

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

G10L 902

Patent

active

052789420

ABSTRACT:
A speech coding apparatus and method for use in a speech recognition apparatus and method. The value of at least one feature of an utterance is measured during each of a series of successive time intervals to produce a series of feature vector signals representing the feature values. A plurality of prototype vector signals, each having at least one parameter value and a unique identification value are stored. The closeness of the feature vector signal is compared to the parameter values of the prototype vector signals to obtain prototype match scores for the feature value signal and each prototype vector signal. The identification value of the prototype vector signal having the best prototype match score is output as a coded representation signal of the feature vector signal. Speaker-dependent prototype vector signals are generated from both synthesized training vector signals and measured training vector signals. The synthesized training vector signals are transformed reference feature vector signals representing the values of features of one or more utterances of one or more speakers in a reference set of speakers. The measured training feature vector signals represent the values of features of one or more utterances of a new speaker/user not in the reference set.

REFERENCES:
patent: 4651289 (1987-03-01), Maeda et al.
patent: 4751737 (1988-06-01), Gerson et al.
patent: 4817156 (1989-03-01), Bahl et al.
patent: 4829577 (1989-05-01), Kuroda et al.
patent: 4980918 (1990-12-01), Bahl et al.
Tappert et al., "Fast Training Method for Speech Recognition Systems," IBM Technical Disclosure Bulletin, vol. 21, No. 8, Jan. 1979, pp. 3413-3414.
Bahl, L. R. et al. "Acoustic Markov Models Used In The Tangora Speech Recognition System." Proceedings of the 1988 IEEE International Conference on Acoustics, Speech, and Signal Processing, S11-3, pp. 497-500, Apr. 1988.
Bahl, L. R., et al. "Fast Algorithm for Deriving Acoustic Prototoypes for Automatic Speech Recognition." U.S. patent application Ser. No. 732,714, filed on Jul. 16, 1991.
Bahl, L. R., et al. "A Maximum Likelihood Approach to Continuous Speech Recognition." IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. PAMI-5, No. 2, pp. 179-190, Mar. 1983.
Bahl, L. R., et al. "Metamorphic Transformations For Speech Recognition." IBM Technical Disclosure Bulletin, vol. 33, No. 1A, Jun. 1990, pp. 291-292.
Bahl, L. R., et al. "Speaker Independent Label Coding Apparatus." U.S. patent application Ser. No. 673,810, filed Mar. 22, 1991.
Bahl, L. R., et al. "Vector Quantization Procedure For Speech Recognition Systems Using Discrete Parameter Phoneme-Based Markov Word Models." IBM Technical Disclosure Bulletin, vol. 34, No. 7, Dec. 1989, pp. 340 and 341.
Hartigan, J. A. "The K-Means Algorithm." Clustering Algorithms, pp. 84-105, John Wiley & Sons, 1975.
Jelinek, F. "Continuous Speech Recognition By Statistical Methods." Proceedings of the IEEE, vol. 64, No. 4, pp. 532-556, Apr. 1976.
Jelinek, F. "The Development of an Experimental Discrete Dictation Recognizer." Proceedings of the IEEE, vol. 73, No. 11, Nov. 1985, pp. 1116-1624.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Speech coding apparatus having speaker dependent prototypes gene does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Speech coding apparatus having speaker dependent prototypes gene, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Speech coding apparatus having speaker dependent prototypes gene will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-1637318

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.