Feature-domain concatenative speech synthesis

Data processing: speech signal processing – linguistics – language – Speech signal processing – For storage or transmission

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C704S205000

Reexamination Certificate

active

07035791

ABSTRACT:
A method for speech synthesis includes receiving an input speech signal containing a set of speech segments, and estimating spectral envelopes of the input speech signal in a succession of time intervals during each of the speech segments. The spectral envelopes are integrated over a plurality of window functions in a frequency domain so as to determine elements of feature vectors corresponding to the speech segments. An output speech signal is reconstructed by concatenating the feature vectors corresponding to a sequence of the speech segments.

REFERENCES:
patent: 4896359 (1990-01-01), Yamamoto et al.
patent: 5165008 (1992-11-01), Hermansky et al.
patent: 5485543 (1996-01-01), Aso
patent: 5528516 (1996-06-01), Yemini et al.
patent: 5740320 (1998-04-01), Itoh
patent: 5751907 (1998-05-01), Moebius et al.
patent: 5774855 (1998-06-01), Foti et al.
patent: 5913193 (1999-06-01), Huang et al.
patent: 5940795 (1999-08-01), Matsumoto
patent: 6041300 (2000-03-01), Ittycheriah et al.
patent: 6076083 (2000-06-01), Baker
patent: 6101470 (2000-08-01), Eide et al.
patent: 6134528 (2000-10-01), Miller et al.
patent: 6195632 (2001-02-01), Pearson
patent: 6266637 (2001-07-01), Donovan et al.
patent: 6334106 (2001-12-01), Mizuno et al.
patent: 6366883 (2002-04-01), Campbell et al.
patent: 6587816 (2003-07-01), Chazan et al.
patent: 6665641 (2003-12-01), Coorman et al.
patent: 6697780 (2004-02-01), Beutnagel et al.
patent: 6725190 (2004-04-01), Chazan et al.
Yoshinori Sagisaka, Speech Synthesis by Rule Using an Optimal Selection of Non-Uniform Synthesis Units 1988, ATR Interpreting Telephony Research Laboratories, pp. 679-682.
Donovan et al., “The IBM Trainable Speech Synthesis System”,Proceedings of ICSLP, (1998), 4 pages.
Rabiner et al., Fundamentals of Speech Recognition (Prentice-Hall), (1993), pp. 125-128.
Davis et al., “Comparison of Parametric Representations for Monosyllabic Word Recognition in Continuously Spoken Sentences”,IEEE Transactions on Acoustics, Speech, and Signal Processing, (1980), vol. ASSP-28, No. 4, pp. 357-366.
Syrdal et al., “TD-PSOLA Versus Harmonic Plus Noise Model in Diphone Based Speech Synthesis”,Proceedings of ICASSP, (1998), 4 pages.
Huang et al., “Recent Improvements on Microsoft's Trainable Text-to-Speech Systems-Whistler”,Proceedings of ICASSP, (1998), 4 pages.
Chazan et al., “Speech Reconstruction from Mel Frequency Cepstral Coefficients and Pitch Frequency”,Proceedings of the International Conference On Acoustics Speech and Signal Processing, (2000), 4 pages.
Hess, “Pitch Determination of Speech Signals”, Printer-Verlag, (1983).
Ramaswamy et al., “Compression of Acoustic Features for Speech Recognition in Network Environments”,Proceedings of ICASSP, (1998).
Hoory et al., “Speech Synthesis for a Specific Speaker Based on a Labeled Speech Database”,Proceedings of the International Conference on Pattern Recognition, (1994), pp. C145-C148.
Donovan, “Segment Pre-Selection in Decision-Tree Based Speech Synthesis Systems”,ICASSP, (2000), 4 pages.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Feature-domain concatenative speech synthesis does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Feature-domain concatenative speech synthesis, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Feature-domain concatenative speech synthesis will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3536043

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.