1. Trang chủ
  2. » Kỹ Thuật - Công Nghệ

Tài liệu Digital Signal Processing Handbook P46 pptx

12 268 1

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 12
Dung lượng 186,94 KB

Nội dung

Sproat, R. & Olive, J. “Text-to-Speech Synthesis” Digital Signal Processing Handbook Ed. Vijay K. Madisetti and Douglas B. Williams Boca Raton: CRC Press LLC, 1999 c  1999byCRCPressLLC 46 Text-to-Speech Synthesis Richard Sproat Bell Laboratories Lucent Technologies Joseph Olive Bell Laboratories Lucent Technologies 46.1 Introduction 46.2 Text Analysis and Linguistic Analysis Text Preprocessing • Accentuation • WordPronunciation • In- tonational Phrasing • Segmental Durations • Intonation 46.3 Speech Synthesis 46.4 The Future of T TS References 46.1 Introduction Text-to-speech synthesis has had a long history, one that can be traced back at least to Dudley’s “Voder”, developed at Bell Laboratories and demonstrated at the 1939 World’s Fair [1]. Practical systems for automatically generating speech parameters from a linguistic representation (such as a phoneme string) were not available until the 1960s, and systems for converting from ordinary text into speech were first completed in the 1970s, with MITalk being the best-known such system [2]. Many projects in text-to-speech conversion have been initiated in the intervening years, and papers on many of these systems have been published. 1 Itistemptingtothinkoftheproblemofconverting writtentextintospeechas“speechrecognition in reverse”: current speech recognition systems are generally deemed successful if they can convert speech input into the sequence of words that was uttered by the speaker, so one might imagine that a text-to-speech (TTS) synthesizer would start with the words in the text, convert each word one-by-oneinto speech(beingcarefulto pronounceeach wordcorrectly),andconcatenatetheresult together. However, when one considers what literate native speakers of a language must do when they read a text aloud, it quickly b ecomes clear that things are much more complicated than this simplistic view suggests. Pronouncing words correctly is only part of the problem faced by human readers: in order to sound natural and to sound as if they understand what they are reading, they mustalsoappropriatelyemphasize(accent)somewords,anddeemphasizeothers; theymust“chunk” thesentenceintomeaningful(intonational)phrases; they mustpickanappropriateF0(fundamental frequency) contour; they must control certain aspects of their voice quality; they must know that a word should be pronounced longer if it appears in some positions in the sentence than if it appears in others because segmental durations are affected by various factors, including phrasal position. 1 For example, [3] gives an overview of recent Dutch efforts in this area. Audio examples of several current projects on TTS can be found at the WWW URL http://www.cs.bham.ac.uk/∼jpi/synth/museum.html. c  1999 by CRC Press LLC What makes reading such a difficult task is that all writing systems systematically fail to specify many kinds of information that are important in speech. While the written form of a sentence (usually) completely specifies the words that are present, it will only partly specify the intonational phrases (typically with some form of punctuation), will usually not indicate which words to accent or deaccent, and hardly ever give information on segmental duration, voice quality, or intonation. (One might think that a question mark “?” indicates that a sentence should be pronounced with a rising intonation: generally, though, a question mark merely indicates that a sentence is a question, leavingituptothereadertojudgewhetherthisquestionshouldberenderedwitharisingintonation.) Theorthographiesofsomelanguages—e.g., Chinese, Japanese,andThai—failtogive information on where word boundaries are, so that even this needs to be figured out by the reader. 2 Humans are able to perform these tasks because, in addition to being knowledgeable about the grammar of their language, they also (usually) understand the content of the text that the y are reading, and can thus appropriately manipulate various extragrammatical “affective” factors, such as appropriate use of intonation and voice quality. The task of a TTS system is thus a complex one that involves mimicking what human readers do. But a machine is hobbled by the fact that it generally “knows” the grammatical facts of the language only imperfectly, and generally can be said to “understand” nothing of what it is reading. TTS algorithms thus have to do the best they can making use, where possible, of purely gr ammatical information to decide on such things as accentuation, phrasing, and intonation — and coming up with a reasonable “middle ground” analysis for aspects of the output that are more dependent on actual understanding. It is natural to divide the TTS problem into two broad subproblems. The first of these is the conversion of text — an imperfect representation of language, as we have seen — into some form of linguistic representation that includes information on the phonemes (sounds) to be produced, their duration, the locations of any pauses, and the F0 contour to be used. The second — the actual synthesis of speech — takes this information and converts it into a speech waveform. Each of these main tasks naturally breaks down into further subtasks, some of which have been alluded to. The first part, text and linguistic analysis, may be broken down as follows: • Textpreprocessing: includingend-of-sentencedetection,“textnormalization”(expansion of numerals and abbreviations), and limited grammatical analysis, such as grammatical part-of-speech assignment. • Accent assignment: the assignment of levels of prominence to various words in the sen- tence. • Word pronunciation: including the pronunciation of names and the disambiguation of homographs. 3 • Intonational phrasing: the breaking of (usually long) stretches of text into one or more intonational units. • Segmental durations: the determination,onthe basis of linguistic informationcomputed thus far, of appropriate durations for phonemes in the input. • F0 contour computation. 2 Even in English, single orthographic words, e.g., AT&T, can actually represent multiple words — A T and T. 3 Ahomograph isasingle writtenwordthat representstwo or more different lexical entries, often having different pronun- ciations: anexample would bebass, which could be thewordfor amusical range —w ith pronunciation/be j s/—orafish — with pronunciation /bæs/. We transcribe pronunciationsusing the International Phonetic Association’s (IPA) symbol set. Symbols used in this chapter are defined in Table 46.1. c  1999 by CRC Press LLC Speech synthesis breaks down into two parts: • The selection and concatenation of appropriate concatenative units given the phoneme string. • The synthesis of a speech waveform given the units, plus a model of the glottal source. 46.2 Text Analysis and Linguistic Analysis 46.2.1 Text Preprocessing The input to TTS systems is text encoded using an electronic coding scheme appropriate for the language,suchasASCII,JIS(Japanese),orBig-5(Chinese). OneofthefirsttasksfacingaTTSsystem is that of dividing the input into reasonable chunks, the most obvious chunk being the sentence. In somewritingsystemsthere is a designated symbol used formarkingthe end of a declarativesentence and for nothing else — in Chinese, for example, a small circle is used — and in such languages end-of-sentencedetection is generally not a problem. For English and other languages we are not so fortunate because a period, in addition to its use as a sentence delimiter, is also used, for example, to mark abbreviations: if one sees the period in Mr., one would not (nor mally) want to analyze this as an end-of-sentence marker. Thus, before one concludes that a period does in fact mark the end of a sentence, one needs to eliminate some other possible analyses. In a typical T TS system, text analysis would include an abbreviation-expansion module; this module is invoked to check for commonabbreviationswhichmightallowonetoeliminateoneormorepossibleperiodsfromfurther consideration. Forexample,ifapreprocessorfor Englishencountersthestring Mr. inanappropriate context (e.g., followed by a capitalized word), it w ill expand it as mister and remove the period. Of course, abbreviation expansion itself is not trivial, since many abbreviations are ambiguous. For example, is St. to be expanded as Street or Saint?IsDr., Doctor or Dr ive? Such cases can be disambiguated via a series of heuristics. For St., for example, the system might first check to see if the abbreviation is followed by a capitalized word (i.e., a potential name), in which case it would be expanded as Saint; otherwise, if it is preceded by a capitalized word, a number, or an alphanumeric (49th), it would be expanded as Street. Another problem that mustbe dealt with is the conversion of numbers into words: 232 should usually be expanded as two hundred thirty two, whereas if the same sequence occurs as part of 232-3142 — a likely telephone number — it would normally be read two three two. In languages like English, tokenization into words can to a large extent be done on the basis of white space. In contrast, in many Asian languages, including Chinese, the situation is not so simple because spaces are never used to delimit words. For the purposes of text analysis it is therefore generally necessary to “reconstruct” word boundary information. A minimal requirement for word segmentation is an on-line dictionary that enumerates the wordforms of the language. This is not enough on its own, however, since there are many words that will not be found in the dictionary; among these are personal names, foreign names in transliteration, and morphological derivatives of words that do not occur in the dictionar y. It is therefore necessary to build models of these non-dictionary words; see [4] for further discussion. In addition to lexical analysis, the text-analysis portion of a TTS system will typically perform syntactic analysis of various kinds. One commonly performed analysis is grammatical part-of- speechassignment, as information on the part of speech of words can be useful for accentuationand phrasing, among other things. Thus, in a sentence likethey can can cans, it is useful for accentuation purposes to know that the first can is a function word — an auxiliary verb, whereas the second and third arecontent words — respectivelya verb and a noun. There are a number of part-of-speech algorithmsavailable,perhapsthebestknownbeingthestochasticmethodof[5],whichcomputesthe c  1999 by CRC Press LLC most likely analysis of a sequence of words, maximizing the product of the lexical probabilities of the parts-of-speechinthesentence(i.e., thepossiblepartsofspeechofeachwordandtheirprobabilities), and the n-gram probabilities (probabilities of n-grams of parts of speech), which provide a model of the context. 46.2.2 Accentuation In languages like English, various words in a sentence are associated with accents, which are usually manifested as upward or downward movements of fundamental frequency. Usually, not every word in the sentence bears an a ccent, however, and the decision on which words should be accented and which should be unaccented is one of the problems that must be addressed as part of text analysis. It is common in prosodic analysis to distinguish three levels of prominence. Two are accented and unaccented, as just described, and the third is cliticized. Cliticized words are unaccented but in addition have lost their word stress, so that they tend to be durationally short: in effect, they behave like unstressed affixes, even though they are w ritten as separate words. A good first step in assig ning accents is to makethe accentual determination on the basis of broad lexicalcategoriesorpartsof speech. Contentwords— nouns,verbs,adjectives,andperhaps adverbs, tend in general to be accented; function words, including auxiliary verbs and prepositions tend to be deaccented; short function words tend to be cliticized. But accenting has a wider function than merely communicating lexical category distinctions between words. In English, one important set of constructions where accenting is more complicated than what might be inferred from the above discussion are complex noun phrases — basically, a noun preceded by one or more adjectival or nominal modifiers. In a “discourse-neutral” context, some constructions are accented on the final word (Madison Avenue), some on the penultimate (Wall Street, kitchen towel rack), and some on an evenearlierword (sump pump factory). Theassignment ofaccenttocomplexnounphrasesdepends on complex lexical and semantic factors; see [6]. Accenting is not only sensitive to syntactic structure and semantics, but also to properties of the discourse. One straightforward effect is contrast, as in the example I didn’t ask for cherry pie, I asked for apple pie. For most speakers, the “discourse neutral” accent would be on pie, but in this example thereisaclear intentiontocontrasttheingredientsinthepies,andpie isthusdeaccentedtoeffectthe contrast between cherry and apple. See [7] for a discussion of how these kind of effects are handled inaTTSsystemforEnglish. Note,whilehumanlikea ccentingcapabilitiesarepossiblein manycases, there are still some intractable problems. For example, just as one would often deaccent a word that had been previously mentioned, so would one often deaccent a word if a supercategory of that word hadbeenmentioned:My son wants a Labrador , but I’m allergic to dogs. Handling such cases in any general way is beyond the capabilities of cur rent TTS systems. 46.2.3 Word Pronunciation The next stage of analysis involves computing pronunciations for the words in the input, given the orthographic representation of those words. The simplest approach is to have a set of “letter-to- sound” rules that simply map sequences of graphemes into sequences of phonemes, along with possible diacritic information, such as stress placement. This approach is naturally best suited to languages wherethereis a relativelysimplerelationbetweenorthogr aphy and phonology: languages such as Spanish or Finnish fall into this category. However, languages likeEnglish manifestly do not, so it has generally been recognized that a highly accurate word pronunciation module must contain c  1999 by CRC Press LLC a pronouncing dictionary that, at the very least, records words whose pronunciation could not be predicted on the basis of general rules. However, having a dictionary that is merely a list of words presentsuswithfamiliarproblemsofcoverage: manytextwordsoccurthat arenottobefoundinthe dictionary, including morphological derivatives from known words, or previously unseen personal names. For morphological derivatives, standard techniques for morphological analysis [2, 8] can be ap- plied to achieve a morphological decomposition for a word. The pronunciation of the whole can then, in general, be computed from the (presumably know n) pronunciation of the morphological parts, applying appropriatephonologicalrulesof thelanguage. Fornovelpersonalnames, additional mechanisms may be necessary since novel names cannot always be related morphologically to previ- ously seen ones. Onesuch additional method involves computing the pronunciation of a new name by analogy with the pronunciation of a similar name [9, 10]. For example, imagine that we have the name Califano in our dictionary and that we know its pronunciation: then we could compute the pronunciation of a hypothetical name Balifano by noting that both names share the “suffix” alifano. The pronunciation of Balifano can then be computed by removing the phoneme /k/, corresponding to the letter C in Califano, and replacing it w ith the phoneme /b/. Therearesomewordformsthatareinherentlyambiguousin pronunciation,andforwhichaword pronunciationmoduleasjustdescribedcanonlyreturn a setofpossiblepronunciations, fromwhich one must then be chosen. A straightforward example is the word Chevy, which is most commonly pronounced /   εvi /, but is /t   εvi / in the name Chevy Chase, so in this case one could succeed by simply storing the bigram Chevy Chase. But n-gram models do not solve all cases of homograph TABLE46.1 IPA Symbols Used in this Chapter disambiguation. So, the word bass, is most likely to be pronounced /bæs/ in a “fishy” context like he wasfishingforbass,but/be j s/inamusicalcontextlikehe playsbass. Whatdefinesthecontextasbeing musical or “fishy” is not characterizable in terms of n-grams, but rather relates to the occurrence of certain words(e.g., fish, lake, boat vs. play, sing, orchestra) in a widercontext. A method proposedby c  1999 by CRC Press LLC Yarowsky [11, 12] allows for both local (n-gram) context and wide context to be used in homograph disambiguation, and excellent results have been achieved using this approach. 4 46.2.4 Intonational Phrasing In reading a long sentence, speakers will typically break the sentenceup into several phrases, each of which can be said to “stand alone” as an intonational unit. If punctuation is used liberally so that there are relatively few words between the commas, semicolons, or periods, then a reasonable guess at an appropriate phrasing would be simply to break the sentenceat the punctuation mar ks (though this is not always appropriate [13]). The real problem comes when long stretches occur without punctuation; in such cases, human readers would normally break the string of words into phrases, and the problem then arises of where to place these breaks. The simplest approach is to havealist of words,typicallyfunction words,that arelikelyindicators ofgoodplacestobreak[1]. Onehastousesomecaution,however,becausewhileaparticularfunction word such as and may coincide with a plausible phrase break in some cases (He got out of the car and walked towards the house), in other examplesitmight coincide with aparticularly poor place to break as in I was forced to sit through a dog and pony show that lasted most of Wednesday afternoon. Other approaches to intonational phrasing have been proposed in the literature, including methods that depend on syntactic parsers of various degrees of sophistication [13, 14]. An alternative approach, described in [15], uses a decision tree model [16, 17] that is trained on a corpus of text annotated with prosodic phrase-boundary information. 46.2.5 Segmental Durations Having computed which phonemes are to be produced by the synthesizer, it is necessary to decide how long to make each one. In this section we briefly describe the methods used for computing segmental durations: the reader is referred to [18] for an extended discussion of this topic. What duration to assign to a phonemic segment depends on many factors, including: • The identity of the segment in question. For example, in many dialects of English, the vowel /æ/ has a longer intrinsic duration than the vowel /ı/. • The stress of the syllable of which the segment is a member. For example, vowels in stressed syllables tend to be longer than vowels in unstressed syllables. • Whetherthesyllableofwhichthesegmentisamemberbearsanaccent. Accentedsyllables tend to be longer than otherwise identical unaccented syllables. • The quality of the surrounding segments. For example, a vowel preceding a voiced consonantinthesamesyllabletendstobelongerthanthesamevowelprecedingavoiceless consonant. • The position of the segment in the phrase: elements close to the ends of phrases tend to be longer than elements more internal to the phrase. VariousapproacheshavebeentakentomodelingsegmentaldurationsinTTSsystems. Onemethod involvesdurationrules,whicharerulesoftheform“ifthesegmentisX anditisinphrase-finalposition, then lengthen X by n milliseconds” [19, 20]. In rule-based systems of this kind, it is not unusual for the duration of a given segment to be rewritten several times as the conditions for the application of 4 Clearlytheabove-describedmethodforhomographdisambiguationcanalsobeappliedtootherformallysimilarproblems in TTS, such as whether St. to be expanded as Saint or Street,or747 is to be read as a number seven hundred and forty seven or the name of an aircraft seven forty seven. c  1999 by CRC Press LLC the various rules are considered. T he r ule-based approach can be formalized explicitly in terms of the second approach — duration models — which are mathematical expressions that prescribe how the various conditioning factors are to be used in computing the duration of a segment [19]; the successiveapplicationoftherulescan,ineffect,be“compiled”into a single mathematical expression that implements the combined effect of the rules. As argued in [18], all extant duration models can be viewed as instances of a more general sum-of-products model, where the duration of a segment is predicted by a formula of the general form: DUR(f) =  i∈T  j∈I i S i,j (f j ) (46.1) Here the duration assigned to a feature vector — DUR(f) — is computed by scaling each factor f j in the ith product term by a factor scale S i,j ; computing the product of all scaled factors within each product term; and then summing over all i product terms. Rather than deciding a priori on a particular sums-of-products model (or set of such models) within the space of all possible models, one approach taken to segmental dur ation is to use exploratory data analysis to arrive at models whose predictions show a good fit to durations from a corpus of labeled speech [18]. More specifically, we start with a text corpus that (ideally) has a good coverage both of various phonemes and of the factors (and their combinations) that are deemed likely to be relevant for du- ration. A native speaker of the language reads this text and the speech is segmented and labeled. Using the text-analysismodules of TTS, with some possible hand correction, weautomatically com- pute the sequenceof phonemes, and the feature vectors (including featureson stress, accent,phrasal position, etc.) associated with each phoneme. Given the feature vectors, various sums-of-products models are compared and their predictions of the values of the observed segmental durations are evaluated. In general, different specific dur ation models may be better suited to different sets of conditions than others: for example, in the English duration system, intervocalic consonants are associated with a different sums-of-products model than consonants that occur in clusters. In the actual implementation of segmental duration predictions, a decision tree is used to determine, on thebasisof contextual factorsappropriatetothesegmentathand, what particular sums-of-products model to use; this model is then used to compute the duration of the segment. Designing a corpus with good coverage of relevant factors is a non-trivial task in itself: the basic problem is to provide a set that has maximal coverage with the minimal amount of text to be read by a speaker, and analyzed. T he method that we use involves starting with a large corpus of text in a languageandautomaticallypredictingthephonemicsegmentsalongwiththeirfeatures(again,using textanalysiscomponentsforthelanguage). A greedy algorithm is then applied to arriveataminimal set of sentences that have good (ideally total) coverage of the desired feature vectors. 46.2.6 Intonation Having computed linguistic information such as the sequence of segments to be produced, their duration, the prominence of the various words, and the locations of prosodic boundaries, the next thingthataTTSsystemneedstocomputeisanintonationcontour. Therearealmostasmanymodels of intonation implemented in TTS systems as there are TTS systems, and we do not have the space to review these different approaches here. Suffice it to say that most intonation models that have actually been incorporated into working TTS systems can be classified into one of three “schools”: • The Fujisaki school [21, 22, inter alia]. An intonation contour for a phrase is computed from a phrase impulse and some number of accent impulses. These impulses are con- volved with a smoothing function to produce phrase and accent curves, which are then summed to produce the final contour. c  1999 by CRC Press LLC • The Dutch school [23]. Intonation contours are represented as sequences of connected line segments which are chosen so as to perceptually closely approximate real (smooth) intonation contours. • The autosegmental/metrical school [24, 25, inter alia]. Intonation contours are repre- sented abstractly as sequences of high and low targets. The computation of an intonation contour from a phonological representation can be illust rated byconsideringtheBellLabsEnglishTTSsystem,whichcurrentlyusesaversionofthePierrehumbert autosegmental model [26, 27, 28]. As the first stage in the computation of an intonation contour, a tone-timing function sets up nominal times for each accent in the sentence. Separate routines are called for initial boundary tones, final boundary tones, pitch accents and phrase accents. Roughly, initial boundary tones are aligned with the silence that is placed at the beginning of each minor phrase, whereas final boundary tones are alig ned with the final vowel of each minor phrase. Phrase accents are aligned after the final word accent of the minor phrase, if there is one; otherwise at the end of the first vowel of the first word, or else at the end of the first phoneme. Finally, accents on words are aligned with their associated syllables using a complex set of contextual factors. These nominal accent times are then converted into actual F0/time pairs, by another function. F0 values are computed dependent on the prominence of the accent (either determined automatically, or else definable by the user), and various phrasal parameters from the intonation model, as well as the particular type of accent involved. Finally, an F0 contour is produced by interpolating the computed pitch/time pairs, and smoothing via convolution with a rectangular window. 46.3 Speech Synthesis Oncethetexthasbeent ransformedintophonemes,andtheirassociateddurationsandafundamental frequency contour have been computed, the system is ready to compute the speech parameters for synthesis. There are two independent variables in the choice of parametric computation in a TTS system. Onevariable is thechoicebetween a rule-based scheme for the computationoftheparametersonthe onehand,andaconcatenativescheme involvingconcatenationofshortsegmentsofpreviouslyuttered speech on the other. The second variable is the actual parametric representation chosen: possible choices include articulatory parameters, formants, LPC (linear predictive coding), spectr al param- eters, or time domain parameters. In a concatenative scheme, any parametric representation that permits independent controlofloudness, F0, voicing, timing, and possibly spectral manipulationsis appropriate. Rule-basedsystems are more restrictive of the choice of parameters since such schemes rely both on our understanding of the relation between the parameters and the acoustic signals they represent, and on our ability to compute the dynamics of the parameters as they move from one sound to another. Thus far only articulatory parameters and formants have been used in rule-based systems. The best-known examples of a formant-based synthesizer are the Klatt synthesizer and its commercial offshoot DECtalk. Rule-based systems are space-efficient because the y eliminate the need to store speech segments. Rule-based approaches also make it easier, in principle, to implement new speaker characteristics for different voices, as well as different phone inventories for new dialects and languages. However, sincethedynamicsofthe parameters arevery difficult to model it requiresa great deal more effort to produce a rule-based system than it does to produce a concatenative system of comparable quality. Given the right choice of units, a concatenative scheme is able to store the dynamics of the speech signal and thus produce high quality synthetic sound. The choice of the exact parameters depends on what the designer values in such a system. Waveform representations — such as PSOLA [29]— have a high sound quality, but they are limiting in terms of the ability to alter the sound, and thus c  1999 by CRC Press LLC FIGURE 46.1: Source-filter model for speech synthesis. far, no one has been able to change the spectral parameters in a time domain system. Articulatory parameters or formants, on the other hand, can be successfully manipulated. However, the speech quality produced by using these parameters is somewhat degraded because there are no reliable methods to extract these parameters and even in a plain coding application (analysis and resynthesis without manipulations) these methods produce degradation of the speech signal. Othersystemsuseaconcatenativeapproach. Inthisapproach,parametrizedshortspeechsegments of natural speech are connected to form a representation of the synthetic speech. The majorit y of the natural speech segments are merely transitions between pairs of phonemes. However, due to the large contextual variation of some phonemes, some segments consisting of three or more phoneme elements are often necessary; such elements consist of the transition from the first phoneme to the second, and a transition from the penultimate phoneme to the last, but the intermediate phonemes arestoredcompletely. FortheBellLabsEnglishsystem,thereareapproximately2900differentspeech elements — also called dyads — in the acoustic inventory, and these elements are sufficient to make up all the legal phoneme combinations for English. The concatenative approach to speech synthesis requires that speech samples be stored in some parametric representationthatwillbesuitableforconnecting thesegments and changing the signal’s characteristicsofloudness,F0,andspectrum. Onemethodforchangingthecharacteristicsofnatural speech is to analyze the speech in terms of a source/filter model, as diagramed in Fig. 46.1. Thismodelofspeechsynthesishasavariety ofindependentinputcontrols. Starting at the leftside ofthefigure,weshowtwopossible sourcegenerators: a noisegeneratoranda simplepulsegenerator. The noise generator has no controlling input whereas the pulse generator is controlled by the F0 parameter; the F0 parameter specifies the distance between any two pulses thus controlling the F0 of the periodic source. These inputs are selected by a switch which is controlled by a voicing flag. It is also possible to have a mixer to control the relative contribution of the noise and pulse source, and to insert a glottal pulse with additional controls for the shape of the glottal source in place of the simple pulse generator. Tothe right of the switch, we have a multiplier that multipliesthe source by an amplitude parameter. This serves as the loudness control for the system. The signal from the multiplier is fed into a filter controlled by the filter coefficients which are varied slowly to shape the speech spect rum. The source/filter model can be used to replicate naturally spoken speech when the parameters are obtained by analysis of natural speech. Speech can be parametrized by an amplitude control, voiced/voicelessflag, F0,andfiltercoefficientsatasmallinterval(ontheorderof5msec. to15msec.) Theloudnesscontrolisdeterminedfromthepowerofthespeechatthetimeframeof theanalysis. F0 extraction algorithms determinethe voicingofthe speech as well as the fundamental frequency. The filter parameters can be determined by various analysis techniques. The parameters obtained from the analysis can be used to drive the source filter model to reproduce the analyzed speech. However, these parameters can also be varied independently to change the speech. The ability to alter the analysis parameters is crucial to a concatenative approach, where the spectral parameters have to be smoothedandinterpolatedwhenevertwoelementsfromdifferentutterancesareconnected,orwhen c  1999 by CRC Press LLC [...]... International Conference on Acoustics, Speech, and Signal Processing, vol.1, ICASSP, San Diego, 2.8.1–2.8.4, 1984 [28] Silverman, K., Utterance-internal prosodic boundaries, in Proceedings of the Second Australian International Conference on Speech Science and Technology, Sydney, Australia, 86–91, 1988 [29] Charpentier, F and Moulines, E., Pitch-synchronous waverform processing techniques for text-to-speech... of the Second ESCA/IEEE Workshop on Speech Synthesis, 1994 [12] Sproat, R., Hirschberg, J and Yarowsky, D., A corpus-based synthesizer, in Proceedings of the International Conference on Spoken Language Processing, Banff, ICSLP, 563–566, Oct 1992 c 1999 by CRC Press LLC [13] O’Shaughnessy, D., Parsing with a small dictionary for applications such as text to speech, Computational Linguistics, 15, 97–108,... Proceedings of 32nd Annual Meeting, 66–73, 1994 [5] Church, K., A stochastic parts program and noun phrase parser for unrestricted text, in Proceedings of the Second Conference on Applied Natural Language Processing, Association for Computational Linguistics, Morristown, NJ, 136-143, 1988 [6] Sproat, R., English noun-phrase accent prediction for text-to-speech, Computer Speech and Language, 8, 79–94, 1994 . Sproat, R. & Olive, J. “Text-to-Speech Synthesis” Digital Signal Processing Handbook Ed. Vijay K. Madisetti and Douglas B. Williams Boca Raton:. in Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, vol.1, ICASSP, San Diego, 2.8.1–2.8.4, 1984. [28] Silverman,K.,Utterance-internalprosodicboundaries,in ProceedingsoftheSecondAustralian International

Ngày đăng: 25/01/2014, 13:20

TỪ KHÓA LIÊN QUAN

w