Tools for mapping between written words and phonetic forms are essential components in many applications of speech technology, such as automatic speech recognition (ASR) and speech synthesis (TTS). Simple converters can be derived from annotated speech corpora using machine learning, and such tools are available for almost all European languages and a great number of others. Whereas their performance is adequate for ASR and for low-quality TTS, their lack of precision makes them unfit for linguistic research purposes such as phonetic annotation of spontaneous speech recordings. A common method of enhancing their predictive power (e.g. faced with out-of-vocabulary tokens) is to include phonetic and lexical rules, and sometimes even semantic and contextual knowledge. In this paper we present some of the principles underlying the typical linguistically informed phonetic converter. We illustrate our points with examples from the Danish grapheme-to-phoneme converter Phonix.
|Number of pages||1|
|Publication status||Published - 2014|
|Event||2014 CRITT - WCRE Conference: Translation in Transition: Between Cognition, Computing and Technology - Copenhagen Business School, Frederiksberg, Denmark|
Duration: 30 Jan 2014 → 31 Jan 2014
|Conference||2014 CRITT - WCRE Conference|
|Location||Copenhagen Business School|
|Period||30/01/2014 → 31/01/2014|