


default search action
ICSLP 1992: Banff, Alberta, Canada
- The Second International Conference on Spoken Language Processing, ICSLP 1992, Banff, Alberta, Canada, October 13-16, 1992. ISCA 1992

- Peter Ladefoged:

Knowing enough to analyze spoken languages. 1-4 - Renato de Mori, Roland Kuhn:

Speech understanding strategies based on string classification trees. 441-448 - Patricia K. Kuhl:

Infants' perception and representation of speech: development of a new theory. 449-456 - Hajime Hirose:

The behavior of the larynx in spoken language production. 457-458
Word Spotting in ASR
- Eduardo Lleida, José B. Mariño, Josep M. Salavedra, Antonio Bonafonte:

Syllabic fillers for Spanish HMM keyword spotting. 5-8 - Yasuhiro Komori, David Rain Ton:

Minimum error classification training for HMM-based keyword spotting. 9-12 - Gregory J. Clary, John H. L. Hansen:

A novel speech recognizer for keyword spotting. 13-16 - Herbert Gish, Kenney Ng, Jan Robin Rohlicek:

Secondary processing using speech segments for an HMM word spotting system. 17-20 - Ming-Whei Feng, Baruch Mazor:

Continuous word spotting for applications in telecommunications. 21-24 - Maurizio Copperi:

A low bit-rate CELP coder based on multi-path search methods. 25-28
Speech Coding
- Katsushi Seza, Hirohisa Tasaki, Shinya Takahashi:

Fully vector quantized arm a analysis combined with glottal model for low bit rate coding. 29-32 - Erdal Paksoy, Wai-Yip Chan, Allen Gersho:

Vector quantization of speech LSF parameters with generalized product codes. 33-36 - Yair Shoham:

Low-rate speech coding based on time-frequency interpolation. 37-40 - Tomohiko Taniguchi, Yoshinori Tanaka, Yasuji Ohta, Fumio Amano:

Improved CELP speech coding at 4 kbit/s and below. 41-44
Speech Production: Coarticulation
- Antonio Bonafonte, José B. Mariño, Montse Pardàs:

Efficient integration of coarticulation and lexical information in a finite state grammar. 45-48 - Herbert A. Leeper, A. P. Rochet, Ian R. A. MacKay:

Characteristics of nasalance in canadian speakers of English and French. 49-52 - Christine H. Shadle, Andre Moulinier, Christian U. Dobelke, Celia Scully:

Ensemble averaging applied to the analysis of fricative consonants. 53-56 - Andrew Slater, Sarah Hawkins:

Effects of stress and vowel context on velar stops in british English. 57-60 - Emanuela Magno Caldognetto, Kyriaki Vagges, Giancarlo Ferrigno, Maria Grazia Busà:

Lip rounding coarticulation in Italian. 61-64
Auditory Models
- Paula M. T. Smeele, Anne C. Sittig, Vincent J. van Heuven:

Intelligibility of audio-visually desynchronised speech: asymmetrical effect of phoneme position. 65-68 - Unto K. Laine:

Speech analysis using complex orthogonal auditory transform (coat). 69-72 - Yuqing Gao, Taiyi Huang, Shaoyan Chen, Jean Paul Haton:

Auditory model based speech processing. 73-76 - Gary N. Tajchman, Nathan Intrator:

Phonetic classification of timit segments preprocessed with lyon's cochlear model using a supervised/unsupervised hybrid neural network. 77-80 - Thomas Holton, Steve Love, Stephen P. Gill:

Formant and pitch-pulse detection using models of auditory signal processing. 81-84
Recognition of Telephone Speech
- Hynek Hermansky, Nelson Morgan:

Towards handling the acoustic environment in spoken language processing. 85-88 - Alberto Ciaramella, Davide Clementino, Roberto Pacifici:

Real-time speaker-independent large-vocabulary CDHMM-based continuous telephonic speech recognizer. 89-92 - Matthew Lennig, Douglas Sharp, Patrick Kenny, Vishwa Gupta, Kristin Precoda:

Flexible vocabulary recognition of speech. 93-96 - Benjamin Chigier, Hong C. Leung:

The effects of signal representations, phonetic classification techniques, and the telephone network. 97-100
Text-to-Speech Synthesis 1, 1
- Leon Gulikers, Rijk Willemse:

A lexicon for a text-to-speech system. 101-104 - Rijk Willemse, Leon Gulikers:

Word class assignment in a text-to-speech system. 105-108 - Gösta Bruce, Björn Granström, Kjell Gustafson, David House:

Aspects of prosodic phrasing in Swedish. 109-112 - Kirk P. H. Sullivan, Robert I. Damper:

Synthesis-by-analogy: a bilingual investigation using German and English. 113-116 - Leonard C. Manzara, David R. Hill:

Degas: a system for rule-based diphone speech synthesis. 117-120 - Shyam Sunder Agrawal, Kenneth N. Stevens:

Towards synthesis of Hindi consonants using KLSYN88. 177-180 - Louis C. W. Pols:

Multi-lingual synthesis evaluation methods. 181-184 - Björn Granström, Petur Helgason, Hoskuldur Thrainsson:

The interaction of phonetics, phonology and morphology in an icelandic text-to-speech system. 185-188
Voice Source Characteristics
- Helmer Strik, Joop Jansen, Louis Boves:

Comparing methods for automatic extraction of voice source parameters from continuous speech. 121-124 - Jacques C. Koreman, Louis Boves, Bert Cranen:

The influence of linguistic variations on the voice source characteristics. 125-128 - Sarah K. Palmer, Jill House:

Dynamic voice source changes in natural and synthetic speech. 129-132 - Satoshi Imaizumi, Jan Gauffin:

Acoustic and perceptual modelling of the voice quality caused by fundamental frequency perturbation. 133-136 - Shigeru Kiritani, Hiroshi Imagawa, Hajime Hirose:

Vocal cord vibration during consonants - high-speed digital imaging using a fiberscope. 1661-1664
Speech Perception: Higher-Order Processes 1, 1
- David R. Traum, James F. Allen:

A "speech acts" approach to grounding in conversation. 137-140 - Sheila Meltzer:

Antecedent activation by empty pronominals in Spanish. 141-144 - Ron Smyth:

Multiple feature matching in pronoun resolution: a new look at parallel function. 145-148 - Keh-Yih Su, Jing-Shin Chang, Yi-Chung Lin:

A discriminative approach for ambiguity resolution based on a semantic score function. 149-152 - Nobuaki Minematsu, Sumio Ohno, Keikichi Hirose, Hiroya Fujisaki:

The influence of semantic and syntactic information on spoken sentence recognition. 153-156 - Lynne C. Nygaard, Mitchell Sommers, David B. Pisoni:

Effects of speaking rate and talker variability on the representation of spoken words in memory. 209-212 - Hugo Quené, Yvette Smits:

On the absence of word segmentation at "weak" syllables. 213-216 - Mitchell Sommers, Lynne C. Nygaard, David B. Pisoni:

Stimulus variability and the perception of spoken words: effects of variations in speaking rate and overall amplitude. 217-220 - James M. McQueen, Anne Cutler:

Words within words: lexical statistics and lexical access. 221-224
Speaker-Independent Word Recognition
- Stephan Euler, Joachim Zinke:

Experiments on the use of the generalized probabilistic descent method in speech recognition. 157-160 - Ricardo de Córdoba, José Manuel Pardo, José Colás:

Improving and optimizing speaker independent, 1000 words speech recognition in Spanish. 161-164 - John F. Pitrelli, David M. Lubensky, Benjamin Chigier, Hong C. Leung:

Multiple-level evaluation of speech recognition systems. 165-168 - Tatsuya Kimura, Mitsuru Endo, Shoji Hiraoka, Katsuyuki Niyada:

Speaker independent word recognition using continuous matching of parameters in time-spectral form based on statistical measure. 169-172 - R. Roddeman, H. Drexler, Louis Boves:

Automatic derivation of lexical models for a very large vocabulary speech recognition system. 173-176
Human Factors
- Anne Cutler, Tony Robinson:

Response time as a metric for comparison of speech recognition by humans and machines. 189-192 - S. M. (Raj) Ulagaraj:

Characterization of directory assistance operator-customer dialogues in AGT limited. 193-196 - Sheri Hunnicutt, Lynette Hirschman, Joseph Polifroni, Stephanie Seneff:

Analysis of the effectiveness of system error messages in a human-machine travel planning task. 197-200 - David Goodine, Lynette Hirschman, Joseph Polifroni, Stephanie Seneff, Victor Zue:

Evaluating interactive spoken language systems. 201-204 - Ute Jekosch:

The cluster-identification test. 205-208
Continuous Speech Recognition 1, 2
- Patrick Kenny, Rene Hollan, Gilles Boulianne, Harinath Garudadri, Yan Ming Cheng, Matthew Lennig, Douglas D. O'Shaughnessy:

Experiments in continuous speech recognition with a 60, 000 word vocabulary. 225-228 - Gilles Boulianne, Patrick Kenny, Matthew Lennig, Douglas D. O'Shaughnessy, Paul Mermelstein:

HMM training on unconstrained speech for large vocabulary, continuous speech recognition. 229-232 - David Rainion, Shigeki Sagayama:

Appropriate error criterion selection for continuous speech HMM minimum error training. 233-236 - Akito Nagai, Kenji Kita, Toshiyuki Hanazawa, Tadashi Suzuki, Tomohiro Iwasaki, Tsuyoshi Kawabata, Kunio Nakajima, Kiyohiro Shikano, Tsuyoshi Morimoto, Shigeki Sagayama, Akira Kurematsu:

Hardware implementation of realtime 1000-word HMM-LR continuous speech recognition. 237-240 - Madeleine Bates, Robert J. Bobrow, Pascale Fung, Robert Ingria, Francis Kubala, John Makhoul, Long Nguyen, Richard M. Schwartz, David Stallard:

Design and performance of HARC, the BBN spoken language understanding system. 241-244 - Otoya Shirotsuka, G. Kawai, Michael Cohen, Jared Bernstein:

Performance of speaker-independent Japanese recognizer as a function of training set size and diversity. 297-300 - Kouichi Yamaguchi, Shigeki Sagayama, Kenji Kita, Frank K. Soong:

Continuous mixture HMM-LR using the a* algorithm for continuous speech recognition. 301-304 - Kenji Kita, Tsuyoshi Morimoto, Kazumi Ohkura, Shigeki Sagayama:

Continuously spoken sentence recognition by HMM-LR. 305-308 - Akinori Ito, Shozo Makino:

Word pre-selection using a redundant hash addressing method for continuous speech recognition. 309-312 - Andrej Ljolje, Michael D. Riley:

Optimal speech recognition using phone recognition and lexical access. 313-316
Natural Language Processing and Speech Understanding 1-3
- Nick Waegner, Steve J. Young:

A trellis-based language model for speech recognition. 245-248 - Carla B. Zoltowski, Mary P. Harper, Leah H. Jamieson, Randall A. Helzerman:

PARSEC: a constraint-based framework for spoken language understanding. 249-252 - Gareth J. F. Jones, Jeremy H. Wright, E. N. Wrigley:

The HMM interface with hybrid grammar-bigram language models for speech recognition. 253-256 - Atsuhiko Kai, Seiichi Nakagawa:

A frame-synchronous continuous speech recognition algorithm using a top-down parsing of context-free grammar. 257-260 - Fernando Pereira, David B. Roe:

Empirical properties of finite state approximations for phrase structure grammars. 261-264 - Stephanie Seneff, Helen M. Meng, Victor Zue:

Language modelling for recognition and understanding using layered bigrams. 317-320 - David Goddeau:

Using probabilistic shift-reduce parsing in speech recognition systems. 321-324 - Tim Howells, David Friedman, Mark A. Fanty:

Broca, an integrated parser for spoken language. 325-328 - P. V. S. Rao, Nandini Bondale:

Blank slate language processor for speech recognition. 329-332 - Eric Jackson:

Integrating two complementary approaches to spoken language understanding. 333-336 - Marcello Pelillo, Mario Refice:

Learning compatibility coefficients for word-class disambiguation relaxation processes. 389-392 - Kaichiro Hatazaki, Jun Noguchi, Akitoshi Okumura, Kazunaga Yoshida, Takao Watanabe:

INTERTALKER: an experimental automatic interpretation system using conceptual representation. 393-396 - Tsuyoshi Morimoto, Toshiyuki Takezawa, Kazumi Ohkura, Masaaki Nagata, Fumihiro Yato, Shigeki Sagayama, Akira Kurematsu:

Enhancement of ATR's spoken language translation system: SL-TRANS2. 397-400 - Tsuyoshi Morimoto:

Continuous speech recognition using a combination of syntactic constraints and dependency relationship. 401-404 - Roberto Pieraccini, Zakhar Gorelov, Esther Levin, Evelyne Tzoukermann:

Automatic learning in spoken language understanding. 405-408
Language Learning and Acquisition 1, 2
- Michael P. Robb, Harold R. Bauer:

Prespeech and early speech coarticulation: american English and Japanese characteristics. 265-268 - Hiroaki Kojima, Kazuyo Tanaka, Satoru Hayamizu:

Formation of phonological concept structures from spoken word samples. 269-272 - Bernard L. Rochet, Fangxin Chen:

Acquisition of the French VOT contrasts by adult speakers of Mandarin Chinese. 273-276 - Michael Gasser:

Phonology as a byproduct of learning to recognize and produce words: a connectionist model. 277-280 - Michael S. Hurlburt, Judith C. Goodman:

The development of lexical effects on children's phoneme identifications. 337-340 - Pierre A. Hallé, Benedicte de Boysson-Bardies:

Word recognition before production of first words? 341-344 - Toshisada Deguchi, Shigeru Kiritani, Akiko Hayashi, Fumi Katoh:

The effect of fundamental frequency for vowel perception in infants. 345-348
Speech Perception: Units of Processing
- William C. Treurniet:

Objective measurement of phoneme similarity. 281-284 - Michael D. Riley, Andrej Ljolje:

Recognizing phonemes vs. recognizing phones: a comparison. 285-288 - Bruce L. Derwing, Terrance M. Nearey, R. A. Beinert, T. A. Bendrien:

On the role of the segment in speech processing by human listeners: evidence from speech perception and from global sound similarity judgments. 289-292 - Grace E. Wiebe, Bruce L. Derwing:

The syllabic status of postvocalic resonants in an unwritten low German dialect. 293-296
Prosody: The Phrase and Beyond 1, 2
- Agaath M. C. Sluijter, Vincent J. van Heuven, Anneke Neijt:

The influence of focus distribution and lexical stress on the temporal organisation of the syllable. 349-352 - Agaath M. C. Sluijter, Jacques M. B. Terken:

The development and perceptive evaluation of a model for paragraph intonation in dutch. 353-356 - Nobuyoshi Kaiki, Yoshinori Sagisaka:

Pause characteristics and local phrase-dependency structure in Japanese. 357-360 - Bernd Möbius, Matthias Pätzold:

F0 synthesis based on a quantitative model of German intonation. 361-364 - Kenneth N. Ross, Mari Ostendorf, Stefanie Shattuck-Hufnagel:

Factors affecting pitch accent placement. 365-368 - Marc Swerts, Ronald Geluykens, Jacques M. B. Terken:

Prosodic correlates of discourse units in spontaneous speech. 421-424 - Shin'ya Nakajima, James F. Allen:

Prosody as a cue for discourse structure. 425-428 - Barbara J. Grosz, Julia Hirschberg:

Some intonational characteristics of discourse structure. 429-432 - Hiroya Fujisaki, Keikichi Hirose, Haitao Lei:

Prosody and syntax in spoken sentences of standard Chinese. 433-436 - Kathleen Bishop:

Modeling sentential stress in the context of a large vocabulary continuous speech recognizer. 437-440
Speaker Adaptation
- Kazumi Ohkura, Masahide Sugiyama, Shigeki Sagayama:

Speaker adaptation based on transfer vector field smoothing with continuous mixture density HMMs. 369-372 - Tatsuo Matsuoka, Kiyohiro Shikano:

Speaker adaptation by modifying mixture coefficients of speaker-independent mixture Gaussian HMMs. 373-376 - Yifan Gong, Olivier Siohan, Jean Paul Haton:

Minimization of speech alignment error by iterative transformation for speaker adaptation. 377-380 - Hiroaki Hattori, Shigeki Sagayama:

Vector field smoothing principle for speaker adaptation. 381-384 - Tetsunori Kobayashi, Katsuhiko Shirai:

Spectral mapping onto probabilistic domain using neural networks and its application to speaker adaptive phoneme recognition. 385-388
Pronunciation Training
- Jean-Paul Lefèvre, Mervyn A. Jack, Claudio Maggio, Mario Refice, Fabio Gabrieli, Michelina Savino, Luigi Santangelo:

An interactive system for automated pronunciation improvement. 409-412 - Edmund Rooney, Steven M. Hiller, John Laver, Mervyn A. Jack:

Prosodic features for automated pronunciation improvement in the spell system. 413-416 - Maria-Gabriella Di Benedetto, Fabrizio Carraro, Steven M. Hiller, Edmund Rooney:

Vowels pronunciation assessment in the spell system. 417-420
Self-Organizing Systems in ASR 1, 2
- Franck Poirier:

Self-organizing map with supervision for speech recognition. 459-462 - Gregory R. De Haan, Ömer Egecioglu:

Topology preservation for speech recognition. 463-466 - Gary Bradshaw, Alan Bell:

Towards the performance limits of connectionist feature detectors. 467-470 - Helge B. D. Sørensen:

Context-dependent and -independent self-structuring hidden control models for speech recognition. 471-474 - Marie-José Caraty, Claude Montacié, Claude Barras:

Integration of frequential and temporal structurations in a symbolic learning system. 475-478 - Enric Monte, José B. Mariño, Eduardo Lleida:

Smoothing hidden Markov models ay means of a self organizing feature map. 535-538 - Jyri Mäntysalo, Kari Torkkola, Teuvo Kohonen:

LVQ-based speech recognition with high-dimensional context vectors. 539-542 - Mikko Kurimo, Kari Torkkola:

Application of self-organizing maps and LVQ in training continuous density hidden Markov models for phonemes. 543-546 - Paul Dalsgaard, Ove Andersen:

Identification of mono- and poly-phonemes using acoustic-phonetic features derived by a self-organising neural network. 547-550 - Pekka Utela, Samuel Kaski, Kari Torkkola:

Using phoneme group specific LVQ-codebooks with HMMs. 551-554
Speech Synthesis 1-3
- Naoto Iwahashi, Yoshinori Sagisaka:

Speech segment network approach for an optimal synthesis unit set. 479-482 - Yoshinori Sagisaka, Nobuyoshi Kaiki, Naoto Iwahashi, Katsuhiko Mimura:

ATR μ-talk speech synthesis system. 483-486 - Bert Van Coile, Steven Leys, Luc Mortier:

On the development of a name pronunciation system. 487-490 - Inger Karlsson:

Consonants for female speech synthesis. 491-494 - Jan P. H. van Santen:

Diagnostic perceptual experiments for text-to-speech system evaluation. 555-558 - Marcello Balestri, Enzo Foti, Luciano Nebbia, Mario Oreglia, Pier Luigi Salza, Stefano Sandri:

Comparison of natural and synthetic speech intelligibility for a reverse telephone directory service. 559-562 - Richard Sproat, Julia Hirschberg, David Yarowsky:

A corpus-based synthesizer. 563-566 - Tomohisa Hirokawa, Kenzo Itoh, Hirokazu Sato:

High quality speech synthesis based on wavelet compilation of phoneme segments. 567-570 - David R. Williams, Corine A. Bickley, Kenneth N. Stevens:

Inventory of phonetic contrasts generated by high-level control of a formant synthesizer. 571-574 - Mikael Goldstein, Ove Till:

Is % overall error rate a valid measure of speech synthesiser and natural speech performance at the segmental level? 1131-1134 - Willy Jongenburger, Renée van Bezooijen:

Text-to-speech conversion for dutch: comprehensibility and acceptability. 1135-1138 - Masayo Katoh, Shin'ichiro Hashimoto:

The rhythm rules in Japanese based on the centers of energy gravity of vowels. 1139-1142 - Kenzo Itoh, Tomohisa Hirokawa, Hirokazu Sato:

Segmental power control for Japanese speech synthesis. 1143-1146 - Jean Schoentgen:

Glottal waveform synthesis with volterra shapers. 1147-1150 - Ken Ceder, Bertil Lyberg:

Yet another rule compiler for text-to-speech conversion? 1151-1154 - Kazuhiko Iwata, Yukio Mitome:

Prosody generation models constructed by considering speech tempo influence on prosody. 1155-1158 - Alex I. C. Monaghan:

Extracting microprosodic information from diphones - a simple way to model segmental effects on prosody for synthetic speech. 1159-1162 - Arjan van Hessen:

Generation of natural sounding speech stimuli by means of linear cepstral interpolation. 1163-1166 - W. Nick Campbell, Colin W. Wightman:

Prosodic encoding of syntactic structure for speech synthesis. 1167-1170 - Susan R. Hertz, Marie K. Huffman:

A nucleus-based timing model applied to multi-dialect speech synthesis by rule. 1171-1174 - Jill House, Nick J. Youd:

Evaluating the prosody of synthesized utterances within a dialogue system. 1175-1178 - Marcel Tatham, Eric Lewis:

Prosodics in a syllable-based text-to-speech synthesis system. 1179-1182 - Rabia Belrhali, Véronique Aubergé, Louis-Jean Boë:

From lexicon to rules: toward a descriptive method of French text-to-phonetics transcription. 1183-1186 - Marianne Elmlund, Ida Frehr, Niels Reinholt Petersen:

Formant transformation from male to female synthetic voices. 1187-1190 - Panagiotis A. Rentzepopoulos, George K. Kokkinakis:

Multilingual phoneme to grapheme conversion system based on HMM. 1191-1194 - Noriyo Hara, Hisayoshi Tsubaki, Hisashi Wakita:

Fundamental frequency control using linguistic information. 1195-1198 - Andrew P. Breen:

A comparison of statistical and rule based methods of determining segmental durations. 1199-1202 - J. R. Andrews, K. Mervyn Curtis, Volker Kraft:

Generation and extraction of high quality synthesis units. 1203-1206 - T. I. Boogaart, Kim E. A. Silverman:

Evaluating the overall comprehensibility of speech synthesizers. 1207-1210 - Olivier Boëffard, Laurent Miclet, S. White:

Automatic generation of optimized unit dictionaries for text to speech synthesis. 1211-1214 - Hideki Kasuya, Seiki Kasuya:

Relationships between syllable, word and sentence intelligibilities of synthetic speech. 1215-1218 - David R. Hill, Craig-Richard Schock, Leonard C. Manzara:

Unrestricted text-to-speech revisited: rhythm and intonation. 1219-1222 - Anton J. Rozsypal:

Wavelet speech synthesizer in the classroom and speech laboratory. 1223-1226 - Thomas Portele, Birgit Steffan, Rainer Preuß, Walter F. Sendlmeier, Wolfgang Hess:

HADIFIX - a speech synthesis system for German. 1227-1230 - Cristina Delogu, Stella Conte, Andrea Paoloni, Ciro Sementina:

Two different methodologies for evaluating the comprehension of synthetic passages. 1231-1234 - Carlos Gussenhoven, Toni C. M. Rietveld:

A target-interpolation model for the intonation of dutch. 1235-1238
Speech Perception: Phonetic Processes 1, 2
- Katharine Davis, Patricia K. Kuhl:

Best exemplars of English velar stops: a first report. 495-498 - Kenneth N. Stevens, Sharon Y. Manuel, Stefanie Shattuck-Hufnagel, Sharlene Liu:

Implementation of a model for lexical access based on features. 499-502 - Dieter Huber:

Perception of aperiodic speech signals. 503-506 - Hiroaki Kato, Minoru Tsuzaki, Yoshinori Sagisaka:

Acceptability and discrimination threshold for distortion of segmental duration in Japanese words. 507-510 - Anne Bonneau, Sylvie Coste, Linda Djezzar, Yves Laprie:

Two level acoustic cues for consistent stop identification. 511-514 - Rolf Carlson, James R. Glass:

Vowel classification based on analysis-by-synthesis. 575-578 - Maria-Gabriella Di Benedetto, Jean-Sylvain Liénard:

Extrinsic normalization of vowel formant values based on cardinal vowels mapping. 579-582 - Terrance M. Nearey:

Applications of generalized linear modeling to vowel data. 583-586 - David B. Pisoni:

Some comments on invariance, variability and perceptual normalization in speech perception. 587-590 - Stephen D. Goldinger, Thomas J. Palmeri, David B. Pisoni:

Words and voices: perceptual details are preserved in lexical representations. 591-594
Speech Enhancement
- Yan Ming Cheng, Douglas D. O'Shaughnessy, Peter Kabal:

Speech enhancement using a statistically derived filter mapping. 515-518 - V. Beattie, Steve J. Young:

Hidden Markov model state-based cepstral noise compensation. 519-522 - Guy J. Brown, Martin P. Cooke:

A computational model of auditory scene analysis. 523-526 - Srinivas Nandkumar, John H. L. Hansen, Robert J. Stets:

A new dual-channel speech enhancement technique with application to CELP coding in noise. 527-530 - Asunción Moreno, José A. R. Fonollosa:

CUMULANT - based voicing decision in noise corrupted speech. 531-534
Speech Recognition 1, 2
- Yolande Anglade, Dominique Fohr, Jean-Claude Junqua:

Selectively trained neural networks for the discrimination of normal and lombard speech. 595-598 - Aaron E. Rosenberg, Joel DeLong, Chin-Hui Lee, Biing-Hwang Juang, Frank K. Soong:

The use of cohort normalized scores for speaker verification. 599-602 - Tomoko Matsui, Sadaoki Furui:

Speaker recognition using concatenated phoneme models. 603-606 - Younès Bennani:

Speaker identification through a modular connectionist architecture: evaluation on the timit database. 607-610 - Claude Montacié, Jean-Luc Le Floch:

AR-vector models for free-text speaker recognition. 611-614 - Florian Schiel:

Rapid non-supervised speaker adaptation of semicontinuous hidden Markov models. 1463-1466 - D. Ederveen, Louis Boves:

Rule-based recognition of phoneme classes. 1467-1470 - Jie Yi, Kei Miki:

A new method of speaker-independent speech recognition using multiphone HMM. 1471-1474 - Myoung-Wan Koo, Chong Kwan Un:

A speaker adaptation based on corrective training and learning vector quantization. 1475-1478 - Katsuhiko Shirai, Shigeki Okawa, Tetsunori Kobayashi:

Phoneme recognition in continuous speech based on mutual information considering phonemic duration and connectivity. 1479-1482 - Shinji Koga, Ryosuke Isotani, Satoshi Tsukada, Kazunaga Yoshida, Kaichiro Hatazaki, Takao Watanabe:

A real-time speaker-independent continuous speech recognition system based on demi-syllable units. 1483-1486 - Saeed Vaseghi, Ben P. Milner:

Speech recognition in noisy environments. 1487-1490 - Fergus R. McInnes:

An enhanced interpolation technique for context-specific probability estimation in speech and language modelling. 1491-1494 - Lorenzo Fissore, Pietro Laface, Giorgio Micca, G. Sperto:

Channel adaptation for a continuous speech recognizer. 1495-1498 - S. Cifuentes, José Colás, Mohammad Hasan Savoji, José Manuel Pardo:

A new algorithm for connected digit recognition. 1499-1502 - Günther Ruske, Bernd Plannerer, Tanja Schultz:

Stochastic modeling of syllable-based units for continuous speech recognition. 1503-1506 - David M. Goblirsch, Toffee A. Albina:

HARK: an experimental speech recognition system. 1507-1510 - Akito Nagai, Jun-ichi Takami, Shigeki Sagayama:

The SSS-LR continuous speech recognition system: integrating SSS-derived allophone models and a phoneme-context-dependent LR parser. 1511-1514 - Shinsuke Sakai, Michael S. Phillips:

J-SUMMIT: a Japanese segment-based speech recognition system. 1515-1518 - Shinobu Mizuta, Kunio Nakajima:

Optimal discriminative training for HMMs to recognize noisy speech. 1519-1522 - Shingo Kuroiwa, Kazuya Takeda, Fumihiro Yato, Seiichi Yamamoto, Kunihiko Owa, Makoto Shozakai, Ryuuji Matsumoto:

Architecture and algorithms of a real-time word recognizer for telephone input. 1523-1526 - Hiroyasu Kuwano, Kazuya Nomura, Atsushi Ookumo, Shoji Hiraoka, Taisuke Watanabe, Katsuyuki Niyada:

Speaker independent speech recognition method using word spotting technique and its application to VCR programming. 1527-1530 - S. Lennon, Eliathamby Ambikairajah:

Transputer implementation of front-end processors for speech recognition systems. 1531-1534 - Yasuhiro Minami, Tatsuo Matsuoka, Kiyohiro Shikano:

Phoneme HMM evaluation algorithm without phoneme labeling. 1535-1538 - Andreas Noll, Henning Bergmann, Hans-Hermann Hamer, Annedore Paeseler, Horst Tomaschewski:

Architecture of a configurable application interface for speech recognition systems. 1539-1542 - Mark A. Fanty, John Pochmara, Ronald A. Cole:

An interactive environment for speech recognition research. 1543-1546 - Yoshiharu Abe, Kunio Nakajima:

An approach to unlimited vocabulary continuous speech recognition based on context-dependent phoneme modeling. 1547-1550 - Chuck Wooters, Nelson Morgan:

Acoustic subword models in the berkeley restaurant project. 1551-1554 - Claus Nedergaard Jacobsen:

SIRtrain, an open standard environment for CHMM recognizer development. 1555-1558 - Yutaka Kobayashi, Yasuhisa Niimi:

Segmented trellis algorithms for the continuous speech recognition. 1559-1562 - Bo Xu, Zhiwei Lin, Taiyi Huang, Dongxin Xu, Yuqing Gao:

A. 46 500 word Chinese speech recognition system. 1563-1566 - Dao Wen Chen:

Study of the time extension flat net for speech recognition. 1567-1568
Hidden Markov Models
- Frank Fallside:

A hidden Markov model structure for the acquisition of speech by machine, ASM. 615-618 - Yasuyuki Masai, Shin'ichi Tanaka, Tsuneo Nitta:

Speaker-independent keyword recognition based on SMQ/HMM. 619-622 - Régis Cardin, Diane Goupil, Roxane Lacouture, Evelyne Millien, Charles Snow, Yves Normandin:

CRIM's spontaneous speech recognition system for the ATIS task. 623-626 - Fabio Brugnara, Renato de Mori, Diego Giuliani, Maurizio Omologo:

Improved connected digit recognition using spectral variation functions. 627-630 - Andrew Tridgell, J. Bruce Millar, Kim-Anh Do:

Alternative preprocessing techniques for discrete hidden Markov model phoneme recognition. 631-634
Dialogue
- Gerhard Th. Niedermair:

Linguistic modelling in the context of oral dialogue. 635-638 - François Andry:

Static and dynamic predictions : a method to improve speech understanding in cooperative dialogues. 639-642 - Paul Heisterkamp, Scott McGlashan, Nick J. Youd:

Dialogue Semantics for an Oral Dialogue System. 643-646 - Masaaki Nagata:

Using pragmatics to rule out recognition errors in cooperative task-oriented dialogues. 647-650 - Yoichi Takebayashi, Hiroyuki Tsuboi, Yoichi Sadamoto, Hideki Hashimoto, Hideaki Shinchi:

A real-time speech dialogue system using spontaneous speech understanding. 651-654
Prosody and Phonology 1, 2
- Li-chiung Yang:

A semantic and pragmatic analysis of tone and intonation in Mandarin Chinese. 655-658 - Yoshimasa Tsukuma:

On prosodic features in speech - comparative studies between Japanese and standard Chinese. 659-662 - W. Nick Campbell:

Prosodic encoding of English speech. 663-666 - Gunnar Fant, Anita Kruckenberg, Lennart Nord:

Prediction of syllable duration, speech rate and tempo. 667-670 - Rolf Carlson, Björn Granström, Lennart Nord:

Experiments with emotive speech - acted utterances and synthesized replicas. 671-674 - J. Caspars, Vincent J. van Heuven:

Phonetic properties of dutch accent lending pitch movements under time pressure. 731-734 - Jacques M. B. Terken, Karin van den Hombergh:

Judgments of relative prominence for adjacent and non-adjacent accents. 735-738 - Frédéric Beaugendre, Christophe d'Alessandro, Anne Lacheret-Dujour, Jacques M. B. Terken:

A perceptual study of French intonation. 739-742 - Mark Y. Liberman, J. Michael Schultz, Soonhyun Hong, Vincent Okeke:

The phonetics of IGBO tone. 743-746 - Stefanie Shattuck-Hufnagel:

Stress shift as pitch accent placement: within-word early accent placement in american English. 747-750
Speaking Styles 1, 2
- Katherine Morton:

Adding emotion to synthetic speech dialogue systems. 675-678 - Cari Spring, Donna Erickson, Thomas Call:

Emotional modalities and intonation in spoken language. 679-682 - Tatiana Slama-Cazacu:

Are any "press-conferences", "interviews" or "dialogues" true dialogues? 683-686 - Arthur J. Bronstein:

The categorization of the dialects and speech styles of north american English. 687-690 - Eleonora Blaauw:

Phonetic differences between read and spontaneous speech. 751-754 - Maxine Eskénazi:

Changing speech styles: strategies in read speech and casual and careful spontaneous speech. 755-758 - Noriko Umeda, Karen Wallace, Josephine Horna:

Usage of words and sentence structures in spontaneous versus text material. 759-762 - Nancy A. Daly, Victor Zue:

Statistical and linguistic analyses of F0 in read and spontaneous speech. 763-766 - Linda Shockey, Edda Farnetani:

Spontaneous speech in English and Italian. 767-770
ASR in Noise
- Claude Lefèbvre, Dariusz A. Zwierzynski, David R. Starks, Gary Birch:

Further optimisation of a robust IMELDA speech recogniser for applications with severely degraded speech. 691-694 - Richard M. Stern, Fu-Hua Liu, Yoshiaki Ohshima, Thomas M. Sullivan, Alejandro Acero:

Multiple approaches to robust speech recognition. 695-698 - Tadashi Kitamura, Satoshi Ando, Etsuro Hayahara:

Speaker-independent spoken digit recognition in noisy environments using dynamic spectral features and neural networks. 699-702 - Douglas A. Cairns, John H. L. Hansen:

ICARUS: an mwave-based real-time speech recognition system in noise and lombard effect. 703-706 - Chafic Mokbel, Laurent Barbier, Y. Kerlou, Gérard Chollet:

Word recognition in the car: adapting recognizers to new environments. 707-710
Dialogues and Applications
- Peter Meyer, Hans-Wilhelm Rühl, L. Vogten:

German announcements using synthetic speech the Gauss system. 711-714 - Mervyn A. Jack, John C. Foster, Fred Stentiford:

Intelligent dialogues in automated telephone services. 715-718 - Palle Bach Nielsen, Anders Baekgaard:

Experience with a dialogue description formalism for realistic applications. 719-722 - Solomon Lerner, Baruch Mazor:

Compensating for additive-noise in automatic speech recognition. 723-726 - Shoichi Matsunaga, Toshiaki Tsuboi, Tomokazu Yamada, Kiyohiro Shikano:

Continuous speech recognition for medical diagnoses using a character trigram model. 727-730
Sampling of Speech Production 1, 2
- Chengxiang Lu, Takayoshi Nakai, Hisayoshi Suzuki:

A three-dimensional FEM simulation of the effects of the vocal tract shape on the transfer function. 771-774 - Kiyoshi Oshimat, Vincent L. Gracco:

Mandibular contributions to speech production. 775-778 - Masafumi Matsumura:

Measurement of three-dimensional shapes of vocal tract and nasal cavity using magnetic resonance imaging technique. 779-782 - Shigeru Kiritani, Hajime Hirose, Kikuo Maekawa, Tsutomu Sato:

Electromyographie studies on the production of pitch contour in accentless dialects in Japanese. 783-786 - Yorinobu Sonoda, Kohichi Ogata:

Improvements of magnetometer sensing system for monitoring tongue point movements during speech. 843-846 - Paavo Alku:

Inverse filtering of the glottal waveform using the Itakura-saito distortion measure. 847-850 - Kunitoshi Motoki, Nobuhiro Miki:

Measurement of intraoral sound pressure distributions of Japanese vowels. 851-854
Labelling of Speech 1, 2
- Alain Marchal, William J. Hardcastle, K. Nicolaidis, Noël Nguyen, Fiona Gibbon:

Non-linear annotation of multi-channel speech data. 787-790 - Shingo Fujiwara, Yasuhiro Komori, Masahide Sugiyama:

A phoneme labelling workbench using HMM and spectrogram reading knowledge. 791-794 - Michael S. Phillips, Victor Zue:

Automatic discovery of acoustic measurements for phonetic classification. 795-798 - Katunobu Itou, Satoru Hayamizu, Hozumi Tanaka:

Detection of unknown words and automatic estimation of their transcriptions in continuous speech recognition. 799-802 - Fabio Brugnara, Daniele Falavigna, Maurizio Omologo:

A HMM-based system for automatic segmentation and labeling of speech. 803-806 - Robert W. P. Luk, Robert I. Damper:

A modification of the viterbi algorithm for stochastic phonographic transduction. 855-858 - Paul C. Bagshaw, Briony J. Williams:

Criteria for labelling prosodic aspects of English speech. 859-862 - Yifan Gong, Jean Paul Haton:

DTW-based phonetic labeling using explicit phoneme duration constraints. 863-866 - Kim E. A. Silverman, Mary E. Beckman, John F. Pitrelli, Mari Ostendorf, Colin W. Wightman, Patti Price, Janet B. Pierrehumbert, Julia Hirschberg:

TOBI: a standard for labeling English prosody. 867-870 - Barbara Eisen, Hans-Günther Tillmann, Christoph Draxler:

Consistency of judgements in manual labelling of phonetic segments: the distinction between clear and unclear cases. 871-874
Speech Production: Models and Theories 1, 2
- Gunnar Fant:

Vocal tract area functions of Swedish vowels and a new three-parameter model. 807-810 - Jean-Claude Junqua:

Acoustic and production pilot studies of speech vowels produced in noise. 811-814 - Yves Laprie, Marie-Odile Berger:

Active models for regularizing formant trajectories. 815-818 - René Carré, Samir Chennoukh, Mohamad Mrayati:

Vowel-consonant-vowel transitions: analysis, modeling, and synthesis. 819-822 - Maureen Stone, Subhash Lele:

Representing the tongue surface with curve fits. 875-878 - Katherine S. Harris, Eric Vatikiotis-Bateson, Peter J. Alfonso:

Muscle forces in vowel vocal tract formation. 879-882 - Makoto Hirayama, Eric Vatikiotis-Bateson, Mitsuo Kawato, Kiyoshi Honda:

Neural network modeling of speech motor control. 883-886 - Eric Vatikiotis-Bateson, Makoto Hirayama, Kiyoshi Honda, Mitsuo Kawato:

The articulatory dynamics of running speech: gestures from phonemes? 887-890
Phonetics
- Patricia A. Keating, B. Blankenship, Dani Byrd, Edward Flemming, Yuichi Todaka:

Phonetic analyses of the TIMIT corpus of american English. 823-826 - Dani Byrd:

Sex, dialects, and reduction. 827-830 - Manjari Ohala, John J. Ohala:

Phonetic universals and hindi segment duration. 831-834 - Donna Erickson, Osamu Fujimura:

Acoustic and articulatory correlates of contrastive emphasis in repeated corrections. 835-838 - Gary N. Tajchman, Marcia A. Bush:

Effects of context and redundancy in the perception of naturally produced English vowels. 839-842
Speech Data Bases
- Ronald A. Cole, Krist Roginski, Mark A. Fanty:

A telephone speech database of spelled and spoken names. 891-894 - Yeshwant K. Muthusamy, Ronald A. Cole, Beatrice T. Oshika:

The OGI multi-language telephone speech corpus. 895-898 - Douglas B. Paul, Janet M. Baker:

The design for the wall street journal-based CSR corpus. 899-902 - Lynette Hirschman:

Multi-site data collection for a spoken language corpus - MAD COW. 903-906 - Michael S. Phillips, James R. Glass, Joseph Polifroni, Victor Zue:

Collection and analyses of WSJ-CSR corpus at MIT. 907-910
Stochastic ASR
- Victor Abrash, Horacio Franco, Michael Cohen, Nelson Morgan, Yochai Konig:

Connectionist gender adaptation in a hybrid neural network / hidden Markov model speech recognition system. 911-914 - Michael Cohen, Horacio Franco, Nelson Morgan, David E. Rumelhart, Victor Abrash:

Hybrid neural network/hidden Markov model continuous-speech recognition. 915-918 - Gernot A. Fink, Franz Kummert, Gerhard Sagerer, Ernst Günter Schukat-Talamazzini, Heinrich Niemann:

Semantic hidden Markov networks. 919-922
Analysis of Disfluencies in Speech 1, 2
- Use Lehiste, Donna Erickson:

Hesitation sounds: is there coarticulation across pause? 923-926 - Corine A. Bickley, Sheri Hunnicutt:

Acoustic analysis of laughter. 927-930 - Douglas D. O'Shaughnessy:

Analysis of false starts in spontaneous speech. 931-934 - Robin J. Lickley, Ellen Gurman Bard:

Processing disfluent speech: recognising disfluency before lexical access. 935-938 - Philippe Morin, Jean-Claude Junqua, Jean-Marie Pierrel:

A flexible multimodal dialogue architecture independent of the application. 939-942 - Catia Cucchiarini, Renée van Bezooijen:

Familiarity with the language transcribed and context as determinants of intratranscriber agreement. 987-990 - Elizabeth Shriberg, Robin J. Lickley:

Intonation of clause-internal filled pauses. 991-994 - Elizabeth Wade, Elizabeth Shriberg, Patti Price:

User behaviors affecting speech recognition. 995-998
Linguistic Phonetics: Reduction
- Sharon Y. Manuel, Stefanie Shattuck-Hufnagel, Marie K. Huffman, Kenneth N. Stevens, Rolf Carlson, Sheri Hunnicutt:

Studies of vowel and consonant reduction. 943-946 - Noriko Umeda:

Formant frequencies of vowels in English function words. 947-950
Hearing/Speech Impaired 1, 2
- Christian Benoît, Tayeb Mohamadi:

The lip benefit: auditory and visual intelligibility of French speech in noise. 951-954 - Anne-Marie Öster:

Phonological assessment of deaf children's productive knowledge as a basis for speech-training. 955-958 - Hideaki Seki, Akiko Hayashi, Satoshi Imaizumi, Takehiko Harada, Hiroshi Hosoi:

Factors affecting voicing distinction of stops for the hearing impaired. 959-962 - Arthur Boothroyd, Robin S. Waldstein, Eddy Yeung:

Investigations into the auditory F0 speechreading enhancement effect using a sinusoidal replica of the F0 contour. 963-966 - Francesco Cutugno:

Some considerations on pitch and timing control in deaf children. 967-970 - Shari R. Baum:

Rate of speech effects in aphasia: an acoustic analysis of voice onset time. 1019-1022 - Parth M. Bhatt:

Fundamental frequency attributes following unilateral left or right temporal lobe lesion. 1023-1026 - Hiroshi Hosoi, Satoshi Imaizumi, Akiko Hayashi, Takehiko Harada, Hideaki Seki:

Cue extraction and integration in speech perception for the hearing impaired. 1027-1030 - Anna K. Nabelek:

The relationship between spectral details in naturally produced vowels and identification errors in noise and reverberation. 1031-1034 - Donald G. Jamieson, Leonard Cornelisse:

Speech processing effects on intelligibility for hearing-impaired listeners. 1035-1038
Unit-Based ASR
- Fuyuan Mo, Changli Li, Tao Chen:

Chinese recognition and synthesis system based on Chinese syllables. 971-974 - Hirofumi Yogo, Naoki Inagaki:

Accelerated stochastic approximation method based parameter estimation of monosyllables and their recognition using a neural network. 975-978 - Toomas Altosaar, Matti Karjalainen:

Diphone-based speech recognition using time-event neural networks. 979-982 - Giovanni Flammia, Paul Dalsgaard, Ove Andersen, Børge Lindberg:

Segment based variable frame rate speech analysis and recognition using a spectral variation function. 983-986
Language and Dialect Characterization
- Christian Benoît:

Intelligibility of the French spoken in France compared across listeners from France and from the Ivory Coast. 999-1002 - Julie Brousseau, Sally Anne Fox:

Dialect-dependent speech recognizers for canadian and european French. 1003-1006 - Yeshwant K. Muthusamy, Ronald A. Cole:

Automatic segmentation and identification of ten languages using telephone speech. 1007-1010 - Seiichi Nakagawa, Yoshio Ueda, Takashi Seino:

Speaker-independent, text-independent language identification by HMM. 1011-1014 - Shuichi Itahashi, Tsutomu Yamashita:

A discrimination method between Japanese dialects. 1015-1018
Speech Production, Parception, and Analysis
- B. Boyanov, Gérard Chollet:

Pathological voice analysis using cepstra, bispectra and group delay functions. 1039-1042 - Qianje Fu, Peyu Xia, Ren-Hua Wang:

Lateralization of speech sounds by binaural distributing processing. 1043-1046 - H. H. Rump:

Timing of pitch movements and perceived vowel duration. 1047-1050 - J. P. Liu, Geneviève Baudoin, Gérard Chollet:

Studies of glottal excitation and vocal tract parameters using inverse filtering and a parameterized input model. 1051-1054 - Dennis Norris, Brit van Ooyen, Anne Cutler:

Speeded detection of vowels and steady-state consonants. 1055-1058 - Elzbieta B. Slawinski:

Temporal factors in the perception of consonants for different age and hearing impairment groups. 1059-1062 - Abeer Alwan:

The role of F3 and F4 in identifying place of articulation for stop consonants. 1063-1066 - Thomas R. Sawallis:

A new measure for perceptual weight of acoustic cues: an experiment on voicing in French intervocalic [t, d]. 1067-1070 - Alan Wrench, Mervyn A. Jack, John Laver, Mary S. Jackson, David S. Soutar, A. Gerry Robertson, Janet MacKenzie Beck:

Objective speech quality assessment in patients with intra-oral cancers: voiceless fricatives. 1071-1074 - Bruce Connell:

Tongue contact, active articulators, and coarticulation. 1075-1078 - Makio Kashino, Astrid van Wieringen, Louis C. W. Pols:

Cross-languages differences in the identification of intervocalic stop consonants by Japanese and dutch listeners. 1079-1082 - Minoru Tsuzaki:

Effects of typicality and interstimulus interval on the discrimination of speech stimuli: within-subject comparison. 1083-1086 - Ronald A. Cole, Yeshwant K. Muthusamy:

Perceptual studies on vowels excised from continuous speech. 1087-1090 - Raymond S. Weitzman:

The relative perceptual salience of spectral and durational differences. 1095-1098 - Florien J. Koopmans-van Beinum:

Can 'level words' from one speaking style become teaks' when spliced into another speaking style? 1099-1102 - Beverley Gable, Helen Nemeth, Martin Haran:

Speech errors and task demand. 1103-1106 - John H. Esling, B. Craig Dickson, Roy C. Snell:

Analysis of phonation type using laryngographic techniques. 1107-1110 - Sumi Shigeno:

Effect of prototypes of vowels on speech perception in Japanese and English. 1111-1114 - Tomo-o Morohashi, Tetsuya Shimamura, Hiroyuki Yashima, Jouji Suzuki:

Characteristics of voice picked up from outer skin of larynx. 1115-1118 - Igor V. Nabelek:

Coding of voicing in whispered plosives. 1119-1122 - Margaret F. Cheesman, Shelly Lawrence, Allison Appleyard:

Performance on a nonsense syllable test using the articulation index. 1123-1126 - Donald G. Jamieson, Ketan Ramji, Issam Kheirallah, Terrance M. Nearey:

CSRE: a speech research environment. 1127-1130
Linguistic Phonetics
- Kazue Hata, Yoko Hasegawa:

A study of F0 reset in naturally-read utterances in Japanese. 1239-1242 - H. Samuel Wang, Fu-Dong Chiu:

On the nature of tone sandhi rules in taiwanese. 1243-1246 - Geoffrey S. Nathan:

How shallow is phonology: declarative phonologies meet fast speech. 1247-1250 - Junko Hosaka, Toshiyuki Takezawa, Noriyoshi Uratani:

Analyzing postposition drops in spoken Japanese. 1251-1254 - Jialu Zhang, Xinghui Hu:

Fundamental frequency patterns of Chinese in different speech modes. 1255-1258 - Knut Kvale, Arne Kjell Foldvik:

The multifarious r-sound. 1259-1262 - Zita McRobbie-Utasi:

The role of preaspiration duration in the voicing contrast in skolt sami. 1263-1266 - Eiji Yamada:

Parameter setting for abstract stress in tokyo Japanese. 1267-1270 - Georg Ottesen:

A method for studying prosody in texts read aloud. 1271-1274 - Vincent J. van Heuven:

Linguistic versus phonetic explanation of consonant lengthening after short vowels: a contrastive study of dutch and English. 1275-1278 - Kjell Elenius, Mats Blomberg:

Comparing phoneme and feature based speech recognition using artificial neural networks. 1279-1282 - Eva Strangert:

Prosodic cues to the perception of syntactic boundaries. 1283-1286 - Paul Taylor, Stephen Isard:

A new model of intonation for use with speech synthesis and recognition. 1287-1290 - Rudolf Weiss:

Computerized error detection/correction in teaching German sounds: some problems and solutions. 1291-1294 - Ahmed M. Elgendy:

Velum and epiglottis behavior during the production of Arabic pharyngeals and laryngeals: a fiberscopic study. 1295-1298 - Kim E. A. Silverman, Eleonora Blaauw, Judith Spitz, John F. Pitrelli:

A prosodic comparison of spontaneous speech and read speech. 1299-1302 - John J. Ohala, Maria Grazia Busà, Karen Harrison:

Phonological and psychological evidence that listeners normalize the speech signal. 1303-1306 - Elizabeth A. Hinkelman:

Intonation and the request/question distinction. 1307-1310 - Robert F. Port, Fred Cummins:

The English voicing contrast as velocity perturbation. 1311-1314 - Michael S. Ziolkowski, Mayumi Usami, Karen L. Landahl, Brenda K. Tunnock:

How many phonologies are there in one speaker? some experimental evidence. 1315-1318 - Hirokazu Sato:

Decomposition into syllable complexes and the accenting of Japanese loanwords. 1319-1322 - Jianfen Cao:

Temporal structure in bisyllabic word frame: an evidence for relational invariance and variability from standard Chinese. 1323-1326 - Shih-ping Wang:

The integration of phonetics and phonology: a case study of taiwanese "gemination" and syllable structure. 1327-1330
Perception and Production
- James H. Bradford:

Towards a robust speech interface for teleoperation systems. 1331-1334 - Piero Cosi, Paolo Frasconi, Marco Gori, N. Griggio:

Phonetic recognition experiments with recurrent neural networks. 1335-1338 - Mikael Goldstein, Björn Lindström, Ove Till:

Some aspects on context and response range effects when assessing naturalness of Swedish sentences generated by 4 synthesiser systems. 1339-1342 - Marcello Pelillo, Franca Moro, Mario Refice:

Probabilistic prediction of parts-of-speech from word spelling using decision trees. 1343-1346 - Dieter Barschdorff, Ulrich Gärtner:

Single word detection system with a neural classifier for recognizing speech at variable levels of background noise. 1347-1350 - Sharon L. Oviatt, Philip R. Cohen, Martin Fong, Michael Frank:

A rapid semi-automatic simulation technique for investigating interactive speech and handwriting. 1351-1354 - Sang-Hwa Chung, Dan I. Moldovan:

Speech understanding on a massively parallel computer. 1355-1358 - Chan-Do Lee:

Rationale for "performance phonology". 1359-1362 - Takuya Koizumi, Jyoji Urata, Shuji Taniguchi:

The effect of information feedback on the performance of a phoneme recognizer using kohonen map. 1363-1366 - Yasuharu Asano, Keikichi Hirose, Hiroya Fujisaki:

A method of dialogue management for the speech response system. 1367-1370 - Yumi Takizawa, Eiichi Tsuboka:

Syllable duration prediction for speech recognition. 1371-1374 - Jun He, Henri Leich:

Extracting fuzzy features from MLP for recognition of speech. 1379-1382 - Keiji Fukuzawa, Yoshinaga Kato, Masahide Sugiyama:

A fuzzy partition model (FPM) neural network architecture for speaker-independent continuous speech recognition. 1383-1386 - A. Ennaji, Jean Rouat:

Conception of speech filters based on a neural network. 1387-1390 - Hong-Kwang Jeff Kuo, Chin-Hui Lee, Aaron E. Rosenberg:

Speaker set identification through speaker group modeling. 1391-1394 - Stephen Springer, Sara Basson, Judith Spitz:

Identification of principal ergonomic requirements for interactive spoken language systems. 1395-1398 - Thomas E. Jacobs, Eric R. Buhrke:

Performance of the united kingdom intelligent network automatic speech recognition system. 1399-1402 - Guy Deville, Pierre Mousel:

Evaluation of parsing strategies in natural language spoken man-machine dialogue. 1403-1406 - Yasuhisa Niimi, Yutaka Kobayashi:

An information retrieval system with a speech interface. 1407-1410 - Julian P. Eatock, John S. D. Mason:

Phoneme performance in speaker recognition. 1411-1414 - Evelyne Tzoukermann, Roberto Pieraccini, Zakhar Gorelov:

Natural language processing in the chronus system. 1415-1418 - Dominique François, Dominique Fohr:

Contribution of neural networks for phoneme identification in the APHODEX expert system. 1419-1422 - Douglas B. Paul:

A CSR-NL interface architecture. 1423-1426 - Pierre Lefebvre, Frank Poirier, G. Duncan:

Speech interface for a man-machine dialog with the unix operating system. 1427-1430 - P. Bardaud, François Capman, Chafic Mokbel, Chakib Tadj, Gérard Chollet:

Transformation of databases for the evaluation of speech recognizers. 1431-1434 - Yoichi Yamashita, Riichiro Mizoguchi:

Dialog management for speech output from concept representation. 1435-1438 - Seiichiro Hangai, Shigetoshi Sugiyama, Kazuhiro Miyauchi:

Speaker verification using locations and sizes of multipulses on neural networks. 1439-1442 - Carlos Teixeira, Isabel Trancoso:

Word rejection using multiple sink models. 1443-1446 - Boerge Lindberg:

Verification of language specific performance factors from recogniser testing on EUROM.1 CVC material. 1447-1450 - Alain Cozannet:

Modeling task driven oral dialogue. 1451-1454 - Wei-ying Li, Kechu Yi, Zheng Hu:

Introducing neural predictor to hidden Markov model for speech recognition. 1455-1458 - Feng Liu, Jianxin Jiang, Jun Cheng, Kechu Yi:

A neural network based on subnets - SNN. 1459-1462
Analysis/Systems
- Ute Ziegenhain:

Syntactic anaphora resolution in a speech understanding system. 1569-1572 - Marion Mast, Ralf Kompe, Franz Kummert, Heinrich Niemann, Elmar Nöth:

The dialog module of the speech recognition and dialog system EVAR. 1573-1576 - Yan Ming Cheng, Douglas D. O'Shaughnessy, Paul Mermelstein:

Statistical recovery of wideband speech from narrowband speech. 1577-1580 - Henk van den Heuvel, Toni C. M. Rietveld:

Speaker related variability in cepstral representations of dutch speech segments. 1581-1584 - Per Rosenbeck, Bo Baungaard:

Experiences from a real-world telephone application: teledialogue. 1585-1588 - K. Y. Lee, P. Ha, J. Rheem, Souguil Ann, Iickho Song:

Robust estimation of time-varying LP parameters on speech. 1589-1592 - Javier Hernando, Climent Nadeu, Eduardo Lleida:

On the AR modelling of the one-sided autocorrelation sequence for noisy speech recognition. 1593-1596 - Hiroshi Shimodaira, Mitsuru Nakai:

Robust pitch detection by narrow band spectrum analysis. 1597-1600 - S. Eady, B. Craig Dickson, Roy C. Snell, J. Woolsey, P. Ollek, A. Wynrib, Jocelyn Clayards:

A microcomputer-based system for real-time analysis and display of laryngograph signals. 1601-1604 - Nanette M. Veilleux, Mari Ostendorf, Colin W. Wightman:

Parse scoring with prosodic information. 1605-1608 - Ying Cheng, Paul Fortier, Yves Normandin:

Topic identification using a neural network with a keyword-spotting preprocessor. 1609-1612 - Shane Switzer, Timothy R. Anderson, Matthew Kabrisky, Steven K. Rogers, Bruce W. Suter:

Frequency domain speech coding. 1613-1616 - Raymond Descout, Robert Bergeron, Bernard Meriald:

MEDIATEX-TASF: a closed captioning real-time service in French. 1617-1620 - S. A. Wilde, K. Mervyn Curtis:

The wavelet transform for speech analysis. 1621-1624 - Pablo Aibar, Andrés Marzal, Enrique Vidal, Francisco Casacuberta:

Problems and algorithms in optimal linguistic decoding: a unified formulation. 1625-1628 - Jean Rouat, Sylvain Lemieux, Alain Migneault:

A spectro-temporal analysis of speech based on nonlinear operators. 1629-1632 - Miguel A. Berrojo, Javier Corrales, Jesus Macias, Santiago Aguilera:

A PC graphic tool for speech research based on a DSP board. 1633-1636 - Satoru Hayamizu, Katunobu Itou, Masafumi Tamoto, Kazuyo Tanaka:

A spoken language dialogue system for automatic collection of spontaneous speech. 1637-1640 - Shingo Nishioka, Yoichi Yamashita, Riichiro Mizoguchi:

A powerful disambiguating mechanism for speech understanding systems based on ATMs. 1641-1644 - Najib Naja, Jean-Marc Boucher, Samir Saoudi:

A mixed Gaussian-stochastic code book for CELP coder in LSP speech coding. 1645-1648 - Hiroyuki Kamata, Yoshihisa Ishida:

A method to estimate the transfer function of ARMA model of speech wave using prony method and homomorphic analysis. 1649-1652 - Børge Lindberg, Bjarne Andersen, Anders Baekgaard, Tom Brøndsted, Paul Dalsgaard, Jan Kristiansen:

An integrated dialogue design and continuous speech recognition system environment. 1653-1656 - Alain Marchal, Christine Meunier, P. Gavarry:

The PSH/DISPE helium speech cdrom. 1657-1660

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.


Google
Google Scholar
Semantic Scholar
Internet Archive Scholar
CiteSeerX
ORCID














