Skip to main content

Joint Audio-Visual Speech Processing

  1. Research Article

    Editorial

    Chalapathy Neti, Gerasimos Potamianos, Juergen Luettin and Eric Vatikiotis-Bateson

    EURASIP Journal on Advances in Signal Processing 2002 2002:475826

    Published on: 28 November 2002

  2. Research Article

    Dynamic Bayesian Networks for Audio-Visual Speech Recognition

    The use of visual features in audio-visual speech recognition (AVSR) is justified by both the speech generation mechanism, which is essentially bimodal in audio and visual representation, and by the need for f...

    Ara V. Nefian, Luhong Liang, Xiaobo Pi, Xiaoxing Liu and Kevin Murphy

    EURASIP Journal on Advances in Signal Processing 2002 2002:783042

    Published on: 28 November 2002

  3. Research Article

    On the Relationship between Face Movements, Tongue Movements, and Speech Acoustics

    This study examines relationships between external face movements, tongue movements, and speech acoustics for consonant-vowel (CV) syllables and sentences spoken by two male and two female talkers with differe...

    Jintao Jiang, Abeer Alwan, Patricia A. Keating, Edward T. Auer and Lynne E. Bernstein

    EURASIP Journal on Advances in Signal Processing 2002 2002:506945

    Published on: 28 November 2002

  4. Research Article

    Audio-Visual Speech Recognition Using MPEG-4 Compliant Visual Features

    We describe an audio-visual automatic continuous speech recognition system, which significantly improves speech recognition performance over a wide range of acoustic noise levels, as well as under clean audio ...

    Petar S. Aleksic, Jay J. Williams, Zhilin Wu and Aggelos K. Katsaggelos

    EURASIP Journal on Advances in Signal Processing 2002 2002:150948

    Published on: 28 November 2002

  5. Research Article

    Joint Audio-Visual Tracking Using Particle Filters

    It is often advantageous to track objects in a scene using multimodal information when such information is available. We use audio as a complementary modality to video data, which, in comparison to vision, can...

    Dmitry N. Zotkin, Ramani Duraiswami and Larry S. Davis

    EURASIP Journal on Advances in Signal Processing 2002 2002:162620

    Published on: 28 November 2002

  6. Research Article

    Automatic Speechreading with Applications to Human-Computer Interfaces

    There has been growing interest in introducing speech as a new modality into the human-computer interface (HCI). Motivated by the multimodal nature of speech, the visual component is considered to yield inform...

    Xiaozheng Zhang, Charles C. Broun, Russell M. Mersereau and Mark A. Clements

    EURASIP Journal on Advances in Signal Processing 2002 2002:240192

    Published on: 28 November 2002

  7. Research Article

    Separation of Audio-Visual Speech Sources: A New Approach Exploiting the Audio-Visual Coherence of Speech Stimuli

    We present a new approach to the source separation problem in the case of multiple speech signals. The method is based on the use of automatic lipreading, the objective is to extract an acoustic speech signal ...

    David Sodoyer, Jean-Luc Schwartz, Laurent Girin, Jacob Klinkisch and Christian Jutten

    EURASIP Journal on Advances in Signal Processing 2002 2002:382823

    Published on: 28 November 2002

  8. Research Article

    Moving-Talker, Speaker-Independent Feature Study, and Baseline Results Using the CUAVE Multimodal Speech Corpus

    Strides in computer technology and the search for deeper, more powerful techniques in signal processing have brought multimodal research to the forefront in recent years. Audio-visual speech processing has bec...

    Eric K. Patterson, Sabri Gurbuz, Zekeriya Tufekci and John N. Gowdy

    EURASIP Journal on Advances in Signal Processing 2002 2002:208541

    Published on: 28 November 2002

  9. Research Article

    Noise Adaptive Stream Weighting in Audio-Visual Speech Recognition

    It has been shown that integration of acoustic and visual information especially in noisy conditions yields improved speech recognition results. This raises the question of how to weight the two modalities in ...

    Martin Heckmann, Frédéric Berthommier and Kristian Kroschel

    EURASIP Journal on Advances in Signal Processing 2002 2002:720764

    Published on: 28 November 2002

  10. Research Article

    Statistical Lip-Appearance Models Trained Automatically Using Audio Information

    We aim at modeling the appearance of the lower face region to assist visual feature extraction for audio-visual speech processing applications. In this paper, we present a neural network based statistical appe...

    Philippe Daubias and Paul Deléglise

    EURASIP Journal on Advances in Signal Processing 2002 2002:720534

    Published on: 28 November 2002