Skip to main content

Advertisement

Intelligent Audio, Speech, and Music Processing Applications

  1. Robust automatic language identification (LID) is a task of identifying the language from a short utterance spoken by an unknown speaker. One of the mainstream approaches named parallel phone recognition langu...

    Authors: Hongbin Suo, Ming Li, Ping Lu and Yonghong Yan

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:674859

    Content type: Research Article

    Published on:

  2. This paper investigates the problem of speaker recognition in noisy conditions. A new approach called nonnegative tensor principal component analysis (NTPCA) with sparse constraint is proposed for speech featu...

    Authors: Qiang Wu and Liqing Zhang

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:578612

    Content type: Research Article

    Published on:

  3. Improving the intelligibility of speech in different environments is one of the main objectives of hearing aid signal processing algorithms. Hearing aids typically employ beamforming techniques using multiple ...

    Authors: Sriram Srinivasan, Ashish Pandharipande and Kees Janse

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:824797

    Content type: Research Article

    Published on:

  4. Online personalization of hearing instruments refers to learning preferred tuning parameter values from user feedback through a control wheel (or remote control), during normal operation of the hearing aid. We...

    Authors: Alexander Ypma, Job Geurts, Serkan Özer, Erik van der Werf and Bert de Vries

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:183456

    Content type: Research Article

    Published on:

  5. A proven method for achieving effective automatic speech recognition (ASR) due to speaker differences is to perform acoustic feature speaker normalization. More effective speaker normalization methods are needed ...

    Authors: Umit H. Yapanel and John H.L. Hansen

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:148967

    Content type: Research Article

    Published on:

  6. Perception of moving sound sources obeys different brain processes from those mediating the localization of static sound events. In view of these specificities, a preprocessing model was designed, based on the...

    Authors: R Kronland-Martinet and T Voinier

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:849696

    Content type: Research Article

    Published on:

  7. The present paper proposes a new approach for detecting music boundaries, such as the boundary between music pieces or the boundary between a music piece and a speech section for automatic segmentation of musi...

    Authors: Yoshiaki Itoh, Akira Iwabuchi, Kazunori Kojima, Masaaki Ishigame, Kazuyo Tanaka and Shi-Wook Lee

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:480786

    Content type: Research Article

    Published on:

  8. This paper proposes a new algorithm for a directional aid with hearing defenders. Users of existing hearing defenders experience distorted information, or in worst cases, directional information may not be per...

    Authors: Benny Sällberg, Farook Sattar and Ingvar Claesson

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:274684

    Content type: Research Article

    Published on:

  9. We propose a new low complexity, low delay, and fast converging frequency-domain adaptive algorithm for network echo cancellation in VoIP exploiting MMax and sparse partial (SP) tap-selection criteria in the f...

    Authors: Xiang(Shawn) Lin, Andy W.H. Khong, Milŏs Doroslovăcki and Patrick A. Naylor

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:156960

    Content type: Research Article

    Published on:

  10. Binaural cue coding (BCC) is an efficient technique for spatial audio rendering by using the side information such as interchannel level difference (ICLD), interchannel time difference (ICTD), and interchannel...

    Authors: Bo Qiu, Yong Xu, Yadong Lu and Jun Yang

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:618104

    Content type: Research Article

    Published on:

  11. The behavior of time delay estimation (TDE) is well understood and therefore attractive to apply in acoustic source localization (ASL). A time delay between microphones maps into a hyperbola. Furthermore, the ...

    Authors: Pasi Pertilä, Teemu Korhonen and Ari Visa

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:278185

    Content type: Research Article

    Published on:

  12. Rhythmic information plays an important role in Music Information Retrieval. Example applications include automatically annotating large databases by genre, meter, ballroom dance style or tempo, fully automate...

    Authors: Björn Schuller, Florian Eyben and Gerhard Rigoll

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:846135

    Content type: Research Article

    Published on:

  13. The phasor representation is introduced to identify the characteristic of the active noise control (ANC) systems. The conventional representation, transfer function, cannot explain the fact that the performanc...

    Authors: Fu-Kun Chen, Ding-Horng Chen and Yue-Dar Jou

    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2008 2008:126859

    Content type: Research Article

    Published on: