Skip to main content

Articles

Page 8 of 11

  1. Many features have been proposed for speech-based emotion recognition, and a majority of them are frame based or statistics estimated from frame-based features. Temporal information is typically modelled on a ...

    Authors: Vidhyasaharan Sethu, Eliathamby Ambikairajah and Julien Epps
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:19
  2. Nonnegative matrix factorization (NMF) is developed for parts-based representation of nonnegative signals with the sparseness constraint. The signals are adequately represented by a set of basis vectors and th...

    Authors: Jen-Tzung Chien and Hsin-Lung Hsieh
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:18
  3. In this study, we focus on the classification of neutral and stressed speech based on a physical model. In order to represent the characteristics of the vocal folds and vocal tract during the process of speech...

    Authors: Xiao Yao, Takatoshi Jitsuhiro, Chiyomi Miyajima, Norihide Kitaoka and Kazuya Takeda
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:17
  4. This paper presents a bimodal acoustic-visual synthesis technique that concurrently generates the acoustic speech signal and a 3D animation of the speaker’s outer face. This is done by concatenating bimodal di...

    Authors: Slim Ouni, Vincent Colotte, Utpala Musti, Asterios Toutios, Brigitte Wrobel-Dautcourt, Marie-Odile Berger and Caroline Lavecchia
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:16
  5. Most existing automatic chord recognition systems use a chromagram in front-end processing and some sort of classifier (e.g., hidden Markov model, Gaussian mixture model (GMM), support vector machine, or other...

    Authors: Maksim Khadkevich and Maurizio Omologo
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:15
  6. Cochannel speech separation aims to separate two speech signals from a single mixture. In a supervised scenario, the identities of two speakers are given, and current methods use pre-trained speaker models for...

    Authors: Ke Hu and DeLiang Wang
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:14
  7. A challenging open question in music classification is which music representation (i.e., audio features) and which machine learning algorithm is appropriate for a specific music classification task. To address...

    Authors: Yannis Panagakis and Constantine Kotropoulos
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:13
  8. Multiple-model based speech recognition (MMSR) has been shown to be quite successful in noisy speech recognition. Since it employs multiple hidden Markov model (HMM) sets that correspond to various noise types...

    Authors: Yongjoo Chung and John HL Hansen
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:12
  9. A novel speech bandwidth extension method based on audio watermark is presented in this paper. The time-domain and frequency-domain envelope parameters are extracted from the high-frequency components of speec...

    Authors: Zhe Chen, Chengyong Zhao, Guosheng Geng and Fuliang Yin
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:10
  10. This paper presents a novel lossless compression technique of the context-based adaptive arithmetic coding which can be used to further compress the quantized parameters in audio codec. The key feature of the ...

    Authors: Jing Wang, Xuan Ji, Shenghui Zhao, Xiang Xie and Jingming Kuang
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:9
  11. This article analyzes and compares influence of different types of spectral and prosodic features for Czech and Slovak emotional speech classification based on Gaussian mixture models (GMM). Influence of initi...

    Authors: Jiří Přibil and Anna Přibilová
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:8
  12. In this article, we describe a speaker adaptation method based on the probabilistic 2-mode analysis of training models. Probabilistic 2-mode analysis is a probabilistic extension of multilinear analysis. We ap...

    Authors: Yongwon Jeong
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:7
  13. Availability of large amounts of raw unlabeled data has sparked the recent surge in semi-supervised learning research. In most works, however, it is assumed that labeled and unlabeled data come from the same d...

    Authors: Konstantin Markov and Tomoko Matsui
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:6
  14. A comprehensive system for facial animation of generic 3D head models driven by speech is presented in this article. In the training stage, audio-visual information is extracted from audio-visual training data...

    Authors: Lucas D Terissi, Mauricio Cerda, Juan C Gómez, Nancy Hitschfeld-Kahler and Bernard Girau
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:5
  15. Blind source separation (BSS) and sound activity detection (SAD) from a sound source mixture with minimum prior information are two major requirements for computational auditory scene analysis that recognizes ...

    Authors: Kohei Nagira, Takuma Otsuka and Hiroshi G Okuno
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:4
  16. We propose an efficient solution to the problem of sparse linear prediction analysis of the speech signal. Our method is based on minimization of a weighted l2-norm of the prediction error. The weighting function...

    Authors: Vahid Khanagha and Khalid Daoudi
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:3
  17. A lot of effort has been made in Computational Auditory Scene Analysis (CASA) to segregate target speech from monaural mixtures. Based on the principle of CASA, this article proposes an improved algorithm for ...

    Authors: Wang Yu, Lin Jiajun, Chen Ning and Yuan Wenhao
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:2
  18. The work presented in this article studies how the context information can be used in the automatic sound event detection process, and how the detection system can benefit from such information. Humans are usi...

    Authors: Toni Heittola, Annamaria Mesaros, Antti Eronen and Tuomas Virtanen
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2013 2013:1
  19. This article describes a modified technique for enhancing noisy speech to improve automatic speech recognition (ASR) performance. The proposed approach improves the widely used spectral subtraction which inher...

    Authors: Hari Krishna Maganti and Marco Matassoni
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:29
  20. Conventional parametric stereo (PS) audio coding employs inter-channel phase difference and overall phase difference as phase parameters. In this article, it is shown that those parameters cannot correctly rep...

    Authors: Dong-il Hyun, Young-cheol Park and Dae Hee Youn
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:27
  21. In this article, the authors propose an optimally designed fixed beamformer (BF) for stereophonic acoustic echo cancelation (SAEC) in real hands-free communication applications. Several contributions related t...

    Authors: Matteo Pirro, Stefano Squartini, Laura Romoli and Francesco Piazza
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:26
  22. The rapid spread in digital data usage in many real life applications have urged new and effective ways to ensure their security. Efficient secrecy can be achieved, at least in part, by implementing steganogra...

    Authors: Fatiha Djebbar, Beghdad Ayad, Karim Abed Meraim and Habib Hamam
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:25
  23. Mood is an important aspect of music and knowledge of mood can be used as a basic feature in music recommender and retrieval systems. A listening experiment was carried out establishing ratings for various moo...

    Authors: Bert den Brinker, Ralph van Dinther and Janto Skowronek
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:24
  24. A vast amount of audio features have been proposed in the literature to characterize the content of audio signals. In order to overcome specific problems related to the existing features (such as lack of discrimi...

    Authors: Toni Mäkinen, Serkan Kiranyaz, Jenni Raitoharju and Moncef Gabbouj
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:23
  25. Humans exhibit a remarkable ability to reliably classify sound sources in the environment even in presence of high levels of noise. In contrast, most engineering systems suffer a drastic drop in performance wh...

    Authors: Sridhar Krishna Nemala, Dmitry N Zotkin, Ramani Duraiswami and Mounya Elhilali
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:22
  26. A new method to secure speech communication using the discrete wavelet transforms (DWT) and the fast Fourier transform is presented in this article. In the first phase of the hiding technique, we separate the ...

    Authors: Siwar Rekik, Driss Guerchi, Sid-Ahmed Selouani and Habib Hamam
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:20
  27. In this article, we present the evaluation results for the task of speaker diarization of broadcast news, which was part of the Albayzin 2010 evaluation campaign of language and speech technologies. The evalua...

    Authors: Martin Zelenák, Henrik Schulz and Javier Hernando
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:19
  28. Dance movements are a complex class of human behavior which convey forms of non-verbal and subjective communication that are performed as cultural vocabularies in all human cultures. The singularity of dance f...

    Authors: João Lobato Oliveira, Luiz Naveda, Fabien Gouyon, Luis Paulo Reis, Paulo Sousa and Marc Leman
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:18
  29. In this article, we propose a new set of acoustic features for automatic emotion recognition from audio. The features are based on the perceptual quality metrics that are given in perceptual evaluation of audi...

    Authors: Mehmet Cenk Sezgin, Bilge Gunsel and Gunes Karabulut Kurt
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:16
  30. An increasing number of multilingual applications require language recognition (LRE) as a frontend, but desire low additional computational cost. This article demonstrates a novel architecture for embedding ph...

    Authors: Yuxiang Shan, Yan Deng, Jia Liu and Michael T Johnson
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:15
  31. The problem of blind source separation (BSS) of convolved acoustic signals is of great interest for many classes of applications. Due to the convolutive mixing process, the source separation is performed in th...

    Authors: Eugen Hoffmann, Dorothea Kolossa, Bert-Uwe Köhler and Reinhold Orglmeister
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:14
  32. A novel approach for robust dialogue act detection in a spoken dialogue system is proposed. Shallow representation named partial sentence trees are employed to represent automatic speech recognition outputs. P...

    Authors: Chia-Ping Chen, Chung-Hsien Wu and Wei-Bin Liang
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:13
  33. As fundamental research for human-robot interaction, this paper addresses the rhythmic reference of a human while turning a rope with another human. We hypothyzed that when interpreting rhythm cues to make a r...

    Authors: Kenta Yonekura, Chyon Hae Kim, Kazuhiro Nakadai, Hiroshi Tsujino and Shigeki Sugano
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:12
  34. This article proposes a new acoustic model using decision trees (DTs) as replacements for Gaussian mixture models (GMM) to compute the observation likelihoods for a given hidden Markov model state in a speech ...

    Authors: Masami Akamine and Jitendra Ajmera
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:10
  35. A study on force-feedback interaction with a model of a neural oscillator provides insight into enhanced human-robot interactions for controlling musical sound. We provide differential equations and discrete-t...

    Authors: Edgar Berdahl, Claude Cadoz and Nicolas Castagné
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:9
  36. Interaction with human musicians is a challenging task for robots as it involves online perception and precise synchronization. In this paper, we present a consistent and theoretically sound framework for comb...

    Authors: Umut Şimşekli, Orhan Sönmez, Barış Kurt Kurt and Ali Taylan Cemgil
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:8
  37. The aim of this paper is to improve beat-tracking for live guitar performances. Beat-tracking is a function to estimate musical measurements, for example musical tempo and phase. This method is critical to ach...

    Authors: Tatsuhiko Itohara, Takuma Otsuka, Takeshi Mizumoto, Angelica Lim, Tetsuya Ogata and Hiroshi G Okuno
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:6
  38. This study proposes a music-aided framework for affective interaction of service robots with humans. The framework consists of three systems, respectively, for perception, memory, and expression on the basis o...

    Authors: Jeong-Sik Park, Gil-Jin Jang and Yong-Ho Seo
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:5
  39. This article studies a vital issue in wireless communications, which is the transmission of audio signals over wireless networks. It presents a novel interleaver scheme for protection against error bursts and ...

    Authors: Mohsen Ahmed Mahmoud Mohamed El-Bendary, Atef E Abou-El-azm, Nawal A El-Fishawy, Farid Shawki, Fathi E Abd-ElSamie, Mostafa Ali Refai El-Tokhy and Hassan B Kazemian
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:4
  40. It has been long speculated that expression of emotions from different modalities have the same underlying 'code', whether it be a dance step, musical phrase, or tone of voice. This is the first attempt to imp...

    Authors: Angelica Lim, Tetsuya Ogata and Hiroshi G Okuno
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:3
  41. The availability of haptic interfaces in music content processing offers interesting possibilities of performer-instrument interaction for musical expression. These new musical instruments can precisely modulate ...

    Authors: Victor Zappi, Antonio Pistillo, Sylvain Calinon, Andrea Brogni and Darwin Caldwell
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2012 2012:2
  42. Most of voice activity detection (VAD) schemes are operated in the discrete Fourier transform (DFT) domain by classifying each sound frame into speech or noise based on the DFT coefficients. These coefficients...

    Authors: Shiwen Deng and Jiqing Han
    Citation: EURASIP Journal on Audio, Speech, and Music Processing 2011 2011:12

Who reads the journal?

Learn more about the impact the EURASIP Journal on Audio, Speech, and Music Processing has worldwide

Annual Journal Metrics

  • 2022 Citation Impact
    2.4 - 2-year Impact Factor
    2.0 - 5-year Impact Factor
    1.081 - SNIP (Source Normalized Impact per Paper)
    0.458 - SJR (SCImago Journal Rank)

    2023 Speed
    17 days submission to first editorial decision for all manuscripts (Median)
    154 days submission to accept (Median)

    2023 Usage 
    368,607 downloads
    70 Altmetric mentions 

Funding your APC

​​​​​​​Open access funding and policy support by SpringerOpen​​

​​​​We offer a free open access support service to make it easier for you to discover and apply for article-processing charge (APC) funding. Learn more here