Publications

Publications

We are pioneers in the field of Audio AI research. AudEERING’s technology is used in many research projects. We provide information about the results of our research in numerous articles, essays, papers and other publications. Take a look at some of our scientific citations as well.

Age Classification: Comparison of Human vs Machine Performance in Prompted and Spontaneous Speech

2021
F. Burkhardt, Markus Brückl and Björn Schuller

Age Classification: Comparison of Human vs Machine Performance in Prompted and Spontaneous Speech, Proc. ESSV, 2021, PDF

Acoustic Correlates of Likable Speakers in the NSC Database

2020
Benjamin Weiss, Jürgen Trouvain and F. Burkhardt

Acoustic Correlates of Likable Speakers in the NSC Database, in book: Voice Attractiveness, Studies on Sexy, Likable, and Charismatic Speakers, DOI: 10.1007/978-981-15-6627-1_13, 2020

How should Pepper Sound – Preliminary Investigations on Robot Vocalizations

2019
F. Burkhardt, Milenko Saponja, Julian Sessner and Benjamin Weiss

How should Pepper Sound - Preliminary Investigations on Robot Vocalizations, Proc. of the ESSV 2019, 2019, PDF

Speech Synthesizing Simultaneous Emotion-Related States

2018
F. Burkhardt and Benjamin Weiss

Speech Synthesizing Simultaneous Emotion-Related States, Proc. of the Specom 2018, 2018, PDF

The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech

2018
Alice Baird, Emilia Parada-Cabaleiro, Simone Hantke, Felix Burkhardt, Nicholas Cummins and Björn Schuller

The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech, Proc. Interspeech, 2018

Robust Speech Emotion Recognition Under Different Encoding Conditions

2019
+ Oates, C., Triantafyllopoulos, A., Steiner, I., & Schuller, B. W.

Robust Speech Emotion Recognition Under Different Encoding Conditions. Proc. Interspeech 2019, 3935-3939.

Towards Robust Speech Emotion Recognition Using Deep Residual Networks for Speech Enhancement

2019
+ Triantafyllopoulos, A., Keren, G., Wagner, J., Steiner, I., & Schuller, B. W.

Towards Robust Speech Emotion Recognition Using Deep Residual Networks for Speech Enhancement. Proc. Interspeech 2019, 1691-1695.

Towards Speech Robustness for Acoustic Scene Classification

2020
+ Liu, S., Triantafyllopoulos, A., Ren, Z., & Schuller, B. W.

Towards Speech Robustness for Acoustic Scene Classification. Proc. Interspeech 2020, 3087-3091.

Spoken Language Identification by Means of Acosutic Mid-level Descriptors

2020
+ Reichel, U. D., Triantafyllopoulos, A., Oates, C., Huber, S., & Schuller, B.

Spoken Language Identification by Means of Acosutic Mid-level Descriptors. Studientexte zur Sprachkommunikation: Elektronische Sprachsignalverarbeitung 2020, 125-132.

Vergleich verschiedener Machine-Learning Ansätze zur kontinuierlichen Schätzung von perzeptivem Sprechtempo

2019
Weiss, B., Michael, T., Reichel, U., Pauly, O.

Vergleich verschiedener Machine-Learning Ansätze zur kontinuierlichen Schätzung von perzeptivem Sprechtempo, In: Birkholz, P., Stone, S. (Eds.): Elektronische Sprachverarbeitung. Studientexte zur Sprachkommunikation 93, pp 164-169, TUDpress, Dresden

Filled pause detection by prosodic discontinuity features

2019
Reichel, U.D., Weiss, B., Michael, T.

Filled pause detection by prosodic discontinuity features, In: Birkholz, P., Stone, S. (Eds.): Elektronische Sprachverarbeitung. Studientexte zur Sprachkommunikation 93, pp 272-279, TUDpress, Dresden

audEERING’s approach to the One-Minute-Gradual Emotion Challenge

2018
A. Triantafyllopoulos, H. Sagha, F. Eyben, B. Schuller

“audEERING’s approach to the One-Minute-Gradual Emotion Challenge,” arXiv preprint arXiv:1805.01222

Detecting Vocal Irony

2017
J. Deng, B. Schuller, “Detecting Vocal Irony

in Language Technologies for the Challenges of the Digital Age: 27th International Conference, GSCL 2017, Vol. 10713, p. 11, Springer

Emotion-awareness for intelligent vehicle assistants: a research agenda

H. J. Vögel, C. Süß, T. Hubregtsen, V. Ghaderi, R. Chadowitz, E. André, … & B. Huet

“Emotion-awareness for intelligent vehicle assistants: a research agenda,” in Proceedings of the 1st International Workshop on Software Engineering for AI in Autonomous Systems, pp. 11-15, ACM

Robust Laughter Detection for Wearable Wellbeing Sensing

2018
G. Hagerer, N. Cummins, F. Eyben, B. Schuller

“Robust Laughter Detection for Wearable Wellbeing Sensing,” in Proceedings of the 2018 International Conference on Digital Health, pp. 156-157, ACM

Deep neural networks for anger detection from real life speech data

2017
J. Deng, F. Eyben, B. Schuller, F. Burkhardt

“Deep neural networks for anger detection from real life speech data,” in Proc. of 2017 Seventh International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW), pp. 1-6, IEEE

Deep recurrent neural network-based autoencoders for acoustic novelty detection

2017
E. Marchi, F. Vesperini, S. Squartini, B. Schuller

“Deep recurrent neural network-based autoencoders for acoustic novelty detection,” in Computational intelligence and neuroscience, 2017

Did you laugh enough today? – Deep Neural Networks for Mobile and Wearable Laughter Trackers

2017
G. Hagerer, N. Cummins, F. Eyben, B. Schuller

“Did you laugh enough today? – Deep Neural Networks for Mobile and Wearable Laughter Trackers,” in Proc. Interspeech 2017, pp. 2044-2045

Automatic speaker analysis 2.0: Hearing the bigger picture

2017
B. Schuller

“Automatic speaker analysis 2.0: Hearing the bigger picture,” in Proc. of 2017 International Conference onSpeech Technology and Human-Computer Dialogue (SpeD), pp. 1-6, IEEE

Seeking the SuperStar: Automatic assessment of perceived singing quality

2017
J. Böhm, F. Eyben, M. Schmitt, H. Kosch, B. Schuller

“Seeking the SuperStar: Automatic assessment of perceived singing quality,” in Proc. of 2017 International Joint Conference on Neural Networks (IJCNN), pp. 1560-1569, IEEE

Enhancing LSTM RNN-Based Speech Overlap Detection by Artificially Mixed Data

2017
G. Hagerer, V. Pandit, F. Eyben, B. Schuller

“Enhancing LSTM RNN-Based Speech Overlap Detection by Artificially Mixed Data,” in Proc. 2017 AES International Conference on Semantic Audio

The effect of personality trait, age, and gender on the performance of automatic speech valence recognition

2017
H. Sagha, J. Deng, B. Schuller

“The effect of personality trait, age, and gender on the performance of automatic speech valence recognition,” in Proc. 7th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2017), San Antonio, Texas, AAAC, IEEE, October 2017

Automatic Multi-lingual Arousal Detection from Voice Applied to Real Product Testing Applications

2017
F. Eyben, M. Unfried, G. Hagerer, B. Schuller

“Automatic Multi-lingual Arousal Detection from Voice Applied to Real Product Testing Applications,” in Proc. 42nd IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2017), New Orleans, LA, IEEE

Real-time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms

2016
E. Marchi, F. Eyben, G. Hagerer, B. Schuller

“Real-time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms,” in Proc. INTERSPEECH 2016, San Francisco, Califorina, USA, pp. 1182-1183

A Paralinguistic Approach To Speaker Diarisation: Using Age, Gender, Voice Likability and Personality Traits

2017
Y. Zhang, F. Weninger, B. Liu, M. Schmitt, F. Eyben, B. Schuller

“A Paralinguistic Approach To Speaker Diarisation: Using Age, Gender, Voice Likability and Personality Traits,” in Proc. 2017 ACM Conference on Multimedia, Mountain View, California, USA, pp. 387-392

An Image-based Deep Spectrum Feature Representation for the Recognition of Emotional Speech

2017
N. Cummins, S. Amiriparian, G. Hagerer, A. Batliner, S. Steidl, B. Schuller

“An Image-based Deep Spectrum Feature Representation for the Recognition of Emotional Speech,” in Proc. 2017 ACM Conference on Multimedia, Mountain View, California, USA, pp. 478-484

Snore sound recognition: On wavelets and classifiers from deep nets to kernels

2017
K. Qian, C. Janott, J. Deng, C. Heiser, W. Hohenhorst, M. Herzog, N. Cummins, B. Schuller

“Snore sound recognition: On wavelets and classifiers from deep nets to kernels,” in Proc. 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 3737-3740

Introducing the Weighted Trustability Evaluator for Crowdsourcing Exemplified by Speaker Likability Classification

2016
S. Hantke, E. Marchi, B. Schuller

“Introducing the Weighted Trustability Evaluator for Crowdsourcing Exemplified by Speaker Likability Classification,” in Proc. LREC 2016

Real-life voice activity detection with LSTM Recurrent Neural Networks and an application to Hollywood movies

2013
F. Eyben, F. Weninger, S. Squartini, B. Schuller

“Real-life voice activity detection with LSTM Recurrent Neural Networks and an application to Hollywood movies,” in Proc. of 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 483-487, 26-31 May 2013. doi: 10.1109/ICASSP.2013.6637694

Affect recognition in real-life acoustic conditions – A new perspective on feature selection

2013
F. Eyben, F. Weninger, B. Schuller

“Affect recognition in real-life acoustic conditions – A new perspective on feature selection,” in Proc. of INTERSPEECH 2013, Lyon, France, pp. 2044-2048

Cross-Language Acoustic Emotion Recognition: An Overview and Some Tendencies

2015
S. Feraru, D. Schuller, B. Schuller

“Cross-Language Acoustic Emotion Recognition: An Overview and Some Tendencies,” in Proc. 6th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2015), (Xi’an, P. R. China), AAAC, IEEE, pp. 125-131, September 2015

Speech Analysis in the Big Data Era

2015
B. Schuller

“Speech Analysis in the Big Data Era,” in Proc. of the 18th International Conference on Text, Speech and Dialogue, TSD 2015, Lecture Notes in Artificial Intelligence (LNAI), Springer, September 2015, Satellite event of INTERSPEECH 2015

The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing

2015
F. Eyben, K. Scherer, B. Schuller, J. Sundberg, E. Andre, C. Busso, L. Devillers, J. Epps, P. Laukka, S. Narayanan, K. Truong

The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing,” IEEE Transactions on Affective Computing, 2015

Building Autonomous Sensitive Artificial Listeners (Extended Abstract)

2015
M. Schröder, E. Bevacqua, R. Cowie, F. Eyben, H. Gunes, D. Heylen, M. ter Maat, G. McKeown, S. Pammi, M. Pantic, C. Pelachaud, B. Schuller, E. de Sevin, M. Valstar, M. Wöllmer

“Building Autonomous Sensitive Artificial Listeners (Extended Abstract),” in Proc. of ACII 2015, Xi’an, China, invited for the Special Session on Most Influential Articles in IEEE Transactions on Affective Computing

Cross-Corpus Acoustic Emotion Recognition: Variances and Strategies (Extended Abstract)

2015
B. Schuller, B. Vlasenko, F. Eyben, M. Wöllmer, A. Stuhlsatz, A. Wendemuth, G. Rigoll

“Cross-Corpus Acoustic Emotion Recognition: Variances and Strategies (Extended Abstract),” in Proc. of ACII 2015, Xi’an, China, invited for the Special Session on Most Influential Articles in IEEE Transactions on Affective Computing

Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification (Extended Abstract)

2015
A. Metallinou, M. Wöllmer, A. Katsamanis, F. Eyben, B. Schuller, S. Narayanan

“Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification (Extended Abstract),” in Proc. of ACII 2015, Xi’an, China, invited for the Special Session on Most Influential Articles in IEEE Transactions on Affective Computing

iHEARu-PLAY: Introducing a game for crowdsourced data collection for affective computing

2015
S. Hantke, T. Appel, F. Eyben, B. Schuller

“iHEARu-PLAY: Introducing a game for crowdsourced data collection for affective computing,” in Proc. 6th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2015), Xi’an, P. R. China, AAAC, IEEE, pp. 891-897, September 2015

Real-time Robust Recognition of Speakers’ Emotions and Characteristics on Mobile Platforms

2015
F. Eyben, B. Huber, E. Marchi, D. Schuller, B. Schuller

“Real-time Robust Recognition of Speakers’ Emotions and Characteristics on Mobile Platforms,” in Proc. 6th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2015), Xi’an, P. R. China, AAAC, IEEE, pp. 778-780, September 2015

Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification (Extended Abstract)

2015
A. Metallinou, M. Wöllmer, A. Katsamanis, F. Eyben, B. Schuller, S. Narayanan

“Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification (Extended Abstract),” in Proc. of ACII 2015, Xi’an, China, invited for the Special Session on Most Influential Articles in IEEE Transactions on Affective Computing

iHEARu-PLAY: Introducing a game for crowdsourced data collection for affective computing

2015
S. Hantke, T. Appel, F. Eyben, B. Schuller

“iHEARu-PLAY: Introducing a game for crowdsourced data collection for affective computing,” in Proc. 6th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2015), Xi’an, P. R. China, AAAC, IEEE, pp. 891-897

Real-time Robust Recognition of Speakers’ Emotions and Characteristics on Mobile Platforms

2015
F. Eyben, B. Huber, E. Marchi, D. Schuller, B. Schuller

“Real-time Robust Recognition of Speakers’ Emotions and Characteristics on Mobile Platforms,” in Proc. 6th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2015), Xi’an, P. R. China, AAAC, IEEE, pp. 778-780