Publications

Publications

We are pioneers in the field of Audio AI research. AudEERING’s technology is used in many research projects. We provide information about the results of our research in numerous articles, essays, papers and other publications. Take a look at some of our scientific citations as well.

Age Classification: Comparison of Human vs Machine Performance in Prompted and Spontaneous Speech

2021
F. Burkhardt, Markus Brückl and Björn Schuller

Age Classification: Comparison of Human vs Machine Performance in Prompted and Spontaneous Speech, Proc. ESSV, 2021, PDF

Acoustic Correlates of Likable Speakers in the NSC Database

2020
Benjamin Weiss, Jürgen Trouvain and F. Burkhardt

Acoustic Correlates of Likable Speakers in the NSC Database, in book: Voice Attractiveness, Studies on Sexy, Likable, and Charismatic Speakers, DOI: 10.1007/978-981-15-6627-1_13, 2020

How should Pepper Sound – Preliminary Investigations on Robot Vocalizations

2019
F. Burkhardt, Milenko Saponja, Julian Sessner and Benjamin Weiss

How should Pepper Sound - Preliminary Investigations on Robot Vocalizations, Proc. of the ESSV 2019, 2019, PDF

Speech Synthesizing Simultaneous Emotion-Related States

2018
F. Burkhardt and Benjamin Weiss

Speech Synthesizing Simultaneous Emotion-Related States, Proc. of the Specom 2018, 2018, PDF

The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech

2018
Alice Baird, Emilia Parada-Cabaleiro, Simone Hantke, Felix Burkhardt, Nicholas Cummins and Björn Schuller

The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech, Proc. Interspeech, 2018

Robust Speech Emotion Recognition Under Different Encoding Conditions

2019
+ Oates, C., Triantafyllopoulos, A., Steiner, I., & Schuller, B. W.

Robust Speech Emotion Recognition Under Different Encoding Conditions. Proc. Interspeech 2019, 3935-3939.

Towards Robust Speech Emotion Recognition Using Deep Residual Networks for Speech Enhancement

2019
+ Triantafyllopoulos, A., Keren, G., Wagner, J., Steiner, I., & Schuller, B. W.

Towards Robust Speech Emotion Recognition Using Deep Residual Networks for Speech Enhancement. Proc. Interspeech 2019, 1691-1695.

Towards Speech Robustness for Acoustic Scene Classification

2020
+ Liu, S., Triantafyllopoulos, A., Ren, Z., & Schuller, B. W.

Towards Speech Robustness for Acoustic Scene Classification. Proc. Interspeech 2020, 3087-3091.

Spoken Language Identification by Means of Acosutic Mid-level Descriptors

2020
+ Reichel, U. D., Triantafyllopoulos, A., Oates, C., Huber, S., & Schuller, B.

Spoken Language Identification by Means of Acosutic Mid-level Descriptors. Studientexte zur Sprachkommunikation: Elektronische Sprachsignalverarbeitung 2020, 125-132.

Vergleich verschiedener Machine-Learning Ansätze zur kontinuierlichen Schätzung von perzeptivem Sprechtempo

2019
Weiss, B., Michael, T., Reichel, U., Pauly, O.

Vergleich verschiedener Machine-Learning Ansätze zur kontinuierlichen Schätzung von perzeptivem Sprechtempo, In: Birkholz, P., Stone, S. (Eds.): Elektronische Sprachverarbeitung. Studientexte zur Sprachkommunikation 93, pp 164-169, TUDpress, Dresden

Filled pause detection by prosodic discontinuity features

2019
Reichel, U.D., Weiss, B., Michael, T.

Filled pause detection by prosodic discontinuity features, In: Birkholz, P., Stone, S. (Eds.): Elektronische Sprachverarbeitung. Studientexte zur Sprachkommunikation 93, pp 272-279, TUDpress, Dresden

audEERING’s approach to the One-Minute-Gradual Emotion Challenge

2018
A. Triantafyllopoulos, H. Sagha, F. Eyben, B. Schuller

“audEERING’s approach to the One-Minute-Gradual Emotion Challenge,” arXiv preprint arXiv:1805.01222

Detecting Vocal Irony

2017
J. Deng, B. Schuller, “Detecting Vocal Irony

in Language Technologies for the Challenges of the Digital Age: 27th International Conference, GSCL 2017, Vol. 10713, p. 11, Springer

Emotion-awareness for intelligent vehicle assistants: a research agenda

H. J. Vögel, C. Süß, T. Hubregtsen, V. Ghaderi, R. Chadowitz, E. André, … & B. Huet

“Emotion-awareness for intelligent vehicle assistants: a research agenda,” in Proceedings of the 1st International Workshop on Software Engineering for AI in Autonomous Systems, pp. 11-15, ACM

Robust Laughter Detection for Wearable Wellbeing Sensing

2018
G. Hagerer, N. Cummins, F. Eyben, B. Schuller

“Robust Laughter Detection for Wearable Wellbeing Sensing,” in Proceedings of the 2018 International Conference on Digital Health, pp. 156-157, ACM

Deep neural networks for anger detection from real life speech data

2017
J. Deng, F. Eyben, B. Schuller, F. Burkhardt

“Deep neural networks for anger detection from real life speech data,” in Proc. of 2017 Seventh International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW), pp. 1-6, IEEE

Deep recurrent neural network-based autoencoders for acoustic novelty detection

2017
E. Marchi, F. Vesperini, S. Squartini, B. Schuller

“Deep recurrent neural network-based autoencoders for acoustic novelty detection,” in Computational intelligence and neuroscience, 2017

Did you laugh enough today? – Deep Neural Networks for Mobile and Wearable Laughter Trackers

2017
G. Hagerer, N. Cummins, F. Eyben, B. Schuller

“Did you laugh enough today? – Deep Neural Networks for Mobile and Wearable Laughter Trackers,” in Proc. Interspeech 2017, pp. 2044-2045

Automatic speaker analysis 2.0: Hearing the bigger picture

2017
B. Schuller

“Automatic speaker analysis 2.0: Hearing the bigger picture,” in Proc. of 2017 International Conference onSpeech Technology and Human-Computer Dialogue (SpeD), pp. 1-6, IEEE

Seeking the SuperStar: Automatic assessment of perceived singing quality

2017
J. Böhm, F. Eyben, M. Schmitt, H. Kosch, B. Schuller

“Seeking the SuperStar: Automatic assessment of perceived singing quality,” in Proc. of 2017 International Joint Conference on Neural Networks (IJCNN), pp. 1560-1569, IEEE

Enhancing LSTM RNN-Based Speech Overlap Detection by Artificially Mixed Data

2017
G. Hagerer, V. Pandit, F. Eyben, B. Schuller

“Enhancing LSTM RNN-Based Speech Overlap Detection by Artificially Mixed Data,” in Proc. 2017 AES International Conference on Semantic Audio

The effect of personality trait, age, and gender on the performance of automatic speech valence recognition

2017
H. Sagha, J. Deng, B. Schuller

“The effect of personality trait, age, and gender on the performance of automatic speech valence recognition,” in Proc. 7th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2017), San Antonio, Texas, AAAC, IEEE, October 2017

Automatic Multi-lingual Arousal Detection from Voice Applied to Real Product Testing Applications

2017
F. Eyben, M. Unfried, G. Hagerer, B. Schuller

“Automatic Multi-lingual Arousal Detection from Voice Applied to Real Product Testing Applications,” in Proc. 42nd IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2017), New Orleans, LA, IEEE

Real-time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms

2016
E. Marchi, F. Eyben, G. Hagerer, B. Schuller

“Real-time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms,” in Proc. INTERSPEECH 2016, San Francisco, Califorina, USA, pp. 1182-1183

A Paralinguistic Approach To Speaker Diarisation: Using Age, Gender, Voice Likability and Personality Traits

2017
Y. Zhang, F. Weninger, B. Liu, M. Schmitt, F. Eyben, B. Schuller

“A Paralinguistic Approach To Speaker Diarisation: Using Age, Gender, Voice Likability and Personality Traits,” in Proc. 2017 ACM Conference on Multimedia, Mountain View, California, USA, pp. 387-392