Veröffentlichungen

Veröffentlichungen

Wir sind Pioniere auf dem Gebiet der Audio-KI-Forschung. Die Technologie von AudEERING wird in vielen Forschungsprojekten eingesetzt. Über die Ergebnisse unserer Forschung informieren wir in zahlreichen Artikeln, Aufsätzen, Abhandlungen und anderen Publikationen. Werfen Sie auch einen Blick auf einige unserer wissenschaftlichen Zitate.

Age Classification: Comparison of Human vs Machine Performance in Prompted and Spontaneous Speech

2021
F. Burkhardt, Markus Brückl und Björn Schuller

Altersklassifizierung: Vergleich zwischen menschlicher und maschineller Leistung bei aufgeforderter und spontaner Sprache, Proc. ESSV, 2021, PDF

Acoustic Correlates of Likable Speakers in the NSC Database

2020
Benjamin Weiss, Jürgen Trouvain und F. Burkhardt

Acoustic Correlates of Likable Speakers in the NSC Database, in book: Voice Attractiveness, Studies on Sexy, Likable, and Charismatic Speakers, DOI: 10.1007/978-981-15-6627-1_13, 2020

How should Pepper Sound – Preliminary Investigations on Robot Vocalizations

2019
F. Burkhardt, Milenko Saponja, Julian Sessner und Benjamin Weiss

Wie soll Pepper klingen - Voruntersuchungen zu Robotervokalisationen, Proc. of the ESSV 2019, 2019, PDF

Speech Synthesizing Simultaneous Emotion-Related States

2018
F. Burkhardt und Benjamin Weiss

Speech Synthesizing Simultaneous Emotion-Related States, Proc. of the Specom 2018, 2018, PDF

The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech

2018
Alice Baird, Emilia Parada-Cabaleiro, Simone Hantke, Felix Burkhardt, Nicholas Cummins und Björn Schuller

The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech, Proc. Interspeech, 2018

Robust Speech Emotion Recognition Under Different Encoding Conditions

2019
Oates, C., Triantafyllopoulos, A., Steiner, I., & Schuller, B. W.

Robuste Sprach-Emotionserkennung unter verschiedenen Kodierungsbedingungen. Proc. Interspeech 2019, 3935-3939.

Towards Robust Speech Emotion Recognition Using Deep Residual Networks for Speech Enhancement

2019
+ Triantafyllopoulos, A., Keren, G., Wagner, J., Steiner, I., & Schuller, B. W.

Auf dem Weg zu einer robusten Emotionserkennung für Sprache mit Hilfe von Deep Residual Networks zur Sprachanhebung. Proc. Interspeech 2019, 1691-1695.

Towards Speech Robustness for Acoustic Scene Classification

2020
Liu, S., Triantafyllopoulos, A., Ren, Z., & Schuller, B. W.

Sprachliche Robustheit für die akustische Szenenklassifikation. Proc. Interspeech 2020, 3087-3091.

Spoken Language Identification by Means of Acosutic Mid-level Descriptors

2020
+ Reichel, U. D., Triantafyllopoulos, A., Oates, C., Huber, S., & Schuller, B.

Identifizierung gesprochener Sprache mit Hilfe von akosutischen Deskriptoren der mittleren Ebene. Studientexte zur Sprachkommunikation: Elektronische Sprachsignalverarbeitung 2020, 125-132.

Vergleich verschiedener Machine-Learning-Ansätze zur kontinuierlichen Schätzung von perzeptivem Sprechtempo

2019
Weiss, B., Michael, T., Reichel, U., Pauly, O.

Vergleich verschiedener Machine-Learning-Ansätze zur kontinuierlichen Schätzung von perzeptivem Sprechtempo, In: Birkholz, P., Stone, S. (Eds.): Elektronische Sprachverarbeitung. Studientexte zur Sprachkommunikation 93, S. 164-169, TUDpress, Dresden

Filled pause detection by prosodic discontinuity features

2019
Reichel, U.D., Weiss, B., Michael, T.

Erkennung gefüllter Pausen durch prosodische Diskontinuitätsmerkmale, In: Birkholz, P., Stone, S. (Eds.): Elektronische Sprachverarbeitung. Studientexte zur Sprachkommunikation 93, S. 272-279, TUDpress, Dresden

audEERING’s approach to the One-Minute-Gradual Emotion Challenge

2018
A. Triantafyllopoulos, H. Sagha, F. Eyben, B. Schuller

"audEERING's approach to the One-Minute-Gradual Emotion Challenge", arXiv preprint arXiv:1805.01222

Detecting Vocal Irony

2017
J. Deng, B. Schuller, "Erkennen von stimmlicher Ironie

in Language Technologies for the Challenges of the Digital Age: 27th International Conference, GSCL 2017, Vol. 10713, S. 11, Springer

Emotion-awareness for intelligent vehicle assistants: a research agenda

H. J. Vögel, C. Süß, T. Hubregtsen, V. Ghaderi, R. Chadowitz, E. André, ... & B. Huet

"Emotion-awareness for intelligent vehicle assistants: a research agenda," in Proceedings of the 1st International Workshop on Software Engineering for AI in Autonomous Systems, pp. 11-15, ACM

Robust Laughter Detection for Wearable Wellbeing Sensing

2018
G. Hagerer, N. Cummins, F. Eyben, B. Schuller

"Robust Laughter Detection for Wearable Wellbeing Sensing," in Proceedings of the 2018 International Conference on Digital Health, pp. 156-157, ACM

Deep neural networks for anger detection from real life speech data

2017
J. Deng, F. Eyben, B. Schuller, F. Burkhardt

"Deep neural networks for anger detection from real life speech data," in Proc. of 2017 Seventh International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW), pp. 1-6, IEEE

Deep recurrent neural network-based autoencoders for acoustic novelty detection

2017
E. Marchi, F. Vesperini, S. Squartini, B. Schuller

"Deep recurrent neural network-based autoencoders for acoustic novelty detection", in Computational intelligence and neuroscience, 2017

Did you laugh enough today? – Deep Neural Networks for Mobile and Wearable Laughter Trackers

2017
G. Hagerer, N. Cummins, F. Eyben, B. Schuller

"Haben Sie heute genug gelacht? - Deep Neural Networks for Mobile and Wearable Laughter Trackers," in Proc. Interspeech 2017, pp. 2044-2045

Automatic speaker analysis 2.0: Hearing the bigger picture

2017
B. Schuller

"Automatische Sprachanalyse 2.0: Hearing the bigger picture," in Proc. of 2017 International Conference onSpeech Technology and Human-Computer Dialogue (SpeD), pp. 1-6, IEEE

Seeking the SuperStar: Automatic assessment of perceived singing quality

2017
J. Böhm, F. Eyben, M. Schmitt, H. Kosch, B. Schuller

"Die Suche nach dem Superstar: Automatic assessment of perceived singing quality," in Proc. of 2017 International Joint Conference on Neural Networks (IJCNN), pp. 1560-1569, IEEE

Enhancing LSTM RNN-Based Speech Overlap Detection by Artificially Mixed Data

2017
G. Hagerer, V. Pandit, F. Eyben, B. Schuller

"Enhancing LSTM RNN-Based Speech Overlap Detection by Artificially Mixed Data," in Proc. 2017 AES International Conference on Semantic Audio

The effect of personality trait, age, and gender on the performance of automatic speech valence recognition

2017
H. Sagha, J. Deng, B. Schuller

"The effect of personality trait, age, and gender on the performance of automatic speech valence recognition," in Proc. 7th biannual Conference on Affective Computing and Intelligent Interaction (ACII 2017), San Antonio, Texas, AAAC, IEEE, October 2017

Automatic Multi-lingual Arousal Detection from Voice Applied to Real Product Testing Applications

2017
F. Eyben, M. Unfried, G. Hagerer, B. Schuller

"Automatic Multi-lingual Arousal Detection from Voice Applied to Real Product Testing Applications," in Proc. 42nd IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2017), New Orleans, LA, IEEE

Real-time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms

2016
E. Marchi, F. Eyben, G. Hagerer, B. Schuller

"Real-time Tracking of Speakers' Emotions, States, and Traits on Mobile Platforms," in Proc. INTERSPEECH 2016, San Francisco, Califorina, USA, pp. 1182-1183

A Paralinguistic Approach To Speaker Diarisation: Using Age, Gender, Voice Likability and Personality Traits

2017
Y. Zhang, F. Weninger, B. Liu, M. Schmitt, F. Eyben, B. Schuller

"A Paralinguistic Approach To Speaker Diarisation: Using Age, Gender, Voice Likability and Personality Traits," in Proc. 2017 ACM Conference on Multimedia, Mountain View, California, USA, S. 387-392