Wissenschaftliche Zitierungen

Wissenschaftliche Zitierungen

Die Technologie von audEERING wird in zahlreichen Forschungsprojekten eingesetzt. Nachfolgend finden Sie eine Auswahl von Referenzen und Zitaten aus verschiedenen Bereichen. Bitte beachten Sie auch unsere eigenen wissenschaftlichen Veröffentlichungen.

  1. Singh, N., Singh, N., & Dhall, A. (2017). Continuous Multimodal Emotion Recognition Approach for AVEC 2017. arXiv preprint arXiv:1709.05861.
  2. Vielzeuf, V., Pateux, S., & Jurie, F. (2017, November). Temporal multimodal fusion for video emotion classification in the wild. In Proceedings of the 19th ACM International Conference on Multimodal Interaction (pp. 569-576). ACM.
  3. Tao, F., & Liu, G. (2018, April). Advanced LSTM: A study about better time dependency modeling in emotion recognition. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 2906-2910). IEEE.
  4. Tian, L., Muszynski, M., Lai, C., Moore, J. D., Kostoulas, T., Lombardo, P., ... & Chanel, G. (2017, Oktober). Erkennen induzierter Emotionen von Kinobesuchern: Are induced and perceived emotions the same? In Affective Computing and Intelligent Interaction (ACII), 2017 Seventh International Conference on (pp. 28-35). IEEE.
  5. Gamage, K. W., Sethu, V., & Ambikairajah, E. (2017, Oktober). Modeling variable length phoneme sequences-A step towards linguistic information for speech emotion recognition in wider world. In Affective Computing and Intelligent Interaction (ACII), 2017 Seventh International Conference on (pp. 518-523). IEEE.
  6. Knyazev, B., Shvetsov, R., Efremova, N., & Kuharenko, A. (2017). Convolutional neural networks pretrained on large face recognition datasets for emotion classification from video. arXiv preprint arXiv:1711.04598.
  7. Gaus, Y. F. A., Meng, H., & Jan, A. (2017, June). Decoupling Temporal Dynamics for Naturalistic Affect Recognition in a Two-Stage Regression Framework. In Cybernetics (CYBCONF), 2017 3rd IEEE International Conference on (pp. 1-6). IEEE.
  8. Cambria, E., Hazarika, D., Poria, S., Hussain, A., & Subramaanyam, R. B. V. (2017). Benchmarking multimodal sentiment analysis. arXiv preprint arXiv:1707.09538.
  9. Torres, J. M. M., & Stepanov, E. A. (2017, August). Enhanced face/audio emotion recognition: video and instance level classification using ConvNets and restricted Boltzmann Machines. In Proceedings of the International Conference on Web Intelligence (S. 939-946). ACM.
  10. Siegert, I., Lotz, A. F., Egorow, O., & Wendemuth, A. (2017, September). Improving Speech-Based Emotion Recognition by Using Psychoacoustic Modeling and Analysis-by-Synthesis. In International Conference on Speech and Computer (S. 445- 455). Springer, Cham.
  11. Huang, C. W., & Narayanan, S. S. (2017, Juli). Deep convolutional recurrent neural network with attention mechanism for robust speech emotion recognition. In Multimedia and Expo (ICME), 2017 IEEE International Conference on (pp. 583- 588). IEEE.
  12. Dhall, A., Goecke, R., Joshi, J., Wagner, M., & Gedeon, T. (2013, Dezember). Emotion recognition in the wild challenge 2013. In Proceedings of the 15th ACM on International conference on multimodal interaction (pp. 509-516). ACM.
  13. Dhall, A., Goecke, R., Joshi, J., Sikka, K., & Gedeon, T. (2014, November). Emotion recognition in the wild challenge 2014: Baseline, data and protocol. In Proceedings of the 16th International Conference on Multimodal Interaction (S. 461-466). ACM.
  14. Liu, M., Wang, R., Li, S., Shan, S., Huang, Z., & Chen, X. (2014, November). Combining multiple kernel methods on riemannian manifold for emotion recognition in the wild. In Proceedings of the 16th International Conference on Multimodal Interaction (pp. 494-501). ACM.
  15. Dhall, A., Ramana Murthy, O. V., Goecke, R., Joshi, J., & Gedeon, T. (2015, November). Video- und bildbasierte Emotionserkennung - Herausforderungen in der freien Wildbahn: Emotiw 2015. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (pp. 423-426). ACM.
  16. Savran, A., Cao, H., Shah, M., Nenkova, A., & Verma, R. (2012, Oktober). Combining video, audio and lexical indicators of affect in spontaneous conversation via particle filtering. In Proceedings of the 14th ACM international conference on Multimodal interaction (pp. 485-492). ACM.
  17. Poria, S., Cambria, E., & Gelbukh, A. F. (2015, September). Deep Convolutional Neural Network Textual Features and Multiple Kernel Learning for Utterance-level Multimodal Sentiment Analysis. In EMNLP (pp. 2539-2544).
  18. Zheng, W., Xin, M., Wang, X., & Wang, B. (2014). A novel speech emotion recognition method via incomplete sparse least square regression. IEEE Signal Processing Letters, 21(5), 569-572.
  19. Bhattacharya, A., Wu, W., & Yang, Z. (2012). Quality of experience evaluation of voice communication: an affect-based approach. Human-centric Computing and Information Sciences, 2(1), 7.
  20. Bone, D., Lee, C. C., & Narayanan, S. (2014). Robust unsupervised arousal rating: A rule-based framework withknowledge-inspired vocal features. IEEE transactions on affective computing, 5(2), 201-213.
  21. Liu, M., Wang, R., Huang, Z., Shan, S., & Chen, X. (2013, Dezember). Partial least squares regression on grassmannian manifold for emotion recognition. In Proceedings of the 15th ACM on International conference on multimodal interaction (pp. 525-530). ACM.
  22. Audhkhasi, K., & Narayanan, S. (2013). A globally-variant locally-constant model for fusion of labels from multiple diverse experts without using reference labels. IEEE transactions on pattern analysis and machine intelligence, 35(4), 769-783.
  23. Mariooryad, S., & Busso, C. (2013). Exploring cross-modality affective reactions for audiovisual emotion recognition. IEEE Transactions on affective computing, 4(2), 183-196.
  24. Chen, J., Chen, Z., Chi, Z., & Fu, H. (2014, November). Emotion recognition in the wild with feature fusion and multiple kernel learning. In Proceedings of the 16th International Conference on Multimodal Interaction (pp. 508-513). ACM.
  25. Rosenberg, A. (2012). Classifying Skewed Data: Importance Weighting to Optimize Average Recall. In Interspeech (pp. 2242-2245).
  26. Sun, R., & Moore, E. (2011). Untersuchung von glottalen Parametern und Teager-Energieoperatoren in der Emotionserkennung. Affective Computing and Intelligent Interaction, 425-434.
  27. Sun, B., Li, L., Zuo, T., Chen, Y., Zhou, G., & Wu, X. (2014, November). Combining multimodal features with hierarchical classifier fusion for emotion recognition in the wild. In Proceedings of the 16th International Conference on Multimodal Interaction (S. 481-486). ACM.
  28. Mariooryad, S., & Busso, C. (2015). Correcting time-continuous emotional labels by modeling the reaction lag of evaluators. IEEE Transactions on Affective Computing, 6(2), 97-108.
  29. Ivanov, A., & Riccardi, G. (2012, März). Kolmogorov-Smirnov test for feature selection in emotion recognition from speech. In Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on (pp. 5125-5128). IEEE.
  30. Mariooryad, S., & Busso, C. (2013, September). Analysis and compensation of the reaction lag of evaluators in continuous emotional annotations. In Affective Computing and Intelligent Interaction (ACII), 2013 Humaine Association Conference on (pp. 85-90). IEEE.
  31. Alonso-Martín, F., Malfaz, M., Sequeira, J., Gorostiza, J. F., & Salichs, M. A. (2013). Ein multimodales System zur Erkennung von Emotionen während der Mensch-Roboter-Interaktion. Sensors, 13(11), 15549-15581.
  32. Moore, J. D., Tian, L., & Lai, C. (2014, April). Word-level emotion recognition using high-level features. In International Conference on Intelligent Text Processing and Computational Linguistics (S. 17-31). Springer Berlin Heidelberg.
  33. Cao, H., Verma, R., & Nenkova, A. (2015). Speaker-sensitive emotion recognition via ranking: Studies on acted and spontaneous speech. Computer Speech & Language, 29(1), 186-202.
  34. Mariooryad, S., & Busso, C. (2014). Compensating for speaker or lexical variabilities in speech for emotion recognition. Speech Communication, 57, 1-12.
  35. Wu, C. H., Lin, J. C., & Wei, W. L. (2014). Survey on audiovisual emotion recognition: databases, features, and data fusion strategies. APSIPA transactions on signal and information processing, 3, e12.
  36. Busso, C., Mariooryad, S., Metallinou, A., & Narayanan, S. (2013). Iterative feature normalization scheme for automatic emotion detection from speech. IEEE transactions on Affective computing, 4(4), 386-397.
  37. Galanis, D., Karabetsos, S., Koutsombogera, M., Papageorgiou, H., Esposito, A., & Riviello, M. T. (2013, Dezember). Classification of emotional speech units in call centre interactions. In Cognitive Infocommunications (CogInfoCom), 2013 IEEE 4th International Conference on (pp. 403-406). IEEE.
  38. Sidorov, M., Brester, C., Minker, W., & Semenkin, E. (2014, Mai). Speech-Based Emotion Recognition: Feature Selection by Self-Adaptive Multi-Criteria Genetic Algorithm. In LREC (pp. 3481-3485).
  39. Oflazoglu, C., & Yildirim, S. (2013). Erkennung von Emotionen aus türkischer Sprache anhand akustischer Merkmale. EURASIP Journal on Audio, Speech, and Music Processing, 2013(1), 26.
  40. Kaya, H., & Salah, A. A. (2016). Combining modality-specific extreme learning machines for emotion recognition in the wild. Journal on Multimodal User Interfaces, 10(2), 139-149.
  41. Amer, M. R., Siddiquie, B., Richey, C., & Divakaran, A. (2014, Mai). Emotion detection in speech using deep networks. In Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on (pp. 3724-3728). IEEE.
  42. Poria, S., Chaturvedi, I., Cambria, E., & Hussain, A. (2016, Dezember). Convolutional MKL based multimodal emotion recognition and sentiment analysis. In Data Mining (ICDM), 2016 IEEE 16th International Conference on (pp. 439-448). IEEE.
  43. Kaya, H., Çilli, F., & Salah, A. A. (2014, November). Ensemble CCA for continuous emotion prediction. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge (pp. 19-26). ACM.
  44. Mariooryad, S., Lotfian, R., & Busso, C. (2014, September). Building a naturalistic emotional speech corpus by retrieving expressive behaviors from existing speech corpora. In INTERSPEECH (pp. 238-242).
  45. Busso, C., Parthasarathy, S., Burmania, A., Abdel-Wahab, M., Sadoughi, N., & Provost, E. M. (2017). MSP-IMPROV: An acted corpus of dyadic interactions to study emotion perception. IEEE Transactions on Affective Computing, 8(1), 67-80.
  46. Jin, Q., Li, C., Chen, S., & Wu, H. (2015, April). Speech emotion recognition with acoustic and lexical features. In Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on (pp. 4749-4753). IEEE.
  47. Peng, S. O. N. G., Yun, J. I. N., Li, Z. H. A. O., & Minghai, X. I. N. (2014). Speech emotion recognition using transfer learning. IEICE TRANSACTIONS on Information and Systems, 97(9), 2530-2532.
  48. Huang, D. Y., Zhang, Z., & Ge, S. S. (2014). Speaker state classification based on fusion of asymmetric simple partial least squares (SIMPLS) and support vector machines. Computer Speech & Language, 28(2), 392-419.
  49. Sun, Y., Wen, G., & Wang, J. (2015). Gewichtete spektrale Merkmale basierend auf lokalen Hu-Momenten für die Sprach-Emotionserkennung. Biomedizinische Signalverarbeitung und Steuerung, 18, 80-90.
  50. Kaya, H., Gürpinar, F., Afshar, S., & Salah, A. A. (2015, November). Contrasting and combining least squares based learners for emotion recognition in the wild. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (S. 459-466). ACM.
  51. Banda, N., & Robinson, P. (2011, November). Noise analysis in audio-visual emotion recognition. In Proceedings of the International Conference on Multimodal Interaction (pp. 1-4).
  52. Chen, S., & Jin, Q. (2015, Oktober). Multi-modal dimensional emotion recognition using recurrent neural networks. In Proceedings of the 5th International Workshop on Audio/Visual Emotion Challenge (pp. 49-56). ACM.
  53. Audhkhasi, K., Sethy, A., Ramabhadran, B., & Narayanan, S. S. (2012, März). Creating ensemble of diverse maximum entropy models. In Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on (pp. 4845-4848). IEEE.
  54. Lubis, N., Sakti, S., Neubig, G., Toda, T., Purwarianti, A., & Nakamura, S. (2016). Emotionen und ihre Auslöser im menschlichen Sprachdialog: Recognition and analysis. In Situated Dialog in Speech-Based Human-Computer Interaction (pp. 103-110). Springer International Publishing.
  55. Song, P., Jin, Y., Zha, C., & Zhao, L. (2014). Speech emotion recognition method based on hidden factor analysis. Electronics Letters, 51(1), 112-114.
  56. Dhall, A., Goecke, R., Joshi, J., Wagner, M., & Gedeon, T. (2013, Dezember). Emotion recognition in the wild challenge (EmotiW) challenge and workshop summary. In Proceedings of the 15th ACM on International conference on multimodal interaction (pp. 371-372). ACM.
  57. Chen, L., Yoon, S. Y., Leong, C. W., Martin, M., & Ma, M. (2014, November). Eine erste Analyse von strukturierten Videointerviews mit Hilfe von multimodaler Emotionserkennung. In Proceedings of the 2014 workshop on Emotion Representation and Modelling in Human-Computer-Interaction-Systems (pp. 1-6). ACM.
  58. Brester, C., Semenkin, E., Sidorov, M., & Minker, W. (2014). Self-adaptive multi-objective genetic algorithms for feature selection. In Proceedings of International Conference on Engineering and Applied Sciences Optimization (pp. 1838-1846).
  59. Tian, L., Lai, C., & Moore, J. (2015, April). Recognizing emotions in dialogues with disfluencies and non-verbal vocalisations. In Proceedings of the 4th Interdisciplinary Workshop on Laughter and Other Nonverbal Vocalisations in Speech (Vol. 14, S. 15).
  60. Lopez-Otero, P., Docio-Fernandez, L., & Garcia-Mateo, C. (2014). iVectors for continuous emotion recognition. Training, 45, 50.
  61. Bojanic, M., Crnojevic, V., & Delic, V. (2012, September). Application of neural networks in emotional speech recognition. In Neural Network Applications in Electrical Engineering (NEUREL), 2012 11th Symposium on (pp. 223-226). IEEE.
  62. Kim, J. C., & Clements, M. A. (2015). Multimodal affect classification at various temporal lengths. IEEE Transactions on Affective Computing, 6(4), 371-384.
  63. Bone, D., Lee, C. C., Potamianos, A., & Narayanan, S. S. (2014). An investigation of vocal arousal dynamics in child-psychologist interactions using synchrony measures and a conversation-based model. In INTERSPEECH (pp. 218-222).
  64. Day, M. (2013, Dezember). Emotion recognition with boosted tree classifiers. In Proceedings of the 15th ACM on International conference on multimodal interaction (pp. 531-534). ACM.
  65. Sidorov, M., Ultes, S., & Schmitt, A. (2014, Mai). Comparison of Gender and Speaker-adaptive Emotion Recognition. In LREC (pp. 3476-3480).
  66. Tian, L., Moore, J. D., & Lai, C. (2015, September). Emotion recognition in spontaneous and acted dialogues. In Affective Computing and Intelligent Interaction (ACII), 2015 International Conference on (pp. 698-704). IEEE.
  67. Sun, B., Li, L., Zhou, G., Wu, X., He, J., Yu, L., ... & Wei, Q. (2015, November). Combining multimodal features within a fusion network for emotion recognition in the wild. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (pp. 497-502). ACM.
  68. Ellis, J. G., Lin, W. S., Lin, C. Y., & Chang, S. F. (2014, Dezember). Predicting evoked emotions in video. In Multimedia (ISM), 2014 IEEE International Symposium on (pp. 287-294). IEEE.
  69. Brester, C., Semenkin, E., Kovalev, I., Zelenkov, P., & Sidorov, M. (2015, Mai). Evolutionäre Merkmalsauswahl für die Emotionserkennung in der mehrsprachigen Sprachanalyse. In Evolutionary Computation (CEC), 2015 IEEE Congress on (pp. 2406-2411). IEEE.
  70. Zhang, B., Provost, E. M., Swedberg, R., & Essl, G. (2015, Januar). Predicting Emotion Perception Across Domains: A Study of Singing and Speaking. In AAAI (S. 1328-1335).
  71. Brester, C., Sidorov, M., & Semenkin, E. (2014). Speech-based emotion recognition: Application of collective decision making concepts. In Proceedings of the 2nd International Conference on Computer Science and Artificial Intelligence (ICCSAI2014) (pp. 216-220).
  72. Cao, H., Savran, A., Verma, R., & Nenkova, A. (2015). Akustische und lexikalische Repräsentationen zur Vorhersage von Affekten in spontanen Gesprächen. Computer Speech & Language, 29(1), 203-217.
  73. Sidorov, M., Brester, C., Semenkin, E., & Minker, W. (2014, September). Speaker state recognition with neural network-based classification and self-adaptive heuristic feature selection. In Informatics in Control, Automation and Robotics (ICINCO), 2014 11th International Conference on (Vol. 1, pp. 699-703). IEEE.
  74. Tickle, A., Raghu, S., & Elshaw, M. (2013). Emotionale Erkennung aus dem Sprachsignal für einen virtuellen Bildungsagenten. In Journal of Physics: Conference Series (Vol. 450, No. 1, p. 012053). IOP Publishing.
  1. Vinciarelli, A., & Mohammadi, G. (2014). A survey of personality computing. IEEE Transactions on Affective Computing, 5(3), 273-291.
  2. Pohjalainen, J., Räsänen, O., & Kadioglu, S. (2015). Methoden der Merkmalsauswahl und ihre Kombinationen bei der hochdimensionalen Klassifizierung von Sprechersympathie, Verständlichkeit und Persönlichkeitsmerkmalen. Computer Speech & Language, 29(1), 145-171.
  3. Ivanov, A. V., Riccardi, G., Sporka, A. J., & Franc, J. (2011). Recognition of Personality Traits from Human Spoken Conversations. In INTERSPEECH (pp. 1549-1552).
  4. Chastagnol, C., & Devillers, L. (2012). Personality traits detection using a parallelized modified SFFS algorithm. computing, 15, 16.
  5. Alam, F., & Riccardi, G. (2013, August). Comparative study of speaker personality traits recognition in conversational and broadcast news speech. In INTERSPEECH (pp. 2851-2855).
  6. Alam, F., & Riccardi, G. (2014, Mai). Fusion of acoustic, linguistic and psycholinguistic features for speaker personality traits recognition. In Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on (pp. 955-959). IEEE.
  7. Wagner, J., Lingenfelser, F., & André, E. (2012). A Frame Pruning Approach for Paralinguistic Recognition Tasks. In INTERSPEECH (pp. 274-277).
  8. Feese, S., Muaremi, A., Arnrich, B., Troster, G., Meyer, B., & Jonas, K. (2011, Oktober). Discriminating individually considerate and authoritarian leaders by speech activity cues. In Privacy, Security, Risk and Trust (PASSAT) and 2011 IEEE Third Inernational Conference on Social Computing (SocialCom), 2011 IEEE Third International Conference on (pp. 1460-1465). IEEE.
  9. Liu, G., & Hansen, J. H. (2014). Supra-segmental feature based speaker trait detection. In Proc. Odyssey.
  10. Liu, C. J., Wu, C. H., & Chiu, Y. H. (2013, Oktober). BFI-based speaker personality perception using acoustic-prosodic features. In Signal and Information Processing Association Annual Summit and Conference (APSIPA), 2013 Asia-Pacific (pp. 1-6). IEEE.
  1. Grünerbl, A., Muaremi, A., Osmani, V., Bahle, G., Oehler, S., Tröster, G., ... & Lukowicz, P. (2015). Smartphone-basierte Erkennung von Zuständen und Zustandsänderungen bei Patienten mit bipolarer Störung. IEEE Journal of Biomedical and Health Informatics, 19(1), 140-148.
  2. Gravenhorst, F., Muaremi, A., Bardram, J., Grünerbl, A., Mayora, O., Wurzer, G., ... & Tröster, G. (2015). Mobiltelefone als medizinische Geräte in der Behandlung psychischer Störungen: ein Überblick. Personal and Ubiquitous Computing, 19(2), 335-353.
  3. Cummins, N., Joshi, J., Dhall, A., Sethu, V., Goecke, R., & Epps, J. (2013, Oktober). Diagnose von Depressionen anhand von Verhaltenssignalen: ein multimodaler Ansatz. In Proceedings of the 3rd ACM international workshop on Audio/visual emotion challenge (pp. 11-20). ACM.
  4. Alghowinem, S., Goecke, R., Wagner, M., Epps, J., Breakspear, M., & Parker, G. (2012, Mai). From Joyous to Clinically Depressed: Mood Detection Using Spontaneous Speech. In FLAIRS Conference.
  5. Joshi, J., Goecke, R., Alghowinem, S., Dhall, A., Wagner, M., Epps, J., ... & Breakspear, M. (2013). Multimodale assistive Technologien für die Diagnose und Überwachung von Depressionen. Journal on Multimodal User Interfaces, 7(3), 217-228.
  6. Alghowinem, S., Goecke, R., Wagner, M., Epps, J., Breakspear, M., & Parker, G. (2013, Mai). Detecting depression: a comparison between spontaneous and read speech. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on (pp. 7547-7551). IEEE.
  7. Cummins, N., Epps, J., Sethu, V., Breakspear, M., & Goecke, R. (2013, August). Modellierung der spektralen Variabilität für die Klassifizierung von depressiver Sprache. In Interspeech (S. 857-861).
  8. Alghowinem, S., Goecke, R., Wagner, M., Epps, J., Gedeon, T., Breakspear, M., & Parker, G. (2013, Mai). Eine vergleichende Studie verschiedener Klassifikatoren zur Erkennung von Depressionen in spontaner Sprache. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on (pp. 8022-8026). IEEE.
  9. Gupta, R., Malandrakis, N., Xiao, B., Guha, T., Van Segbroeck, M., Black, M., ... & Narayanan, S. (2014, November). Multimodale Vorhersage von affektiven Dimensionen und Depression in Mensch-Computer-Interaktionen. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge (pp. 33-40). ACM.
  10. Karam, Z. N., Provost, E. M., Singh, S., Montgomery, J., Archer, C., Harrington, G., & Mcinnis, M. G. (2014, Mai). Ökologisch valide Langzeit-Stimmungsüberwachung von Personen mit bipolarer Störung mittels Sprache. In Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on (pp. 4858-4862). IEEE.
  11. Mitra, V., Shriberg, E., McLaren, M., Kathol, A., Richey, C., Vergyri, D., & Graciarena, M. (2014, November). The SRI AVEC-2014 evaluation system. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge (pp. 93-101). ACM.
  12. Sidorov, M., & Minker, W. (2014, November). Emotionserkennung und Depressionsdiagnose durch akustische und visuelle Merkmale: A multimodal approach. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge (pp. 81-86). ACM.
  13. Kaya, H., & Salah, A. A. (2014, November). Eyes whisper depression: A cca based multimodal approach. In Proceedings of the 22nd ACM international conference on Multimedia (pp. 961-964). ACM.
  14. Hönig, F., Batliner, A., Nöth, E., Schnieder, S., & Krajewski, J. (2014, September). Automatische Modellierung von depressiver Sprache: relevante Merkmale und Relevanz des Geschlechts. In INTERSPEECH (pp. 1248-1252).
  15. Alghowinem, S., Goecke, R., Wagner, M., Epps, J., Parker, G., & Breakspear, M. (2013). Charakterisierung depressiver Sprache für die Klassifizierung. In Interspeech (pp. 2534-2538).
  16. Asgari, M., Shafran, I., & Sheeber, L. B. (2014, September). Inferring clinical depression from speech and spoken utterances. In Machine Learning for Signal Processing (MLSP), 2014 IEEE International Workshop on (pp. 1-5). IEEE.
  17. Lopez-Otero, P., Docio-Fernandez, L., & Garcia-Mateo, C. (2014, Mai). Eine Studie über akustische Merkmale für die Klassifizierung von depressiver Sprache. In Information and Communication Technology, Electronics and Microelectronics (MIPRO), 2014 37th International Convention on (pp. 1331-1335). IEEE.
  18. Lopez-Otero, P., Dacia-Fernandez, L., & Garcia-Mateo, C. (2014, März). A study of acoustic features for depression detection. In Biometrics and Forensics (IWBF), 2014 International Workshop on (pp. 1-6). IEEE.
  1. Nasir, M., Baucom, B. R., Georgiou, P., & Narayanan, S. (2017). Predicting couple therapy outcomes based on speech acoustic features. PloS one, 12(9), e0185123.
  2. Rao, H., Clements, M. A., Li, Y., Swanson, M. R., Piven, J., & Messinger, D. S. (2017). Paralinguistic Analysis of Children's Speech in Natural Environments. In Mobile Health (pp. 219-238). Springer, Cham.
  3. Chowdhury, S. A. (2017). Computational modeling of turn-taking dynamics in spoken conversations (Dissertation, Universität Trient).
  4. Silber-Varod, V., Lerner, A., & Jokisch, O. (2017). Automatic Speaker's Role Classification with a Bottom-up Acoustic Feature Selection. In Proc. GLU 2017 International Workshop on Grounding Language Understanding (pp. 52-56).
  5. Rehg, J., Abowd, G., Rozga, A., Romero, M., Clements, M., Sclaroff, S., ... & Rao, H. (2013). Decodierung des Sozialverhaltens von Kindern. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3414-3421).
  6. Wagner, J., Lingenfelser, F., Baur, T., Damian, I., Kistler, F., & André, E. (2013, Oktober). The social signal interpretation (SSI) framework: multimodal signal processing and recognition in real-time. In Proceedings of the 21st ACM international conference on Multimedia (pp. 831-834). ACM.
  7. Black, M. P., Katsamanis, A., Baucom, B. R., Lee, C. C., Lammert, A. C., Christensen, A., ... & Narayanan, S. S. (2013). Auf dem Weg zur Automatisierung eines Systems zur Kodierung des menschlichen Verhaltens bei der Interaktion von Ehepaaren anhand von akustischen Sprachmerkmalen. Speech Communication, 55(1), 1-21.
  8. Lee, C. C., Katsamanis, A., Black, M. P., Baucom, B. R., Christensen, A., Georgiou, P. G., & Narayanan, S. S. (2014). Computing vocal entrainment: A signal-derived PCA-based quantification scheme with application to affect analysis in married couple interactions. Computer Speech & Language, 28(2), 518-539.
  9. Black, M., Georgiou, P. G., Katsamanis, A., Baucom, B. R., & Narayanan, S. S. (2011, August). " You made me do it": Classification of Blame in Married Couples' Interactions by Fusing Automatically Derived Speech and Language Information. In Interspeech (pp. 89-92).
  10. Lubold, N., & Pon-Barry, H. (2014, November). Acoustic-prosodic entrainment and rapport in collaborative learning dialogues. In Proceedings of the 2014 ACM workshop on Multimodal Learning Analytics Workshop and Grand Challenge (pp. 5-12). ACM.
  11. Neiberg, D., & Gustafson, J. (2011). Predicting Speaker Changes and Listener Responses with and without Eye-Contact. In INTERSPEECH (pp. 1565-1568).
  12. Wagner, J., Lingenfelser, F., & André, E. (2013). Verwendung phonetischer Muster zur Erkennung sozialer Hinweise in natürlichen Gesprächen. In INTERSPEECH (pp. 168-172).
  13. Avril, M., Leclère, C., Viaux, S., Michelet, S., Achard, C., Missonnier, S., ... & Chetouani, M. (2014). Soziale Signalverarbeitung zur Untersuchung der Eltern-Kind-Interaktion. Frontiers in Psychology, 5, 1437.
  14. Jones, H. E., Sabouret, N., Damian, I., Baur, T., André, E., Porayska-Pomsta, K., & Rizzo, P. (2014). Interpreting social cues to generate credible affective reactions of virtual job interviewers. arXiv preprint arXiv:1402.5039.
  15. Zhao, R., Sinha, T., Black, A. W., & Cassell, J. (2016, September). Automatic recognition of conversational strategies in the service of a socially-aware dialog system. In 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue (S. 381).
  16. Rasheed, U., Tahir, Y., Dauwels, S., Dauwels, J., Thalmann, D., & Magnenat-Thalmann, N. (2013, Oktober). Real-Time Comprehensive Sociometrics for Two-Person Dialogs. In HBU (pp. 196-208).
  17. Sapru, A., & Bourlard, H. (2015). Automatic recognition of emergent social roles in small group interactions. IEEE Transactions on Multimedia, 17(5), 746-760.
  1. Muaremi, A., Arnrich, B., & Tröster, G. (2013). Zur Messung von Stress mit Smartphones und tragbaren Geräten während des Arbeitstages und des Schlafes. BioNanoScience, 3(2), 172-183.
  2. Van Segbroeck, M., Travadi, R., Vaz, C., Kim, J., Black, M. P., Potamianos, A., & Narayanan, S. S. (2014, September). Classification of cognitive load from speech using an i-vector framework. In INTERSPEECH (pp. 751-755).
  3. Aguiar, A. C., Kaiseler, M., Meinedo, H., Abrudan, T. E., & Almeida, P. R. (2013, September). Bewertung von Sprachstress durch physiologische und psychologische Messungen. In Proceedings of the 2013 ACM conference on Pervasive and ubiquitous computing adjunct publication (pp. 921-930). ACM.
  4. Li, M. (2014). Automatische Erkennung der physischen Belastung des Sprechers mit Hilfe von auf der Posteriorwahrscheinlichkeit basierenden Merkmalen aus akustischen und phonetischen Token.
  1. Niewiadomski, R., Hofmann, J., Urbain, J., Platt, T., Wagner, J., Piot, B., ... & Geist, M. (2013, Mai). Laugh-aware virtual agent and its impact on user amusement. In Proceedings of the 2013 international conference on Autonomous agents
  2. Gupta, R., Audhkhasi, K., Lee, S., & Narayanan, S. (2013). Paralinguistic event detection from speech using probabilistic time-series smoothing and masking. In Interspeech (pp. 173-177).
  3. Oh, J., Cho, E., & Slaney, M. (2013, August). Characteristic contours of syllabic-level units in laughter. In Interspeech (pp. 158-162).
  1. Pohjalainen, J., Räsänen, O., & Kadioglu, S. (2015). Methoden der Merkmalsauswahl und ihre Kombinationen bei der hochdimensionalen Klassifizierung von Sprechersympathie, Verständlichkeit und Persönlichkeitsmerkmalen. Computer Speech & Language, 29(1), 145-171.
  2. Carlson, N. A. (2017, September). Simple Acoustic-Prosodic Models of Confidence and Likability are Associated with Long-Term Funding Outcomes for Entrepreneurs. In International Conference on Social Informatics (pp. 3-16). Springer, Cham.
  1. Bone, D., Lee, C. C., Black, M. P., Williams, M. E., Lee, S., Levitt, P., & Narayanan, S. (2014). Der Psychologe als Gesprächspartner bei der Beurteilung von Autismus-Spektrum-Störungen: Insights from a study of spontaneous prosody. Journal of Speech, Language, and Hearing Research, 57(4), 1162-1177.
  2. Räsänen, O., & Pohjalainen, J. (2013, August). Random subset feature selection in automatic recognition of developmental disorders, affective states, and level of conflict from speech. In INTERSPEECH (pp. 210-214).
  3. Bone, D., Chaspari, T., Audhkhasi, K., Gibson, J., Tsiartas, A., Van Segbroeck, M., ... & Narayanan, S. (2013). Klassifizierung sprachbezogener Entwicklungsstörungen anhand von Sprachmerkmalen: das Versprechen und die potenziellen Hindernisse. In INTERSPEECH (S. 182-186).
  1. Reidsma, D., de Kok, I., Neiberg, D., Pammi, S. C., van Straalen, B., Truong, K., & van Welbergen, H. (2011). Kontinuierliche Interaktion mit einem virtuellen Menschen. Journal on Multimodal User Interfaces, 4(2), 97-118.
  2. Bevacqua, E., De Sevin, E., Hyniewska, S. J., & Pelachaud, C. (2012). A listener model: introducing personality traits. Journal on Multimodal User Interfaces, 6(1-2), 27-38.
  3. Kopp, S., van Welbergen, H., Yaghoubzadeh, R., & Buschmeier, H. (2014). An architecture for fluid real-time conversational agents: integrating incremental output generation and input processing. Journal on Multimodal User Interfaces, 8(1), 97-108.
  4. Neiberg, D., & Truong, K. P. (2011, Mai). Online-Erkennung von vokalen Hörerantworten mit maximaler Latenzzeit. In Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on (pp. 5836-5839). IEEE.
  5. Maat, M. (2011). Antwortauswahl und Turn-Taking für einen sensiblen künstlichen Höragenten. Universität Twente.
  6. Gebhard, P., Baur, T., Damian, I., Mehlmann, G., Wagner, J., & André, E. (2014, Mai). Exploring interaction strategies for virtual characters to induce stress in simulated job interviews. In Proceedings of the 2014 international conference on Autonomous agents and multi-agent systems (pp. 661-668). Internationale Stiftung für autonome Agenten und Multiagentensysteme.
  1. Potamitis, I., Ntalampiras, S., Jahn, O., & Riede, K. (2014). Automatische Vogelstimmenerkennung in langen Realfeldaufnahmen: Applications and tools. Applied Acoustics, 80, 1-9.
  2. Goëau, H., Glotin, H., Vellinga, W. P., Planqué, R., Rauber, A., & Joly, A. (2014, September). LifeCLEF bird identification task 2014. In CLEF2014.
  3. Lasseck, M. (2014). Large-scale Identification of Birds in Audio Recordings. In CLEF (Working Notes) (S. 643-653).
  4. Lasseck, M. (2015, September). Towards automatic large-scale identification of birds in audio recordings. In International Conference of the Cross-Language Evaluation Forum for European Languages (pp. 364-375). Springer International Publishing.

Lukacs, G., Jani, M., & Takacs, G. (2013, September). Acoustic feature mining for mixed speech and music playlist generation. In ELMAR, 2013 55th International Symposium (pp. 275-278). IEEE.

  1. Black, A. W., Bunnell, H. T., Dou, Y., Muthukumar, P. K., Metze, F., Perry, D., ... & Vaughn, C. (2012, March). Artikulatorische Merkmale für ausdrucksstarke Sprachsynthese. In Acoustics, Speech and Signal Processing (ICASSP), 2012 IEEE International Conference on (pp. 4005-4008). IEEE.
  2. Steidl, S., Polzehl, T., Bunnell, H. T., Dou, Y., Muthukumar, P. K., Perry, D., ... & Metze, F. (2012). Emotionsidentifikation zur Bewertung von synthetisierter emotionaler Sprache.
  3. Gallardo-Antolín, A., Montero, J. M., & King, S. (2014). Ein Vergleich von Open-Source-Segmentierungsarchitekturen für den Umgang mit unvollkommenen Daten aus den Medien in der Sprachsynthese.
  1. Alhanai, T., Au, R., & Glass, J. (2017, Dezember). Spoken language biomarkers for detecting cognitive impairment. In Automatic Speech Recognition and Understanding Workshop (ASRU), 2017 IEEE (pp. 409-416). IEEE.
  2. Bayestehtashk, A., Asgari, M., Shafran, I., & McNames, J. (2015). Vollautomatische Bewertung des Schweregrads der Parkinson-Krankheit anhand von Sprache. Computer Speech & Language, 29(1), 172-185.
  3. Bocklet, T., Steidl, S., Nöth, E., & Skodda, S. (2013). Automatic evaluation of parkinson's speech-acoustic, prosodic and voice related cues. In Interspeech (pp. 1149-1153).
  4. Orozco-Arroyave, J. R., Hönig, F., Arias-Londoño, J. D., Vargas-Bonilla, J. F., Daqrouq, K., Skodda, S., ... & Nöth, E. (2016). Automatische Erkennung der Parkinson-Krankheit in laufender Sprache in drei verschiedenen Sprachen. The Journal of the Acoustical Society of America, 139(1), 481-500.
  5. Kim, J., Nasir, M., Gupta, R., Van Segbroeck, M., Bone, D., Black, M. P., ... & Narayanan, S. S. (2015, September). Automatic estimation of parkinson's disease severity from various speech tasks. In INTERSPEECH (pp. 914-918).
  6. Pompili, A., Abad, A., Romano, P., Martins, I. P., Cardoso, R., Santos, H., ... & Ferreira, J. J. (2017, August). Automatic Detection of Parkinson's Disease: An Experimental Analysis of Common Speech Production Tasks Used for Diagnosis.
    In International Conference on Text, Speech, and Dialogue (pp. 411-419). Springer, Cham.
  1. Gajšek, R., Mihelic, F., & Dobrišek, S. (2013). Speaker State Recognition using an HMM-based feature extraction method. Computer Speech & Language, 27(1), 135-150.
  2. Bone, D., Li, M., Black, M. P., & Narayanan, S. S. (2014). Intoxicated speech detection: A fusion framework with speaker-normalized hierarchical functionalals and GMM supervectors. Computer Speech & Language, 28(2), 375-391.
  3. Suendermann-Oeft, D., Ramanarayanan, V., Teckenbrock, M., Neutatz, F., & Schmidt, D. (2015). HALEF: An Open-Source Standard-Compliant Telephony-Based Modular Spoken Dialog System: A Review and An Outlook. In Natural Language Dialog Systems and Intelligent Assistants (pp. 53-61). Springer International Publishing.
  4. Huang, C. L., Tsao, Y., Hori, C., & Kashioka, H. (2011, Oktober). Feature normalization and selection for robust speaker state recognition. In Speech Database and Assessments (Oriental COCOSDA), 2011 International Conference on (pp. 102-105). IEEE.

Kim, J., Kumar, N., Tsiartas, A., Li, M., & Narayanan, S. S. (2015). Automatische Verständlichkeitsklassifikation von pathologischer Sprache auf Satzebene. Computer Speech & Language, 29(1), 132-144.

  1. Lefter, I., Rothkrantz, L. J., & Burghouts, G. J. (2013). A comparative study on automatic audio-visual fusion for aggression detection using meta-information. Pattern Recognition Letters, 34(15), 1953-1963.
  2. Gosztolya, G., & Tóth, L. (2017). DNN-Based Feature Extraction for Conflict Intensity Estimation From Speech. IEEE Signal Processing Letters, 24(12), 1837-1841.

Audhkhasi, K., Zavou, A. M., Georgiou, P. G., & Narayanan, S. S. (2014). Theoretische Analyse der Diversität in einem Ensemble von automatischen Spracherkennungssystemen. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 22(3), 711-726.

  1. Forbes-Riley, K., Litman, D., Friedberg, H., & Drummond, J. (2012, June). Intrinsic and extrinsic evaluation of an automatic user disengagement detector for an uncertainty-adaptive spoken dialogue system. In Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (pp. 91-102). Vereinigung für Computerlinguistik
  2. Litman, D. J., Friedberg, H., & Forbes-Riley, K. (2012). Prosodic Cues to Disengagement and Uncertainty in Physics Tutorial Dialogues. In INTERSPEECH (pp. 755-758).
  1. Cmejla, R., Rusz, J., Bergl, P., & Vokral, J. (2013). Bayesian changepoint detection for the automatic assessment of fluency and articulatory disorders. Speech Communication, 55(1), 178-189.
  2. Chalasani, T. (2017). AUTOMATISIERTE BEWERTUNG DES THERAPIEERFOLGS BEIM FREMDAKZENTSYNDROM: Based on Emotional Temperature.

Kalantarian, H., & Sarrafzadeh, M. (2015). Audiobasierte Erkennung und Bewertung des Essverhaltens mithilfe der Smartwatch-Plattform. Computer in Biologie und Medizin, 65, 1-9.

  1. Metze, F., Rawat, S., & Wang, Y. (2014, Juli). Improved audio features for large-scale multimedia event detection. In Multimedia and Expo (ICME), 2014 IEEE International Conference on (pp. 1-6). IEEE.
  2. Rawat, S., Schulam, P. F., Burger, S., Ding, D., Wang, Y., & Metze, F. (2013). Robuste Audio-Codebücher für groß angelegte Ereigniserkennung in Verbrauchervideos.
  3. Avila, S., Moreira, D., Perez, M., Moraes, D., Cota, I., Testoni, V., ... & Rocha, A. (2014). RECOD bei MediaEval 2014: Violent Scenes Detection Task. In CEUR Workshop Proceedings. CEUR-WS.

Tran, T., Mariooryad, S., & Busso, C. (2013, Mai). Audiovisueller Korpus zur Analyse von Flüstersprache. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on (pp. 8101-8105). IEEE.

  1. Mariooryad, S., Kannan, A., Hakkani-Tur, D., & Shriberg, E. (2014, Mai). Automatic characterization of speaking styles in educational videos. In Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on (pp. 4848-4852). IEEE.
  2. Verkhodanova, V., Shapranov, V., & Kipyatkova, I. (2017, September). Hesitations in Spontaneous Speech: Acoustic Analysis and Detection. In International Conference on Speech and Computer (pp. 398-406). Springer, Cham.
  3. Lee, M., Kim, J., Truong, K., de Kort, Y., Beute, F., & IJsselsteijn, W. (2017, October). Exploring moral conflicts in speech: Multidisciplinary analysis of affect and stress. In Affective Computing and Intelligent Interaction (ACII), 2017 Seventh International Conference on (pp. 407-414). IEEE.

Ben Youssef, A., Shimodaira, H., & Braude, D. A. (2013). Artikulatorische Merkmale für sprachgesteuerte Kopfbewegungssynthese. Proceedings of Interspeech, Lyon, Frankreich.

Fan, Y., & Xu, M. (2014, Oktober). MediaEval 2014: THU-HCSIL Approach to Emotion in Music Task using Multi-level Regression. In MediaEval.

Heckmann, M. (2014, September). Steps towards more natural human-machine interaction via audio-visual word prominence detection. In International Workshop on Multimodal Analyses Enabling Artificial Agents in Human-Machine Interaction (pp. 15-24). Springer International Publishing.

  1. Hönig, F., Bocklet, T., Riedhammer, K., Batliner, A., & Nöth, E. (2012). The Automatic Assessment of Non-native Prosody: Combining Classical Prosodic Analysis with Acoustic Modelling. In INTERSPEECH (pp. 823-826).
  2. Finkelstein, S., Ogan, A., Vaughn, C., & Cassell, J. (2013). Alex: Ein virtueller Peer, der den Dialekt von Schülern identifiziert. In Proc. Workshop on Culturally-aware Technology Enhanced Learning in conjuction with EC-TEL 2013, Paphos, Cyprus, September 17.
  1. Weng, S., Chen, S., Yu, L., Wu, X., Cai, W., Liu, Z., ... & Li, M. (2015, Dezember). The SYSU system for the interspeech 2015 automatic speaker verification spoofing and countermeasures challenge. In Signal and Information Processing Association Annual Summit and Conference (APSIPA), 2015 Asia-Pacific (pp. 152-155). IEEE.
  2. Parthasarathy, S., & Busso, C. (2017, Oktober). Predicting speaker recognition reliability by considering emotional content. In Affective Computing and Intelligent Interaction (ACII), 2017 Seventh International Conference on (pp. 434 - 439). IEEE.
  1. Lehner, B., Widmer, G., & Bock, S. (2015, August). A low-latency, real-time-capable singing voice detection method with LSTM recurrent neural networks. In Signal Processing Conference (EUSIPCO), 2015 23rd European (pp. 21-25). IEEE.
  2. Sha, C. Y., Yang, Y. H., Lin, Y. C., & Chen, H. H. (2013, Mai). Klassifizierung der Klangfarbe von Gesangsstimmen in chinesischer populärer Musik. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on (pp. 734-738). IEEE.

Ghosh, A., & Riccardi, G. (2014, November). Recognizing human activities from smartphone sensor signals. In Proceedings of the 22nd ACM international conference on Multimedia (pp. 865-868). ACM.

Fehlt Ihre Veröffentlichung
?

Sie haben einen wissenschaftlichen Artikel, eine Diplomarbeit, ein Paper oder ähnliches mit Bezug zu audEERING veröffentlicht und möchten hier gelistet werden? Dann kontaktieren Sie uns info@audeering.com