PL EN


Preferencje help
Widoczny [Schowaj] Abstrakt
Liczba wyników
Tytuł artykułu

Speech emotion recognition system for social robots

Autorzy
Treść / Zawartość
Identyfikatory
Warianty tytułu
Konferencja
National Conference on Robotics (12, 12-16.2012, Świeradów-Zdrój, Poland)
Języki publikacji
EN
Abstrakty
EN
The paper presents a speech emotion recognition system for social robots. Emotions are recognised using global acoustic features of the speech. The system implements the speech parameters calculation, features extraction, features selection and classification. All these phases are described. The system was verified using the two emotional speech databases: Polish and German. Perspectives for using such system in the social robots are presented.
Twórcy
  • Wrocław University of Technology, Institute of Computer Engineering, Control and Robotics, 50-370 Wrocław, Wybrzeże Wyspiańskiego 27
Bibliografia
  • [1] P. Boersma, “Accurate Short-Term Analysis of the Fundamental Frequency and the Harmonics-to-Noise Ratio of a Sampled Sound”, Institute of Phonetic Sciences, University of Amsterdam, Proceedings, vol. 17, 1993, pp. 97–110.
  • [2] R. R. Bouckaert, “Bayesian network classifiers in weka for version 3-5-7”, Network, 2008.
  • [3] R. Budziński, J. Kędzierski, and B. Weselak. “Head of social robot Samuel – construction (in Polish)”. In: Prace Naukowe Politechniki Warszawskiej, Elektronika, pp. 185–194, z. 175, t. I. Oficyna Wydawnicza Politechniki Warszawskiej, 2010.
  • [4] F. Burkhardt, A. Paeschke, M. Rolfes, W. F. Sendlmeier, and B. Weiss. A database of german emotional speech, volume 2005, pp. 3–6. Citeseer, 2005.
  • [5] S. Casale, A. Russo, G. Scebba, and S. Serrano, “Speech emotion classifiication using machine learning algorithms”. In: Proceedings of the 2008 IEEE International Conference on Semantic Computing, Washington, DC, USA, 2008, pp. 158–165.
  • [6] J. Cichosz and Ślot K., “Emotion recognition in speech signal using emotion-extracting binary decision trees”. In: Proceedings of the 2nd International Conference on Affective Computing and Intelligent Interaction (ACII): Doctoral Consortium, 2007.
  • [7] G. F. Cooper and T. Dietterich, “A bayesian method for the induction of probabilistic networks from data”. In: Machine Learning, 1992, pp. 309–347.
  • [8] R. Cowie, E. Douglas-Cowie, N. Tsapatsoulis, G. Votsis, S. Kollias, W. Fellenz, and J. G. Taylor, “Emotion recognition in human-computer interaction”, IEEE Signal Processing Magazine, vol. 18, no. 1, 2001, pp. 32–80.
  • [9] K. Dautenhahn. Socially intelligent agents: creating relationships with computers and robots, chapter Creating emotion recognition agents for speech signal. Multiagent systems, artifiicial societies, and simulated organizations. Kluwer Academic Publishers, 2002.
  • [10] M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. H. Witten, “The weka data mining software: an update”, SIGKDD Explor. Newsl., vol. 11, 2009, pp. 10–18.
  • [11] M. A. Hall. Correlation-based Feature Subset Selection for Machine Learning. PhD thesis, Department of Computer Science, University of Waikato, Hamilton, New Zealand, April 1999.
  • [12] S. Haykin, Advances in spectrum analysis and array processing, Prentice Hall advanced reference series: Engineering, Prentice Hall, 1995.
  • [13] E. Keller, “The analysis of voice quality in speech processing”. In: Lecture Notes in Computer Science, 2005, pp. 54–73.
  • [14] P. Langley and S. Sage, “Induction of selective bayesian classifiers”. In: Conference on uncertainity in artificial intelligence, 1994, pp. 399–406.
  • [15] Lodz University of Technology, Medical Electronics Division. “Database of Polish Emotional Speech”. Http://www.eletel.p..lodz.pl/bronakowski/med_catalog/docs/licence.txt
  • [16] X. Mao, B. Zhang, and Y. Luo, “Speech emotion recognition based on a hybrid of HMM/ANN”. In: Proceedings of the 7th Conference on 7th WSEAS International Conference on Applied Informatics and Communications - Volume 7, Stevens Point, Wisconsin, USA, 2007, pp. 367–370.
  • [17] I. Murray and J. Arnott, “Toward the simulation of emotion in synthetic speech: a review of the literature on human vocal emotion”, Journal of the Acoustic Society of America, vol. 93, no. 2, 1993, p. 1097–1108.
  • [18] T. L. Nwe, S. W. Foo, and L. C. D. Silva, “Speech emotion recognition using hidden markov models”, Speech Communication, vol. 41, 2003, pp. 603–623.
  • [19] A. Osherenko and E. André, “Differentiated semantic analysis in lexical affect sensing”, 2009 3rd International Conference on Affective Computingand Intelligent Interaction and Workshops, 2009, pp. 1–6.
  • [20] D. W. Paul Boersma. “Praat: doing phonetics by computer (version 5.2.05)”, 2010.
  • [21] S. Samad, A. Hussain, and L. K. Fah, “Pitch detection of speech signals using the cross-correlation technique”. In: TENCON 2000. Proceedings, vol. 1, 2000, pp. 283 – 286.
  • [22] S. Schnall, “The pragmatics of emotion language”, Psychological Inquiry, vol. 16, no. 1, 2005, pp. 28–31.
  • [23] B. Schuller, A. Batliner, D. Seppi, S. Steidl, T. Vogt, J. Wagner, L. Devillers, L. Vidrascu, N. Amir, L. Kossous, and V. Aharonson, “The relevance of feature type for the automatic classification of emotional user states: Low level descriptors and functionals”. In: Proceedings of Interspeech, Antwerp, Belgium, 2007.
  • [24] B. Schuller, S. Reiter, R. Muller, M. Al-Hames, M. Lang, and G. Rigoll, “Speaker independent speech emotion recognition by ensemble classification”. In: Multimedia and Expo, 2005. ICME 2005. IEEE International Conference on, 2005, pp. 864–867.
  • [25] B. Siciliano and O. Khatib, eds., Springer Handbook of Robotics, Springer, 2008.
  • [26] J. Sidorova, “Speech emotion recognition with TGI+.2 classifier”. In: Proceedings of the 12th Conference
  • of the European Chapter of the Association for Computational Linguistics: Student Research Workshop, 2009, pp. 54–60.
  • [27] S. S. Stevens, J. Volkmann, and E. B. Newman, “A scale for the measurement of the psychological magnitude pitch”, Journal of the Acoustical Society of America, vol. 8, no. 3, 1937, pp. 185–190.
  • [28] Łukasz Juszkiewicz. Postępy robotyki, chapter Speech emotion recognition system for social robots (in Polish), pp. 695–704. Oficyna wydawnicza PW, 2012.
  • [29] T. Vogt. Real-time automatic emotion recognition from speech. PhD thesis, Technischen Fakultät der Universität Bielefeld, 2010.
  • [30] Z. Xiao, E. Dellandréa, W. Dou, and L. Chen. “Hierarchical Classification of Emotional Speech”. Technical Report RR-LIRIS-2007-006, LIRIS UMR 5205 CNRS/INSA de Lyon/UniversitéClaude Bernard Lyon 1/Université Lumiére Lyon 2/École Centrale de Lyon, March 2007.
  • [31] M. You, C. Chen, J. Bu, J. Liu, and J. Tao, “A hierarchical framework for speech emotion recognition”. In: Industrial Electronics, 2006 IEEE International Symposium on, vol. 1, 2006, pp. 515–519.
  • [32] S. Zhang and Z. Zhao, “Feature selection filtering methods for emotion recognition in chinese speech signal”. In: Signal Processing, 2008. ICSP 2008. 9th International Conference on, 2008, pp. 1699–1702.
  • [33] J. Zhou, G. Wang, Y. Yang, and P. Chen, “Speech emotion recognition based on rough set and svm.”. In: Y. Yao, Z. Shi, Y. Wang, and W. Kinsner, eds., IEEE ICCI, 2006, pp. 53–61.
Typ dokumentu
Bibliografia
Identyfikator YADDA
bwmeta1.element.baztech-7e94690a-1db1-4cf4-8af6-86f3316de51b
JavaScript jest wyłączony w Twojej przeglądarce internetowej. Włącz go, a następnie odśwież stronę, aby móc w pełni z niej korzystać.