PL EN


Preferencje help
Widoczny [Schowaj] Abstrakt
Liczba wyników
Tytuł artykułu

Integration of Gesture Planning System with Social Robot Hand

Identyfikatory
Warianty tytułu
Języki publikacji
EN
Abstrakty
EN
Nonverbal communication is one of the key issues in social robotics. This work presents an adaptation of a virtual agent gesture engine to a real social companion. The system performance is illustrated with an exemplary simulation. The solution quality is discussed.
Rocznik
Strony
703--712
Opis fizyczny
Bibliogr. 18 poz., rys., wykr.
Twórcy
autor
  • Institute of Computer Engineering, Control and Robotics, Wrocław University of Technology, ul. Janiszewskiego 11/17, 50-372 Wrocław, adam.oleksy@pwr.wroc.pl
Bibliografia
  • [1] C. Breazeal. Designing Sociable Robots. Cambridge, The MIT Press 2002.
  • [2] J. Kedzierski et al. Foundations of embodied companions. Technical report, 7FP LIREC GROUP Deliverable 6.2, 2009.
  • [3] Prillwitz et al. HamNoSys. Version 2.0. Hamburg Notation System for Sign Language. An Introductory Guide. Hamburg, Signum 1992.
  • [4] B. Hartmann, Mancini M., Pelachaud C. Formational parameters and adaptive prototype instantiation for MPEG-4 compliant gesture synthesis. In: Computer Animation, IEEE. Proceedings. IEEE Comp. Society Press, 2002, pp. 111-119.
  • [5] B. Hartmann, M. Mancini, C. Pelachaud. Towards affective agent action: Modelling expressive ECA gestures. In: Int. Conf. on Intelligent User Interfaces - Workshop on Affective Interaction. Proceedings, 2005.
  • [6] K. Hirai et al. The development of Honda humanoid robot. In: Int. Conf. on Robotics and Automation. Proceedings, 1998. Vol. 2, pp. 1321-1326.
  • [7] M. Kallmann, S. Marsella. Hierarchical motion controllers for real-time autonomous virtual humans. In: 5th Int. Conf. on Interactive Virtual Agents. Proceedings. Springer, 2005, pp. 12-14.
  • [8] A. Kendon. Movement coordination in social interaction: Some examples described. Acta Psychologica, 1970, Vol. 32, pp. 101-125.
  • [9] D. Kochanek, R. Bartels. Interpolating splines with local tension, continuity, and bias control. SIGGRAPH Comput. Graph., 1984, Vol. 18, No. 3, pp. 33-41.
  • [10] M. Koga, Y. Hosoda, T. Moriya. Humanoid robots. Hitachi Review, 2009, Vol. 58, No. 4, pp. 151-156.
  • [11] S. Kopp, I. Wachsmuth. A knowledge-based approach for lifelike gesture animation. In: ECAI. Proceedings, 2000, pp. 663-667.
  • [12] S. Kopp, I. Wachsmuth. Planning and motion control in lifelike gesture: A refined approach. In: Computer Animation. Proceedings. IEEE Computer Society, 2000, pp. 92-97.
  • [13] D. McNeill. What Gestures Reveal about Thought. Chicago, The University of Chicago Press 1992.
  • [14] M. Salem et al. Towards meaningful robot gesture. In: 3rd Int. Workshop on Human-Centered Robotic Systems. Proceedings. Springer, 2009. To appear.
  • [15] M. Thiebaux, S. Marsella. Smartbody: Behavior realization for embodied conversational agents. In: 7th Int. Joint Conf. on Autonomous Agents and Multi-Agent Systems. Proceedings, 2008, pp. 151-158.
  • [16] Humaine. Project website, http://emotion-research.net/
  • [17] LIREC. Project website, http://www.lirec.eu
  • [18] MIT Media Lab. MDS, http://robotic.media.mit.edu/projects/robots/mds/overview/overview.html
Typ dokumentu
Bibliografia
Identyfikator YADDA
bwmeta1.element.baztech-article-PWA9-0046-0034
JavaScript jest wyłączony w Twojej przeglądarce internetowej. Włącz go, a następnie odśwież stronę, aby móc w pełni z niej korzystać.