Abstract
A social agent such as a receptionist or an escort robot encounters challenges when communicating with people in open areas. The agent must know not to react to distracting acoustic and visual events and it needs to appropriately handle situations that include multiple humans, being able to to focus on active interlocutors and appropriately shift attention based on the context. We describe a multiparty interaction agent that helps multiple users arrange a common activity. From the user study we conducted, we found that the agent can discriminate between active and inactive interlocutors well by using the skeletal and azimuth information. Participants found the addressee much clearer when an animated talking head was used.
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Traum, D.: Issues in multiparty dialogues. Advances in agent communication, 1954–1954 (2004)
Clark, H.H., Carlson, T.B.: Hearers and speech acts. Language, 332–373 (1982)
Akker, R., Traum, D.: A comparison of addressee detection methods for multiparty conversations (2009)
Nakano, Y., Ishii, R.: Estimating user’s engagement from eye-gaze behaviors in human-agent conversations. In: International Conference on Intelligent user Interfaces, pp. 139–148. ACM (2010)
Bohus, D., Horvitz, E.: Learning to predict engagement with a spoken dialog system in open-world settings. In: Proceedings of the 2009 Annual Meeting of the Special Interest Group on Discourse and Dialogue, pp. 244–252 (2009)
Shotton, J., Fitzgibbon, A., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., Blake, A.: Real-time human pose recognition in parts from a single depth image. In: Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition (2011)
Foster, M., Gaschler, A., Giuliani, M., Isard, A., Pateraki, M., Petrick, R.: “two people walk into a bar”: Dynamic multi-party social interaction with a robot agent. In: Proc. of the 14th ACM International Conference on Multimodal Interaction ICMI (2012)
Fukuda, T., Taguri, J., Arai, F., Nakashima, M., Tachibana, D., Hasegawa, Y.: Facial expression of robot face for human-robot mutual communication. In: Proceedings of 2002 IEEE International Conference on Robotics and Automation, vol. 1, pp. 46–51. IEEE (2002)
Al Moubayed, S., Beskow, J., Skantze, G., Granström, B.: Furhat: A back-projected human-like robot head for multiparty human-machine interaction. In: Esposito, A., Esposito, A.M., Vinciarelli, A., Hoffmann, R., Müller, V.C. (eds.) COST 2102. LNCS, vol. 7403, pp. 114–130. Springer, Heidelberg (2012)
Bohus, D., Rudnicky, A.: The ravenclaw dialog management framework: architecture and systems. In: Proceedings of the 2008 Computer Speech and Language, pp. 332–361 (2008)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Pappu, A., Sun, M., Sridharan, S., Rudnicky, A. (2013). Situated Multiparty Interaction between Humans and Agents. In: Kurosu, M. (eds) Human-Computer Interaction. Interaction Modalities and Techniques. HCI 2013. Lecture Notes in Computer Science, vol 8007. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-39330-3_12
Download citation
DOI: https://doi.org/10.1007/978-3-642-39330-3_12
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-39329-7
Online ISBN: 978-3-642-39330-3
eBook Packages: Computer ScienceComputer Science (R0)