Abstract
Capturing and making use of observable actions and behaviours presents compelling opportunities for allowing end-users to interact with such data and eachother. For example, simple visualisations based on on detected behaviour or context allow users to interpret this data based on their existing knowledge and awarness of social cues. This paper presents one such “remote awareness” application where users can interpret a visualization based on simple behaviours to gain a sense of awareness of other users’ current context or actions. Using a prop embedded with sensors, users could control the visualisation using gesture and voice-based input. The results of this work describe the kinds of performances users generated during the trial, how they imagined the actions of their fellow participants based on the visualisation, and how the props containing sensors were used to support, or in some cases hinder, successful performance and interaction.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Benford, S., Greenhalgh, C., Giannachi, G., Walker, B., Marshall, J., Rodden, T.: Uncomfortable interactions. In: Proceedings of the 2012 ACM Annual Conference on Human Factors in Computing Systems, CHI 2012, pp. 2005–2014. ACM, New York (2012)
Brewer, J., Williams, A., Dourish, P.: A handle on what’s going on: combining tangible interfaces and ambient displays for collaborative groups. In: Proceedings of the 1st International Conference on Tangible and Embedded Interaction, TEI 2007, pp. 3–10. ACM, New York (2007)
Crossan, A., Murray-Smith, R., Brewster, S., Kelly, J., Musizza, B.: Gait phase effects in mobile interaction. In: CHI 2005 Extended Abstracts on Human Factors in Computing Systems, CHI EA 2005, pp. 1312–1315. ACM, New York (2005)
Crossan, A., Williamson, J., Brewster, S., Murray-Smith, R.: Wrist rotation for interaction in mobile contexts. In: Proceedings of the 10th International Conference on Human Computer Interaction with Mobile Devices and Services, MobileHCI 2008, pp. 435–438. ACM, New York (2008)
Dalsgaard, P., Hansen, L.K.: Performing perception—staging aesthetics of interaction. ACM Trans. Comput.-Hum. Interact. 15, 13:1–13:33 (2008)
Dey, A.K., de Guzman, E.: From awareness to connectedness: the design and deployment of presence displays. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI 2006, pp. 899–908. ACM, New York (2006)
Goffman, E.: Behavior in public places: notes on the social organization of gatherings. Free press paperback. Free Press (1966)
Goffman, E.: The presentation of self in everyday life. Penguin Psychology. Penguin (1990)
Hoggan, E., Brewster, S.A.: Crosstrainer: testing the use of multimodal interfaces in situ. In: Proceedings of the 28th International Conference on Human Factors in Computing Systems, CHI 2010, pp. 333–342. ACM, New York (2010)
Jacucci, G.: Interaction as Performance. Cases of configuring physical interfaces in mixed media. PhD thesis, University of Oulu (2004)
Jones, E., Alexander, J., Andreou, A., Irani, P., Subramanian, S.: Gestext: accelerometer-based gestural text-entry systems. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI 2010, pp. 2173–2182. ACM, New York (2010)
Jones, M.: Journeying toward extravagant, expressive, place-based computing. Interactions 18, 26–31 (2011)
Perry, M., Beckett, S., O’Hara, K., Subramanian, S.: Wavewindow: public, performative gestural interaction. In: ACM International Conference on Interactive Tabletops and Surfaces, ITS 2010, pp. 109–112. ACM, New York (2010)
Reeves, S., Benford, S., O’Malley, C., Fraser, M.: Designing the spectator experience. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI 2005, pp. 741–750. ACM, New York (2005)
Scheible, J., Ojala, T., Coulton, P.: Mobitoss: a novel gesture based interface for creating and sharing mobile multimedia art on large public displays. In: Proceedings of the 16th ACM International Conference on Multimedia, MM 2008, pp. 957–960. ACM, New York (2008)
Sheridan, J., Bryan-Kinns, N., Reeves, S., Marshall, J., Lane, G.: Graffito: crowd-based performative interaction at festivals. In: Proceedings of the 2011 Annual Conference Extended Abstracts on Human Factors in Computing Systems, CHI EA 2011, pp. 1129–1134. ACM, New York (2011)
Strachan, S., Murray-Smith, R., O’Modhrain, S.: Bodyspace: inferring body pose for natural control of a music player. In: CHI 2007 Extended Abstracts on Human Factors in Computing Systems, CHI EA 2007, pp. 2001–2006. ACM, New York (2007)
Vinciarelli, A., Murray-Smith, R., Bourlard, H.: Mobile social signal processing: vision and research issues. In: Proceedings of the 12th International Conference on Human Computer Interaction with Mobile Devices and Services, MobileHCI 2010, pp. 513–516. ACM, New York (2010)
Vinciarelli, A., Pantic, M., Bourlard, H.: Social signal processing: Survey of an emerging domain. Image Vision Comput. 27(12), 1743–1759 (2009)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Williamson, J.R., Brewster, S. (2014). Capturing Performative Actions for Interaction and Social Awareness. In: Murray-Smith, R. (eds) Mobile Social Signal Processing. MSSP 2010. Lecture Notes in Computer Science, vol 8045. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-54325-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-642-54325-8_6
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-54324-1
Online ISBN: 978-3-642-54325-8
eBook Packages: Computer ScienceComputer Science (R0)