Keywords

2.1 Introduction

The realization of a robotic agent capable of natural interactions with humans raises a number of issues and problems related to (a) the acquisition of sufficient competences to afford a full description and understanding of the user’s perspective, and thus the creation of reliable and functional user models, (b) the agent’s appearance and design, including the to be implemented interface, and (c) the reliability and credibility of a such artificial agent under the variety of application areas its tasks are devoted to [7, 12, 14]. The application areas may vary from training/education systems, to human-computer interfaces, entertainment, computer games, and several more. Each may require different agent’s abilities, at a different level of complexity, ranging from natural speech communication to autonomous behaviors, detection and interpretation of user moods, personalities, and emotional behaviors, adaptation to the physical, organizational, and socio-cultural context, as well as, user needs and requirements.

Robotic socially believable behaving agents must be adaptive systems programmed to interact with humans by simulating typical spontaneous human interactions. To this aim it is needed a comprehensive model of the possible sets of “user” final conducts (the “user model”), as well as, in order for the agent to properly act, it is necessary to provide a proprioceptive description of the agent itself (the “model of itself”, [5]. In human-machine interaction, a user model is what the system “knows “ about the perspective and/or the expectations of the actual user, normally including both psychological data and a user profile that estimates the preferences of a specific class or a wide range of users. Psychological data tend to function effectively as a paradigm for a wider set of users, and are a reliable source for modeling the user’s prospective [24]. For example, in the case of agents devoted to act a “butler” for guiding users in a foreign city, the user model needs to provide a comprehensive inventory of possible human requests and actions/reactions to expected and unexpected situations in terms of the reliability, confidence, facility of interaction/communication, as well as, trustworthiness, satisfaction, and credibility the users ascribe to the agent under different circumstances. Moreover, together a model of the possible user behavioral response, it is also important to anticipate such responses by means of a motivated design in order to minimize negative effects that can produce the user rejection of the agent.

Since emotions play a very important role in many aspects of our lives, including decision making, perception, learning, and behavior, and emotional skills are an important component of intelligence, especially for human-human interactions, the next generation of cognitive architectures must integrate and incorporate principles of emotionally colored interactions to define realistic models of human-machine interaction and suggest novel computational approaches for the implementation of real-time believable, autonomous, adaptive, and context-aware robotic agents [10, 13]. To this aim, fair efforts have been devoted to recognize and synthesize human emotional states exploiting different modalities, including: speech, facial expressions, gestures, and physiological signals as EEG, ECG, and others [13, 2023, 26]. Some works investigated on the possibility to combine signals from multiple sources, with the general aim of improving the emotion classification accuracy and synthesis [6, 11]. However, a truly cognitive real-time multimodal system capable of working with spontaneous emotional material is still missing. The chapters proposed in this book aim to further progress in this direction, even though automatic synthesis and recognition of human emotions remain an example of computational task that cannot be perfectly solved by classical pattern recognition and machine learning algorithms.

2.2 Content of the Book

Sophisticated and functional computational instruments able to recognize, process, store, and synthesize relevant emotional signals, as well as interact with people, displaying reactions that show abilities of appropriately sensing and understanding (under conditions of limited time) environmental changes and producing suitable, autonomous, and adaptable responses created great promises in Information Communication Technology (ICT). However, Human Computer Interaction (HCI) paradigms that account for users’ emotional states and emotional reactions are still far from being implemented. This book is signaling advances in these directions taking into account the multifunctional role of emotions in driving communication and interactional exchanges. To this aim, the book includes nine investigations on the role of emotions in shaping intentions, motivations, learning, and decision making. The first chapter by Vernon et al. [25] affords the problem of modeling (in socially believable behaving systems) actions, attentions, goals, and intentions, particularly intentions considered by the authors as the ability of the agent to read others minds and understand others’ perspectives and beliefs. In this context, emotions will drive the efficient behavior of the agents, since as stated by the authors, the perception of social stimuli produces “bodily states in the perceiving agent [and] trigger [in it] affective states [which in turn affect] the agent’s physical and cognitive performance” [25, p. 3]. Corrigan et al. [8] faces similar problems from an implementation point of view considering aspects of human engagements with robotic and virtual agents, which are still actions guided by intentions and abilities to understand others’ perspectives. The contribution of Belpaeme et al. [4] approaches similar problematics at a developmental level. The authors show that “cognition emerges from the interaction between the brain, the body and the physical environment” exploiting the iCub humanoid platform. The interesting paradigm emerging from their results is that “artificial cognition, just as its natural counterpart, benefits from being grounded and embodied” in a body, a brain, a physical, social context. The missing of one of these four constituents still allow to create artificial cognition, however, the complete system seems to be more efficient and effective. Meudt’s et al. [19] contribution intend to show that the ability of an agent to recognize users’ emotions will facilitate the users’ adaptation process and improve the human interaction with the agent, making its function as companion or assistive technology more reliable. The contribution of Lewandowska-Tomaszczyk and Wilson [17] underline the physical and moral role of the disgust across different cultures, showing how important is the social setting and why “an emotion-sensitive socially interacting robot would need to encode and decode [such emotion] in order to competently and appropriately interact with the environmental culture [17, p. 1]. On a similar theme is the contribution of Maricchiolo et al. [18], which affords the analysis of nonverbal (gestural) and physiological (heart rate and skin conductance) reactions to disagreeable (disgusting?) messages. Surely very far from disgust is the contribution of Dupont et al. [9] which reports on the result of a four-year EU project investigation on the “laughter”: the ILHAIRE Project (http://www.ilhaire.eu/project). The authors describe the collected data and the multi-determined role of the “laughter” in social interaction. Finally, the last two contributions of this book by Hunyadi et al. [16] and Gangamohan et al. [15] are dedicated to emotional speech and in particular to emotional prosodic features extracted from the HuComTech Corpus and all the emotional speech features able to recognize emotional vocal expressions.

2.3 Conclusions

The readers of this book will get a taste of the major research areas in modeling emotions and of the multifunctional role of emotions in generating actions, goals, attentions, and intentions, as well as on the different paradigms to model emotions by analyzing interactional exchanges. The research topics afforded by the book cover research fields related to psychology, sociology, philosophy, computer science, robotics, signal processing and human-computer interaction. The contributors to this volume are leading authorities in their respective fields. The book captures and presents interesting aspects of communicative exchanges and is fundamental in studies, such as robotics, where multidisciplinary facets need to be considered in order to succeed in the implementation of robotic companions, and assistive technologies. In particular the book covers aspects of emotional information processing during interactional exchanges which would lead to the implementation of socially believable, autonomous, adaptive, context-aware situated HCI systems.