Keywords

1 Introduction

In social communication activities, understanding the counterpart’s expressions of agreement through social signals (such as vocal-verbal unimodal feedback hm, yeah, no, gestural unimodal feedback a head nod/shake, or multimodal feedback yes in combination with a head nod) is a fundamental aspect for achieving mutual understanding in interaction. Expressing agreement indicates “the situation in which people have the same opinion, or in which they approve or accept something” (Cambridge University Press 2020). People can share convergent or divergent opinions, proposals, goals, reciprocal relations, attitudes and feelings through social signals of agreement and disagreement (Khaki et al. 2016). Recognizing and interpreting such social signals is an essential capacity for social life, and for developing Intelligent Systems, from dialogue systems to Embodied Agents capable of sensing, interpreting and delivering social signals (Poggi et al. 2011). Thus, research on multimodal human interaction in naturalistic settings from various communication contexts is essential to enhance understanding of social signals and its applications in digital technology development.

Earlier research has shown that people express agreement in different ways, and culture is one of the factors which influences it. In high-context cultures such as China and Malaysia, face saving leads to avoidance of expressing disagreement and saying ‘yes’ doesn’t necessarily mean agreement, which is the opposite to low context cultures such as Germany and the U.S. (Hall 1959; Kevin 2004). The research of Swedes’ communication has shown that they give much feedback in interactions, signaling seeking consensus and avoiding conflict (Pedersen 2010). Gender is another factor influencing agreement expression. The research of Malaysian Chinese agreement strategies shows that females tend to express more agreement in interactions than the males (Azlina Abdul 2017).

Few studies in cross-cultural communication research focus on analysis of real-time interactions. Currently, there is little research with a focus on multimodal communication and cross-cultural comparison between Sweden and China, although they have increasing contacts and collaborations in both academia and industry, which motivates why more research is needed in this area.

2 Background

Social signals are perceivable stimuli that, either directly or indirectly, convey information concerning social actions, social interaction, attitudes, social emotions and social relations (Vinciarelli et al. 2012). Expressing agreement indicates “the situation in which people have the same opinion, or in which they approve or accept something” (Cambridge University Press 2020). Agreement and disagreement are expressed by specific social signals, through which people in the interaction express, for example, whether they share the same opinions, whether they accept each other’s proposals, whether they have convergent or divergent goals, attitudes and feelings, how their reciprocal relations are, and how to predict the development and outcome of the discussion or negotiation.When people argue in a discussion or negotiation, it is important for people to understand whether, what, and how they agree or disagree.

In order to perceive and understand the social signals of agreement and disagreement, one must be able to catch and interpret the relevant feedback expressions, which are often expressed in different communicative modalities – words, gestures, intonation, facial expressions, gaze, head movements, and posture (Gander 2018; Poggi et al. 2011). One may look at the other while smiling and nodding to express agreement, while shaking head and saying “no” to signal disagreement. Gestural unimodal expressions of agreement (such as head nod, finger thumb up) have been researched in different datasets such as TV live political debate (Poggi et al. 2011), dyadic dialogues about the likes and dislikes on common topics such as sports, movies, and music (Khaki et al. 2016), and human-computer interaction through gestures via touchless interfaces (Madapana et al. 2020). Expressions of agreement have also been explored for developing an implementation and training model of agreement negotiation dialogue, for example, with a dataset of dialogues where people have contradictory communicative goals and try to reach agreement, in particular focusing on the sequence of communicative acts and the vocal-verbal unimodal means (Koit 2016). Mehu and van der Maaten (2014) stressed the importance of using multimodal and dynamic features to investigate the cooperative role of pitch, vocal intensity and speed of speech to verbal utterance of agreement and disagreement. Although researching both visual and auditory social signals of agreement and disagreement has obtained growing attention, studies on multimodal communication of agreement and disagreement still focus more on the role of nonverbal behavior in social perception (e.g., Mehu and van der Maaten 2014) rather than the combination of multimodalities in expressing agreement and disagreement. This paper will investigate both gestural unimodal, vocal-verbal unimodal, and their combined multimodal communicative behaviors to contribute to big data corpus of naturalistic interactions. Such empirical research findings can be used for developing and evaluating statistical models of agreement and disagreement in interactions for instance, and also for designing technology of autonomous speech and gesture recognition of agreement communication.

Besides modality issues, there have been also cultural and gender concerns of agreement communication. As known, understanding the counterpart’s expressions of agreement is a fundamental aspect for achieving mutual understanding in cross-cultural communication activities. Across cultures, people express agreement in different ways. The research about the Malaysian Chinese agreement strategies shows that interlocutors tend to complete and repeat part of the previous speaker’s utterance and give positive feedback, which is explained by striving for supporting speaker’s face and harmony rooted in Chinese cultural values (Azlina Abdul 2017). The research also emphasizes that female speakers tend to express more agreement than male speakers. In high-context cultures, for example, China and Malaysia, face saving leads to avoiding expressing open disagreement with interlocutors and saying ‘yes’ doesn’t necessarily mean agreement; whereas, a more direct communication like “no, I don’t think so” or “sorry, I don’t agree” is common in low context cultures, such as Germany and the U.S. (Hall 1959; Kevin 2004). The Swedes are known for giving much feedback in interactions, signaling consensus seeking and conflict avoidance (Pedersen 2010). Few studies in cross-cultural communication research focus on analysis of real-time interactions. Also, there is little research focus on multimodal communication and cross-cultural comparison between Swedish and Chinese communication patterns, though contacts between Chinese and Swedes are increasing. In recent years, Chinese direct investments in Sweden have tripled and are expected to increase further (Weissmann and Rappe 2017). Sweden and China have many international collaborations in both academia and industry. However, few studies of real-time interactions have assessed similarities and differences between Swedish and Chinese cultures in general, with little focus on expressing agreement in particular. We are going to address this in the present paper.

3 Research Aim and Questions

This paper aims at exploring and comparing how Chinese and Swedes express agreement in interactions. More specifically, it focuses on the multimodal communicative feedback for expressing agreement, including vocal, gestural and combination of vocal and verbal feedback expressions.

Two specific research questions will be addressed:

  1. 1.

    What feedback expressions are used for expressing agreement in the Swedish and the Chinese conversations?

  2. 2.

    What similarities and differences can be observed between the Swedish and the Chinese expressions of agreement?

4 Methodology

4.1 Data Collection and Participants

Our data consist of four Swedish-Swedish and four Chinese-Chinese first-time encounters’ dyadic conversations between eight university students in Sweden. The conversations were in languages of Swedish and Mandarin Chinese respectively and were video- and audio-recorded in a university setting. Each conversation lasts approximately ten minutes. The participants were 23–30 years old, four males and four females. They were native speakers of Swedish and Mandarin Chinese. The participants did not know each other at the time of the recordings. They were invited to get acquainted and interact with one another, with no limitation on the topics. Anonymity was emphasized in the research project. The participants were asked for permission to use the pictures from the recordings for publications. All participants gave written consent for their participation in the research and the use of their data.

Fig. 1.
figure 1

Recording of the encounter.

Two participants (see, e.g., Fig. 1) were paired in a classroom and asked to interact with each other for about 10 min after the research assistant switched video-cameras on and left the room. The interactions were audio- and video-recorded by three video cameras (left-, center-, and right-position) with interlocutors standing face-to-face to capture the participant’s communicative movement as much as possible. The cameras were placed off the participants’ field of vision to minimize possible effects of video recording on participants’ behavior.

Each participant was involved in two mono-cultural dialogues, in which the gender of the other interlocutor varied. The first seven minutes of each dialogue were studied, because the shortest dialogue was around seven minutes long. Within the time given, the use of agreement expressions was compared between the Chinese and the Swedish speakers. An overview of the data is presented in Table 1.

Table 1. Overview of the recordings (C = Chinese, S = Swedish, f = female, and m = male).

4.2 Transcription, Coding, and Inter-coder Rating

The recordings were transcribed and annotated by native speakers of Swedish and Chinese, according to the Gothenburg Transcription Standard (GTS) version 6.4 (Nivre et al. 2004) and the MUMIN Coding Scheme for the Annotation of Feedback, Turn management and Sequencing Phenomena (Allwood et al. 2007). This allows inclusion of such features of spoken language as pauses and overlaps as well as comments on communicative body movements and other events. Feedback expressions through various modalities, their communicative act units, and their communicative functions were annotated. The following conventions from GTS and MUMIN were used in the paper (see Table 2).

Table 2. Transcription and MUMIN coding conventions.

Given the differences and connections between the perception modalities (e.g., visual modality, auditory modality, and haptic modality) and production modalities (e.g., gestural input is perceived through visual modality, vocal-verbal input through auditory modality, tactile input through haptic modality) (cf. Gander 2018), this paper focuses on the perception-sensory perspective. Thus, the researched agreement expressions are classified into three categories:

  1. 1.

    Vocal-verbal unimodal: verbal expressions only

  2. 2.

    Gestural unimodal: bodily communication, including head nods, facial displays, etc.

  3. 3.

    Vocal-verbal + gestural multimodal: combinations of (1) and (2)

The annotation was carried out by one annotator, and inter-coder reliability checking was performed independently by another annotator. The inter-coder agreement on the annotation of feedback and its communicative functions is moderate, with a free-marginal κ = 0.58.

By using the coding scheme (Allwood et al. 2007), the basic communicative function of feedback was coded as CPU (i.e., contact, perception and understanding). A code of CPU means ‘I want to and am able to continue the conversation and I perceive and understand what you have communicated.’ In addition, we also coded other communicative functions of feedback in our data, such as effective and epistemic stances which include emotion, attitude, agreement, acceptance, and evaluation etc. (Allwood et al. 2000; Paggio et al. 2017; Wessel-Tolvig and Paggio 2013; Gander 2018). There, agreement differs from acceptance in that agreement indicates acceptance, whereas acceptance does not indicate agreement. One can accept but may not necessarily agree. Acceptance is the necessary but not sufficient condition for agreement. In this paper, we focus specifically on agreement. CPU + agreement refers to that ‘I want to and am able to continue the conversation and I perceive and understand what you have communicated’ which is meant by CPU (only) plus ‘I agree with you on what you have said’.’

During the coding work, we identified CPU + confirmation too, and distinguished it from CPU + agreement, with a consideration of whether it is confirming a truth or fact or agreeing on an opinion. We do not want to simplify a truth or a fact as an opinion, and we do not want to simply take ‘confirmation’ as one type of ‘agreement’. Accordingly, we define the criterion of identifying CPU + agreement and CPU + confirmation as below.

The code of CPU + agreement occurs when the interlocutor is agreeing on some subjective opinions that the other interlocutor has expressed in the previous utterance which is not necessarily the truth or fact; while, CPU + confirmation is coded when the interlocutor is confirming what the other interlocutor has said is true. As has been stated and reasoned above, in this research we do not consider confirmation as one type of agreement. Instead, we take agreement to be based on shared views or opinions (not necessarily truth or fact), whereas confirmation to be associated with facts or truths.

4.3 Data Analysis

Occurrences of agreement feedback were identified and categorized according to modality (see Sect. 4.2, vocal-verbal unimodal, gestural unimodal, and vocal-verbal + gestural multimodal). Further, interactional sequences were qualitatively analyzed in order to examine the communicative functions of agreement feedback in the interactions. Additionally, quantitative analysis of frequencies and percentages of feedback occurrences were performed, comparing cultures (Swedish/Chinese) and genders. Because of the limited size of the samples, analysis of differences was made directly on the observed frequencies in the samples rather than being statistically tested, and generalizations were not attempted.

5 Results

The first research question aims to identify the agreement expressions in Swedish and Chinese. The second research question aims to explore the similarities and differences between the Swedish and the Chinese agreement expressions. Results are presented using raw frequencies and percentages of agreement feedback. Excerpts are extracted from the transcriptions.

5.1 General Overview of Agreement Feedback

This section presents a general overview of the types of agreement feedback in the data. Table 3 shows that the Swedish participants in the sample use more of both unimodal and multimodal agreement feedback than the Chinese participants (110 compared to 54 occurrences). Compared to the Chinese, the Swedes also use more vocal-verbal expressions of agreement (52 versus 17) and more multimodal feedback (41 compared to 27).

Table 3. Overview of agreement feedback used by the Swedish and Chinese participants (FB = Feedback, M = Male, F = Female).

Regarding gender, both the Chinese and the Swedish female speakers use more agreement expressions than male (81% versus 19% in the Chinese sample and 67% versus 33% in the Swedish sample), including all the three types of modalities of agreement expressions. In the Chinese sample, the share of gestural unimodal and multimodal feedback expressions is much higher in the females than in the males (90% and 89% compared to 10% and 11% respectively). In addition, Table 3 also shows that the Swedish females use more agreement expressions than the Chinese females (74 compared to 44), so as the Swedish males compared to the Chinese males (36 versus 10).

5.2 Swedish and Chinese Unimodal Vocal-Verbal Feedback Expressions for Agreement

Regarding unimodal agreement feedback expressions, we will look at the Swedish and the Chinese data respectively.

In the Swedish data, the identified Swedish unimodal vocal-verbal feedback expressions can be grouped into three categories: 1) yes and its variations, for example, ja (yes), {j}a (yeah), repetitions of {j}a (yeah); {j}a: precis (yes exactly); jo (sure); 2) m and its variations, for example, m, m: (lengthening); 3) no and its variations (e.g., when interlocutor agrees with preceding negative statement), such as no, no no (see Table 4). The Swedish participants show a tendency of expressing agreement using primarily {j}a and its variations, with an occurrence of 69% of all the unimodal vocal-verbal feedback expressions. The Swedish participants also use unimodal words no and m with their variations to express agreement, with frequencies of 17% and 14% respectively (see Table 4).

Table 4. Unimodal vocal-verbal feedback (VFB) agreement expressions used by the Swedish participants

The data presented in Table 4 also shows that the Swedish females use more vocal-verbal feedback expressions of yes and m and the variations than the Swedish males, with occurrences of 50% versus 19% and 12% versus 2%, respectively. In the meanwhile, the Swedish females and males show similar preferences in using no and its variations, with frequencies of 9% and 8%.

To present an example from our data, in Excerpt 1 below, Swedish female 2 expresses agreement while listening to Swedish female 1 about Swedish female 1’s experiences with home care. Swedish female 2 agrees with Swedish female 1 on that some people one meet at the workplace people can be nice while others can be avoided, by using unimodal vocal-verbal feedback expression {j}a.

figure a

Then, turning to the Chinese data, Table 5 presents the unimodal vocal-verbal feedback expressions used by the Chinese participants when expressing agreement.

Table 5. Unimodal vocal-verbal feedback (VFB) agreement expressions made by the Chinese participants.

Dui and its variation dui na shi (equivalent of yes and right in English) are the most common (82%) agreement expressions in the Chinese data. The Chinese female speakers give more agreement feedback than the Chinese males (65% compared to 35%).

Excerpt 2 below illustrates the Chinese female speaker 1 agreeing with the Chinese female 2 on that the work in China is faster tempo and more task oriented compared to that in Sweden:

figure b

Comparing the data of Table 4 and Table 5, one can observe that in both Swedish and Chinese conversations the vocal-verbal feedback yes and its variations are the most common unimodal agreement expressions. Also, in both Swedish and Chinese data, the female speakers give more feedback than the males. Besides, there is more variety in feedback expressions used in the Swedish sample compared to the Chinese data.

5.3 Swedish and Chinese Unimodal Gestural Feedback Expressions for Agreement

The Chinese and the Swedish unimodal gestural feedback expressions for agreement will be analyzed in this section. An overview of the data is presented in Table 6.

Table 6. Swedish and Chinese unimodal gestural feedback (GFB) agreement expressions

As can be seen from Table 6, both Swedish and Chinese speakers use unimodal gestural expressions relatively seldom, compared to how they use unimodal vocal-verbal expressions (presented in the previous section). Nod(s) are the most common unimodal feedback expression to communicate agreement in both Swedish and Chinese samples.

Excerpt 3 below illustrates the Swedish male speaker 2 agreeing with the Swedish male speaker 1 on that travelling is better to do after having completed education. A possible reason for just nodding (of Sm2) in this excerpt can be that the speaker (Sm2) does not want to interrupt an apparently lively monologue of the other speaker (Sm1).

figure c

Similarly, in the Chinese data, by nodding the female speaker 2 shows agreement on that if there is an opportunity to work then one should work rather than studying (see Excerpt 4).

figure d

5.4 Swedish and Chinese Multimodal Feedback Expressions for Agreement

In this section, the Swedish and the Chinese multimodal feedback expressions for agreement will be presented respectively. We will start with the Swedish data, followed by the Chinese data.

The Swedish multimodal feedback expressions are grouped in the same categories as the vocal-verbal unimodal expressions (see Sect. 5.2), namely, (1) yes and its variations, (2) no and its variations, and (3) m and its variations. Bodily movements used with each vocal-verbal expression comprise the corresponding multimodal unit. Table 7 presents an overview of the Swedish multimodal expressions identified in the data.

Table 7. Multimodal feedback (VFB + GFB) agreement expressions used by the Swedish speakers.

The most common multimodal feedback expressions for agreement employed by the Swedish participants are yes and its variations + nods as well as no and its variations + bodily movements including shakes, eyebrow movements and tilts (with frequencies of 70% and 20%).

An example is presented in Excerpt 5. The Swedish male speaker 1 expressed agreement with the Swedish male speaker 2 on randomly selected group work, by using the multimodal expression of yeah combined with head tilt.

figure e

Then, in the Chinese data, the Chinese multimodal feedback expressions for agreement can be grouped in two categories: 1) yes and its variations + bodily movement and 2) no and its variations + bodily movement (see Table 8).

Table 8. Multimodal feedback (VFB + GFB) agreement expressions used by the Chinese speakers

The majority of multimodal agreement expressions in the Chinese sample are contributed by the Chinese female speakers. As a variation of yes (see Table 8), the vocal-verbal component en (equivalent to yeah in English) combined with the gestural component nod(s) comprised the most common multimodal feedback expression for agreement, which is only used by the females. Chuckle, nod and smile gestural feedback expressions combined with the vocal-verbal feedback dui, shi or en (equivalent to yeah in English) comprise the most common multimodal agreement expressions.

Excerpt 6 presents an example of the Chinese female speaker 2 agreeing with the Chinese male speaker 1 on the open mindedness of their families regarding what they are allowed to do or not to do. The speaker Cf2 first provided a verbal feedback en (equivalent to yeah in English), followed by head nods and a lengthy pause which reinforced her vocal-verbal expression of agreement.

figure f

6 Discussion

Regarding modalities, the data has shown that both the Swedish and the Chinese speakers use vocal-verbal unimodal, gestural unimodal and multimodal expressions for expressing agreement. Unimodal gestural feedback expressions (e.g., head, shake, and smile) are less frequently used, compared to unimodal vocal-verbal and multimodal feedback expressions. This possibly indicates that vocal-verbal means are prioritized over the gestural means when people express agreement, and that simply nodding, for example, is not enough to ensure or reinforce agreeing in both cultures. Nevertheless, in both Swedish and Chinese samples, nodding is still the most common unimodal gestural expression of agreement, which is in line with previous research (Andonova and Taylor 2012). The findings also show that both Swedish and Chinese participants use head nod(s) in combination with yes of their own respective languages as the most common multimodal feedback expressions for showing agreement, which is consistent with earlier research (Lu and Allwood 2011). Also, the Swedish participants use more agreement feedback expressions both unimodally and multimodally than the Chinese participants. This is consistent with the prior studies which have noted the importance of consensus in Sweden (Havaleschka 2002) and the tendency that giving much feedback in interactions indicates listening and agreeing (Allwood et al. 2000; Tronnier and Allwood 2004) and avoiding conflict (Daun 2005; Pedersen 2010).

Besides the features of modalities for expressing agreement, both Swedish and Chinese female participants in this research use more agreement feedback than the males. This finding supports the research of Deng et al. (2016) on that females tend to have higher emotional expressivity and provide more feedback than males. In addition, although it has been found that both the Swedish and the Chinese female speakers express agreement more often than the males, this gender difference was bigger in the Chinese sample compared to the Swedish sample. This finding is in line with that of Azlina Abdul’s (2017) research, in which the Chinese females are found giving considerably more agreement feedback than the Chinese males. Further, as Sweden is considered to be a society in which gender roles overlap to a higher extent than in China, it might be reflected in more similarities between the Swedish male and female participants in expressing agreement (Hofstede 1994).

The data has also shown that the Chinese participants provide considerably fewer agreement feedback expressions than Swedes. It can potentially be explained by that in Sweden interactions are often informal in particular among students at a university context, due to a relatively low power distance (Hofstede 1994). On the contrary, in China, a differentiation is clearly made between polite/impolite and insider/outsider communication, the former often occurring between strangers while the latter between close acquaintances (Fang and Faure 2011). As Chinese culture is a high context culture, many Chinese people are cautious in communicating with ‘outsiders’ by, for example, limiting their contact to brief and functional communicative exchanges (Hall 1959), which might be reflected in the Chinese speakers in our research being less willing to express themselves compared to the Swedes.

Because these findings are based on a particular communication context of first encounters and between a limited number of participants, the present research can be seen as a case study for the future work with more participants and varied contexts to investigate agreement expressions, patterns and strategies in Swedish and Chinese or Western and Eastern cultures further.

7 Conclusion

This paper explored and compared how Chinese and Swedes express agreement in first-time encounter’s conversations by using social signals of communicative feedback. In this paper, eight audio- and video-recorded interactions between four Swedish and four Chinese participants were analyzed. The identified agreement feedback expressions were classified into three types according to which modalities are involved, namely, the vocal-verbal unimodal, the gestural unimodal, and the vocal-verbal + gestural multimodal expressions. Two research questions were investigated, regarding respectively what feedback expressions are used for expressing agreement in the Swedish and the Chinese conversations and what similarities and differences can be observed between the Swedish and the Chinese expressions of agreement.

About the feedback expressions used for expressing agreement, we have found that participants from both samples use both unimodal and multimodal feedback expressions. The Swedish participants most frequently use unimodal vocal-verbal feedback ja and nej (equivalent to yes and no in English respectively), unimodal gestural feedback, primarily head nod(s), multimodal feedback ja in combination with head nod(s) and up-nod(s). In the Chinese data, the most commonly used feedback is unimodal vocal-verbal expression dui and shi (equivalent to yes in English), followed by the unimodal gestural expression head nod(s), and multimodal expression dui and shi in combination with head nod(s) to express agreement.

Regarding the similarities between the Swedish and the Chinese samples, this case study suggests that both Swedish and Chinese speakers express agreement more multimodally than unimodally, most often with yes in combination with nod(s). The infrequent use of unimodal agreement expressions possibly indicates that simply nodding, for example, is not enough to reinforce agreeing. Also, in both samples the females showed more agreement than males. On the differences between the Swedish and the Chinese data, the Swedish participants expressed more agreement than the Chinese, which can be explained by the value the Swedes put on reaching consensus and avoiding conflict. Also, the gender difference between males and females in expressing agreement has been found bigger in the Chinese sample than the Swedish one. This might be because that Sweden is considered to be a society in which gender roles overlap to a higher extent, which could be reflected in more similarities between male and female swedes in social behaviors.

This case study suggests that the expression of agreement varies between cultures and genders. Both Swedish and Chinese speakers express agreement more multimodally than unimodally. Females express agreement more than males. Swedes express agreement more than Chinese. In spite of the limitations of single communication context and number of participants, this paper contributes to an increased understanding of Swedish and Chinese cultural similarities and differences in expressing agreement. The empirical findings can contribute to cross-cultural multimodal communication research and intercultural practices such as business negotiation and educational collaboration. It can also contribute to big data corpus and automatic analysis of agreement communication, which may help designing technology of autonomous speech and gesture recognition, and developing machine socio-cultural training.