Keywords

1 Introduction

Autism spectrum disorder (ASD) is a neurodevelopmental disorder more common in males. It is likely four times more prevalent than in females [1]. Because of this, the number of female ASD individuals included in the existing studies is insufficient, so that difficult to accurately evaluate the potential gender effect on symptom presentation. To date, we remain puzzled about the nature of gender differences in ASD, and thus, it is warranted to continue the research.

Gender differences are not only reflected in the diagnostic characteristics of ASD. Regarding social cognition, individuals with ASD generally have impaired attention to social stimuli, especially face stimuli, which has been well demonstrated by eye-tracking technology [2]. Individuals with ASD fixate less on the core features of the emotional face, especially the eye region [3, 4], leading researchers to speculate that abnormal visual processing may be a major reason for poor emotional discrimination in ASD, which in turn aggravates their difficulty in interpreting facial expressions. Existing research found that visual attention to faces is also modulated by gender in ASD: females focus more on faces in the social scene than males [5]. Also, the gender effects are found in typically developing (TD) individuals: females are more sensitive to emotional faces than males, and as age increases, they still perform better in emotional recognition [6, 7]. Therefore, it remains to be studied whether the gender differences in the visual processing of emotional faces are due to the pathology of ASD or to differences in social attention between males and females.

Based on the above points, to provide intuitional evidence, we employed eye-tracking technology, designed a dynamic facial emotion discrimination task, and recruited TD adults as participants. We compared and analyzed behavioral performance, visual fixation patterns, and the pupillary response of emotional faces between genders, exploring whether the visual processing of emotional faces is modulated by gender.

2 Method

2.1 Participants

Fifty-one TD university students participated in our study, of whom three participants were excluded due to poor data quality, and finally, 48 participants were included in further analysis. All participants have normal or corrected-to-normal vision. Of the participants, 24 were male, ranging in age from 19 to 29 years (M = 22.65, SD = 2.27), and 24 were female, ranging in age from 19 to 28 years (M = 22.74, SD = 2.29). No significant difference in age was found between males and females by the two-sample t-test. This research was approved by the Ethics Committee of Tianjin Children’s Hospital, and all participants provided written informed consent.

2.2 Stimuli and Experimental Paradigm

Face stimuli were chosen from the Karolinska Directed Emotional Faces (KDEF) database [8], containing the emotional facial photo of 10 male and 10 female actors.

Fig. 1.
figure 1

Examples of face stimuli presentation

The stimuli presented four emotions: neutral, happy, sad, and angry, and included three angles: 0° (left or right), 45° (left or right), and 90°. All photos were the same size (width: 17° visual angle, height: 23° visual angle), and the critical face features were standardized to the same location. We designed two orders of stimuli presentation, as shown in Fig. 1. Among them, order 1 presents face photos from angle 0° to 90° to 0°, and order 2 presents from 90° to 0° to 90°. This dynamic stimulus presentation imitated the pattern of viewing faces in daily life, improving the ecological validity of the experiment.

2.3 Procedure

The experiment was conducted in a quiet, soft lightroom. Participants were guided to sit approximately 60cm away from a 27-inch monitor comfortably, which resolution is 1920*1080 pixels. Eye movement data were recorded by Tobii Pro Spectrum eye tracker; the sampling rate is 1200 Hz. Before the experiment, participants conducted a 5-point calibration with the Tobii Pro Lab software. After participants were fully aware of the emotion-discrimination task, the experiment was started by the experimenter pressing the space key.

The experiment consisted of 5 blocks, a total of 80 trials. A cartoon character was first presented in each trial to attract participants’ visual attention to the centre of the screen, lasting for 3s. Then five photos of the same actor, the same emotion, but different angles of the faces were presented in a specific order (as shown in Fig. 1), each lasting for 2s, a total of 10s. The participants were required to discriminate the emotion shown by the actor within 10s of the presentation and give feedback with the mouse by clicking the corresponding icon on the screen at the moment of recognition. The response time and whether correct were recorded, and the participants themselves would not get the result. Across the 80 trials, the two orders of presentation and the four emotional conditions were balanced in number and presented in random order. The eye movement data were recorded throughout the whole process for each block.

2.4 Data Analysis

Data were exported for offline analysis with MATLAB. For behavioral data, the average accuracy and average response time (in the case of correct discrimination) of 48 participants were calculated in two orders of presentation and four emotional conditions, respectively.

For the eye movement data, we calculated the preference index and pupil diameter variation as features. First, taking the nose as the boundary, we divided the face stimuli into eye and mouth ROI. Then we calculated the proportional fixation time for each ROI. And we subtracted the proportional fixation time of the mouth from that of the eye, which is the preference index. If positive, the participant is fixated more on the eyes; if negative, the participant is fixated more on the mouth. For the pupil diameter variation, we filtrated the pupil data during the fixation of the eye ROI. Subsequently, the pupil diameter of the first 3s was used as the baseline, and the baseline was subtracted from the diameter of the facial photos, obtaining the pupil diameter variation. We averaged the two features according to the corresponding presented order and emotional conditions.

In this article, we used behavioral characteristics and eye-movement characteristics as dependent variables, gender as between-group variables, order (1 and 2), and emotion conditions (neutral, happy, sad, angry) as within-group variables. As the dependent variables did not follow the normal distribution, statistical analysis was performed using the generalized estimating equation (GEE) with a significance level of 0.05, adjusting for multiple comparisons using the Bonferroni method.

3 Results

3.1 Behavioral Results

For accuracy, after statistical testing, only the main effect of emotion was significant (p < 0.001). There was no significant difference in accuracy between males (M = 93.45, SD = 3.89) and females (M = 93.55, SD = 3.79). Figure 2 shows that the happy condition has the highest accuracy (M = 98.07, SD = 3.88), followed by the neutral condition (M = 96.61, SD = 4.20). Angry was third (M = 91.92, SD = 7.67) and significantly less accurate than happy (p < 0.001) and neutral (p = 0.001) conditions. The sad condition (M = 87.06, SD = 8.95) was the lowest and significantly lower than happy (p < 0.001), neutral (p < 0.001), and angry (p = 0.001) conditions.

The response time is influenced by order, emotion, and gender, as shown in Fig. 3. The statistical analysis showed that the main effect of order (p < 0.001), emotion.

Fig. 2.
figure 2

Comparison of response accuracy under four emotion conditions (*** p < 0.001)

Fig. 3.
figure 3

Comparison of male and female response times under different presented order and emotion conditions (* p < 0.05)

(p < 0.001), gender (p = 0.034), and the interaction effect of order*emotion (p < 0.001) were significant. Further analysis found that males took more time to discriminate emotions than females, regardless of the stimuli. While for emotion, we found that when the stimuli were presented in order 1, the response time of happy (M = 1.57, SD = 0.42) was the lowest, followed by neutral (M = 2.00, SD = 0.97), angry (M = 2.26, SD = 0.80), and sad (M = 2.51, SD = 0.74) was the highest. There were significant differences between each two of these four emotions. When the stimuli were presented in order 2, the happy condition (M = 2.19, SD = 0.82) was the lowest, which was significantly lower than the neutral (p < 0.001), sad (p < 0.001), and angry (p < 0.001). But there were no significant differences between the remaining three emotions (neutral: M = 3.47, SD = 1.23; sad: M = 3.42, SD = 1.05; angry: M = 3.47, SD = 1.08).

Unexpectedly, we found that response times were significantly lower in order 1 than in order 2, regardless of emotions. This may be due to the 0° faces presented first, providing more emotional information and helping participants quickly discriminate.

Overall, males and females have similar discriminated accuracy for the task, but males spent more time. In addition, they both discriminate happy emotions accurately and quickly, while two negative emotions, especially sad, were discriminated less accurately and slower.

3.2 Preference Index Results

Fig. 4.
figure 4

Gender contrast of the preference index under different emotion conditions (*** p < 0.001, ** p < 0.01, * p < 0.05)

To explore the fixation patterns of males and females to emotional faces, we calculated the preference index to reflect the difference in fixation preference between the two groups, and the results are shown in Fig. 4. Statistical analysis showed that the main effects of gender (p = 0.002), order (p < 0.001), and emotion (p < 0.001) were all significant, and there was a significant interaction effect of gender*emotion (p = 0.006). From Fig. 4, we can clearly see that the preference index of males was significantly lower than that of females (neutral: p = 0.001; happy: p = 0.001; sad: p = 0.002; angry: p = 0.012). This result indicates that during visual scanning of emotional faces, males tend to focus on the mouth region and females tend to focus on the eye region. Both males and females tended to focus more on the mouth region for happy faces and more on the eye region for negative emotional faces. In addition, we were surprised to find out that the preference index in order 1 was significantly higher (p < 0.001), suggesting that viewing the 0° face first may increase subsequent attention to the eye region.

3.3 Pupil Diameter Variation Results

Fig. 5.
figure 5

Comparison of male and female pupil diameter variation under different presented order and emotion conditions (** p < 0.01, * p < 0.05)

Pupil diameter can reflect participants’ physiological responses to emotional faces. To further investigate gender differences, we calculated the pupil diameter variation of the eye region, as shown in Fig. 5. Statistical analysis found that the main effect of gender (p = 0.026) and emotion (p = 0.021) was significant, and the interaction effect of order*emotion*gender (p = 0.003) was significant. The further specific analysis found that the pupil variation of males was less than that of females. In order 1, the male pupil diameter variation in the eye region of neutral and sad faces was significantly lower than that of females (neutral: p = 0.017; sad: p = 0.019). And in order 2, the male pupil diameter variation in the eye region of happy and angry faces was significantly lower than that of females (happy: p = 0.006; angry: p = 0.026). Overall, females responded greater to emotional faces than males, suggesting more sensitivity to the emotional information conveyed by eyes.

4 Discussion

This study verifies the behavioral and visual performance differences when TD males and females in the facial emotion discrimination task. Specifically, we find that males required more time to achieve similar discrimination accuracy as females. Further, males were more preferred to focus on the mouth region and showed a weaker pupillary response to the eye region. On the other hand, females are more preferred to view the eye region and respond greater.

Attention to the eye region shows salient importance to face processing. In our study, females paid more attention to the eyes than males, suggesting that females are more efficient in capturing facial information. According to neuroimaging studies, in the facial perception stage, females have greater limbic system activation, facilitating primary emotion processing. But males’ emotion processing style tends to be analytical, which may potentially be slower in perception [9]. Meanwhile, the limbic system structurally overlaps with the central autonomic network (CAN), which can modulate the pupillary response through the noradrenergic sympathetic pathway [10]. Therefore, females may experience emotions with greater intensity when viewing emotional faces, have higher emotional arousal, and eventually show larger pupil dilation.

Overall, we found apparent gender effects on visual processing strategies for emotional faces and the sensitivity to emotional information conveyed by the eyes. In the subsequent studies, we will include clinical data to verify whether the gender effect can be extended in the ASD population.

5 Conclusions

The current work focused on gender effects on emotional face processing. In general, we have three main findings: (A). Males required more time to achieve similar emotional discrimination accuracy as females. (B). Males preferred to fixate on the mouth region during processing of emotional faces, while females preferred the eye region. (C). Females had greater pupillary responses to eye region than males. These findings prove that TD females have stronger social attention for emotional faces and are more sensitive to emotional information conveyed by the eyes. In the future, we will verify these results in the ASD group and make a comparison between the ASD and TD groups.