Buscar en
International Journal of Clinical and Health Psychology
Toda la web
Inicio International Journal of Clinical and Health Psychology Diminished emotion recognition with reduced face gaze in complex situation in in...
Journal Information
Vol. 23. Issue 4.
(October - December 2023)
Share
Share
Download PDF
More article options
Visits
469
Vol. 23. Issue 4.
(October - December 2023)
Full text access
Diminished emotion recognition with reduced face gaze in complex situation in individuals with broad autism phenotype
Disminución del reconocimiento de emociones en situaciones complejas debido a la reducción de la mirada a la cara en individuos con fenotipo de autismo amplio
Visits
469
Woo-Jin Chaa,1, Kiho Kimb,
Corresponding author
kihohere@sjcu.ac.kr

Corresponding author.
a Department of Psychology, Chung-Ang University, 84, Heukseok-ro, Dongjak-gu, Seoul 06974, Republic of Korea
b Department of Psychology of Counseling, Sejong Cyber University, Cheonho-daero 680, Gwangjingu, Seoul 04992, Republic of Korea
This item has received
Article information
Abstract
Full Text
Bibliography
Download PDF
Statistics
Figures (3)
Show moreShow less
Tables (3)
Table 1. Demographic and clinical characteristic for each group.
Table 2. Reaction time and accuracy under condition where facial stimuli were presented with mild emotional intensity.
Table 3. Gaze proportions under condition where facial stimuli were presented with mild emotional intensity.
Show moreShow less
Additional material (1)
Abstract
Background/objective

Individuals with broad autism phenotype (BAP) showed a diminished ability to recognize emotion. This study aims to examine whether their decline in emotion recognition ability could be more clearly identified as task complexity increased and whether their decline could be influenced by their eye-gaze patterns.

Method

41 individuals with BAP and 40 healthy controls performed two types of emotion recognition tasks. After confirming conditions wherein the BAP group did not perform well compared to the control group, we compared gaze proportion on faces and context between groups when performing the conditions.

Results

The more difficult the task, the clearer the significant relationships between the level of autistic traits and emotion recognition ability. The BAP group showed lower accuracy compared to the control group when a face with mild emotional intensity was presented with context. In terms of gaze proportion, the BAP group looked less at faces when recognizing emotions compared to the control group.

Conclusion

These findings indicate that diminished emotion recognition ability in individuals with BAP may be influenced by face gaze.

Keywords:
Broad autism phenotype
Autism spectrum disorder
Emotion recognition
Weak central coherence theory
Eye-tracking
Full Text
Introduction

The broad autism phenotype (BAP) has been defined as a range of traits qualitatively similar to autism spectrum disorder (ASD) but with a mild level of symptoms. While individuals with BAP experience higher clinical distress compared to the general population, they do not show clinically severe symptoms that cause social and occupational impairment. These individuals experience significantly higher levels of depression and anxiety compared to the general population and these symptoms could have been caused by their reduced social interaction and social skills (Kulasinghe et al., 2021). Previous studies have focused on the defects in the social interaction of the BAP group. Relationships between defects in social interaction and BAP level were reported to be mediated by social cognition, particularly facial emotion recognition (Sasson et al., 2013). Compared to the general population, the BAP group has a lower ability to recognize facial expressions (Stewart et al., 2020). As people reveal their emotions and intentions through their faces primarily, recognizing emotions through others’ faces in social interaction is critical (Costa-Cordella, 2021).

In individuals with BAP, their diminished ability to recognize emotion could be caused by their unique cognitive patterns. On emotion recognition, facial expressions rarely appear without complex environments. In complex situations, people recognize emotions correctly by considering faces and contexts simultaneously (Le et al., 2022). However, individuals with BAP may find considering faces and contexts simultaneously difficult because of their unique cognitive patterns, as described by weak central coherence theory (WCCT). The WCCT indicates that individuals with ASD have cognitive patterns that may prioritize local and detailed parts, as opposed to the general population's cognitive patterns, which aim to integrate multiple components (Happe & Frith, 2006). The cognitive pattern related to the WCCT was also found in the BAP group (Nayar et al., 2022). Moreover, during social interaction, individuals in the BAP and ASD groups tend to mainly look at hands or other areas instead of faces. These characteristics result in reduced effectiveness to detect a target's emotion (Rigby et al., 2016).

While previous studies have reported that the BAP group showed a diminished ability to recognize other people's emotions, the diminished ability does not appear severe enough to directly cause clinical impairment. Most studies suggest that the BAP group has a diminished ability to recognize emotions, but some have reported that there is no difference in recognizing emotion between the BAP group and the general population (Pisula & Ziegart-Sadowska, 2015). Combining these mixed results, a diminished ability to recognize emotions in the BAP group could be identified in complex conditions. For example, when a BAP group was asked to match the emotion of the facial stimuli with the presented context or match the face with ambiguous emotions (e.g., low emotional intensity), the group showed a significantly diminished ability to recognize emotion (Cha & Lee, 2022; Stewart et al., 2020). However, when a BAP group was asked to identify a face's emotion presented alone or to match the basic emotion including happiness, they exhibited a similar emotion recognition ability to the general population (Holt et al., 2014; Soto-Icaza et al., 2022). In summary, the more difficult the measurement of emotion recognition ability, the clearer the decrease in the emotion recognition ability of the BAP.

First, this study aims to investigate whether the diminished emotion recognition ability of the BAP group can be identified more clearly as the task becomes more difficult. Hence, we investigate the conditions wherein the BAP group shows a diminished ability to recognize facial expressions. To diversify the conditions of emotion recognition tasks, we divide two types of emotion recognition tasks into three conditions. Moreover, we also use four types of emotions (i.e., fear, anger, sadness, happiness) and two types of emotional intensity (i.e., full-blown, mild). Second, this study aims to identify whether their diminished emotion recognition ability could be influenced by their unique cognitive patterns measured as eye-gaze patterns. Hence, we use an eye tracker to determine the gaze duration on the face during the task wherein the BAP group showed a decline in emotion recognition ability. This study presents the following hypotheses: (1) the BAP group would show a diminished ability to recognize emotions in faces compared to the general population as the task becomes more difficult and (2) the BAP group would look less at faces when they recognize emotions in tasks where they exhibit diminished ability.

MethodParticipants

In this study, we recruited 512 participants from several universities in Seoul, Korea, through advertisements in online communities and internet bulletin boards. To screen for individuals with BAP, participants were asked to complete the Korean Version of the Broad Autism Phenotype Questionnaire (BAPQ) and Autism-Spectrum Quotient (AQ) (Kim & Kim, 2022; Ko et al., 2018). Subjects with scores of 3.15 or higher and 23 or higher on the BAPQ and AQ, respectively, were classified under the BAP group. Individuals with low scores on the BAPQ and AQ were preferentially selected as the control group. We used the following exclusion criteria for the present study: (1) diagnosis of other psychiatric disorders and (2) participation in any other pharmacological treatment. Finally, we sorted 81 participants into the following two groups: (a) the BAP group (n = 41) and (b) control group (n = 40).

Questionnaires and measurementBroad autism phenotype questionnaire (BAPQ)

The BAPQ is a 36-item self-reported scale developed to assess autistic traits (Hurley et al., 2007; Kim & Kim, 2022). BAPQ provides quantitative information associated with the core symptoms of ASD: aloof, pragmatic language, and rigid. We calculated the total score by averaging all 36 items. Each domain's scores were calculated by averaging the 12 items for each subscale. A higher BAPQ score corresponds to a higher level of autistic traits. Cronbach's α was 0.89 and 0.94 in the validation and present studies, respectively.

Autism-spectrum quotient (AQ)

The AQ was developed to assess the five domains of autistic traits: social skills, communication, imagination, attention to detail, and attention switching (Baron-Cohen et al., 2001; Ko et al., 2018). The AQ is a 50-item self-report questionnaire. Scores above 32 indicate clinically significant levels of autistic traits in the original scale. Scores above 23 could be used as an effective cut-off point for accurately classifying the maximum number of individuals with BAP in the Korean version of the AQ. Cronbach's α was 0.85 and 0.88 in the validation and present studies, respectively.

Beck depression inventory-second edition (BDI-II)

The BDI-II was developed to assess the severity of cognitive and physical symptoms resulting from depression in the 2 weeks prior to testing (Beck et al., 1996; Lee et al., 2017). BDI-II is composed of 21 items. A higher BDI-II score indicates a higher level of depression. Cronbach's α was 0.89 and 0.91 in the validation and present studies, respectively.

Beck anxiety inventory (BAI)

The BAI was developed to measure physical and cognitive symptoms of anxiety in the week prior to testing (Beck et al., 1988; Yook & Kim, 1997). BAI is composed of 21 items. A higher BAI score indicates a higher anxiety level. Cronbach's α was 0.91 and 0.88 in the validation and present studies, respectively.

Behavioral taskPhase 1: emotion recognition task

The emotion recognition task (ERT) comprised eight colored facial stimuli from the “Extended ChaeLee Korean facial expressions of emotions: ChaeLee-E” (Lee et al., 2013). Facial stimuli were classified according to four emotions (i.e., fear, anger, sadness, happiness) and two emotional intensities (i.e., full-blown, mild). We created facial stimuli with mild intensity by morphing the neutral and full-blown expressions using computer software (Sqirlz Morph: www.xiberpix.com). Through the program, we generated 51 images with intensity levels ranging from neutral to full-blown emotion with 2% increment steps (Supplementary Fig. 1A). Among the 51 images, we defined facial stimulus with a 50% intensity level as mild. Facial stimuli at a size of 4.7 × 5.7 cm were placed at the center of a white background of 25.4 × 16.9 cm (Supplementary Fig. 1B). The question “What is the emotion of the face in the image?” was presented for 5 s before facial stimuli were presented. We then required participants to respond as quickly and accurately as possible to the question. To select fear, anger, sadness, or happiness, participants were to type 1, 3, 5, or 7, respectively, on the keyboard. After participants pushed one of the four buttons, the trial moved on to the next stimulus. If participants did not select a facial emotion for 3 s, the trial was treated as an incorrect answer and then moved to the next. Between each trial was a 2-s interval. This task comprised four blocks, with each block composed of eight trials. We calculated accuracy as the percentage of correct answers in the entire trial. We evaluated reaction time as the period from the time the stimulus was presented to the time the participants responded.

Phase 2: emotions in context task

The emotions in context task (ECT) comprised 32 colored facial stimuli within the emotional contexts (Sasson et al., 2016). Facial stimuli were classified according to four emotions (i.e., fear, anger, sadness, happiness), two emotional intensities (i.e., full-blown, mild), and four emotional contexts (i.e., fear, anger, sadness, happiness). Pictures of emotional context were taken from “CanStockPhoto” (https://www.canstockphoto.co.kr). We replaced face parts in the original pictures with the eight facial stimuli used in Phase 1 (four emotions and two intensities). To validate the types and intensities of the emotional context pictures, 15 graduate students rated the relevance, arousal, and valence of the emotions using a 7-point Likert scale (0 = Not at all; 6 = Extremely). We commissioned a professional editor to integrate each of the eight facial stimuli with four emotional contexts naturally. Facial stimuli were presented at a size of 4.7 × 5.7 cm in the emotional context of 25.4 × 16.9 cm (Supplementary Fig. 1C). We divided ECT into two conditions according to congruence: congruent and incongruent. We defined the congruent condition of ECT as matching the emotion of the facial stimulus with the emotion of the context (e.g., a happy face in a happy context). A mismatch between the emotion of the facial stimulus and of the context (e.g., a happy face in an angry context) would be an incongruent condition. We asked participants to respond as quickly and accurately as possible to the emotion of the facial stimulus in the presented image. The selection had to be made by considering only the emotion of the facial stimulus without considering that the context. To select fear, anger, sadness, or happiness, participants were to type answers as 1, 3, 5, and 7 on the keyboard. If participants did not select a facial emotion for 3 s, the trial was treated as an incorrect answer and then moved to the next. Between each trial was a 2-s interval. This task comprised two blocks and each block comprised 32 trials. Accuracy was calculated as the percentage of correct answers in the entire trial. We evaluated reaction time as the period from the time the stimulus was presented to the time participants responded. We measured accuracy and reaction time by dividing congruent and incongruent conditions.

Apparatus

Detailed information on the apparatus was presented in supplementary material.

Procedure

We performed this study while adhering to the principles of the Declaration of Helsinki. All participants voluntarily participated in the experiment and signed a consent form approved by the Institutional Review Board of Chung-Ang University (NO. 1,041,078–202,108-HRSB- 273–01). Detailed information on the procedure was presented in supplementary material.

Data analysis

The required sample size was calculated using G*power 3.1.9.7 (Faul et al., 2007), that estimated a sample size of 54 participants as adequate for a design with mixed-design analysis of variance (ANOVA), an alpha error probability of 0.05, a power of 0.95, and a medium effect size (ηp2= 0.25).

In our data analysis, we performed an independent t-test and chi-square test to analyze differences in group characteristics between the BAP and control groups. To investigate the relationship between performance ability (reaction time and accuracy) and level of autistic traits (BAPQ and AQ), we performed a correlation analysis (Pearson's r). To examine the difference in emotion recognition ability between BAP and control groups, we conducted 2 (group: BAP, control) × 2 (intensity of emotion: full-blown, mild) × 4 (emotion of target: fear, anger, sadness, happiness) mixed-design analysis of covariance (ANCOVA) to analyze the data in each task. Accuracy or reaction time was used as dependent variable, and BDI-II and BAI scores as covariates. When the ANCOVA showed a significant main effect for emotion of target with 4 levels, Tukey's HSD test was conducted for a post-hoc test. When analyzing ECT results, we analyzed the congruent and incongruent conditions separately.

After confirming the diminished ability to recognize facial emotions in the BAP group, we compared gaze proportions of ROIs between both groups to infer cause of the diminished ability in the BAP. The gaze proportion of face, eyes, mouth, or context was the dependent variable, and the BDI-II and BAI scores were the covariates. Gaze proportion of ROI was calculated as the percentage of duration spent gazing the ROI out of the total gazing duration. The ROIs of context were defined as the parts with valence in the background (Supplementary Fig. 2). We investigated differences in gaze proportion of the ROIs including face, eyes, mouth, and context between groups under conditions that the BAP group showed decreased ability to recognize emotions. All analyses were performed using R Statistical Software (v4.1.2; R Core Team 2021).

Of all participants, nine were removed from the analysis through the quality check: nine participants’ error rate was more than 2 SD from the average. Ultimately, we selected 37 BAP and 35 control groups for the final analysis.

ResultsGroup characteristics

Table 1 shows the demographic and clinical characteristics of participants acquired from the results of the t-test and the chi-square test.

Table 1.

Demographic and clinical characteristic for each group.

MeasureBAP (n = 37)Control (n = 35)Test Statistics
t / χ2  p 
Age (years)22.189 (2.885)  22.629 (2.224)  −0.726  0.470 
Sex (male/female)12 / 25  15 / 20  0.448  0.503 
BAPQ3.678 (0.320)  2.451 (0.363)  15.179  < 0.001 
  Aloof  3.829 (0.492)  2.451 (0.480)  11.944  < 0.001 
  Pragmatic language  3.370 (0.414)  2.343 (0.470)  9.757  < 0.001 
  Rigid  3.718 (0.534)  2.590 (0.436)  9.759  < 0.001 
AQ30.405 (4.885)  14.143 (3.353)  16.546  < 0.001 
  Social skill  7.194 (1.600)  2.257 (1.379)  13.937  < 0.001 
  Communication  5.444 (1.715)  0.857 (0.944)  14.017  < 0.001 
  Imagination  4.472 (2.035)  1.971 (1.403)  6.042  < 0.001 
  Attention for detail  5.167 (2.490)  4.486 (2.293)  1.1992  0.235 
  Attention Switching  7.667 (1.454)  4.571 (1.703)  8.227  < 0.001 
BDI-II16.162 (9.060)  9.000 (7.276)  3.708  < 0.001 
BAI10.027 (6.300)  6.600 (6.713)  2.231  0.029 

Note. Mean (standard deviation); BAP: Broad Autism Phenotype Group, BAPQ: Broad Autism Phenotype Questionnaire; AQ: Autism Spectrum Quotient; BDI-II: Beck Depression Inventory-II; BAI: Beck Anxiety Inventory.

We found significant differences in average BAPQ and total AQ scores between groups. Hence, groups were properly divided according to their autistic traits. No significant difference was found in age and sex. Significant differences in BDI-II and BAI were also found. These variables could affect the dependent variables such as reaction time or accuracy. We analyzed these after adding BDI-II and BAI scores as covariates in each analysis.

Correlation analysis

We conducted a Pearson correlation analysis to examine the relationship between level of autistic traits and performance ability (Fig. 1; Supplementary Table 1).

Fig. 1.

Correlation heatmap between ability to recognize emotions and autistic traits. P1: Phase 1 (emotion recognition task); P2: Phase 2 (emotions in context task); con: congruent condition; incon: incongruent condition. *p < .05. **p < .01.

(0.22MB).

No significant correlation between the level of autistic traits and reaction time or accuracy was confirmed in Phase 1.

We found significant positive correlations between some subscales of autistic traits and reaction time when examining the relationship in the congruent condition of Phase 2. Hence, the higher level of autistic traits, the slower the emotion recognition speed. However, no significant correlation was found between the level of autistic traits and accuracy in the congruent condition of Phase 2.

We found significant positive correlations between some subscales of autistic traits and reaction time when investigating the relationship in an incongruent condition of Phase 2. Significant negative correlations were also found between some subscales of autistic traits and accuracy in the incongruent condition of Phase 2. These results indicated that the higher the level of autistic traits, the slower the emotion recognition speed, and the lower the accuracy of emotion recognition.

In summary, the level of autistic traits can be most significantly correlated with emotion recognition ability when performing the most difficult task wherein the emotion of the facial stimulus and emotion of the context do not match.

Comparing performance ability between groups in each taskPhase 1: emotion recognition task

To examine reaction time and accuracy of emotion recognition, we performed a 2 × 2 × 4 mixed-design ANCOVA on reaction time and accuracy (Supplementary Table 2).

When examining reaction time, we found a significant main effect of intensity. Hence, stimuli with mild emotional intensity were recognized significantly slower than the stimuli with full-blown emotional intensity. A significant main effect of emotion was identified. The happy emotion was recognized significantly faster than the sad emotion (Supplementary Table 3).

When investigating the accuracy, a significant interaction effect between intensity × emotion was found. When post hoc tests were performed, stimuli with full-blown emotional intensity were more accurately recognized than stimuli with mild emotional intensity when recognizing the emotions happy, sad, and fearful. However, no difference between intensities was found when recognizing the angry emotion (Supplementary Table 4). We identified a significant main effect of emotion. However, no significant difference was found between emotions when post hoc tests was conducted (Supplementary Table 3).

In summary, when performing Phase 1 (ERT), no difference in emotion recognition ability between groups was found. Hence, when performing a simple task where facial stimuli are presented alone, the BAP group can recognize emotions as accurately and quickly as the general population.

Phase 2: emotions in context task (Congruent condition)

To examine the reaction time and accuracy of emotion recognition, we performed a 2 × 2 × 4 mixed-design ANCOVA on reaction time and accuracy (Supplementary Table 2).

When examining the reaction time, we confirmed a significant main effect of intensity. Hence, stimuli with mild emotional intensity were recognized slower than stimuli with full-blown emotional intensity. Moreover, a significant main effect of emotion was found. Happy emotions were recognized faster than other emotions, and angry emotions were recognized faster than sad emotions (Supplementary Table 3).

When investigating the accuracy, we confirmed a significant two-way interaction effect between group × intensity. To examine the interaction effect in terms of our study objectives, we analyzed the main effects of the group in each intensity. While we identified no difference between groups in full-blown emotional intensity [F (1, 68) = 0.067, p = .797], we confirmed a significant difference between groups in mild emotional intensity [F (1, 68) = 6.230, p = .015, ηp2 = 0.084] (Fig. 2; Table 2). Hence, while the BAP group could recognize emotions with high intensity accurately as the control group, the BAP group were less accurate in recognizing the emotions than the control group wherein facial expressions are represented with mild intensity. We also confirmed a significant interaction effect between intensity × emotion, indicating that stimuli with full-blown emotional intensity were more accurately recognized than stimuli with mild emotional intensity when recognizing the emotions happy, sad, and fearful. However, no difference between both intensities when recognizing angry emotion was found (Supplementary Table 4). A significant main effect of intensity was identified, indicating that emotions with mild intensity were recognized less accurately than emotions with full-blown intensity. A significant main effect of emotion was found. However, when post hoc tests were conducted, no significant difference between emotions were found (Supplementary Table 3).

Fig. 2.

Comparison of accuracy between groups under the condition of intensity of emotion in congruent and incongruent conditions of ECT. BAP, broad autism phenotype group; Control, control group. *p < .05. **p < .01. ***p < .001.

(0.13MB).
Table 2.

Reaction time and accuracy under condition where facial stimuli were presented with mild emotional intensity.

Dependent variableBAP (n = 37)Control (n = 35)Test Statistics
F  p  ηp2 
Phase2: Congruent condition
Reaction Time  1.520 (0.467)  1.381 (0.434)  0.213  0.646  0.001 
Accuracy  0.814 (0.310)  0.911 (0.210)  6.230  0.015  0.084 
Phase2: Incongruent condition
Reaction Time  1.579 (0.440)  1.407 (0.364)  0.651  0.423  0.011 
Accuracy  0.693 (0.317)  0.801 (0.254)  5.009  0.028  0.069 

Note. BAP: Broad Autism Phenotype Group; Bold value indicates that the value is statistically significant.

In summary, when performing the congruent condition of Phase 2 (ECT), the BAP group could not recognize the emotions with mild intensity accurately compared to the control group.

Phase 2: emotions in context task (Incongruent condition)

To examine reaction time and accuracy of recognition of facial expressions, we performed a 2 × 2 × 4 mixed-design ANCOVA on reaction time and accuracy (Supplementary Table 2).

When examining reaction time, the main effect of emotion was significant. Fearful emotions were recognized slower than other emotions, and sad emotions were recognized slower than happy emotions (Supplementary Table 3).

When investigating the accuracy, between group × intensity had a significant two-way interaction effect. To examine the interaction effect in terms of the purpose of our study, we analyzed the main effects of the group in each intensity. While no difference between BAP and control groups in faces with high intensity was found [F (1, 68) = 0.380, p = .539], a significant main effect between groups in faces with mild intensity was confirmed [F (1, 68) = 5.009, p = .028, ηp2 = 0.069] (Fig. 2; Table 2). Basically, while the BAP group could recognize full-blown emotions as accurately as the control group, they could not recognize mild intensity emotions as accurately as the control group. The interaction effect between intensity × emotion was also significant, indicating that emotions with full-blown intensity were more accurately recognized than emotions with mild intensity when recognizing happy, sad, and fearful emotions. However, no difference between both intensities was found when recognizing angry emotions (Supplementary Table 4). The main effect of intensity was significant. Hence, we recognized stimuli with mild emotional intensity less accurately compared to stimuli with full-blown emotional intensity. The main effect of emotion was also significant, indicating that fearful emotion was recognized less accurately than other emotions (Supplementary Table 3).

In summary, when performing the incongruent condition of Phase 2 (ECT), the BAP group could not recognize the emotion with mild intensity accurately compared to the control group.

Therefore, when performing a complex task where facial stimuli with mild emotional intensity were presented with a context, the BAP group were less accurate in recognizing the emotions.

Phase 2: emotions in context task (Congruent & incongruent condition)

From the previous analyses, we confirmed that the BAP group was less accurate than the control group in recognizing emotions with mild intensity in both congruent and incongruent conditions in phase 2. To investigate whether these results could vary by congruence, we performed a 2 (group; BAP, control) × 2 (congruence; congruent, incongruent) mixed-design ANCOVA on accuracy in phase 2.

There was no interaction effect of group × congruence [F (1, 68) = 0.002, p = .965]. However, there were significant main effects of group [F (1, 68) = 8.195, p = .006, ηp2 = 0.076] and congruence [F (1, 68) = 5.617, p = .021, ηp2 = 0.026]. These indicate that the BAP group couldn't recognize facial stimuli with mild emotional intensity compared to the control group, regardless of emotional congruence between target and context, and both groups could recognize better when the targets were presented in congruent context than incongruent context.

Investigating the difference of gaze proportion between groups

We found that the BAP group were less accurate in recognizing the emotions than the control group when the emotions are represented with mild intensity in the congruent and incongruent conditions of ECT. To infer the cause of this discrepancy, we examined the differences in gaze proportion of the face, eyes, mouth, and context between groups when emotions are represented with mild intensity in the ECT (Table 3).

Table 3.

Gaze proportions under condition where facial stimuli were presented with mild emotional intensity.

Dependent variableBAP (n = 37)Control (n = 35)Test Statistics
F  p  ηp2 
Phase2: Congruent condition
Gaze proportion of face  0.786 (0.163)  0.887 (0.045)  8.532  0.005  0.146 
Gaze proportion of eyes  0.255 (0.208)  0.325 (0.224)  1.795  0.186  0.035 
Gaze proportion of mouth  0.203 (0.162)  0.221 (0.234)  0.109  0.743  0.002 
Gaze proportion of context  0.049 (0.043)  0.037 (0.029)  2.161  0.148  0.041 
Phase2: Incongruent condition
Gaze proportion of face  0.796 (0.124)  0.791 (0.157)  0.670  0.416  0.011 
Gaze proportion of eyes  0.269 (0.203)  0.255 (0.212)  <0.001  0.996  <0.001 
Gaze proportion of mouth  0.214 (0.180)  0.216 (0.237)  0.058  0.811  <0.001 
Gaze proportion of context  0.053 (0.035)  0.035 (0.018)  2.400  0.128  0.046 

Note. BAP: Broad Autism Phenotype Group; Bold value indicates that the value is statistically significant.

When examining the congruent condition of ECT, we confirmed a significant difference in gaze proportion of the face between groups. However, no difference in gaze proportion of the eyes, mouth, and context between groups was found (Fig. 3).

Fig. 3.

Comparison of gaze proportion between groups under the condition of mild emotional intensity in congruent and incongruent conditions of ECT. BAP, broad autism phenotype group; Control, control group. **p < .01.

(0.28MB).

When investigating the incongruent condition of ECT, no difference in gaze proportion of the face, eyes, mouth, and context between groups was found (Fig. 3).

Since there was a significant difference of accuracy between the BAP and control groups in both the congruent and incongruent conditions, we combined the conditions and conducted a 2 (group; BAP, control) × 2 (congruence; congruent, incongruent) mixed-design ANCOVA on gaze proportion of the face, eyes, mouth, and context (Supplementary Table 5). There was no interaction effect of group × congruence and no main effect of congruence. However, there was a significant main effect of group for face, indicating that the BAP group looked at the face shorter than the control group, regardless of emotional congruence between target and context.

Discussion

This study examined whether individuals with BAP showed a diminished ability to recognize emotions compared to the general population as the task becomes more difficult. This study also identified whether the reduced ability to recognize emotions was related to the gaze pattern on the face. This study corroborated that our hypotheses were supported. First, the more difficult the task, the clearer the significant relationships between the level of autistic traits and emotion recognition ability. Especially, the BAP group showed lower accuracy of emotion recognition when recognizing emotions with low emotional intensity in complex situations where faces and contexts were presented together. Second, the reduced emotion recognition ability of the BAP group could be accompanied by decreased gaze proportion on face.

We divided tasks into three types according to difficulty level and investigated correlation analysis between the level of autistic traits and emotion recognition ability in each type. Additionally, difference in emotion recognition ability between BAP and control groups was checked in each type. Through these processes, we confirmed that the first hypothesis of this study was supported. These are comprehensive results that can explain why some studies have shown that the BAP group had similar emotion recognition ability to the general population (Soto-Icaza et al., 2022), while some found that the BAP group had lower emotion recognition ability than the general population (Stewart et al., 2020). While individuals with BAP have mild autistic characteristics that do not show clinical impairment, the BAP group was confirmed to be different from the general population. By varying task difficulty, we could clearly confirm this characteristic of the individuals with BAP. The BAP group showed a normal level of emotion recognition ability in the easy task, but they showed diminished emotion recognition ability in the difficult task. People usually reveal their emotions along with various external factors including body gestures or voices, and they rarely express their emotions in a full-blown manner (Mondada, 2016). Social interaction with others is composed of various emotional elements and successful social interaction is possible only when these are fully understood. The reason that the BAP group is unable to adequately engage in social interaction may be from their inability to identify the complex social interaction related to emotion recognition. As this inability can induce depressive and anxiety disorders, our results suggest that the BAP group should be considered as a priority group in need and an extension of ASD.

Notably, this study corroborated that the diminished ability to recognize emotion in the BAP group could be accompanied by their diminished gaze proportion on face. We confirmed the BAP group looked less at face than the control group in the congruent condition of Phase 2 and there was a main effect of group in group × congruence ANCOVA on the gaze proportion of face. These results raise the possibility that a decrease in the gaze proportion on faces may be associated with a decrease in the accuracy of emotion recognition. People can express their emotions primarily through their faces, so it is effective to recognize others’ emotions through their faces in social interaction (Costa-Cordella, 2021; Garcia-Garcia et al., 2022; Mondada, 2016). Additionally, for effective emotion perception, participants were required to select a part of the face in a complex social interaction environment, sustain attention, and obtain information about another's emotional state (Morrison et al., 1988). Indeed, it has been reported that gaze duration on faces is associated with facial emotion recognition (Stanley et al., 2013). Previous eye-tracking studies in ASD group have reported that individuals with ASD look at face shorter than general population (Kaliukhovich et al., 2020; Rigby et al., 2016), and that the shorter gaze on face is associated with poorer emotion recognition and deficits in social interaction (Murias et al., 2018; Sasson et al., 2016). According to our results, the BAP group also looked at the face for a shorter time than the general population in emotion recognition. This tendency of the BAP group may contribute to their diminished emotion recognition ability.

However, there was no significant difference in gaze proportion on the face between the BAP and control groups when performing the incongruent condition of Phase 2 (ECT), even though the BAP group showed lower emotion recognition accuracy than the control group. This result may be due to the fact that emotion recognition is influenced not just by the cognitive patterns, but also by the process of interpreting emotions. People process faces based on two procedures (Haxby et al., 2000). The core system, the first stage of face-processing, is related to the visual analysis of faces. The extended system, the second stage, is associated with processing information (e.g., emotions or identity) from the face. The face gaze is related to the core system and the ability to recognize facial emotions is related to the extended system. While the face gaze is critical for face-processing in the core system aspect, the process of interpreting the emotion of the gazed face is also an important factor in face-processing in the extended system aspect. Our results indicate that the BAP group's lower ability to interpret faces within the extended system compared to the general population could also affect their lower emotion recognition accuracy.

This study didn't find a significant difference in gaze proportion on context between the BAP and control groups. Based on previous studies, individuals with BAP as well as ASD gazed less at face and more at other contexts compared to general population (Kaliukhovich et al., 2020; Nayar et al., 2022; Rigby et al., 2016; Shic et al., 2011). The reason why there was no difference in gaze proportion on context between groups in this study may be related to the instruction of the task. Previous studies reporting higher gaze proportion on context in autism groups used free-viewing task, so that all participants gazed the social images free. On the contrary, this study used emotion recognition tasks that instructed participants to respond as quickly and accurately as possible to the emotion of the face presented in the center. Both the BAP and control groups would have been looking at the face and trying to determine the emotion of the face as quickly as possible, rather than paying attention to the context. Future research is needed to investigate whether the BAP group may have different patterns of gazing at faces and contexts depending on the type of task (free-viewing task and emotion recognition task).

In this study, the BAP as well as control groups recognized happy emotion faster than other emotions, and recognized fearful emotion slower and less accurately than other emotions. These results indicate that although the BAP group showed a decrease in emotion recognition compared to the control group, but the pattern of difference in recognition ability between emotion types were similar to the control group. Previous studies reporting the diminished ability to recognize emotion in individuals with BAP showed that the BAP group recognized happy faces faster than sad and fearful faces, like the control group, and other study found no interaction effect of group × emotion on the accuracy of emotion recognition (Cha & Lee, 2022; Sasson et al., 2016). Moreover, previous study in the general population reported that people are able to recognize happy faces faster and more accurately than other emotions, while recognizing fearful emotions slower and more inaccurately (Wells et al., 2016). Taken together with these studies, the results of the present study suggest that the BAP group may not be specifically impaired for certain emotions compared to the control group, but rather show an overall impairment in emotion recognition, and that the differences in recognition across emotions exhibited by the individuals with BAP may be similar to the pattern seen in the general population.

This study had a few limitations. First, while this study suggested that the decreased emotion recognition in the BAP group might have resulted from a decrease in gaze proportion on face, it had not examined the direct relationship between the two variables by setting gaze proportion on face as an independent variable. Follow-up studies are needed to verify that reduced gaze proportion on faces resulted in diminished emotion recognition in the individuals with BAP based on direct comparison. Second, although the levels of anxiety and depression were statistically controlled when analyzing, the anxiety and depression may have induced an attention bias toward certain emotions, especially negative emotions (Bergman et al., 2021). In future studies, depression and anxiety levels should also be matched when recruiting BAP and control groups for analysis. Third, in this study, individuals with high and low BAPQ/AQ scores were selected for the study. However, since individuals with BAP are located between the ASD group and the general population at various levels of the spectrum, it may be more effective to identify the characteristics of the BAP population by taking a dimensional approach rather than categorizing them as high or low. Future research should attempt to characterize the BAP population using a dimensional approach based on a larger sample size. Fourth, this study confirmed that no significant difference between the BAP and control groups could be found in the gaze proportion at the eyes and mouth, contrary to previous studies based on ASD (Setien-Ramos et al., 2022; Riddiford et al., 2022). As some studies have reported that the BAP group also looked at the eyes less and mouth longer than the general population, follow-up studies will be needed to confirm the relationship between the BAP and the gaze pattern on the eyes and mouth (Merin et al., 2007). Fifth, in this study, the BAP and control groups didn't show a significant difference between mild and full-blown intensity when recognizing angry emotion, unlike other emotions. Based on the current literature, there is no clear explanation for why this result was only found for angry emotion, and further research is needed to clarify the reason.

This study highlights the decline in the emotion recognition ability of the BAP by presenting context and using facial stimuli with mild intensity. Additionally, this study confirmed that this decline could be influenced by their gaze patterns. Previous studies reported that individuals with BAP had depression and anxiety owing to difficulties in social interaction. This study can provide evidence for why individuals with BAP experience difficulties in social interactions. Moreover, these results offer evidence that intervention which provides feedback to the BAP group to look at faces in the emotion recognition process will be effective. We can expect that intervention based on the real-time eye-feedback skill can modify eye movements and improve individuals’ ability to recognize emotions by increasing their gaze proportion on faces.

Funding

This work was supported by the Ministry of Education of the Republic of Korea and the National Research Foundation of Korea (2019R1I1A1A01062451).

References
[Baron-Cohen et al., 2001]
S. Baron-Cohen, S. Wheelwright, R. Skinner, J. Martin, E. Clubley.
The autism-spectrum quotient (AQ): Evidence from asperger syndrome/high-functioning autism, malesand females, scientists and mathematicians.
Journal of Autism and Developmental Disorders, 31 (2001), pp. 5-17
[Beck et al., 1988]
A.T. Beck, N. Epstein, G. Brown, R.A. Steer.
An inventory for measuring clinical anxiety: Psychometric properties.
Journal of Consulting and Clinical Psychology, 56 (1988), pp. 893-897
[Beck et al., 1996]
A.T. Beck, R.A. Steer, R. Ball, W.F. Ranieri.
Comparison of beck depression inventories-IA and-II in psychiatric outpatients.
Journal of Personality Assessment, 67 (1996), pp. 588-597
[Bergman et al., 2021]
M.A. Bergman, J.N. Vrijsen, M. Rinck, I. van Oostrom, C.C. Kan, R.M. Collard, et al.
Is a negative attentional bias in individuals with autism spectrum disorder explained by comorbid depression? An eye-tracking study.
Journal of Autism and Developmental Disorders, 51 (2021), pp. 4213-4226
[Cha & Lee, 2022]
W.J. Cha, J.H. Lee.
Diminished ability to integrate target stimuli with context during emotional recognition in individuals with broad autism phenotype.
Frontiers in Psychology, 13 (2022), pp. 934385
[Costa-Cordella et al., 2021]
S. Costa-Cordella, C. Arevalo-Romero, F.J. Parada, A. Rossi.
Social support and cognition: A systematic review.
Frontiers in Psychology, 12 (2021),
[Faul et al., 2007]
F. Faul, E. Erdfelder, A.G. Lang, A. Buchner.
G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences.
Behavior Research Methods, 39 (2007), pp. 175-191
[Garcia-Garcia et al., 2022]
J.M. Garcia-Garcia, V.M. Penichet, M.D. Lozano, A. Fernando.
Using emotion recognition technologies to teach children with autism spectrum disorder how to identify and express emotions.
Universal Access in the Information Society, 21 (2022), pp. 809-825
[Happe and Frith, 2006]
F. Happe, U. Frith.
The weak coherence account: Detail-focused cognitive style in autism spectrum disorders.
Journal of Autism and Developmental Disorders, 36 (2006), pp. 5-25
[Haxby et al., 2000]
J.V. Haxby, E.A. Hoffman, M.I. Gobbini.
The distributed human neural system for face perception.
Trends in Cognitive Sciences, 4 (2000), pp. 223-233
[Holt et al., 2014]
R.J. Holt, L.R. Chura, M.C. Lai, J. Suckling, E. Von Dem Hagen, A.J. Calder, et al.
Reading the mind in the eyes’: An fMRI study of adolescents with autism and their siblings.
Psychological Medicine, 44 (2014), pp. 3215-3227
[Hurley et al., 2007]
R.S. Hurley, M. Losh, M. Parlier, J.S. Reznick, J. Piven.
The broad autism phenotype questionnaire.
Journal of Autism and Developmental Disorders, 37 (2007), pp. 1679-1690
[Kaliukhovich et al., 2020]
D.A. Kaliukhovich, N.V. Manyakov, A. Bangerter, S. Ness, A. Skalkin, M.S. Goodwin, et al.
Social attention to activities in children and adults with autism spectrum disorder: Effects of context and age.
Molecular Autism, 11 (2020), pp. 1-14
[Kim and Kim, 2022]
H.J. Kim, K. Kim.
Psychometric properties of the Korean version of the broad autism phenotype questionnaire.
Social Behavior and Personality: An International Journal, 50 (2022), pp. 1-13
[Ko et al., 2018]
H.Y. Ko, W.H. Lee, E.K. Won, J.J. Ban, D.E. Jung, Y. Kim.
The reliability and validity of the Korean version of the autism-spectrum quotient.
Psychiatry Investigation, 15 (2018), pp. 783-789
[Kulasinghe et al., 2021]
K. Kulasinghe, K. Whittingham, A.E. Mitchell.
Mental health, broad autism phenotype and psychological inflexibility in mothers of young children with autism spectrum disorder in Australia: A cross-sectional survey.
Autism: The International Journal of Research and Practice, 25 (2021), pp. 1187-1202
[Le et al., 2022]
N. Le, K. Nguyen, A. Nguyen, B. Le.
Global-local attention for emotion recognition.
Neural Computing and Applications, 34 (2022), pp. 21625-21639
[Lee et al., 2017]
E.H. Lee, S.J. Lee, S.T. Hwang, S.H. Hong, J.H. Kim.
Reliability and validity of the beck depression inventory-II among Korean adolescents.
Psychiatry Investigation, 14 (2017), pp. 30-36
[Lee et al., 2013]
K.U. Lee, J. Kim, B. Yeon, S.H. Kim, J.H. Chae.
Development and standardization of extended ChaeLee Korean facial expressions of emotions.
Psychiatry Investigation, 10 (2013), pp. 155-163
[Merin et al., 2007]
N. Merin, G.S. Young, S. Ozonoff, S.J. Rogers.
Visual fixation patterns during reciprocal social interaction distinguish a subgroup of 6-month-old infants at-risk for autism from comparison infants.
Journal of Autism and Developmental Disorders, 37 (2007), pp. 108-121
[Mondada, 2016]
L. Mondada.
Challenges of multimodality: Language and the body in social interaction.
Journal of Sociolinguistics, 20 (2016), pp. 336-366
[Morrison et al., 1988]
R.L. Morrison, A.S. Bellack, K.T. Mueser.
Deficits in facial-affect recognition and schizophrenia.
Schizophrenia Bulletin, 14 (1988), pp. 67-83
[Murias et al., 2018]
M. Murias, S. Major, K. Davlantis, L. Franz, A. Harris, B. Rardin, et al.
Validation of eye-tracking measures of social attention as a potential biomarker for autism clinical trials.
Autism Research, 11 (2018), pp. 166-174
[Nayar et al., 2022]
K. Nayar, F. Shic, M. Winston, M. Losh.
A constellation of eye-tracking measures reveals social attention differences in ASD and the broad autism phenotype.
Molecular Autism, 13 (2022), pp. 18
[Pisula and Ziegart-Sadowska, 2015]
E. Pisula, K. Ziegart-Sadowska.
Broader autism phenotype in siblings of children with ASD—a review.
International Journal of Molecular Sciences, 16 (2015), pp. 13217-13258
[Riddiford et al., 2022]
J.A. Riddiford, P.G. Enticott, A. Lavale, C. Gurvich.
Gaze and social functioning associations in autism spectrum disorder: A systematic review and meta-analysis.
Autism Research, 15 (2022), pp. 1380-1446
[Rigby et al., 2016]
S.N. Rigby, B.M. Stoesz, L.S. Jakobson.
Gaze patterns during scene processing in typical adults and adults with autism spectrum disorders.
Research in Autism Spectrum Disorders, 25 (2016), pp. 24-36
[Sasson et al., 2013]
N.J. Sasson, R.B. Nowlin, A.E. Pinkham.
Social cognition, social skill, and the broad autism phenotype.
Autism: The International Journal of Research and Practice, 17 (2013), pp. 655-667
[Sasson et al., 2016]
N.J. Sasson, A.E. Pinkham, L.P. Weittenhiller, D.J. Faso, C. Simpson.
Context effects on facial affect recognition in schizophrenia and autism: Behavioral and eye-tracking evidence.
Schizophrenia Bulletin, 42 (2016), pp. 675-683
[Setien-Ramos et al., 2022]
I. Setien-Ramos, J. Lugo-Marín, L. Gisbert-Gustemps, E. Díez-Villoria, M. Magán-Maganto, R. Canal-Bedia, et al.
Eye-tracking studies in adults with autism spectrum disorder: A systematic review and meta-analysis.
Journal of Autism and Developmental Disorders, (2022), pp. 1-14
[Shic et al., 2011]
F. Shic, J. Bradshaw, A. Klin, B. Scassellati, K. Chawarska.
Limited activity monitoring in toddlers with autism spec- trum disorder.
Brain Research, 1380 (2011), pp. 246-254
[Soto-Icaza et al., 2022]
P. Soto-Icaza, B. Beffara-Bret, L. Vargas, F. Aboitiz, P. Billeke.
Differences in cortical processing of facial emotions in broader autism phenotype.
[Stanley et al., 2013]
J.T. Stanley, X. Zhang, H.H. Fung, D.M. Isaacowitz.
Cultural differences in gaze and emotion recognition: Americans contrast more than Chinese.
Emotion (Washington, D.C.), 13 (2013), pp. 36
[Stewart et al., 2020]
G.R. Stewart, G.L. Wallace, M. Cottam, R.A. Charlton.
Theory of mind performance in younger and older adults with elevated autistic traits.
Autism Research, 13 (2020), pp. 751-762
[Wells et al., 2016]
L.J. Wells, S.M. Gillespie, P. Rotshtein.
Identification of emotional facial expressions: Effects of expression, intensity, and sex on eye gaze.
PloS One, 11 (2016),
[Yook and Kim, 1997]
S.P. Yook, Z.S. Kim.
A clinical study on the Korean version of Beck anxiety inventory: Comparative study of patient and non-patient.
Korean Journal of Clinical Psychology, 16 (1997), pp. 185-197

Present address: Department of Neuropsychiatry, Seoul National University Hospital, Seoul 03080, Republic of Korea.

Copyright © 2023. The Authors
Article options
Tools
Supplemental materials