Introduction
Perception of emotion from facial information is vital for effective social and relational functioning; misinterpretation of emotional expressions can lead to uncomfortable social situations and reduce appropriate social communication. There is growing evidence that impaired facial emotion processing ability is a feature of the social cognitive profile of bipolar disorder (BD: see Van Rheenen & Rossell, Reference Van Rheenen and Rossell2013b). This impairment is likely to be a factor in the problematic psychosocial functioning and reduced quality of life seen in BD (Hoertnagl et al., Reference Hoertnagl, Muehlbacher, Biedermann, Yalcin, Baumgartner, Schwitzer and Hofer2011; Martino, Strejilevich, Fassi, Marengo, & Igoa, Reference Martino, Strejilevich, Fassi, Marengo and Igoa2011).
Studies investigating emotion processing have shown that patients with BD have an impaired capacity to recognize and discriminate facial emotions. This effect has been demonstrated in both symptomatic and euthymic samples as well as in at risk groups (e.g., Bozikas, Tonia, Fokas, Karavatos, & Kosmidis, Reference Bozikas, Tonia, Fokas, Karavatos and Kosmidis2006; Brotman, Guyer, et al., Reference Brotman, Guyer, Lawson, Horsey, Rich, Dickstein and Leibenluft2008; Getz, Shear, & Strakowski, Reference Getz, Shear and Strakowski2003; Lembke & Ketter, Reference Lembke and Ketter2002; Vederman et al., Reference Vederman, Weisenbach, Rapport, Leon, Haase, Franti and McInnis2012). Although this body of research is growing, there is still little clarity around the nature of the deficit. For example, although many studies have reported a general reduction in emotion processing accuracy (Brotman, Skup, et al., Reference Brotman, Skup, Rich, Blair, Pine, Blair and Leibenluft2008; Derntl, Seidel, Kryspin-Exner, Hasmann, & Dobmeier, Reference Derntl, Seidel, Kryspin-Exner, Hasmann and Dobmeier2009; Getz, et al., Reference Getz, Shear and Strakowski2003; Gray et al., Reference Gray, Venn, Montagne, Murray, Burt, Frigerio and Young2006; Guyer et al., Reference Guyer, McClure, Adler, Brotman, Rich, Kimes and Leibenluft2007), there are findings suggesting that abnormalities are more heavily weighted toward the processing of fear (Lembke & Ketter, Reference Lembke and Ketter2002; Vederman et al., Reference Vederman, Weisenbach, Rapport, Leon, Haase, Franti and McInnis2012), sadness (Derntl et al., Reference Derntl, Seidel, Kryspin-Exner, Hasmann and Dobmeier2009; Schenkel, Pavuluri, Herbener, Harral, & Sweeney, Reference Schenkel, Pavuluri, Herbener, Harral and Sweeney2007; Vederman, et al., Reference Vederman, Weisenbach, Rapport, Leon, Haase, Franti and McInnis2012), or surprise (Summers, Papadopoulou, Bruno, Cipolotti, & Ron, Reference Summers, Papadopoulou, Bruno, Cipolotti and Ron2006). Some studies have found no deficit (Vaskinn et al., Reference Vaskinn, Sundet, Friis, Simonsen, Birkenæs, Engh and Andreassen2007), and some have found impairments of emotion discrimination but not labeling (Addington & Addington, Reference Addington and Addington1998; Rossell, Van Rheenen, Groot, Gogos, & Joshua, Reference Rossell, Van Rheenen, Groot, Gogos and Joshua2013).
Progress toward adequately profiling facial emotion processing abilities in BD is currently hampered by several factors. First, existing studies differ in terms of task stimuli, with some using still photographic stimuli (static tasks), and some using morphing facial expressions (dynamic tasks). As there has been no study directly comparing performance across static and dynamic stimuli in the same BD cohort, the impact of subtle differences between task stimuli designs is not known. Dynamic tasks are arguably more ecologically valid; in the healthy population accuracy rates for facial emotion recognition are better for dynamic task designs which suggests that motion has a facilitatory effect on facial emotion perception (Ambadar, Schooler, & Cohn, Reference Ambadar, Schooler and Cohn2005). Whether this effect extends to BD, however, remains to be seen.
Second, very few studies in BD have investigated both emotion labeling and discrimination performance. However, as these separate, albeit related abilities require different skills (Feinberg, Rifkin, Schaffer, & Walker, Reference Feinberg, Rifkin, Schaffer and Walker1986; Walker, McGuire, & Bettes, Reference Walker, McGuire and Bettes1984), it is uncertain as to whether facial emotion processing problems in BD reflect a specific difficulty in recognizing emotions on the basis of an impairment in matching emotional facial cues (discrimination), a specific difficulty in applying linguistic labels to different expressions (labeling), or a more generalized impairment involving both processes. Third, given the generally small effect sizes for facial emotion processing differences between patients and controls (Samamé, Martino, & Strejilevich, Reference Samamé, Martino and Strejilevich2012; Vaskinn et al., Reference Vaskinn, Sundet, Friis, Simonsen, Birkenæs, Engh and Andreassen2007), it is likely that emotion processing impairments in BD are subtle. Therefore, slight alterations in the intensity of a facial expression stimulus could well influence observed group related differences in identification or discrimination of expressions. Although some studies have attempted to profile sensitivity thresholds for emotion in BD, there are few that have assessed the threshold of intensity at which emotions are most consistently identified. The former use paradigms in which respondents themselves alter the intensity of an expression until it reaches a level at which it is recognizable (Gray et al., Reference Gray, Venn, Montagne, Murray, Burt, Frigerio and Young2006; Schaefer, Baumann, Rich, Luckenbaugh, & Zarate, Reference Schaefer, Baumann, Rich, Luckenbaugh and Zarate2010; Summers et al., Reference Summers, Papadopoulou, Bruno, Cipolotti and Ron2006; Venn et al., Reference Venn, Gray, Montagne, Murray, Michael Burt, Frigerio and Young2004). However, these paradigms merely permit the assessment of how much intensity is required to perceive an emotion, rather than how reliably emotions are perceived at different intensities. To determine if BD is associated with reduced perceptual processing of emotional cues, research designs manipulating stimulus intensity are required. Demonstration that people with BD require higher levels of stimulus intensity to reliably label and discriminate facial emotions would constitute evidence for this hypothesized subtle but functionally important deficit.
In light of these factors, we set out to comprehensively examine facial emotion processing in a group of patients with BD compared to controls in a single experiment in which multiple variables were manipulated sequentially. Our objectives were fourfold in nature; first, we aimed to establish the comparability of two emotion labeling task stimuli designs (dynamic and static) by directly contrasting performance between them. Second, we aimed to test for group related differences in the consistency with which facial emotions are processed at different levels of intensity. Third, we aimed to determine the specificity of potential emotion processing deficits as a function of task type (labeling vs. discrimination). Finally, we aimed to determine whether emotion labeling performance varied as a function of emotion types between groups, to establish whether impairments generalize to a range of basic emotions, or are more weighted toward a single emotion or subset of emotions in BD. Broadly we expected that BD-related impairments in facial emotion processing would be seen in both labeling and discrimination tasks. The following specific hypotheses were made; BD-related impairments will be seen irrespective of stimulus type (dynamic versus static: Hypothesis 1), BD-related impairments will be significant across all emotions (Hypothesis 2), BD-related impairments will be seen at all levels of stimulus intensity for both emotion labeling (Hypothesis 3a) and emotion discrimination (Hypothesis 3b), BD will not be associated with a deficiency in non-emotional identification of faces (Control task: Hypothesis 4).
Method
This study was approved by the Alfred Hospital and Swinburne University Human Ethics Review Boards and abided by the Declaration of Helsinki. Written informed consent was obtained from each participant before the study began.
Participants
The clinical sample comprised 50 patients (16 male, 34 female) diagnosed as having DSM-IV-TR BD (39 BD I, 12 BD II) using the Mini International Neuropsychiatric Interview (MINI: Sheehan et al., Reference Sheehan, Lecrubier, Harnett Sheehan, Amorim, Janavs, Weiller and Dunbar1998). Patients were recruited via community support groups and general advertisements and were all out-patients. Current symptomology was assessed using the Young Mania Rating Scale (YMRS: Young, Biggs, Ziegler, & Meyer, Reference Young, Biggs, Ziegler and Meyer1978) and the Montgomery Asberg Depression Rating Scale (MADRS: Montgomery & Asberg, Reference Montgomery and Asberg1979); there were 17 depressed (defined as those that met strict criteria for MADRS scores > 8), 12 mixed (defined as those that met strict criteria for YMRS and MADRS scores > 8), 4 (hypo)manic (defined as those that met strict criteria for YMRS scores > 8) and 17 euthymic (defined as those that met strict criteria for YMRS and MADRS scores≤8) patients (i.e., 33 that were symptomatic). Patients with current psychosis, co-morbid psychotic disorders, visual impairments, neurological disorder and/or a history of substance/alcohol abuse or dependence during the past six months were excluded. Thirty-two patients were taking antipsychotics, 16 were taking antidepressants, 16 were taking mood stabilizers and 10 were taking benzodiazepines.Footnote 1 Demographic and clinical characteristics are presented in Table 1.
Note. ^ Group comparisons all independent samples t-tests except gender and education which was chi-squared; M/F = Male/Female, WTAR = Wechsler Test of Adult Intelligence, YMRS = Young Mania Rating Scale, MADRS = Montgomery Asberg Depression Rating Scale.
A control sample of 52 healthy participants (20 male, 32 female) were recruited for comparison purposes by general advertisement and contacts of the authors. Using the MINI screen, no control participant had a current diagnosis or previous history of psychiatric illness (Axis I). An immediate family history of mood and psychiatric disorder, in addition to a personal history of neurological disorder, current or previous alcohol/substance dependence or abuse, visual impairments and current psychiatric medication use was exclusion criteria for all controls.
All participants were fluent in English, were between the ages of 18 and 65 years and had an estimated pre-morbid IQ as scored by the Wechsler Test Of Adult Reading (WTAR) of >90.
Materials
All participants completed three computerized tasks (designed by the author) that were used to measure (1) emotion labeling performance across stimulus type (dynamic versus static) and emotion type (happy, sad, angry, fear, neutral), (2) emotion discrimination, and (3) emotion labeling and discrimination performance across three levels of intensity. Participants also completed a control task (designed by the authors) to test whether potential impairments on the facial emotion processing tasks were reflective of a generalized performance deficit or were specific to facial emotion processing.
The face stimuli were taken from the widely used and well validated Ekman and Friesen series known as the Pictures of Facial Affect (POFA: Ekman & Friesen, Reference Ekman and Friesen1976). The stimuli comprised black and white photographs of faces free of jewelry, spectacles, make up and facial hair (five female and five male) and expressing the emotions happy, sad, fear, angry and neutral. The faces were cropped to an oval shape spanning the top of the forehead to the bottom of the chin and excluding any hair and the ears on either side of the face. Thirteen faces were used in total. However, given that we endeavored to have an equal presentation of emotional expressions of different genders in the labeling tasks across ten different trials for each emotion and neutral, and that some expressions were not available for all faces, the tasks presented below varied in the exact faces used.
A morphing program called Fantamorph (Abrosoft, 2012) was used to reduce the intensity of the POFA stimuli's emotional expression by 25% decrements to create static intensity varied stimuli. This resulted in static stimulus expressions at 100%, 75%, 50%, and 25% intensity. Pilot testing revealed that emotions presented at 25% intensity elicited floor effects in controls, and they were therefore excluded from the task. The final stimulus set for the latter two tasks comprised static faces displaying 100%, 75%, and 50% (high, medium, and low) emotional intensity only. The dynamic stimuli were created by morphing the low, medium, and high intensity static faces through quick successive frames from a neutral expression (0%) to the final emotional expression (100%), such that they appeared as a moving image. All tasks were presented on a 14′′ Lenovo laptop computer and were run through Presentation (Neurobehavioral Systems Inc, 2012). These tasks are described below.
The dynamic facial emotion labeling task required participants to view dynamic facial images (i.e., morphs) and identify the emotion being expressed. Forty randomized dynamic display trials comprising 10 presentations (5 male and 5 female faces) each for happy, sad, fear, and angry expressions were presented one at a time for 1500 msFootnote 2 followed by an inter-stimulus interval of 1500 ms. Participants were instructed to press a labeled keyboard button corresponding to the emotion that they believed the face was expressing as soon as they recognized it. The averaged accuracy percentage for each emotion was taken as the primary dependent variable. This task was used in the analyses testing Hypothesis 1 and 2.
The static facial emotion labeling task was designed to assess participants’ ability to identify emotional expressions from static facial stimuli (i.e., a photograph). It required participants to view an image of a male or female face, and identify the emotional expression exhibited by that face. The faces were presented one at a time on a black background for 2000 ms followed by an inter-stimulus interval of 1500 ms. The task involves 130 randomized trials in total, including 10 presentations (5 male and 5 female faces) for each of the emotions happy, sad, angry, and fear at each of the three levels of emotional intensity (high, medium, and low) and 10 presentations of neutral. Participants were instructed to press a labeled keyboard button corresponding to the emotion that they believed the face was expressing as soon as they recognized it. The averaged accuracy percentage and response time for each emotion at each level of intensity was taken as the primary dependent variable. This task was used in analyses testing Hypotheses 1 and 3a.
The static facial emotion discrimination task was designed to assess participants’ ability to differentiate between static facial emotional expressions at different levels of intensity. It required participants to view two simultaneously presented images of human faces and identify whether the emotion that the two faces were showing was the same or different. The task represents the emotions happy, sad, angry, fear, and neutral over 135 randomized paired stimulus trials at the three different intensities.Footnote 3 Emotional expressions were paired only with those expressing the same level of intensity (i.e., high intensity expressions paired together) or a neutral expression, but never with an emotion of a different expressive intensity. There were 31 incongruent paired trials (representing pairings across the emotions happy, sad, angry, fear, and neutral) and 11 congruent paired trials (representing pairings across happy, sad, angry, and fear) for each level of intensity, with 9 additional trials representing paired neutral expressions that were used as fillers. One face in each pair was presented to the left visual field, and the other was presented to the right visual field on a black background for 2000 ms followed by an inter-stimulus interval of 1500 ms. Participants were instructed to respond via two button keyboard press (same or different) as soon as they could discriminate the emotional expressions. Responses made from 200 ms onward were recorded. Written instructions, an example and a set of practice trials were provided to participants before commencing the task. The averaged accuracy percentage and response times across congruent and incongruent trials at each level of intensity was taken as the primary dependent variable for this task. This task was used in analyses testing Hypothesis 3b.
To rule out the possibility that BD is associated with a more fundamental deficit in facial processing (i.e., not specific to higher order facial emotion processing), a static identity discrimination task was designed. The tasks assessed participant's ability to determine whether two simultaneously presented static facial stimuli were identical or not. The task comprised 55 (45 incongruent and 10 congruent) randomized trials made up of couplings between six male and six female faces displaying neutral expressions only. One face in each pair was presented to the left visual field, and the other was presented to the right visual field on a black background for 2000 ms followed by an inter-stimulus interval of 1500 ms. The task involved 65 randomized trials with the same pairs being presented twice. Participants were instructed to respond via two button keyboard press (same or different) as soon as they could discriminate the faces. Responses made from 200 ms onward were recorded with the averaged accuracy percentage and response times across same/different pairs taken as the dependent variables. This task was used in the analysis testing Hypothesis 4.
Statistical Analysis
Demographic and clinical group differences were assessed via independent samples t tests or χ2 tests. We conducted a series of analyses to address our hypotheses: we used a four (emotion: happy, sad, angry, fear) * two (stimuli type: static, dynamic) * two (group: control, BD) repeated measures analysis of variance (ANOVA) of the accuracy data to address Hypothesis 1 and 2. Significant main effects of group, emotion, and of stimuli type were expected to show support for this hypothesis. Only responses to high intensity conditions were used as the static emotion labeling variables in this analysis. Due to the response time windows differing between dynamic and static stimuli tasks, we were unable to analyze response time differences across these tasks.Footnote 4 Two repeated measures ANOVAs using a four (emotion: happy, sad, angry, fear) * three (intensity: high, medium, low) * two (group: control, BD) design with the accuracy and response time data of the static emotion labeling task were used to address Hypothesis 3a. To address Hypothesis 3b we also completed two, three (intensity: high, medium, low) * two (group: control, BD) repeated measures ANOVAs on the static emotion discrimination task, separately for the accuracy and response time data. Support for these hypotheses was expected to be shown through main effects of group and intensity. Two one-way ANOVAs were used to compare accuracy and response time performance between groups on the control task (i.e., the static identity discrimination task), and thus to address Hypothesis 4. The absence of a significant group effects would demonstrate support for this hypothesis.
To better understand the effects of mood and diagnostic status on facial emotion processing performance, all analyses were re-run in the patient group: in the first series of analyses, diagnosis (BD I; n = 38 or BD II; n = 12) was entered as the between-groups factor for all tasks. For the second series of analyses, current mood state was entered as the between-groups factor; however, given that the sample size of some of the mood state subgroups was too small for meaningful analysis, we collapsed the mixed (n = 12) and manic (n = 4) groups into one (resulting n = 16) and compared this to patients meeting criteria for euthymia (n = 17) or depression (n = 17). Bivariate correlations were also conducted to examine the relationship between emotion labeling and discrimination performance and symptom severity on the YMRS and MADRS. All analyses were corrected for multiple testing using a conservative α set at .01.
Results
No significant differences in age, gender, education level completed, or pre-morbid IQ were found between the two groups (see Table 1).
Emotion Labeling Accuracy as a Function of Task Stimuli Type (Hypotheses 1 and 2)
Investigation of the determinants of emotion labeling accuracy found main effects of emotion (Greenhouse Geisser corrected F(2.715, 271.538) = 71.51; p < .001, partial η 2 = .42), stimuli type: static versus dynamic (F(1,100) = 8.21; p < .01; partial η 2 = .08) and group (F(1,100) = 7.04; p < .01; partial η 2 = .07), but no two- or three-way interactions reached significance. Accuracy across emotions occurred in the descending order of happy (M = 98.43; SD = 2.90), fear (M = 86.32; SD = 12.60), angry (M = 81.76; SD = 13.67), and sad (M = 79.59; SD = 13.93), with performance being slightly better for the dynamic stimuli task relative to the static stimuli task (dynamic: M = 87.60; SD = 8.02; static M = 85.44; SD = 9.20; d = −0.25) and BD patients performing less accurately than controls overall (Control: M = 88.43; SD = 6.73; BD: M = 84.50; SD = 8.30; d = −0.52). Figure 1 presents emotion labeling accuracy performance for dynamic and static stimuli tasks as a function of emotion across groups.
Static Emotion Labeling as a Function of Emotion and Intensity (Hypothesis 3a)
Table 2 presents the accuracy and response time means and standard deviations for static emotion labeling as a function of intensity across groups. Analyses were conducted separately for the two dependent variables.
Note: BD = bipolar disorder; d = Cohen's d; static labelling group accuracy effect p = .07; static labelling group response time effect p < .05.
Accuracy
There was a main effect of emotion (F(3,297) = 152.37; p < .001, partial η 2 = .61) and intensity (Greenhouse Geisser corrected F(1.86, 184.22) = 621.83; p < .001, partial η 2 = .86), but no group effect or two- or three-way interactions. Accuracy across emotions occurred in the descending order of happy (M = 91.74; SD = 6.57), fear (M = 75.52; SD = 13.82), sad (M = 69.51; SD = 14.94) and angry (M = 60.98; SD = 14.05), with performance best in high intensity (M = 85.44; SD = 9.20), followed by medium intensity (M = 80.02; SD = 9.52) and low intensity (M = 57.72; SD = 11.62) conditions. Although not significant, BD patients performed less accurately than controls and the effect size difference was in the medium range (Control: M = 76.01; SD = 8.18; BD: M = 72.73; SD = 9.56; d = −0.37).
Response time
There was a main effect of emotion (Greenhouse Geisser corrected F (2.65; 265.30) = 190.03; p < .001; partial η 2 = .66), intensity (Greenhouse Geisser corrected F(1.62, 161.94) = 98.901; p < .001; partial η 2 = .50), and group (F(1,100) = 5.61; p < .05; partial η 2 = .05), although the latter did not survive statistical correction. There were no two- or three-way interactions. Response latencies across emotions occurred in the ascending order of happy (M = 1233.55, SD = 224.86), sad (M = 1471.30, SD = 238.63), angry (M = 1522.61, SD = 245.00), and fear (M = 1617.83, SD = 235.61), with performance best in high intensity (M = 1385.78, SD = 229.36), followed by medium intensity (M = 1436.21, SD = 221.60) and low intensity (M = 1562.30, SD = 224.10) conditions. Although not significant, BD patients had longer latencies than controls and the effect size difference was in the medium range (Control: M = 1413.72, SD = 209.27; BD: M = 1511.05, SD = 205.63; d = 0.47).
Static Emotion Discrimination as a Function of Intensity (Hypothesis 3b)
Table 3 presents the accuracy and response time means and standard deviations for static emotion discrimination as a function of intensity across groups. Analyses were conducted separately for the two dependent variables.
Note: BD = bipolar disorder; RT = response time; d = Cohen's d; static discrimination group accuracy effect p < .01; static labelling group response time effect p = .75.
Accuracy
There was a main effect of intensity (F(2,200) = 97.78; p < .001; partial η 2 = .49) and group (F(1,100) = 9.04; p < .01; partial η 2 = .08), but no two-way interaction. Performance accuracy was best in high intensity (M = 78.85; SD = 8.89) followed by medium intensity (M = 69.40; SD = 10.88) and low intensity (M = 65.59; SD = 9.55) conditions, with overall performance being worse in patients than controls (BD: M = 68.94; SD = 7.94; Controls M =73.52; SD = 7.43; d = −0.60).
Response time
There was a main effect of intensity (F(2,200) = 69.33; p < .001; partial η 2 = .41), but no effect of group and no two-way interaction. Response latencies were shortest in medium intensity (M = 1544.58; SD = 198.71) followed by high intensity (M = 1598.50; SD = 281.22) and low intensity (M = 1700.73; SD = 286.84) conditions.
Control Task Performance (Hypothesis 4)
There were no significant accuracy or response time differences between groups on the identity labeling control task (both p's > .05).
Subgroup Analyses
There were no between-group main effects or interactions on any of the tasks for patients diagnosed as having BD I versus BD II (all p's > .05), nor were there any between-group main effects or interactions on any of the tasks for patients classified as euthymic, depressed or mixed/manic (all p's > .05). Furthermore, bivariate correlation analyses found no significant associations between accuracy or response time performance (where applicable) on any measure and severity of current depression (MADRS score) or mania (YMRS score).
Discussion
The current study examined facial emotion processing in a cohort of BD patients compared to controls in a complex experiment whereby multiple variables were manipulated. In addition to examining whether emotion labeling performance varied as a function of specific emotions between groups, we aimed to establish the comparability of two commonly used emotion labeling task stimuli designs, determine the existence of differences in facial emotion processing at varying levels of intensity, and establish the specificity of emotion processing deficits as a function of task types; labeling versus discrimination. These aims were formulated to further understandings of the emotion processing profile of BD.
Our results indicated that emotion labeling performance was better for the dynamic task stimuli relative to the static task stimuli, with patients performing worse overall (supporting Hypothesis 1). This is consistent with previous research in the healthy population indicating that emotion processing performance is more accurately assessed using tasks that enable the facilitatory effect of motion to guide emotion recognition (Ambadar et al., Reference Ambadar, Schooler and Cohn2005). Furthermore, given the significant group and emotion effects in the absence of interactions between them, there was no evidence that BD-related impairments were more heavily weighted toward a particular emotion. Rather BD patients appear to be globally compromised in processing a range of emotional expressions (supporting Hypothesis 2). We also found that the accuracy with which facial emotions were labeled and discriminated diminished in line with the degradation of stimulus intensity across groups. However, BD patients did exhibit overall deficits in discriminating emotions across these intensities compared to their control counterparts (supporting Hypothesis 3b). Contrary to expectations, this group effect was not evident for intensity labeling performance (i.e., there was no support for Hypothesis 3a), although the effect size difference between groups was still in the medium range. Finally, as no group difference was observed on the control task, it appears that the general processing of facial information was not compromised in this cohort (supporting Hypothesis 4).
Taken together, this pattern of findings suggests that BD patients have a relatively generalized impairment for the labeling of facially conveyed emotional expressions, which cannot be attributed to a pervasive impairment in the general processing of faces. These results accord with several studies in which emotion labeling (Derntl et al., Reference Derntl, Seidel, Kryspin-Exner, Hasmann and Dobmeier2009; Getz et al., Reference Getz, Shear and Strakowski2003; Vederman et al., Reference Vederman, Weisenbach, Rapport, Leon, Haase, Franti and McInnis2012) but not face processing itself (e.g., Bozikas et al., Reference Bozikas, Tonia, Fokas, Karavatos and Kosmidis2006; Getz et al., Reference Getz, Shear and Strakowski2003) has been shown to be impaired in BD. Thus it appears that BD related facial emotion processing difficulties reflect inabilities in both adequately matching facial cues of emotion to recognize expressions and adequately applying or understanding the linguistic labels used to identify them. Moreover, as diagnostic subtype and current mood state did not have any influence on the present findings, the generalized facial emotion processing impairment we have observed here is likely to be reflective of a trait-like feature of the disorder, which is consistent with past research (Bozikas et al., Reference Bozikas, Tonia, Fokas, Karavatos and Kosmidis2006; Vederman et al., Reference Vederman, Weisenbach, Rapport, Leon, Haase, Franti and McInnis2012).
Importantly, it appears that whilst variability between tasks designs and intensities may subtly influence the strength at which emotion processing abilities are apparent (i.e., performance accuracy is better at higher compared to lower levels of intensity, and in tasks using dynamic instead of static stimuli), these procedural factors are unlikely to significantly impact the detection of a facial emotion processing effect in BD (as evidenced by the main effects of group, but not interactions across analyses). Thus, it is improbable that inconsistencies evident in the current BD literature are by-products of emotion processing impairments being masked by task related artifacts (as BD performance on all emotion processing tasks was impaired here). Rather, they may be a function of other factors such as differences amongst study cohorts with regard to clinical history, or the use and dosage of medications. Alternatively, the null effects of past research may represent a result of poor statistical power (see Vaskinn et al., Reference Vaskinn, Sundet, Friis, Simonsen, Birkenæs, Engh and Andreassen2007 who's BD sample comprised only 21 patients).
The present results should be interpreted within the confines of several limitations. First, as emotion processing under time pressure relies on general processing speed which is known to be compromised in BD (see Van Rheenen & Rossell, Reference Van Rheenen and Rossell2013a), it is possible that the effects we have observed here are confounded by generic BD-related cognitive impairments. Second, given that there was overlap in the stimuli used across the different tasks, it is possible that our results are partly attributable to cross-contamination effects whereby responses on earlier trials affected responses on later trials using the same face and facial expression. Third, as abnormalities of disgust and surprised expressions have been demonstrated in some studies (Gray et al., Reference Gray, Venn, Montagne, Murray, Burt, Frigerio and Young2006; Harmer, Grayson, & Goodwin, Reference Harmer, Grayson and Goodwin2002), our omission of these emotions from the battery limited our understanding of how accurately and quickly patients in this cohort were able to process these emotions. Fourth, as we did not explicitly counterbalance the presentation of faces and emotions across visual fields, we cannot account for hemisphere specific laterality effects. Finally, we were unable to directly compare mood subgroups to controls due to the restricted power after stratification into mixed/manic, depressed and euthymic subgroups. Although within group analyses failed to differentiate performance across patients in these current states for all tasks, it is still possible that mood may have had an effect on performance. Thus, given the rather heterogeneous nature of our BD sample, our results should be interpreted with caution.
Nevertheless, as this study is the first of its kind to provide insight into the nature of emotion processing impairments in BD while paying attention to a range of potential confounds including task stimuli designs, intensity, and emotion specific factors in a single experiment, these findings do add substantially to the existing literature on facial emotion processing in BD. Future studies would certainly do well to address the present limitations however, with a view to providing greater clarity with regard to the impact of pre-existing cognitive impairments, cross-contamination effects and mood state related factors on facial emotion processing.
In summary, this study is the first of its kind to comprehensively examine emotion processing performance in a battery that controlled for subtle differences in task stimuli, and investigated the specificity of impairments across task types, emotions, and intensities in persons with BD. Our primary finding of a generalized patient impairment in the ability to label facial expressions and to make use of available emotional facial cues to differentiate them, suggests that facial emotion processing is considerably more challenging for people with the disorder than for those without. This may have direct impact on the significant psychosocial burden carried by patients, although this remains to be seen.
Acknowledgements and Conflicts of Interest
The authors have no conflicts of interest but would like to acknowledge the Australian Rotary Health/Bipolar Expedition, the Helen McPherson Smith Trust and an Australian Postgraduate Award for providing financial assistance for the completion of this work.