Key Points
Question
Does face transplant restore the possibility of facial emotional expression, and can software-based video analysis be used to track progress over time?
Findings
In this case-control study including 6 patients who underwent face transplant, all emotions were detectable, but only expression of happiness was reliably restored to 43% of the level of healthy controls and showed statistically significant improvement 1 year after transplant.
Meaning
Software-based video analysis can be used as an objective, noninvasive, and nonobtrusive method of detecting and tracking facial emotional expression restoration after face transplant.
This case-control study uses software-based video analysis to detect and track restoration of facial emotional expression after face transplant.
Abstract
Importance
Limited quantitative data exist on the restoration of nonverbal communication via facial emotional expression after face transplant. Objective and noninvasive methods for measuring outcomes and tracking rehabilitation after face transplant are lacking.
Objective
To measure emotional expression as an indicator of functional outcomes and rehabilitation after face transplant via objective, noninvasive, and nonobtrusive software-based video analysis.
Design, Setting, and Participants
This single-center case-control study analyzed videos with commercially available video analysis software capable of detecting emotional expression. The study participants were 6 patients who underwent face transplant at Brigham and Women’s Hospital between April 2009 and March 2014. They were matched by age, race/ethnicity, culture, and sex to 6 healthy controls with no prior facial surgical procedures. Participants were asked to perform either emotional expressions (direct evaluation) or standardized facial movements (indirect evaluation). Videos were obtained in a clinical setting, except for direct evaluation videos of 3 patients that were recorded at the patients’ residences. Data analysis was performed from June 2018 to November 2018.
Main Outcomes and Measures
The possibility of detecting the emotional expressions of happiness, sadness, anger, fear, surprise, and disgust was evaluated using intensity score values between 0 and 1, representing expressions that are absent or fully present, respectively.
Results
Six patients underwent face transplant (4 men; mean [SD] age, 42 [14] years). Four underwent full face transplants, and 2 underwent partial face transplants of the middle and lower two-thirds of the face. In healthy controls, happiness was the only emotion reliably recognized in both indirect (mean [SD] intensity score, 0.92 [0.05]) and direct (mean [SD] intensity score, 0.91 [0.04]) evaluation. Indirect evaluation showed that expression of happiness significantly improved 1 year after transplant (0.04 point per year; 95% CI, 0.02 to 0.06 point per year; P = .002). Expression of happiness was restored to a mean of 43% (range, 14% to 75%) of that of healthy controls after face transplant. The expression of sadness showed a significant change only during the first year after transplant (−0.53 point per year; 95% CI, −0.82 to −0.24 point per year; P = .005). All other emotions were detectable with no significant change after transplant. Nearly all emotions were detectable in long-term direct evaluation of 3 patients, with expression of happiness restored to a mean of 26% (range, 5% to 59%) of that of healthy controls.
Conclusions and Relevance
Partial restoration of facial emotional expression is possible after face transplant. Video analysis software may provide useful clinical information and aid rehabilitation after face transplant.
Introduction
Face transplant is a viable reconstructive option for patients with severe facial deformity that shows promising long-term results in improving functionality and quality of life.1 Outcome measures of face transplant have traditionally assessed the recovery of vital functions (eg, ability to breathe,2 eat, and speak3) and independent functions (eg, motor movement and protective and discriminative sensation),4,5,6,7 as well as the procedure’s functional psychological impact on quality of life and mental health.8,9,10 Measuring the restoration of these functions is necessary to determine the value of face transplant to the individual patient, but their recovery alone is not sufficient to achieve or explain societal reintegration after face transplant.
Nonverbal communication via facial emotional expression, a social function of the face, has evolved under the pressures of interacting in a social environment.11 Six specific emotional expressions—happiness, sadness, anger, surprise, fear, and disgust—are recognized across cultures and are the focus of social psychology research.12,13,14 Despite their high relevance, limited quantitative data are available on the restoration of facial emotional expression after face transplant. Existing evidence comes from methods such as facial surface electromyography,15 which is sensitive but requires painstaking placement of several electrodes on the skin,16 and appearance-based facial feature extraction, which is similar to facial recognition technology but requires significant data processing that limits reproducibility.17 These methods are obtrusive and prone to human instrumentation error. Their clinical implementation would be time-consuming and would bind patients to laboratory settings, which could affect medical adherence over time.18 The need to find a less obtrusive and more reliable method for evaluating emotional expression as an outcome measure of face transplant remains.
Software-based video analysis, a merger of facial recognition technology and deep learning, has proven to be capable of assessing facial motor movement functions after face transplant.19 We used a commercially available video analysis software that automatically analyzes facial movement for emotional expression to evaluate recovery of social functions after face transplant because it has been shown to do so in a manner similar to that of an objective human observer.20 Simultaneously, this method remains unobtrusive and capable of producing standardized measurements to track an individual’s rehabilitation progress or for group comparison.
We hypothesize that face transplant restores the possibility of emotional expression because it restores both the face’s human aesthetic and its underlying musculature, allowing for nonverbal communication perceivable to human observers. We believe that quantitative evaluation of emotional expression recovery in patients with face transplants could provide another objective outcome measure of face transplant. To our knowledge, this is the first study to detect and track facial emotional expression in patients with face transplants via an objective, noninvasive, and nonobtrusive method.
Methods
Study Participants
This study was approved by the institutional review board of the Partners Human Research Committee. Written informed consent was obtained from all study participants. This study follows the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) reporting guideline for case-control studies.
This retrospective case-control study was performed using 44 videos from 6 patients with face transplants (representing 15% of patients with face transplants worldwide) taken at regular intervals over a maximal posttransplant period of 9.5 years. Also used were 12 videos from 6 healthy controls who were matched according to the age of the donor and sex and cultural ethnicity of the recipient and who had no history of previous reconstructive or cosmetic facial procedures.
Study Tools
Videos were acquired using a commercially available camera (EOS Rebel T3i; Canon) and tripod. Analysis was based on conventional video formats; thus, no special equipment or extra processing was required. FaceReader facial expression recognition software version 6.1 (Noldus) was used to detect and track faces, extract facial features, and analyze facial expressions.21 The software uses the Viola-Jones cascaded classifier algorithm22 to identify facial features and create a neutral face state. Then using the Active Appearance Method,23 an artificial face model is synthesized to compare vector variations between baseline and simulated faces with a database of annotated images.
The video analysis software achieves this by relying on the Facial Action Coding System,24 which taxonomizes visibly different facial movements on the basis of underlying anatomical structures into individual action units. For example, when people produce the prototypical facial expression for happiness, their cheeks raise (action unit 6) and the commissures of their mouth are pulled laterally (action unit 12). These facial expressions performed in concert effectively increase the width of the lower two-thirds of the face and may increase the distance from the chin to the brows, depending on the intensity of expression (Figure 1). The resultant smile is detectable and perceivable as the emotional expression for happiness, objectively, by untrained human observers25 and trained video analysis software.26
Figure 1. Action Units and Their Facial Regions.
The simple facial action units, as first described by Ekman et al,12 are presented with the muscles responsible for their motor movement. We assigned them into regions of the face commonly amenable to restoration by face transplant. Notably, there are no action units 3, 8, 19, and 21. Action units 25 and 26 are not represented visually because they are dependent on muscle relaxation. Action unit 27 is not represented visually because the force vector for the pterygoid muscles are not in the same plane of the image. Image by visual artist Coralie Vogelaar, 2018, used with permission.
The video analysis software determines the magnitude of vector variation between neutral and simulated facial expressions using a trained artificial neural network27 and then compares them with prototypical features of 6 basic emotions to produce an intensity score for each. This intensity score value ranges from 0 to 1, depending on whether the emotional expression is entirely absent or fully present, respectively. According to the FaceReader software manufacturer, intensity score values greater than 0.5 are detectable by objective human observers.21
Study Design
All study participants were recorded performing commands from 2 different protocols to either indirectly or directly evaluate emotional expression. For indirect evaluation, patients with face transplants were filmed every 6 months after transplant. All study participants performed a series of 12 facial movements: smile, frown, purse lips, open mouth wide, shut mouth tight, open eyes wide, close eyes tight, wrinkle nose, pucker lips, wink with right eye, wink with left eye, and puff cheeks. For direct evaluation, all healthy controls and 3 patients with face transplants (patient 1 at 9.5 years, patient 4 at 7.5 years, and patient 5 at 5.5 years) performed a series of 6 simulated faces for when they feel happy, sad, angry, surprised, scared, and disgusted. For both protocols, all study participants were asked to return to their neutral resting face between commands. Each video was less than 2 minutes long, and we attempted to standardize the background and lighting implemented.
Video analysis was performed after individual video calibration to correct for participant-specific biases toward certain facial expressions. Baseline emotional expressions were set to 0 (ie, intensity score equal to 0) using the first neutral resting face in each video. For indirect evaluation, maximum intensity score values for each emotion were used, and the possibility of expression was verified by correlating the protocol command, with correspondent action units, to the emotional state detected for consistency. For direct evaluation, the maximal intensity score values were extracted from the video sequence dedicated to the performed emotion. Data from all study participants were used for analysis of happiness. For all other emotions, patients with partial face transplants were excluded because not all action units necessary for the emotional expression were transplanted. For indirect evaluation, the highest intensity score value after the first year was chosen for each patient with face transplant to allow comparison with healthy controls.
Statistical Analysis
To study changes in emotional expression over time in patients with face transplants, a piecewise linear regression model with random slope and intercept was fitted to the data with a knot at 1 year after transplant to allow for changes in the slope. The knot location was chosen because previous scientific literature28,29 reported that motor recovery improvements occurred mostly during the first year after transplant and because quantitative data beyond that time frame are mostly lacking. Statistical significance of the model was calculated from the comparison with hypothetical model, with a slope of 0. Two-sided P values less than .05 were considered statistically significant and were calculated using the exact sum-of-squares F test. The continuous parametric variables are presented as mean (95% CI) or mean (SD). All statistical analysis was performed using Prism statistical software version 8.02 (GraphPad Software). Data analysis was performed from June 2018 to November 2018.
Results
Study Participant Characteristics
Six patients underwent face transplant (4 men; mean [SD] age, 42 [14] years). Four patients received all action unit regions after full face transplant,30,31,32,33 whereas the 2 patients with partial face transplants received all and nearly all action units of the entire middle and lower two-thirds of the face (Table 1). The mean (SD) donor allograft age and mean healthy control age were both 48 (10) years. All patients underwent pertinent facial nerve and facial sensory nerve neurorrhaphies, bilaterally, with 3 patients requiring intraoperative nerve grafts, 1 of which required a revision nerve transfer (Table 1).
Table 1. Characteristics of Patients With Face Transplants and Healthy Controls.
Patient No. | Date of Transplant | Age and Sex | Indication | Allograft Structures | Facial Nerve Neurorrhaphies | Sensory Neurorrhaphies | Nerve Transfers and Grafts | Transplanted Action Unit Regions | |||
---|---|---|---|---|---|---|---|---|---|---|---|
Donor | Recipient | Healthy Controls | Intraoperative | Revision | |||||||
1 | April 2009 | Early 60s, male | Late 50s, male | Early 60s, male | High-voltage burn | Nose, cheeks, upper lip, midface bone5 | 5 Branches of the facial nerve, bilaterally30 | Bilateral infraorbital and buccal nerves30 | NA | NA | Midface plus lower two-thirds (except action unit 17) |
2 | March 2011 | Late 40s, male | Mid 20s, male | Late 40s, male | High-voltage burn | Full face, partial scalp31 | Only upper and lower divisions of the facial nerve on the left; frontal, zygomatic, buccal and marginal mandibular branches on the right31 | Left mental nerve; right supraorbital, infraorbital, and mental nerves31 | Recipient thoracodorsal nerve graft for superior and inferior divisions of facial nerve on the left, and marginal mandibular on the right7 | NA | All |
3 | April 2011 | Early 30s, male | Early 30s, male | Early 30s, male | High-voltage burn | Full face | Buccal and marginal mandibular branches of the facial nerve, bilaterally31 | Bilateral supraorbital, infraorbital, and mental nerves31 | NA | NA | All |
4 | May 2011 | Early 40s, female | Late 50s, female | Late 40s, female | Animal attack | Full face, partial scalp, midface bone | 6 Branches of the facial nerve, bilaterally, including frontal, zygomatic, buccal, and marginal mandibular31 | Bilateral supraorbital, supratrochlear, and mental nerves31 | Recipient great auricular nerve graft for 2 inferior buccal branches of facial nerve on the left7 | 11 mo after surgery; masseter to facial nerve transfer with recipient great auricular nerve interposition graft on the right7 | All |
5 | February 2013 | Mid 50s, female | Mid 40s, female | Early 50s, female | Chemical burn | Full face | 5 Branches of the facial nerve, bilaterally32 | Bilateral supraorbital, buccal, and mental nerves32 | Nerve grafts had to be used on the left | NA | All |
6 | March 2014 | Early 50s, male | Late 30s, male | Late 40s, male | Self-inflicted gunshot wound | Nose, lower two-thirds, maxilla, mandible | Buccal and marginal mandibular branches of the facial nerve, bilaterally33 | Infraorbital nerves, bilaterally33 | NA | NA | Midface plus lower two-thirds |
Abbreviation: NA, not applicable.
Evaluation of Healthy Controls
The healthy control videos were analyzed first to validate the sensitivity and specificity of both protocols and the video analysis software. Only the emotion of happiness could be reliably detected, with mean (SD) intensity score values of 0.92 (0.05) for indirect evaluation and 0.91 (0.04) for direct evaluation (Table 2 and eFigure 1 in the Supplement). All other emotions were detectable, but mean intensity score values did not pass the threshold for objective observer detection during both indirect and direct evaluation (eFigure 1 in the Supplement).
Table 2. Evaluation of Emotional Expressiona.
Emotional Expression | Action Units Involved | Patients With Face Transplant, Mean (SD), Intensity Score Value | Healthy Controls, Mean (SD), Intensity Score Value | ||
---|---|---|---|---|---|
Indirect Evaluation (n = 6) | Direct Evaluation (n = 3) | Indirect Evaluation (n = 6) | Direct Evaluation (n = 6) | ||
Happiness | 6 + 12 | 0.38 (0.24) | 0.24 (0.26) | 0.92 (0.05) | 0.91 (0.04) |
Sadness | 1 + 4 + 15 | 0.34 (0.16) | 0.13 (0.11) | 0.44 (0.28) | 0.34 (0.31) |
Anger | 1 + 2 + 5 + 26 | 0.17 (0.21) | 0 | 0.33 (0.16) | 0.28 (0.21) |
Surprise | 1 + 2 + 4 + 5 + 7 + 20 + 26 | 0.28 (0.23) | 0.22 (0.37) | 0.47 (0.14) | 0.19 (0.14) |
Fear | 4 + 5 + 7 + 23 | 0.24 (0.16) | 0.06 (0.09) | 0.23 (0.16) | 0.11 (0.15) |
Disgust | 9 + 15 + 16 | 0.09 (0.10) | 0.02 (0.02) | 0.34 (0.26) | 0.19 (0.25) |
The results for the indirect and direct evaluation of emotional expression in patients with face transplants and healthy controls are displayed. For each emotion, their corresponding action units are included and results display the mean (SD) intensity score values. For indirect evaluation, the highest intensity score value after the first year was chosen for each patient with a face transplant to allow comparison with healthy controls.
Indirect Evaluation of Patients With Face Transplant
We found that the emotional expression of happiness, sadness, anger, surprise, fear, and disgust was possible after face transplant with nonzero intensity score values detectable in all patients with face transplants. The mean (SD) group intensity score values were 0.38 (0.24) for happiness, 0.34 (0.16) for sadness, 0.17 (0.21) for anger, 0.28 (0.23) for surprise, 0.24 (0.16) for fear, and 0.09 (0.10) for disgust (Table 2). To determine percentage recovery, maximum intensity score values for each patient were compared with mean intensity score values of healthy controls for every emotion. This made recovery greater than 100% of the level of healthy controls possible for individual patients with face transplants. Happiness expression after face transplant was found to be restored to 43% (range, 14%-75%) of that of healthy controls (Figure 2), with other emotions compared in similar fashion but yielding unreliable results (eFigure 2 in the Supplement).
Figure 2. Restoration of Happiness Expression After Face Transplant Compared With Healthy Controls.
For indirect evaluation of expression of happiness, the maximum intensity score values of each patient with face transplant after the first posttransplant year were compared with the mean intensity score values of healthy controls. We found that expression of happiness, based on the recovery of the ability to smile (action units 6 + 12), was restored to a mean of 43% of that of healthy controls (dashed line) in the first 5 years after transplant.
Intensity score values for indirect evaluation of patients with face transplants were tracked over time. During posttransplant year 1, the intensity score values for happiness decreased nonsignificantly by 0.06 point per year (95% CI, −0.34 to 0.23 point per year; P = .66). Afterward, intensity score values for happiness increased significantly by 0.04 point per year (95% CI, 0.02 to 0.06 point per year; P = .002) (Figure 3). The intensity score values for sadness decreased significantly by 0.53 point during posttransplant year 1 (95% CI, −0.82 to −0.24 point per year; P = .005), with negligible changes afterward (0.01 point per year; 95% CI, −0.01 to 0.03 point per year; P = .48) (Figure 3). Individual patient trends for happiness and sadness are displayed in eFigure 3 in the Supplement. The remaining emotions of anger, surprise, fear, and disgust had intensity score values with nonsignificant changes (P > .05) after transplant (eFigure 4 in the Supplement).
Figure 3. Longitudinal Evaluation of Expression of Happiness and Sadness After Face Transplant.
The intensity score values for expression of happiness and sadness during longitudinal indirect evaluation were modeled using piecewise linear regression with a knot at posttransplant year 1 (dashed lines). A, Expression of happiness was found to increase significantly by 0.04 point per year (95% CI, 0.02 to 0.06 point per year; P = .002) after the knot at year 1 (before year 1, −0.06 point per year; 95% CI, −0.34 to 0.23 point per year). B, The intensity score values for expression of sadness decreased significantly by 0.53 point per year in posttransplant year 1 (95% CI, −0.82 to −0.24 point per year; P = .005), but afterward the change was negligible (0.01 point per year, 95% CI, −0.01 to 0.03 point per year; P = .48).
Long-term Direct Evaluation of Patients With Face Transplant
Three of 6 patients with face transplants were able to be filmed for the direct evaluation protocol. The mean (SD) group intensity score values were 0.24 (0.26) for happiness, 0.13 (0.11) for sadness, 0 for anger, 0.22 (0.37) for surprise, 0.06 (0.09) for fear, and 0.02 (0.02) for disgust (Table 2). On the basis of limited cohort data, expression of happiness was restored to a mean of 26% (range, 5%-59%) of that of healthy controls in the long term (eFigure 5 in the Supplement).
Discussion
Facial recognition technology has been proposed as a method to improve performance metrics in vascularized composite allotransplant.34 In this study, we show that using software-based video analysis to detect and track nonverbal communication via facial emotional expression after face transplant is feasible. This method is noninvasive, nonobtrusive, able to be implemented widely, and capable of producing objective intensity score values that are amenable to standardization. All 6 basic emotions—happiness, sadness, anger, surprise, fear, and disgust—were detectable with nonzero intensity score values in patients with face transplants during indirect evaluation. This finding suggests that restoration of functional human aesthetic and underlying facial musculature after face transplant surpasses the threshold necessary for objective perception by software.
A prior study19 using software-based analysis showed that smile significantly improves over time after face transplant and is comparable with that of the healthy population, consistent with reported outcomes from most other face transplant teams.5 The current study found that the potential to express happiness after face transplant recovers to 43% of that of healthy controls in the first 5 years; after year 1, expression of happiness significantly improves by 0.04 intensity score point per year. These findings are significant because they are the first objective values on smile restoration after face transplant that can be standardized and tracked over time. Despite having a mean intensity score value less than 0.50, or below the threshold for recognition by objective human observers, happiness may be the only valid and reliable marker after face transplant. This emotional expression is uniquely and reliably recognized in healthy controls with high specificity and sensitivity. Furthermore, given that the video analysis software recognizes happiness similarly to objective human observers with accuracy greater than 90%,25,35,36,37 its ability to detect happiness after face transplant at subclinical levels not perceptible to human observers highlights the potential of video analysis software as a rehabilitative tool.
Although reliable interpretation is possible for happiness only, sadness was the only other expression that significantly changed after transplant. We believe that the high intensity score values before year 1 are the result of tissues drooping because of their weight and from incomplete recovery of muscle tone. Increased intensity score values for sadness in all patients with face transplants before calibration before year 1 further support this theory. Given no significant change in sadness expression thereafter, the theory that neuromuscular recovery after year 1 creates a true baseline expression of the new face is supported.
Both the present study and our previous study19 using facial recognition technology independently validate our findings that mean motor function at 5 years after transplant reaches 60% of maximal possible recovery.1 Currently, we evaluate motor recovery using the Daniels and Worthingham manual muscle testing technique,38 which is the standard for measurement but is liable to subjective evaluation, time-consuming, and exhausting for patients. Using software-based video analysis to measure underlying muscle recovery indirectly via emotional expression could address these limitations and provide novel information about a patient’s recovery. In essence, this tool measures the return of recognizable human aesthetic and nonverbal communication, both social functions of the face, in an unbiased manner. It can objectively provide data helpful to determine whether face transplant meets its ethical goal of restoring functions necessary for societal reintegration.
Some studies have attempted to measure emotional expression after face transplant. Topçu et al15 found that the frequency and spatial distribution of facial surface electromyography data recorded during emotional expression were significantly different between patients with face transplants and healthy controls after 2 years. One of their 3 patients with full face transplants had high-frequency firing distribution similar to that of healthy controls for happiness, whereas other expressions had similar distribution patterns but lower frequency. This could be due to limitations in electromyography measurement of emotional expression.39 Supporting this claim are findings from De Letter et al40 showing that electromyography detected signs of remyelinization after face transplant without clinically meaningful return of function. Bedeloglu et al17 acknowledged this limitation, performed image-based analysis of emotional expression in 2 patients with full face transplants, and found that happiness could be detected to 45% of that of healthy controls after 3 years. Their methods, despite being similar to processes required for software analysis and yielding results comparable to ours, require trained interpretation of the data. The video analysis software in our study has sensitivity on par with that of electromyography,41 but because it depends on visual data necessary for human visual system processing,26,42 clinical specificity for emotional expression detection is greater. Reliability of the tool improves with optimization of lighting and background, and when it is combined with easily interpretable intensity score values amenable to standardization and reduced sensitivity to human instrumentation error, the argument for video analysis software as a clinical assessment tool is stronger than that for electromyography.
Future research should prioritize face transplant rehabilitation programs. Topçu et al43 observed gradual recovery of emotional expressions after rehabilitation with functional electrical stimulation in 2 patients with full face transplants after 3 years. Their findings are subject to the same pitfalls of using electromyography and would make long-term comparison unreliable given the high likelihood of human instrumentation error. Incorporating software-based video analysis into posttransplant rehabilitation could allow for both intensity score value tracking and personalized rehabilitation goals based on expected face transplant cohort data. A second research focus should seek to improve on the Cleveland Clinic FACES Scoring System for Face Transplant Candidate Evaluation.44,45 Development of a specific functional outcome scale for face transplant should rely on objective measures of vital, independent, psychological, and social functions of the face. Software-based video analysis could provide objective intensity score values to standardize a grading system for the recovery of emotional expression potential after face transplant. Another area for implementing software-based video analysis is the detection of allograft rejection. Visible redness and swelling of the facial allograft are associated with rejection episodes within the first 2 years.46 These may be features amenable to software detection that could provide objective data to supplement more-specific biomarkers of rejection.47
Limitations
Motivation of patients to perform facial expressions and their comfort in expressing emotions is a limitation of this study. Rehabilitation after face transplant is strenuous because it burdens the patient with numerous, lengthy appointments that can result in follow-up fatigue. This could affect participation in research and may be responsible for intensity score variability within the same patient and between patients. Software-based video analysis could address follow-up fatigue by reducing the need for long-distance travel to the transplant hospital.48 Videos could be filmed at a patient’s home, analyzed, and then shared with the transplant team, as was done for long-term direct evaluation in patients with face transplants for this study. This could facilitate a more honest appreciation of functional recovery outside of a laboratory setting,49 aid rehabilitation,50 and, on the basis of solid-organ transplantation data, may improve medical adherence and outcomes while decreasing the economic cost after transplant.51
Objective evaluation of emotional expressions after face transplant is a challenge because of cultural, regional,52 and personal variance in expression and presents another limitation to this study. The deep learning algorithm in the video analysis software is trained on images displaying prototypical emotions specific to Western cultures, with exaggerated facial expressions that may not depict realistic expression by individuals. It is possible that not every study participant performed all prototypical features while making facial emotional expressions under direct evaluation. This could explain why intensity score values for healthy controls under direct evaluation demonstrated large variability due to personal differences in expressivity. These biases could also explain why only happiness was reliably recognized in healthy controls, because its recognition is dependent on a smile that requires 2 action units to perform. More-complex emotional expressions, such as fear, require a greater number of units firing in concert to be detected, allowing for greater likelihood of personal and cultural biases to affect individual emotional expression.
Along these lines, new research shows that Eastern and Western cultures appreciate facial emotional expressions differently,53,54,55 which suggests that universality in expression relies more on valence, arousal, and dominance.56,57 We believe our findings support this latter point in 2 ways. Happiness and sadness (positive and negative valence, respectively) were the only emotions that were detected with statistical significance and were the 2 emotions with the largest mean intensity score values under direct evaluation in healthy controls. Interestingly, for all study participants, the mean intensity score values were greater for indirect evaluation of every emotional expression than for direct evaluation while maintaining similar variability. This could be explained by decreased effort and recruitment of motor units necessary to activate single action units. However, this would only explain these findings for simple emotional expressions, such as happiness and sadness, but not for more complex emotional expressions, such as anger, surprise, fear, and disgust, which require multiple action units firing simultaneously to be detected. This suggests that action units necessary for emotional expression detection are functional after face transplant and cultural bias in emotional expression was mitigated under indirect evaluation.
There is evidence that face transplant restores the potential to perform a function important for social integration,58 one that provides clues on biological and mental health.59,60 Although our results are reflective of only 15% of patients with face transplants worldwide, future collaborative studies between face transplantation centers should validate our results. If and when face transplantation occurs in Eastern societies, software will have to be trained on a more culturally appropriate set of images to accurately detect emotional expression as an outcome.
Conclusions
This study demonstrates the potential for quantitative evaluation of emotional expression after face transplant to complement existing outcome measures. We believe that vital and independent functions restore the hardware of the face, with psychological and social functions acting as its necessary software. The use of video analysis software—an objective, noninvasive, and nonobtrusive method—to detect and track emotional expression in patients with face transplants provides novel evidence supporting the procedure’s viability as a treatment modality for severe facial deformity. Results from our case-control study suggest that partial restoration of facial emotional expression is possible after face transplant. The results indicate the potential for video analysis software to provide useful clinical information and aid rehabilitation after face transplant.
eFigure 1. Emotional Expression in Healthy Controls
eFigure 2. Comparison of Emotions
eFigure 3. Individual Patient Trends of Longitudinal Evaluation of Happiness and Sadness After Face Transplantation
eFigure 4. Longitudinal Evaluation of Emotions After Face Transplantation
eFigure 5. Long-term Comparison of Happiness
References
- 1.Tasigiorgos S, Kollar B, Turk M, et al. Five-year follow-up after face transplantation. N Engl J Med. 2019;380(26):-. doi: 10.1056/NEJMc1810468 [DOI] [PubMed] [Google Scholar]
- 2.Fischer S, Wallins JS, Bueno EM, et al. Airway recovery after face transplantation. Plast Reconstr Surg. 2014;134(6):946e-954e. doi: 10.1097/PRS.0000000000000752 [DOI] [PubMed] [Google Scholar]
- 3.Grigos MI, LeBlanc É, Hagedorn C, Diaz-Siso JR, Plana N, Rodriguez ED. Changes in articulatory control pre- and post-facial transplant: a case report. J Speech Lang Hear Res. 2019;62(2):297-306. doi: 10.1044/2018_JSLHR-S-18-0147 [DOI] [PubMed] [Google Scholar]
- 4.Petruzzo P, Dubernard JM. The International Registry on Hand and Composite Tissue Allotransplantation. Clin Transpl. 2011;247-253. [PubMed] [Google Scholar]
- 5.Fischer S, Kueckelhaus M, Pauzenberger R, Bueno EM, Pomahac B. Functional outcomes of face transplantation. Am J Transplant. 2015;15(1):220-233. doi: 10.1111/ajt.12956 [DOI] [PubMed] [Google Scholar]
- 6.Siemionow M. The decade of face transplant outcomes. J Mater Sci Mater Med. 2017;28(5):64. doi: 10.1007/s10856-017-5873-z [DOI] [PubMed] [Google Scholar]
- 7.Aycart MA, Perry B, Alhefzi M, et al. Surgical optimization of motor recovery in face transplantation. J Craniofac Surg. 2016;27(2):286–292. doi: 10.1097/scs.0000000000002305 [DOI] [PubMed] [Google Scholar]
- 8.Furr LA, Wiggins O, Cunningham M, et al. Psychosocial implications of disfigurement and the future of human face transplantation. Plast Reconstr Surg. 2007;120(2):559-565. doi: 10.1097/01.prs.0000267584.66732.e5 [DOI] [PubMed] [Google Scholar]
- 9.Aycart MA, Kiwanuka H, Krezdorn N, et al. Quality of life after face transplantation: outcomes, assessment tools, and future directions. Plast Reconstr Surg. 2017;139(1):194-203. doi: 10.1097/PRS.0000000000002890 [DOI] [PubMed] [Google Scholar]
- 10.Nizzi M-C, Tasigiorgos S, Turk M, Moroni C, Bueno E, Pomahac B. Psychological outcomes in face transplant recipients: a literature review. Curr Surg Rep. 2017;5(10):26. doi: 10.1007/s40137-017-0189-y [DOI] [Google Scholar]
- 11.Darwin C. The Expression of the Emotions in Man and Animals. 3rd ed London, UK: Fontana Press; 1999. [Google Scholar]
- 12.Ekman P, Sorenson ER, Friesen WV. Pan-cultural elements in facial displays of emotion. Science. 1969;164(3875):86-88. doi: 10.1126/science.164.3875.86 [DOI] [PubMed] [Google Scholar]
- 13.Ekman P. Are there basic emotions? Psychol Rev. 1992;99(3):550-553. doi: 10.1037/0033-295X.99.3.550 [DOI] [PubMed] [Google Scholar]
- 14.Ekman P. What scientists who study emotion agree about. Perspect Psychol Sci. 2016;11(1):31-34. doi: 10.1177/1745691615596992 [DOI] [PubMed] [Google Scholar]
- 15.Topçu Ç, Uysal H, Özkan Ö, et al. Assessment of emotional expressions after full-face transplantation. Neural Plast. 2017;2017:8789724. doi: 10.1155/2017/8789724 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Mallat S. Wavelet packet and local cosine bases In: Mallat S, ed. A Wavelet Tour of Signal Processing: The Sparse Way. 3rd ed Cambridge, UK: Academic Press; 2008:377-431. [Google Scholar]
- 17.Bedeloglu M, Topçu Ç, Akgul A, et al. Image-based analysis of emotional facial expressions in full face transplants. J Med Syst. 2018;42(3):42. doi: 10.1007/s10916-018-0895-8 [DOI] [PubMed] [Google Scholar]
- 18.Tasigiorgos S, Dorante MI, Devine E, et al. American Society for Reconstructive Transplantation Conference Abstracts 2018: can we predict adherence in face transplantation recipients? the Brigham and Women’s Hospital experience. SAGE Open Med. 2018;6:42-43. doi: 10.1177/2050312118808661 [DOI] [Google Scholar]
- 19.Fischer S, Diehm Y, Dorante MI, et al. Software-based video analysis of functional outcomes of face transplantation. Microsurgery. 2019;39(1):53-61. doi: 10.1002/micr.30360 [DOI] [PubMed] [Google Scholar]
- 20.Wegrzyn M, Vogt M, Kireclioglu B, Schneider J, Kissler J. Mapping the emotional face: how individual face parts contribute to successful emotion recognition. PLoS One. 2017;12(5):e0177239. doi: 10.1371/journal.pone.0177239 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Noldus Information Technology FaceReader version 6.1, reference manual. https://www.noldus.com/human-behavior-research/products/facereader. Published 2015. Accessed November 26, 2019.
- 22.Viola P, Jones MJ. Robust real-time face detection. Int J Comput Vis. 2004;57(2):137-154. doi: 10.1023/B:VISI.0000013087.49260.fb [DOI] [Google Scholar]
- 23.Cootes TF, Taylor CJ Statistical models of appearance for computer vision. http://www.face-rec.org/algorithms/AAM/app_models.pdf. Published March 8, 2004. Accessed November 26, 2019.
- 24.Ekman P, Friesen WV. Facial Action Coding System: A Technique for the Measurement of Facial Movement. Sunnyvale, CA: Consulting Psychologists Press; 1978. [Google Scholar]
- 25.Terzis V, Moridis CN, Economides AA Measuring instant emotions during a self-assessment test: the use of FaceReader. In: Proceedings of the 7th International Conference on Methods and Techniques in Behavioral Research Eindhoven, the Netherlands: ACM Press; 2010:1-4. doi: 10.1145/1931344.1931362 [DOI] [Google Scholar]
- 26.Martinez A, Du S. A model of the perception of facial expressions of emotion by humans: research overview and perspectives. J Mach Learn Res. 2012;13:1589-1608. [PMC free article] [PubMed] [Google Scholar]
- 27.Bishop CM. Neural Networks for Pattern Recognition. Oxford, UK: Oxford University Press; 1995. [Google Scholar]
- 28.Khalifian S, Brazio PS, Mohan R, et al. Facial transplantation: the first 9 years. Lancet. 2014;384(9960):2153-2163. doi: 10.1016/S0140-6736(13)62632-X [DOI] [PubMed] [Google Scholar]
- 29.Rifkin WJ, David JA, Plana NM, et al. Achievements and challenges in facial transplantation. Ann Surg. 2018;268(2):260-270. doi: 10.1097/SLA.0000000000002723 [DOI] [PubMed] [Google Scholar]
- 30.Pomahac B, Lengele B, Ridgway EB, et al. Vascular considerations in composite midfacial allotransplantation. Plast Reconstr Surg. 2010;125(2):517-522. doi: 10.1097/PRS.0b013e3181c82e6f [DOI] [PubMed] [Google Scholar]
- 31.Pomahac B, Pribaz J, Eriksson E, et al. Three patients with full facial transplantation. N Engl J Med. 2012;366(8):715-722. doi: 10.1056/NEJMoa1111432 [DOI] [PubMed] [Google Scholar]
- 32.Perry BJ, Richburg BD, Pomahac B, Bueno EM, Green JR. The effects of lip-closure exercise on lip strength and function following full facial transplantation: a case report. Am J Speech Lang Pathol. 2017;26(2):682-686. doi: 10.1044/2017_AJSLP-16-0101 [DOI] [PubMed] [Google Scholar]
- 33.Kiwanuka H, Aycart MA, Gitlin DF, et al. The role of face transplantation in the self-inflicted gunshot wound. J Plast Reconstr Aesthet Surg. 2016;69(12):1636-1647. doi: 10.1016/j.bjps.2016.08.014 [DOI] [PubMed] [Google Scholar]
- 34.Zuo KJ, Saun TJ, Forrest CR. Facial recognition technology: a primer for plastic surgeons. Plast Reconstr Surg. 2019;143(6):1298e-1306e. doi: 10.1097/PRS.0000000000005673 [DOI] [PubMed] [Google Scholar]
- 35.van Kuilenburg H, Wiering M, den Uyl M. A model based method for automatic facial expression recognition In: Gama J, Camacho R, Brazdil PB, Jorge AM, Torgo L, eds. Machine Learning: ECML 2005. Berlin, Germany: Springer; 2005;3720:194-205. doi: 10.1007/11564096_22. [DOI] [Google Scholar]
- 36.den Uyl MJ, van Kuilenburg H The FaceReader: online facial expression recognition. http://www.vicarvision.nl/pub/fc_denuyl_and_vankuilenburg_2005.pdf. Published September 2008. Accessed November 26, 2019.
- 37.Langner O, Dotsch R, Bijlstra G, Wigboldus DHJ, Hawk ST, van Knippenberg A. Presentation and validation of the Radboud Faces Database. Cogn Emotion. 2010;24(8):1377-1388. doi: 10.1080/02699930903485076 [DOI] [Google Scholar]
- 38.Hislop HJ, Avers D, Brown M. Daniels and Worthingham’s Muscle Testing: Techniques of Manual Examination and Performance Testing. 9th ed St. Louis, MO: Elsevier Saunders; 2014. [Google Scholar]
- 39.van Boxtel A. Facial EMG as a tool for inferring affective states. https://www.measuringbehavior.org/files/ProceedingsPDF(website)/Boxtel_Symposium6.4.pdf. Published 2010. Accessed November 26, 2019.
- 40.De Letter M, Vanhoutte S, Aerts A, et al. Facial nerve regeneration after facial allotransplantation: a longitudinal clinical and electromyographic follow-up of lip movements during speech. J Plast Reconstr Aesthet Surg. 2017;70(6):729-733. doi: 10.1016/j.bjps.2017.02.025 [DOI] [PubMed] [Google Scholar]
- 41.D’Arcey T, Johnson M, Ennis M Assessing the validity of FaceReader using facial electromyography. http://www.darcey.us/pdf/facereader.pdf. Published 2012. November 26, 2019.
- 42.Martinez AM. Visual perception of facial expressions of emotion. Curr Opin Psychol. 2017;17:27-33. doi: 10.1016/j.copsyc.2017.06.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Topçu Ç, Uysal H, Özkan Ö, et al. Recovery of facial expressions using functional electrical stimulation after full-face transplantation. J Neuroeng Rehabil. 2018;15(1):15. doi: 10.1186/s12984-018-0356-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Gordon CR, Siemionow M, Coffman K, et al. The Cleveland Clinic FACES Score: a preliminary assessment tool for identifying the optimal face transplant candidate. J Craniofac Surg. 2009;20(6):1969-1974. doi: 10.1097/SCS.0b013e3181bd2c86 [DOI] [PubMed] [Google Scholar]
- 45.Chopra K, Susarla SM, Goodrich D, et al. Clinical application of the FACES score for face transplantation. J Craniofac Surg. 2014;25(1):64-69. doi: 10.1097/SCS.0b013e3182a2dda9 [DOI] [PubMed] [Google Scholar]
- 46.Haug V, Kollar B, Obed D, et al. The evolving clinical presentation of acute rejection in facial transplantation. JAMA Facial Plast Surg. 2019;21(4):278-285. doi: 10.1001/jamafacial.2019.0076 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Kollar B, Shubin A, Borges TJ, et al. Increased levels of circulating MMP3 correlate with severe rejection in face transplantation. Sci Rep. 2018;8(1):14915. doi: 10.1038/s41598-018-33272-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Rifkin WJ, Manjunath A, Kimberly LL, et al. Long-distance care of face transplant recipients in the United States. J Plast Reconstr Aesthet Surg. 2018;71(10):1383-1391. doi: 10.1016/j.bjps.2018.05.019 [DOI] [PubMed] [Google Scholar]
- 49.Spokas M, Luterek JA, Heimberg RG. Social anxiety and emotional suppression: the mediating role of beliefs. J Behav Ther Exp Psychiatry. 2009;40(2):283-291. doi: 10.1016/j.jbtep.2008.12.004 [DOI] [PubMed] [Google Scholar]
- 50.Nakamura K, Toda N, Sakamaki K, Kashima K, Takeda N. Biofeedback rehabilitation for prevention of synkinesis after facial palsy. Otolaryngol Head Neck Surg. 2003;128(4):539-543. doi: 10.1016/S0194-5998(02)23254-4 [DOI] [PubMed] [Google Scholar]
- 51.Schmid A, Hils S, Kramer-Zucker A, et al. Telemedically supported case management of living-donor renal transplant recipients to optimize routine evidence-based aftercare: a single-center randomized controlled trial. Am J Transplant. 2017;17(6):1594-1605. doi: 10.1111/ajt.14138 [DOI] [PubMed] [Google Scholar]
- 52.Elfenbein HA. Nonverbal dialects and accents in facial expressions of emotion. Emotion Rev. 2013;5(1):90-96. doi: 10.1177/1754073912451332 [DOI] [Google Scholar]
- 53.Masuda T, Ellsworth PC, Mesquita B, Leu J, Tanida S, Van de Veerdonk E. Placing the face in context: cultural differences in the perception of facial emotion. J Pers Soc Psychol. 2008;94(3):365-381. doi: 10.1037/0022-3514.94.3.365 [DOI] [PubMed] [Google Scholar]
- 54.Chen C, Jack RE. Discovering cultural differences (and similarities) in facial expressions of emotion. Curr Opin Psychol. 2017;17:61-66. doi: 10.1016/j.copsyc.2017.06.010 [DOI] [PubMed] [Google Scholar]
- 55.Niedenthal PM, Rychlowska M, Wood A. Feelings and contexts: socioecological influences on the nonverbal expression of emotion. Curr Opin Psychol. 2017;17:170-175. doi: 10.1016/j.copsyc.2017.07.025 [DOI] [PubMed] [Google Scholar]
- 56.Jack RE, Sun W, Delis I, Garrod OG, Schyns PG. Four not six: revealing culturally common facial expressions of emotion. J Exp Psychol Gen. 2016;145(6):708-730. doi: 10.1037/xge0000162 [DOI] [PubMed] [Google Scholar]
- 57.Calvo MG, Nummenmaa L. Perceptual and affective mechanisms in facial expression recognition: an integrative review. Cogn Emot. 2016;30(6):1081-1106. doi: 10.1080/02699931.2015.1049124 [DOI] [PubMed] [Google Scholar]
- 58.Chanes L, Wormwood JB, Betz N, Barrett LF. Facial expression predictions as drivers of social perception. J Pers Soc Psychol. 2018;114(3):380-396. doi: 10.1037/pspa0000108 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Kitayama S, Park J. Emotion and biological health: the socio-cultural moderation. Curr Opin Psychol. 2017;17:99-105. doi: 10.1016/j.copsyc.2017.06.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Baudry A-S, Grynberg D, Dassonneville C, Lelorain S, Christophe V. Sub-dimensions of trait emotional intelligence and health: a critical and systematic review of the literature. Scand J Psychol. 2018;59(2):206-222. doi: 10.1111/sjop.12424 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
eFigure 1. Emotional Expression in Healthy Controls
eFigure 2. Comparison of Emotions
eFigure 3. Individual Patient Trends of Longitudinal Evaluation of Happiness and Sadness After Face Transplantation
eFigure 4. Longitudinal Evaluation of Emotions After Face Transplantation
eFigure 5. Long-term Comparison of Happiness