Abstract
Simulating emotional experience, emotional empathy is the fundamental ingredient of interpersonal communication. In the speaker-listener scenario, the speaker is always a child, the listener is a human or a toy robot. Two groups of neurotypical children aged 6 years on average composed the population: one Japanese (n = 20) and one French (n = 20). Revealing potential similarities in communicative exchanges in both groups when in contact with a human or a toy robot, the results might signify that emotional empathy requires the implication of an automatic identification. In this sense, emotional empathy might be considered a broad idiosyncrasy, a kind of synchronisation, offering the mind a peculiar form of communication. Our findings seem to be consistent with the assumption that children’s brains would be constructed to simulate the feelings of others in order to ensure interpersonal synchronisation.
Keywords: child, emotional empathy, interactor robot, heart rate, synchronisation
Introduction
The term empathy derives from the Ancient Greek word “εμπάθεια” (“εν πάθoς,” i.e., in passion), which has been used to create the German word “Einfühlung” that means “feeling into.” Empathy has been defined as a multifaceted process encompassing emotional and cognitive components (Christov-Moore et al., 2014). In this study, we focus mainly on emotional empathy. In a human-human interactive process, emotional empathy leads one to discern the feelings of another and reflects the matching between “self” and the “other’s” feelings (Kanske, 2018). As such, emotional empathy embraces the similar sharing of the positive or negative emotional state of the other person which can be generated by direct interpersonal communication (Lamm et al., 2017).
Emotional empathy has been shown to be a premise of the synchronisation between children and their mothers. When 2 to 3 years of age, children seem enable to anticipate the emotional states of others (Tramacere and Ferrari, 2016). Developmental references suggest that children aged 6 to 7 years old can recognise and respond empathically to positive and negative emotions of their animate or inanimate companions in complex situations (Winnicott, 1971). As children grow into different environments with contrasting conventions, including emotion, it might make sense that they build different ideas about the meaning that is associated with those conventions. Many options have been utilised to explore interpersonal communication. Whereas contextual factors are thought to be crucial for designing better interpersonal interaction, theories eulogise that children are born with an integrated feeling and expression of their emotional selves in their relationship with the others (Frith and Frith, 2003). In human-human interaction, emotional adjustment between companions is based on synchronisation and relationship (Lischke et al., 2018). Therefore, it has been demonstrated that when people interact verbally and non-verbally face to face, they naturally synchronise their reactions (Llobera et al., 2016; Cornejo et al., 2017).
Beyond interpersonal interaction studies, some cross-educational investigations exist on how children interact with robots. Even so, these investigations are not convincing. Some studies have indicated that when robots were used as companions several aspects of human–human communication were directly replicated in human–robot communication (Audrey, 2009). Some others have shown that the way humans communicate with robots is education-dependent (Castellano et al., 2010; Shahid et al., 2014). Based on observations or on self-report questionnaires, most of the above-mentioned studies were conducted more frequently with adults than with children using anthropomorphic or zoomorphic robots (Mitchell and Hamm, 1997). Only marginal attention was committed to the comparison between human-human and human-robot interaction using toy robots (i.e., non-anthropomorphic or non-zoomorphic robots). Moreover, in the aforementioned cross-educational studies, the question of a synchronisation mechanism between companions (human and/or robots) has never been investigated. However, this question is fundamental when we explore human-robot interaction (Giannopulu, 2016a,b, 2018).
In highlighting the importance of synchronisation, our view is congruent with the neuroconstructivist position. According to this position, emotional development, including emotional empathy, arises from dynamic contextual changes in neural structures leading to construal representations across multiple brain regions (Marschal et al., 2010). As such, these representations not only depend on the neural context but also on the physical context (Cacioppo et al., 2014). The analogy between neural activity during experiences has further motivated the interpretation of emotional empathy as a simulation process, associated with a robust biomarker: the mirror neuron system (Rizzolatti and Craighero, 2008). Neuroscientific evidence indicates that there are suggestive parallels between the emotional experience of “self” and “others” (Lamm et al., 2011). Developing areas such as the amygdala, posterior insula, and ventromedial pre-frontal cortex share emotional empathy in children aged 6 to 7 years old even if they show changes in functionality over a lifespan (Decety and Michalska, 2010; Steinbeis et al., 2015; Tramacere and Ferrari, 2016). Subcortical areas (i.e., midbrain areas), develop in association with these other areas (Fan et al., 2011) emphasising the possibility of automatic and unconscious functioning (Giannopulu and Watanabe, 2015, 2018; Giannopulu et al., 2016, 2018; Giannopulu, 2018). Actually, it has been revealed that emotional empathy is directly associated with an autonomic activity mediated by heart rate (Müller and Lindenberger, 2011). Physiological techniques have therefore reported evidence of heart rate synchronisation between adults partners (Levenson and Gottman, 1985) and mother-child dyads (Feldman et al., 2011). Individuals with a high level of emotional empathy showed high heart rates (Stellar et al., 2015; Lischke et al., 2018) and reported hardly any difficulties in recognising and expressing their own emotional feelings (Panayiotou and Constantinou, 2017).
In the context of an international and interdisciplinary project on human-human and human-robot interaction using a verbal and non-verbal communication between two actors, a speaker and a listener, we investigated emotional empathy in two groups of children: French and Japanese. The speaker was always a neurotypical child, and the listener was a human or a toy robot named “Pekoppa” that reacted to speech sounds by nodding (Watanabe, 2011; Giannopulu, 2016a,b; Giannopulu et al., 2016, 2018). Based on a mathematical construction, the toy robot automatically generated nodding motion from speech input which promoted the synchronisation with the speaker (Watanabe, 2011). Specially designed for human-robot intercommunication, this toy robot is a universal listener. The human in the study was the same in France and in Japan, and carried out the same procedure in both countries. The range of research discussed so far lends implicit support to the inference that the use of a universal synchroniser such as the toy robot “Pekoppa” would lead to similarities in communicative exchanges between French and Japanese children. Namely, in each group, children would discern all communicative signs in terms of their understanding of the given empathetic state behind the behaviour of others: human or robot. Accordingly, we have hypothesised that emotional empathy, which is a mechanism of synchronisation, would lead to potential similarities between companions (human-human and human-robots) both in France and in Japan.
Materials and Methods
Participants
Two groups of 6-year old children participated in the study. Twenty children (10 boys and 10 girls) composed the “French Group”; twenty children (10 boys and 10 girls) composed the “Japanese Group.” The developmental age of the first group ranged from 6 to 7 years old (mean = 6.3 years; s.d = 4 months). The developmental age of the second group ranged from 6 to 7 years old (mean = 6.4 years; s.d = 2.4 months). The children came from the same class both in Paris and Gifu. As reported by their parents and themselves, none of them had previous experience with robots. All the children were healthy. As reported by their teachers, children attended regular schools and had no learning disorders, neurodevelopmental diseases or cardiac or psychiatric problems. Their academic achievements were standard in their schools. The study was approved by the local ethical committee in Paris (Scientific Committee of Individual’s Protection), France and Gifu (Medical Review Board of Gifu University Graduate School of Medicine) Japan and was in accordance with Helsinki convention 2.0. Anonymity was guaranteed. In both countries, parents gave their informed consent both verbally and in writing for the participation of their children in the study as well as for the analysis of the data; however, they did not allow the authors to send out the raw data. In addition, in Paris and Gifu, each child was asked to give his/her verbal consent before the study began.
Robot
An InterActor toy robot, called “Pekoppa,” was used as a listener (Watanabe, 2011). Pekoppa is the simplest expression of Sakura which is a humanoid robot that reacts to speech sounds by only nodding in the same way as humans do. Pekoppa is shaped like a bilobed plant and its leaves and stem make a nodding response based on speech input and supports the sharing of mutual embodiment in communication (see Figure 1). It uses a material called BioMetal made of a shape-memory alloy that acts as its driving force.
Heart Rate Device
A Mio Alpha watch heart-rate monitor was used to record heart rate. It was systematically placed on the left hand of each participant in Paris and Gifu. Mio Alpha measures the heart rate on-line using two green LED and a photo-electric cell. The LEDs are integrated in the back plate of the watch. They project light on the skin, which allows the photo-electric cell to detect the volume of blood flow. The optical sensor displays an accuracy of -01 ± 0.3 bpm. Universal in nature, the device can be used across the lifespan. However, the physiological heart rate limits differ according to the age of individuals. At the age of 6 to 7 years, the heart rate corresponds to 95 bpm (± 30).
Procedure
For both groups, the study took place in a room with which the children were familiar. The room was located in a school context both in Paris and Gifu. We defined three conditions: the first was called “rest condition,” the second was called “with human,” i.e., child-adult, and the third was called “with robot” (i.e., child-robot). The second and third conditions were counterbalanced across the children. The duration of the “rest condition” was 1 min; the second and third conditions each lasted approximately 7 min. The inter-condition interval was approximately 30 s. For each child, the whole experimental session lasted 15 min (see Figure 2).
At the beginning of each session, the experimenter presented the robot to the child explaining that the robot nods whenever the child speaks. Then, the experimenter hid the robot. The session was run as follows: during the “rest condition,” the heart rate of each child was measured in silence. At the end of that condition, the child was also asked to estimate her/his own emotional feeling on a scale ranging from 1 (the lowest level) to 5 (the highest level) (Giannopulu and Sagot, 2010; Giannopulu, 2011, 2013; 2016a; 2016b; Giannopulu and Watanabe, 2014; Giannopulu et al., 2016, 2018). Each level corresponded to a specific emotional state depicted by a child’s face as follows: 1st fair, 2nd moderately good, 3rd good, 4th very good, 5th excellent. During the “with human” condition, the child was invited to talk to the experimenter. To that end, the experimenter asked the child “what did you do at school since this morning.” As such, the experimenter initiated discussion and then listened by only nodding to the child. The heart rate of each child was measured meanwhile. During the “with robot” condition, the robot was set to nod; the experimenter gave the robot to the child and invited the child to use it. As previously, the experimenter asked the child to tell the robot “what he or she did at school since this morning.” The robot was the listener, the child was the speaker and the experimenter remained silent and discreet. The heart rate was recorded at the same time once again. The study started around 2.00pm Parisian and Gifu time for all the children. At the end of the session, the child was invited to estimate his/her own emotion on the same above-mentioned scale. More particularly, each child was asked to report his/her own emotional feeling after contact with the robot (Giannopulu and Watanabe, 2015; Giannopulu, 2016a,b, 2018; Giannopulu et al., 2016, 2018).
Analysis
The heart rate served as the first dependent variable in a 3 (“Rest,” “Human InterActor,” and “Robot InterActor”) × 2 (“French” vs. “Japanese”) mixed-model ANOVA. The emotional feeling reported served as the second dependent variable using the Wilcoxon matched-pairs signed-ranks test. We also performed a statistic of comparisons using the student t-test to examine differences in heart rate and chi-square test to analyse the emotional feeling reported. The obtained results were very similar. We present below the results of ANOVA and Wilcoxon matched-pairs signed-ranks test. The data analysis was performed with SPSS Statistics 24.
Results
First, we present the results for the heart rate of both groups in three conditions: “rest,” “with human,” and “with robot.” We then examine the emotional feeling reported for each group.
As illustrated in Figure 3, the mean heart rate of Japanese children was higher (M = 94.90, s.d = 7.60) than the mean heart rate of French children (M = 89.95, s.d = 5.1) in the “rest condition” (F(1,38) = 5.86, p < 0.05, = 0.1336). Conversely, the mean heart rate of Japanese children (M = 95.56, s.d = 7.87), did not differ from the mean heart rate of French children (M = 98.2, s.d = 9.39) when the InterActor was the human (F(1,38) = 0.93, p = 0.3409, = 0.0238). In the same vein, the mean heart rate of Japanese children (M = 104.06, s.d = 4.08) did not differ from the mean heart rate of French children (M = 105, s.d = 6.97) when the InterActor was the robot (F(1,38) = 0.22, p = 0.6417, = 0.0057). Likewise, the mean heart rate in the “rest condition” did not differ from the mean heart rate in the “with human” condition (F(1,38) = 0.03, p = 0.8633, = 0.0007 for the Japanese and F(1,38) = 3.67, p = 0.0746, = 0.0880 for the French children). However, the mean heart rate in the “with robot” condition was higher than the mean heart rate of the “rest condition” for both groups (F(1,38) = 17.35, p < 0.01, = 0.3134 for the Japanese and F(1,38) = 23.54, p < 0.01, = 0.3825 for the French children). Finally, the heart rate of Japanese children was higher when they interacted with the robot than when they interacted with the human (F(1,38) = 18.04, p < 0.01, = 0.3219), but it did not significantly differ for French children (F(1,38) = 4.01, p = 0.0524, = 0.0954).
Figure 4 shows that the initial emotional state of French and Japanese children did not differ (Mann–Whitney U = 129.5, p = 0.0623). In the same vein, the final emotional state of both groups did not differ (Mann–Whitney U = 167.5, p = 0.3843). The interaction with the InterActor robot did not have any significant effect in the initial emotional state of French and Japanese children (Wilcoxon two-tailed test p > 0.05, T = 45, n = 20 and Wilcoxon two-tailed test p > 0.05, T = 9, n = 20 respectively).
Discussion and Conclusion
This study dealt with emotional empathy as a mechanism of synchronisation, using the same speaker-listener paradigm and the same human (i.e., experimenter) in two groups of children, one French and one Japanese. It revealed that despite a more important non-verbal emotional empathetic expression in Japanese than French children in the rest condition, a similar mechanism of synchronisation characterised non-verbal communicative exchange when both groups of children interacted with a human or a toy robot. When Japanese children interacted with the robot, their heart rate was higher than when they interacted with the human. Moreover, the initial emotional state reported did not differ for Japanese and French children. The interaction with the toy robot did not affect the final emotional state for either groups of children.
Our results are consistent with recent findings that reported significant interdependence between emotional empathy and synchronisation in adults (Levenson and Gottman, 1985; Stellar et al., 2015; Lischke et al., 2018) and in mother-child dyads (Feldman et al., 2011). But these findings also concur with our hypothesis that emotional empathy which is a mechanism of synchronisation, leads to potential similarities between two different groups of children: French and Japanese. This appears to be reflected in the adjustment of a given automatic physiological state: the heart rate. Physiologically speaking, the heart rate is automatically controlled by both the sympathetic nervous system (SNS) and the parasympathetic nervous (PNS) system of the autonomic nervous system (ANS) and provides a measure of autonomic functioning (i.e., unconscious functioning) (Porges, 2007). The PNS is activated during rest in order to maintain homeostasis; the SNS is activated during periods of perceived change by increasing, the heart rate and mobilising emotional functioning (Suurland et al., 2016). Both physiological systems act in complementary ways to respond and adapt to internal and external changes; namely, both systems are based on synchronisation. Note that the SNS is controlled by the spinal cord, the PNS is controlled by the spinal cord and the brain. Ostensibly, in the rest condition, the heart rate of Japanese children was higher than the heart rate of French children. This automatic activity would provide support for the engagement of the children and would indicate a given emotional state. When Japanese children were in contact with the human or the toy robot, their heart rate was similar to the heart rate of French children.
Contrary to data claiming that autonomic functions in the cardiovascular system largely depend on genetic factors (Tanaka et al., 1994), the present study indicates that these functions seem more likely to rely upon inter communication that is possible via a human or a toy robot in our situation. Essentially, both groups of children expressed potentially similar profiles of heart rate in all conditions, with the exception of a higher heart rate in interaction with the robot than in interaction with the human for the Japanese children. A similar profile was observed for the French children even if the data are not statistically significant. Note that the heart rate of Japanese children was very similar during the “rest” and the “human” conditions where the human was the main passive or active “actor.” Note as well, that the human, and the toy robot, were the same in France and in Japan. In both groups, the heart rate was quasi-identical when both groups of children interacted with the robot.
The sharing and transformation of children’s emotional states would emanate from their understanding of the emotional experience that typifies the other (Giannopulu et al., 2016, 2018; Giannopulu, 2018). Based on interpersonal synchronisation, this is the essence of the speaker-listener condition (Tatsukawa et al., 2016). In this context, both interlocutors are performing a scenario of communication trying various verbal and non-verbal emotional reactions. Verbal reaction necessitates the elaboration of coherent sentences; non-verbal reaction takes the form of head nods and/or various kinds of facial expressions (Giannopulu, 2016a,b, 2018; Giannopulu et al., 2016, 2018; Giannopulu and Watanabe, 2018). Intimately connected with the state of the speaker, these responses signify that everything (or a part) is being integrated (Clark, 1996; Bavelas et al., 2002). Successful communication requires that both speaker and listener accurately interpret (via verbal and non-verbal emotional processes) the meaning of each other’s emotional statement. It seems that the expressive non-verbal emotional nature of a human action (i.e., the heart rate associated with the autonomic nervous system) depicts an indication of the emotional statement of the other at least. This might be considered as an emotional resonance process, or a kind of unconscious synchronisation mechanism that seems to be similar between Japanese and French children. As the emotional state reported is analogous in both populations, it seems that verbal emotional expression would be also associated with this mechanism. This latter result would be consistent with theories in which basic emotional concepts as such “want,” “feel,” or “sense” are common in all educational contexts (Wierzbicka, 1992).
Given the current findings, it seems that both in French and Japanese populations emotional empathy would require the implication of an automatic unconscious identification without any intermediary cognitive empathy (Gallese, 2003; Asada, 2014). Such identification is active with humans and with toy robots. Recent data are in accordance: emotionally empathic children exhibit unconscious non-verbal expressions (Giannopulu and Watanabe, 2018; Giannopulu et al., 2018). Neuroimaging evidence also supports such a process suggesting that the mirror neuron system is involved not only in the intersubjectivity of actions but also in the emotional empathy that allows one to feel related to others (i.e., intersubjective synchronisation) (Carr et al., 2003). Such a neural mechanism authorises one, in essence, to understand the feelings of others as well as to express his or her own feelings (Gallese, 2003). A shared representation of emotional empathetic state underlies the process. In this sense, emotional empathy would be considered a broad unconscious idiosyncrasy offering the mind a specific form of communication: a way of automatic simulation of emotional experiences that is analogous between French and Japanese children. Therefore, our findings are in accordance with the assumption that children’s brains simulate other’s feeling at an unconscious level. This might be valuable for both groups of neurotypical children analysed in our study. Given the above, emotional empathy might be considered as a synchronisation mechanism that supports human-human and human-robot interactions and predicts future emotional behaviours.
Limitations
One main limitation of our study is the absence of ecological validity. Even if our experimental approach is indispensable to establish valuable relationships between emotional empthy, synchronisation and heart rate, we would suggest that future studies should explore such relationships in more naturalistic contexts. Another limitation of the study is the lack of cortical activity and its relationship with the peripheral activity during the synchronisation process. A possibility for future research could be to investigate whether brain central activity is related to peripheral activity when synchronisation operates in human-human interaction and in child-robot interaction. Finally, we would agree that in our study we did not include a clinical population. Future studies should examine one clinical group (ASD or head injury) at least in comparison with a typical one.
Author Contributions
IG developed the method, performed the experiment, collected the data, analysed and prepared the manuscript. TW is the creator of the robot. IG, KT, and TW discussed the paper.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.The reviewer EF and handling Editor declared their shared affiliation.
Acknowledgments
We are grateful to all the participants and their parents, the Pedagogical Inspector, the Educational Advisor, the Director and the team of principal elementary school of the first district of Paris, France and Gifu, Japan.
Footnotes
Funding. The study belongs to an international project on toy robots child interactions sponsored by the Franco-Japanese Foundation of Paris (FJF16P29).
References
- Asada M. (2014). Towards artificial empathy. How can artificial empathy follow the developmental pathway of natural empathy? Int. J. Soc. Robot. 7 19–33. 10.1007/s12369-014-0253-z [DOI] [Google Scholar]
- Audrey S. (2009). A cross-cultural discussion of Japan and south Korea and how differences are manifested in the ESL/EFL classroom. Asian Soc. Sci. 5 34–39. [Google Scholar]
- Bavelas J. B., Coates L., Johnson T. (2002). Listener responses as a collaborative process: the role of gaze. J. Commun. 52 566–580. 10.1111/j.1460-2466.2002.tb02562.x [DOI] [Google Scholar]
- Cacioppo S., Zhou H., Monteleone G., Majka E. A., Quinn K. A., Ball A. B., et al. (2014). You are in sync with me: neural correlates of interpersonal synchrony with a partner. Neuroscience 277 852–858. 10.1016/j.neuroscience.2014.07.051 [DOI] [PubMed] [Google Scholar]
- Carr L., Iacoboni M., Dubeau M. C., Mazzotta C., Lenzi G. L. (2003). Neural mechanisms of empathy in humans: a relay from neural systems for imitation to limbic areas. Proc. Natl. Acad. Sci. U.S.A. 9 5497–5502. 10.1073/pnas.0935845100 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Castellano G., Leite L., Pereira A., Martinho C., Paiva A., McOwan P. (2010). Affect recognition for interactive companions: challenges and design in real world scenarios. J. Multimodal User Interfaces 3 89–98. 10.1007/s12193-009-0033-5 [DOI] [Google Scholar]
- Christov-Moore L., Simpson E. A., Coude G., Grigaityre K., Iacoboni M., Ferrari P. F. (2014). Emapthy: gender effects in brain and behavior. Neurosci. Biobehav. Rev. 46(Pt 4), 604–627. 10.1016/j.neubiorev.2014.09.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Clark H. H. (1996). Using Language. Cambridge: Cambridge University Press; 10.1017/CBO9780511620539 [DOI] [Google Scholar]
- Cornejo C., Cuadros Z., Morales R., Paredes J. (2017). Interpersonal coordination: methods, achievements, and challenges. Front. Psychol. 8:1685. 10.3389/fpsyg.2017.01685 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Decety J., Michalska K. J. (2010). Neurodevelopmental changes in the circuits underlying empathy and sympathy from childhood to adulthood. Dev. Sci. 13 886–899. 10.1111/j.1467-7687.2009.00940.x [DOI] [PubMed] [Google Scholar]
- Fan Y., Duncan N. W., de Greck M., Northoff G. (2011). Is there a core neural network in empathy? An fMRI based quantitative meta-analysis. Neurosci. Biobehav. Rev. 35 903–911. 10.1016/j.neubiorev.2010.10.009 [DOI] [PubMed] [Google Scholar]
- Feldman R., Galili G., Singer M., Louzoun Y. (2011). Mother and infant coordinate heart rhythms through episodes interaction synchrony. Infant Behav. Dev. 34 569–577. 10.1016/j.infbeh.2011.06.008 [DOI] [PubMed] [Google Scholar]
- Frith C. D., Frith U. (2003). The neural basis of mentalizing. Neuron 50 531–534. 10.1016/j.neuron.2006.05.001 [DOI] [PubMed] [Google Scholar]
- Gallese V. (2003). The roots of empathy: the shared manifold hypothesis and the neural basis of intersubjectivity. Psychopathology 36 171–180. 10.1159/000072786 [DOI] [PubMed] [Google Scholar]
- Giannopulu I. (2011). Cognitive and emotional interactions between autistic child, mobile robot and therapist. Front. Comput. Neurosci. 10.3389/conf.fncom.2011.52.00002 [DOI] [Google Scholar]
- Giannopulu I. (2013). Multimodal interactions in typically and atypically developing children: natural vs. artificial environments. Cogn. Process. 14 323–331. 10.1007/s10339-013-0566-0 [DOI] [PubMed] [Google Scholar]
- Giannopulu I. (2016a). “Enrobotment: toy robots in the developing brain,” in Handbook of Digital Games and Entertainment Technologies, eds Nakatsu R., Rauterberg M., Ciancarini P. (Berlin: Springer Science; ),1–29. [Google Scholar]
- Giannopulu I. (2016b). Contribution to the Comprehension of Multimodal Representations. Saarbrücken: European Editions. [Google Scholar]
- Giannopulu I. (2018). Neuroscience, Robotics and Virtual reality: Internalised vs. Externalised Mind/Brain. Berlin: Springer. [Google Scholar]
- Giannopulu I., Montreynaud V., Watanabe T. (2016). Minimalistic toy robot to analyse a scenery of speaker-listener condition in autism. Cogn. Process. 17 195–203. 10.1007/s10339-016-0752-y [DOI] [PubMed] [Google Scholar]
- Giannopulu I., Sagot I. (2010). Positive emotion in the course of an experimental task in children. (Ressenti émotionnel positif dans une tâche expérimentale chez l’enfant). Ann. Med. Psychol. 168 740–745. 10.1016/j.amp.2009.10.014 [DOI] [Google Scholar]
- Giannopulu I., Terada K., Watanabe T. (2018). Communication using robots: a action-perception scenario in ASD children. J. Exp. Theor. Artif. Intell. 30 603–613. 10.1080/0952813X.2018.1430865 [DOI] [Google Scholar]
- Giannopulu I., Watanabe T. (2015). “Conscious/unconscious emotional dialogues in typical children in the presence of an interactor robot,” in Proceedings of the 24th IEEE International Symposium on Robot and Human Interactive Communication, Nanjing, 264–270. [Google Scholar]
- Giannopulu I., Watanabe T. (2018). “Inter-individual differences in conscious and unconscious processes during robot-child interaction,” in New Trends in Medical and Service Robots - Design, Analysis and Control, eds Husty M., Hofbaur M. (Berlin: Springer; ), 147–158. [Google Scholar]
- Giannopulu I., Watanabe T. (2014). Give Children toys robots to educate and/or neuroreeducate: the example of PEKOPPA. New Trends Med. Serv. Robots 205–216. 10.1007/978-3-319-23832-6_17 [DOI] [Google Scholar]
- Kanske P. (2018). The social mind: disentangling affective and cognitive routes to understanding others. Interdiscip. Sci. Rev. 43 115–124. 10.1016/j.neuroimage.2015.07.082 [DOI] [PubMed] [Google Scholar]
- Lamm C., Decety J., Singer T. (2011). Meta-analytic evidence for common and distinct neural networks associated with directly experienced pain and empathy for pain. Neuroimage 54 2492–2502. 10.1016/j.neuroimage.2010.10.014 [DOI] [PubMed] [Google Scholar]
- Lamm C., Rütgen M., Wagner I. C. (2017). Imaging empathy and prosocial emotions. Neurosci. Lett. 10.1016/j.neulet.2017.06.054 [Epub ahead of print]. [DOI] [PubMed] [Google Scholar]
- Levenson R. W., Gottman J. M. (1985). Physiological and affective predictors of change in relationship satisfaction. J. Pers. Soc. Psychol. 49 85–94. 10.1037/0022-3514.49.1.85 [DOI] [PubMed] [Google Scholar]
- Lischke A., Pahnke R., Mau-Moeller A., Behrens M., Grabe H., Freyberger H. J., et al. (2018). Inter-individual differences in heart rate variability are associated with inter-individual differences in empathy and alexithymia. Front. Psychol. 9:229. 10.3389/fpsyg.2018.00229 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Llobera J., Charbonnier C., Chagué S., Preissmann D., Antonietti J. P., Ansermet F., et al. (2016). The subjective sensation of synchrony: an experimental study. PLoS One 11:e0147008. 10.1371/journal.pone.0147008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marschal D., Quinn P. C., Lea S. E. G. (2010). The Making of Human Concepts. Oxford: Oxford University Press; 10.1093/acprof:oso/9780199549221.001.0001 [DOI] [Google Scholar]
- Mitchell R. W., Hamm M. (1997). The interpretation of animal psychology; anthropomorphism or reading? Behaviour 134 173–204. 10.1163/156853997X00449 [DOI] [Google Scholar]
- Müller V., Lindenberger U. (2011). Cardiac and respiratory patterns synchronize between persons during choir singing. PLoS One 6:e24893. 10.1371/journal.pone.0024893 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Panayiotou G., Constantinou E. (2017). Emotion dysregulation in alexithymia: startle reactivity to fearful affective imagery and its relation to heart rate variability. Psychophysiology 54 1323–1334. 10.1111/psyp.12887 [DOI] [PubMed] [Google Scholar]
- Porges S. (2007). A phylogenetic journey through the vague and ambiguous Xth cranial nerve: a commentary on contemporary heart rate variability research. Biol. Psychol. 74 301–307. 10.1016/j.biopsycho.2006.08.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rizzolatti G., Craighero L. (2008). The mirror-neutron system. Annu. Rev. Neurosci. 27 169–192. 10.1146/annurev.neuro.27.070203.144230 [DOI] [PubMed] [Google Scholar]
- Shahid S., Kramer E., Swerts M. (2014). Child-robot interaction across cultures: how does playing a game with a social robot compare to playing a game alone or with a friend. Comput. Hum. Behav. 40 86–100. 10.1016/j.chb.2014.07.043 [DOI] [Google Scholar]
- Steinbeis N., Bernhardt B. C., Singer T. (2015). Age-related differences in function and structure of rSMG and reduced functional connectivity with DLPFC explains heightened emotional egocentricity bias in childhood. Soc. Cogn. Affect. Neurosci. 10 302–310. 10.1093/scan/nsu057 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stellar J. E., Cohen A., Oveis C., Keltner D. (2015). Affective and physiological responses to the suffering of others: compassion and vagal activity. J. Pers. Soc. Psychol. 108 572–585. 10.1037/pspi0000010 [DOI] [PubMed] [Google Scholar]
- Suurland J., Kristiaan B., Heijden V. D., Smaling H. J. A., Huijbregts S. C. J., Van Goozen S. H. M., et al. (2016). Infant autonomic nervous system response and recovery: associations with maternal risk status and infant emotion regulation. Dev. Pathol. 29 759–773. 10.1017/S0954579416000456 [DOI] [PubMed] [Google Scholar]
- Tanaka H., Thulesius O., Borres M., Yamaguchi H., Mino M. (1994). Blood pressure responses in Japanese and Swedish children in the supine and standing position. Eur. Heart J. 15 1011–1019. 10.1093/oxfordjournals.eurheartj.a060622 [DOI] [PubMed] [Google Scholar]
- Tatsukawa K., Nakano T., Ishiguro H., Yoshikawa Y. (2016). Eyeblink synchrony in multimodal huma-android interaction. Sci. Rep. 6:39718. 10.1038/srep39718 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tramacere A., Ferrari P. F. (2016). Faces in the mirror, from the neuroscience of mimicry to the emergence of mentalizing. J. Anthropol. Sci. 94 113–126. [DOI] [PubMed] [Google Scholar]
- Watanabe T. (2011). Human-entrained embodied interaction and communication technology. Emot. Eng. 9 161–177. 10.1007/978-1-84996-423-4_9 [DOI] [Google Scholar]
- Wierzbicka A. (1992). Semantics, Culture and Cognition: Universal Human Concepts in Culture-Specific Configurations. Oxford: Oxford Press. [Google Scholar]
- Winnicott D. W. (1971). Playing and Reality. Taylor, MI: Tavistock Publications. [Google Scholar]