Forman, Susan G.; Forman, Bruce D.
The application of Rational-Emotive Therapy principles and techniques in in-service education for school personnel is discussed. Teacher and counselor participation in a staff development program is described. (Author)
Abstract On the broadness of the vast field called “Expressions of Emotions” this study focuses on the whole bodily emotional expression. The main question posed is: Whether there are movement patterns specific to each emotion?. I carried out a thorough review of the theories of emotion...... and of expressions of emotions and movement notation that provided the sources for a careful research plan for the empirical process of this study. On this basis I chose to record onto video the four previously choreographed movements that I considered to correspond each of the following emotions: joy, fear, sadness...... emotional display. The observers closely matched the investigator’s own parameters of what was expressed in the video. Other conditions which this observing system was designed to fulfil: to use simple and “objective” terms, only a short training period, and not use any special symbols. The results obtained...
Hess, Ursula; Thibault, Pascal
In his book "The Expression of the Emotions in Man and Animals," Charles Darwin (1872/1965) defended the argument that emotion expressions are evolved and adaptive (at least at some point in the past) and serve an important communicative function. The ideas he developed in his book had an important impact on the field and spawned rich domains of…
Vieillard, Sandrine; Roy, Mathieu; Peretz, Isabelle
This study was designed to investigate how emotion category, characterized by distinct musical structures (happiness, sadness, threat) and expressiveness (mechanical, expressive) may influence overt and covert behavioral judgments and physiological responses in musically trained and untrained listeners. Mechanical and expressive versions of happy, sad and scary excerpts were presented while physiological measures were recorded. Participants rated the intensity of the emotion they felt. In addition, they monitored excerpts for the presence of brief breaths. Results showed that the emotion categories were rated higher in the expressive than in the mechanical versions and that this effect was larger in musicians. Moreover, expressive excerpts were found to increase skin conductance level more than the mechanical ones, independently of their arousal value, and to slow down response times in the breath detection task relative to the mechanical versions, suggesting enhanced capture of attention by expressiveness. Altogether, the results support the key role of the performer's expression in the listener's emotional response to music.
Tsiourti, Christiana; Weiss, Astrid; Wac, Katarzyna
Socially assistive agents, be it virtual avatars or robots, need to engage in social interactions with humans and express their internal emotional states, goals, and desires. In this work, we conducted a comparative study to investigate how humans perceive emotional cues expressed by humanoid...... with abstract humanlike features. A qualitative and quantitative data analysis confirmed the expressive power of the face, but also demonstrated that body expressions or even simple head and locomotion movements could convey emotional information. These findings suggest that emotion recognition accuracy varies...... robots through five communication modalities (face, head, body, voice, locomotion) and examined whether the degree of a robot's human-like embodiment affects this perception. In an online survey, we asked people to identify emotions communicated by Pepper -a highly human-like robot and Hobbit – a robot...
Rodriguez-Hidalgo, Carmina Rodriguez-Hidalgo; Tan, Ed S.; Verlegh, Peeter
of emotion (SSE, Rimé, 2009). This study content-analyzed Live Journal blogposts for the occurrence of TPC in three phases of online SSE: initiation, feedback and repost. We compared these to TPC on a second type of emotional expression, emotional venting. Based on Social Information processing theory (SIP......Textual paralanguage cues (TPC) have been signaled as effective emotion transmitters online. Though several studies have investigated their properties and occurrence, there remains a gap concerning their communicative impact within specific psychological processes, such as the social sharing......, Walther, 1992), and on the Emotional Mimicry in Context (EMC, Hess & Fischer, 2013) framework, we study predictive relationships in TPC usage in our phased model of online SSE. Results showed that TPC prevailed in SSE blogposts and strongly dominated in emotional venting posts. TPC was more common...
Research into emotions has increased in recent decades, especially on the subject of recognition of emotions. However, studies of the facial expressions of emotion were compromised by technical problems with visible video analysis and electromyography in experimental settings. These have only recently been overcome. There have been new developments in the field of automated computerized facial recognition; allowing real-time identification of facial expression in social environments. This review addresses three approaches to measuring facial expression of emotion and describes their specific contributions to understanding emotion in the healthy population and in persons with mental illness. Despite recent progress, studies on human emotions have been hindered by the lack of consensus on an emotion theory suited to examining the dynamic aspects of emotion and its expression. Studying expression of emotion in patients with mental health conditions for diagnostic and therapeutic purposes will profit from theoretical and methodological progress.
Codier, Estelle; Kamikawa, Cindy; Kooker, Barbara M; Shoultz, Jan
Emotional intelligence has been correlated with performance, retention, and organizational commitment in professions other than nursing. A 2006 pilot study provided the first evidence of a correlation between emotional intelligence and performance in clinical staff nurses. A follow-up study was completed, the purpose of which was to explore emotional intelligence, performance level, organizational commitment, and retention. A convenience sample of 350 nurses in a large medical center in urban Hawaii participated in this study. This article reports the findings pertaining to the subset of 193 clinical staff nurses who responded. The Mayer-Salovey-Caruso Emotional Intelligence Test instrument was used to measure emotional intelligence abilities. Performance was defined as ranking on a clinical ladder. Commitment was scored on a Likert scale. The following variables measured retention: total years in nursing, years in current job, total years anticipated in current job, and total anticipated career length. Emotional intelligence scores in clinical staff nurses correlated positively with both performance level and retention variables. Clinical staff nurses with higher emotional intelligence scores demonstrated higher performance, had longer careers, and greater job retention.
Eric G. Lambert
Full Text Available The vast majority of past correctional staff burnout studies have focused on the possible antecedents of job burnout. Far fewer studies have been published on the possible outcomes of burnout among correctional staff. This study examined the effects of the emotional exhaustion dimension of burnout on life satisfaction, support for treatment, support for punishment, absenteeism, views on use of sick leave, and turnover intent among 272 staff at a state-run Midwestern maximum security prison. Ordinary least squares (OLS regression analysis of survey data indicated that emotional burnout had significant negative associations with life satisfaction and support for treatment and significant positive relationships with support for punishment, absenteeism, views on use of sick leave (i.e., a right to be used however the employee wishes, and turnover intent. The results indicate that job burnout has negative outcomes for both staff and correctional institutions.
Rasmussen, Tove Arendt
are being treated as ordinary people. My article will discuss different presentations of selves and especially the emotional verbal and nonverbal expressions in reality TV communication. Aspects of the intimate self and its emotional expressions seem to be strategically managed in reality TV and even......, that information ‘given off’ in lying behavior will must often be found in non-verbal mikro expressions and mikro gestures. I will end by discussing the strategic emotional expressions as production of floating identities in a broader framework on the basis of among others Gergen, Lipovetsky and Baumann...
Emotions – a part of person life. We always feel something and express in different ways: body language, mime, or words (our language). In my research are analyzed emotions language in weblogs, because weblogs like public notebooks, which shows a really language usage. It can to see in weblogs, how do people perceive emotions, what criteria assigns them. The aim of research: to determine, what kind of words and how do people express own emotions in weblogs. There was reviewed 15 weblogs and g...
Kao, Min C; Chen, Yung Y
This study investigated Emotional Intelligence (EI) as a moderator for the association between emotional expression and adaptive trauma processing, as measured by depressive symptoms. Using Pennebaker's written emotional expression paradigm, 105 participants were assigned to either a conventional trauma-writing or religious trauma-writing condition. Depressive symptoms were assessed at baseline and again at one-month post writing. No significant association between EI and religiousness was found at baseline. Results indicated a three-way interaction among EI, trauma severity, and writing condition on depressive symptoms at follow-up. For the religious trauma-writing condition only, there was a significant difference between high- versus low-EI participants who experienced more severe trauma in depressive symptoms at follow-up, such that low-EI participants registered less depressive symptoms than high-EI participants; while there was no significant difference between low versus high EI for participants with less severe trauma. These findings encourage further investigation of the conditions under which religion may be a beneficial factor in trauma adaptation.
Harper, Mary G; Jones-Schenk, Jan
This study investigated the emotional intelligence (EI) profile of successful staff nurses to examine correlations among EI and demographic variables. This descriptive, exploratory study examined the EI of 42 participants using the Bar-On Emotional Quotient Inventory. Mean scores for total EI, scales, and subscales were all in the average range, indicating an ability to successfully navigate relationships in work and life. Nineteen percent of the participants scored below average on total EI, whereas 31% scored above average. A negative correlation between age and empathy was found. Relative areas of strength included stress tolerance, problem solving, self-regard, and self-actualization. The study findings suggested that successful staff nurses have average or higher levels of EI and that empathy among these nurses declines with age. Research on how empathy evolves, factors that influence empathy, and strategies to enhance EI among nurses is warranted. Copyright 2012, SLACK Incorporated.
Explores the importance of emotional expression in music performance. Performers played music to express different emotions and then listening tests were conducted in order to determine whether the intended expressions were perceived. Presents and discusses the results. (CMK)
Immordino-Yang, Mary Helen; Yang, Xiao-Fei; Damasio, Hanna
The brain’s mapping of bodily responses during emotion contributes to emotional experiences, or feelings. Culture influences emotional expressiveness, i.e. the magnitude of individuals’ bodily responses during emotion. So, are cultural influences on behavioral expressiveness associated with differences in how individuals experience emotion? Chinese and American young adults reported how strongly admiration and compassion-inducing stories made them feel, first in a private interview and then d...
Zijlmans, L. J. M.; Embregts, P. J. C. M.; Gerits, L.; Bosman, A. M. T.; Derksen, J. J. L.
Background: Recent research addressed the relationship between staff behaviour and challenging behaviour of individuals with an intellectual disability (ID). Consequently, research on interventions aimed at staff is warranted. The present study focused on the effectiveness of a staff training aimed at emotional intelligence and interactions…
Maârof Bin Redzuan Haslinda Abdullah, Aida Mehrad Hanina Halimatussadiah
One of the great positive behavioral factors among staff at university is emotional intelligence. In reality, emotional intelligence is cause of different reaction at workplace that was appeared by staff and also controlled most of moods in various situations. Moreover, knowing factors that impact on emotional intelligence is very vital and lead to different positive and negative behavior among staff. Reward is one of these external effective factors that influence on emotional intelligence. ...
Rosner, R; Beutler, L E; Daldrup, R J
Emotional arousal is a key concept in most theories of change. To be able to understand the role of emotional expression better, two treatments, cognitive therapy (CT) and focused expressive psychotherapy (FEP; a manualized form of Gestalt therapy), with opposite process assumptions about the expression of emotions were compared. Additionally vicarious emotional experience in the sense of an underlying emotional contagion was examined. Clients suffering from major depression were rated for the expression of emotion in three randomly selected sessions of a 20-session treatment course. While the types of emotions generally experienced by CT clients and FEP clients did not differ significantly, differences in the subgroups of active and observing-group members were found. This indicated that the process assumptions made by the respective treatments were only valid for the actively participating clients and not for the observing group members. Emotional contagion as a process was not supported.
Stordeur, S; D'hoore, W; Vandenberghe, C
STUDY'S RATIONALE AND OBJECTIVES: We examined the effect of work stressors and head nurses' transactional and transformational leadership on the levels of emotional exhaustion experienced among their staff. A questionnaire was sent to all nurses of a university hospital. Usable returns were received from 625 nurses, giving a response rate of 39.2%. Data were treated using correlational analyses and multiple regression. The latter modelled stressors and leadership as predictors of nurses' reported emotional exhaustion. Work stressors were assessed using the Nursing Stress Scale (NSS) which comprises 34 items divided into three subscales (referring to stress from the physical, psychological, and social environment), and the role ambiguity (three items) and conflict (three items) scales. Leadership was measured with the Multifactor Leadership Questionnaire. In regression analyses, work stressors as a whole were found to explain 22% of the variance in emotional exhaustion whereas leadership dimensions explained 9% of the variance in that outcome measure. Stress emanating from the physical and social environment, role ambiguity, and active management-by-exception leadership were significantly associated with increased levels of emotional exhaustion. Transformational and contingent reward leadership did not influence emotional exhaustion. A limitation of this study is that it considered only the emotional exhaustion dimension of burnout. Also, as data were cross-sectional in nature, conclusions regarding the direction of causality among variables cannot be drawn. This study provided, for the first time, a test of the influence of leadership on burnout among nurses, taking into account the role of work stressors. Future research is needed to examine if the effects reported herein can be replicated using the two other dimensions of burnout (depersonalization and reduced personal accomplishment).
Caballero Meneses, Jonathan Azael; Menez Díaz, Judith Marina
Emotional expressions have been proposed to be important for regulating social interaction as they can serve as cues for behavioral intentions. The issue has been mainly addressed analyzing the effects of facial emotional expressions in cooperation behavior, but there are contradictory results regarding the impact of emotional expressions on that…
Zijlmans, Linda J M; Embregts, Petri J C M; Bosman, Anna M T
Working with clients who show challenging behavior can be emotionally demanding and stressful for support staff, because this behavior may cause a range of negative emotional reactions and feelings. These reactions are of negative influence on staff wellbeing and behavior. Research has focused on negative emotions of staff. However, a distinction between emotions and feelings has never been made in the research field of intellectual disabilities. Negative emotions and feelings may be regulated by emotional intelligence, a psychological construct that takes into account personal style and individual differences. The purpose of this study was to explore the relationship between emotional intelligence on the one hand and emotions and feelings on the other. Participants were 207 support staff serving clients with moderate to borderline intellectual disabilities and challenging behavior. Emotional intelligence, emotions, and feelings were measured with questionnaires. The results show that emotional intelligence, emotions, and feelings are related. However, found relationships were weak. Most significant relations were found between feelings and stress management and adaptation elements of emotional intelligence. Because the explored variables can change over time they call for a longitudinal research approach. Copyright © 2013 Elsevier Ltd. All rights reserved.
Hufnagel, Elizabeth; Kelly, Gregory J.
This methodological paper presents an approach for examining emotional expressions through discourse analysis and ethnographic methods. Drawing on trends in the current literature in science education, we briefly explain the importance of emotions in science education and examine the current research methodologies used in interactional emotion studies. We put forth and substantiate a methodological approach that attends to the interactional, contextual, intertextual, and consequential aspects of emotional expressions. By examining emotional expressions in the discourse in which they are constructed, emotional expressions are identified through semantics, contextualization, and linguistic features. These features make salient four dimensions of emotional expressions: aboutness, frequency, type, and ownership. Drawing on data from a large empirical study of pre-service elementary teachers' emotional expressions about climate change in a science course, we provide illustrative examples to describe what counts as emotional expressions in situ. In doing so we explain how our approach makes salient the nuanced nature of such expressions as well as the broader discourse in which they are constructed and the implications for researching emotional expressions in science education discourse. We suggest reasons why this discourse orientated research methodology can contribute to the interactional study of emotions in science education contexts.
fundamental reason. (Mis)matching peoples mood with the emotions expressed in music was found to be an essential underlying mechanism, people use to regulate their emotions. This formed the basis and overall goal of the thesis, to investigate how to create a predictive model of emotions expressed in music....... To use in the next generation of music systems. The thesis was divided into three main topics involved in creating a predictive model 1) Elicitation of emotion, 2) Audio representation and 3) Modelling framework, associating the emotion and audio representation, allowing to predict the emotions expressed...... in the form of pairwise comparisons. One issue with pairwise comparisons is the scaling, this was solved using an active learning approach through a Gaussian Process model. Traditional audio representation disregards all temporal information in audio features used for modelling the emotions expressed in music...
Are, Funlola; Shaffer, Anne
While there is a growing body of literature examining the influence of emotion socialization on children's emotional and social development, there is less research on what predicts emotion socialization behaviors among parents. The current study explores maternal emotion regulation difficulties as a predictor of emotion socialization practices, specifically, family emotion expressiveness. Further, the current study examines the role of family emotion expressiveness as a possible mediator of the relations between maternal and child emotion regulation in a community sample of 110 mother-child dyads with preschool-aged children. Analyses revealed that positive family expressiveness mediated the relations between maternal emotion dysregulation and child emotion regulation and thus presents important clinical implications for existing emotion socialization interventions.
Akkoç, Betül; Arslan, Ahmet
Eyes play an important role in expressing emotions in nonverbal communication. In the present study, emotional expression classification was performed based on the features that were automatically extracted from the eye area. Fırst, the face area and the eye area were automatically extracted from the captured image. Afterwards, the parameters to be used for the analysis through discrete wavelet transformation were obtained from the eye area. Using these parameters, emotional expression analysis was performed through artificial intelligence techniques. As the result of the experimental studies, 6 universal emotions consisting of expressions of happiness, sadness, surprise, disgust, anger and fear were classified at a success rate of 84% using artificial neural networks.
Full Text Available A robot emotional expression model based on Hidden Markov Model (HMM is built to enable robots which have different personalities to response in a more satisfactory emotional level. Gross emotion regulation theory and Five Factors Model (FFM which are the theoretical basis are firstly described. And then the importance of the personality effect on the emotion expression process is proposed, and how to make the effect quantization is discussed. After that, the algorithm of HMM is used to describe the process of emotional state transition and expression, and the performance transferring probability affected by personality is calculated. At last, the algorithm model is simulated and applied in a robot platform. The results prove that the emotional expression model can acquire humanlike expressions and improve the human-computer interaction.
In her thoughtful work regarding various aspects of emotion and emotion related words, Pavlenko explores a variety of perspectives on how we might characterize and conceptualize expressions of emotion. It is a work that is quite rich in breadth--one that leads to a variety of different thoughts on this topic, many of which are amenable to…
Rubita Sudirman; Ching Yee Yong; Kim Mey Chew
This study is to investigate human perceptions on pairing of facial expressions of emotion with colours. A group of 27 subjects consisting mainly of younger and Malaysian had participated in this study. For each of the seven faces, which expresses the basic emotions neutral, happiness, surprise, anger, disgust, fear and sadness, a single colour is chosen from the eight basic colours for the “match” of best visual look to the face accordingly. The different emotions appear well characterized b...
Baldwin, Carol May
Studies in the perceptual identification of emotional states suggested that listeners seemed to depend on a limited set of vocal cues to distinguish among emotions. Linguistics and speech science literatures have indicated that this small set of cues included intensity, fundamental frequency, and temporal properties such as speech rate and duration. Little research has been done, however, to validate these cues in the production of emotional speech, or to determine if specific dimensions of each cue are associated with the production of a particular emotion for a variety of speakers. This study addressed deficiencies in understanding of the acoustical properties of duration and intensity as components of emotional speech by means of speech science instrumentation. Acoustic data were conveyed in a brief sentence spoken by twelve English speaking adult male and female subjects, half with dramatic training, and half without such training. Simulated expressions included: happiness, surprise, sadness, fear, anger, and disgust. The study demonstrated that the acoustic property of mean intensity served as an important cue for a vocal taxonomy. Overall duration was rejected as an element for a general taxonomy due to interactions involving gender and role. Findings suggested a gender-related taxonomy, however, based on differences in the ways in which men and women use the duration cue in their emotional expressions. Results also indicated that speaker training may influence greater use of the duration cue in expressions of emotion, particularly for male actors. Discussion of these results provided linkages to (1) practical management of emotional interactions in clinical and interpersonal environments, (2) implications for differences in the ways in which males and females may be socialized to express emotions, and (3) guidelines for future perceptual studies of emotional sensitivity.
Dael, Nele; Mortillaro, Marcello; Scherer, Klaus R
Emotion communication research strongly focuses on the face and voice as expressive modalities, leaving the rest of the body relatively understudied. Contrary to the early assumption that body movement only indicates emotional intensity, recent studies have shown that body movement and posture also conveys emotion specific information. However, a deeper understanding of the underlying mechanisms is hampered by a lack of production studies informed by a theoretical framework. In this research we adopted the Body Action and Posture (BAP) coding system to examine the types and patterns of body movement that are employed by 10 professional actors to portray a set of 12 emotions. We investigated to what extent these expression patterns support explicit or implicit predictions from basic emotion theory, bidimensional theory, and componential appraisal theory. The overall results showed partial support for the different theoretical approaches. They revealed that several patterns of body movement systematically occur in portrayals of specific emotions, allowing emotion differentiation. Although a few emotions were prototypically expressed by one particular pattern, most emotions were variably expressed by multiple patterns, many of which can be explained as reflecting functional components of emotion such as modes of appraisal and action readiness. It is concluded that further work in this largely underdeveloped area should be guided by an appropriate theoretical framework to allow a more systematic design of experiments and clear hypothesis testing.
Kniazev, G G; Mitrofanova, L G; Bocharov, A V
Emotional intelligence-related differences in oscillatory responses to emotional facial expressions were investigated in 48 subjects (26 men and 22 women) in age 18-30 years. Participants were instructed to evaluate emotional expression (angry, happy and neutral) of each presented face on an analog scale ranging from -100 (very hostile) to + 100 (very friendly). High emotional intelligence (EI) participants were found to be more sensitive to the emotional content of the stimuli. It showed up both in their subjective evaluation of the stimuli and in a stronger EEG theta synchronization at an earlier (between 100 and 500 ms after face presentation) processing stage. Source localization using sLORETA showed that this effect was localized in the fusiform gyrus upon the presentation of angry faces and in the posterior cingulate gyrus upon the presentation of happy faces. At a later processing stage (500-870 ms) event-related theta synchronization in high emotional intelligence subject was higher in the left prefrontal cortex upon the presentation of happy faces, but it was lower in the anterior cingulate cortex upon presentation of angry faces. This suggests the existence of a mechanism that can be selectively increase the positive emotions and reduce negative emotions.
Monin, Joan K.; Martire, Lynn M.; Schulz, Richard; Clark, Margaret S.
This study examined the association between care-recipients’ willingness to express emotions to spousal caregivers and caregiver’s well-being and support behaviors. Using self-report measures in the context of a larger study, 262 care-recipients with osteoarthritis reported on their willingness to express emotions to caregivers, and caregivers reported on their stress and insensitive responding to care-recipients. Results revealed that care-recipients’ willingness to express happiness was ass...
Full Text Available The present paper scrutinizes the relationship between family emotional expressiveness (i.e., the tendency to express dominant and/or submissive positive and negative emotions and components of family structure as proposed in Olson’s Circumplex model (i.e., cohesion and flexibility, family communication, and satisfaction in families with adolescents. The study was conducted on a sample of 514 Slovenian adolescents, who filled out two questionnaires: the Slovenian version of Family Emotional Expressiveness - FEQ and FACES IV. The results revealed that all four basic dimensions of family functioning were significantly associated with higher/more frequent expressions of positive submissive emotions, as well as with lower/less frequent expressions of negative dominant emotions. Moreover, expressions of negative submissive emotions explained a small, but significant amount of variance in three out of four family functioning variables (satisfaction, flexibility, and communication. The importance of particular aspects of emotional expressiveness for family cohesion, flexibility, communication, and satisfaction is discussed, and the relevance of present findings for family counselling is outlined.
de Gelder, B; de Borst, A W; Watson, R
During communication, we perceive and express emotional information through many different channels, including facial expressions, prosody, body motion, and posture. Although historically the human body has been perceived primarily as a tool for actions, there is now increased understanding that the body is also an important medium for emotional expression. Indeed, research on emotional body language is rapidly emerging as a new field in cognitive and affective neuroscience. This article reviews how whole-body signals are processed and understood, at the behavioral and neural levels, with specific reference to their role in emotional communication. The first part of this review outlines brain regions and spectrotemporal dynamics underlying perception of isolated neutral and affective bodies, the second part details the contextual effects on body emotion recognition, and final part discusses body processing on a subconscious level. More specifically, research has shown that body expressions as compared with neutral bodies draw upon a larger network of regions responsible for action observation and preparation, emotion processing, body processing, and integrative processes. Results from neurotypical populations and masking paradigms suggest that subconscious processing of affective bodies relies on a specific subset of these regions. Moreover, recent evidence has shown that emotional information from the face, voice, and body all interact, with body motion and posture often highlighting and intensifying the emotion expressed in the face and voice. © 2014 John Wiley & Sons, Ltd.
Monin, Joan K; Martire, Lynn M; Schulz, Richard; Clark, Margaret S
This study examined the association between care-recipients' willingness to express emotions to spousal caregivers and caregiver's well-being and support behaviors. Using self-report measures in the context of a larger study, 262 care-recipients with osteoarthritis reported on their willingness to express emotions to caregivers, and caregivers reported on their stress and insensitive responding to care-recipients. Results revealed that care-recipients' willingness to express happiness was associated with less insensitive caregiver responding, and willingness to express interpersonal emotions (e.g., compassion, guilt) was associated with less caregiving stress. There were also gender differences, such that caregiving wives, in particular, benefited from their husband's willingness to express vulnerable (e.g., anxiety, sadness) and interpersonal emotions. (c) 2009 APA, all rights reserved
Eglee Duran Rodríguez
Full Text Available The article reports the results of the project “Mental health and emotional expression in Facebook”. The research was approached from the qualitative paradigm under virtual ethnographic approach, interpreting the findings through their own players and triangulated with the views of researchers and experts in the area of mental health, emotions and information technology and communication. We concluded that a good part of users vented their secrets on Facebook, where they are able to confide and express a range of emotions and intimacies that in the real context is unlikely to give. Along these findings show that the use of Facebook serves as a space for emotional expression impacting the mental and emotional health.
Matsumoto, David; Hwang, Hyisung C
Most studies on judgments of facial expressions of emotion have primarily utilized prototypical, high-intensity expressions. This paper examines judgments of subtle facial expressions of emotion, including not only low-intensity versions of full-face prototypes but also variants of those prototypes. A dynamic paradigm was used in which observers were shown a neutral expression followed by the target expression to judge, and then the neutral expression again, allowing for a simulation of the emergence of the expression from and then return to a baseline. We also examined how signal and intensity clarities of the expressions (explained more fully in the Introduction) were associated with judgment agreement levels. Low-intensity, full-face prototypical expressions of emotion were judged as the intended emotion at rates significantly greater than chance. A number of the proposed variants were also judged as the intended emotions. Both signal and intensity clarities were individually associated with agreement rates; when their interrelationships were taken into account, signal clarity independently predicted agreement rates but intensity clarity did not. The presence or absence of specific muscles appeared to be more important to agreement rates than their intensity levels, with the exception of the intensity of zygomatic major, which was positively correlated with agreement rates for judgments of joy.
Øverland, Klara; Størksen, Ingunn; Bru, Edvin; Thorsen, Arlene Arstad
This Q methodological study explores emotional experiences and coping of daycare staff when working with children of divorce and their families. Two main coping strategies among daycare staff were identified: 1) Confident copers, and 2) Non-confident copers. Interviews exemplify the two main experiences. Both groups may struggle with coping in…
Bui, T.D.; Magnenat-Thalmann, N.; Heylen, Dirk K.J.; Joslin, C.; Nijholt, Antinus; Kim, H.S.
In this paper we presented Obie, an embodied agent that experiences and expresses emotions. Obie has an adaptive, quantitative and domain-independent emotion component which appraises events to trigger emotions. Obie’s emotions are expressed via his utterances or his facial expressions. The expression via utterances is done by a simple mapping from emotions to text fragments. The mapping from emotions to facial expressions is done by a fuzzy rule-based system. Obie’s utterances and facial exp...
Waaramaa, Teija; Laukkanen, Anne-Maria; Alku, Paavo; Väyrynen, Eero
Fundamental frequency (F(0)) and intensity are known to be important variables in the communication of emotions in speech. In singing, however, pitch is predetermined and yet the voice should convey emotions. Hence, other vocal parameters are needed to express emotions. This study investigated the role of voice source characteristics and formant frequencies in the communication of emotions in monopitched vowel samples [a:], [i:] and [u:]. Student actors (5 males, 8 females) produced the emotional samples simulating joy, tenderness, sadness, anger and a neutral emotional state. Equivalent sound level (L(eq)), alpha ratio [SPL (1-5 kHz) - SPL (50 Hz-1 kHz)] and formant frequencies F1-F4 were measured. The [a:] samples were inverse filtered and the estimated glottal flows were parameterized with the normalized amplitude quotient [NAQ = f(AC)/(d(peak)T)]. Interrelations of acoustic variables were studied by ANCOVA, considering the valence and psychophysiological activity of the expressions. Forty participants listened to the randomized samples (n = 210) for identification of the emotions. The capacity of monopitched vowels for conveying emotions differed. L(eq) and NAQ differentiated activity levels. NAQ also varied independently of L(eq). In [a:], filter (formant frequencies F1-F4) was related to valence. The interplay between voice source and F1-F4 warrants a synthesis study. Copyright 2008 S. Karger AG, Basel.
GERLSMA, C; VANDERLUBBE, PM; VANNIEUWENHUIZEN, C
When the factor structure and psychometric qualities of the Level of Expressed Emotion scale, an instrument intended to assess patient's perceptions of expressed emotion, were evaluated, three moderately intercorrelated factors emerged, with good internal consistency; these were lack of emotional
Doroftei, I.; Adascalitei, F.; Lefeber, D.; Vanderborght, B.; Doroftei, I. A.
The purpose of this study is to present the preliminary steps in facial expressions recognition with a new version of an expressive social robotic head. So, in a first phase, our main goal was to reach a minimum level of emotional expressiveness in order to obtain nonverbal communication between the robot and human by building six basic facial expressions. To evaluate the facial expressions, the robot was used in some preliminary user studies, among children and adults.
Mackay, Nadine; Barrowclough, Christine
The study applied Weiner's (1980, 1986) attributional model of helping behaviour to Accident and Emergency (A&E) staff's care of patients presenting with deliberate self-harm. It was hypothesized that where staff attributed precipitants of the act of deliberate self-harm to controllable, internal, and stable patient factors, then staff would display greater negative affect, less optimism, and less willingness to help the patient. Using four hypothetical scenarios in a two-factor between-subjects design, contextual factors describing a self-harm patient were manipulated. Participants were 89 A&E medical and nursing staff. They were asked to rate attributions for the cause of the deliberate self-harm and their emotional responses, optimism for change, and willingness to help change the behaviour. Their general attitudes towards deliberate self-harm patients and perceived needs for training in the care of these patients were also assessed. The findings were consistent with Weiner's attributional model of helping. The greater attributions of controllability, the greater the negative affect of staff towards the person, and the less the propensity to help. The higher the ratings of stability of outcome, the less staff optimism for the success of their input. Male staff and medical staff had more negative attitudes, and medical staff saw less need for further training. Formulating A&E staff's responses to deliberate self-harm using a cognitive-emotional model offers the possibility of working with staffs' beliefs, emotions, and behaviour to improve the care and treatment of deliberate self-harm patients.
Full Text Available Objectives: The purpose of this study was to determine the relationship between emotional intelligence and occupational stress among rehabilitation staffs in Tehran’s training hospitals . Methods: A cross-sectional study was conducted on a sample of 169 staff members selected from a total of 300 rehabilitation staffs working in Tehran’s training hospitals, recruited by random cluster sampling. Two questionnaires were used: The emotional intelligence questionnaire designed by Petrides and Furnham and HSE occupational stress questionnaire. Data obtained from this study were analyzed using Pearson’s correlation and multiple regression tests. Results: An inverse significant relationship existed between occupational stress and emotional intelligence (P<0.001, r=-0.33. There are, also, significant relationships between subscales of emotional intelligence including self-awareness (P=0.031, r=-0.18, social skills (P<0.001, r=-0.302, empathy (P=0.006, r=-0.238 and occupational stress. The results of multiple regressions indicated that the two subscales of ‘understanding other’s emotions’ and ‘social skills’ can be used for predicting occupational stress. Discussion: This study confirmed the relationship between emotional intelligence and occupational stress. Promotion of emotional intelligence through implementing training courses may lower rehabilitation staffs occupational stress or prevent it.
Pochedly, Joseph T; Widen, Sherri C; Russell, James A
The emotion attributed to the prototypical "facial expression of disgust" (a nose scrunch) depended on what facial expressions preceded it. In two studies, the majority of 120 children (5-14 years) and 135 adults (16-58 years) judged the nose scrunch as expressing disgust when the preceding set included an anger scowl, but as angry when the anger scowl was omitted. An even greater proportion of observers judged the nose scrunch as angry when the preceding set also included a facial expression of someone about to be sick. The emotion attributed to the nose scrunch therefore varies with experimental context. PsycINFO Database Record (c) 2012 APA, all rights reserved.
Bui, T.D.; Magnenat-Thalmann, N.; Heylen, Dirk K.J.; Joslin, C.; Nijholt, Antinus; Kim, H.S.
In this paper we presented Obie, an embodied agent that experiences and expresses emotions. Obie has an adaptive, quantitative and domain-independent emotion component which appraises events to trigger emotions. Obie’s emotions are expressed via his utterances or his facial expressions. The
Röndahl, Gerd; Innala, Sune; Carlsson, Marianne
Studies have reported that homosexual patients fear they will not receive adequate care if they openly show their sexual orientation, for example, when introducing their partner. The aims of this study were to investigate the emotions of nursing staff and nursing students, and possible relations to cultural background and gender, towards homosexual patients; whether nursing staff and nursing students would choose to refrain from nursing homosexual patients, if the option existed; and, if so, how they express their wish to refrain from nursing this group of patients. All participants received verbal and written information before the study started. Returning a completed questionnaire indicated a participant's tacit consent. Approval was obtained from the heads of departments and persons in charge of nursing and nursing assistant programmes. The study had a descriptive, comparative design, and an Affect Adjective Checklist (AAC) and specially designed Nursing Behaviour Questionnaire (NBQ) were used. The participants included nurses and assistant nurses from an infectious disease clinic, and students enrolled in a university nursing programme and upper secondary assistant nurses' training, all in central Sweden. The findings showed that both professional nursing staff (response rate 67%, n = 57), and students (response rate 62%, n = 165), expressed emotions of homophobic anger, homophobic guilt and delight. Groups with a cultural background other than Swedish expressed more homophobia. No gender differences were indicated for homophobic emotions. In the professional group, 36% would refrain from nursing for homosexual patients if given the option. The corresponding figure for the students was 9%. The limitations were that the sample was small and not randomly selected, and as participation was anonymous no follow-up could be done. It was concluded that the emotional factors of homosexual anger and homosexual guilt might be of value in helping to explain and predict
Becker, S; Konrad, A; Zimmermann, L; Müller, C; Tomczyk, S; Reichler, L; Körner, M
Aim: In the context of high prevalence rates of mental and psychosomatic disorders in the medical staff, emotional exhaustion and wellbeing are often considered as important indicators. Teamwork can have a positive influence on wellbeing of staff members. In the sector of rehabilitation, however, this is not sufficiently investigated. The aim of this study was to investigate aspects of teamwork as predictors of wellbeing and emotional exhaustion in staff at rehabilitation clinics in Germany. Methods: Data was collected in 10 rehabilitation clinics, 9 of them could be included in the data analysis (n=306, 70% female, 68% age 40-59). Data was analyzed with multiple linear regression analyses. Results: Staff reported moderate rates of emotional exhaustion and good rates of overall wellbeing. Results of the regression analysis show that cohesion (β=0.27, pteamwork are discussed as well as the potential relevance of interventions to improve teamwork, in order to enhance wellbeing and counteract emotional exhaustion of staff members. © Georg Thieme Verlag KG Stuttgart · New York.
Polanco-Roman, Lillian; Moore, Alyssa; Tsypes, Aliona; Jacobson, Colleen; Miranda, Regina
Emotion reactivity and difficulties in expressing emotions have been implicated in risk for suicidal behavior. This study examined comfort in expressing emotions (positive vs. negative) and depressive symptoms as mediators of the prospective relation between emotion reactivity and suicidal ideation. Emerging adults (N = 143; 72% female; 28% White) completed measures of emotion reactivity, comfort expressing emotions, and suicidal ideation at baseline and of depressive symptoms and suicidal ideation 12 months later. Emotion reactivity predicted suicidal ideation at follow-up through depressive symptoms. Difficulty expressing love-but not happiness, sadness, and anger-partially mediated the relationship between emotion reactivity and suicidal ideation at follow-up before but not after adjusting for baseline ideation. The relation between high emotion reactivity and suicidal ideation may be explained by discomfort in the expression of positive emotions and by depressive symptoms. Promotion of comfort in positive emotion expression may reduce vulnerability to suicidal ideation. © 2017 Wiley Periodicals, Inc.
Ali Reza Salar
Full Text Available Nowadays, companies and organizations are paying more attention to the emotional intelligence in an increasingly rapid pace. The reason for such an attention is that emotional intelligence reflects the favorable and optimum administrative capabilities in controlling the psychological dispositions and behaviors and tensions and it is deemed as a factor that creates motivation and hope in the individual when the time for failure in achieving the objective s and goals arrives and because there was not information regarding emotional intelligence in and among Zahedan medical sciences university staff members the present study was undertaken to survey the emotional intelligence among Zahedan medical sciences university staff. The present is a cross-sectional descriptive-analytical research which has been conducted on 160 individuals from Zahedan medical sciences university staff members who were selected randomly in 2016. To gather the information required for the current study, a questionnaire comprised of two parts was applied the first part of which pertained to the demographic characteristics and the second part was related to the emotional intelligence standard questionnaire. Data were analyzed by the use of SPSS 19 and descriptive statistics, Pierson correlation and independent t-test. The findings of the present study indicated that the participants’ average age was 36.54 ± 10.03, 98 individuals were women, 137 individuals were married. The emotional intelligence total mean score was 114.11 ± 14.07 which is ranked as high according to the qu estionnaire classification. The relationship between the age and marital status with emotional intelligence total mean score and each of its components was not statistically significant. Although the comparison between the results obtained in the present study and the other studies’ results indicated the emotional intelligence is in an acceptable level among Zahedan Medical sciences university staff
Full Text Available This study is to investigate human perceptions on pairing of facial expressions of emotion with colours. A group of 27 subjects consisting mainly of younger and Malaysian had participated in this study. For each of the seven faces, which expresses the basic emotions neutral, happiness, surprise, anger, disgust, fear and sadness, a single colour is chosen from the eight basic colours for the match of best visual look to the face accordingly. The different emotions appear well characterized by a single colour. The approaches used in this experiment for analysis are psychology disciplines and colours engineering. These seven emotions are being matched by the subjects with their perceptions and feeling. Then, 12 male and 12 female data are randomly chosen from among the previous data to make a colour perception comparison between genders. The successes or failures in running of this test depend on the possibility of subjects to propose their every single colour for each expression. The result will translate into number and percentage as a guide for colours designers and psychology field.
Trondhjem, Naja Blytmann
How are emotions expressed in Kalaallisut In Kalaallisut the expressions of emotions are expressed by verbal stems, verbal as well as nominal affixes, and of ‘metaphorical’ expressions. Within verbal stems there are both mono morphemic ‘basic’ emotion terms and also ‘metaphorical’ terms which...... consist of compound verbal stems. Besides these it seems there are specific aspectual markers with concrete emotional meanings, and some verbal modifying affixes with emotional meanings also, and finally a group of nominal modifiers which express the speakers emotional evaluation of another person....... In this paper I’ll consider these questions: -do emotions terms have special linguistic constructions or not? - do emotions verbs have specificity in comparison with other verbs, for example in telicity? -are special affixes linked with the verbs of emotion? - do emotions verbs have special temporal...
Kwon, Kyongboon; Hanrahan, Amanda R.; Kupzyk, Kevin A.
We examined emotional expressivity (i.e., happiness, sadness, and anger) and emotion regulation (regulation of exuberance, sadness, and anger) as they relate to academic functioning (motivation, engagement, and achievement). Also, we tested the premise that emotional expressivity and emotion regulation are indirectly associated with achievement…
Coats, Abby Heckman; Blanchard-Fields, Fredda
Young, middle-aged, and older adults' emotion regulation strategies in interpersonal problems were examined. Participants imagined themselves in anger- or sadness-eliciting situations with a close friend. Factor analyses of a new questionnaire supported a 4-factor model of emotion regulation strategies, including passivity, expressing emotions, seeking emotional information or support, and solving the problem. Results suggest that age differences in emotion regulation (such as older adults' increased endorsement of passive emotion regulation relative to young adults) are partially due to older adults' decreased ability to integrate emotion and cognition, increased prioritization of emotion regulation goals, and decreased tendency to express anger. (c) 2008 APA, all rights reserved.
Chu, Min-Yi; Li, Xu; Lv, Qin-Yu; Yl, Zheng-Hui; Cheung, Eric F C; Chan, Raymond C K
Impairments in emotional experience and expression have been observed in patients with schizophrenia. However, most previous studies have been limited to either emotional experience (especially anhedonia) or expression. Few studies have examined both the experience and expression of emotion in schizophrenia patients at the same time. The present study aimed to examine pleasure experience and emotion expression in patients with schizophrenia. In particular, we specifically examined the relationship between emotion impairments (both pleasure experience and expression) and negative symptoms. One hundred and fifty patients completed the Temporal Experience of Pleasure Scale and Emotional Expressivity Scale. Schizophrenia patients exhibited deficits in experiencing pleasure, but showed intact reported emotion expression. Patients with prominent negative symptoms showed reduced anticipatory pleasure, especially in abstract anticipatory pleasure. The present findings suggest that patients with schizophrenia have deficits in pleasure experience, while their abilities to express emotion appear intact. Such deficits are more severe in patients with prominent negative symptoms.
Coates, Dominiek D; Howe, Deborah
Mental health work presents problems for staff over and above those encountered in other organisations, including other areas of healthcare. Healthcare workers, in particular mental health workers, have poorer job satisfaction and higher job burnout and turnover compared with established norms for other occupational groups. To make sense of why healthcare workers experience high levels of burnout, a strong body of literature points to the emotionally demanding nature of people-work. The negative effects of mental health work on employee health can be mitigated by the provision of appropriate job resources and wellbeing initiatives. As to develop initiatives that appropriately target staff sources of stress and needs, it is important to engage staff in this process. As such, Children and Young People's Mental Health (CYPMH) and headspace Gosford, in Australia, New South Wales (NSW), developed a survey to identify how staff experience and manage the emotional demands of mental health work, what they identify as key stressors and which initiatives they would like to see implemented. Fifty-five staff (response rate of 73 %) completed the survey, and the results suggest that while staff find the work emotionally demanding, they do not appear to be emotionally exhausted and report administrative rather than client issues as their primary concerns. While a strong body of literature identifies the management of emotions in the workplace as a significant cause of stress, organisational stressors such as working in a bureaucratic environment are also important to understanding staff wellbeing.
Sandvoll, Anne Marie; Grov, Ellen Karine; Kristoffersen, Kjell; Hauge, Solveig
Caring practice in nursing homes is a complex topic, especially the challenges of meeting the basic needs of residents when their behaviour evokes difficult emotions. Cognitive and physical changes related to aging and disability can contribute to behaviours considered to be unacceptable. For example, resident behaviours such as spitting, making a mess with food or grinding teeth are behaviours that most people do not want to see, hear or experience. The aim of this study was to gain a deeper understanding of how nursing home staff members deal with such behaviours in care situations. This article draws on ethnographic data to describe how nursing home staff members manage unpleasant resident behaviours. The study was based on two long-term units in two Norwegian public nursing homes. The Region's Medical Ethics Committee and the Norwegian Social Science Data Services granted approval. In total, 45 participants (37 nursing aides and eight nurses) agreed to participate in this study. Ten of the participants were interviewed at the end of the field study. This study indicates that nursing home staff members experience difficult emotions related to some residents' behaviours. However, they found these feelings difficult to express and rarely verbalized them openly. In addition, they were characterized by a strong obligation to help all residents, despite their own feelings. Therefore, it appears that an inner struggle occurs as a part of everyday practice. Despite these difficult emotions, nursing staff members believed that they needed to manage their responses and continued to offer good care to all residents. These findings extend our understanding of this unarticulated part of nursing home practice.
Haider, Badeea; Khan, Sumaira; Anwar, Kanwal
There are clear differences in the extent to which people express their emotions. These differences in emotional expressions have long interested researchers and are relevant to several areas of psychology. The research topic is emotional expressions between hostalize male and female in. The sum of 24 students (12 male and 12 female) were selected…
Chen, Antao; Cui, Qian; Zhang, Qinglin
The present study investigated whether emotional conflict and emotional conflict adaptation could be triggered by unconscious emotional information as assessed in a backward-masked affective priming task. Participants were instructed to identify the valence of a face (e.g., happy or sad) preceded by a masked happy or sad face. The results of two experiments revealed the emotional conflict effect but no emotional conflict adaptation effect. This demonstrates that emotional conflict can be triggered by unconsciously presented emotional information, but participants may not adjust their subsequent performance trial-by trial to reduce this conflict. PMID:23409084
Camras, Linda A.
To make the point that infant emotions are more dynamic than suggested by Differential Emotions Theory, which maintains that infants show the same prototypical facial expressions for emotions as adults do, this paper explores two questions: (1) when infants experience an emotion, do they always show the corresponding prototypical facial…
Lundqvist, Lars-Olov; Carlsson, Fredrik; Hilmersson, Per; Juslin, Patrik N.
A crucial issue in research on music and emotion is whether music evokes genuine emotional responses in listeners (the emotivist position) or whether listeners merely perceive emotions expressed by the music (the cognitivist position). To investigate this issue, we measured self-reported emotion, facial muscle activity, and autonomic activity in…
McCloughen, Andrea; Foster, Kim
To identify challenging interpersonal interactions experienced by nursing and pharmacy students during clinical placement, and strategies used to manage those situations. Healthcare students and staff experience elevated stress when exposed to dynamic clinical environments, complex care and challenging professional relationships. Emotionally intelligent behaviours are associated with appropriate recognition and management of emotions evoked by stressful experiences and development of effective relationships. Nursing and pharmacy students' use of emotionally intelligent behaviours to manage challenging interpersonal situations is not well known. A qualitative design, using semi-structured interviews to explore experiences of challenging interpersonal situations during clinical placement (Phase two of a larger mixed-methods study). Final-year Australian university nursing and pharmacy students (n = 20) were purposefully recruited using a range of Emotional Intelligence scores (derived in Phase one), measured using the GENOS Emotional intelligence Inventory (concise version). Challenging interpersonal situations involving student-staff and intrastaff conflict, discourteous behaviour and criticism occurred during clinical placement. Students used personal and relational strategies, incorporating emotionally intelligent behaviours, to manage these encounters. Strategies included reflecting and reframing, being calm, controlling discomfort and expressing emotions appropriately. Emotionally intelligent behaviours are effective to manage stressful interpersonal interactions. Methods for strengthening these behaviours should be integrated into education of nursing and pharmacy students and qualified professionals. Education within the clinical/workplace environment can incorporate key interpersonal skills of collaboration, social interaction and reflection, while also attending to sociocultural contexts of the healthcare setting. Students and staff are frequently exposed
Full Text Available Emotional processing without conscious awareness plays an important role in human social interaction. Several behavioral studies reported that subliminal presentation of photographs of emotional facial expressions induces unconscious emotional processing. However, it was difficult to elicit strong and robust effects using this method. We hypothesized that dynamic presentations of facial expressions would enhance subliminal emotional effects and tested this hypothesis with two experiments. Fearful or happy facial expressions were presented dynamically or statically in either the left or the right visual field for 20 (Experiment 1 and 30 (Experiment 2 ms. Nonsense target ideographs were then presented, and participants reported their preference for them. The results consistently showed that dynamic presentations of emotional facial expressions induced more evident emotional biases toward subsequent targets than did static ones. These results indicate that dynamic presentations of emotional facial expressions induce more evident unconscious emotional processing.
Gibson, Donald E.
This exercise explores how organizations affect individuals' feelings and expressions of emotion. Although recent attention by management theorists suggests that emotions are an important aspect of organizational life, people's actual experience of emotions at work often do not reflect this emphasis: Work-place emotions remain, in large part,…
Begeer, Sander; Rieffe, Carolien; Terwogt, Mark Meerum; Stockmann, Lex
High-functioning children in the autism spectrum are frequently noted for their impaired attention to facial expressions of emotions. In this study, we examined whether attention to emotion cues in others could be enhanced in children with autism, by varying the relevance of children's attention to emotion expressions. Twenty-eight…
Tielman, M.; Neerincx, M.A.; Meuer, J.J.; Looije, R.
Expressive behaviour is a vital aspect of human interaction. A model for adaptive emotion expression was developed for the Nao robot. The robot has an internal arousal and va- lence value, which are in uenced by the emotional state of its interaction partner and emotional occurrences such as win-
Background: The prevention of relapse is one of the major aims of treatment of emotional disorders. Expressed emotion (EE) is one concept that has been associated with relapse. The study is aimed at studying the relationship between expressed emotion and the clinical characteristics of patients with schizophrenia.
Tielman, Myrthe; Neerincx, Mark A.; Meyer, John-Jules Ch.; Looije, Rosemarijn
Expressive behaviour is a vital aspect of human interaction. A model for adaptive emotion expression was developed for the Nao robot. The robot has an internal arousal and va- lence value, which are in uenced by the emotional state of its interaction partner and emotional occurrences such as win-
Bowman, Deborah; Hughes, Patricia
Problem-based learning (PBL) is a method of teaching and learning that is used increasingly in medical and health care curricula worldwide. The literature on PBL is considerable and continues to develop. One important aspect of PBL is that students and tutors spend a lot of time together and this fosters an informal atmosphere that may encourage intimacy. The existing literature on PBL has not considered the emotional and psychological aspects of PBL nor the concomitant need for staff support and development. We present a discussion paper considering the ways in which educationalists using or considering using PBL could be informed by the psychological and psychotherapeutic literature on groups and group dynamics, in particular the work of Wilfred Bion. We discuss how PBL tutorials may arouse emotional responses that could result in unconsidered behaviours that impede student learning. We argue that faculty and PBL tutors need to agree and remain alert to the primary task of the group. Faculty should develop professional standards for tutors to use as reference points to ensure the group stays on course and achieves its intended outcomes. We conclude that greater attention should be paid by educationalists and faculty to identifying possible tutor emotional responses as part of initial PBL tutor training and ongoing staff development. We offer vignettes that have been successfully used in training and staff development at a UK medical school to demonstrate the practical application of our theoretical discussion.
Fang, X.; Sauter, D.A.; van Kleef, G.A.
Although perceivers often agree about the primary emotion that is conveyed by a particular expression, observers may concurrently perceive several additional emotions from a given facial expression. In the present research, we compared the perception of two types of nonintended emotions in Chinese
Fang, Xia; Sauter, Disa A.; Van Kleef, Gerben A.
Although perceivers often agree about the primary emotion that is conveyed by a particular expression, observers may concurrently perceive several additional emotions from a given facial expression. In the present research, we compared the perception of two types of nonintended emotions in Chinese and Dutch observers viewing facial expressions: emotions which were morphologically similar to the intended emotion and emotions which were morphologically dissimilar to the intended emotion. Findings were consistent across two studies and showed that (a) morphologically similar emotions were endorsed to a greater extent than dissimilar emotions and (b) Chinese observers endorsed nonintended emotions more than did Dutch observers. Furthermore, the difference between Chinese and Dutch observers was more pronounced for the endorsement of morphologically similar emotions than of dissimilar emotions. We also obtained consistent evidence that Dutch observers endorsed nonintended emotions that were congruent with the preceding expressions to a greater degree. These findings suggest that culture and morphological similarity both influence the extent to which perceivers see several emotions in a facial expression. PMID:29386689
Fang, Xia; Sauter, Disa A; Van Kleef, Gerben A
Although perceivers often agree about the primary emotion that is conveyed by a particular expression, observers may concurrently perceive several additional emotions from a given facial expression. In the present research, we compared the perception of two types of nonintended emotions in Chinese and Dutch observers viewing facial expressions: emotions which were morphologically similar to the intended emotion and emotions which were morphologically dissimilar to the intended emotion. Findings were consistent across two studies and showed that (a) morphologically similar emotions were endorsed to a greater extent than dissimilar emotions and (b) Chinese observers endorsed nonintended emotions more than did Dutch observers. Furthermore, the difference between Chinese and Dutch observers was more pronounced for the endorsement of morphologically similar emotions than of dissimilar emotions. We also obtained consistent evidence that Dutch observers endorsed nonintended emotions that were congruent with the preceding expressions to a greater degree. These findings suggest that culture and morphological similarity both influence the extent to which perceivers see several emotions in a facial expression.
The conventional GUI is more mechanical and does not recognize or communicate emotions. The modern GUIs are trying to infer the likely emotional state and personality of the user and communicate through a corresponding emotional state. Emotions are expressed in graphical icons, sounds, pictures and other means. The emotions are found to be useful in especially in communication software, interactive learning systems, robotics and other adaptive environments. Various mechanisms have been develo...
The main purpose of the diploma thesis is to research how often, in which situations and how do preschool girls and boys express their emotions. I start the theoretical part by explaining general characteristics of emotions, following by characteristics of individual basic emotions, namely joy, anger, sadness and fear. After that I present an overview of the development of emotions and the role of preschool teacher in the emotional development of children. At the end I focus on gender di...
Full Text Available The mechanisms to communicate emotions have dramatically changed in the last 10 years with social networks, where users massively communicate their emotional states by using the Internet. However, people with socialization problems have difficulty expressing their emotions verbally or interpreting the environment and providing an appropriate emotional response. In this paper, a novel solution called the Emotion-Bracelet is presented that combines a hardware device and a software system. The proposed approach identifies the polarity and emotional intensity of texts published on a social network site by performing real-time processing using a web service. It also shows emotions with a LED matrix using five emoticons that represent positive, very positive, negative, very negative, and neutral states. The Emotion-Bracelet is designed to help people express their emotions in a non-intrusive way, thereby expanding the social aspect of human emotions.
Martinez, Alicia; Estrada, Hugo; Molina, Alejandra; Mejia, Manuel; Perez, Joaquin
The mechanisms to communicate emotions have dramatically changed in the last 10 years with social networks, where users massively communicate their emotional states by using the Internet. However, people with socialization problems have difficulty expressing their emotions verbally or interpreting the environment and providing an appropriate emotional response. In this paper, a novel solution called the Emotion-Bracelet is presented that combines a hardware device and a software system. The proposed approach identifies the polarity and emotional intensity of texts published on a social network site by performing real-time processing using a web service. It also shows emotions with a LED matrix using five emoticons that represent positive, very positive, negative, very negative, and neutral states. The Emotion-Bracelet is designed to help people express their emotions in a non-intrusive way, thereby expanding the social aspect of human emotions.
Zijlmans, L.J.M.; Embregts, P.J.C.M.; Gerits, L.; Bosman, A.M.T.; Derksen, J.
Background Recent research addressed the relationship between staff behaviour and challenging behaviour of individuals with an intellectual disability (ID). Consequently, research on interventions aimed at staff is warranted. The present study focused on the effectiveness of a staff training aimed
Recent studies have substantiated that preschool children are able to perceive the emotional meaning in music. However, to date, research has not paid enough attention to children’s emotional expression in music. The current study investigated 3- (N=18) and 5- (N=19) year-old children’s ability to express emotion with music by manipulating musical factors (pitch, loudness and tempo) to indicate one of three basic emotions (happiness, sadness or angriness). In order to facilitate the recogn...
Xue, Henry; Gertner, Izidor
In the human-computer interaction (HCI) process it is desirable to have an artificial intelligent (AI) system that can identify and categorize human emotions from facial expressions. Such systems can be used in security, in entertainment industries, and also to study visual perception, social interactions and disorders (e.g. schizophrenia and autism). In this work we survey and compare the performance of different feature extraction algorithms and classification schemes. We introduce a faster feature extraction method that resizes and applies a set of filters to the data images without sacrificing the accuracy. In addition, we have enhanced SVM to multiple dimensions while retaining the high accuracy rate of SVM. The algorithms were tested using the Japanese Female Facial Expression (JAFFE) Database and the Database of Faces (AT&T Faces).
Kalokerinos, Elise K; Greenaway, Katharine H; Casey, James P
It is generally considered socially undesirable to suppress the expression of positive emotion. However, previous research has not considered the role that social context plays in governing appropriate emotion regulation. We investigated a context in which it may be more appropriate to suppress than express positive emotion, hypothesizing that positive emotion expressions would be considered inappropriate when the valence of the expressed emotion (i.e., positive) did not match the valence of the context (i.e., negative). Six experiments (N = 1,621) supported this hypothesis: when there was a positive emotion-context mismatch, participants rated targets who suppressed positive emotion as more appropriate, and evaluated them more positively than targets who expressed positive emotion. This effect occurred even when participants were explicitly made aware that suppressing targets were experiencing mismatched emotion for the context (e.g., feeling positive in a negative context), suggesting that appropriate emotional expression is key to these effects. These studies are among the first to provide empirical evidence that social costs to suppression are not inevitable, but instead are dependent on context. Expressive suppression can be a socially useful emotion regulation strategy in situations that call for it. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Full Text Available How universal are men and women’s attitudes toward the expression of emotion? How similar are the emotions that men and women from various ethnic groups experience and express in their close love relationships? In this study, 144 men and 307 women of European, Chinese, Filipino, Hawaiian, and Japanese ancestry were asked about their ideologies as to how people ought to deal with strong emotions in close relationships, how often they themselves felt a variety of emotions, and how they dealt with such feelings in close relationships. Finally, they were asked how satisfied they were with their close relationships. Men and women appeared to possess different emotional ideologies. Women tended to favor direct expression of emotion; men to favor emotional management. People of Chinese, European, Filipino, Hawaiian, and Japanese ancestry also possessed different ideologies as to how people ought to deal with strong emotions in intimate relationships.
Gaspar, Tina; Labor, Marina; Jurić, Iva; Dumancić, Dijana; Ilakovac, Vesna; Heffer, Marija
The recognition of basic emotions in everyday communication involves interpretation of different visual and auditory clues. The ability to recognize emotions is not clearly determined as their presentation is usually very short (micro expressions), whereas the recognition itself does not have to be a conscious process. We assumed that the recognition from facial expressions is selected over the recognition of emotions communicated through music. In order to compare the success rate in recognizing emotions presented as facial expressions or in classical music works we conducted a survey which included 90 elementary school and 87 high school students from Osijek (Croatia). The participants had to match 8 photographs of different emotions expressed on the face and 8 pieces of classical music works with 8 offered emotions. The recognition of emotions expressed through classical music pieces was significantly less successful than the recognition of emotional facial expressions. The high school students were significantly better at recognizing facial emotions than the elementary school students, whereas girls were better than boys. The success rate in recognizing emotions from music pieces was associated with higher grades in mathematics. Basic emotions are far better recognized if presented on human faces than in music, possibly because the understanding of facial emotions is one of the oldest communication skills in human society. Female advantage in emotion recognition was selected due to the necessity of their communication with the newborns during early development. The proficiency in recognizing emotional content of music and mathematical skills probably share some general cognitive skills like attention, memory and motivation. Music pieces were differently processed in brain than facial expressions and consequently, probably differently evaluated as relevant emotional clues.
Volkova, Ekaterina; de la Rosa, Stephan; Bülthoff, Heinrich H.; Mohler, Betty
Emotion expression in human-human interaction takes place via various types of information, including body motion. Research on the perceptual-cognitive mechanisms underlying the processing of natural emotional body language can benefit greatly from datasets of natural emotional body expressions that facilitate stimulus manipulation and analysis. The existing databases have so far focused on few emotion categories which display predominantly prototypical, exaggerated emotion expressions. Moreover, many of these databases consist of video recordings which limit the ability to manipulate and analyse the physical properties of these stimuli. We present a new database consisting of a large set (over 1400) of natural emotional body expressions typical of monologues. To achieve close-to-natural emotional body expressions, amateur actors were narrating coherent stories while their body movements were recorded with motion capture technology. The resulting 3-dimensional motion data recorded at a high frame rate (120 frames per second) provides fine-grained information about body movements and allows the manipulation of movement on a body joint basis. For each expression it gives the positions and orientations in space of 23 body joints for every frame. We report the results of physical motion properties analysis and of an emotion categorisation study. The reactions of observers from the emotion categorisation study are included in the database. Moreover, we recorded the intended emotion expression for each motion sequence from the actor to allow for investigations regarding the link between intended and perceived emotions. The motion sequences along with the accompanying information are made available in a searchable MPI Emotional Body Expression Database. We hope that this database will enable researchers to study expression and perception of naturally occurring emotional body expressions in greater depth. PMID:25461382
Full Text Available Emotion expression in human-human interaction takes place via various types of information, including body motion. Research on the perceptual-cognitive mechanisms underlying the processing of natural emotional body language can benefit greatly from datasets of natural emotional body expressions that facilitate stimulus manipulation and analysis. The existing databases have so far focused on few emotion categories which display predominantly prototypical, exaggerated emotion expressions. Moreover, many of these databases consist of video recordings which limit the ability to manipulate and analyse the physical properties of these stimuli. We present a new database consisting of a large set (over 1400 of natural emotional body expressions typical of monologues. To achieve close-to-natural emotional body expressions, amateur actors were narrating coherent stories while their body movements were recorded with motion capture technology. The resulting 3-dimensional motion data recorded at a high frame rate (120 frames per second provides fine-grained information about body movements and allows the manipulation of movement on a body joint basis. For each expression it gives the positions and orientations in space of 23 body joints for every frame. We report the results of physical motion properties analysis and of an emotion categorisation study. The reactions of observers from the emotion categorisation study are included in the database. Moreover, we recorded the intended emotion expression for each motion sequence from the actor to allow for investigations regarding the link between intended and perceived emotions. The motion sequences along with the accompanying information are made available in a searchable MPI Emotional Body Expression Database. We hope that this database will enable researchers to study expression and perception of naturally occurring emotional body expressions in greater depth.
Deady, D K; North, N T; Allan, D; Smith, M J Law; O'Carroll, R E
The prevailing view on the effects of spinal cord injury (SCI) on emotion is that it dampens emotional experience due to a loss of peripheral bodily feedback, with the higher the lesion on the spinal cord the greater the reduction in the intensity of emotional experience. This view persists despite many studies showing an absence of such an emotional impairment in people with SCI. This study specifically aimed to investigate whether total cervical-6 spinal cord transection (i) reduces emotional expressivity and emotional awareness (ii) impairs memory for emotional material. The study contained three groups: 24 patients with SCI, 20 orthopaedic injury control (OIC) patients and 20 young adult controls. A mixed factor design was employed to examine between group and within subject differences. Participants completed the Levels of Emotional Awareness Scale (LEAS), the Berkeley Expressivity Questionnaire (BEQ), and viewed an emotionally arousing slide presentation. Thirty minutes post viewing, participants completed memory tests for the presentation. SCI patients reported greater present levels of emotional expressivity compared with perceived levels prior to their injuries. SCI and OIC groups did not differ on any of the emotional awareness variables. There was also no evidence that SCI leads to impairment in memory for emotional events. This study's findings contradict the mainstream view in the cognitive neuroscience of emotion that SCI dampens emotional experience.
Zijlmans, L. J. M.; Embregts, P. J. C. M.; Gerits, L.; Bosman, A. M. T.; Derksen, J. J. L.
Background: Staff working with clients with intellectual disabilities (ID) who display challenging behaviour may contribute to the continuation of this behaviour, because it causes emotional reactions such as anxiety, anger and annoyance, which may prohibit adequate response behaviour. To enhance staff behaviour and treatment skills a training…
Full Text Available Although most people can identify facial expressions of emotions well, they still differ in this ability. According to embodied simulation theories understanding emotions of others is fostered by involuntarily mimicking the perceived expressions, causing a "reactivation" of the corresponding mental state. Some studies suggest automatic facial mimicry during expression viewing; however, findings on the relationship between mimicry and emotion perception abilities are equivocal. The present study investigated individual differences in emotion perception and its relationship to facial muscle responses - recorded with electromyogram (EMG--in response to emotional facial expressions. N° = °269 participants completed multiple tasks measuring face and emotion perception. EMG recordings were taken from a subsample (N° = °110 in an independent emotion classification task of short videos displaying six emotions. Confirmatory factor analyses of the m. corrugator supercilii in response to angry, happy, sad, and neutral expressions showed that individual differences in corrugator activity can be separated into a general response to all faces and an emotion-related response. Structural equation modeling revealed a substantial relationship between the emotion-related response and emotion perception ability, providing evidence for the role of facial muscle activation in emotion perception from an individual differences perspective.
Künecke, Janina; Hildebrandt, Andrea; Recio, Guillermo; Sommer, Werner; Wilhelm, Oliver
Although most people can identify facial expressions of emotions well, they still differ in this ability. According to embodied simulation theories understanding emotions of others is fostered by involuntarily mimicking the perceived expressions, causing a "reactivation" of the corresponding mental state. Some studies suggest automatic facial mimicry during expression viewing; however, findings on the relationship between mimicry and emotion perception abilities are equivocal. The present study investigated individual differences in emotion perception and its relationship to facial muscle responses - recorded with electromyogram (EMG)--in response to emotional facial expressions. N° = °269 participants completed multiple tasks measuring face and emotion perception. EMG recordings were taken from a subsample (N° = °110) in an independent emotion classification task of short videos displaying six emotions. Confirmatory factor analyses of the m. corrugator supercilii in response to angry, happy, sad, and neutral expressions showed that individual differences in corrugator activity can be separated into a general response to all faces and an emotion-related response. Structural equation modeling revealed a substantial relationship between the emotion-related response and emotion perception ability, providing evidence for the role of facial muscle activation in emotion perception from an individual differences perspective.
Muhammed Tayyib Kadak
Full Text Available Autism is a genetically transferred neurodevelopmental disorder characterized by severe and permanent deficits in many interpersonal relation areas like communication, social interaction and emotional responsiveness. Patients with autism have deficits in face recognition, eye contact and recognition of emotional expression. Both recognition of face and expression of facial emotion carried on face processing. Structural and functional impairment in fusiform gyrus, amygdala, superior temporal sulcus and other brain regions lead to deficits in recognition of face and facial emotion. Therefore studies suggest that face processing deficits resulted in problems in areas of social interaction and emotion in autism. Studies revealed that children with autism had problems in recognition of facial expression and used mouth region more than eye region. It was also shown that autistic patients interpreted ambiguous expressions as negative emotion. In autism, deficits related in various stages of face processing like detection of gaze, face identity, recognition of emotional expression were determined, so far. Social interaction impairments in autistic spectrum disorders originated from face processing deficits during the periods of infancy, childhood and adolescence. Recognition of face and expression of facial emotion could be affected either automatically by orienting towards faces after birth, or by “learning” processes in developmental periods such as identity and emotion processing. This article aimed to review neurobiological basis of face processing and recognition of emotional facial expressions during normal development and in autism.
Prosen, Simona; Vitulic, Helena Smrtnik; Škraban, Olga Poljšak
Emotions are an integral part of "classroom life" and are experienced in teacher-student interactions quite often (Hosotani & Imai-Matsumura, 2011). The present study focuses on teachers' emotions in classrooms. Its purpose is to establish which emotions are expressed by teachers in their interactions with students, the triggering…
Elliott, David J.
What do musicians, critics, and listeners mean when they use emotion-words to describe a piece of instrumental music? How can "pure" musical sounds "express" emotions such as joyfulness, sadness, anguish, optimism, and anger? Sounds are not living organisms; sounds cannot feel emotions. Yet many people around the world believe they hear emotions…
Full Text Available Emotions are an integral part of “classroom life” and are experienced in teacher-student interactions quite often (Hosotani & Imai-Matsumura, 2011. The present study focuses on teachers’ emotions in classrooms. Its purpose is to establish which emotions are expressed by teachers in their interactions with students, the triggering situations of the two most frequent emotions, and their level of intensity and suitability. Teachers’ emotions were observed by students of primary education during their practical experience work, in grades one to five. They used a scheme constructed for observing different aspects of emotions. The observations of 108 teachers in 93 primary schools from various Slovenian regions were gathered. The results show that primary school teachers express various pleasant and unpleasant emotions, with unpleasant emotions prevailing. The average frequency of teachers’ emotion expression decreased from grade one to five. Anger was the most frequently expressed emotion (N = 261, followed by joy (N = 151. Teachers’ anger and joy were triggered in different situations: anger predominantly when students lacked discipline and joy predominantly in situations of students’ academic achievement. The intensity of expressed anger and joy was moderate in all five grades, while the assessed suitability of these two emotions was high.
Hess, Ursula; Adams, Reginald B; Kleck, Robert E
Western gender stereotypes describe women as affiliative and more likely to show happiness and men as dominant and more likely to show anger. The authors assessed the hypothesis that the gender-stereotypic effects on perceptions of anger and happiness are partially mediated by facial appearance markers of dominance and affiliation by equating men's and women's faces for these cues. In 2 studies, women were rated as more angry and men as more happy-a reversal of the stereotype. Ratings of sadness, however, were not systematically affected. It is posited that markers of affiliation and dominance, themselves confounded with gender, interact with the expressive cues for anger and happiness to produce emotional perceptions that have been viewed as simple gender stereotypes. copyright (c) 2004 APA, all rights reserved.
Van Kleef, Gerben A; van den Berg, Helma; Heerdink, Marc W
Despite a long-standing interest in the intrapersonal role of affect in persuasion, the interpersonal effects of emotions on persuasion remain poorly understood-how do one person's emotional expressions shape others' attitudes? Drawing on emotions as social information (EASI) theory (Van Kleef, 2009), we hypothesized that people use the emotional expressions of others to inform their own attitudes, but only when they are sufficiently motivated and able to process those expressions. Five experiments support these ideas. Participants reported more positive attitudes about various topics after seeing a source's sad (rather than happy) expressions when topics were negatively framed (e.g., abandoning bobsleighing from the Olympics). Conversely, participants reported more positive attitudes after seeing happy (rather than sad) expressions when topics were positively framed (e.g., introducing kite surfing at the Olympics). This suggests that participants used the source's emotional expressions as information when forming their own attitudes. Supporting this interpretation, effects were mitigated when participants' information processing was undermined by cognitive load or was chronically low. Moreover, a source's anger expressions engendered negative attitude change when directed at the attitude object and positive change when directed at the recipient's attitude. Effects occurred regardless of whether emotional expressions were manipulated through written words, pictures of facial expressions, film clips containing both facial and vocal emotional expressions, or emoticons. The findings support EASI theory and indicate that emotional expressions are a powerful source of social influence. (c) 2015 APA, all rights reserved).
Full Text Available We report the development of two simple, objective, psychophysical measures of the ability to discriminate facial expressions of emotion that vary in intensity from a neutral facial expression and to discriminate between varying intensities of emotional facial expression. The stimuli were created by morphing photographs of models expressing four basic emotions, anger, disgust, happiness and sadness with neutral expressions. Psychometric functions were obtained for 15 healthy young adults using the Method of Constant Stimuli with a two-interval forced-choice procedure. Individual data points were fitted by Quick functions for each task and each emotion, allowing estimates of absolute thresholds and slopes. The tasks give objective and sensitive measures of the basic perceptual abilities required for perceiving and interpreting emotional facial expressions.
Yoon, HeungSik; Kim, Shin Ah; Kim, Sang Hee
An individual's responses to emotional information are influenced not only by the emotional quality of the information, but also by the context in which the information is presented. We hypothesized that facial expressions of happiness and anger would serve as primes to modulate subjective and neural responses to subsequently presented negative information. To test this hypothesis, we conducted a functional MRI study in which the brains of healthy adults were scanned while they performed an emotion-rating task. During the task, participants viewed a series of negative and neutral photos, one at a time; each photo was presented after a picture showing a face expressing a happy, angry, or neutral emotion. Brain imaging results showed that compared with neutral primes, happy facial primes increased activation during negative emotion in the dorsal anterior cingulated cortex and the right ventrolateral prefrontal cortex, which are typically implicated in conflict detection and implicit emotion control, respectively. Conversely, relative to neutral primes, angry primes activated the right middle temporal gyrus and the left supramarginal gyrus during the experience of negative emotion. Activity in the amygdala in response to negative emotion was marginally reduced after exposure to happy primes compared with angry primes. Relative to neutral primes, angry facial primes increased the subjectively experienced intensity of negative emotion. The current study results suggest that prior exposure to facial expressions of emotions modulates the subsequent experience of negative emotion by implicitly activating the emotion-regulation system.
Deng, Yaling; Chang, Lei; Yang, Meng; Huo, Meng
The present study investigated gender differences in both emotional experience and expressivity. Heart rate (HR) was recorded as an indicator of emotional experience while the participants watched 16 video clips that induced eight types of emotion (sadness, anger, horror, disgust, neutrality, amusement, surprise, and pleasure). We also asked the participants to report valence, arousal, and motivation as indicators of emotional expressivity. Overall, the results revealed gender differences in emotional experience and emotional expressivity. When watching videos that induced anger, amusement, and pleasure, men showed larger decreases in HR, whereas women reported higher levels of arousal. There was no gender difference in HR when the participants watched videos that induced horror and disgust, but women reported lower valence, higher arousal, and stronger avoidance motivation than did men. Finally, no gender difference was observed in sadness or surprise, although there was one exception—women reported higher arousal when watching videos that induced sadness. The findings suggest that, when watching videos that induce an emotional response, men often have more intense emotional experiences, whereas women have higher emotional expressivity, particularly for negative emotions. In addition, gender differences depend on the specific emotion type but not the valence. PMID:27362361
Leppanen, Jenni; Dapelo, Marcela Marin; Davies, Helen; Lang, Katie; Treasure, Janet; Tchanturia, Kate
Background:Problems with social-emotional processing are known to be an important contributor to the development and maintenance of eating disorders (EDs). Diminished facial communication of emotion has been frequently reported in individuals with anorexia nervosa (AN). Less is known about facial expressivity in bulimia nervosa (BN) and in people who have recovered from AN (RecAN). This study aimed to pilot the use of computerised facial expression analysis software to investigate emotion exp...
Gašpar, Tina; Labor, Marina; Jurić, Iva; Dumančić, Dijana; Ilakovac, Vesna; Heffer, Marija
The recognition of basic emotions in everyday communication involves interpretation of different visual and auditory clues. The ability to recognize emotions is not clearly determined as their presentation is usually very short (micro expressions), whereas the recognition itself does not have to be a conscious process. We assumed that the recognition from facial expressions is selected over the recognition of emotions communicated through music. In order to compare the success rate in recogni...
Davies, Helen; Schmidt, Ulrike; Stahl, Daniel; Tchanturia, Kate
To use an experimental paradigm to assess facial expression, subjective experience of emotion and the relationship between them in people with anorexia nervosa (AN). Film clips are used to elicit emotion and participants' facial expression and subjective experience are recorded. Thirty inpatients with AN and 34 healthy control (HC) women are included in the study. People with AN are less facially expressive than HC while watching positive and negative film clips and report feeling less positive emotion than HC but not less negative emotion. People with AN look away significantly more than HC during the negative film clip. Duration of illness and depression relate to attenuated positive facial expression and eating pathology to attenuated negative facial expression. This experimental study supports self report studies showing people with AN attenuate emotional expression and avoid negative affect. Such behavior may affect social interaction and contribute to the maintenance of the disorder. Copyright © 2010 Wiley Periodicals, Inc.
Leppanen, Jenni; Dapelo, Marcela Marin; Davies, Helen; Lang, Katie; Treasure, Janet; Tchanturia, Kate
Problems with social-emotional processing are known to be an important contributor to the development and maintenance of eating disorders (EDs). Diminished facial communication of emotion has been frequently reported in individuals with anorexia nervosa (AN). Less is known about facial expressivity in bulimia nervosa (BN) and in people who have recovered from AN (RecAN). This study aimed to pilot the use of computerised facial expression analysis software to investigate emotion expression across the ED spectrum and recovery in a large sample of participants. 297 participants with AN, BN, RecAN, and healthy controls were recruited. Participants watched film clips designed to elicit happy or sad emotions, and facial expressions were then analysed using FaceReader. The finding mirrored those from previous work showing that healthy control and RecAN participants expressed significantly more positive emotions during the positive clip compared to the AN group. There were no differences in emotion expression during the sad film clip. These findings support the use of computerised methods to analyse emotion expression in EDs. The findings also demonstrate that reduced positive emotion expression is likely to be associated with the acute stage of AN illness, with individuals with BN showing an intermediate profile.
Herndon, Kristina J.; Bailey, Craig S.; Shewark, Elizabeth A.; Denham, Susanne A.; Bassett, Hideko H.
Children's expression and regulation of emotions are building blocks of their experiences in classrooms. Thus, the authors' primary goal was to investigate whether preschoolers' expression or ability to regulate emotions were associated with teachers' ratings of school adjustment. A secondary goal was to investigate how boys and girls differed…
Full Text Available Problems with social-emotional processing are known to be an important contributor to the development and maintenance of eating disorders (EDs. Diminished facial communication of emotion has been frequently reported in individuals with anorexia nervosa (AN. Less is known about facial expressivity in bulimia nervosa (BN and in people who have recovered from AN (RecAN. This study aimed to pilot the use of computerised facial expression analysis software to investigate emotion expression across the ED spectrum and recovery in a large sample of participants.297 participants with AN, BN, RecAN, and healthy controls were recruited. Participants watched film clips designed to elicit happy or sad emotions, and facial expressions were then analysed using FaceReader.The finding mirrored those from previous work showing that healthy control and RecAN participants expressed significantly more positive emotions during the positive clip compared to the AN group. There were no differences in emotion expression during the sad film clip.These findings support the use of computerised methods to analyse emotion expression in EDs. The findings also demonstrate that reduced positive emotion expression is likely to be associated with the acute stage of AN illness, with individuals with BN showing an intermediate profile.
Bolzani Dinehart, Laura H.; Messinger, Daniel S.; Acosta, Susan I.; Cassel, Tricia; Ambadar, Zara; Cohn, Jeffrey
Adults' perceptions provide information about the emotional meaning of infant facial expressions. This study asks whether similar facial movements influence adult perceptions of emotional intensity in both infant positive (smile) and negative (cry face) facial expressions. Ninety-five college students rated a series of naturally occurring and…
Ioannou, Korina; Fox, John R E
Recent research has documented links between eating disorder (ED) symptomatology and emotional expression deficits. Relevant theoretical models have alluded to the functional role of disordered eating in alleviating affect that is felt to be otherwise unmanageable and threatening. Nevertheless, research examining ED individuals' perceptions of emotional states has been sparse, while empirical studies have predominantly focused on global conceptualizations of emotion, failing to address discrete affect states. The current study had three aims: (a) to determine the relation between ED symptomatology and emotional expression, in a sample of women with ED, in an attempt to confirm previous findings of an inverse relation; (b) to test the hypothesis that women with ED inhibit the expression of emotions perceived as threatening, by examining the relation between emotional expression and perceptions of threat from emotion, while partialling out the effects of depression; and (c) to determine whether, amongst women with ED, perceptions of threat from anger are uniquely associated with emotional inhibition, when the effects of depression and body dissatisfaction are controlled for. Results demonstrated that (a) emotional expression was negatively related with the three Eating Disorders Inventory-3 subscales (drive for thinness, bulimia and body dissatisfaction); (b) perceived threat from emotion, particularly anger, was negatively correlated with emotional expression, when depression was partialled out in the analysis; and (c) perceived threat from anger significantly and uniquely predicted emotional inhibition, over and above the effects of body dissatisfaction and depression, in a sample of women with ED symptomatology. It is suggested that anger may be perceived as particularly threatening amongst women with ED, and play a significant role in the emotional expression difficulties that this population experiences. The implications of the current findings are discussed in
Nao Shikanai; Kozaburo Hachimura
Effects of facial expressions on recognizing emotions expressed in dance movements were investigated. Dancers expressed three emotions: joy, sadness, and anger through dance movements. We used digital video cameras and a 3D motion capturing system to record and capture the movements. We then created full-video displays with an expressive face, full-video displays with an unexpressive face, stick figure displays (no face), or point-light displays (no face) from these data using 3D animation so...
Full Text Available The most important issues of psychology is psychosomatic disease. This study aimed to compare spiritual intelligence and emotional expression in patients with irritable bowel syndrome, coronary heart disease and asthma. This research was a post-event descriptive study. The statistical population included patients with coronary heart disease, irritable bowel syndrome and asthma that attended Imam Khomeini hospital, Tehran. The participants consisted of 150 participants (86 women, 64 men with irritable bowel syndrome (n=50, coronary heart disease (n=50 and asthma (n=50. They answered King and Emmons’ emotional expressiveness questionnaire and King’s spiritual intelligence questionnaire. The results showed a significant relationship between spiritual intelligence and emotional expressiveness subscales and a low level of spiritual intelligence and emotional expressiveness in all three groups of patients. Comparing the three groups showed that spiritual intelligence and emotional expression were low in all of them and coronary heart disease was the lowest in three group of patient.
Ricciardi, Lucia; Visco-Comandini, Federica; Erro, Roberto; Morgante, Francesca; Bologna, Matteo; Fasano, Alfonso; Ricciardi, Diego; Edwards, Mark J; Kilner, James
Parkinson's disease (PD) patients have impairment of facial expressivity (hypomimia) and difficulties in interpreting the emotional facial expressions produced by others, especially for aversive emotions. We aimed to evaluate the ability to produce facial emotional expressions and to recognize facial emotional expressions produced by others in a group of PD patients and a group of healthy participants in order to explore the relationship between these two abilities and any differences between the two groups of participants. Twenty non-demented, non-depressed PD patients and twenty healthy participants (HC) matched for demographic characteristics were studied. The ability of recognizing emotional facial expressions was assessed with the Ekman 60-faces test (Emotion recognition task). Participants were video-recorded while posing facial expressions of 6 primary emotions (happiness, sadness, surprise, disgust, fear and anger). The most expressive pictures for each emotion were derived from the videos. Ten healthy raters were asked to look at the pictures displayed on a computer-screen in pseudo-random fashion and to identify the emotional label in a six-forced-choice response format (Emotion expressivity task). Reaction time (RT) and accuracy of responses were recorded. At the end of each trial the participant was asked to rate his/her confidence in his/her perceived accuracy of response. For emotion recognition, PD reported lower score than HC for Ekman total score (pemotions sub-scores happiness, fear, anger, sadness (pemotion expressivity task, PD and HC significantly differed in the total score (p = 0.05) and in the sub-scores for happiness, sadness, anger (all pemotions. There was a significant positive correlation between the emotion facial recognition and expressivity in both groups; the correlation was even stronger when ranking emotions from the best recognized to the worst (R = 0.75, p = 0.004). PD patients showed difficulties in recognizing emotional facial
Koval, Peter; Butler, Emily A; Hollenstein, Tom; Lanteigne, Dianna; Kuppens, Peter
The tendency for emotions to be predictable over time, labelled emotional inertia, has been linked to low well-being and is thought to reflect impaired emotion regulation. However, almost no studies have examined how emotion regulation relates to emotional inertia. We examined the effects of cognitive reappraisal and expressive suppression on the inertia of behavioural, subjective and physiological measures of emotion. In Study 1 (N = 111), trait suppression was associated with higher inertia of negative behaviours. We replicated this finding experimentally in Study 2 (N = 186). Furthermore, in Study 2, instructed suppressors and reappraisers both showed higher inertia of positive behaviours, and reappraisers displayed higher inertia of heart rate. Neither suppression nor reappraisal were associated with the inertia of subjective feelings in either study. Thus, the effects of suppression and reappraisal on the temporal dynamics of emotions depend on the valence and emotional response component in question.
Full Text Available Psychopathic individuals show selfish, manipulative, and antisocial behavior in addition to emotional detachment and reduced empathy. Their empathic deficits are thought to be associated with a reduced responsiveness to emotional stimuli. Immediate facial muscle responses to the emotional expressions of others reflect the expressive part of emotional responsiveness and are positively related to trait empathy. Empirical evidence for reduced facial muscle responses in adult psychopathic individuals to the emotional expressions of others is rare. In the present study, 261 male criminal offenders and non-offenders categorized dynamically presented facial emotion expressions (angry, happy, sad, and neutral during facial electromyography recording of their corrugator muscle activity. We replicated a measurement model of facial muscle activity, which controls for general facial responsiveness to face stimuli, and modeled three correlated emotion-specific factors (i.e., anger, happiness, and sadness representing emotion specific activity. In a multi-group confirmatory factor analysis, we compared the means of the anger, happiness, and sadness latent factors between three groups: 1 non-offenders, 2 low, and 3 high psychopathic offenders. There were no significant mean differences between groups. Our results challenge current theories that focus on deficits in emotional responsiveness as leading to the development of psychopathy and encourage further theoretical development on deviant emotional processes in psychopathic individuals.
Oh, Seong Il; Oh, Ki Wook; Kim, Hee Jin; Park, Jin Seok; Kim, Seung Hyun
The increasing recognition that deficits in social emotions occur in amyotrophic lateral sclerosis (ALS) is helping to explain the spectrum of neuropsychological dysfunctions, thus supporting the view of ALS as a multisystem disorder involving neuropsychological deficits as well as motor deficits. The aim of this study was to characterize the emotion perception abilities of Korean patients with ALS based on the recognition of facial expressions. Twenty-four patients with ALS and 24 age- and sex-matched healthy controls completed neuropsychological tests and facial emotion recognition tasks [ChaeLee Korean Facial Expressions of Emotions (ChaeLee-E)]. The ChaeLee-E test includes facial expressions for seven emotions: happiness, sadness, anger, disgust, fear, surprise, and neutral. The ability to perceive facial emotions was significantly worse among ALS patients performed than among healthy controls [65.2±18.0% vs. 77.1±6.6% (mean±SD), p=0.009]. Eight of the 24 patients (33%) scored below the 5th percentile score of controls for recognizing facial emotions. Emotion perception deficits occur in Korean ALS patients, particularly regarding facial expressions of emotion. These findings expand the spectrum of cognitive and behavioral dysfunction associated with ALS into emotion processing dysfunction.
Graham, Steven M; Huang, Julie Y; Clark, Margaret S; Helgeson, Vicki S
Four studies support the hypothesis that expressing negative emotion is associated with positive relationship outcomes, including elicitation of support, building of new close relationships, and heightening of intimacy in the closest of those relationships. In Study 1, participants read vignettes in which another person was experiencing a negative emotion. Participants reported they would provide more help when the person chose to express the negative emotion. In Study 2, participants watched a confederate preparing for a speech. Participants provided more help to her when she expressed nervousness. In Study 3, self-reports of willingness to express negative emotions predicted having more friends, controlling for demographic variables and extraversion. In Study 4, self-reports of willingness to express negative emotion measured prior to arrival at college predicted formation of more relationships, greater intimacy in the closest of those relationships, and greater received support from roommates across participants' first semester of college.
Full Text Available Background & objectives: Students bullying, especially in the adolescence period, is a prevalent problem in the schools, that emotional dysregulation is posed as a one cause of it. Considering this issue, the aim of this study was to investigate the relationships between emotion regulation and emotion expression styles with bullying behaviors in adolescent students. Methods: The method of this study was correlation. Whole male students of secondary and high schools in Ardabil at 90-91 educational year comprised statistical population of this research. Two hundred thirty students, were selected by multistage cluster sampling method, responded to the questionnaires of bullying/victimization, emotion regulation and emotion expression. Gathered data were analyzed by Pearson correlation and multiple regression tests. Results: The results showed that victimization by bullying has positive relationship with cognitive reappraisal (r= 0.15, p<0.02, emotion suppression (r= 0.47, p<0.001, and positive expression (r= 0.25, p<0.02, but has negative relationship with impulse severity (r= -0.35, p<0.001, and negative emotion expression (r= -0.43, p<0.001. Furthermore bullying has a positive relationship with cognitive reappraisal (r= 0.14, p<0.03, impulse severity (r= 0.31, p<0.003, and negative expression (r= 0.29, p<0.001, but has negative relationship with emotion suppression (r= 0.28, p<0.001, and positive expression (r= 0.24, p<0.001. In sum emotion regulation and emotion expression styles explained 36 percent of the variance of the victimization by bullying and 19 percent of the variance of the bullying. Conclusion: This research demonstrated that emotion dysregulation at the adolescent period plays important role in bullying and victimization, thus the training of emotion regulation abilities is suggested as the one of interventions methods for this behavioral problems.
Thonse, Umesh; Behere, Rishikesh V; Praharaj, Samir Kumar; Sharma, Podila Sathya Venkata Narasimha
Facial emotion recognition deficits have been consistently demonstrated in patients with severe mental disorders. Expressed emotion is found to be an important predictor of relapse. However, the relationship between facial emotion recognition abilities and expressed emotions and its influence on socio-occupational functioning in schizophrenia versus bipolar disorder has not been studied. In this study we examined 91 patients with schizophrenia and 71 with bipolar disorder for psychopathology, socio occupational functioning and emotion recognition abilities. Primary caregivers of 62 patients with schizophrenia and 49 with bipolar disorder were assessed on Family Attitude Questionnaire to assess their expressed emotions. Patients of schizophrenia and bipolar disorder performed similarly on the emotion recognition task. Patients with schizophrenia group experienced higher critical comments and had a poorer socio-occupational functioning as compared to patients with bipolar disorder. Poorer socio-occupational functioning in patients with schizophrenia was significantly associated with greater dissatisfaction in their caregivers. In patients with bipolar disorder, poorer emotion recognition scores significantly correlated with poorer adaptive living skills and greater hostility and dissatisfaction in their caregivers. The findings of our study suggest that emotion recognition abilities in patients with bipolar disorder are associated with negative expressed emotions leading to problems in adaptive living skills. Copyright © 2018 Elsevier B.V. All rights reserved.
Craig, Belinda M; Lipp, Ottmar V
Facial attributes such as race, sex, and age can interact with emotional expressions; however, only a couple of studies have investigated the nature of the interaction between facial age cues and emotional expressions and these have produced inconsistent results. Additionally, these studies have not addressed the mechanism/s driving the influence of facial age cues on emotional expression or vice versa. In the current study, participants categorised young and older adult faces expressing happiness and anger (Experiment 1) or sadness (Experiment 2) by their age and their emotional expression. Age cues moderated categorisation of happiness vs. anger and sadness in the absence of an influence of emotional expression on age categorisation times. This asymmetrical interaction suggests that facial age cues are obligatorily processed prior to emotional expressions. Finding a categorisation advantage for happiness expressed on young faces relative to both anger and sadness which are negative in valence but different in their congruence with old age stereotypes or structural overlap with age cues suggests that the observed influence of facial age cues on emotion perception is due to the congruence between relatively positive evaluations of young faces and happy expressions.
Full Text Available Effects of facial expressions on recognizing emotions expressed in dance movements were investigated. Dancers expressed three emotions: joy, sadness, and anger through dance movements. We used digital video cameras and a 3D motion capturing system to record and capture the movements. We then created full-video displays with an expressive face, full-video displays with an unexpressive face, stick figure displays (no face, or point-light displays (no face from these data using 3D animation software. To make point-light displays, 13 markers were attached to the body of each dancer. We examined how accurately observers were able to identify the expression that the dancers intended to create through their dance movements. Dance experienced and inexperienced observers participated in the experiment. They watched the movements and rated the compatibility of each emotion with each movement on a 5-point Likert scale. The results indicated that both experienced and inexperienced observers could identify all the emotions that dancers intended to express. Identification scores for dance movements with an expressive face were higher than for other expressions. This finding indicates that facial expressions affect the identification of emotions in dance movements, whereas only bodily expressions provide sufficient information to recognize emotions.
Visch, Valentijn T; Goudbeek, Martijn B; Mortillaro, Marcello
In two studies, the robustness of anger recognition of bodily expressions is tested. In the first study, video recordings of an actor expressing four distinct emotions (anger, despair, fear, and joy) were structurally manipulated as to image impairment and body segmentation. The results show that anger recognition is more robust than other emotions to image impairment and to body segmentation. Moreover, the study showed that arms expressing anger were more robustly recognised than arms expressing other emotions. Study 2 added face blurring as a variable to the bodily expressions and showed that it decreased accurate emotion recognition-but more for recognition of joy and despair than for anger and fear. In sum, the paper indicates the robustness of anger recognition in multileveled deteriorated bodily expressions.
Dael, Nele; Goudbeek, Martijn; Scherer, K R
Recent judgment studies have shown that people are able to fairly correctly attribute emotional states to others' bodily expressions. It is, however, not clear which movement qualities are salient, and how this applies to emotional gesture during speech-based interaction. In this study we investigated how the expression of emotions that vary on three major emotion dimensions-that is, arousal, valence, and potency-affects the perception of dynamic arm gestures. Ten professional actors enacted 12 emotions in a scenario-based social interaction setting. Participants (N = 43) rated all emotional expressions with muted sound and blurred faces on six spatiotemporal characteristics of gestural arm movement that were found to be related to emotion in previous research (amount of movement, movement speed, force, fluency, size, and height/vertical position). Arousal and potency were found to be strong determinants of the perception of gestural dynamics, whereas the differences between positive or negative emotions were less pronounced. These results confirm the importance of arm movement in communicating major emotion dimensions and show that gesture forms an integrated part of multimodal nonverbal emotion communication.
Kostić Aleksandra P.
Full Text Available The results of a study on the accuracy of intensity ratings of emotion from facial expressions are reported. The so far research into the field has shown that spontaneous facial expressions of basic emotions are a reliable source of information about the category of emotion. The question is raised of whether this can be true for the intensity of emotion as well and whether the accuracy of intensity ratings is dependent on the observer’s sex and vocational orientation. A total of 228 observers of both sexes and of various vocational orientations rated the emotional intensity of presented facial expressions on a scale-range from 0 to 8. The results have supported the hypothesis that spontaneous facial expressions of basic emotions do provide sufficient information about emotional intensity. The hypothesis on the interdependence between the accuracy of intensity ratings of emotion and the observer’s sex and vocational orientation has not been confirmed. However, the accuracy of intensity rating has been proved to vary with the category of the emotion presented.
Alberto eDi Domenico
Full Text Available In this study, we investigated whether age-related differences in emotion regulation priorities influence online dynamic emotional facial discrimination. A group of 40 younger and a group of 40 older adults were invited to recognize a positive or negative expression as soon as the expression slowly emerged and subsequently rate it in terms of intensity. Our findings show that older adults recognized happy expressions faster than angry ones, while the direction of emotional expression does not seem to affect younger adults’ performance. Furthermore, older adults rated both negative and positive emotional faces as more intense compared to younger controls. This study detects age-related differences with a dynamic online paradigm and suggests that different regulation strategies may shape emotional face recognition.
Bowling, Daniel Liu; Sundararajan, Janani; Han, Shui'er; Purves, Dale
In Western music, the major mode is typically used to convey excited, happy, bright or martial emotions, whereas the minor mode typically conveys subdued, sad or dark emotions. Recent studies indicate that the differences between these modes parallel differences between the prosodic and spectral characteristics of voiced speech sounds uttered in corresponding emotional states. Here we ask whether tonality and emotion are similarly linked in an Eastern musical tradition. The results show that the tonal relationships used to express positive/excited and negative/subdued emotions in classical South Indian music are much the same as those used in Western music. Moreover, tonal variations in the prosody of English and Tamil speech uttered in different emotional states are parallel to the tonal trends in music. These results are consistent with the hypothesis that the association between musical tonality and emotion is based on universal vocal characteristics of different affective states.
Daniel Liu Bowling
Full Text Available In Western music, the major mode is typically used to convey excited, happy, bright or martial emotions, whereas the minor mode typically conveys subdued, sad or dark emotions. Recent studies indicate that the differences between these modes parallel differences between the prosodic and spectral characteristics of voiced speech sounds uttered in corresponding emotional states. Here we ask whether tonality and emotion are similarly linked in an Eastern musical tradition. The results show that the tonal relationships used to express positive/excited and negative/subdued emotions in classical South Indian music are much the same as those used in Western music. Moreover, tonal variations in the prosody of English and Tamil speech uttered in different emotional states are parallel to the tonal trends in music. These results are consistent with the hypothesis that the association between musical tonality and emotion is based on universal vocal characteristics of different affective states.
Volkova, Ekaterina P; Mohler, Betty J; Dodds, Trevor J; Tesch, Joachim; Bülthoff, Heinrich H
Humans can recognize emotions expressed through body motion with high accuracy even when the stimuli are impoverished. However, most of the research on body motion has relied on exaggerated displays of emotions. In this paper we present two experiments where we investigated whether emotional body expressions could be recognized when they were recorded during natural narration. Our actors were free to use their entire body, face, and voice to express emotions, but our resulting visual stimuli used only the upper body motion trajectories in the form of animated stick figures. Observers were asked to perform an emotion recognition task on short motion sequences using a large and balanced set of emotions (amusement, joy, pride, relief, surprise, anger, disgust, fear, sadness, shame, and neutral). Even with only upper body motion available, our results show recognition accuracy significantly above chance level and high consistency rates among observers. In our first experiment, that used more classic emotion induction setup, all emotions were well recognized. In the second study that employed narrations, four basic emotion categories (joy, anger, fear, and sadness), three non-basic emotion categories (amusement, pride, and shame) and the "neutral" category were recognized above chance. Interestingly, especially in the second experiment, observers showed a bias toward anger when recognizing the motion sequences for emotions. We discovered that similarities between motion sequences across the emotions along such properties as mean motion speed, number of peaks in the motion trajectory and mean motion span can explain a large percent of the variation in observers' responses. Overall, our results show that upper body motion is informative for emotion recognition in narrative scenarios.
Ekaterina P. Volkova
Full Text Available Humans can recognise emotions expressed through body motion with high accuracy even when the stimuli are impoverished. However, most of the research on body motion has relied on exaggerated displays of emotions. In this paper we present two experiments where we investigated whether emotional body expressions could be recognised when they were recorded during natural narration. Our actors were free to use their entire body, face and voice to express emotions, but our resulting visual stimuli used only the upper body motion trajectories in the form of animated stick figures. Observers were asked to perform an emotion recognition task on short motion sequences using a large and balanced set of emotions (amusement, joy, pride, relief, surprise, anger, disgust, fear, sadness, shame and neutral. Even with only upper body motion available, our results show recognition accuracy significantly above chance level and high consistency rates among observers. In our first experiment, that used more classic emotion induction setup, all emotions were well recognised. In the second study that employed narrations, four basic emotion categories (joy, anger, fear and sadness, three non-basic emotion categories (amusement, pride and shame and the neutral category were recognised above chance. Interestingly, especially in the second experiment, observers showed a bias towards anger when recognising the motion sequences for emotions. We discovered that similarities between motion sequences across the emotions along such properties as mean motion speed, number of peaks in the motion trajectory and mean motion span can explain a large percent of the variation in observers' responses. Overall, our results show that upper body motion is informative for emotion recognition in narrative scenarios.
Alvino, Christopher; Kohler, Christian; Barrett, Frederick; Gur, Raquel E; Gur, Ruben C; Verma, Ragini
Deficits in the ability to express emotions characterize several neuropsychiatric disorders and are a hallmark of schizophrenia, and there is need for a method of quantifying expression, which is currently done by clinical ratings. This paper presents the development and validation of a computational framework for quantifying emotional expression differences between patients with schizophrenia and healthy controls. Each face is modeled as a combination of elastic regions, and expression changes are modeled as a deformation between a neutral face and an expressive face. Functions of these deformations, known as the regional volumetric difference (RVD) functions, form distinctive quantitative profiles of expressions. Employing pattern classification techniques, we have designed expression classifiers for the four universal emotions of happiness, sadness, anger and fear by training on RVD functions of expression changes. The classifiers were cross-validated and then applied to facial expression images of patients with schizophrenia and healthy controls. The classification score for each image reflects the extent to which the expressed emotion matches the intended emotion. Group-wise statistical analysis revealed this score to be significantly different between healthy controls and patients, especially in the case of anger. This score correlated with clinical severity of flat affect. These results encourage the use of such deformation based expression quantification measures for research in clinical applications that require the automated measurement of facial affect.
Full Text Available Abstract Background Schools have been identified as an important place in which to support adolescent emotional health, although evidence as to which interventions are effective remains limited. Relatively little is known about student and staff views regarding current school-based emotional health provision and what they would like to see in the future, and this is what this study explored. Methods A random sample of 296 English secondary schools were surveyed to quantify current level of emotional health provision. Qualitative student focus groups (27 groups, 154 students aged 12-14 and staff interviews (12 interviews, 15 individuals were conducted in eight schools, purposively sampled from the survey respondents to ensure a range of emotional health activity, free school meal eligibility and location. Data were analysed thematically, following a constant comparison approach. Results Emergent themes were grouped into three areas in which participants felt schools did or could intervene: emotional health in the curriculum, support for those in distress, and the physical and psychosocial environment. Little time was spent teaching about emotional health in the curriculum, and most staff and students wanted more. Opportunities to explore emotions in other curriculum subjects were valued. All schools provided some support for students experiencing emotional distress, but the type and quality varied a great deal. Students wanted an increase in school-based help sources that were confidential, available to all and sympathetic, and were concerned that accessing support should not lead to stigma. Finally, staff and students emphasised the need to consider the whole school environment in order to address sources of distress such as bullying and teacher-student relationships, but also to increase activities that enhanced emotional health. Conclusion Staff and students identified several ways in which schools can improve their support of adolescent
This article addresses the debate between emotion-expression and motive-communication approaches to facial movements, focusing on Ekman's (1972) and Fridlund's (1994) contrasting models and their historical antecedents. Available evidence suggests that the presence of others either reduces or increases facial responses, depending on the quality and strength of the emotional manipulation and on the nature of the relationship between interactants. Although both display rules and social motives provide viable explanations of audience "inhibition" effects, some audience facilitation effects are less easily accommodated within an emotion-expression perspective. In particular, emotion is not a sufficient condition for a corresponding "expression," even discounting explicit regulation, and, apparently, "spontaneous" facial movements may be facilitated by the presence of others. Further, there is no direct evidence that any particular facial movement provides an unambiguous expression of a specific emotion. However, information communicated by facial movements is not necessarily extrinsic to emotion. Facial movements not only transmit emotion-relevant information but also contribute to ongoing processes of emotional action in accordance with pragmatic theories.
Buttelmann, David; Call, Josep; Tomasello, Michael
Although apes understand others' goals and perceptions, little is known about their understanding of others' emotional expressions. We conducted three studies following the general paradigm of Repacholi and colleagues (1997, 1998). In Study 1, a human reacted emotionally to the hidden contents of two boxes, after which the ape was allowed to…
Castillo, Yuleinys; Fischer, Jerome M.
The skills and tasks in the human services field are highly connected to emotional intelligence abilities. The purpose of the study was to investigate the effect of an expressive writing program involving human service students in an undergraduate rehabilitation services course. The program was developed to enhance their emotional intelligence.…
Soleymani, Mohammad; Asghari-Esfeden, Sadjad; Pantic, Maja; Fu, Yun
Emotions play an important role in how we select and consume multimedia. Recent advances on affect detection are focused on detecting emotions continuously. In this paper, for the first time, we continuously detect valence from electroencephalogram (EEG) signals and facial expressions in response to
Full Text Available Although the expression of emotions in humans is considered to be largely universal, cultural effects contribute to both emotion expression and recognition. To disentangle the interplay between these factors, play-acted and authentic (non-instructed vocal expressions of emotions were used, on the assumption that cultural effects may contribute differentially to the recognition of staged and spontaneous emotions. Speech tokens depicting four emotions (anger, sadness, joy, fear were obtained from German radio archives and reenacted by professional actors, and presented to 120 participants from Germany, Romania, and Indonesia. Participants in all three countries were poor at distinguishing between play-acted and spontaneous emotional utterances (58.73% correct on average with only marginal cultural differences. Nevertheless, authenticity influenced emotion recognition: across cultures, anger was recognized more accurately when play-acted (z = 15.06, p < .001 and sadness when authentic (z = 6.63, p < .001, replicating previous findings from German populations. German subjects revealed a slight advantage in recognizing emotions, indicating a moderate in-group advantage. There was no difference between Romanian and Indonesian subjects in the overall emotion recognition. Differential cultural effects became particularly apparent in terms of differential biases in emotion attribution. While all participants labeled play-acted expressions as anger more frequently than expected, German participants exhibited a further bias towards choosing anger for spontaneous stimuli. In contrast to the German sample, Romanian and Indonesian participants were biased towards choosing sadness. These results support the view that emotion recognition rests on a complex interaction of human universals and cultural specificities. Whether and in which way the observed biases are linked to cultural differences in self-construal remains an issue for further investigation.
Full Text Available This paper proposes two multimodal fusion methods between brain and peripheral signals for emotion recognition. The input signals are electroencephalogram and facial expression. The stimuli are based on a subset of movie clips that correspond to four specific areas of valance-arousal emotional space (happiness, neutral, sadness, and fear. For facial expression detection, four basic emotion states (happiness, neutral, sadness, and fear are detected by a neural network classifier. For EEG detection, four basic emotion states and three emotion intensity levels (strong, ordinary, and weak are detected by two support vector machines (SVM classifiers, respectively. Emotion recognition is based on two decision-level fusion methods of both EEG and facial expression detections by using a sum rule or a production rule. Twenty healthy subjects attended two experiments. The results show that the accuracies of two multimodal fusion detections are 81.25% and 82.75%, respectively, which are both higher than that of facial expression (74.38% or EEG detection (66.88%. The combination of facial expressions and EEG information for emotion recognition compensates for their defects as single information sources.
Martin, Caren McHenry
Sudden outbursts of laughing, crying, or other emotional expression without an apparent triggering stimulus have been recorded in the literature for decades. Confusing nomenclature and a paucity of clinical research, however, have had clinicians wondering whether this syndrome may be both under-recognized and undertreated. Treatment options, including antidepressants, levodopa, and dextromethorphan/quinidine, may show promise in ameliorating symptoms for patients plagued with disorders of emotional expression.
Merola, I; Prato-Previde, E; Lazzaroni, M; Marshall-Pescini, S
Dogs have been shown to discriminate between human facial expressions, and they seem to use human emotional communication to regulate their behaviour towards an external object/situation. However, it is still not clear (1) whether they just respond to the emotional message received with a corresponding increase/decrease in their level of activation or whether they perceive that the emotional message refers to a specific object, (2) which emotional message they use to modify their behaviour (i.e. whether they are following the positive message or avoiding the negative one) and (3) whether their familiarity with the informant has an effect on the dogs' behaviour. To address these issues, five groups of dogs were tested in two experiments. The first group observed the owner delivering two different emotional messages (happiness and fear) towards two identical objects hidden behind barriers, and the second group observed the owner delivering the same emotional messages but with no-objects present in the room. The third and the fourth groups observed the owner delivering a happy versus a neutral, and a negative versus a neutral emotional message towards the hidden objects. Finally, the fifth group observed a stranger acting like the owner of the first group. When the owner was acting as the informant, dogs seemed to be capable of distinguishing between a fearful and happy emotional expression and preferentially chose to investigate a box eliciting an expression of happiness rather than of fear or neutrality. Dogs, however, seemed to have greater difficulty in distinguishing between the fearful and neutral emotional messages delivered by the owner and between the happy and fearful expressions delivered by the stranger. Results suggest that dogs have learned to associate their owners' positive emotional messages to positive outcomes, and hence use their communicative messages to guide their actions. However, negative emotional messages and those delivered by strangers are
This paper presents an investigation of mirroring facial expressions and the emotions which they convey in dyadic naturally occurring first encounters. Mirroring facial expressions are a common phenomenon in face-to-face interactions, and they are due to the mirror neuron system which has been...... and overlapping facial expressions are very frequent. In this study, we want to determine whether the overlapping facial expressions are mirrored or are otherwise correlated in the encounters, and to what extent mirroring facial expressions convey the same emotion. The results of our study show that the majority...... of smiles and laughs, and one fifth of the occurrences of raised eyebrows are mirrored in the data. Moreover some facial traits in co-occurring expressions co-occur more often than it would be expected by chance. Finally, amusement, and to a lesser extent friendliness, are often emotions shared by both...
de Melo, Celso M; Carnevale, Peter J; Read, Stephen J; Gratch, Jonathan
How do people make inferences about other people's minds from their emotion displays? The ability to infer others' beliefs, desires, and intentions from their facial expressions should be especially important in interdependent decision making when people make decisions from beliefs about the others' intention to cooperate. Five experiments tested the general proposition that people follow principles of appraisal when making inferences from emotion displays, in context. Experiment 1 revealed that the same emotion display produced opposite effects depending on context: When the other was competitive, a smile on the other's face evoked a more negative response than when the other was cooperative. Experiment 2 revealed that the essential information from emotion displays was derived from appraisals (e.g., Is the current state of affairs conducive to my goals? Who is to blame for it?); facial displays of emotion had the same impact on people's decision making as textual expressions of the corresponding appraisals. Experiments 3, 4, and 5 used multiple mediation analyses and a causal-chain design: Results supported the proposition that beliefs about others' appraisals mediate the effects of emotion displays on expectations about others' intentions. We suggest a model based on appraisal theories of emotion that posits an inferential mechanism whereby people retrieve, from emotion expressions, information about others' appraisals, which then lead to inferences about others' mental states. This work has implications for the design of algorithms that drive agent behavior in human-agent strategic interaction, an emerging domain at the interface of computer science and social psychology.
Beaudry, Olivia; Roy-Charland, Annie; Perron, Melanie; Cormier, Isabelle; Tapp, Roxane
The present study aimed to clarify the role played by the eye/brow and mouth areas in the recognition of the six basic emotions. In Experiment 1, accuracy was examined while participants viewed partial and full facial expressions; in Experiment 2, participants viewed full facial expressions while their eye movements were recorded. Recognition rates were consistent with previous research: happiness was highest and fear was lowest. The mouth and eye/brow areas were not equally important for the recognition of all emotions. More precisely, while the mouth was revealed to be important in the recognition of happiness and the eye/brow area of sadness, results are not as consistent for the other emotions. In Experiment 2, consistent with previous studies, the eyes/brows were fixated for longer periods than the mouth for all emotions. Again, variations occurred as a function of the emotions, the mouth having an important role in happiness and the eyes/brows in sadness. The general pattern of results for the other four emotions was inconsistent between the experiments as well as across different measures. The complexity of the results suggests that the recognition process of emotional facial expressions cannot be reduced to a simple feature processing or holistic processing for all emotions.
Tang, Yulong; Harris, Paul L; Zou, Hong; Xu, Qunxia
Research on the development of selective trust has shown that young children do not indiscriminately trust all potential informants. They are likely to seek and endorse information from individuals who have proven competent or benign in the past. However, research on trust among adults raises the possibility that children might also be influenced by the emotions expressed by potential informants. In particular, they might trust individuals expressing more positive emotion. Indeed, young children's trust in particular informants based on their past behaviour might be undermined by their currently expressed emotions. To examine this possibility, we tested the selective trust of fifty 4- and 5-year-olds in two steps. We first confirmed that children are likely to invest more trust in individuals expressing more positive emotion. We then showed that even if children have already formed an impression of two potential informants based on their behavioural record, their choices about whose claims to trust are markedly influenced by the degree of positive emotion currently expressed by the two informants. By implication, the facial emotions expressed by potential informants can undermine young children's selective trust based on the behavioural record of those informants.
Turchet, Luca; Bresin, Roberto
in the music performance domain, as well as the absence of an influence of musical expertise lend support to the “motor origin hypothesis of emotional expression in music” according to which a motor origin for the expression of emotions is common in all those domains of human activity that result......This paper describes two experiments conducted to investigate the role of sonically simulated ground materials in modulating both production and recognition of walks performed with emotional intentions. The results of the first experiment showed that the involved auditory feedbacks affected...... the pattern of emotional walking in different ways, although such an influence manifested itself in more than one direction. The results of the second experiment showed the absence of an influence of the sound conditions on the recognition of the emotions from acoustic information alone. Similar results were...
Kayyal, Mary H; Russell, James A
The claim that certain emotions are universally recognized from facial expressions is based primarily on the study of expressions that were posed. The current study was of spontaneous facial expressions shown by aborigines in Papua New Guinea (Ekman, 1980); 17 faces claimed to convey one (or, in the case of blends, two) basic emotions and five faces claimed to show other universal feelings. For each face, participants rated the degree to which each of the 12 predicted emotions or feelings was conveyed. The modal choice for English-speaking Americans (n = 60), English-speaking Palestinians (n = 60), and Arabic-speaking Palestinians (n = 44) was the predicted label for only 4, 5, and 4, respectively, of the 17 faces for basic emotions, and for only 2, 2, and 2, respectively, of the 5 faces for other feelings. Observers endorsed the predicted emotion or feeling moderately often (65%, 55%, and 44%), but also denied it moderately often (35%, 45%, and 56%). They also endorsed more than one (or, for blends, two) label(s) in each face-on average, 2.3, 2.3, and 1.5 of basic emotions and 2.6, 2.2, and 1.5 of other feelings. There were both similarities and differences across culture and language, but the emotional meaning of a facial expression is not well captured by the predicted label(s) or, indeed, by any single label.
Brandão, Tânia; Tavares, Rita; Schulz, Marc S; Matos, Paula Mena
The important role of emotion regulation and expression in adaptation to breast cancer is now widely recognized. Studies have shown that optimal emotion regulation strategies, including less constrained emotional expression, are associated with better adaptation. Our objective was to systematically review measures used to assess the way women with breast cancer regulate their emotions. This systematic review was conducted in accordance with PRISMA guidelines. Nine different databases were searched. Data were independently extracted and assessed by two researchers. English-language articles that used at least one instrument to measure strategies to regulate emotions in women with breast cancer were included. Of 679 abstracts identified 59 studies were deemed eligible for inclusion. Studies were coded regarding their objectives, methods, and results. We identified 16 instruments used to measure strategies of emotion regulation and expression. The most frequently employed instrument was the Courtauld Emotional Control Scale. Few psychometric proprieties other than internal consistency were reported for most instruments. Many studies did not include important information regarding descriptive characteristics and psychometric properties of the instruments used. The instruments used tap different aspects of emotion regulation. Specific instruments should be explored further with regard to content, validity, and reliability in the context of breast cancer. Copyright © 2015 Elsevier Ltd. All rights reserved.
Kostov, Vlaho; Yanagisawa, Hideyoshi; Johansson, Martin; Fukuda, Shuichi
A simple method for extracting emotion from a human face, as a form of non-verbal communication, was developed to cope with and optimize mobile communication in a globalized and diversified society. A cartoon face based model was developed and used to evaluate emotional content of real faces. After a pilot survey, basic rules were defined and student subjects were asked to express emotion using the cartoon face. Their face samples were then analyzed using principal component analysis and the Mahalanobis distance method. Feature parameters considered as having relations with emotions were extracted and new cartoon faces (based on these parameters) were generated. The subjects evaluated emotion of these cartoon faces again and we confirmed these parameters were suitable. To confirm how these parameters could be applied to real faces, we asked subjects to express the same emotions which were then captured electronically. Simple image processing techniques were also developed to extract these features from real faces and we then compared them with the cartoon face parameters. It is demonstrated via the cartoon face that we are able to express the emotions from very small amounts of information. As a result, real and cartoon faces correspond to each other. It is also shown that emotion could be extracted from still and dynamic real face images using these cartoon-based features.
Munsch, Simone; Dremmel, Daniela; Kurz, Susanne; De Albuquerque, Jiske; Meyer, Andrea H; Hilbert, Anja
We investigated whether parental expressed emotion (criticism and emotional overinvolvement) is related to children's emotional eating and whether this relationship is mediated by children's negative urgency. One hundred children, aged 8 to 13 years, either healthy or have binge-eating disorder and/or attention-deficit/hyperactivity disorder, completed the questionnaires, along with their parents. Parental criticism and, to a lesser extent, parental emotional overinvolvement were both positively related to children's emotional eating, and this relationship was mediated by children's negative urgency. Further exploratory analyses revealed that the mediating role of children's negative urgency in the relationship between parental criticism and children's emotional eating was pronounced in the clinical group of children with binge-eating disorder and attention-deficit/hyperactivity disorder but almost absent in the healthy control group. Copyright © 2016 John Wiley & Sons, Ltd and Eating Disorders Association. © 2016 John Wiley & Sons, Ltd and Eating Disorders Association.
Watling, Dawn; Bourne, Victoria J
Understanding of emotions has been shown to develop between the ages of 4 and 10 years; however, individual differences exist in this development. While previous research has typically examined these differences in terms of developmental and/or social factors, little research has considered the possible impact of neuropsychological development on the behavioural understanding of emotions. Emotion processing tends to be lateralised to the right hemisphere of the brain in adults, yet this pattern is not as evident in children until around the age of 10 years. In this study 136 children between 5 and 10 years were given both behavioural and neuropsychological tests of emotion processing. The behavioural task examined expression regulation knowledge (ERK) for prosocial and self-presentational hypothetical interactions. The chimeric faces test was given as a measure of lateralisation for processing positive facial emotion. An interaction between age and lateralisation for emotion processing was predictive of children's ERK for only the self-presentational interactions. The relationships between children's ERK and lateralisation for emotion processing changes across the three age groups, emerging as a positive relationship in the 10-year-olds. The 10-years-olds who were more lateralised to the right hemisphere for emotion processing tended to show greater understanding of the need for regulating negative emotions during interactions that would have a self-presentational motivation. This finding suggests an association between the behavioural and neuropsychological development of emotion processing.
Madsen, Jens; Jensen, Bjørn Sand; Larsen, Jan
We introduce a two-alternative forced-choice (2AFC) experimental paradigm to quantify expressed emotions in music using the arousal and valence (AV) dimensions. A wide range of well-known audio features are investigated for predicting the expressed emotions in music using learning curves...... and essential baselines. We furthermore investigate the scalability issues of using 2AFC in quantifying emotions expressed in music on large-scale music databases. The possibility of dividing the annotation task between multiple individuals, while pooling individuals’ comparisons is investigated by looking...... at the subjective differences of ranking emotion in the AV space. We find this to be problematic due to the large variation in subjects’ rankings of excerpts. Finally, solving scalability issues by reducing the number of pairwise comparisons is analyzed. We compare two active learning schemes to selecting...
Juslin, Patrik N
Numerous studies have investigated whether music can reliably convey emotions to listeners, and-if so-what musical parameters might carry this information. Far less attention has been devoted to the actual contents of the communicative process. The goal of this article is thus to consider what types of emotional content are possible to convey in music. I will argue that the content is mainly constrained by the type of coding involved, and that distinct types of content are related to different types of coding. Based on these premises, I suggest a conceptualization in terms of "multiple layers" of musical expression of emotions. The "core" layer is constituted by iconically-coded basic emotions. I attempt to clarify the meaning of this concept, dispel the myths that surround it, and provide examples of how it can be heuristic in explaining findings in this domain. However, I also propose that this "core" layer may be extended, qualified, and even modified by additional layers of expression that involve intrinsic and associative coding. These layers enable listeners to perceive more complex emotions-though the expressions are less cross-culturally invariant and more dependent on the social context and/or the individual listener. This multiple-layer conceptualization of expression in music can help to explain both similarities and differences between vocal and musical expression of emotions.
Full Text Available Do our background beliefs, desires, and mental images influence our perceptual experience of the emotions of others? In this paper, we will address the possibility of cognitive penetration of perceptual experience in the domain of social cognition. In particular, we focus on emotion recognition based on the visual experience of facial expressions. After introducing the current debate on cognitive penetration, we review examples of perceptual adaptation for facial expressions of emotion. This evidence supports the idea that facial expressions are perceptually processed as wholes. That is, the perceptual system integrates lower-level facial features, such as eyebrow orientation, mouth angle etc., into facial compounds. We then present additional experimental evidence showing that in some cases, emotion recognition on the basis of facial expression is sensitive to and modified by the background knowledge of the subject. We argue that such sensitivity is best explained as a difference in the visual experience of the facial expression, not just as a modification of the judgment based on this experience. The difference in experience is characterized as the result of the interference of background knowledge with the perceptual integration process for faces. Thus, according to the best explanation, we have to accept cognitive penetration in some cases of emotion recognition. Finally, we highlight a recent model of social vision in order to propose a mechanism for cognitive penetration used in the face-based recognition of emotion.
Dael, Nele; Perseguers, Marie-Noëlle; Marchand, Cynthia; Antonietti, Jean-Philippe; Mohr, Christine
People associate affective meaning with colour, and this may influence decisions about colours. Hue is traditionally considered the most salient descriptor of colour and colour-affect associations, although colour brightness and saturation seem to have particularly strong affective connotations. To test whether colour choices can be driven by emotion, we investigated whether and how colour hue, brightness, and saturation are systematically associated with bodily expressions of positive (joy) and negative (fear) emotions. Twenty-five non-colour-blind participants viewed videos of these expressions and selected for each video the most appropriate colour using colour sliders providing values for hue, brightness, and saturation. The overall colour choices were congruent with the expressed emotion--that is, participants selected brighter and more saturated colours for joy expressions than for fear expressions. Also, colours along the red-yellow spectrum were deemed more appropriate for joy expressions and cyan-bluish hues for fear expressions. The current study adds further support to the role of emotion in colour choices by (a) showing that emotional information is spontaneously used in an unconstrained choice setting, (b) extending to ecologically valid stimuli occurring in everyday encounters (dressed bodies), and (c) suggesting that all colour parameters are likely to be important when processing affective nonverbal person information, though not independently from each other.
Von Culin, Katherine R; Hirsch, Jennifer L; Clark, Margaret S
Two studies document that people are more willing to express emotions that reveal vulnerabilities to partners when they perceive those partners to be more communally responsive to them. In Study 1, participants rated the communal strength they thought various partners felt toward them and their own willingness to express happiness, sadness and anxiety to each partner. Individuals who generally perceive high communal strength from their partners were also generally most willing to express emotion to partners. Independently, participants were more willing to express emotion to particular partners whom they perceived felt more communal strength toward them. In Study 2, members of romantic couples independently reported their own felt communal strength toward one another, perceptions of their partners' felt communal strength toward them, and willingness to express emotions (happiness, sadness, anxiety, disgust, anger, hurt and guilt) to each other. The communal strength partners reported feeling toward the participants predicted the participants' willingness to express emotion to those partners. This link was mediated by participants' perceptions of the partner's communal strength toward them which, itself, was a joint function of accurate perceptions of the communal strength partners had reported feeling toward them and projections of their own felt communal strength for their partners onto those partners.
Nyklicek, I.; Vingerhoets, A.J.J.M.; Denollet, J.K.L.
The available evidence on the relationship between expression and non-expression of emotions (E/NE) and health is selectively and critically reviewed. It is concluded that research in this field still lacks conceptual lucidity with regard to the many existing E/NE concepts. Despite the fact that few
In this paper we discuss the aspects of designing facial expressions for Virtual Humans with a specific culture. First we explore the notion of cultures and its relevance for applications with a Virtual Human. Then we give a general scheme of designing emotional facial expressions, and identify the
I. S. Ivanova
Full Text Available The paper emphasizes the need for studying the subjective effectiveness criteria of interpersonal communication and importance of effective communication for personality development in adolescence. The problemof undeveloped representation of positive emotions in communication process is discussed. Both the identification and verbalization of emotions are regarded by the author as the basic communication skills. The experimental data regarding the longitude and age levels are described, the gender differences in identification and verbalization of emotions considered. The outcomes of experimental study demonstrate that the accuracy of facial emotional expressions of teenage boys and girls changes at different rates. The prospects of defining the age norms for identification and verbalization of emotions are analyzed.
Tobin, Kenneth; King, Donna; Henderson, Senka; Bellocchi, Alberto; Ritchie, Stephen M.
We investigated the expression of emotions while teaching in relation to a teacher's physiological changes. We used polyvagal theory (PVT) to frame the study of teaching in a teacher education program. Donna, a teacher-researcher, experienced high levels of stress and anxiety prior to beginning to teach and throughout the lesson we used her expressed emotions as a focus for this research. We adopted event-oriented inquiry in a study of heart rate, oxygenation of the blood, and expressed emotions. Five events were identified for multilevel analysis in which we used narrative, prosodic analysis, and hermeneutic-phenomenological methods to learn more about the expression of emotions when Donna had: high heart rate (before and while teaching); low blood oxygenation (before and while teaching); and high blood oxygenation (while teaching). What we learned was consistent with the body's monitoring system recognizing social harm and switching to the control of the unmyelinated vagus nerve, thereby shutting down organs and muscles associated with social communication—leading to irregularities in prosody and expression of emotion. In events involving high heart rate and low blood oxygenation the physiological environment was associated with less effective and sometimes confusing patterns in prosody, including intonation, pace of speaking, and pausing. In a low blood oxygenation environment there was evidence of rapid speech and shallow, irregular breathing. In contrast, during an event in which 100 % blood oxygenation occurred, prosody was perceived to be conducive to engagement and teacher expressed positive emotions, such as satisfaction, while teaching. Becoming aware of the purposes of the research and the results we obtained provided the teacher with tools to enact changes to her teaching practice, especially prosody of the voice. We regard it as a high priority to create tools to allow teachers and students, if and as necessary, to ameliorate excess emotions, and
Juslin, Patrik N.
Numerous studies have investigated whether music can reliably convey emotions to listeners, and—if so—what musical parameters might carry this information. Far less attention has been devoted to the actual contents of the communicative process. The goal of this article is thus to consider what types of emotional content are possible to convey in music. I will argue that the content is mainly constrained by the type of coding involved, and that distinct types of content are related to different types of coding. Based on these premises, I suggest a conceptualization in terms of “multiple layers” of musical expression of emotions. The “core” layer is constituted by iconically-coded basic emotions. I attempt to clarify the meaning of this concept, dispel the myths that surround it, and provide examples of how it can be heuristic in explaining findings in this domain. However, I also propose that this “core” layer may be extended, qualified, and even modified by additional layers of expression that involve intrinsic and associative coding. These layers enable listeners to perceive more complex emotions—though the expressions are less cross-culturally invariant and more dependent on the social context and/or the individual listener. This multiple-layer conceptualization of expression in music can help to explain both similarities and differences between vocal and musical expression of emotions. PMID:24046758
Marin Dapelo, Marcela; Hart, Sharon; Hale, Christiane; Morris, Robin; Lynch, Thomas R.; Tchanturia, Kate
A large body of research has associated Eating Disorders with difficulties in socio-emotional functioning and it has been argued that they may serve to maintain the illness. This study aimed to explore facial expressions of positive emotions in individuals with Anorexia Nervosa (AN) and Bulimia Nervosa (BN) compared to healthy controls (HC), through an examination of the Duchenne smile (DS), which has been associated with feelings of enjoyment, amusement and happiness ( Ekman et al., 1990 ). ...
The present study investigates the link between gender, emotional intelligence (EI), personality traits and self-reported emotional expression in the second language (L2). Data analysis suggests that gender might not influence self-perceived emotional expression in the L2, as the results of the t-test show that both males and females declare…
Olderbak, Sally; Hildebrandt, Andrea; Pinkpank, Thomas; Sommer, Werner; Wilhelm, Oliver
Coding of facial emotion expressions is increasingly performed by automated emotion expression scoring software; however, there is limited discussion on how best to score the resulting codes. We present a discussion of facial emotion expression theories and a review of contemporary emotion expression coding methodology. We highlight methodological challenges pertinent to scoring software-coded facial emotion expression codes and present important psychometric research questions centered on co...
The author's purpose was to examine children's recognition of emotional facial expressions, by comparing two types of stimulus: photographs and drawings. The author aimed to investigate whether drawings could be considered as a more evocative material than photographs, as a function of age and emotion. Five- and 7-year-old children were presented with photographs and drawings displaying facial expressions of 4 basic emotions (i.e., happiness, sadness, anger, and fear) and were asked to perform a matching task by pointing to the face corresponding to the target emotion labeled by the experimenter. The photographs we used were selected from the Radboud Faces Database and the drawings were designed on the basis of both the facial components involved in the expression of these emotions and the graphic cues children tend to use when asked to depict these emotions in their own drawings. Our results show that drawings are better recognized than photographs, for sadness, anger, and fear (with no difference for happiness, due to a ceiling effect). And that the difference between the 2 types of stimuli tends to be more important for 5-year-olds compared to 7-year-olds. These results are discussed in view of their implications, both for future research and for practical application.
Perry, Anat; Aviezer, Hillel; Goldstein, Pavel; Palgi, Sharon; Klein, Ehud; Shamay-Tsoory, Simone G
The neuropeptide oxytocin (OT) has been repeatedly reported to play an essential role in the regulation of social cognition in humans in general, and specifically in enhancing the recognition of emotions from facial expressions. The later was assessed in different paradigms that rely primarily on isolated and decontextualized emotional faces. However, recent evidence has indicated that the perception of basic facial expressions is not context invariant and can be categorically altered by context, especially body context, at early perceptual levels. Body context has a strong effect on our perception of emotional expressions, especially when the actual target face and the contextually expected face are perceptually similar. To examine whether and how OT affects emotion recognition, we investigated the role of OT in categorizing facial expressions in incongruent body contexts. Our results show that in the combined process of deciphering emotions from facial expressions and from context, OT gives an advantage to the face. This advantage is most evident when the target face and the contextually expected face are perceptually similar. Copyright © 2013 Elsevier Ltd. All rights reserved.
van der Mey, A.; Smit, F; Droog, K.J.; Visser, A.
Following a soccer game is an example where clear emotions are displayed. This example is worked out for a humanoid robot which can express emotions with body language. The emotions expressed by the robot are not just stimuli-response, but are based on an affective state which shows dynamic behavior
Yuasa, Masahide; Saito, Keiichi; Mukawa, Naoki
In this paper, we describe the brain activity associated with kanji characters expressing emotion, which are places at the end of a sentence. Japanese people use a special kanji character in brackets at the end of sentences in text messages such as those sent through e-mail and messenger tools. Such kanji characters plays a role to expresses the sender's emotion (such as fun, laughter, sadness, tears), like emoticons. It is a very simple and effective way to convey the senders' emotions and his/her thoughts to the receiver. In this research, we investigate the effects of emotional kanji characters by using an fMRI study. The experimental results show that both the right and left inferior frontal gyrus, which have been implicated on verbal and nonverbal information, were activated. We found that we detect a sentence with an emotional kanji character as the verbal and nonverval information, and a sentence with emotional kanji characters enrich communication between the sender and the reciever.
Alexander, Stewart C; Garner, David Kirkland; Somoroff, Matthew; Gramling, David J; Norton, Sally A; Gramling, Robert
Being able to identify expressions of emotion is crucial to effective clinical communication research. However, traditional linguistic coding systems often cannot represent emotions that are expressed nonlexically or phonologically (i.e., not through words themselves but through vocal pitch, speed/rhythm/tempo, and volume). Using audio recording of a palliative care consultation in the natural hospital setting, two experienced music scholars employed Western musical notation, as well as the graphic realization of a digital audio program (Piano roll visualization), to visually represent the sonic features of conversation where a patient has an emotional "choke" moment. Western musical notation showed the ways that changes in pitch and rate correspond to the patient's emotion: rising sharply in intensity before slowly fading away. Piano roll visualization is a helpful supplement. Using musical notation to illustrate palliative care conversations in the hospital setting can render visible for analysis several aspects of emotional expression that researchers otherwise experience as intuitive or subjective. Various forms and formats of musical notation techniques and sonic visualization technologies should be considered as fruitful and complementary alternatives to traditional coding tools in clinical communications research. Musical notation offers opportunity for both researchers and learners to "see" how communication evolves in clinical encounters, particularly where the lexical and phonological features of interpersonal communication are concordant and discordant with one another. Copyright © 2015. Published by Elsevier Ireland Ltd.
Paulus, Andrea; Wentura, Dirk
Studies examining approach and avoidance reactions to emotional expressions have yielded conflicting results. For example, expressions of anger have been reported to elicit approach reactions in some studies but avoidance reactions in others. Nonetheless, the results were often explained by the same general underlying process, namely the influence that the social message signaled by the expression has on motivational responses. It is therefore unclear which reaction is triggered by which emotional expression, and which underlying process is responsible for these reactions. In order to address this issue, we examined the role of a potential moderator on approach and avoidance reactions to emotional expressions, namely the contrast emotion used in the task. We believe that different approach and avoidance reactions occur depending on the congruency or incongruency of the evaluation of the 2 emotions presented in the task. The results from a series of experiments supported these assumptions: Negative emotional expressions (anger, fear, sadness) elicited avoidance reactions if contrasted with expressions of happiness. However, if contrasted with a different negative emotional expression, anger and sadness triggered approach reactions and fear activated avoidance reactions. Importantly, these results also emerged if the emotional expression was not task-relevant. We propose that approach and avoidance reactions to emotional expressions are triggered by their evaluation if the 2 emotions presented in a task differ in evaluative connotation. If they have the same evaluative connotation, however, reactions are determined by their social message. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Haltom, Kate E.; Mulvenna, Catherine M.; Lieberman, Matthew D.; Stanton, Annette L.
The current study assessed main effects and moderators (including emotional expressiveness, emotional processing and ambivalence over emotional expression) of the effects of expressive writing in a sample of healthy adults. Young adult participants (N = 116) were randomly assigned to write for 20 minutes on four occasions about deepest thoughts and feelings regarding their most stressful/traumatic event in the past five years (expressive writing) or about a control topic (control). Dependent variables were indicators of anxiety, depression, and physical symptoms. No significant effects of writing condition were evident on anxiety, depressive symptoms, or physical symptoms. Emotional expressiveness emerged as a significant moderator of anxiety outcomes, however. Within the expressive writing group, participants high in expressiveness evidenced a significant reduction in anxiety at three-month follow-up, and participants low in expressiveness showed a significant increase in anxiety. Expressiveness did not predict change in anxiety in the control group. These findings on anxiety are consistent with the matching hypothesis, which suggests that matching a person’s naturally elected coping approach with an assigned intervention is beneficial. These findings also suggest that expressive writing about a stressful event may be contraindicated for individuals who do not typically express emotions. PMID:23742666
Ji, Luyan; Pourtois, Gilles
We examined the processing capacity and the role of emotion variance in ensemble representation for multiple facial expressions shown concurrently. A standard set size manipulation was used, whereby the sets consisted of 4, 8, or 16 morphed faces each uniquely varying along a happy-angry continuum (Experiment 1) or a neutral-happy/angry continuum (Experiments 2 & 3). Across the three experiments, we reduced the amount of emotion variance in the sets to explore the boundaries of this process. Participants judged the perceived average emotion from each set on a continuous scale. We computed and compared objective and subjective difference scores, using the morph units and post-experiment ratings, respectively. Results of the subjective scores were more consistent than the objective ones across the first two experiments where the variance was relatively large, and revealed each time that increasing set size led to a poorer averaging ability, suggesting capacity limitations in establishing ensemble representations for multiple facial expressions. However, when the emotion variance in the sets was reduced in Experiment 3, both subjective and objective scores remained unaffected by set size, suggesting that the emotion averaging process was unlimited in these conditions. Collectively, these results suggest that extracting mean emotion from a set composed of multiple faces depends on both structural (attentional) and stimulus-related effects. Copyright © 2018 Elsevier Ltd. All rights reserved.
Laura Jean Wells
Full Text Available The identification of emotional expressions is vital for social interaction, and can be affected by various factors, including the expressed emotion, the intensity of the expression, the sex of the face, and the gender of the observer. This study investigates how these factors affect the speed and accuracy of expression recognition, as well as dwell time on the two most significant areas of the face: the eyes and the mouth. Participants were asked to identify expressions from female and male faces displaying six expressions (anger, disgust, fear, happiness, sadness, and surprise, each with three levels of intensity (low, moderate, and normal. Overall, responses were fastest and most accurate for happy expressions, but slowest and least accurate for fearful expressions. More intense expressions were also classified most accurately. Reaction time showed a different pattern, with slowest response times recorded for expressions of moderate intensity. Overall, responses were slowest, but also most accurate, for female faces. Relative to male observers, women showed greater accuracy and speed when recognizing female expressions. Dwell time analyses revealed that attention to the eyes was about three times greater than on the mouth, with fearful eyes in particular attracting longer dwell times. The mouth region was attended to the most for fearful, angry, and disgusted expressions and least for surprise. These results extend upon previous findings to show important effects of expression, emotion intensity, and sex on expression recognition and gaze behaviour, and may have implications for understanding the ways in which emotion recognition abilities break down.
Affect-adaptive systems have the potential to assist users that experience systematically negative moods. This thesis aims at building a platform for predicting automatically a person’s mood from his/her visual expressions. The key word is mood, namely a relatively long-term, stable and diffused
McKinnon, Jacqueline M; Greenberg, Leslie S
This study examined whether interactions characterized by high expression of emotional vulnerability in one partner followed by a highly supportive response style by the other partner predicted greater improvement on domains of forgiveness, unfinished business, trust, and relationship satisfaction in a sample of 32 couples presenting for Emotion Focused Couples Therapy with unresolved emotional injuries. For each outcome measure, two separate hierarchical regression models were tested (injured partner vulnerability and offending partner supportiveness; offending partner vulnerability and injured partner supportiveness). Both models significantly predicted improvement on the majority of outcome measures. Practice suggestions for working with emotionally injured couples are provided in light of the findings. © 2017 American Association for Marriage and Family Therapy.
Sachs, Matthew E; Habibi, Assal; Damasio, Antonio; Kaplan, Jonas T
Effective social functioning relies in part on the ability to identify emotions from auditory stimuli and respond appropriately. Previous studies have uncovered brain regions engaged by the affective information conveyed by sound. But some of the acoustical properties of sounds that express certain emotions vary remarkably with the instrument used to produce them, for example the human voice or a violin. Do these brain regions respond in the same way to different emotions regardless of the sound source? To address this question, we had participants (N = 38, 20 females) listen to brief audio excerpts produced by the violin, clarinet, and human voice, each conveying one of three target emotions-happiness, sadness, and fear-while brain activity was measured with fMRI. We used multivoxel pattern analysis to test whether emotion-specific neural responses to the voice could predict emotion-specific neural responses to musical instruments and vice-versa. A whole-brain searchlight analysis revealed that patterns of activity within the primary and secondary auditory cortex, posterior insula, and parietal operculum were predictive of the affective content of sound both within and across instruments. Furthermore, classification accuracy within the anterior insula was correlated with behavioral measures of empathy. The findings suggest that these brain regions carry emotion-specific patterns that generalize across sounds with different acoustical properties. Also, individuals with greater empathic ability have more distinct neural patterns related to perceiving emotions. These results extend previous knowledge regarding how the human brain extracts emotional meaning from auditory stimuli and enables us to understand and connect with others effectively. Copyright © 2018 Elsevier Inc. All rights reserved.
Meaux, Emilie; Vuilleumier, Patrik
The ability to decode facial emotions is of primary importance for human social interactions; yet, it is still debated how we analyze faces to determine their expression. Here we compared the processing of emotional face expressions through holistic integration and/or local analysis of visual features, and determined which brain systems mediate these distinct processes. Behavioral, physiological, and brain responses to happy and angry faces were assessed by presenting congruent global configurations of expressions (e.g., happy top+happy bottom), incongruent composite configurations (e.g., angry top+happy bottom), and isolated features (e.g. happy top only). Top and bottom parts were always from the same individual. Twenty-six healthy volunteers were scanned using fMRI while they classified the expression in either the top or the bottom face part but ignored information in the other non-target part. Results indicate that the recognition of happy and anger expressions is neither strictly holistic nor analytic Both routes were involved, but with a different role for analytic and holistic information depending on the emotion type, and different weights of local features between happy and anger expressions. Dissociable neural pathways were engaged depending on emotional face configurations. In particular, regions within the face processing network differed in their sensitivity to holistic expression information, which predominantly activated fusiform, inferior occipital areas and amygdala when internal features were congruent (i.e. template matching), whereas more local analysis of independent features preferentially engaged STS and prefrontal areas (IFG/OFC) in the context of full face configurations, but early visual areas and pulvinar when seen in isolated parts. Collectively, these findings suggest that facial emotion recognition recruits separate, but interactive dorsal and ventral routes within the face processing networks, whose engagement may be shaped by
Manolete S. Moscoso
Full Text Available The purpose of this article is to focus on unique issues that are encountered in the crosscultural adaptation of measures of emotions. We take into consideration the cross-cultural equivalence of the concept of emotion, and how cultural differences influence the meaning of words that are utilized to describe these concepts. The critical need to take the state-trait distinction into account in adapting measures of emotional states and personality traits is then discussed. The effects of language and culture in adapting measures of the experience, expression, and control of anger in Latin-America are also reviewed. The construction of the Latin American Multicultural State-Trait Anger Expression Inventory is described.
Cook, Richard; Brewer, Rebecca; Shah, Punit; Bird, Geoffrey
Despite considerable research into whether face perception is impaired in autistic individuals, clear answers have proved elusive. In the present study, we sought to determine whether co-occurring alexithymia (characterized by difficulties interpreting emotional states) may be responsible for face-perception deficits previously attributed to autism. Two experiments were conducted using psychophysical procedures to determine the relative contributions of alexithymia and autism to identity and expression recognition. Experiment 1 showed that alexithymia correlates strongly with the precision of expression attributions, whereas autism severity was unrelated to expression-recognition ability. Experiment 2 confirmed that alexithymia is not associated with impaired ability to detect expression variation; instead, results suggested that alexithymia is associated with difficulties interpreting intact sensory descriptions. Neither alexithymia nor autism was associated with biased or imprecise identity attributions. These findings accord with the hypothesis that the emotional symptoms of autism are in fact due to co-occurring alexithymia and that existing diagnostic criteria may need to be revised.
Demenescu, Liliana R.; Kortekaas, Rudie; den Boer, Johan A.; Aleman, Andre
Background: Recognition of others' emotions is an important aspect of interpersonal communication. In major depression, a significant emotion recognition impairment has been reported. It remains unclear whether the ability to recognize emotion from facial expressions is also impaired in anxiety
Tobin, Kenneth; King, Donna; Henderson, Senka; Bellocchi, Alberto; Ritchie, Stephen M.
We investigated the expression of emotions while teaching in relation to a teacher's physiological changes. We used polyvagal theory (PVT) to frame the study of teaching in a teacher education program. Donna, a teacher-researcher, experienced high levels of stress and anxiety prior to beginning to teach and throughout the lesson we used her…
Madsen, Jens; Nielsen, Jens Brehm; Jensen, Bjørn Sand
We introduce a two-alternative forced-choice experimental paradigm to quantify expressed emotions in music using the two wellknown arousal and valence (AV) dimensions. In order to produce AV scores from the pairwise comparisons and to visualize the locations of excerpts in the AV space, we...
Hayashi, Hajimu; Shiomi, Yuki
This study examined whether children understand that people selectively conceal or express emotion depending upon the context. We prepared two contexts for a verbal display task for 70 first-graders, 80 third-graders, 64 fifth-graders, and 71 adults. In both contexts, protagonists had negative feelings because of the behavior of the other…
Madsen, Jens; Jensen, Bjørn Sand; Larsen, Jan
We introduce five regression models for the modeling of expressed emotion in music using data obtained in a two alternative forced choice listening experiment. The predictive performance of the proposed models is compared using learning curves, showing that all models converge to produce a similar...
Abstract ance is the outward rhythmic expression of inner emotion that compliments a good rhythm and is performed within time and space. Although some dance scholars and performers, notable among who is Susanne Kircher, have tried to experiment with unaccompanied dance movements; the notion of dance without.
Trauffer, Nicole M.; Widen, Sherri C.; Russell, James A.
Certain facial expressions have been proposed to be signals evolved to communicate a single specific emotion. Evidence to support this view is based primarily on university-educated Western adults. In the current study (N=96), university-educated and non-university-educated Americans were asked to label purported facial expressions of happiness, sadness, anger, fear, surprise, and disgust. Participants with no university education were significantly less likely to label the "fear face" as sca...
Владимир Иванович Озюменко
Full Text Available Peculiarities of expressing emotions in different languages and cultures are nowadays of great interest among researchers. However, traditionally, more attention is paid to lexis, phraseology, prosody, and nonverbal signs of communication and grammatical means are ignored. Nevertherless, they also possess significant emotive potential. The goal of this article is to systematize some facts showing that grammar can implicitly convey various emotions of the speaker and impact the hearer, as well as to draw attention of language teachers and translators to this phenomenon. I will mainly focus on modal verbs and expressions, nontraditional use of tenses, nontraditional use of some adverbs, inverse word order and some other cases summing up the earlier described facts (Ozyumenko 2005-2006, 2007 and supplementing them with new observations. The article emphasizes that discovering emotive implicature, i.e. implicit emotive intention of the speaker, and revealing its meaning is sine qua non for successful intercultural communication and equivalent translation. The data for the research were taken from English language text books, grammar books, dictionaries and fiction. The study implements contextual, definitive, pragmatic, discourse and contrastive analyses.
Benitez-Quiroz, C Fabian; Wilbur, Ronnie B; Martinez, Aleix M
Facial expressions of emotion are thought to have evolved from the development of facial muscles used in sensory regulation and later adapted to express moral judgment. Negative moral judgment includes the expressions of anger, disgust and contempt. Here, we study the hypothesis that these facial expressions of negative moral judgment have further evolved into a facial expression of negation regularly used as a grammatical marker in human language. Specifically, we show that people from different cultures expressing negation use the same facial muscles as those employed to express negative moral judgment. We then show that this nonverbal signal is used as a co-articulator in speech and that, in American Sign Language, it has been grammaticalized as a non-manual marker. Furthermore, this facial expression of negation exhibits the theta oscillation (3-8 Hz) universally seen in syllable and mouthing production in speech and signing. These results provide evidence for the hypothesis that some components of human language have evolved from facial expressions of emotion, and suggest an evolutionary route for the emergence of grammatical markers. Copyright © 2016 Elsevier B.V. All rights reserved.
Is teaching emotional labor? Are teachers selling their own emotions in exchange for money? To examine these questions, this paper examines teachers' emotion expression in teaching. Most previous studies have assessed teachers' emotional experience using interviews, and have reported that because teachers manage their own emotions, teaching is…
Müller, Corsin A; Schmitt, Kira; Barber, Anjuli L A; Huber, Ludwig
The question of whether animals have emotions and respond to the emotional expressions of others has become a focus of research in the last decade [1-9]. However, to date, no study has convincingly shown that animals discriminate between emotional expressions of heterospecifics, excluding the possibility that they respond to simple cues. Here, we show that dogs use the emotion of a heterospecific as a discriminative cue. After learning to discriminate between happy and angry human faces in 15 picture pairs, whereby for one group only the upper halves of the faces were shown and for the other group only the lower halves of the faces were shown, dogs were tested with four types of probe trials: (1) the same half of the faces as in the training but of novel faces, (2) the other half of the faces used in training, (3) the other half of novel faces, and (4) the left half of the faces used in training. We found that dogs for which the happy faces were rewarded learned the discrimination more quickly than dogs for which the angry faces were rewarded. This would be predicted if the dogs recognized an angry face as an aversive stimulus. Furthermore, the dogs performed significantly above chance level in all four probe conditions and thus transferred the training contingency to novel stimuli that shared with the training set only the emotional expression as a distinguishing feature. We conclude that the dogs used their memories of real emotional human faces to accomplish the discrimination task. Copyright © 2015 Elsevier Ltd. All rights reserved.
Lawton, Julia; Kirkham, Jackie; White, David; Rankin, David; Cooper, Cindy; Heller, Simon
The perspectives and experiences of trial staff are increasingly being investigated as these can be used to improve recruitment, adherence to trial protocols and support given to future staff. We interviewed staff working on a type 1 diabetes trial in order to aid interpretation of trial findings, inform recommendations for the rollout of the treatments investigated and provide recommendations for the conduct of future trials. However, our interviews uncovered aspects of trial work erstwhile unrecognised or underreported in the trials literature, and it is these which form the focus of this paper. In-depth interviews were conducted with (n = 18) staff, recruited from seven centres, who were involved in recruitment and trial delivery. Data were analysed thematically. Alongside logistical and practical issues which made trial work challenging, staff often talked spontaneously and at length about how trial work had affected them emotionally. Staff not only described the emotional stresses arising from having to meet recruitment targets and from balancing research roles with clinical responsibilities, they also discussed having to emotionally manage patients and their colleagues. The emotional aspects of trial work particularly came to the fore when staff notified patients about their treatment allocation. On such occasions, staff described having to employ emotional strategies to pre-empt and manage potential patient disappointment and anger. Staff also described having to manage their own emotions when patients withdrew from the trial or were not randomised to the treatment arm which, in their clinical judgment, would have been in their best interests. To help address the emotional challenges they encountered, staff highlighted a need for more practical, emotional and specialist psychological support. More attention should be paid to the emotional aspects of trial work to help ensure trial staff are adequately supported. Such support could comprise: increased
Park, Mona; Hennig-Fast, Kristina; Bao, Yan; Carl, Petra; Pöppel, Ernst; Welker, Lorenz; Reiser, Maximilian; Meindl, Thomas; Gutyrchik, Evgeny
Music communicates and evokes emotions. The number of studies on the neural correlates of musical emotion processing is increasing but few have investigated the factors that modulate these neural activations. Previous research has shown that personality traits account for individual variability of neural responses. In this study, we used functional magnetic resonance imaging (fMRI) to investigate how the dimensions Extraversion and Neuroticism are related to differences in brain reactivity to musical stimuli expressing the emotions happiness, sadness and fear. 12 participants (7 female, M=20.33 years) completed the NEO-Five Factor Inventory (NEO-FFI) and were scanned while performing a passive listening task. Neurofunctional analyses revealed significant positive correlations between Neuroticism scores and activations in bilateral basal ganglia, insula and orbitofrontal cortex in response to music expressing happiness. Extraversion scores were marginally negatively correlated with activations in the right amygdala in response to music expressing fear. Our findings show that subjects' personality may have a predictive power in the neural correlates of musical emotion processing and should be considered in the context of experimental group homogeneity. Copyright © 2013 Elsevier B.V. All rights reserved.
Vigil's socio-relational framework of sex differences in emotional expressiveness emphasizes general sex differences in emotional responding, but largely ignores the social context in which emotions are expressed. There is much empirical evidence showing that sex differences in emotion displays are
ÇETİNKAYA DUMAN, Zekiye; KUŞCU, M. Kemal; ÖZGÜN, Serkan
Introduction The aim of the study was to compare the Camberwell Family Interview (CFI) and the Expressed Emotion Scale (EES) in determining the level of expressed emotion in caregivers of patients with schizophrenia. Method The study sample included caregivers of 22 schizophrenic patients followed in two psychiatric clinics. The level of expressed emotion in the caregivers was assessed by the CFI and the EES. CFI was applied to caregivers of the inpatients and the procedure was audio recorded. These records were later used for the ratings. EES was completed by the caregivers. Total EES scores were used to determine the level of expressed emotion in the caregivers. Results Forty point nine percent and 50% of the caregivers had high level of expressed emotion based on the analysis of the data obtained from the CFI and EES. Fifty-nine percent and 50% of the caregivers had low level of expressed emotion based on the data obtained from the CFI and EES. The proportion of the caregivers with high level of expressed emotion as measured by the CFI and the EES were not statistically significantly different within the sample (χ2= 0.727). Conclusion The CFI and the EES were similar in determining the level of expressed emotion in caregivers of schizophrenic patients. It can be suggested that the EES, a user friendly tool, may be preferred to determine the level of expressed emotion in caregivers of patients with schizophrenia. However, further studies with larger samples are needed to obtain more reliable results. PMID:28360561
Chen, Stephen H; Zhou, Qing; Main, Alexandra; Lee, Erica H
The present study examined 2 measures of Chinese American immigrant parents' emotional expression in the family context: self-reported emotional expressivity and observed emotional expression during a parent-child interaction task. Path analyses were conducted to examine the concurrent associations between measures of emotional expression and (a) parents' American and Chinese cultural orientations in language proficiency, media use, and social affiliation domains, and (b) parents' and teachers' ratings of children's emotion-related regulation. Results suggested that cultural orientations were primarily associated with parents' self-reported expressivity (rather than observed emotional expression), such that higher American orientations were generally associated with higher expressivity. Although parents' self-reported expressivity was only related to their own reports of children's regulation, parents' observed emotional expression was related to both parents' and teachers' reports of children's regulation. These results suggest that self-reported expressivity and observed emotional expression reflect different constructs and have differential relations to parents' cultural orientations and children's regulation. (c) 2015 APA, all rights reserved).
Tell, Dina; Davidson, Denise
In this research, the emotion recognition abilities of children with autism spectrum disorder and typically developing children were compared. When facial expressions and situational cues of emotion were congruent, accuracy in recognizing emotions was good for both children with autism spectrum disorder and typically developing children. When…
Evers, Kris; Steyaert, Jean; Noens, Ilse; Wagemans, Johan
Emotion labelling was evaluated in two matched samples of 6-14-year old children with and without an autism spectrum disorder (ASD; N = 45 and N = 50, resp.), using six dynamic facial expressions. The Emotion Recognition Task proved to be valuable demonstrating subtle emotion recognition difficulties in ASD, as we showed a general poorer emotion…
Pauletto, Sandra; Balentine, Bruce; Pidcock, Chris; Jones, Kevin; Bottaci, Leonardo; Aretoulaki, Maria; Wells, Jez; Mundy, Darren P; Balentine, James
Emotion in audio-voice signals, as synthesized by text-to-speech (TTS) technologies, was investigated to formulate a theory of expression for user interface design. Emotional parameters were specified with markup tags, and the resulting audio was further modulated with post-processing techniques. Software was then developed to link a selected TTS synthesizer with an automatic speech recognition (ASR) engine, producing a chatbot that could speak and listen. Using these two artificial voice subsystems, investigators explored both artistic and psychological implications of artificial speech emotion. Goals of the investigation were interdisciplinary, with interest in musical composition, augmentative and alternative communication (AAC), commercial voice announcement applications, human-computer interaction (HCI), and artificial intelligence (AI). The work-in-progress points towards an emerging interdisciplinary ontology for artificial voices. As one study output, HCI tools are proposed for future collaboration.
Jackman, Louisa; Fielden, Amy; Pearson, Steven
The help of specialist clinicians is often sought to advise staff in residential and nursing care homes about how to work with people with dementia whose behaviour is challenging. The Newcastle Model ( James, 2011 ) is a framework and a process developed to help care staff understand and improve their care of this group. The model emphasises the use of sharing information with staff to develop effective care plans. In the Shared Formulation Sessions characteristic of the Newcastle Model, clinicians take the role of a group facilitator, helping the staff reach a consensus about what needs to change. These sessions can be difficult to manage as intra and inter-group processes emerge and the group express their anxieties. This paper aims to explore the processes that might be in play Shared Formulation Sessions and to suggest ways in which the facilitator might approach this to manage effective collaborative working.
Vatne, Torun M; Ruland, Cornelia M; Ørnes, Knut; Finset, Arnstein
One function of expressing emotion is to receive support. The aim of this study was to assess how children with heart disease express negative emotions during routine consultations, and examine the interaction between children's expressions and adults' responses. Seventy children, aged 7-13 years, completed measures of anxiety and were videotaped during cardiology visits. Adult-child interactions were analyzed using the Verona Definitions of Emotional Sequences. Children expressed negative emotion, mainly in subtle ways; however, adults rarely recognized and responded to these expressions. The frequency of children's expressions and adults' responses were related to the child's age, level of anxiety, and verbal participation. Children do not openly express negative emotions frequently during routine cardiac consultations; they are more likely to provide subtle cues of negative emotion. When expression of negative emotions does occur, adults may consider using the opportunity to explore the child's emotional experiences.
Van Doorn, E.A.; van Kleef, G.A.; van der Pligt, J.
How do instructors' emotional expressions influence students' learning performance? Scholars and practitioners alike have emphasized the importance of positive, nurturing emotions for successful learning. However, teachers may sometimes lose their temper and express anger at their pupils. Drawing on
Kessels, R.P.C.; Spee, P.S.; Hendriks, A.W.C.J.
Previous studies have shown deficits in the perception of static emotional facial expressions in individuals with autism spectrum disorders (ASD), but results are inconclusive. Possibly, using dynamic facial stimuli expressing emotions at different levels of intensities may produce more robust
Kessels, R.P.C.; Spee, P.; Hendriks, A.W.
Previous studies have shown deficits in the perception of static emotional facial expressions in individuals with autism spectrum disorders (ASD), but results are inconclusive. Possibly, using dynamic facial stimuli expressing emotions at different levels of intensities may produce more robust
Long, Kristin A; Lobato, Debra; Kao, Barbara; Plante, Wendy; Grullón, Edicta; Cheas, Lydia; Houck, Christopher; Seifer, Ronald
Examine general emotion expression and sibling-parent emotion communication among Latino and non-Latino white (NLW) siblings of children with intellectual disabilities (ID) and matched comparisons. 200 siblings (ages 8-15 years) completed the newly developed Sibling-Parent Emotion Communication Scale and existing measures of general emotion expression and psychosocial functioning. Preliminary analyses evaluated scale psychometrics across ethnicity. Structure and internal consistency of the emotion expression and communication measures differed by respondent ethnicity. Latino siblings endorsed more general emotion expression problems and marginally lower sibling-parent emotion communication than NLW siblings. Siblings of children with ID reported marginally more general emotion expression problems than comparisons. Emotion expression problems and lower sibling-parent emotion communication predicted more internalizing and somatic symptoms and poorer personal adjustment, regardless of ID status. Siblings of children with ID endorsed poorer personal adjustment. Cultural differences in emotion expression and communication may increase Latino siblings' risk for emotional adjustment difficulties.
Full Text Available Background: Recognition of emotional facial expressions is one of the psychological factors which involve in obsessive-compulsive disorder (OCD and major depressive disorder (MDD. The aim of present study was to compare the ability of recognizing emotional facial expressions in patients with Obsessive-Compulsive Disorder and major depressive disorder. Materials and Methods: The present study is a cross-sectional and ex-post facto investigation (causal-comparative method. Forty participants (20 patients with OCD, 20 patients with MDD were selected through available sampling method from the clients referred to Tabriz Bozorgmehr clinic. Data were collected through Structured Clinical Interview and Recognition of Emotional Facial States test. The data were analyzed utilizing MANOVA. Results: The obtained results showed that there is no significant difference between groups in the mean score of recognition emotional states of surprise, sadness, happiness and fear; but groups had a significant difference in the mean score of diagnosing disgust and anger states (p<0.05. Conclusion: Patients suffering from both OCD and MDD show equal ability to recognize surprise, sadness, happiness and fear. However, the former are less competent in recognizing disgust and anger than the latter.
Polce-Lynch, Mary; Myers, Barbara J.; Kilmartin, Christopher T.; Forssmann-Falck, Renate; Kliewer, Wendy
Used written narratives to examine gender and age patterns in body image, emotional expression, and self-esteem for 209 students in grades 5, 8, and 12. Results indicate that boys restrict emotional expression in adolescence, whereas girls increase emotional expression in the same period. Girls also are more influenced by body image. (SLD)
van Kleef, G.A.
Emotions have a pervasive impact on organizational behavior. They do not just influence people’s own actions; when expressed, emotions may also exert influence on other organization members who perceive the expressions. Sometimes emotional expressions have ‘symmetrical’ effects, in that positive
Schmidt, Ricarda; Tetzlaff, Anne; Hilbert, Anja
A sizeable body of research has documented Expressed Emotion (EE) to predict clinical outcomes in various psychiatric disorders, including eating disorders. Patients’ perceptions of relative’s EE, however, were found to play an important role in the processing of EE. This study aimed to examine the level of perceived EE in adolescent binge-eating disorder (BED) and its impact on eating disorder psychopathology. Adolescents (12 – 20 years) seeking treatment for BED (n = 40) were compared to...
Tadmor, Tamar; Dolev, Niva; Attias, Dina; Reuven-Lelong, Ayalla; Rofe, Amnon
During the last decade, medical organizations have undergone major changes worldwide and these continue to evolve at a rapid pace. Today the medical profession faces many new challenges that will eventually have an impact on almost every aspect of daily hospital routine. To a large extent, these issues arise from emerging new technologies, the entry of a new generation of trained workers who have different views and characteristics than previous generations, and the introduction of stricter regulations and accreditation procedures in recent years. In addition, the various hospital staff members now have different professional expectations and demands; there is also an important need to reduce costs, accompanied by a shift towards the concept of patients perceiving themselves as clients rather than only as people needing medical assistance. Facing all these challenges, undoubtedly, medical teams will need to acquire a more comprehensive set of professional skills critical for their continued success in the 21st century. These skills will have to include the ability to be more flexible, so as to be able to adapt to changing environments, to remain effective at work under stress, to develop positive personal interactive working relationships, while providing excellent service to patients, and to maintain the ability to guide and lead others in a changing medical environment. People with the above skills reflect the positive attributes of high emotional intelligence. Recent studies show that emotional intelligence plays an important role in the success of the entire medical staff and particularly for those in management roles. Hospitals will have to take into consideration all the necessary characteristics, if they wish to maintain and further consolidate their previous achievements in the 21st century. In particular, they will need to pay attention to the EQ of both new and existing staff, using it as a meaningful parameter for new recruits and for the further
Clément, Fabrice; Bernard, Stéphane; Grandjean, Didier; Sander, David
A great deal of what we know about the world has not been learned via first-hand observation but thanks to others' testimony. A crucial issue is to know which kind of cues people use to evaluate information provided by others. In this context, recent studies in adults and children underline that informants' facial expressions could play an essential role. To test the importance of the other's emotions in vocabulary learning, we used two avatars expressing happiness, anger or neutral emotions when proposing different verbal labels for an unknown object. Experiment 1 revealed that adult participants were significantly more likely than chance to choose the label suggested by the avatar displaying a happy face over the label suggested by the avatar displaying an angry face. Experiment 2 extended these results by showing that both adults and children as young as 3 years old showed this effect. These data suggest that decision making concerning newly acquired information depends on informant's expressions of emotions, a finding that is consistent with the idea that behavioural intents have facial signatures that can be used to detect another's intention to cooperate.
Algoe, Sara B; Fredrickson, Barbara L; Gable, Shelly L
Recent theory posits that the emotion of gratitude uniquely functions to build a high-quality relationship between a grateful person and the target of his or her gratitude, that is, the person who performed a kind action (Algoe et al., 2008). Therefore, gratitude is a prime candidate for testing the dyadic question of whether one person's grateful emotion has consequences for the other half of the relational unit, the person who is the target of that gratitude. The current study tests the critical hypothesis that being the target of gratitude forecasts one's relational growth with the person who expresses gratitude. The study employed a novel behavioral task in which members of romantic relationships expressed gratitude to one another in a laboratory paradigm. As predicted, the target's greater perceptions of the expresser's responsiveness after the interaction significantly predicted improvements in relationship quality over 6 months. These effects were independent from perceptions of responsiveness following two other types of relationally important and emotionally evocative social interactions in the lab, suggesting the unique weight that gratitude carries in cultivating social bonds. PsycINFO Database Record (c) 2013 APA, all rights reserved.
Sebastián Eduardo Piemontesi
Full Text Available Emotional Expressivity, defined as the ability to express emotional states in ob-servable behaviors, is essential for individuals healthy functioning, and was po -sitively associated with wellbeing, self-esteem, life satisfaction and negatively related with diseases such as schizophrenia, depression, personality disorders and post traumatic stress disorder. To answer the need for an instrument which can evaluate this construct in a valid and reliable manner, this study explored the psychometric properties of the Emotional Expressivity Scale adapted into Spa-nish. For this reason, an exploratory factor analysis replicating the one-dimension solution was performed, a coefficient alpha of .94 was obtained, gender differen -ces with higher scores in women, and test-retest coefficients for a 4-week interval with values of .88 in women and .86 in men. Additionally, confirmatory factor analyzes were performed separately for each gender obtaining appropriate va-lues for all fit indices, but not in men. Finally the results, scope and limitations of this paper are discussed.
De Maesschalck, Stéphanie; Deveugele, Myriam; Willems, Sara
This study explores ethnic minority patients' expression of emotional cues and concerns in primary healthcare, and examines relationships with patient, provider and consultation attributes. 191 video-recorded consultations were analyzed using the VR-CoDES. Patients were interviewed before the consultation. Generalized Estimating Equations models (GEE) were used to test for associations. Psychosocial versus bio-medically oriented encounters contained significantly more cues (p≤0.05). Patients with poor versus good language proficiency expressed significantly less cues (p≤0.001). No significant correlations were found with patients' cultural values, patients' or physicians' gender or the presence of an interpreter. Female patients express more concerns (p≤0.05), female physicians have a higher number of concerns expressed by patients (p≤0.02). This study shows that independent of physician and diagnosis, patients' language proficiency has a more important impact on the number of cues expressed by the patient than cultural difference. Medical schools and Continuing Medical Education should focus on training programs for recognizing and handling linguistic barriers between physicians and patients. Patient education programs should encourage patients who experience language barriers to open up to physicians. In situations where language is a barrier, physicians and patients should be encouraged to use interpreters to enhance the expression of emotions. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.
Full Text Available We studied discrimination of briefly presented Upright vs. Inverted emotional facial expressions (FEs, hypothesising that inversion would impair emotion decoding by disrupting holistic FE processing. Stimuli were photographs of seven emotion prototypes, of a male and female poser (Ekman and Friesen, 1976, and eight intermediate morphs in each set. Subjects made speeded Same/Different judgements of emotional content for all Upright (U or Inverted (I pairs of FEs, presented for 500 ms, 100 times each pair. Signal Detection Theory revealed the sensitivity measure d' to be slightly but significantly higher for the Upright FEs. In further analysis using multidimensional scaling (MDS, percentages of Same judgements were taken as an index of pairwise perceptual similarity, separately for U and I presentation mode. The outcome was a 4D ‘emotion expression space’, with FEs represented as points and the dimensions identified as Happy–Sad, Surprise/Fear, Disgust and Anger. The solutions for U and I FEs were compared by means of cophenetic and canonical correlation, Procrustes analysis and weighted-Euclidean analysis of individual difference. Differences in discrimination produced by inverting FE stimuli were found to be small and manifested as minor changes in the MDS structure or weights of the dimensions. Solutions differed substantially more between the two posers, however. Notably, for stimuli containing elements of Happiness (whether U or I, the MDS structure revealed some signs of categorical perception, indicating that mouth curvature – the dominant feature conveying Happiness – is visually salient and receives early processing. The findings suggest that for briefly-presented FEs, Same/Different decisions are dominated by low-level visual analysis of abstract patterns of lightness and edge filters, but also reflect emerging featural analysis. These analyses, insensitive to face orientation, enable initial positive/negative Valence
Aragón, Oriana R; Clark, Margaret S; Dyer, Rebecca L; Bargh, John A
Extremely positive experiences, and positive appraisals thereof, produce intense positive emotions that often generate both positive expressions (e.g., smiles) and expressions normatively reserved for negative emotions (e.g., tears). We developed a definition of these dimorphous expressions and tested the proposal that their function is to regulate emotions. We showed that individuals who express emotions in this dimorphous manner do so as a general response across a variety of emotionally provoking situations, which suggests that these expressions are responses to intense positive emotion rather than unique to one particular situation. We used cute stimuli (an elicitor of positive emotion) to demonstrate both the existence of these dimorphous expressions and to provide preliminary evidence of their function as regulators of emotion. © The Author(s) 2015.
Full Text Available The ability to communicate is one of the core aspects of human life. For this, we use not only verbal but also nonverbal signals of remarkable complexity. Among the latter, facial expressions belong to the most important information channels. Despite the large variety of facial expressions we use in daily life, research on facial expressions has so far mostly focused on the emotional aspect. Consequently, most databases of facial expressions available to the research community also include only emotional expressions, neglecting the largely unexplored aspect of conversational expressions. To fill this gap, we present the MPI facial expression database, which contains a large variety of natural emotional and conversational expressions. The database contains 55 different facial expressions performed by 19 German participants. Expressions were elicited with the help of a method-acting protocol, which guarantees both well-defined and natural facial expressions. The method-acting protocol was based on every-day scenarios, which are used to define the necessary context information for each expression. All facial expressions are available in three repetitions, in two intensities, as well as from three different camera angles. A detailed frame annotation is provided, from which a dynamic and a static version of the database have been created. In addition to describing the database in detail, we also present the results of an experiment with two conditions that serve to validate the context scenarios as well as the naturalness and recognizability of the video sequences. Our results provide clear evidence that conversational expressions can be recognized surprisingly well from visual information alone. The MPI facial expression database will enable researchers from different research fields (including the perceptual and cognitive sciences, but also affective computing, as well as computer vision to investigate the processing of a wider range of natural
Hess, Ursula; Kafetsios, Konstantinos; Mauersberger, Heidi; Blaison, Christophe; Kessler, Carolin-Louisa
Human interactions are replete with emotional exchanges, and hence, the ability to decode others' emotional expressions is of great importance. The present research distinguishes between the emotional signal (the intended emotion) and noise (perception of secondary emotions) in social emotion perception and investigates whether these predict the quality of social interactions. In three studies, participants completed laboratory-based assessments of emotion recognition ability and later reported their perceptions of naturally occurring social interactions. Overall, noise perception in the recognition task was associated with perceiving more negative emotions in others and perceiving interactions more negatively. Conversely, signal perception of facial emotion expressions was associated with higher quality in social interactions. These effects were moderated by relationship closeness in Greece but not in Germany. These findings suggest that emotion recognition as assessed in the laboratory is a valid predictor of social interaction quality. Thus, emotion recognition generalizes from the laboratory to everyday life. © 2016 by the Society for Personality and Social Psychology, Inc.
Lively, Kathryn J.; Powell, Brian
Using the emotions module of the 1996 General Social Survey, we examine strategies that individuals use to express emotion. We focus on anger, one of the emotions most problematic or potentially disruptive to human interaction. Relying on insights from three theoretical approaches to emotion--the cultural perspective, the structural perspective,…
Strand, Paul S.; Downs, Andrew; Barbosa-Leiker, Celestina
The authors explored predictions from basic emotion theory (BET) that facial emotion expression recognition skills are insular with respect to their own development, and yet foundational to the development of emotional perspective-taking skills. Participants included 417 preschool children for whom estimates of these 2 emotion understanding…
Full Text Available Recently, real-time facial expression recognition has attracted more and more research. In this study, an automatic facial expression real-time system was built and tested. Firstly, the system and model were designed and tested on a MATLAB environment followed by a MATLAB Simulink environment that is capable of recognizing continuous facial expressions in real-time with a rate of 1 frame per second and that is implemented on a desktop PC. They have been evaluated in a public dataset, and the experimental results were promising. The dataset and labels used in this study were made from videos, which were recorded twice from five participants while watching a video. Secondly, in order to implement in real-time at a faster frame rate, the facial expression recognition system was built on the field-programmable gate array (FPGA. The camera sensor used in this work was a Digilent VmodCAM — stereo camera module. The model was built on the Atlys™ Spartan-6 FPGA development board. It can continuously perform emotional state recognition in real-time at a frame rate of 30. A graphical user interface was designed to display the participant’s video in real-time and two-dimensional predict labels of the emotion at the same time.
Wingenbach, Tanja S H; Ashwin, Chris; Brosnan, Mark
There has been much research on sex differences in the ability to recognise facial expressions of emotions, with results generally showing a female advantage in reading emotional expressions from the face. However, most of the research to date has used static images and/or 'extreme' examples of facial expressions. Therefore, little is known about how expression intensity and dynamic stimuli might affect the commonly reported female advantage in facial emotion recognition. The current study investigated sex differences in accuracy of response (Hu; unbiased hit rates) and response latencies for emotion recognition using short video stimuli (1sec) of 10 different facial emotion expressions (anger, disgust, fear, sadness, surprise, happiness, contempt, pride, embarrassment, neutral) across three variations in the intensity of the emotional expression (low, intermediate, high) in an adolescent and adult sample (N = 111; 51 male, 60 female) aged between 16 and 45 (M = 22.2, SD = 5.7). Overall, females showed more accurate facial emotion recognition compared to males and were faster in correctly recognising facial emotions. The female advantage in reading expressions from the faces of others was unaffected by expression intensity levels and emotion categories used in the study. The effects were specific to recognition of emotions, as males and females did not differ in the recognition of neutral faces. Together, the results showed a robust sex difference favouring females in facial emotion recognition using video stimuli of a wide range of emotions and expression intensity variations.
Tanja S H Wingenbach
Full Text Available There has been much research on sex differences in the ability to recognise facial expressions of emotions, with results generally showing a female advantage in reading emotional expressions from the face. However, most of the research to date has used static images and/or 'extreme' examples of facial expressions. Therefore, little is known about how expression intensity and dynamic stimuli might affect the commonly reported female advantage in facial emotion recognition. The current study investigated sex differences in accuracy of response (Hu; unbiased hit rates and response latencies for emotion recognition using short video stimuli (1sec of 10 different facial emotion expressions (anger, disgust, fear, sadness, surprise, happiness, contempt, pride, embarrassment, neutral across three variations in the intensity of the emotional expression (low, intermediate, high in an adolescent and adult sample (N = 111; 51 male, 60 female aged between 16 and 45 (M = 22.2, SD = 5.7. Overall, females showed more accurate facial emotion recognition compared to males and were faster in correctly recognising facial emotions. The female advantage in reading expressions from the faces of others was unaffected by expression intensity levels and emotion categories used in the study. The effects were specific to recognition of emotions, as males and females did not differ in the recognition of neutral faces. Together, the results showed a robust sex difference favouring females in facial emotion recognition using video stimuli of a wide range of emotions and expression intensity variations.
There has been much research on sex differences in the ability to recognise facial expressions of emotions, with results generally showing a female advantage in reading emotional expressions from the face. However, most of the research to date has used static images and/or ‘extreme’ examples of facial expressions. Therefore, little is known about how expression intensity and dynamic stimuli might affect the commonly reported female advantage in facial emotion recognition. The current study investigated sex differences in accuracy of response (Hu; unbiased hit rates) and response latencies for emotion recognition using short video stimuli (1sec) of 10 different facial emotion expressions (anger, disgust, fear, sadness, surprise, happiness, contempt, pride, embarrassment, neutral) across three variations in the intensity of the emotional expression (low, intermediate, high) in an adolescent and adult sample (N = 111; 51 male, 60 female) aged between 16 and 45 (M = 22.2, SD = 5.7). Overall, females showed more accurate facial emotion recognition compared to males and were faster in correctly recognising facial emotions. The female advantage in reading expressions from the faces of others was unaffected by expression intensity levels and emotion categories used in the study. The effects were specific to recognition of emotions, as males and females did not differ in the recognition of neutral faces. Together, the results showed a robust sex difference favouring females in facial emotion recognition using video stimuli of a wide range of emotions and expression intensity variations. PMID:29293674
Lieberman, Morton A; Goldstein, Benjamin A
The repression/suppression of negative emotions has long been considered detrimental for breast cancer (BC) patients, leading to poor coping, progression of symptoms, and general lower quality of life. Therapies have focused on encouraging the expression of negative emotions. While group therapies have proven to be successful for BC patients, no study has looked at the role of expressing negative emotions during the therapeutic interaction. We examined written expressed emotions by women participating in a common form of psychosocial support, Internet based bulletin boards (BBs). Fifty-two new members to BC BBs were studied. They completed measures of quality of life and depression. After 6 months the measures were again assessed and messages during that time were collected and analyzed for emotional content. For the 52 women, results showed that greater expression of anger was associated with higher quality of life and lower depression, while the expression of fear and anxiety was associated with lower quality of life and higher depression. The expression of sadness was unrelated to change scores. Our results serve to challenge the commonly held belief that the expression of all negative emotions are beneficial for BC patients. Instead, expressing specific negative emotions are beneficial, while others are not. Copyright 2005 John Wiley & Sons, Ltd.
Montepare, Joann M; Dobish, Heidi
Although theorists acknowledge that beliefs about emotions may play a role in age-related emotion behavior, no research has explored these beliefs. This research examined beliefs about the experience and expression of emotions across the life span, especially across the adult years. Younger and older adults rated the extent to which infants, children, adolescents, young adults, middle-aged adults, and older adults were likely to experience and express a range of emotions. Younger and older adults held similar beliefs about the course of emotions across the life span. Moreover, these beliefs differed across emotion categories. In particular, although older adults were believed to experience and express fewer highly charged, negative emotions, they were expected to be more likely to experience and express positive, low arousal emotions, as well as negative, low arousal emotions. The experience and expression of positive, high arousal emotions were seen as more characteristic of very young age groups as opposed to older age groups. These findings beg questions about if and how beliefs about emotion may affect age-related emotion regulation strategies and other everyday emotion-focused behaviors, as well as social reactions to older adults observed experiencing and expressing particular types of emotions. © The Author 2013. Published by Oxford University Press on behalf of The Gerontological Society of America. All rights reserved. For permissions, please e-mail: firstname.lastname@example.org.
Romero, Enilda J.; Watson, Ginger S.
The use of animated pedagogical agents or avatars in instruction has lagged behind their use in entertainment. This is due in part to the cost and complexity of development and implementation of agents in educational settings, but also results from a lack of research to understand how emotions from animated agents influence instructional effectiveness. The phenomenological study presented here assesses the perceptions of eight learners interacting with low and high intensity emotionally expressive pedagogical agents in a computer-mediated environment. Research methods include maximum variation and snowball sampling with random assignment to treatment. The resulting themes incorporate perceptions of importance, agent humanness, enjoyment, implementation barriers, and suggested improvements. Design recommendations and implications for future research are presented.
Campos, Belinda; Shiota, Michelle N; Keltner, Dacher; Gonzaga, Gian C; Goetz, Jennifer L
Understanding positive emotions' shared and differentiating features can yield valuable insight into the structure of positive emotion space and identify emotion states, or aspects of emotion states, that are most relevant for particular psychological processes and outcomes. We report two studies that examined core relational themes (Study 1) and expressive displays (Study 2) for eight positive emotion constructs--amusement, awe, contentment, gratitude, interest, joy, love, and pride. Across studies, all eight emotions shared one quality: high positive valence. Distinctive core relational theme and expressive display patterns were found for four emotions--amusement, awe, interest, and pride. Gratitude was associated with a distinct core relational theme but not an expressive display. Joy and love were each associated with a distinct expressive display but their core relational themes also characterised pride and gratitude, respectively. Contentment was associated with a distinct expressive display but not a core relational theme. The implications of this work for the study of positive emotion are discussed.
Diéguez-Risco, Teresa; Aguado, Luis; Albert, Jacobo; Hinojosa, José Antonio
The influence of explicit evaluative processes on the contextual integration of facial expressions of emotion was studied in a procedure that required the participants to judge the congruency of happy and angry faces with preceding sentences describing emotion-inducing situations. Judgments were faster on congruent trials in the case of happy faces and on incongruent trials in the case of angry faces. At the electrophysiological level, a congruency effect was observed in the face-sensitive N170 component that showed larger amplitudes on incongruent trials. An interactive effect of congruency and emotion appeared on the LPP (late positive potential), with larger amplitudes in response to happy faces that followed anger-inducing situations. These results show that the deliberate intention to judge the contextual congruency of facial expressions influences not only processes involved in affective evaluation such as those indexed by the LPP but also earlier processing stages that are involved in face perception. Copyright © 2015. Published by Elsevier B.V.
Hogg, Rhona; Hanley, Janet; Smith, Pam
This article explores the content of letters of complaint by patients and carers about the behaviour, attitudes and communication of healthcare staff. The most common focus of patient complaints in the UK and other high-income countries is staff attitudes, behaviour and communication. There is a move to learn lessons from patient complaints, which can be used to improve patient care and experience. Fifty letters of complaint made by patients and carers relating to the behaviour, attitudes and communication of healthcare staff were analysed. Poor attitudes, behaviours and communication have significant negative impact on the emotional well-being of patients and carers. Many patients and carers have heightened sensitivities due to both health-related stresses and also other factors. The healthcare role is expected to include compassion and kindness. The concept of emotional labour is useful in explaining the skills and effort required of staff in this often invisible and undervalued aspect of health care. Given the increasing focus on patient experience, it is important that the importance of good staff attitudes, behaviours and communication is understood and that the emotional labour associated with this is recognised. An understanding of emotional intelligence can protect staff from burnout and other negative outcomes which those in a caring role can experience. © 2017 John Wiley & Sons Ltd.
Li, Chao-Yin; Murray, MaryAnne
To clarify the concept of 'expressed emotion' and its application to caregivers of older adults with dementia. Expressed emotion has been a useful construct for understanding the quality of family relationships affecting patients with mental illness and their caregivers. However, this concept has been developed without precisely defining 'expressed emotion' as it pertains to dementia patients. Clarity regarding expressed emotion will enable nurses to apply knowledge of expressed emotion and provide important information for the development of new clinical interventions for this specific population. Integrative review. A review of literature on expressed emotion by caregivers of older adults with dementia. The inclusion criteria were: (1) published in English or Chinese during 1970-2012; (2) included both research and theoretical review articles on expressed emotion in nursing and other disciplines such as psychology, psychiatry and sociology. Initially, 236 articles were screened, and finally, 32 articles were evaluated for this review. Emotional expression and expressed emotion were discussed to clarify the distinctions and address overlap between these two similar terms. In addition, expressed emotion was examined further from three different aspects: trait or state, social control and cross-cultural. Finally, the results of reviewed papers for expressed emotion on dementia patients were explored and synthesised. A conceptual definition and a theoretical framework for the concept of expressed emotion are urgently needed to further our understanding of this critical phenomenon. With increasing attention to caregiving for patients with dementia, including the concept of expressed emotion in the research of this field may accelerate understanding of the importance of the family dynamics in advanced ageing caregiving. The expressed emotion concept could guide much of current clinical practice and help professional nurses understand the family's experience and
Gaspar, Augusta; Esteves, Francisco G.
Prototypical facial expressions of emotion, also known as universal facial expressions, are the underpinnings of most research concerning recognition of emotions in both adults and children. Data on natural occurrences of these prototypes in natural emotional contexts are rare and difficult to obtain in adults. By recording naturalistic…
Doi, Hirokazu; Fujisawa, Takashi X.; Kanai, Chieko; Ohta, Haruhisa; Yokoi, Hideki; Iwanami, Akira; Kato, Nobumasa; Shinohara, Kazuyuki
This study investigated the ability of adults with Asperger syndrome to recognize emotional categories of facial expressions and emotional prosodies with graded emotional intensities. The individuals with Asperger syndrome showed poorer recognition performance for angry and sad expressions from both facial and vocal information. The group…
Bos, P.A.; Jap-Tjong, Nadine; Spencer, H.; Hofman, D.
Children use emotional facial expressions of others for guiding their behavior, a process which is important to a child’s social-emotional development. Earlier studies on facial interaction demonstrate that imitation of emotional expressions of others is automatic, yet can be dynamically modulated
Missana, Manuela; Grossmann, Tobias
Sensitive responding to others' emotional body expressions is an essential social skill in humans. Using event-related brain potentials, it has recently been shown that the ability to discriminate between emotional body expressions develops between 4 and 8 months of age. However, it is not clear whether the perception of emotional body expressions…
Ewing, Donna L.; Monsen, Jeremy J.; Kwoka, Maria
This action research project considered whether significant improvements in child and young person behavioural and emotional mental health could be achieved using school-based play workers as opposed to qualified therapists. This was seen as being an important practice question as access to qualified play therapists was severely restricted with…
Scheller, Elisa; Büchel, Christian; Gamer, Matthias
Diagnostic features of emotional expressions are differentially distributed across the face. The current study examined whether these diagnostic features are preferentially attended to even when they are irrelevant for the task at hand or when faces appear at different locations in the visual field. To this aim, fearful, happy and neutral faces were presented to healthy individuals in two experiments while measuring eye movements. In Experiment 1, participants had to accomplish an emotion classification, a gender discrimination or a passive viewing task. To differentiate fast, potentially reflexive, eye movements from a more elaborate scanning of faces, stimuli were either presented for 150 or 2000 ms. In Experiment 2, similar faces were presented at different spatial positions to rule out the possibility that eye movements only reflect a general bias for certain visual field locations. In both experiments, participants fixated the eye region much longer than any other region in the face. Furthermore, the eye region was attended to more pronouncedly when fearful or neutral faces were shown whereas more attention was directed toward the mouth of happy facial expressions. Since these results were similar across the other experimental manipulations, they indicate that diagnostic features of emotional expressions are preferentially processed irrespective of task demands and spatial locations. Saliency analyses revealed that a computational model of bottom-up visual attention could not explain these results. Furthermore, as these gaze preferences were evident very early after stimulus onset and occurred even when saccades did not allow for extracting further information from these stimuli, they may reflect a preattentive mechanism that automatically detects relevant facial features in the visual field and facilitates the orientation of attention towards them. This mechanism might crucially depend on amygdala functioning and it is potentially impaired in a number of
Anderson, Wendy G; Alexander, Stewart C; Rodriguez, Keri L; Jeffreys, Amy S; Olsen, Maren K; Pollak, Kathryn I; Tulsky, James A; Arnold, Robert M
Cancer patients have high levels of distress, yet oncologists often do not recognize patients' concerns. We sought to describe how patients with advanced cancer verbally express negative emotion to their oncologists. As part of the Studying Communication in Oncologist-Patient Encounters Trial, we audio-recorded 415 visits that 281 patients with advanced cancer made to their oncologists at three US cancer centers. Using qualitative methodology, we coded for verbal expressions of negative emotion, identified words patients used to express emotion, and categorized emotions by type and content. Patients verbally expressed negative emotion in 17% of the visits. The most commonly used words were: "concern," "scared," "worried," "depressed," and "nervous." Types of emotion expressed were: anxiety (46%), fear (25%), depression (12%), anger (9%), and other (8%). Topics about which emotion was expressed were: symptoms and functional concerns (66%), medical diagnoses and treatments (54%), social issues (14%), and the health care system (9%). Although all patients had terminal cancer, they expressed negative emotion overtly related to death and dying only 2% of the time. Patients infrequently expressed negative emotion to their oncologists. When they did, they typically expressed anxiety and fear, indicating concern about the future. When patients use emotionally expressive words such as those we described, oncologists should respond empathically, allowing patients to express their distress and concerns more fully.
Wingenbach, Tanja S. H.; Ashwin, Chris; Brosnan, Mark
There has been much research on sex differences in the ability to recognise facial expressions of emotions, with results generally showing a female advantage in reading emotional expressions from the face. However, most of the research to date has used static images and/or ‘extreme’ examples of facial expressions. Therefore, little is known about how expression intensity and dynamic stimuli might affect the commonly reported female advantage in facial emotion recognition. The current study in...
Roy P. C. Kessels
Full Text Available Behavioural problems are a key feature of frontotemporal lobar degeneration (FTLD. Also, FTLD patients show impairments in emotion processing. Specifically, the perception of negative emotional facial expressions is affected. Generally, however, negative emotional expressions are regarded as more difficult to recognize than positive ones, which thus may have been a confounding factor in previous studies. Also, ceiling effects are often present on emotion recognition tasks using full-blown emotional facial expressions. In the present study with FTLD patients, we examined the perception of sadness, anger, fear, happiness, surprise and disgust at different emotional intensities on morphed facial expressions to take task difficulty into account. Results showed that our FTLD patients were specifically impaired at the recognition of the emotion anger. Also, the patients performed worse than the controls on recognition of surprise, but performed at control levels on disgust, happiness, sadness and fear. These findings corroborate and extend previous results showing deficits in emotion perception in FTLD.
Full Text Available Integration of information from face and voice plays a central role in social interactions. The present study investigated the modulation of emotional intensity on the integration of facial-vocal emotional cues by recording EEG for participants while they were performing emotion identification task on facial, vocal, and bimodal angry expressions varying in emotional intensity. Behavioral results showed the rates of anger and reaction speed increased as emotional intensity across modalities. Critically, the P2 amplitudes were larger for bimodal expressions than for the sum of facial and vocal expressions for low emotional intensity stimuli, but not for middle and high emotional intensity stimuli. These findings suggested that emotional intensity modulates the integration of facial-vocal angry expressions, following the principle of Inverse Effectiveness (IE in multimodal sensory integration.
Dapelo, Marcela M; Hart, Sharon; Hale, Christiane; Morris, Robin; Lynch, Thomas R; Tchanturia, Kate
A large body of research has associated Eating Disorders with difficulties in socio-emotional functioning and it has been argued that they may serve to maintain the illness. This study aimed to explore facial expressions of positive emotions in individuals with Anorexia Nervosa (AN) and Bulimia Nervosa (BN) compared to healthy controls (HC), through an examination of the Duchenne smile (DS), which has been associated with feelings of enjoyment, amusement and happiness (Ekman et al., 1990). Sixty participants (AN=20; BN=20; HC=20) were videotaped while watching a humorous film clip. The duration and intensity of DS were subsequently analyzed using the facial action coding system (FACS) (Ekman and Friesen, 2003). Participants with AN displayed DS for shorter durations than BN and HC participants, and their DS had lower intensity. In the clinical groups, lower duration and intensity of DS were associated with lower BMI, and use of psychotropic medication. The study is the first to explore DS in people with eating disorders, providing further evidence of difficulties in the socio-emotional domain in people with AN. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Keenan, Kate; Hipwell, Alison; Hinze, Amanda; Babinski, Dara
Emotion dysregulation is often invoked as an important construct for understanding risk for psychopathology, but specificity of domains of emotion regulation in clinically relevant research is often lacking. In the present study Gross’ (2001) model of emotion regulation is used to generate hypotheses regarding the relative contribution of two specific types of deficits in emotion regulation, inhibited and disinhibited expression of negative emotion, to individual differences in depressive sym...
Eack, Shaun M.; MAZEFSKY, CARLA A.; Minshew, Nancy J.
Facial emotion perception is significantly affected in autism spectrum disorder (ASD), yet little is known about how individuals with ASD misinterpret facial expressions that result in their difficulty in accurately recognizing emotion in faces. This study examined facial emotion perception in 45 verbal adults with ASD and 30 age- and gender-matched volunteers without ASD to identify patterns of emotion misinterpretation during face processing that contribute to emotion recognition impairment...
Full Text Available Traditional emotion theories stress the importance of the face in the expression of emotions but bodily expressions are becoming increasingly important. Here we tested the hypothesis that similar physiological responses can be evoked by observing emotional face and body signals and that the reaction to angry signals is amplified in anxious individuals. We designed three experiments in which participants categorized emotional expressions from isolated facial and bodily expressions and from emotionally congruent and incongruent face-body compounds. Participants’ fixations were measured and their pupil size recorded with eye-tracking equipment, and their facial reactions measured with electromyography (EMG. The behavioral results support our prediction that the recognition of a facial expression is improved in the context of a matching posture and importantly, also vice versa. From their facial expression, it appeared that observers acted with signs of negative emotionality (increased corrugator activity to angry and fearful facial expressions and with positive emotionality (increased zygomaticus to happy facial expressions. What we predicted and found, was that angry and fearful cues from the face or the body, attracted more attention than happy cues. We further observed that responses evoked by angry cues were amplified in individuals with high anxiety scores. In sum, we show that people process bodily expressions of emotion in a similar fashion as facial expressions and that the congruency between the emotional signals from the face and body ameliorates the recognition of the emotion.
Full Text Available Yazhou Jin,* Zhiqi Mao,* Zhipei Ling, Xin Xu, Zhiyuan Zhang, Xinguang Yu Department of Neurosurgery, People’s Liberation Army General Hospital, Beijing, People’s Republic of China *These authors contributed equally to this work Background: Parkinson’s disease (PD patients exhibit deficits in emotional recognition and expression abilities, including emotional faces and voices. The aim of this study was to explore emotional processing in pre-deep brain stimulation (pre-DBS PD patients using two sensory modalities (visual and auditory. Methods: Fifteen PD patients who needed DBS surgery and 15 healthy, age- and gender-matched controls were recruited as participants. All participants were assessed by the Karolinska Directed Emotional Faces database 50 Faces Recognition test. Vocal recognition was evaluated by the Montreal Affective Voices database 50 Voices Recognition test. For emotional facial expression, the participants were asked to imitate five basic emotions (neutral, happiness, anger, fear, and sadness. The subjects were required to express nonverbal vocalizations of the five basic emotions. Fifteen Chinese native speakers were recruited as decoders. We recorded the accuracy of the responses, reaction time, and confidence level. Results: For emotional recognition and expression, the PD group scored lower on both facial and vocal emotional processing than did the healthy control group. There were significant differences between the two groups in both reaction time and confidence level. A significant relationship was also found between emotional recognition and emotional expression when considering all participants between the two groups together. Conclusion: The PD group exhibited poorer performance on both the recognition and expression tasks. Facial emotion deficits and vocal emotion abnormalities were associated with each other. In addition, our data allow us to speculate that emotional recognition and expression may share a common
Johnson, Stacy R.; Finlon, Kristy J.; Izard, Carroll E.
Research Findings: This article describes the development and evaluation of the Behavior and Emotion Expression Observation System (BEEOS), a direct observation tool to characterize preschoolers' social and emotion behaviors during semistructured activities in the classroom. The BEEOS was used to observe 148 Head Start preschoolers, and…
Hejmadi, A; Davidson, R J; Rozin, P
Subjects were presented with videotaped expressions of 10 classic Hindu emotions. The 10 emotions were (in rough translation from Sanskrit) anger, disgust, fear, heroism, humor-amusement, love, peace, sadness, shame-embarrassment, and wonder. These emotions (except for shame) and their portrayal were described about 2,000 years ago in the Natyasastra, and are enacted in the contemporary Hindu classical dance. The expressions are dynamic and include both the face and the body, especially the hands. Three different expressive versions of each emotion were presented, along with 15 neutral expressions. American and Indian college students responded to each of these 45 expressions using either a fixed-response format (10 emotion names and "neutral/no emotion") or a totally free response format. Participants from both countries were quite accurate in identifying emotions correctly using both fixed-choice (65% correct, expected value of 9%) and free-response (61% correct, expected value close to zero) methods.
Aguilera, Adrian; López, Steven R.; Breitborde, Nicholas J. K.; Kopelowicz, Alex; Zarate, Roberto
This study examines whether the sociocultural context moderates the relationship between families' expressed emotion and clinical outcomes in schizophrenia. In a sample of 60 Mexican-American caregivers and their ill relatives, we first assessed whether expressed emotion (EE), and its indices (criticism, EOI and warmth), relate to relapse. Secondly, we extended the analysis of EE and its indices to a longitudinal assessment of symptomatology. Last, we tested whether bidimensional acculturation moderates the relationship between EE (and its indices) and both relapse and symptom trajectory over time. Results indicated that EOI was associated with increased relapse and that criticism was associated with increased symptomatology. Additionally as patients' Mexican enculturation (Spanish language and media involvement) decreased, EE was increasingly related to relapse. For symptomatology, as the patient's U.S. acculturation (English language and media involvement) increased, EE was associated with increased symptoms longitudinally. Our results replicate and extend past research on how culture might shape the way family factors relate to the course of schizophrenia. PMID:21090883
Fang, Xia; van Kleef, Gerben A; Sauter, Disa A
Dynamic changes in emotional expressions are a valuable source of information in social interactions. As the expressive behaviour of a person changes, the inferences drawn from the behaviour may also change. Here, we test the possibility that dynamic changes in emotional expressions affect person perception in terms of stable trait attributions. Across three experiments, we examined perceivers' inferences about others' personality traits from changing emotional expressions. Expressions changed from one emotion ("start emotion") to another emotion ("end emotion"), allowing us to disentangle potential primacy, recency, and averaging effects. Drawing on three influential models of person perception, we examined perceptions of dominance and affiliation (Experiment 1a), competence and warmth (Experiment 1b), and dominance and trustworthiness (Experiment 2). A strong recency effect was consistently found across all trait judgments, that is, the end emotion of dynamic expressions had a strong impact on trait ratings. Evidence for a primacy effect was also observed (i.e. the information of start emotions was integrated), but less pronounced, and only for trait ratings relating to affiliation, warmth, and trustworthiness. Taken together, these findings suggest that, when making trait judgements about others, observers weigh the most recently displayed emotion in dynamic expressions more heavily than the preceding emotion.
Poole, Ernest Andre
In this review I explore and discuss the use of micro and subtle expression training in the social sciences. These trainings, offered commercially, are designed and endorsed by noted psychologist Paul Ekman, co-author of the Facial Action Coding System, a comprehensive system of measuring muscular movement in the face and its relationship to the expression of emotions. The trainings build upon that seminal work and present them in a way for either the layperson or researcher to easily add to their personal toolbox for a variety of purposes. Outlined are my experiences across the training products, how they could be used in social science research, a brief comparison to automated systems, and possible next steps.
Lang, Katie; Larsson, Emma E C; Mavromara, Liza; Simic, Mima; Treasure, Janet; Tchanturia, Kate
This study aimed to investigate emotion expression in a large group of children, adolescents and adults with Anorexia Nervosa (AN), and investigate the associated clinical correlates. One hundred and forty-one participants (AN=66, HC= 75) were recruited and positive and negative film clips were used to elicit emotion expressions. The Facial Activation Coding system (FACES) was used to code emotion expression. Subjective ratings of emotion were collected. Individuals with AN displayed less positive emotions during the positive film clip compared to healthy controls (HC). There was no significant difference between the groups on the Positive and Negative Affect Scale (PANAS). The AN group displayed emotional incongruence (reporting a different emotion to what would be expected given the stimuli, with limited facial affect to signal the emotion experienced), whereby they reported feeling significantly higher rates of negative emotion during the positive clip. There were no differences in emotion expression between the groups during the negative film clip. Despite this individuals with AN reported feeling significantly higher levels of negative emotions during the negative clip. Diminished positive emotion expression was associated with more severe clinical symptoms, which could suggest that these individuals represent a group with serious social difficulties, which may require specific attention in treatment. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
This paper examines emotional expression experienced by female principals in the Arab school system in Israel over their managerial careers--role-related emotions that they choose to express or repress before others. I employed narrative methodology, interviewing nine female principals from the Arab school system to investigate expression of…
Hernández, Maciel M.; Eisenberg, Nancy; Valiente, Carlos; Spinrad, Tracy L.; Berger, Rebecca H.; VanSchyndel, Sarah K.; Thompson, Marilyn S.; Southworth, Jody; Silva, Kassondra M.
Positive emotional expressivity has been associated with increased social competence and decreased maladjustment in childhood. However, a few researchers have found null or even positive associations between positive emotional expressivity and maladjustment, which suggests that there may be nuanced associations of positive expressivity, perhaps as…
H. S. Asthana
Full Text Available This study examined hemispatial bias in a free-viewing condition of the judgement of facial expressions of emotions. Right-handed male subjects were asked to judge the intensity, in terms of expressiveness, of facial emotion in normal and mirror-reversed orientations. Expressions in mirror-reversed orientation were perceived as more intense than in normal orientation.
Sorber, Anne Verbeck; Cunningham, Joseph G.
This study investigated effects of gender, emotion, and family expressiveness on preschool children's reactions to narrative characters' emotion expressions. Forty-five preschool children rank-ordered playmate preferences for male and female story characters who expressed happiness, anger, sadness, fear, and neutrality and indicated how much they…
Panjwani, Naaila; Chaplin, Tara M.; Sinha, Rajita; Mayes, Linda C.
Gender roles in mainstream U.S. culture suggest that girls express more happiness, sadness, anxiety, and shame/embarrassment than boys, while boys express more anger and externalizing emotions, such as contempt. However, gender roles and emotion expression may be different in low-income and ethnically diverse families, as children and parents are often faced with greater environmental stressors and may have different gender expectations. This study examined gender differences in emotion expre...
Full Text Available Emotions are adaptive reaction to events from the environment and they represent the central part of every person's life. Willing emotional expression influences people from the environment according to our expectations if they recognize our emotions. Studies on expressing emotions in persons with visual impairments indicate the existence of the same type of spontaneous emotional expressions as in typically developing population. Also, these studies point to difficulties in presenting willing emotional expressions. The aim of this research was to determine the difference in emotional expression between high school students with visual impairments and their typically developing peers. The sample consisted of 33 students with visual impairments and the same number of students with no developmental disabilities. Emotional states simulation scenario by Friedman et al. was used in this research. Emotional expression was assessed with regard to the level of success in simulating seven emotions (happiness, sadness, anger, disgust, surprise, fear, and neutral state. The participants' task was to simulate the given emotional states in three structured situations (uttering two sentences and a series of vowels. The simulation of emotions was recorded. On the basis of video recordings, three independent assessors measured the success in simulating emotions on a nine-point scale. By analyzing the obtained results, a statistically significant difference in emotional expression was determined between the participants with visual impairments and their peers with no developmental disabilities (F(1=3.692; p=0.05.Arithmetic mean differences are statistically significant for simulating disgust (p=0.002 and surprise (p=0.01. In the group of participants with visual impairments, gender was a significant factor in simulating the emotional state of happiness (p=0.024, while type of school was a significant factor in simulating sadness (p=0.027.
Wong, Elena; Tschan, Franziska; Messerli, Laurence; Semmer, Norbert K.
Expressing emotions has social functions; it provides information, affects social interactions, and shapes relationships with others. Expressing positive emotions could be a strategic tool for improving goal attainment during social interactions at work. Such effects have been found in research on social contagion, impression management, and emotion work. However, expressing emotions one does not feel entails the risk of being perceived as inauthentic. This risk may well be worth taking when the emotions felt are negative, as expressing negative emotions usually has negative effects. When experiencing positive emotions, however, expressing them authentically promises benefits, and the advantage of amplifying them is not so obvious. We postulated that expressing, and amplifying, positive emotions would foster goal attainment in social interactions at work, particularly when dealing with superiors. Analyses are based on 494 interactions involving the pursuit of a goal by 113 employes. Multilevel analyses, including polynomial analyses, show that authentic display of positive emotions supported goal attainment throughout. However, amplifying felt positive emotions promoted goal attainment only in interactions with superiors, but not with colleagues. Results are discussed with regard to the importance of hierarchy for detecting, and interpreting, signs of strategic display of positive emotions. PMID:23675358
Full Text Available Expressing emotions has social functions; it provides information, affects social interactions, and shapes relationships with others. Expressing positive emotions could be a strategic tool for improving goal attainment during social interactions at work. Such effects have been found in research on social contagion, impression management, and emotion work. However, expressing emotions one does not feel entails the risk of being perceived as inauthentic. This risk may well be worth taking when the emotions felt are negative, as expressing negative emotions usually has negative effects. When experiencing positive emotions, however, expressing them authentically promises benefits, and the advantage of amplifying them is not so obvious. We postulated that expressing, and amplifying, positive emotions would foster goal attainment in social interactions at work, particularly when dealing with superiors. Analyses are based on 494 interactions involving the pursuit of a goal by 113 employees. Multilevel analyses, including polynomial analyses, show that authentic display of positive emotions supported goal attainment throughout. However, amplifying felt positive emotions promoted goal attainment only in interactions with superiors, but not with colleagues. Results are discussed with regard to the importance of hierarchy for detecting, and interpreting, signs of strategic display of positive emotions.
Trinkler, Iris; Cleret de Langavant, Laurent; Bachoud-Lévi, Anne-Catherine
Patients with Huntington's disease (HD), a neurodegenerative disorder that causes major motor impairments, also show cognitive and emotional deficits. While their deficit in recognising emotions has been explored in depth, little is known about their ability to express emotions and understand their feelings. If these faculties were impaired, patients might not only mis-read emotion expressions in others but their own emotions might be mis-interpreted by others as well, or thirdly, they might have difficulties understanding and describing their feelings. We compared the performance of recognition and expression of facial emotions in 13 HD patients with mild motor impairments but without significant bucco-facial abnormalities, and 13 controls matched for age and education. Emotion recognition was investigated in a forced-choice recognition test (FCR), and emotion expression by filming participants while they mimed the six basic emotional facial expressions (anger, disgust, fear, surprise, sadness and joy) to the experimenter. The films were then segmented into 60 stimuli per participant and four external raters performed a FCR on this material. Further, we tested understanding of feelings in self (alexithymia) and others (empathy) using questionnaires. Both recognition and expression were impaired across different emotions in HD compared to controls and recognition and expression scores were correlated. By contrast, alexithymia and empathy scores were very similar in HD and controls. This might suggest that emotion deficits in HD might be tied to the expression itself. Because similar emotion recognition-expression deficits are also found in Parkinson's Disease and vascular lesions of the striatum, our results further confirm the importance of the striatum for emotion recognition and expression, while access to the meaning of feelings relies on a different brain network, and is spared in HD. Copyright © 2011 Elsevier Ltd. All rights reserved.
Schmidt, Ricarda; Tetzlaff, Anne; Hilbert, Anja
A sizeable body of research has documented Expressed Emotion (EE) to predict clinical outcomes in various psychiatric disorders, including eating disorders. Patients' perceptions of relative's EE, however, were found to play an important role in the processing of EE. This study aimed to examine the level of perceived EE in adolescent binge-eating disorder (BED) and its impact on eating disorder psychopathology. Adolescents (12-20 years) seeking treatment for BED (n = 40) were compared to adolescents without current or lifetime eating disorder (CG; n = 40). Both groups were stratified according to age, sex, body mass index (BMI, kg/m(2)), and socio-economic status. The Five Minute Speech Sample (FMSS) and the Brief Dyadic Scale of EE were administered to assess patients' perceived maternal EE. Additionally, adolescents and mothers completed questionnaires on eating disorder and general psychopathology. On the FMSS, 37.5 % of patients with BED perceived their mothers as high EE (vs. 12.5 % in the CG). On the Brief Dyadic Scale of EE, patients with BED reported significantly higher levels of perceived maternal criticism, emotional overinvolvement, and lower levels of perceived warmth than controls. After controlling for the diagnosis of BED, perceived criticism and warmth, as assessed by questionnaire, significantly explained adolescents' global eating disorder psychopathology. Negative perceptions of maternal behavior and emotional atmosphere towards the child are characteristic of adolescent BED. As documented for other eating disorders, family factors are likely to have substantial implications for the maintenance and treatment of adolescent BED.
Ivanete Ribeiro do Nascimento
Full Text Available Breast cancer is one of the cancers most feared by women for its high incidence and its psychological effects that affect the perception of sexuality and self-image. Objective: To identify the difficulties of nursing professionals in the treatment of patients with cancer, from the standpoint of a terminally ill patient of breast cancer. Methodology: This is a case study of a patient who is in the terminal stages of breast cancer. We carried out the survey of literature in journals indexed the databases LILACS and SciELO Open Access and English, on terminally ill cancer. Results: Feelings of loneliness and sadness were softened and smoothed by the attitude and disposition of nursing professionals. In moments of intervention needs of physical care, nursing care was provided. Conclusion: The nursing staff has always demonstrated skills in treating patients with cancer, providing quality care, humane and comprehensive, meeting all your needs biopsicoespiritual.
Juslin, Patrik N; Laukka, Petri; Bänziger, Tanja
It has been the subject of much debate in the study of vocal expression of emotions whether posed expressions (e.g., actor portrayals) are different from spontaneous expressions. In the present investigation, we assembled a new database consisting of 1877 voice clips from 23 datasets, and used it to systematically compare spontaneous and posed expressions across 3 experiments. Results showed that (a) spontaneous expressions were generally rated as more genuinely emotional than were posed expressions, even when controlling for differences in emotion intensity, (b) there were differences between the two stimulus types with regard to their acoustic characteristics, and (c) spontaneous expressions with a high emotion intensity conveyed discrete emotions to listeners to a similar degree as has previously been found for posed expressions, supporting a dose-response relationship between intensity of expression and discreteness in perceived emotions. Our conclusion is that there are reliable differences between spontaneous and posed expressions, though not necessarily in the ways commonly assumed. Implications for emotion theories and the use of emotion portrayals in studies of vocal expression are discussed.
Chang, Valerie T; Overall, Nickola C; Madden, Helen; Low, Rachel S T
The current research extends prior research linking negative emotions and emotion regulation tendencies to memory by investigating whether (a) naturally occurring negative emotions during routine weekly life are associated with more negatively biased memories of prior emotional experiences-a bias called projection; (b) tendencies to regulate emotions via expressive suppression are associated with greater projection bias in memory of negative emotions; and (c) greater projection bias in memory is associated with poorer future well-being. Participants (N = 308) completed a questionnaire assessing their general tendencies to engage in expressive suppression. Then, every week for 7 weeks, participants reported on (a) the negative emotions they experienced across the current week (e.g., "This week, I felt 'sad'"), (b) their memories of the negative emotions they experienced the prior week (e.g., "Last week, I felt 'sad'"), and (c) their well-being. First, participants demonstrated significant projection bias in memory: Greater negative emotions in a given week were associated with remembering emotions in the prior week more negatively than those prior emotions were originally reported. Second, projection bias in memory of negative emotions was greater for individuals who reported greater tendencies to regulate emotions via expressive suppression. Third, greater projection bias in memory of negative emotions was associated with reductions in well-being across weeks. These 3 novel findings indicate that (a) current negative emotions bias memory of past emotions, (b) this memory bias is magnified for people who habitually use expressive suppression to regulate emotions, and (c) this memory bias may undermine well-being over time. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Haberman, Jason; Whitney, David
There has been a recent surge in the study of ensemble coding, the idea that the visual system represents a set of similar items using summary statistics (Alvarez & Oliva, 2008; Ariely, 2001; Chong & Treisman, 2003; Parkes, Lund, Angelucci, Solomon, & Morgan, 2001). We previously demonstrated that this ability extends to faces and thus requires a high level of object processing (Haberman & Whitney, 2007, 2009). Recent debate has centered on the nature of the summary representation of size (e.g., Myczek & Simons, 2008) and whether the perceived average simply reflects the sampling of a very small subset of the items in a set. In the present study, we explored this further in the context of faces, asking observers to judge the average expressions of sets of faces containing emotional outliers. Our results suggest that the visual system implicitly and unintentionally discounts the emotional outliers, thereby computing a summary representation that encompasses the vast majority of the information present. Additional computational modeling and behavioral results reveal that an intentional cognitive sampling strategy does not accurately capture observer performance. Observers derive precise ensemble information given a 250-msec exposure, suggesting a rapid and flexible system not bound by the limits of serial attention.
Aguilera, Adrian; López, Steven R; Breitborde, Nicholas J K; Kopelowicz, Alex; Zarate, Roberto
This study examined whether the sociocultural context moderates the relationship between families' expressed emotion (EE) and clinical outcomes in schizophrenia. In a sample of 60 Mexican American caregivers and their ill relatives, we first assessed whether EE and its indices (criticism, emotional overinvolvement [EOI], and warmth) related to relapse. Second, we extended the analysis of EE and its indices to a longitudinal assessment of symptomatology. Last, we tested whether bidimensional acculturation moderated the relationship between EE (and its indices) and both relapse and symptom trajectory over time. Results indicated that EOI was associated with increased relapse and that criticism was associated with increased symptomatology. Additionally, as patients' Mexican enculturation (Spanish language and media involvement) decreased, EE was increasingly related to relapse. For symptomatology, as patients' U.S. acculturation (English language and media involvement) increased, EE was associated with increased symptoms longitudinally. Our results replicate and extend past research on how culture might shape the way family factors relate to the course of schizophrenia. PsycINFO Database Record (c) 2010 APA, all rights reserved
Bonanno, George A; Colak, Deniz M; Keltner, Dacher; Shiota, Michelle N; Papa, Anthony; Noll, Jennie G; Putnam, Frank W; Trickett, Penelope K
Positive emotions promote adjustment to aversive life events. However, evolutionary theory and empirical research on trauma disclosure suggest that in the context of stigmatized events, expressing positive emotions might incur social costs. To test this thesis, the authors coded genuine (Duchenne) smiling and laughter and also non-Duchenne smiling from videotapes of late-adolescent and young adult women, approximately half with documented histories of childhood sexual abuse (CSA), as they described the most distressing event of their lives. Consistent with previous studies, genuine positive emotional expression was generally associated with better social adjustment two years later. However, as anticipated, CSA survivors who expressed positive emotion in the context of describing a past CSA experience had poorer long-term social adjustment, whereas CSA survivors who expressed positive emotion while describing a nonabuse experience had improved social adjustment. These findings suggest that the benefits of positive emotional expression may often be context specific.
Hosotani, Rika; Imai-Matsumura, Kyoko
The present study investigates the emotional experience, expression, and regulation processes of high-quality Japanese elementary school teachers while they interact with children, in terms of teachers' emotional competence. Qualitative analysis of interview data demonstrated that teachers had various emotional experiences including self-elicited…
Eack, Shaun M.; Mazefsky, Carla A.; Minshew, Nancy J.
Facial emotion perception is significantly affected in autism spectrum disorder, yet little is known about how individuals with autism spectrum disorder misinterpret facial expressions that result in their difficulty in accurately recognizing emotion in faces. This study examined facial emotion perception in 45 verbal adults with autism spectrum…
Chickerur, Satyadhyan; Joshi, Kartik
Emotion detection using facial images is a technique that researchers have been using for the last two decades to try to analyze a person's emotional state given his/her image. Detection of various kinds of emotion using facial expressions of students in educational environment is useful in providing insight into the effectiveness of tutoring…
Huis In 't Veld, E.M.J.; van Boxtel, G.J.M.; de Gelder, B.
Body postures provide clear signals about emotional expressions, but so far it is not clear what muscle patterns are associated with specific emotions. This study lays the groundwork for a Body Action Coding System by investigating what combinations of muscles are used for emotional bodily
Vaiman, M.; Wagner, M.A.; Caicedo, E.; Pereno, G.L.
Pictures of facial expressions of emotion are used in a wide range of experiments. The last decade has seen an increase in the number of studies presenting local sets of emotion stimuli. However, only a few existing sets contain pictures of Latin Americans, despite the growing attention emotion
Sharma, Jyoti; Dhar, Rajib Lochan; Tyagi, Akansha
The study examined the extent to which work-family conflicts cause stress among nursing staff and its subsequent impact on their psychological health. It also examined if the emotional intelligence level of the nursing staff acted as a moderator between their level of stress and psychological health. A survey was carried out on 693 nursing staff associated with 33 healthcare institutions in Uttarakhand, India. A hierarchical multiple regression analysis was carried out to understand the relationships shared by independent (work-family conflicts) and dependent (psychological health) constructs with the mediator (stress) as well as the moderator (emotional intelligence). The results revealed that stress acted as a mediator between work-family conflict of the nursing staff and their psychological health. However, their emotional intelligence level acted as a moderator between their stress level and psychological health. To conclude, the crucial roles of emotional intelligence in controlling the impact of stress on psychological health along with the practical as well as theoretical implications are also discussed. Copyright © 2015 Elsevier Inc. All rights reserved.
Huston, Sally A; Blount, Ronald L; Heidesch, Troy; Southwood, Robin
Poor adherence to self-care among youth with type-1 diabetes (YWD) can lead to significant long-term health problems. Negative diabetes-related emotions (NDRE) are common, and are significantly correlated with poor/deteriorating A1c. Resilient youth handle diabetes self-care challenges, such as adjusting for diabetes in public, better. Resiliency skills and perceptions include benefit finding (BF), fitting in with friends (FI), diabetes acceptance (DA), emotion processing (EP) and emotion expression (EE). First study goal: to verify structure of underlying measurement variables: NDRE, EP, EE, BF, DA, FI and comfort in adjusting for diabetes in public (CA) among youth 11-16 yr of age with diabetes. We also hypothesize: (i) YWD who engage in EP and EE will have higher levels of BF, FI, DA, (ii) EP and EE will moderate NDRE impact and (iii) higher levels of EP, EE, BF, FI and DA will be associated with higher CA. 243 summer diabetes campers between 11-16 yr of age. Pre-camp survey. Measurement variables were verified. EP and EE to friends were positively associated with BF, FI and DA for most YWD. NDRE was negatively associated with FI and DA, and for YWD aged 14-16 yr with CA. FI was positively associated with CA. EE moderated the impact of NDRE on CA among youth 11-13 yr. R 2 for CA in youth 14-16 yr was 48.2%, for 11-13 yr was 38.3%. DA was positively associated with CA for youth 14-16 yr. Resilience factors appear to influence CA either directly or indirectly. © 2016 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.
Brewer, Rebecca; Biotti, Federica; Catmur, Caroline; Press, Clare; Happé, Francesca; Cook, Richard; Bird, Geoffrey
The difficulties encountered by individuals with autism spectrum disorder (ASD) when interacting with neurotypical (NT, i.e. nonautistic) individuals are usually attributed to failure to recognize the emotions and mental states of their NT interaction partner. It is also possible, however, that at least some of the difficulty is due to a failure of NT individuals to read the mental and emotional states of ASD interaction partners. Previous research has frequently observed deficits of typical facial emotion recognition in individuals with ASD, suggesting atypical representations of emotional expressions. Relatively little research, however, has investigated the ability of individuals with ASD to produce recognizable emotional expressions, and thus, whether NT individuals can recognize autistic emotional expressions. The few studies which have investigated this have used only NT observers, making it impossible to determine whether atypical representations are shared among individuals with ASD, or idiosyncratic. This study investigated NT and ASD participants' ability to recognize emotional expressions produced by NT and ASD posers. Three posing conditions were included, to determine whether potential group differences are due to atypical cognitive representations of emotion, impaired understanding of the communicative value of expressions, or poor proprioceptive feedback. Results indicated that ASD expressions were recognized less well than NT expressions, and that this is likely due to a genuine deficit in the representation of typical emotional expressions in this population. Further, ASD expressions were equally poorly recognized by NT individuals and those with ASD, implicating idiosyncratic, rather than common, atypical representations of emotional expressions in ASD. © 2015 The Authors Autism Research published by Wiley Periodicals, Inc. on behalf of International Society for Autism Research.
Aviezer, Hillel; Trope, Yaacov; Todorov, Alexander
The distinction between positive and negative emotions is fundamental in emotion models. Intriguingly, neurobiological work suggests shared mechanisms across positive and negative emotions. We tested whether similar overlap occurs in real-life facial expressions. During peak intensities of emotion, positive and negative situations were successfully discriminated from isolated bodies but not faces. Nevertheless, viewers perceived illusory positivity or negativity in the nondiagnostic faces when seen with bodies. To reveal the underlying mechanisms, we created compounds of intense negative faces combined with positive bodies, and vice versa. Perceived affect and mimicry of the faces shifted systematically as a function of their contextual body emotion. These findings challenge standard models of emotion expression and highlight the role of the body in expressing and perceiving emotions.
Ullrich, Philip M; Lutgendorf, Susan K
The effects of two journaling interventions, one focusing on emotional expression and the other on both cognitive processing and emotional expression, were compared during 1 month of journaling about a stressful or traumatic event. One hundred twenty-two students were randomly assigned to one of three writing conditions: (a) focusing on emotions related to a trauma or stressor, (b) focusing on cognitions and emotions related to a trauma or stressor, or (c) writing factually about media events. Writers focusing on cognitions and emotions developed greater awareness of the positive benefits of the stressful event than the other two groups. This effect was apparently mediated by greater cognitive processing during writing. Writers focusing on emotions alone reported more severe illness symptoms during the study than those in other conditions. This effect appeared to be mediated by a greater focus on negative emotional expression during writing.
Full Text Available When presented with emotional visual scenes, older adults have been found to be equally capable to regulate emotion expression as younger adults, corroborating the view that emotion regulation skills are maintained or even improved in later adulthood. However, the possibility that gaze direction might help achieve an emotion control goal has not been taken into account, raising the question whether the effortful processing of expressive regulation is really spared from the general age-related decline. Since it does not allow perceptual attention to be redirected away from the emotional source, music provides a useful way to address this question. In the present study, affective, behavioral and physiological consequences of free expression of emotion, expressive suppression and expressive enhancement were measured in 31 younger and 30 older adults while they listened to positive and negative musical excerpts. The main results indicated that compared to younger adults, older adults reported experiencing less emotional intensity in response to negative music during the free expression of emotion condition. No age difference was found in the ability to amplify or reduce emotional expressions. However, an age-related decline in the ability to reduce the intensity of emotional state and an age-related increase in physiological reactivity were found when participants were instructed to suppress negative expression. Taken together, the current data support previous findings suggesting an age-related change in response to music. They also corroborate the observation that older adults are as efficient as younger adults at controlling behavioral expression. But most importantly, they suggest that when faced with auditory sources of negative emotion, older age does not always confer a better ability to regulate emotions.
Vieillard, Sandrine; Harm, Jonathan; Bigand, Emmanuel
When presented with emotional visual scenes, older adults have been found to be equally capable to regulate emotion expression as younger adults, corroborating the view that emotion regulation skills are maintained or even improved in later adulthood. However, the possibility that gaze direction might help achieve an emotion control goal has not been taken into account, raising the question whether the effortful processing of expressive regulation is really spared from the general age-related decline. Since it does not allow perceptual attention to be redirected away from the emotional source, music provides a useful way to address this question. In the present study, affective, behavioral, and physiological consequences of free expression of emotion, expressive suppression and expressive enhancement were measured in 31 younger and 30 older adults while they listened to positive and negative musical excerpts. The main results indicated that compared to younger adults, older adults reported experiencing less emotional intensity in response to negative music during the free expression of emotion condition. No age difference was found in the ability to amplify or reduce emotional expressions. However, an age-related decline in the ability to reduce the intensity of emotional state and an age-related increase in physiological reactivity were found when participants were instructed to suppress negative expression. Taken together, the current data support previous findings suggesting an age-related change in response to music. They also corroborate the observation that older adults are as efficient as younger adults at controlling behavioral expression. But most importantly, they suggest that when faced with auditory sources of negative emotion, older age does not always confer a better ability to regulate emotions.
Bronson Blake Harry
Full Text Available It is widely assumed that the fusiform face area (FFA, a brain region specialised for face perception, is not involved in processing emotional expressions. This assumption is based on the proposition that the FFA is involved in face identification and only processes features that are invariant across changes due to head movements, speaking and expressing emotions. The present study tested this proposition by examining whether the response in the human FFA varies across emotional expressions with functional magnetic resonance imaging and brain decoding analysis techniques (n = 11. A one versus all classification analysis showed that most emotional expressions that participants perceived could be reliably predicted from the neural pattern of activity in left and the right FFA, suggesting that the perception of different emotional expressions recruit partially non-overlaping neural mechanisms. In addition, emotional expressions could also be decoded from the pattern of activity in the early visual cortex (EVC, indicating that retinotopic cortex also shows a differential response to emotional expressions. These results cast doubt on the idea that the FFA is involved in expression invariant face processing, and instead indicate that emotional expressions evoke partially de-correlated signals throughout occipital and posterior temporal cortex.
Miyata, Hiromitsu; Nishimura, Ritsuko; Okanoya, Kazuo; Kawai, Nobuyuki
A Noh mask worn by expert actors when performing on a Japanese traditional Noh drama is suggested to convey countless different facial expressions according to different angles of head/body orientation. The present study addressed the question of how different facial parts of a Noh mask, including the eyebrows, the eyes, and the mouth, may contribute to different emotional expressions. Both experimental situations of active creation and passive recognition of emotional facial expressions were introduced. In Experiment 1, participants either created happy or sad facial expressions, or imitated a face that looked up or down, by actively changing each facial part of a Noh mask image presented on a computer screen. For an upward tilted mask, the eyebrows and the mouth shared common features with sad expressions, whereas the eyes with happy expressions. This contingency tended to be reversed for a downward tilted mask. Experiment 2 further examined which facial parts of a Noh mask are crucial in determining emotional expressions. Participants were exposed to the synthesized Noh mask images with different facial parts expressing different emotions. Results clearly revealed that participants primarily used the shape of the mouth in judging emotions. The facial images having the mouth of an upward/downward tilted Noh mask strongly tended to be evaluated as sad/happy, respectively. The results suggest that Noh masks express chimeric emotional patterns, with different facial parts conveying different emotions This appears consistent with the principles of Noh which highly appreciate subtle and composite emotional expressions, as well as with the mysterious facial expressions observed in Western art. It was further demonstrated that the mouth serves as a diagnostic feature in characterizing the emotional expressions. This indicates the superiority of biologically-driven factors over the traditionally formulated performing styles when evaluating the emotions of the Noh masks.
Miyata, Hiromitsu; Nishimura, Ritsuko; Okanoya, Kazuo; Kawai, Nobuyuki
Background A Noh mask worn by expert actors when performing on a Japanese traditional Noh drama is suggested to convey countless different facial expressions according to different angles of head/body orientation. The present study addressed the question of how different facial parts of a Noh mask, including the eyebrows, the eyes, and the mouth, may contribute to different emotional expressions. Both experimental situations of active creation and passive recognition of emotional facial expressions were introduced. Methodology/Principal Findings In Experiment 1, participants either created happy or sad facial expressions, or imitated a face that looked up or down, by actively changing each facial part of a Noh mask image presented on a computer screen. For an upward tilted mask, the eyebrows and the mouth shared common features with sad expressions, whereas the eyes with happy expressions. This contingency tended to be reversed for a downward tilted mask. Experiment 2 further examined which facial parts of a Noh mask are crucial in determining emotional expressions. Participants were exposed to the synthesized Noh mask images with different facial parts expressing different emotions. Results clearly revealed that participants primarily used the shape of the mouth in judging emotions. The facial images having the mouth of an upward/downward tilted Noh mask strongly tended to be evaluated as sad/happy, respectively. Conclusions/Significance The results suggest that Noh masks express chimeric emotional patterns, with different facial parts conveying different emotions This appears consistent with the principles of Noh which highly appreciate subtle and composite emotional expressions, as well as with the mysterious facial expressions observed in Western art. It was further demonstrated that the mouth serves as a diagnostic feature in characterizing the emotional expressions. This indicates the superiority of biologically-driven factors over the traditionally
Leppanen, Jenni; Ng, Kah Wee; Tchanturia, Kate; Treasure, Janet
Accurate interpretation and appropriate expression of emotions are key aspects of social-cognition. Several mental disorders are characterised by transdiagnostic difficulties in these areas and, recently, there has been increasing interest in exploring the effects of oxytocin on social-emotional functioning. This review consists of 33 studies. Fifteen of the studies included people with autism spectrum disorder, schizophrenia, borderline personality disorder, frontotemporal dementia, anorexia nervosa, bulimia nervosa, post-traumatic stress disorder, depression, and opioid and alcohol dependence. We conducted ten meta-analyses examining the effects of intranasal oxytocin on expression of emotions, emotional theory of mind, sensitivity to recognise basic emotions, and recognition of basic emotions. A single dose of intranasal oxytocin significantly improved the recognition of basic emotions, particularly fear, and increased the expression of positive emotions among the healthy individuals. Oxytocin did not significantly influence theory of mind or the expression of negative emotions among the healthy individuals. Finally, intranasal oxytocin did not significantly influence interpretation or expression of emotions among the clinical populations. Copyright © 2017. Published by Elsevier Ltd.
Hufnagel, Elizabeth J.
As we face the challenges of serious environmental issues, science education has made a commitment to improving environmental literacy, in particular climate literacy (NRC, 2012; 2013). With an increased focus on climate change education in the United States, more research on the teaching and learning of this problem in science classrooms is occurring (e.g. Arslan, Cigdemoglu, & Moseley, 2012; Svihla & Linn, 2012). However, even though people experience a range of emotions about global problems like climate change (Hicks & Holden, 2007; Ojala, 2012; Rickinson, 2001), little attention is given to their emotions about the problem in science classrooms. Because emotions are evaluative (Boler, 1999; Keltner & Gross, 1999), they provided a lens for understanding how students engage personally with climate change. In this study, I drew from sociolinguistics, social psychology, and the sociology of emotions to examine a) the social interactions that allowed for emotional expressions to be constructed and b) the ways in which pre-service elementary teachers constructed emotional expressions about climate change in a science course. Three overall findings emerged: 1) emotions provided a means of understanding how students' conceptualized climate to be relevant to their lives, 2) emotional expressions and the aboutness of these expressions indicated that the students conceptualized climate change as distanced, both temporally and spatially, and 3) although most emotional constructions were distanced, there were multiple instances of emotional expressions in which students took climate change personally. Following a discussion of the findings, implications, limitations, and directions for future research are also described.
Whitehouse, Peter J; Waller, Sara
Understanding why we produce labels for neuropsychiatric conditions, such as Alzheimer's disease (AD), and how we use those words to tell stories about our brain, as well as which groups control such diagnostic discourse, is important to a wise understanding of our cognitive abilities, their limitations, and even our very human nature. Here, we explore the history and current focus of a newly emerging field called neuroethics and explore its relationship (or lack thereof) to a newly created clinical syndrome called involuntary emotional expressive disorder (IEED). The main argument concerns the lack of neuroethical discussion of issues pertinent to social influences on disease and the construction of professional specialization. We are critical of the processes associated with the creation of both the field and the syndrome, and express concern about their eventual outcomes. The interaction of social, political, and business institutions, the inherent interests of the advancement of larger research projects (and the individuals that compose them), their potential for profit, and other incentives to enhance marketability and public attention toward certain research programs will be examined as we discuss the development of the field of neuroethics. Similarly, we argue that these social factors and forces are instrumental in the development of IEED as a recognizable category and condition. Our critique is guided by the hope that through such analyses we can improve our understanding of how we go about our academic activities in cognitive neuroscience and also improve our efforts to help people suffering from neuropsychiatric conditions, such as dementia.
Green, S.; Baker, B.
Background: Parents' expression of positive emotion towards children who are typically developing (TD) is generally associated with better social development. However, the association between parents' negative emotion expression and social development can be positive or negative depending upon a number of factors, including the child's emotion…
Widen, Sherri C.; Russell, James A.
Past research has shown that children recognize emotions from facial expressions poorly and improve only gradually with age, but the stimuli in such studies have been static faces. Because dynamic faces include more information, it may well be that children more readily recognize emotions from dynamic facial expressions. The current study of…
Chamcharatsri, Pisarn Bee
Writing to express emotions can be a challenging task for second language (L2) writers, especially because it tends to be a process that is less addressed in language classrooms. This paper aims to expand thinking on L2 literacy and writing by exploring how L2 writers can express emotion (fear) through narratives both in their first language (L1)…
Kheirzadeh, Shiela; Hajiabed, Mohammadreza
The present interdisciplinary research investigates the differential emotional expression between Persian monolinguals and Persian-English bilinguals. In other words, the article was an attempt to answer the questions whether bilinguals and monolinguals differ in the expression of positive and negative emotions elicited through sad and happy…
Truong, K.P.; Leeuwen, D.A. van; Neerincx, M.A.
Two unobtrusive modalities for automatic emotion recognition are discussed: speech and facial expressions. First, an overview is given of emotion recognition studies based on a combination of speech and facial expressions. We will identify difficulties concerning data collection, data fusion, system
Froyen, Laura C.; Skibbe, Lori E.; Bowles, Ryan P.; Blow, Adrian J.; Gerde, Hope K.
The current study investigates associations among marital satisfaction, family emotional expressiveness, the home learning environment, and preschool-aged children's emergent literacy skills among 385 Midwestern mothers and their children. Path analyses examined how marital satisfaction related to emotional expressiveness in the home and whether…
Uchida, Yukiko; Townsend, Sarah S M; Rose Markus, Hazel; Bergsieker, Hilary B
Four studies using open-ended and experimental methods test the hypothesis that in Japanese contexts, emotions are understood as between people, whereas in American contexts, emotions are understood as primarily within people. Study 1 analyzed television interviews of Olympic athletes. When asked about their relationships, Japanese athletes used significantly more emotion words than American athletes. This difference was not significant when questions asked directly about athletes' feelings. In Study 2, when describing an athlete's emotional reaction to winning, Japanese participants implicated others more often than American participants. After reading an athlete's self-description, Japanese participants inferred more emotions when the athlete mentioned relationships, whereas American participants inferred more emotions when the athlete focused only on herself (Study 3). Finally, when viewing images of athletes, Japanese participants inferred more emotions for athletes pictured with teammates, whereas American participants inferred more emotions for athletes pictured alone (Studies 4a and 4b).
Cebula, Katie R.; Wishart, Jennifer G.; Willis, Diane S.; Pitcairn, Tom K.
Some children with Down syndrome may experience difficulties in recognising facial emotions, particularly fear, but it is not clear why, nor how such skills can best be facilitated. Using a photo-matching task, emotion recognition was tested in children with Down syndrome, children with non-specific intellectual disabilities and cognitively-matched typically-developing children (all groups N = 21) under four conditions: veridical vs exaggerated emotions and emotion-labelling vs generic task i...
Kelly A. Gola
Full Text Available Intentional facial expression of emotion is critical to healthy social interactions. Patients with neurodegenerative disease, particularly those with right temporal or prefrontal atrophy, show dramatic socioemotional impairment. This was an exploratory study examining the neural and behavioral correlates of intentional facial expression of emotion in neurodegenerative disease patients and healthy controls. One hundred and thirty three participants (45 Alzheimer's disease, 16 behavioral variant frontotemporal dementia, 8 non-fluent primary progressive aphasia, 10 progressive supranuclear palsy, 11 right-temporal frontotemporal dementia, 9 semantic variant primary progressive aphasia patients and 34 healthy controls were video recorded while imitating static images of emotional faces and producing emotional expressions based on verbal command; the accuracy of their expression was rated by blinded raters. Participants also underwent face-to-face socioemotional testing and informants described participants' typical socioemotional behavior. Patients' performance on emotion expression tasks was correlated with gray matter volume using voxel-based morphometry (VBM across the entire sample. We found that intentional emotional imitation scores were related to fundamental socioemotional deficits; patients with known socioemotional deficits performed worse than controls on intentional emotion imitation; and intentional emotional expression predicted caregiver ratings of empathy and interpersonal warmth. Whole brain VBMs revealed a rightward cortical atrophy pattern homologous to the left lateralized speech production network was associated with intentional emotional imitation deficits. Results point to a possible neural mechanisms underlying complex socioemotional communication deficits in neurodegenerative disease patients.
Gola, Kelly A; Shany-Ur, Tal; Pressman, Peter; Sulman, Isa; Galeana, Eduardo; Paulsen, Hillary; Nguyen, Lauren; Wu, Teresa; Adhimoolam, Babu; Poorzand, Pardis; Miller, Bruce L; Rankin, Katherine P
Intentional facial expression of emotion is critical to healthy social interactions. Patients with neurodegenerative disease, particularly those with right temporal or prefrontal atrophy, show dramatic socioemotional impairment. This was an exploratory study examining the neural and behavioral correlates of intentional facial expression of emotion in neurodegenerative disease patients and healthy controls. One hundred and thirty three participants (45 Alzheimer's disease, 16 behavioral variant frontotemporal dementia, 8 non-fluent primary progressive aphasia, 10 progressive supranuclear palsy, 11 right-temporal frontotemporal dementia, 9 semantic variant primary progressive aphasia patients and 34 healthy controls) were video recorded while imitating static images of emotional faces and producing emotional expressions based on verbal command; the accuracy of their expression was rated by blinded raters. Participants also underwent face-to-face socioemotional testing and informants described participants' typical socioemotional behavior. Patients' performance on emotion expression tasks was correlated with gray matter volume using voxel-based morphometry (VBM) across the entire sample. We found that intentional emotional imitation scores were related to fundamental socioemotional deficits; patients with known socioemotional deficits performed worse than controls on intentional emotion imitation; and intentional emotional expression predicted caregiver ratings of empathy and interpersonal warmth. Whole brain VBMs revealed a rightward cortical atrophy pattern homologous to the left lateralized speech production network was associated with intentional emotional imitation deficits. Results point to a possible neural mechanisms underlying complex socioemotional communication deficits in neurodegenerative disease patients.
This paper deals with the automatic identification of emotions from the manual annotations of the shape and functions of facial expressions in a Danish corpus of video recorded naturally occurring first encounters. More specifically, a support vector classified is trained on the corpus annotations...... to identify emotions in facial expressions. In the classification experiments, we test to what extent emotions expressed in naturally-occurring conversations can be identified automatically by a classifier trained on the manual annotations of the shape of facial expressions and co-occurring speech tokens. We...... also investigate the relation between emotions and the communicative functions of facial expressions. Both emotion labels and their values in a three dimensional space are identified. The three dimensions are Pleasure, Arousal and Dominance. The results of our experiments indicate that the classifiers...
Murray, John C.; Cañamero, Lola; Bard, Kim A.; Ross, Marina Davila; Thorsteinsson, Kate
In this paper we focus primarily on the influence that socio-emotional interaction has on the perception of emotional expression by a robot. We also investigate and discuss the importance of emotion expression in socially interactive situations involving human robot interaction (HRI), and show the importance of utilising emotion expression when dealing with interactive robots, that are to learn and develop in socially situated environments. We discuss early expressional development and the function of emotion in communication in humans and how this can improve HRI communications. Finally we provide experimental results showing how emotion-rich interaction via emotion expression can affect the HRI process by providing additional information.
Bebko, Genna M; Franconeri, Steven L; Ochsner, Kevin N; Chiao, Joan Y
According to appraisal theories of emotion, cognitive reappraisal is a successful emotion regulation strategy because it involves cognitively changing our thoughts, which, in turn, change our emotions. However, recent evidence has challenged the importance of cognitive change and, instead, has suggested that attentional deployment may at least partly explain the emotion regulation success of cognitive reappraisal. The purpose of the current study was to examine the causal relationship between attentional deployment and emotion regulation success. We examined 2 commonly used emotion regulation strategies--cognitive reappraisal and expressive suppression-because both depend on attention but have divergent behavioral, experiential, and physiological outcomes. Participants were either instructed to regulate emotions during free-viewing (unrestricted image viewing) or gaze-controlled (restricted image viewing) conditions and to self-report negative emotional experience. For both emotion regulation strategies, emotion regulation success was not altered by changes in participant control over the (a) direction of attention (free-viewing vs. gaze-controlled) during image viewing and (b) valence (negative vs. neutral) of visual stimuli viewed when gaze was controlled. Taken together, these findings provide convergent evidence that attentional deployment does not alter subjective negative emotional experience during either cognitive reappraisal or expressive suppression, suggesting that strategy-specific processes, such as cognitive appraisal and response modulation, respectively, may have a greater impact on emotional regulation success than processes common to both strategies, such as attention.
Lawrence Ian Reed
Full Text Available What is the function of emotional tearing? Previous work has found a tear effect, which resolves ambiguity in neutral expressions and increases perceptions of sadness in sad expressions. Tearing, however, is associated with a variety of emotional states, and it remains unclear how the tear effect generalizes to other emotion expressions. Here we expand upon previous works by examining ratings of video clips depicting posed facial expressions presented with and without tears. We replicate Provine et al.’s (2009 findings that tearing increases perceptions of sadness in sad expressions. Furthermore, we find that tearing has specific effects on ratings of emotion (happiness, sadness, anger, and fear and ratings of intensity and valence in neutral, positive, and negative expressions. These results suggest that tearing may serve a specific and independent communicative function, interacting with those of various expressions.
Chervonsky, Elizabeth; Hunt, Caroline
Emotion expression is critical for the communication of important social information, such as emotional states and behavioral intentions. However, people tend to vary in their level of emotional expression. This meta-analysis investigated the relationships between levels of emotion expression and suppression, and social and interpersonal outcomes. PsycINFO databases, as well as reference lists were searched. Forty-three papers from a total of 3,200 papers met inclusion criteria, allowing for 105 effect sizes to be calculated. Meta-analyses revealed that greater suppression of emotion was significantly associated with poorer social wellbeing, including more negative first impressions, lower social support, lower social satisfaction and quality, and poorer romantic relationship quality. Furthermore, the expression of positive and general/nonspecific emotion was related to better social outcomes, while the expression of anger was associated with poorer social wellbeing. Expression of negative emotion generally was also associated with poorer social outcomes, although this effect size was very small and consisted of mixed results. These findings highlight the importance of considering the role that regulation of emotional expression can play in the development of social dysfunction and interpersonal problems. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Niles, Andrea N; Haltom, Kate E Byrne; Mulvenna, Catherine M; Lieberman, Matthew D; Stanton, Annette L
The current study assessed main effects and moderators (including emotional expressiveness, emotional processing, and ambivalence over emotional expression) of the effects of expressive writing in a sample of healthy adults. Young adult participants (N=116) were randomly assigned to write for 20 minutes on four occasions about deepest thoughts and feelings regarding their most stressful/traumatic event in the past five years (expressive writing) or about a control topic (control). Dependent variables were indicators of anxiety, depression, and physical symptoms. No significant effects of writing condition were evident on anxiety, depressive symptoms, or physical symptoms. Emotional expressiveness emerged as a significant moderator of anxiety outcomes, however. Within the expressive writing group, participants high in expressiveness evidenced a significant reduction in anxiety at three-month follow-up, and participants low in expressiveness showed a significant increase in anxiety. Expressiveness did not predict change in anxiety in the control group. These findings on anxiety are consistent with the matching hypothesis, which suggests that matching a person's naturally elected coping approach with an assigned intervention is beneficial. These findings also suggest that expressive writing about a stressful event may be contraindicated for individuals who do not typically express emotions.
Hareli, Shlomo; David, Shlomo
Social perception of emotions is influenced by the context in which it occurs. One such context is a social interaction involving an exchange of emotions. The way parties to the interaction are perceived is shaped by the combination of emotions exchanged. This idea was examined by assessing the extent to which expressions of anger toward a target-which, in isolation, are perceived as signals of high social power-are influenced by the target's emotional reaction to it (i.e., reactive emotions). Three studies show that the angry person was perceived as having a higher level of social power when this anger was responded by fear or sadness than when it was responded by neutrality or anger. Study 1 indicated that reactive emotions have a stronger effect on perceived social power when emotions were incongruent with gender stereotypes. Study 2 indicated that these effects are a result of these emotions serving as reactive emotions rather than a benchmark against which the angry person's power is assessed. Study 3 showed that reactive emotions affect perceived social power by serving as signals of the level to which the high social power suggested by the first person's expression is confirmed by its target. Comparing effects of reactive emotions to anger with reactive emotions to sadness, showed that perceived social power of the expresser is determined by the nature of the expression, with some adjustment caused by the reactive emotions. This underscores the importance of social interaction as a context for the social perception of emotions. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Khan, Masood Mehmood; Ward, Robert D.; Ingleby, Michael
The ability to distinguish feigned from involuntary expressions of emotions could help in the investigation and treatment of neuropsychiatric and affective disorders and in the detection of malingering. This work investigates differences in emotion-specific patterns of thermal variations along the major facial muscles. Using experimental data extracted from 156 images, we attempted to classify patterns of emotion-specific thermal variations into neutral, and voluntary and involuntary expressions of positive and negative emotive states. Initial results suggest (i) each facial muscle exhibits a unique thermal response to various emotive states; (ii) the pattern of thermal variances along the facial muscles may assist in classifying voluntary and involuntary facial expressions; and (iii) facial skin temperature measurements along the major facial muscles may be used in automated emotion assessment.
Cebula, Katie R; Wishart, Jennifer G; Willis, Diane S; Pitcairn, Tom K
Some children with Down syndrome may experience difficulties in recognizing facial emotions, particularly fear, but it is not clear why, nor how such skills can best be facilitated. Using a photo-matching task, emotion recognition was tested in children with Down syndrome, children with nonspecific intellectual disability and cognitively matched, typically developing children (all groups N = 21) under four conditions: veridical vs. exaggerated emotions and emotion-labelling vs. generic task instructions. In all groups, exaggerating emotions facilitated recognition accuracy and speed, with emotion labelling facilitating recognition accuracy. Overall accuracy and speed did not differ in the children with Down syndrome, although recognition of fear was poorer than in the typically developing children and unrelated to emotion label use. Implications for interventions are considered.
Helga Cristina Santos Sartori
Full Text Available CONTEXTO: O transtorno da expressão emocional involuntária (involuntary emotional expression disorder ou IEED consiste em um transtorno do afeto, caracterizado por uma dificuldade em controlar a expressão emocional, que se apresenta por episódios breves e estereotipados de riso e/ou choro incontroláveis. Pode estar relacionado a diversas patologias encefálicas, em variadas localizações anatômicas. OBJETIVOS: Revisar aspectos clínicos, epidemiológicos e fisiopatológicos envolvidos no transtorno da expressão emocional involuntária e apresentar as opções atuais e futuras na abordagem terapêutica. MÉTODOS: Pesquisa de base de dados MEDLINE/PUBMED e LILACS utilizando os termos transtorno da expressão emocional involuntária, afeto pseudobulbar, riso e choro patológicos, acidente vascular cerebral, doença de Alzheimer, esclerose múltipla, esclerose lateral amiotrófica. RESULTADOS: No trantorno da expressão emocional involuntária, as crises de choro e/ou riso, além de serem incontroláveis, tendem a ser desproporcionais ao estímulo recebido, podendo estar completamente dissociada do estado de humor do paciente ou mesmo ser contraditória ao contexto no qual o estímulo está inserido. Outros termos são usados na nosografia desse transtorno, como afeto pseudobulbar, riso e choro patológicos, labilidade emocional, emocionalismo e desregulação emocional. Termos como choro forçado, choro involuntário, emocionalidade patológica e incontinência emocional também têm sido utilizados com menor freqüência. Os mecanismos fisiopatológicos específicos envolvidos nesse transtorno ainda não estão bem esclarecidos. Lesões que podem causá-lo estão amplamente distribuídas no encéfalo, mas parecem envolver o lobo frontal, o sistema límbico, o tronco cerebral e o cerebelo, assim como a substância branca que interconecta essa rede. Seu principal diagnóstico diferencial é a depressão. As terapias farmacológicas hoje
Escoffier, Nicolas; Zhong, Jidan; Schirmer, Annett; Qiu, Anqi
Scholars have documented similarities in the way voice and music convey emotions. By using functional magnetic resonance imaging (fMRI) we explored whether these similarities imply overlapping processing substrates. We asked participants to trace changes in either the emotion or pitch of vocalizations and music using a joystick. Compared to music, vocalizations more strongly activated superior and middle temporal cortex, cuneus, and precuneus. However, despite these differences, overlapping rather than differing regions emerged when comparing emotion with pitch tracing for music and vocalizations, respectively. Relative to pitch tracing, emotion tracing activated medial superior frontal and anterior cingulate cortex regardless of stimulus type. Additionally, we observed emotion specific effects in primary and secondary auditory cortex as well as in medial frontal cortex that were comparable for voice and music. Together these results indicate that similar mechanisms support emotional inferences from vocalizations and music and that these mechanisms tap on a general system involved in social cognition. Copyright © 2011 Wiley Periodicals, Inc.
Chen, Stephen H; Kennedy, Morgan; Zhou, Qing
Parents regularly use words to express and discuss emotion with their children, but does it matter which language they use to do so? In this article, we examine this question in the multilingual family context by integrating findings from both psychological and linguistic research. We propose that parents' use of different languages for emotional expression or discussion holds significant implications for children's emotional experience, understanding, and regulation. Finally, we suggest that an understanding of the implications of emotion-related language shifts is critical, particularly in adapting interventions within a rapidly diversifying society. © The Author(s) 2012.
Widen, Sherri C; Christy, Anita M; Hewett, Kristen; Russell, James A
Shame, embarrassment, compassion, and contempt have been considered candidates for the status of basic emotions on the grounds that each has a recognisable facial expression. In two studies (N=88, N=60) on recognition of these four facial expressions, observers showed moderate agreement on the predicted emotion when assessed with forced choice (58%; 42%), but low agreement when assessed with free labelling (18%; 16%). Thus, even though some observers endorsed the predicted emotion when it was presented in a list, over 80% spontaneously interpreted these faces in a way other than the predicted emotion.
Long, Kristin A.; Lobato, Debra; Kao, Barbara; Plante, Wendy; Grullón, Edicta; Cheas, Lydia; Houck, Christopher; Seifer, Ronald
Objective Examine general emotion expression and sibling–parent emotion communication among Latino and non-Latino white (NLW) siblings of children with intellectual disabilities (ID) and matched comparisons. Methods 200 siblings (ages 8–15 years) completed the newly developed Sibling–Parent Emotion Communication Scale and existing measures of general emotion expression and psychosocial functioning. Preliminary analyses evaluated scale psychometrics across ethnicity. Results Structure and inte...
Safar, Kristina; Moulson, Margaret C
Infants may recognize facial expressions of emotion more readily when familiar faces express the emotions. Studies 1 and 2 investigated whether familiarity influences two metrics of emotion processing: Categorization and spontaneous preference. In Study 1 (n = 32), we replicated previous findings showing an asymmetrical pattern of categorization of happy and fearful faces in 6.5-month-old infants, and extended these findings by demonstrating that infants' categorization did not differ when emotions were expressed by familiar (i.e., caregiver) faces. In Study 2 (n = 34), we replicated the spontaneous preference for fearful over happy expressions in 6.5-month-old infants, and extended these findings by demonstrating that the spontaneous preference for fear was also present for familiar faces. Thus, infants' performance on two metrics of emotion processing did not differ depending on face familiarity. © 2017 Wiley Periodicals, Inc.
Chaplin, Tara M; Aldao, Amelia
Emotion expression is an important feature of healthy child development that has been found to show gender differences. However, there has been no empirical review of the literature on gender and facial, vocal, and behavioral expressions of different types of emotions in children. The present study constitutes a comprehensive meta-analytic review of gender differences and moderators of differences in emotion expression from infancy through adolescence. We analyzed 555 effect sizes from 166 studies with a total of 21,709 participants. Significant but very small gender differences were found overall, with girls showing more positive emotions (g = -.08) and internalizing emotions (e.g., sadness, anxiety, sympathy; g = -.10) than boys, and boys showing more externalizing emotions (e.g., anger; g = .09) than girls. Notably, gender differences were moderated by age, interpersonal context, and task valence, underscoring the importance of contextual factors in gender differences. Gender differences in positive emotions were more pronounced with increasing age, with girls showing more positive emotions than boys in middle childhood (g = -.20) and adolescence (g = -.28). Boys showed more externalizing emotions than girls at toddler/preschool age (g = .17) and middle childhood (g = .13) and fewer externalizing emotions than girls in adolescence (g = -.27). Gender differences were less pronounced with parents and were more pronounced with unfamiliar adults (for positive emotions) and with peers/when alone (for externalizing emotions). Our findings of gender differences in emotion expression in specific contexts have important implications for gender differences in children's healthy and maladaptive development. 2013 APA, all rights reserved
Chaplin, Tara M.; Aldao, Amelia
Emotion expression is an important feature of healthy child development that has been found to show gender differences. However, there has been no empirical review of the literature on gender and facial, vocal, and behavioral expressions of different types of emotions in children. The present study constitutes a comprehensive meta-analytic review of gender differences, and moderators of differences, in emotion expression from infancy through adolescence. We analyzed 555 effect sizes from 166 studies with a total of 21,709 participants. Significant, but very small, gender differences were found overall, with girls showing more positive emotions (g = −.08) and internalizing emotions (e.g., sadness, anxiety, sympathy; g = −.10) than boys, and boys showing more externalizing emotions (e.g., anger; g = .09) than girls. Notably, gender differences were moderated by age, interpersonal context, and task valence, underscoring the importance of contextual factors in gender differences. Gender differences in positive emotions were more pronounced with increasing age, with girls showing more positive emotions than boys in middle childhood (g = −.20) and adolescence (g = −.28). Boys showed more externalizing emotions than girls at toddler/preschool age (g = .17) and middle childhood (g = .13) and fewer externalizing emotions than girls in adolescence (g = −.27). Gender differences were less pronounced with parents and were more pronounced with unfamiliar adults (for positive emotions) and with peers/when alone (for externalizing emotions). Our findings of gender differences in emotion expression in specific contexts have important implications for gender differences in children’s healthy and maladaptive development. PMID:23231534
Eack, Shaun M; Mazefsky, Carla A; Minshew, Nancy J
Facial emotion perception is significantly affected in autism spectrum disorder, yet little is known about how individuals with autism spectrum disorder misinterpret facial expressions that result in their difficulty in accurately recognizing emotion in faces. This study examined facial emotion perception in 45 verbal adults with autism spectrum disorder and 30 age- and gender-matched volunteers without autism spectrum disorder to identify patterns of emotion misinterpretation during face processing that contribute to emotion recognition impairments in autism. Results revealed that difficulty distinguishing emotional from neutral facial expressions characterized much of the emotion perception impairments exhibited by participants with autism spectrum disorder. In particular, adults with autism spectrum disorder uniquely misinterpreted happy faces as neutral, and were significantly more likely than typical volunteers to attribute negative valence to nonemotional faces. The over-attribution of emotions to neutral faces was significantly related to greater communication and emotional intelligence impairments in individuals with autism spectrum disorder. These findings suggest a potential negative bias toward the interpretation of facial expressions and may have implications for interventions designed to remediate emotion perception in autism spectrum disorder. © The Author(s) 2014.
Wingenbach, Tanja S. H.; Ashwin, Chris; Brosnan, Mark
Most of the existing sets of facial expressions of emotion contain static photographs. While increasing demand for stimuli with enhanced ecological validity in facial emotion recognition research has led to the development of video stimuli, these typically involve full-blown (apex) expressions. However, variations of intensity in emotional facial expressions occur in real life social interactions, with low intensity expressions of emotions frequently occurring. The current study therefore dev...
Full Text Available There is a growing appreciation that individuals differ systematically in their use of particular emotion regulation strategies. Our aim was to examine the structural correlates of the habitual use of expressive suppression of emotions. Based on our previous research on the voluntary suppression of actions we expected this response-focused emotion regulation strategy to be associated with increased grey matter volume in the dorsomedial prefrontal cortex (dmPFC. On high-resolution MRI scans of 42 college-aged healthy adults we computed optimized voxel-based-morphometry (VBM to explore the correlation between grey matter volume and inter-individual differences in the tendency to suppress the expression of emotions assessed by means of the Emotion Regulation Questionnaire (Gross & John, 2003. We found a positive correlation between the habitual use of expressive suppression as an emotion regulation strategy and grey matter volume in the dmPFC. No other brain area showed a significant positive or negative correlation with the Emotion Regulation Questionnaire scores. The association between the suppression of expression of emotions and volume in the dmPFC supports the behavioural stability and biological foundation of the concept of this particular emotion regulation strategy within an age-homogenous sample of adults.
Kühn, Simone; Gallinat, Jürgen; Brass, Marcel
There is a growing appreciation that individuals differ systematically in their use of particular emotion regulation strategies. Our aim was to examine the structural correlates of the habitual use of expressive suppression of emotions. Based on our previous research on the voluntary suppression of actions we expected this response-focused emotion regulation strategy to be associated with increased grey matter volume in the dorsomedial prefrontal cortex (dmPFC). On high-resolution MRI scans of 42 college-aged healthy adults we computed optimized voxel-based-morphometry (VBM) to explore the correlation between grey matter volume and inter-individual differences in the tendency to suppress the expression of emotions assessed by means of the Emotion Regulation Questionnaire (Gross & John, 2003). We found a positive correlation between the habitual use of expressive suppression as an emotion regulation strategy and grey matter volume in the dmPFC. No other brain area showed a significant positive or negative correlation with the Emotion Regulation Questionnaire scores. The association between the suppression of expression of emotions and volume in the dmPFC supports the behavioural stability and biological foundation of the concept of this particular emotion regulation strategy within an age-homogenous sample of adults.
Watling, Dawn; Bourne, Victoria
Understanding of emotions has been shown to develop between the ages of 4 and 10 years; however, individual differences exist in this development. While previous research has typically examined these differences in terms of developmental and/or social factors, little research has considered the possible impact of neuropsychological development on the behavioural understanding of emotions. Emotion processing tends to be lateralised to the right hemisphere of the brain in adults, yet this patt...
Full Text Available Facial expressions are one of the most important means of nonverbal communication transporting both emotional and conversational content. For investigating this large space of expressions we recently developed a large database containing dynamic emotional and conversational expressions in Germany (MPI facial expression database. As facial expressions crucially depend on the cultural context, however, a similar resource is needed for studies outside of Germany. Here, we introduce and validate a new, extensive Korean facial expression database containing dynamic emotional and conversational information. Ten individuals performed 62 expressions following a method-acting protocol, in which each person was asked to imagine themselves in one of 62 corresponding everyday scenarios and to react accordingly. To validate this database, we conducted two experiments: 20 participants were asked to name the appropriate expression for each of the 62 everyday scenarios shown as text. Ten additional participants were asked to name each of the 62 expression videos from 10 actors in addition to rating its naturalness. All naming answers were then rated as valid or invalid. Scenario validation yielded 89% valid answers showing that the scenarios are effective in eliciting appropriate expressions. Video sequences were judged as natural with an average of 66% valid answers. This is an excellent result considering that videos were seen without any conversational context and that 62 expressions were to be recognized. These results validate our Korean database and, as they also parallel the German validation results, will enable detailed cross-cultural comparisons of the complex space of emotional and conversational expressions.
Leppanen, J; Cardi, V; Ng, K W; Paloyelis, Y; Stein, D; Tchanturia, K; Treasure, J
Altered social-emotional functioning is considered to play an important role in the development and maintenance of anorexia nervosa (AN). Recently, there has been increasing interest in investigating the role of intranasal oxytocin in social-emotional processing. The present study aimed to investigate the effects of intranasal oxytocin on the interpretation and expression of emotions among people with AN. Thirty women with AN and 29 age-matched healthy women took part in the present study, which used a double-blind, placebo-controlled, cross-over design. The participants received a single dose of 40 IU of intranasal oxytocin in one session and a placebo spray in the other. Fifteen minutes after administration, the participants completed the Reading the Mind in the Eyes Test to assess the interpretation of complex emotions and mental states followed by a video task, which assessed expressions of facial affect when they were viewing humorous and sad film clips. The intranasal oxytocin did not significantly influence the expression or interpretation of emotions in the AN or healthy comparison groups. The AN group expressed significantly less positive emotion, spent more time looking away and reported experiencing a significantly more negative affect in response to the film clips. The finding that intranasal oxytocin had little to no effect on the interpretation or expression of emotions in either group supports the notion that the effects of oxytocin on social-emotional processing are not straightforward and may depend on individual and environmental differences, as well as the emotion being processed. Replication of these findings is necessary to explore the effect of timing on the effects of oxytocin before firm conclusions can be drawn. Nonetheless, these findings add to the steady accumulation of evidence that people with AN have reduced emotional expression and avoidance of emotionally provoking stimuli. © 2017 The Authors. Journal of Neuroendocrinology
Milojevich, Helen M; Haskett, Mary E
The present study took a developmental psychopathology approach to examine the longitudinal association between parents' emotional expressiveness and children's self-regulation. Data collection spanned from 2004 to 2008. Ninety-two physically abusive parents completed yearly assessments of their emotional expressiveness, as well as their children's self-regulation abilities. Observational and behavioral measures were also obtained yearly to capture both parents' emotional expressiveness and children's self-regulation. Specifically, parents participated in a parent-child interaction task, which provided insight into their levels of flat affect. A puzzle box task was completed by each child to assess self-regulation. Results indicated, first, that greater parental expression of negative emotions predicted poorer self-regulation in children, both concurrently and across time. Second, parental expressions of positive emotions and parents' flat affect were unrelated to children's self-regulation. Findings inform our understanding of parental socialization of self-regulation and provide insight into the roles of distinct components of emotional expressiveness. Moreover, findings have crucial implications for understanding emotional expressiveness in high-risk samples and increase our understanding of within-group functioning among maltreating families that may serve as a means to direct intervention efforts. Copyright © 2018 Elsevier Ltd. All rights reserved.
Horowitz, Mardi J.; And Others
Single-case study examined frank disclosure of important topics in brief exploratory psychotherapy, including topics closely related to recent, unintegrated stressor life event. Quantitative measures of emotion and control variables showed heightened levels of both emotionality and defensive control during discourse on topic of stressor event.…
Full Text Available Altered facial expressions of emotions are characteristic impairments in schizophrenia. Ratings of affect have traditionally been limited to clinical rating scales and facial muscle movement analysis, which require extensive training and have limitations based on methodology and ecological validity. To improve reliable assessment of dynamic facial expression changes, we have developed automated measurements of facial emotion expressions based on information-theoretic measures of expressivity of ambiguity and distinctiveness of facial expressions. These measures were examined in matched groups of persons with schizophrenia (n=28 and healthy controls (n=26 who underwent video acquisition to assess expressivity of basic emotions (happiness, sadness, anger, fear, and disgust in evoked conditions. Persons with schizophrenia scored higher on ambiguity, the measure of conditional entropy within the expression of a single emotion, and they scored lower on distinctiveness, the measure of mutual information across expressions of different emotions. The automated measures compared favorably with observer-based ratings. This method can be applied for delineating dynamic emotional expressivity in healthy and clinical populations.
Gendron, Maria; Roberson, Debi; van der Vyver, Jacoba Marietta; Barrett, Lisa Feldman
It is widely believed that certain emotions are universally recognized in facial expressions. Recent evidence indicates that Western perceptions (e.g., scowls as anger) depend on cues to US emotion concepts embedded in experiments. Since such cues are standard feature in methods used in cross-cultural experiments, we hypothesized that evidence of universality depends on this conceptual context. In our study, participants from the US and the Himba ethnic group sorted images of posed facial expressions into piles by emotion type. Without cues to emotion concepts, Himba participants did not show the presumed “universal” pattern, whereas US participants produced a pattern with presumed universal features. With cues to emotion concepts, participants in both cultures produced sorts that were closer to the presumed “universal” pattern, although substantial cultural variation persisted. Our findings indicate that perceptions of emotion are not universal, but depend on cultural and conceptual contexts. PMID:24708506
Gendron, Maria; Roberson, Debi; van der Vyver, Jacoba Marietta; Barrett, Lisa Feldman
It is widely believed that certain emotions are universally recognized in facial expressions. Recent evidence indicates that Western perceptions (e.g., scowls as anger) depend on cues to U.S. emotion concepts embedded in experiments. Because such cues are standard features in methods used in cross-cultural experiments, we hypothesized that evidence of universality depends on this conceptual context. In our study, participants from the United States and the Himba ethnic group from the Keunene region of northwestern Namibia sorted images of posed facial expressions into piles by emotion type. Without cues to emotion concepts, Himba participants did not show the presumed "universal" pattern, whereas U.S. participants produced a pattern with presumed universal features. With cues to emotion concepts, participants in both cultures produced sorts that were closer to the presumed "universal" pattern, although substantial cultural variation persisted. Our findings indicate that perceptions of emotion are not universal, but depend on cultural and conceptual contexts.
Goldblatt, Hadass; Cohen, Miri; Azaiza, Faisal
Researchers have suggested that older adults express less negative emotions. Yet, emotional expression patterns in older and younger breast cancer survivors, have barely been examined. This study aimed to explore types and intensity of negative and positive emotional expression related to the breast cancer experience by younger and older Arab breast cancer survivors. Participants were 20 younger (aged 32-50) and 20 older (aged 51-75) Muslim and Christian Arab breast cancer survivors (stages I-III), currently free of disease. Data were gathered through in-depth semi-structured interviews. Mixed methods analyses were conducted, including: (1) frequency analysis of participants' emotional expressions; (2) content analysis of emotional expressions, categorized according to negative and positive emotions. Three emotional expression modalities were revealed: (1) Succinct versus comprehensive accounts; (2) expression of emotions versus avoidance of emotions; (3) patterns of expression of positive emotions and a sense of personal growth. Younger women provided more detailed accounts about their illness experiences than older women. Older women's accounts were succinct, action-focused, and included more emotion-avoiding expressions than younger women. Understanding the relationships between emotional expression, emotional experience, and cancer survivors' quality of life, specifically of those from traditional communities, is necessary for developing effective psycho-social interventions.
da Silva, Flávio Altinier Maximiano; Pedrini, Helio
Facial expressions are an important demonstration of humanity's humors and emotions. Algorithms capable of recognizing facial expressions and associating them with emotions were developed and employed to compare the expressions that different cultural groups use to show their emotions. Static pictures of predominantly occidental and oriental subjects from public datasets were used to train machine learning algorithms, whereas local binary patterns, histogram of oriented gradients (HOGs), and Gabor filters were employed to describe the facial expressions for six different basic emotions. The most consistent combination, formed by the association of HOG filter and support vector machines, was then used to classify the other cultural group: there was a strong drop in accuracy, meaning that the subtle differences of facial expressions of each culture affected the classifier performance. Finally, a classifier was trained with images from both occidental and oriental subjects and its accuracy was higher on multicultural data, evidencing the need of a multicultural training set to build an efficient classifier.
Lipp, Marilda E Novaes; Pereira, Márcia M Bignotto; Justo, Ana Paula; de Matos, Thania M Gomes
This study investigated cardiovascular reactivity of hypertensive adults during periods of emotional stress. Two types of instructions were given at different moments, to the same subject, either to express or to suppress feelings during role-play. Expressing, but not inhibiting, emotions elicited significantly higher reactivity during responding to negative scenes, followed by responding during the positive interactions. Blood pressure increases in both expressing and inhibiting conditions, were also found during the instruction periods. Results indicated that socially demanding situations represent a stressor whose effects may vary depending on whether or not respondents regulate expression of emotions. It is suggested that the difficulty in expressing emotions found in some hypertensive individuals may have the function of controlling or reducing blood pressure reactivity.
Alexander, Stewart C; Pollak, Kathryn I; Morgan, Perri A; Strand, Justine; Abernethy, Amy P; Jeffreys, Amy S; Arnold, Robert M; Olsen, Maren; Rodriguez, Keri L; Garrigues, Sarah K; Manusov, Justin R E; Tulsky, James A
Patients with advanced cancer often experience negative emotion; clinicians' empathic responses can alleviate patient distress. Much is known about how physicians respond to patient emotion; less is known about non-physician clinicians. Given that oncology care is increasingly provided by an interdisciplinary team, it is important to know more about how patients with advanced cancer express emotions to non-physician clinicians (NPCs) and how NPCs respond to those empathic opportunities. We audio recorded conversations between non-physician clinicians and patients with advanced cancer. We analyzed 45 conversations between patients and oncology physician assistants, nurse practitioners, and nurse clinicians in which patients or their loved ones expressed at least one negative emotion to the NPC (i.e., an empathic opportunity). Empathic opportunities were coded three ways: type of emotion (anger, sadness, or fear), severity of emotion (least, moderate, or most severe), and NPC response to emotion (not empathic, on-topic medical response, and empathic response). We identified 103 empathic opportunities presented to 25 different NPCs during 45 visits. Approximately half of the empathic opportunities contained anger (53%), followed by sadness (25%) and fear (21%). The majority of emotions expressed were moderately severe (73%), followed by most severe (16%), and least severe (12%). The severity of emotions presented was not found to be statistically different between types of NPCs. NPCs responded to empathic opportunities with empathic statements 30% of the time. Additionally, 40% of the time, NPCs responded to empathic opportunities with on-topic, medical explanations and 30% of the responses were not empathic. Patients expressed emotional concerns to NPCs typically in the form of anger; most emotions were moderately severe, with no statistical differences among types of NPC. On average, NPCs responded to patient emotion with empathic language only 30% of the time. A
Spagnuolo, Antonella; De Santis, Marco; Torretta, Claudia; Filippi, Mauro; Talucci, Carlo
The emotional leadership applied to nursing management is a new topic in the Italian nursing literature, but of great interest internationally. There is a close correlation between nursing leaders with a well-developed emotional intelligence and nurses working well-being. This study investigates knowledge about the emotional leadership and emotional competence in nursing management. The survey was conducted using a questionnaire devised for the purpose, validated and administered to 130 managers, head nurses and nurses in a hospital in Rome. Analysis of data shows a great interest in the subject. 90% of the sample showed that it is essential for managerial roles, be aware and able to manage their own and others' emotions to generate wellbeing at work. Emotional competencies are considered important just as theoretical, technical and social skills to a effective leadership on nursing. This study is one of the first Italian survey on the importance of the development of emotional intelligence in nursing leadership to improve wellbeing at work. Results of the survey should be confirmed by further studies. The emotional skills could be improved in nursing education programs and used as a yardstick for the nursing managers selection.
Cebula, Katie R.; Wishart, Jennifer G.; Willis, Diane S.; Pitcairn, Tom K.
Some children with Down syndrome may experience difficulties in recognizing facial emotions, particularly fear, but it is not clear why, nor how such skills can best be facilitated. Using a photo-matching task, emotion recognition was tested in children with Down syndrome, children with nonspecific intellectual disability and cognitively matched,…
Scherer, Klaus R; Ellgring, Heiner
The different assumptions made by discrete and componential emotion theories about the nature of the facial expression of emotion and the underlying mechanisms are reviewed. Explicit and implicit predictions are derived from each model. It is argued that experimental expression-production paradigms rather than recognition studies are required to critically test these differential predictions. Data from a large-scale actor portrayal study are reported to demonstrate the utility of this approach. The frequencies with which 12 professional actors use major facial muscle actions individually and in combination to express 14 major emotions show little evidence for emotion-specific prototypical affect programs. Rather, the results encourage empirical investigation of componential emotion model predictions of dynamic configurations of appraisal-driven adaptive facial actions. (c) 2007 APA, all rights reserved.
Vaiman, Marcelo; Wagner, Mónica Anna; Caicedo, Estefanía; Pereno, Germán Leandro
Pictures of facial expressions of emotion are used in a wide range of experiments. The last decade has seen an increase in the number of studies presenting local sets of emotion stimuli. However, only a few existing sets contain pictures of Latin Americans, despite the growing attention emotion research is receiving in this region. Here we present the development and validation of the Universidad Nacional de Cordoba, Expresiones de Emociones Faciales (UNCEEF), a Facial Action Coding System (FACS)-verified set of pictures of Argentineans expressing the six basic emotions, plus neutral expressions. FACS scores, recognition rates, Hu scores, and discrimination indices are reported. Evidence of convergent validity was obtained using the Pictures of Facial Affect in an Argentine sample. However, recognition accuracy was greater for UNCEEF. The importance of local sets of emotion pictures is discussed.
Full Text Available This article focuses on bodily memory and emotional expressions of the Russian male militaries who participated in the Afghanistan war. The research aims to reveal the axis structure of bodily memory and to analyze the structural elements of emotionality objectified in narrations of the male Afghan veterans. The author reflects the loci of concentrated bodily memory, such as corporal inscriptions and skin writings which are optical, perceptible and indelible. They function as “the truth” about the past imprinted on the male body. The following questions are discussed in the article. How does bodily memory function in male militaries’ narrations? What are the results of bodily memory work? What role do emotions play during the process of remembering? What kind of emotions do male veterans express in their narrations? What type of connection is there between bodily memory and emotional expressing?
Liu, Bin; Govindan, Ramesh; Uzzi, Brian
Emotions are increasingly inferred linguistically from online data with a goal of predicting off-line behavior. Yet, it is unknown whether emotions inferred linguistically from online communications correlate with actual changes in off-line activity. We analyzed all 886,000 trading decisions and 1,234,822 instant messages of 30 professional day traders over a continuous 2 year period. Linguistically inferring the traders’ emotional states from instant messages, we find that emotions expressed in online communications reflect the same distributions of emotions found in controlled experiments done on traders. Further, we find that expressed online emotions predict the profitability of actual trading behavior. Relative to their baselines, traders who expressed little emotion or traders that expressed high levels of emotion made relatively unprofitable trades. Conversely, traders expressing moderate levels of emotional activation made relatively profitable trades. PMID:26765539
Knyazev, Gennady G; Barchard, Kimberly A; Razumnikova, Olga M; Mitrofanova, Larisa G
The tendency to express emotions non-verbally is positively related to perception of emotions in oneself. This study examined its relationship to perception of emotions in others. In 40 healthy adults, EEG theta synchronization was used to indicate emotion processing following presentation of happy, angry, and neutral faces. Both positive and negative expressiveness were associated with higher emotional sensitivity, as shown by cortical responses to facial expressions during the early, unconscious processing stage. At the late, conscious processing stage, positive expressiveness was associated with higher sensitivity to happy faces but lower sensitivity to angry faces. Thus, positive expressiveness predisposes people to allocate fewer attentional resources for conscious perception of angry faces. In contrast, negative expressiveness was consistently associated with higher sensitivity. The effects of positive expressiveness occurred in cortical areas that deal with emotions, but the effects of negative expressiveness occurred in areas engaged in self-referential processes in the context of social relationships. © 2012 The Authors. Scandinavian Journal of Psychology © 2012 The Scandinavian Psychological Associations.
In his seminal paper, Gabrielsson (2002) distinguishes between emotion felt by the listener, here: "internal locus of emotion" (IL), and the emotion the music is expressing, here: "external locus of emotion" (EL). This paper tabulates 16 comparisons of felt versus expressed emotions in music published in the decade 2003-2012 consisting of 19 studies/experiments and provides some theoretical perspectives. The key findings were that (1) IL rating was frequently rated statistically the same or lower than the corresponding EL rating (e.g., lower felt happiness rating compared to the apparent happiness of the music), and that (2) self-select and preferred music had a smaller gap across the emotion loci than experimenter-selected and disliked music. These key findings were explained by an "inhibited" emotional contagion mechanism, where the otherwise matching felt emotion may have been attenuated by some other factor such as social context. Matching between EL and IL for loved and self-selected pieces was explained by the activation of "contagion" circuits. Physiological arousal, personality and age, as well as musical features (tempo, mode, putative emotions) also influenced perceived and felt emotion distinctions. A variety of data collection formats were identified, but mostly using rating items. In conclusion, a more systematic use of terminology appears desirable. Two broad categories, namely matched and unmatched, are proposed as being sufficient to capture the relationships between EL and IL, instead of four categories as suggested by Gabrielsson.
Balconi, Michela; Carrera, Alba
The paper explored conceptual and lexical skills with regard to emotional correlates of facial stimuli and scripts. In two different experimental phases normal and autistic children observed six facial expressions of emotions (happiness, anger, fear, sadness, surprise, and disgust) and six emotional scripts (contextualized facial expressions). In the second place, the effect of emotional domain (different emotions) in decoding was explored. A semantic grid was applied to conversational line, including two levels of data: the lexical adequacy index (correct decoding of emotion) and the emotional vocabulary (such as the causal representation and the hedonic valence of the stimulus). Log-linear analysis showed different representations across the subjects, as a function of emotion, task and pathology. Specifically, childrens' lexical competence was well developed for some emotions (such as happiness, anger, and fear), and as a function of type of task, that is script was better represented than face. Between the main linguistic indexes, causal relation was a prototypical index for emotional conceptualization. Finally, pathology affected children's performance, with an increased "facilitation effect" for autistic children in the script condition.
E Glenn eSchellenberg
Full Text Available Theories of aesthetic appreciation propose that (1 a stimulus is liked because it is expected or familiar, (2 a stimulus is liked most when it is neither too familiar nor too novel, or (3 a novel stimulus is liked because it elicits an intensified emotional response. We tested the third hypothesis by examining liking for music as a function of whether the emotion it expressed contrasted with the emotion expressed by music heard previously. Stimuli were 30-s happy- or sad-sounding excerpts from recordings of classical piano music. On each trial, listeners heard a different excerpt and made liking and emotion-intensity ratings. The emotional character of consecutive excerpts was repeated with varying frequencies, followed by an excerpt that expressed a contrasting emotion. As the number of presentations of the background emotion increased, liking and intensity ratings became lower compared to those for the contrasting emotion. Consequently, when the emotional character of the music was relatively novel, listeners’ responses intensified and their appreciation increased.
Volkman, Julie E.; Parrott, Roxanne L.
This study examined the use of different narratives expressing positive or negative emotions, and varying the narrator's perspective on the arousal of discrete emotions, dominant cognitions, perceived evidence quality, and perceived message effectiveness related to osteoporosis behavioral intentions. Formative research led to the creation of…
Widen, Sherri C.; Russell, James A.
Understanding and recognition of emotions relies on emotion concepts, which are narrative structures (scripts) specifying facial expressions, causes, consequences, label, etc. organized in a temporal and causal order. Scripts and their development are revealed by examining which components better tap which concepts at which ages. This study…
Larsen, Junilla K.; Vermulst, Ad A.; Geenen, Rinie; van Middendorp, Henriet; English, Tammy; Gross, James J.; Ha, Thao; Evers, Catharine; Engels, Rutger C. M. E.
Cross-sectional studies have shown a positive association between expressive suppression and depressive symptoms. These results have been interpreted as reflecting the impact of emotion regulation efforts on depression. However, it is also possible that depression may alter emotion regulation tendencies. The goal of the present study was to…
Wishart, J. G.; Cebula, K. R.; Willis, D. S.; Pitcairn, T. K.
Background: Interpreting emotional expressions is a socio-cognitive skill central to interpersonal interaction. Poor emotion recognition has been reported in autism but is less well understood in other kinds of intellectual disabilities (ID), with procedural differences making comparisons across studies and syndromes difficult. This study aimed to…
Dennis, Tracy A; Cole, Pamela M; Wiggins, Crystal N; Cohen, Laura H; Zalewski, Maureen
Although functional links between emotion and action are implied in emotion regulation research, there is limited evidence that specific adaptive actions for coping with a challenge are more probable when certain negative emotions are expressed. The current study examined this question among 3- and 4-year-olds (N = 113; M age = 47.84 months, SD = 6.19). Emotion expressions and actions were observed during 2 challenging tasks: children waited for a gift while the mother worked, and children worked alone to retrieve a prize from a locked box with the wrong key. Angry and happy expressions, compared with sad expressions, were associated with more actions. These actions varied with the nature of the task, reflecting appreciation of situational appropriateness. In addition, when waiting with the mother, happiness was associated with the broadest range of actions, whereas when working alone on the locked box, anger was associated with the broadest range of actions. Results are discussed in terms of the adaptive function of negative emotions and in terms of functional and dimensional models of emotion. Findings have implications for the development of emotion regulation and social-emotional competence. 2009 APA, all rights reserved.
Bui, T.D.; Heylen, Dirk K.J.; Poel, Mannes; Nijholt, Antinus; Stumptner, Markus; Corbett, Dan; Brooks, Mike
We propose a fuzzy rule-based system to map representations of the emotional state of an animated agent onto muscle contraction values for the appropriate facial expressions. Our implementation pays special attention to the way in which continuous changes in the intensity of emotions can be
Bericat Alastuey, Eduardo
The emotions that human beings experience play a fundamental role in all social phenomena. As a result, sociology needs to incorporate the analysis of emotions into its objects of study. This process began three decades ago with the birth of the sociology of emotions. This article offers an introductory and critical overview of the work sociologists of emotions have carried out so far.
This PhD is an investigation of vocal expressions of emotions, mainly focusing on non-verbal sounds such as laughter, cries and sighs. The research examines the roles of categorical and dimensional factors, the contributions of a number of acoustic cues, and the influence of culture. A series of studies established that naive listeners can reliably identify non-verbal vocalisations of positive and negative emotions in forced-choice and rating tasks. Some evidence for underlying dimensions of arousal and valence is found, although each emotion had a discrete expression. The role of acoustic characteristics of the sounds is investigated experimentally and analytically. This work shows that the cues used to identify different emotions vary, although pitch and pitch variation play a central role. The cues used to identify emotions in non-verbal vocalisations differ from the cues used when comprehending speech. An additional set of studies using stimuli consisting of emotional speech demonstrates that these sounds can also be reliably identified, and rely on similar acoustic cues. A series of studies with a pre-literate Namibian tribe shows that non-verbal vocalisations can be recognized across cultures. An fMRI study carried out to investigate the neural processing of non-verbal vocalisations of emotions is presented. The results show activation in pre-motor regions arising from passive listening to non-verbal emotional vocalisations, suggesting neural auditory-motor interactions in the perception of these sounds. In sum, this thesis demonstrates that non-verbal vocalisations of emotions are reliably identifiable tokens of information that belong to discrete categories. These vocalisations are recognisable across vastly different cultures and thus seem to, like facial expressions of emotions, comprise human universals. Listeners rely mainly on pitch and pitch variation to identify emotions in non verbal vocalisations, which differs with the cues used to comprehend
Laukka, Petri; Neiberg, Daniel; Elfenbein, Hillary Anger
The possibility of cultural differences in the fundamental acoustic patterns used to express emotion through the voice is an unanswered question central to the larger debate about the universality versus cultural specificity of emotion. This study used emotionally inflected standard-content speech segments expressing 11 emotions produced by 100 professional actors from 5 English-speaking cultures. Machine learning simulations were employed to classify expressions based on their acoustic features, using conditions where training and testing were conducted on stimuli coming from either the same or different cultures. A wide range of emotions were classified with above-chance accuracy in cross-cultural conditions, suggesting vocal expressions share important characteristics across cultures. However, classification showed an in-group advantage with higher accuracy in within- versus cross-cultural conditions. This finding demonstrates cultural differences in expressive vocal style, and supports the dialect theory of emotions according to which greater recognition of expressions from in-group members results from greater familiarity with culturally specific expressive styles.
Kashdan, Todd B.; Breen, William E.
The purpose of the present study was to examine social anxiety as a predictor of positive emotions using a short-term prospective design. We examined whether the effects of social anxiety on positive emotions are moderated by tendencies to openly express or suppress emotions. Over the course of a 3-month interval, people with excessive social…
Full Text Available When we do not know how to correctly behave in a new context, the emotions that people familiar with the context show in response to the behaviors of others, can help us understand what to do or not to do. The present study examined cross-cultural differences in how group emotional expressions (anger, sadness, neutral can be used to deduce a norm violation in four cultures (Germany, Israel, Greece and the US, which differ in terms of decoding rules for negative emotions. As expected, in all four countries, anger was a stronger norm violation signal than sadness or neutral expressions. However, angry and sad expressions were perceived as more intense and the relevant norm was learned better in Germany and Israel than in Greece and the US. Participants in Greece were relatively better at using sadness as a sign of a likely norm violation. The results demonstrate both cultural universality and cultural differences in the use of group emotion expressions in norm learning. In terms of cultural differences they underscore that the social signal value of emotional expressions may vary with culture as a function of cultural differences, both in emotion perception, and as a function of a differential use of emotions.
Atadokht, Akbar; Hajloo, Nader; Karimi, Masoud; Narimani, Mohammad
Background: Emotional conditions governing the family and patients’ perceived social support play important roles in the treatment or relapse process of the chronic disease. Objectives: The current study aimed to investigate the role of family expressed emotion and perceived social support in prediction of addiction relapse. Patients and Methods: The descriptive-correlation method was used in the current study. The study population consisted of the individuals referred to the addiction treatment centers in Ardabil from October 2013 to January 2014. The subjects (n = 80) were randomly selected using cluster sampling method. To collect data, expressed emotion test by Cole and Kazaryan, and Multidimensional Scale of Perceived Social Support (MSPSS) were used, and the obtained data was analyzed using the Pearson's correlation coefficient and multiple regression analyses. Results: Results showed a positive relationship between family expressed emotions and the frequency of relapse (r = 0.26, P = 0.011) and a significant negative relationship between perceived social support and the frequency of relapse (r = -0.34, P = 0.001). Multiple regression analysis also showed that perceived social support from family and the family expressed emotions significantly explained 12% of the total variance of relapse frequency. Conclusions: These results have implications for addicted people, their families and professionals working in addiction centers to use the emotional potential of families especially their expressed emotions and the perceived social support of addicts to increase the success rate of addiction treatment. PMID:25883918
Snyder, Peter J; Kaufman, Rebecca; Harrison, John; Maruff, Paul
In the late 1860s and early 1870s, Darwin had corresponded with the French physician and physiologist, G. B. A. Duchenne, regarding Duchenne's experimental manipulation of human facial expression of emotion, by applying Galvanic electrical stimulation directly to facial muscles. Duchenne had produced a set of over 60 photographic plates to illustrate his view that there are different muscles in the human face that are separately responsible for each individual emotion. Darwin studied this material very carefully and he received permission from Duchenne in 1871 to reproduce several of these images in The Expression of the Emotions in Man and Animals (1872). Darwin had doubted Duchenne's view that there were individual muscle groups that mediate the expression of dozens of separable emotions, and he wondered whether there might instead be a fewer set of core emotions that are expressed with great stability worldwide and across cultures. Prompted by his doubts regarding the veracity of Duchenne's model, Darwin conducted what may have been the first-ever single-blind study of the recognition of human facial expression of emotion. This single experiment was a little-known forerunner for an entire modern field of study with contemporary clinical relevance. Moreover, his specific question about cross-cultural recognition of the cardinal emotions in faces is a topic that is being actively studied (in the twenty-first century) with the hope of developing novel biomarkers to aid the discovery of new therapies for the treatment of schizophrenia, autism, and other neuropsychiatric diseases.
Hareli, Shlomo; Kafetsios, Konstantinos; Hess, Ursula
When we do not know how to correctly behave in a new context, the emotions that people familiar with the context show in response to the behaviors of others, can help us understand what to do or not to do. The present study examined cross-cultural differences in how group emotional expressions (anger, sadness, neutral) can be used to deduce a norm violation in four cultures (Germany, Israel, Greece, and the US), which differ in terms of decoding rules for negative emotions. As expected, in all four countries, anger was a stronger norm violation signal than sadness or neutral expressions. However, angry and sad expressions were perceived as more intense and the relevant norm was learned better in Germany and Israel than in Greece and the US. Participants in Greece were relatively better at using sadness as a sign of a likely norm violation. The results demonstrate both cultural universality and cultural differences in the use of group emotion expressions in norm learning. In terms of cultural differences they underscore that the social signal value of emotional expressions may vary with culture as a function of cultural differences, both in emotion perception, and as a function of a differential use of emotions. PMID:26483744
Giner-Bartolomé, Cristina; Steward, Trevor; Wolz, Ines; Jiménez-Murcia, Susana; Granero, Roser; Tárrega, Salomé; Fernández-Formoso, José Antonio; Soriano-Mas, Carles; Menchón, José M; Fernández-Aranda, Fernando
Facial expressions are critical in forming social bonds and in signalling one's emotional state to others. In eating disorder patients, impairments in facial emotion recognition have been associated with eating psychopathology severity. Little research however has been carried out on how bulimic spectrum disorder (BSD) patients spontaneously express emotions. Our aim was to investigate emotion expression in BSD patients and to explore the influence of personality traits. Our study comprised 28 BSD women and 15 healthy controls. Facial expressions were recorded while participants played a serious video game. Expressions of anger and joy were used as outcome measures. Overall, BSD participants displayed less facial expressiveness than controls. Among BSD women, expressions of joy were positively associated with reward dependence, novelty seeking and self-directedness, whereas expressions of anger were associated with lower self-directedness. Our findings suggest that specific personality traits are associated with altered emotion facial expression in patients with BSD. Copyright © 2016 John Wiley & Sons, Ltd and Eating Disorders Association. Copyright © 2016 John Wiley & Sons, Ltd and Eating Disorders Association.
Han, Meng-Ju; Lin, Chia-How; Song, Kai-Tai
This paper presents a method of mood transition design of a robot for autonomous emotional interaction with humans. A 2-D emotional model is proposed to combine robot emotion, mood, and personality in order to generate emotional expressions. In this design, the robot personality is programmed by adjusting the factors of the five factor model proposed by psychologists. From Big Five personality traits, the influence factors of robot mood transition are determined. Furthermore, a method to fuse basic robotic emotional behaviors is proposed in order to manifest robotic emotional states via continuous facial expressions. An artificial face on a screen is a way to provide a robot with a humanlike appearance, which might be useful for human-robot interaction. An artificial face simulator has been implemented to show the effectiveness of the proposed methods. Questionnaire surveys have been carried out to evaluate the effectiveness of the proposed method by observing robotic responses to a user's emotional expressions. Preliminary experimental results on a robotic head show that the proposed mood state transition scheme appropriately responds to a user's emotional changes in a continuous manner.
Johnson, Kareem J.; Waugh, Christian E.; Fredrickson, Barbara L.
The broaden hypothesis, part of Fredrickson’s (1998, 2001) broaden-and-build theory, proposes that positive emotions lead to broadened cognitive states. Here, we present evidence that cognitive broadening can be produced by frequent facial expressions of positive emotion. Additionally, we present a novel method of using facial electromyography (EMG) to discriminate between Duchenne (genuine) and non-Duchenne (non-genuine) smiles. Across experiments, Duchenne smiles occurred more frequently during positive emotion inductions than neutral or negative inductions. Across experiments, Duchenne smiles correlated with self-reports of specific positive emotions. In Experiment 1, high frequencies of Duchenne smiles predicted increased attentional breadth on a global–local visual processing task. In Experiment 2, high frequencies of Duchenne smiles predicted increased attentional flexibility on a covert attentional orienting task. These data underscore the value of using multiple methods to measure emotional experience in studies of emotion and cognition. PMID:23275681
Desrosiers, Alethea; Sipsma, Heather; Divney, Anna; Magriples, Urania; Kershaw, Trace
Deficits in emotion expression skills have been associated with alcohol and substance use, but the mechanisms through which these associations occur are not well understood. The current study investigated (a) associations between emotion expression and substance use (i.e., alcohol, cigarettes, and marijuana) in newly parenting adolescents and young adults and (b) whether symptoms of depression and stress mediate these associations in young mothers and fathers. Participants recruited from obstetrics and gynecology clinics completed the Center for Epidemiological Studies-Depression Scale, Perceived Stress Scale, Emotion Expression Scale for Children, and substance use items. Path analysis indicated that lower emotion expression at 6 months postpartum was significantly associated with more alcohol and marijuana use at 12 months postpartum for males but not females. Also among males, stress levels at 6 months postpartum partially mediated associations between emotion expression and alcohol and marijuana use at 12 months postpartum. Findings suggest that poor emotion expression skills are related to more substance use in young fathers, and levels of stress may partially account for this association. © 2015 Wiley Periodicals, Inc.
Bobkowska, Katarzyna; Przyborski, Marek; Skorupka, Dariusz
This article shows how complex emotions are. This has been proven by the analysis of the changes that occur on the face. The authors present the problem of image analysis for the purpose of identifying emotions. In addition, they point out the importance of recording the phenomenon of the development of emotions on the human face with the use of high-speed cameras, which allows the detection of micro expression. The work that was prepared for this article was based on analyzing the parallax pair correlation coefficients for specific faces. In the article authors proposed to divide the facial image into 8 characteristic segments. With this approach, it was confirmed that at different moments of emotion the pace of expression and the maximum change characteristic of a particular emotion, for each part of the face is different.
Wilhelm, Oliver; Hildebrandt, Andrea; Manske, Karsten; Schacht, Annekathrin; Sommer, Werner
Despite the importance of perceiving and recognizing facial expressions in everyday life, there is no comprehensive test battery for the multivariate assessment of these abilities. As a first step toward such a compilation, we present 16 tasks that measure the perception and recognition of facial emotion expressions, and data illustrating each task's difficulty and reliability. The scoring of these tasks focuses on either the speed or accuracy of performance. A sample of 269 healthy young adults completed all tasks. In general, accuracy and reaction time measures for emotion-general scores showed acceptable and high estimates of internal consistency and factor reliability. Emotion-specific scores yielded lower reliabilities, yet high enough to encourage further studies with such measures. Analyses of task difficulty revealed that all tasks are suitable for measuring emotion perception and emotion recognition related abilities in normal populations. PMID:24860528
Full Text Available This article shows how complex emotions are. This has been proven by the analysis of the changes that occur on the face. The authors present the problem of image analysis for the purpose of identifying emotions. In addition, they point out the importance of recording the phenomenon of the development of emotions on the human face with the use of high-speed cameras, which allows the detection of micro expression. The work that was prepared for this article was based on analyzing the parallax pair correlation coefficients for specific faces. In the article authors proposed to divide the facial image into 8 characteristic segments. With this approach, it was confirmed that at different moments of emotion the pace of expression and the maximum change characteristic of a particular emotion, for each part of the face is different.
Leleu, Arnaud; Demily, Caroline; Franck, Nicolas; Durand, Karine; Schaal, Benoist; Baudouin, Jean-Yves
It has been established that the recognition of facial expressions integrates contextual information. In this study, we aimed to clarify the influence of contextual odors. The participants were asked to match a target face varying in expression intensity with non-ambiguous expressive faces. Intensity variations in the target faces were designed by morphing expressive faces with neutral faces. In addition, the influence of verbal information was assessed by providing half the participants with the emotion names. Odor cues were manipulated by placing participants in a pleasant (strawberry), aversive (butyric acid), or no-odor control context. The results showed two main effects of the odor context. First, the minimum amount of visual information required to perceive an expression was lowered when the odor context was emotionally congruent: happiness was correctly perceived at lower intensities in the faces displayed in the pleasant odor context, and the same phenomenon occurred for disgust and anger in the aversive odor context. Second, the odor context influenced the false perception of expressions that were not used in target faces, with distinct patterns according to the presence of emotion names. When emotion names were provided, the aversive odor context decreased intrusions for disgust ambiguous faces but increased them for anger. When the emotion names were not provided, this effect did not occur and the pleasant odor context elicited an overall increase in intrusions for negative expressions. We conclude that olfaction plays a role in the way facial expressions are perceived in interaction with other contextual influences such as verbal information.
Full Text Available Individual genetic differences in the serotonin transporter-linked polymorphic region (5-HTTLPR have been associated with variations in the sensitivity to social and emotional cues as well as altered amygdala reactivity to facial expressions of emotion. Amygdala activation has further been shown to trigger gaze changes towards diagnostically relevant facial features. The current study examined whether altered socio-emotional reactivity in variants of the 5-HTTLPR promoter polymorphism reflects individual differences in attending to diagnostic features of facial expressions. For this purpose, visual exploration of emotional facial expressions was compared between a low (n=39 and a high (n=40 5-HTT expressing group of healthy human volunteers in an eye tracking paradigm. Emotional faces were presented while manipulating the initial fixation such that saccadic changes towards the eyes and towards the mouth could be identified. We found that the low versus the high 5-HTT group demonstrated greater accuracy with regard to emotion classifications, particularly when faces were presented for a longer duration. No group differences in gaze orientation towards diagnostic facial features could be observed. However, participants in the low 5-HTT group exhibited more and faster fixation changes for certain emotions when faces were presented for a longer duration and overall face fixation times were reduced for this genotype group. These results suggest that the 5-HTT gene influences social perception by modulating the general vigilance to social cues rather than selectively affecting the pre-attentive detection of diagnostic facial features.
Full Text Available Body expressions exert strong contextual effects on facial emotion perception in adults. Specifically, conflicting body cues hamper the recognition of emotion from faces, as evident on both the behavioral and neural level. We examined the developmental origins of the neural processes involved in emotion perception across body and face in 8-month-old infants by measuring event-related brain potentials (ERPs. We primed infants with body postures (fearful, happy that were followed by either congruent or incongruent facial expressions. Our results revealed that body expressions impact facial emotion processing and that incongruent body cues impair the neural discrimination of emotional facial expressions. Priming effects were associated with attentional and recognition memory processes, as reflected in a modulation of the Nc and Pc evoked at anterior electrodes. These findings demonstrate that 8-month-old infants possess neural mechanisms that allow for the integration of emotion across body and face, providing evidence for the early developmental emergence of context-sensitive facial emotion perception.
Mitchell, Rachel L C; Jazdzyk, Agnieszka; Stets, Manuela; Kotz, Sonja A
We aimed to progress understanding of prosodic emotion expression by establishing brain regions active when expressing specific emotions, those activated irrespective of the target emotion, and those whose activation intensity varied depending on individual performance. BOLD contrast data were acquired whilst participants spoke non-sense words in happy, angry or neutral tones, or performed jaw-movements. Emotion-specific analyses demonstrated that when expressing angry prosody, activated brain regions included the inferior frontal and superior temporal gyri, the insula, and the basal ganglia. When expressing happy prosody, the activated brain regions also included the superior temporal gyrus, insula, and basal ganglia, with additional activation in the anterior cingulate. Conjunction analysis confirmed that the superior temporal gyrus and basal ganglia were activated regardless of the specific emotion concerned. Nevertheless, disjunctive comparisons between the expression of angry and happy prosody established that anterior cingulate activity was significantly higher for angry prosody than for happy prosody production. Degree of inferior frontal gyrus activity correlated with the ability to express the target emotion through prosody. We conclude that expressing prosodic emotions (vs. neutral intonation) requires generic brain regions involved in comprehending numerous aspects of language, emotion-related processes such as experiencing emotions, and in the time-critical integration of speech information.
Rachel L. C. Mitchell
Full Text Available We aimed to progress understanding of prosodic emotion expression by establishing brain regions active when expressing specific emotions, those activated irrespective of the target emotion, and those whose activation intensity varied depending on individual performance. BOLD contrast data were acquired whilst participants spoke nonsense words in happy, angry or neutral tones, or performed jaw-movements. Emotion-specific analyses demonstrated that when expressing angry prosody, activated brain regions included the inferior frontal and superior temporal gyri, the insula, and the basal ganglia. When expressing happy prosody, the activated brain regions also included the superior temporal gyrus, insula, and basal ganglia, with additional activation in the anterior cingulate. Conjunction analysis confirmed that the superior temporal gyrus and basal ganglia were activated regardless of the specific emotion concerned. Nevertheless, disjunctive comparisons between the expression of angry and happy prosody established that anterior cingulate activity was significantly higher for angry prosody than for happy prosody production. Degree of inferior frontal gyrus activity correlated with the ability to express the target emotion through prosody. We conclude that expressing prosodic emotions (vs neutral intonation requires generic brain regions involved in comprehending numerous aspects of language, emotion-related processes such as experiencing emotions, and in the time-critical integration of speech information.
Kavallakis, George; Vidakis, Nikolaos; Triantafyllidis, Georgios
This paper presents a scheme of creating an emotion index of cover song music video clips by recognizing and classifying facial expressions of the artist in the video. More specifically, it fuses effective and robust algorithms which are employed for expression recognition, along with the use...... of a neural network system using the features extracted by the SIFT algorithm. Also we support the need of this fusion of different expression recognition algorithms, because of the way that emotions are linked to facial expressions in music video clips....
Acerbi, Alberto; Lampos, Vasileios; Garnett, Philip; Bentley, R Alexander
We report here trends in the usage of "mood" words, that is, words carrying emotional content, in 20th century English language books, using the data set provided by Google that includes word frequencies in roughly 4% of all books published up to the year 2008. We find evidence for distinct historical periods of positive and negative moods, underlain by a general decrease in the use of emotion-related words through time. Finally, we show that, in books, American English has become decidedly more "emotional" than British English in the last half-century, as a part of a more general increase of the stylistic divergence between the two variants of English language.
Gillespie, Steven M; Rotshtein, Pia; Satherley, Rose-Marie; Beech, Anthony R; Mitchell, Ian J
Research with violent offenders has consistently shown impaired recognition of other's facial expressions of emotion. However, the extent to which similar problems can be observed among sexual offenders remains unknown. Using a computerized task, we presented sexual and violent offenders, and non-offenders, with male and female expressions of anger, disgust, fear, happiness, sadness, and surprise, morphed with neutral expressions at varying levels of intensity (10, 55, and 90% expressive). Based on signal detection theory, we used hit rates and false alarms to calculate the sensitivity index d-prime (d') and criterion (c) for each emotional expression. Overall, sexual offenders showed reduced sensitivity to emotional expressions across intensity, sex, and type of expression, compared with non-offenders, while both sexual and violent offenders showed particular reduced sensitivity to fearful expressions. We also observed specific effects for high (90%) intensity female faces, with sexual offenders showing reduced sensitivity to anger compared with non-offenders and violent offenders, and reduced sensitivity to disgust compared with non-offenders. Furthermore, both sexual and violent offenders showed impaired sensitivity to high intensity female fearful expressions compared with non-offenders. Violent offenders also showed a higher criterion for classifying moderate and high intensity male expressions as fearful, indicative of a more conservative response style, compared with angry, happy, or sad. These results suggest that both types of offender show problems in emotion recognition, and may have implications for understanding the inhibition of violent and sexually violent behaviors.
Steven Mark Gillespie
Full Text Available Research with violent offenders has consistently shown impaired recognition of other’s facial expressions of emotion. However, the extent to which similar problems can be observed among sexual offenders remains unknown. Using a computerized task, we presented sexual and violent offenders, and non-offenders, with male and female expressions of anger, disgust, fear, happiness, sadness, and surprise, morphed with neutral expressions at varying levels of intensity (10%, 55%, and 90% expressive. Based on signal detection theory, we used hit rates and false alarms to calculate the sensitivity index d-prime (d’ and criterion (c for each emotional expression. Overall, sexual offenders showed reduced sensitivity to emotional expressions across intensity, sex, and type of expression, compared with non-offenders, while both sexual and violent offenders showed particular reduced sensitivity to fearful expressions. We also observed specific effects for high (90% intensity female faces, with sexual offenders showing reduced sensitivity to anger compared with non-offenders and violent offenders, and reduced sensitivity to disgust compared with non-offenders. Furthermore, both sexual and violent offenders showed impaired sensitivity to high intensity female fearful expressions compared with non-offenders. Violent offenders also showed a higher criterion for classifying moderate and high intensity male expressions as fearful, indicative of a more conservative response style, compared with angry, happy, or sad. These results suggest that both types of offender show problems in emotion recognition, and may have implications for understanding the inhibition of violent and sexually violent behaviors.
van Doorn, Evert A; van Kleef, Gerben A; van der Pligt, Joop
How do instructors' emotional expressions influence students' learning performance? Scholars and practitioners alike have emphasized the importance of positive, nurturing emotions for successful learning. However, teachers may sometimes lose their temper and express anger at their pupils. Drawing on emotions as social information (EASI) theory, we hypothesized that expressions of anger can benefit learning performance. In Experiment 1, participants who were confronted with an angry instructor exhibited more accurate recognition of word pairs after a week of learning, compared with those who were confronted with a happy instructor. In Experiment 2, we conceptually replicated this effect on a recall task, but only among participants in a promotion rather than prevention focus. Present findings thus show, for the 1st time, that instructor anger can enhance students' performance. Findings are consistent with a conceptualization of emotion as social information and call into question the generally endorsed positivity paradigm. PsycINFO Database Record (c) 2014 APA, all rights reserved.
Laukka, Petri; Audibert, Nicolas; Aubergé, Véronique
We examined what determines the typicality, or graded structure, of vocal emotion expressions. Separate groups of judges rated acted and spontaneous expressions of anger, fear, and joy with regard to their typicality and three main determinants of the graded structure of categories: category members' similarity to the central tendency of their category (CT); category members' frequency of instantiation, i.e., how often they are encountered as category members (FI); and category members' similarity to ideals associated with the goals served by its category, i.e., suitability to express particular emotions. Partial correlations and multiple regression analysis revealed that similarity to ideals, rather than CT or FI, explained most variance in judged typicality. Results thus suggest that vocal emotion expressions constitute ideal-based goal-derived categories, rather than taxonomic categories based on CT and FI. This could explain how prototypical expressions can be acoustically distinct and highly recognisable but occur relatively rarely in everyday speech.
Pennebaker, J W; Rimé, B; Blankenship, V E
Montesquieu argued that residents of warmer climates are more emotionally expressive than those living in cooler ones. More than 2,900 college students from 26 countries completed a brief questionnaire assessing the degree to which they considered Northerners and Southerners within their own countries to be emotionally expressive. In addition, individuals rated themselves on their own degree of expressiveness. In partial confirmation of Montesquieu's hypothesis, it was found that large within-country North-South stereotypes exist. Especially in Old World countries, Northerners are viewed as less emotionally expressive than Southerners. Regression and other analyses revealed that self-ratings of expressiveness were, in fact, related to being from the South and to warmer mean temperatures. Several possible explanations for these effects are discussed.
Heshmati, Saeideh; Sbarra, David A; Mason, Ashley E
The importance of studying specific and expressed emotions after a stressful life event is well known, yet few studies have moved beyond assessing self-reported emotional responses to a romantic breakup. This study examined associations between computer-recognized facial expressions and self-reported breakup-related distress among recently separated college-aged young adults ( N = 135; 37 men) on four visits across 9 weeks. Participants' facial expressions were coded using the Computer Expression Recognition Toolbox while participants spoke about their breakups. Of the seven expressed emotions studied, only Contempt showed a unique association with breakup-related distress over time. At baseline, greater Contempt was associated with less breakup-related distress; however, over time, greater Contempt was associated with greater breakup-related distress.
Nelson, Nicole L; Russell, James A
In a classic study, children were shown an array of facial expressions and asked to choose the person who expressed a specific emotion. Children were later asked to name the emotion in the face with any label they wanted. Subsequent research often relied on the same two tasks--choice from array and free labeling--to support the conclusion that children recognize basic emotions from facial expressions. Here five studies (N=120, 2- to 10-year-olds) showed that these two tasks produce illusory recognition; a novel nonsense facial expression was included in the array. Children "recognized" a nonsense emotion (pax or tolen) and two familiar emotions (fear and jealousy) from the same nonsense face. Children likely used a process of elimination; they paired the unknown facial expression with a label given in the choice-from-array task and, after just two trials, freely labeled the new facial expression with the new label. These data indicate that past studies using this method may have overestimated children's expression knowledge. Copyright © 2015 Elsevier Inc. All rights reserved.
Wieckowski, Andrea Trubanova; White, Susan W
Impaired emotion recognition and expression in individuals with autism spectrum disorder (ASD) may contribute to observed social impairment. The aim of this study was to examine the role of visual attention directed toward nonsocial aspects of a scene as a possible mechanism underlying recognition and expressive ability deficiency in ASD. One recognition and two expression tasks were administered. Recognition was assessed in force-choice paradigm, and expression was assessed during scripted and free-choice response (in response to emotional stimuli) tasks in youth with ASD (n = 20) and an age-matched sample of typically developing youth (n = 20). During stimulus presentation prior to response in each task, participants' eye gaze was tracked. Youth with ASD were less accurate at identifying disgust and sadness in the recognition task. They fixated less to the eye region of stimuli showing surprise. A group difference was found during the free-choice response task, such that those with ASD expressed emotion less clearly but not during the scripted task. Results suggest altered eye gaze to the mouth region but not the eye region as a candidate mechanism for decreased ability to recognize or express emotion. Findings inform our understanding of the association between social attention and emotion recognition and expression deficits.
Dapelo, Marcela Marin; Bodas, Sergio; Morris, Robin; Tchanturia, Kate
People with eating disorders have difficulties in socio emotional functioning that could contribute to maintaining the functional consequences of the disorder. This study aimed to explore the ability to deliberately generate (i.e., pose) and imitate facial expressions of emotions in women with anorexia (AN) and bulimia nervosa (BN), compared to healthy controls (HC). One hundred and three participants (36 AN, 25 BN, and 42 HC) were asked to pose and imitate facial expressions of anger, disgust, fear, happiness, and sadness. Their facial expressions were recorded and coded. Participants with eating disorders (both AN and BN) were less accurate than HC when posing facial expressions of emotions. Participants with AN were less accurate compared to HC imitating facial expressions, whilst BN participants had a middle range performance. All results remained significant after controlling for anxiety, depression and autistic features. The relatively small number of BN participants recruited for this study. The study findings suggest that people with eating disorders, particularly those with AN, have difficulties posing and imitating facial expressions of emotions. These difficulties could have an impact in social communication and social functioning. This is the first study to investigate the ability to pose and imitate facial expressions of emotions in people with eating disorders, and the findings suggest this area should be further explored in future studies. Copyright © 2015. Published by Elsevier B.V.
Charbonneau, Geneviève; Bertone, Armando; Lepore, Franco; Nassim, Marouane; Lassonde, Maryse; Mottron, Laurent; Collignon, Olivier
The abilities to recognize and integrate emotions from another person's facial and vocal expressions are fundamental cognitive skills involved in the effective regulation of social interactions. Deficits in such abilities have been suggested as a possible source for certain atypical social behaviors manifested by persons with autism spectrum disorders (ASD). In the present study, we assessed the recognition and integration of emotional expressions in ASD using a validated set of ecological stimuli comprised of dynamic visual and auditory (non-verbal) vocal clips. Autistic participants and typically developing controls (TD) were asked to discriminate between clips depicting expressions of disgust and fear presented either visually, auditorily or audio-visually. The group of autistic participants was less efficient to discriminate emotional expressions across all conditions (unimodal and bimodal). Moreover, they necessitated a higher signal-to-noise ratio for the discrimination of visual or auditory presentations of disgust versus fear expressions. These results suggest an altered sensitivity to emotion expressions in this population that is not modality-specific. In addition, the group of autistic participants benefited from exposure to bimodal information to a lesser extent than did the TD group, indicative of a decreased multisensory gain in this population. These results are the first to compellingly demonstrate joint alterations for both the perception and the integration of multisensory emotion expressions in ASD. Copyright © 2013 Elsevier Ltd. All rights reserved.
Full Text Available The present paper explored the relationship between emotional facial response and electromyographic modulation in children when they observe facial expression of emotions. Facial responsiveness (evaluated by arousal and valence ratings and psychophysiological correlates (facial electromyography, EMG were analyzed when children looked at six facial expressions of emotions (happiness, anger, fear, sadness, surprise and disgust. About EMG measure, corrugator and zygomatic muscle activity was monitored in response to different emotional types. ANOVAs showed differences for both EMG and facial response across the subjects, as a function of different emotions. Specifically, some emotions were well expressed by all the subjects (such as happiness, anger and fear in terms of high arousal, whereas some others were less level arousal (such as sadness. Zygomatic activity was increased mainly for happiness, from one hand, corrugator activity was increased mainly for anger, fear and surprise, from the other hand. More generally, EMG and facial behavior were highly correlated each other, showing a “mirror” effect with respect of the observed faces.
Del Piccolo, Lidia; Pietrolongo, Erika; Radice, Davide; Tortorella, Carla; Confalonieri, Paolo; Pugliatti, Maura; Lugaresi, Alessandra; Giordano, Andrea; Heesen, Christoph; Solari, Alessandra
Background Anxiety and depression are common in people with multiple sclerosis (MS), but data on emotional communication during MS consultations are lacking. We assessed patient expressions of emotion and neurologist responses during first-ever MS consultations using the Verona Coding Definitions of Emotional Sequences (VR-CoDES). Methods We applied VR-CoDES to recordings/transcripts of 88 outpatient consultations (10 neurologists, four MS Italian centers). Before consultation, patients completed the Hospital Anxiety and Depression Scale (HADS). Multilevel sequential analysis was performed on the number of cues/concerns expressed by patients, and the proportion of reduce space responses by neurologists. Results Patients expressed 492 cues and 45 concerns (median 4 cues and 1 concern per consultation). The commonest cues were verbal hints of hidden worries (cue type b, 41%) and references to stressful life events (type d, 26%). Variables independently associated with number of cues/concerns were: anxiety (HADS-Anxiety score >8) (incidence risk ratio, IRR 1.08, 95% CI 1.06-1.09; pAnxiety was the only variable significantly associated with ‘reduce space’ responses (odds ratio 2.17, 95% CI 1.32-3.57; p=0.003). Conclusions Patient emotional expressions varied widely, but VR-CoDES cues b and d were expressed most often. Patient anxiety was directly associated with emotional expressions; older age of patients and neurologists, and second opinion consultations were inversely associated with patient emotional expression. In over 50% of instances, neurologists responded to these expressions by reducing space, more so in anxious patients. These findings suggest that neurologists need to improve their skills in dealing with patient emotions. PMID:26030822
Martinez, Aleix; Du, Shichuan
In cognitive science and neuroscience, there have been two leading models describing how humans perceive and classify facial expressions of emotion-the continuous and the categorical model. The continuous model defines each facial expression of emotion as a feature vector in a face space. This model explains, for example, how expressions of emotion can be seen at different intensities. In contrast, the categorical model consists of C classifiers, each tuned to a specific emotion category. This model explains, among other findings, why the images in a morphing sequence between a happy and a surprise face are perceived as either happy or surprise but not something in between. While the continuous model has a more difficult time justifying this latter finding, the categorical model is not as good when it comes to explaining how expressions are recognized at different intensities or modes. Most importantly, both models have problems explaining how one can recognize combinations of emotion categories such as happily surprised versus angrily surprised versus surprise. To resolve these issues, in the past several years, we have worked on a revised model that justifies the results reported in the cognitive science and neuroscience literature. This model consists of C distinct continuous spaces. Multiple (compound) emotion categories can be recognized by linearly combining these C face spaces. The dimensions of these spaces are shown to be mostly configural. According to this model, the major task for the classification of facial expressions of emotion is precise, detailed detection of facial landmarks rather than recognition. We provide an overview of the literature justifying the model, show how the resulting model can be employed to build algorithms for the recognition of facial expression of emotion, and propose research directions in machine learning and computer vision researchers to keep pushing the state of the art in these areas. We also discuss how the model can
Sandberg, Kristian; Bibby, Bo Martin; Overgaard, Morten
with emotional content (fearful vs. neutral faces). Although we find the study interesting, we disagree with the conclusion that CR is superior to PAS because of two methodological issues. First, the conclusion is not based on a formal test. We performed this test and found no evidence that CR predicted accuracy...
Hunter, Patrick G.; Schellenberg, E. Glenn; Stalinski, Stephanie M.
Adults and children 5, 8, and 11 years of age listened to short excerpts of unfamiliar music that sounded happy, scary, peaceful, or sad. Listeners initially rated how much they liked each excerpt. They subsequently made a forced-choice judgment about the emotion that each excerpt conveyed. Identification accuracy was higher for young girls than…
Louie, Jennifer Yu; Wang, Shu-wen; Fung, Joey; Lau, Anna
Previous research suggests that adult perceptions of children's social competence may vary depending on the socialization goals in a given cultural context. There is also ample evidence of cultural differences in values concerning emotional display, with East Asian collectivistic contexts favoring restraint and Western individualistic contexts…
Faliagkas, L.; Rao-Ruiz, P.; Kindt, M.
The hypothesis that fear memory is not necessarily permanent but can change when retrieved opens avenues to develop revolutionary treatments for emotional memory disorders. Memory reconsolidation is however only one of several mnemonic processes that may be triggered by memory reactivation and
In his seminal paper, Gabrielsson (2002) distinguishes between emotion felt by the listener, here: “internal locus of emotion” (IL), and the emotion the music is expressing, here: “external locus of emotion” (EL). This paper tabulates 16 comparisons of felt versus expressed emotions in music published in the decade 2003–2012 consisting of 19 studies/experiments and provides some theoretical perspectives. The key findings were that (1) IL rating was frequently rated statistically the same or lower than the corresponding EL rating (e.g., lower felt happiness rating compared to the apparent happiness of the music), and that (2) self-select and preferred music had a smaller gap across the emotion loci than experimenter-selected and disliked music. These key findings were explained by an “inhibited” emotional contagion mechanism, where the otherwise matching felt emotion may have been attenuated by some other factor such as social context. Matching between EL and IL for loved and self-selected pieces was explained by the activation of “contagion” circuits. Physiological arousal, personality and age, as well as musical features (tempo, mode, putative emotions) also influenced perceived and felt emotion distinctions. A variety of data collection formats were identified, but mostly using rating items. In conclusion, a more systematic use of terminology appears desirable. Two broad categories, namely matched and unmatched, are proposed as being sufficient to capture the relationships between EL and IL, instead of four categories as suggested by Gabrielsson. PMID:24381565
Gillespie, Steven M.; Rotshtein, Pia; Satherley, Rose-Marie; Beech, Anthony R.; Mitchell, Ian J.
Research with violent offenders has consistently shown impaired recognition of other’s facial expressions of emotion. However, the extent to which similar problems can be observed among sexual offenders remains unknown. Using a computerized task, we presented sexual and violent offenders, and non-offenders, with male and female expressions of anger, disgust, fear, happiness, sadness, and surprise, morphed with neutral expressions at varying levels of intensity (10%, 55%, and 90% expressive). ...
Schwartz, Rachel; Pell, Marc D
Patients with Parkinson's disease (PD) are perceived more negatively than their healthy peers, yet it remains unclear what factors contribute to this negative social perception. Based on a cohort of 17 PD patients and 20 healthy controls, we assessed how naïve raters judge the emotion and emotional intensity displayed in dynamic facial expressions as adults with and without PD watched emotionally evocative films (Experiment 1), and how age-matched peers naïve to patients' disease status judge their social desirability along various dimensions from audiovisual stimuli (interview excerpts) recorded after certain films (Experiment 2). In Experiment 1, participants with PD were rated as significantly more facially expressive than healthy controls; moreover, ratings demonstrated that PD patients were routinely mistaken for experiencing a negative emotion, whereas controls were rated as displaying a more positive emotion than they reported feeling. In Experiment 2, results showed that age-peers rated PD patients as significantly less socially desirable than control participants. Specifically, PD patients were rated as less involved, interested, friendly, intelligent, optimistic, attentive, and physically attractive than healthy controls. Taken together, our results point to a disconnect between how PD patients report feeling and attributions that others make about their emotions and social characteristics, underlining significant social challenges of the disease. In particular, changes in the ability to modulate the expression of negative emotions may contribute to the negative social impressions that many PD patients face.
Palm, M E; Elliott, R; McKie, S; Deakin, J F W; Anderson, I M
Generalized anxiety disorder (GAD) is under-researched despite its high prevalence and large impact on the healthcare system. There is a paucity of functional magnetic resonance imaging (fMRI) studies that explore the neural correlates of emotional processing in GAD. The present study investigated the blood oxygen level dependent (BOLD) response to processing positive and negative facial emotions in patients with GAD. A total of 15 female GAD patients and 16 female controls undertook an implicit face emotion task during fMRI scanning. They also performed a face emotion recognition task outside the scanner. The only behavioural difference observed in GAD patients was less accurate detection of sad facial expressions compared with control participants. However, GAD patients showed an attenuated BOLD signal in the prefrontal cortex to fearful, sad, angry and happy facial expressions and an attenuated signal in the anterior cingulate cortex to happy and fearful facial expressions. No differences were found in amygdala response. In contrast with previous research, this study found BOLD signal attenuation in the ventrolateral and medial prefrontal cortex and the anterior cingulate cortex during face emotion processing, consistent with a hypothesis of hypo-responsivity to external emotional stimuli in GAD. These decreases were in areas that have been implicated in emotion and cognition and may reflect an altered balance between internally and externally directed attentional processes.
Stewart, Suzanne L K; Schepman, Astrid; Haigh, Matthew; McHugh, Rhian; Stewart, Andrew J
The recognition of emotional facial expressions is often subject to contextual influence, particularly when the face and the context convey similar emotions. We investigated whether spontaneous, incidental affective theory of mind inferences made while reading vignettes describing social situations would produce context effects on the identification of same-valenced emotions (Experiment 1) as well as differently-valenced emotions (Experiment 2) conveyed by subsequently presented faces. Crucially, we found an effect of context on reaction times in both experiments while, in line with previous work, we found evidence for a context effect on accuracy only in Experiment 1. This demonstrates that affective theory of mind inferences made at the pragmatic level of a text can automatically, contextually influence the perceptual processing of emotional facial expressions in a separate task even when those emotions are of a distinctive valence. Thus, our novel findings suggest that language acts as a contextual influence to the recognition of emotional facial expressions for both same and different valences.
Full Text Available Autism is characterized by difficulties in social interaction, communication, restrictive and repetitive behaviours and specific impairments in emotional processing. The present study employed The Autism Spectrum Quotient (Baron-Cohen et al. 2006 to quantify autistic traits in a group of 260 healthy individuals and to investigate whether this measure is related to the perception of facial emotional expressions. The emotional processing of twelve participants that scored significantly higher than the average on the AQ was compared to twelve participants with significantly lower AQ scores. Perception of emotional expressions was estimated by The Facial Recognition Task (Montagne et al. 2007. There were significant differences between the two groups with regard to accuracy and sensitivity of the perception of emotional facial expressions. Specifically, the group with high AQ score was less accurate and needed higher emotional content to recognize emotions of anger, disgust, happiness and sadness. This result implies a selective impairment that might be helpful in understanding the psychopathology of autism spectrum disorders.
Individuals regulate their emotions in a wide variety of ways. In the present review it has been addressed the issue of whether some forms of emotion regulation are healthier than others by focusing on two commonly used emotion regulation strategies: cognitive reappraisal (changing the way one thinks about potentially emotion-eliciting events) and expressive suppression (changing the way one behaviorally responds to emotion-eliciting events). In the first section, experimental findings showin...
Moreno, Patricia I; Bauer, Margaret R; Yanez, Betina; Jorge, Alexandra; Maggard-Gibbons, Melinda; Stanton, Annette L
Coping processes directed toward avoiding and approaching stressor-related thoughts and emotions predict psychological adjustment. However, few studies have examined how the relationship between dispositional emotional tendencies and stressor-specific coping affects outcomes. The aim of the current study was to examine the association of dispositional emotional expressivity (i.e., the propensity to experience and express emotions strongly) with cancer-specific coping through avoidance and emotional approach to predict intrusive thoughts and depressive symptoms in Latinas with breast cancer. Recently diagnosed Latina breast cancer patients receiving treatment completed standardized assessments via interview at 2 time points: within 18 months of diagnosis (Time 1; N = 95) and 3 months later (Time 2; N = 79). Most women were immigrants (93%), reported a combined household income of $20,000 or less (75%), did not graduate from high school (59%), and primarily spoke Spanish (88%). In path analyses, more recent immigration was associated with greater dispositional expressivity, which in turn was associated with coping with the cancer experience using both greater avoidance and emotional approach strategies. Only avoidance-oriented strategies predicted an increase in intrusive thoughts at 3 months. No significant effects on depressive symptoms were observed. Findings suggest that Latina breast cancer patients who have a propensity to experience and express emotions strongly may be initially overwhelmed by their cancer-related emotions and consequently turn to avoidance-oriented and emotional approach strategies to cope with their diagnosis. Avoidance-oriented coping in turn may uniquely predict an increase in cancer-related intrusive thoughts 3 months later. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Diana S Cortes
Full Text Available We investigated how memory for faces and voices (presented separately and in combination varies as a function of sex and emotional expression (anger, disgust, fear, happiness, sadness, and neutral. At encoding, participants judged the expressed emotion of items in forced-choice tasks, followed by incidental Remember/Know recognition tasks. Results from 600 participants showed that accuracy (hits minus false alarms was consistently higher for neutral compared to emotional items, whereas accuracy for specific emotions varied across the presentation modalities (i.e., faces, voices, and face-voice combinations. For the subjective sense of recollection ("remember" hits, neutral items received the highest hit rates only for faces, whereas for voices and face-voice combinations anger and fear expressions instead received the highest recollection rates. We also observed better accuracy for items by female expressers, and own-sex bias where female participants displayed memory advantage for female faces and face-voice combinations. Results further suggest that own-sex bias can be explained by recollection, rather than familiarity, rates. Overall, results show that memory for faces and voices may be influenced by the expressions that they carry, as well as by the sex of both items and participants. Emotion expressions may also enhance the subjective sense of recollection without enhancing memory accuracy.
Panjwani, Naaila; Chaplin, Tara M; Sinha, Rajita; Mayes, Linda C
Gender roles in mainstream U.S. culture suggest that girls express more happiness, sadness, anxiety, and shame/embarrassment than boys, while boys express more anger and externalizing emotions, such as contempt. However, gender roles and emotion expression may be different in low-income and ethnically diverse families, as children and parents are often faced with greater environmental stressors and may have different gender expectations. This study examined gender differences in emotion expression in low-income adolescents, an understudied population. One hundred and seventy nine adolescents (aged 14-17) participated in the Trier Social Stress Test (TSST). Trained coders rated adolescents' expressions of happiness, sadness, anxiety, shame/embarrassment, anger, and contempt during the TSST using a micro-analytic coding system. Analyses showed that, consistent with gender roles, girls expressed higher levels of happiness and shame than boys; however, contrary to traditional gender roles, girls showed higher levels of contempt than boys. Also, in contrast to cultural stereotypes, there were no differences in anger between boys and girls. Findings suggest gender-role inconsistent displays of externalizing emotions in low-income adolescents under acute stress, and may reflect different emotion socialization experiences in this group.
Tuck, Natalie L; Grant, Rosemary C I; Sollers, John J; Booth, Roger J; Consedine, Nathan S
Vagally mediated heart rate variability (vmHRV) is a measure of cardiac vagal tone, and is widely viewed as a physiological index of the capacity to regulate emotions. However, studies have not directly tested whether vmHRV is associated with the ability to facially express emotions. In extending prior work, the current report tested links between resting vmHRV and the objectively assessed ability to facially express emotions, hypothesizing that higher vmHRV would predict greater expressive skill. Eighty healthy women completed self-reported measures, before attending a laboratory session in which vmHRV and the ability to express six emotions in the face were assessed. A repeated measures analysis of variance revealed a marginal main effect for vmHRV on skill overall; individuals with higher resting vmHRV were only better able to deliberately facially express anger and interest. Findings suggest that differences in resting vmHRV are associated with the objectively assessed ability to facially express some, but not all, emotions, with potential implications for health and well-being. © 2016 Society for Psychophysiological Research.
Laukka, Petri; Lindahl, Christina; Fischer, Håkan
We investigated how memory for faces and voices (presented separately and in combination) varies as a function of sex and emotional expression (anger, disgust, fear, happiness, sadness, and neutral). At encoding, participants judged the expressed emotion of items in forced-choice tasks, followed by incidental Remember/Know recognition tasks. Results from 600 participants showed that accuracy (hits minus false alarms) was consistently higher for neutral compared to emotional items, whereas accuracy for specific emotions varied across the presentation modalities (i.e., faces, voices, and face-voice combinations). For the subjective sense of recollection (“remember” hits), neutral items received the highest hit rates only for faces, whereas for voices and face-voice combinations anger and fear expressions instead received the highest recollection rates. We also observed better accuracy for items by female expressers, and own-sex bias where female participants displayed memory advantage for female faces and face-voice combinations. Results further suggest that own-sex bias can be explained by recollection, rather than familiarity, rates. Overall, results show that memory for faces and voices may be influenced by the expressions that they carry, as well as by the sex of both items and participants. Emotion expressions may also enhance the subjective sense of recollection without enhancing memory accuracy. PMID:28570691
S Cortes, Diana; Laukka, Petri; Lindahl, Christina; Fischer, Håkan
We investigated how memory for faces and voices (presented separately and in combination) varies as a function of sex and emotional expression (anger, disgust, fear, happiness, sadness, and neutral). At encoding, participants judged the expressed emotion of items in forced-choice tasks, followed by incidental Remember/Know recognition tasks. Results from 600 participants showed that accuracy (hits minus false alarms) was consistently higher for neutral compared to emotional items, whereas accuracy for specific emotions varied across the presentation modalities (i.e., faces, voices, and face-voice combinations). For the subjective sense of recollection ("remember" hits), neutral items received the highest hit rates only for faces, whereas for voices and face-voice combinations anger and fear expressions instead received the highest recollection rates. We also observed better accuracy for items by female expressers, and own-sex bias where female participants displayed memory advantage for female faces and face-voice combinations. Results further suggest that own-sex bias can be explained by recollection, rather than familiarity, rates. Overall, results show that memory for faces and voices may be influenced by the expressions that they carry, as well as by the sex of both items and participants. Emotion expressions may also enhance the subjective sense of recollection without enhancing memory accuracy.
Möller-Leimkühler, Anne Maria; Jandl, Mitja
While the impact of mentally ill patients' perceptions of their key relatives' expressed emotion is well examined with regard to relapse, there is a paucity of evidence concerning the impact on their key relatives' burden. The present study aims to evaluate the relative prognostic value of expressed and perceived emotion on caregivers' stress outcome within a 3-year follow-up period. Yearly follow-up data of the key relatives of 16 first-hospitalized schizophrenic and 34 depressed patients were available including expressed and perceived emotion and different dimensions of caregivers' stress outcome: objective and subjective burden, well-being, psychological symptoms and subjective quality of life. Multiple linear regression analyses were computed to assess the relative impact of expressed and perceived emotion. All dimensions of burden were significantly and consistently correlated with caregivers' expressed emotion and patients' perceived criticism on the bivariate level. On the multivariate level, however, expressed criticism appeared to be the most relevant predictor, followed by perceived resignation. Data indicate that the impact of the patients' perceived criticism on caregivers' stress outcome is limited. More attention should be paid to patients' perceived resignation which may be an unidentified stress contributor for caregivers so far.
Csukly, Gábor; Czobor, Pál; Simon, Lajos; Takács, Barnabás
Cross-cultural studies have demonstrated universal similarity in the recognition and expression of basic emotions in facial expressions. The so-called mood congruency effect, observed primarily in clinical populations, implies that subjects with depressed mood tend to judge positive emotions as neutral and neutral faces as negative. The objective was to investigate whether a mood congruency effect can be detected in case of mild impairments among healthy subjects. First, it was hypothesized that subjects with mild psychiatric symptom distress have poorer performance in affective facial recognition in general. Second, it was also hypothesized that these subjects have poorer functioning in neutral face recognition and that they are prone to attribute negative emotions, for example, sadness and fear to neutral faces. Third, it was also assumed that people with mild psychiatric symptom distress have poor performance in recognizing positive emotions. Pictures representing the basic emotions were used to examine the recognition of facial emotions; the Symptom Checklist-90 was obtained to quantify overall psychological distress and the severity of psychiatric symptoms on 9 primary symptom dimensions, including somatization, obsessive-compulsive, interpersonal sensitivity, depression, anxiety, hostility, phobic anxiety, paranoid ideation, and psychoticism. One hundred seventeen healthy volunteers were recruited for the purpose of the study. Consistent with the first hypothesis, results indicated a significant negative association between the overall recognition rate of facial expressions and the level of psychiatric symptoms in a healthy population. Consistent with the second hypothesis, the level of psychiatric symptoms was related inversely with the neutral facial expression recognition and directly with the negative bias in neutral facial expressions. However, our data did not support the assumption that people with mild psychiatric symptom distress would have a poorer
Buttelmann, David; Tomasello, Michael
Although many studies have investigated domestic dogs' (Canis familiaris) use of human communicative cues, little is known about their use of humans' emotional expressions. We conducted a study following the general paradigm of Repacholi in Dev Psychol 34:1017-1025, (1998) and tested four breeds of dogs in the laboratory and another breed in the open air. In our study, a human reacted emotionally (happy, neutral or disgust) to the hidden contents of two boxes, after which the dog was then allowed to choose one of the boxes. Dogs tested in the laboratory distinguished between the most distinct of the expressed emotions (Happy-Disgust condition) by choosing appropriately, but performed at chance level when the two emotions were less distinct (Happy-Neutral condition). The breed tested in the open air passed both conditions, but this breed's differing testing setup might have been responsible for their success. Although without meaningful emotional expressions, when given a choice, these subjects chose randomly, their performance did not differ from that in the experimental conditions. Based on the findings revealed in the laboratory, we suggest that some domestic dogs recognize both the directedness and the valence of some human emotional expressions.
Faso, Daniel J.; Sasson, Noah J.; Pinkham, Amy E.
Though many studies have examined facial affect perception by individuals with autism spectrum disorder (ASD), little research has investigated how facial expressivity in ASD is perceived by others. Here, naïve female observers (n = 38) judged the intensity, naturalness and emotional category of expressions produced by adults with ASD (n = 6) and…
van Kleef, G.A.; Anastasopoulou, C.; Nijstad, B.A.
We investigated whether expressions of anger can enhance creative performance. Building on the emotions as social information (EASI) model (Van Kleef, 2009), we predicted that the interpersonal effects of anger expressions on creativity depend on the target's epistemic motivation (EM)—the desire to
Brechet, Claire; Jolley, Richard P.
The purpose of the present study was to investigate the roles of emotional comprehension and representational drawing skill in children's expressive drawing. Fifty 7- to 10-year-olds were asked to produce two (happy and sad) expressive drawings, two representational drawings (drawing of a man running and drawing of a house) and to answer the…
aan het Rot, Marije; Enea, Violeta; Dafinoiu, Ion; Iancu, Sorina; Taftă, Steluţa A; Bărbuşelu, Mariana
While the recognition of emotional expressions has been extensively studied, the behavioural response to these expressions has not. In the interpersonal circumplex, behaviour is defined in terms of communion and agency. In this study, we examined behavioural responses to both facial and postural
A.E.R. Bos (Arjan); A.J. Dijker (Anton); W. Koomen (Willem)
textabstractTwo studies examined the influence of HIV+ individual’s expression of distress on perceivers’ emotional and behavioral reactions. In Study 1 (N = 224), HIV+ individuals’ expression of distress was experimentally manipulated by means of vignettes. Men and women reacted differently when
Du, Shichuan; Martinez, Aleix M
Emotions are sometimes revealed through facial expressions. When these natural facial articulations involve the contraction of the same muscle groups in people of distinct cultural upbringings, this is taken as evidence of a biological origin of these emotions. While past research had identified facial expressions associated with a single internally felt category (eg, the facial expression of happiness when we feel joyful), we have recently studied facial expressions observed when people experience compound emotions (eg, the facial expression of happy surprise when we feel joyful in a surprised way, as, for example, at a surprise birthday party). Our research has identified 17 compound expressions consistently produced across cultures, suggesting that the number of facial expressions of emotion of biological origin is much larger than previously believed. The present paper provides an overview of these findings and shows evidence supporting the view that spontaneous expressions are produced using the same facial articulations previously identified in laboratory experiments. We also discuss the implications of our results in the study of psychopathologies, and consider several open research questions.
van Kleef, Gerben A.; Anastasopoulou, Christina; Nijstad, Bernard A.
We investigated whether expressions of anger can enhance creative performance. Building on the emotions as social information (EASI) model (Van Kleef, 2009), we predicted that the interpersonal effects of anger expressions on creativity depend on the target's epistemic motivation (EM) the desire to
Smith, Pam; Pearson, Pauline H; Ross, Fiona
This paper sets the discussion of emotions at work within the modern NHS and the current prioritisation of creating a safety culture within the service. The paper focuses on the work of students, frontline nurses and their managers drawing on recent studies of patient safety in the curriculum, and governance and incentives in the care of patients with complex long term conditions. The primary research featured in the paper combined a case study design with focus groups, interviews and observation. In the patient safety research the importance of physical and emotional safety emerged as a key finding both for users and professionals. In the governance and incentives research, risk emerged as a key concern for managers, frontline workers and users. The recognition of emotions and the importance of emotional labour at an individual and organizational level managed by emotionally intelligent leaders played an important role in promoting worker and patient safety and reducing workplace risk. Nurse managers need to be aware of the emotional complexities of their organizations in order to set up systems to support the emotional wellbeing of professionals and users which in turn ensures safety and reduces risk.
Full Text Available In order to overcome the limitation of single mode emotion recognition. This paper describes a novel multimodal emotion recognition algorithm, and takes speech signal and facial expression signal as the research subjects. First, fuse the speech signal feature and facial expression signal feature, get sample sets by putting back sampling, and then get classifiers by BP neural network (BPNN. Second, measure the difference between two classifiers by double error difference selection strategy. Finally, get the final recognition result by the majority voting rule. Experiments show the method improves the accuracy of emotion recognition by giving full play to the advantages of decision level fusion and feature level fusion, and makes the whole fusion process close to human emotion recognition more, with a recognition rate 90.4%.
Chapin, Heather; Jantzen, Kelly; Kelso, J A Scott; Steinberg, Fred; Large, Edward
Apart from its natural relevance to cognition, music provides a window into the intimate relationships between production, perception, experience, and emotion. Here, emotional responses and neural activity were observed as they evolved together with stimulus parameters over several minutes. Participants listened to a skilled music performance that included the natural fluctuations in timing and sound intensity that musicians use to evoke emotional responses. A mechanical performance of the same piece served as a control. Before and after fMRI scanning, participants reported real-time emotional responses on a 2-dimensional rating scale (arousal and valence) as they listened to each performance. During fMRI scanning, participants listened without reporting emotional responses. Limbic and paralimbic brain areas responded to the expressive dynamics of human music performance, and both emotion and reward related activations during music listening were dependent upon musical training. Moreover, dynamic changes in timing predicted ratings of emotional arousal, as well as real-time changes in neural activity. BOLD signal changes correlated with expressive timing fluctuations in cortical and subcortical motor areas consistent with pulse perception, and in a network consistent with the human mirror neuron system. These findings show that expressive music performance evokes emotion and reward related neural activations, and that music's affective impact on the brains of listeners is altered by musical training. Our observations are consistent with the idea that music performance evokes an emotional response through a form of empathy that is based, at least in part, on the perception of movement and on violations of pulse-based temporal expectancies.
Aho-Özhan, Helena E A; Keller, Jürgen; Heimrath, Johanna; Uttner, Ingo; Kassubek, Jan; Birbaumer, Niels; Ludolph, Albert C; Lulé, Dorothée
Amyotrophic lateral sclerosis (ALS) primarily impairs motor abilities but also affects cognition and emotional processing. We hypothesise that subjective ratings of emotional stimuli depicting social interactions and facial expressions is changed in ALS. It was found that recognition of negative emotions and ability to mentalize other's intentions is reduced. Processing of emotions in faces was investigated. A behavioural test of Ekman faces expressing six basic emotions was presented to 30 ALS patients and 29 age-, gender and education matched healthy controls. Additionally, a subgroup of 15 ALS patients that were able to lie supine in the scanner and 14 matched healthy controls viewed the Ekman faces during functional magnetic resonance imaging (fMRI). Affective state and a number of daily social contacts were measured. ALS patients recognized disgust and fear less accurately than healthy controls. In fMRI, reduced brain activity was seen in areas involved in processing of negative emotions replicating our previous results. During processing of sad faces, increased brain activity was seen in areas associated with social emotions in right inferior frontal gyrus and reduced activity in hippocampus bilaterally. No differences in brain activity were seen for any of the other emotional expressions. Inferior frontal gyrus activity for sad faces was associated with increased amount of social contacts of ALS patients. ALS patients showed decreased brain and behavioural responses in processing of disgust and fear and an altered brain response pattern for sadness. The negative consequences of neurodegenerative processes in the course of ALS might be counteracted by positive emotional activity and positive social interactions.
Full Text Available Apart from its natural relevance to cognition, music provides a window into the intimate relationships between production, perception, experience, and emotion. Here, emotional responses and neural activity were observed as they evolved together with stimulus parameters over several minutes. Participants listened to a skilled music performance that included the natural fluctuations in timing and sound intensity that musicians use to evoke emotional responses. A mechanical performance of the same piece served as a control. Before and after fMRI scanning, participants reported real-time emotional responses on a 2-dimensional rating scale (arousal and valence as they listened to each performance. During fMRI scanning, participants listened without reporting emotional responses. Limbic and paralimbic brain areas responded to the expressive dynamics of human music performance, and both emotion and reward related activations during music listening were dependent upon musical training. Moreover, dynamic changes in timing predicted ratings of emotional arousal, as well as real-time changes in neural activity. BOLD signal changes correlated with expressive timing fluctuations in cortical and subcortical motor areas consistent with pulse perception, and in a network consistent with the human mirror neuron system. These findings show that expressive music performance evokes emotion and reward related neural activations, and that music's affective impact on the brains of listeners is altered by musical training. Our observations are consistent with the idea that music performance evokes an emotional response through a form of empathy that is based, at least in part, on the perception of movement and on violations of pulse-based temporal expectancies.
Pioggia, Giovanni; Hanson, David; Dinelli, Serena; Di Francesco, Fabio; Francesconi, R.; De Rossi, Danilo
The nonverbal expression of the emotions, especially in the human face, has rapidly become an area of intense interest in computer science and robotics. Exploring the emotions as a link between external events and behavioural responses, artificial intelligence designers and psychologists are approaching a theoretical understanding of foundational principles which will be key to the physical embodiment of artificial intelligence. In fact, it has been well demonstrated that many important aspects of intelligence are grounded in intimate communication with the physical world- so-called embodied intelligence . It follows naturally, then, that recent advances in emotive artificial intelligence show clear and undeniable broadening in the capacities of biologically-inspired robots to survive and thrive in a social environment. The means by which AI may express its foundling emotions are clearly integral to such capacities. In effect: powerful facial expressions are critical to the development of intelligent, sociable robots. Following discussion the importance of the nonverbal expression of emotions in humans and robots, this paper describes methods used in robotically emulating nonverbal expressions using human-like robotic faces. Furthermore, it describes the potentially revolutionary impact of electroactive polymer (EAP) actuators as artificial muscles for such robotic devices.
Cheng, Feng; Wang, Yifang; Wu, Xixian; Su, Zhuqing
The current study shows that parental punitive discipline places children at risk of developing internalizing and externalizing problems. Although some studies have analyzed the reasons for the use of discipline methods, little to no research has analyzed the moderating effects. In this study, we examine the relationship between maternal negative emotional expression and mothers' use of disciplinary methods (psychological aggression, corporal punishment and physical maltreatment) and the moderating effects of educational attainment in Chinese societies. Five hundred and sixteen mothers with preschool-aged children were recruited to participate in this research. The Chinese versions of the Self-Expressiveness in the Family Questionnaire (SEFQ) and the Parent-Child Conflict Tactics Scales (CTSPC) were used to measure the mothers' negative emotional expression and discipline, respectively. The results suggested that the mothers' negative emotional expression was positively related to their disciplinary behaviors. Moreover, maternal educational attainment moderated the association between negative emotional expression and discipline. The findings of the current study highlight the importance of considering how mothers' educational backgrounds may interact with their emotions to influence maternal disciplinary behaviors. Copyright © 2017 Elsevier Ltd. All rights reserved.
Lu, Qian; Yeung, Nelson; Man, Jenny; Gallagher, Matthew W; Chu, Qiao; Deen, Sidra H
Posttraumatic stress symptoms (PTSS) are common among breast cancer survivors. However, the association and the underlying mediating mechanism between psychosocial factors and PTSS were rarely investigated among breast cancer survivors. Previous studies have suggested the importance of emotional expression in cancer survivors' PTSS. This study examined the association between ambivalence over emotional expression (AEE; defined as the conflict between the desire to express feelings and the fear of its consequences) and PTSS, and proposed intrusive thoughts as the mediators in such an association. We tested this proposed mediation model among Chinese breast cancer survivors whose culture discourages emotional expression. Participants were 118 Chinese-speaking breast cancer survivors in the USA, who were diagnosed with breast cancer of stages 0-III within the past 5 years. They completed questionnaires measuring their levels of AEE, PTSS, and intrusive thoughts. AEE was positively associated with intrusive thoughts (r = 0.43, p expression tend to have higher PTSS, and this may be partially due to the lack of opportunities to discuss emotional events, thereby increasing the repetitive cancer-related negative thoughts. Intervention for PTSS should consider helping cancer patients to develop adaptive emotional regulation strategies to reduce the detrimental effects of cancer-related intrusive thoughts.
Graf, Maria C; Gaudiano, Brandon A; Geller, Pamela A
The current study investigated the extent to which outpatient psychotherapy clients benefited from Pennebaker's expressive writing protocol (Pennebaker & Beall, 1986) adapted for use as a homework intervention. Participants were randomly assigned to written emotional disclosure or writing control conditions. Pre- and postintervention outcome measures were collected for three consecutive therapy sessions. Clients in the written emotional disclosure group showed significantly greater reductions in anxiety and depressive symptoms as well as greater overall progress in psychotherapy in comparison to the writing control group. Results suggest that emotional disclosure writing homework, in conjunction with outpatient psychotherapy, facilitates therapeutic process and outcome.
Tuck, Natalie L; Adams, Kathryn S; Pressman, Sarah D; Consedine, Nathan S
Positive emotion is associated with lower cardiovascular disease (CVD) risk, yet some mechanisms remain unclear. One potential pathway is via emotional competencies/skills. The present study tests whether the ability to facially express positive emotion is associated with CVD risk scores, while controlling for potential confounds and testing for sex moderation. Eighty-two men and women underwent blood draws before completing self-report assessments and a performance test of expressive skill. Positive expressions were scored for degree of 'happiness' using expression coding software. CVD risk scores were calculated using established algorithms based on biological, demographic, and behavioral risk factors. Linear regressions revealed a main effect for skill, with skill in expressing positive emotion associated with lower CVD risk scores. Analyses also revealed a sex-by-skill interaction whereby links between expressive skill and CVD risk scores were stronger among men. Objective tests of expressive skill have methodological advantages, appear to have links to physical health, and offer a novel avenue for research and intervention.
Full Text Available In his seminal paper, Gabrielsson (2002 distinguishes between emotion felt by the listener, here: ‘internal locus of emotion’ (IL, and the emotion the music is expressing, here: 'external locus emotion' (EL. This paper tabulates 16 such publications published in the decade 2003-2012 consisting of 19 studies/experiments and provides some theoretical perspectives. The key findings were that (1 IL ratings was frequently rated statistically the same or lower than the corresponding EL rating (e.g. lower felt happiness rating compared to the apparent happiness of the music, and that (2 self-select and preferred music had a smaller gap across the emotion loci than experimenter selected and disliked music. These key findings were explained by an ‘inhibited’ emotional contagion mechanism, where the otherwise matching felt emotion may have been attenuated by some other factor such as social context. Matching between EL and IL for loved and self-selected pieces was explained by the activation of ‘contagion’ circuits. Physiological arousal, personality and age, as well as musical features (tempo, mode, putative emotions were observed to influence perceived and felt emotion distinctions. A variety of data collection formats were identified, but mostly using continuous rating scales. In conclusion, a more systematic use of terminology appears desirable with respect to theory-building. Whether two broad categories, namely matched and unmatched, are sufficient to capture the relationships between EL and IL, instead of four categories as suggested by Gabrielsson, is subject to future research.
McRae, Kateri; Rhee, Soo Hyun; Gatt, Justine M; Godinez, Detre; Williams, Leanne M; Gross, James J
Previous studies have established that personality traits related to emotionality are moderately heritable. However, the relative heritability of the strategies people use to regulate emotions is unknown. The present study compared the magnitude of additive genetic, shared environmental, and nonshared environmental influences on 2 commonly used emotion regulation strategies: cognitive reappraisal and expressive suppression. In 743 twin pairs (1,486 twins), we replicated previous estimates of heritability of neuroticism (a2 = .41). Furthermore, cognitive reappraisal was significantly less heritable and more influenced by nonshared environment (a2 = .20; e2 = .80) than either neuroticism or suppression (a2 = .35; e2 = .65), another emotion regulation strategy. Finally, Cholesky decomposition modeling suggested that while there were common genetic and environmental influences on neuroticism, reappraisal and suppression, there were also significant nonshared environmental influences common between reappraisal and adaptive emotional functioning after controlling for neuroticism and suppression. These findings highlight that different aspects of emotional processing, even the use of different emotion regulation strategies, are differentially heritable. The importance of the nonshared environmental influences specific to reappraisal and adaptive emotional functioning speaks to the potential impact of social context, social partners, and psychosocial interventions on reappraisal habits. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Görlach, Mirja Gianna; Kohlmann, Sebastian; Shedden-Mora, Meike; Rief, Winfried; Westermann, Stefan
Emotions have a considerable impact on eating behaviour; however, research addressing emotion regulation in obesity is rare. The present study is the first to investigate the association between emotional suppression and overeating in individuals with overweight. In total, 314 participants including 190 individuals with obesity filled in a cross-sectional online survey, which assessed emotional suppression, eating behaviour and psychopathology. A hierarchical linear regression analysis was conducted to identify factors associated with overeating. Individuals with obesity reported more frequent overeating compared with individuals without obesity. The habitual use of emotional suppression was associated with more overeating; however, this link was moderated by increased body mass index (BMI). The results suggest that suppression of emotional expression contributes to overeating and is maladaptive especially in individuals with obesity. Further research should longitudinally investigate the predictive value of emotional suppression on overweight, as the training of emotion regulation could contribute to treating obesity. Copyright © 2016 John Wiley & Sons, Ltd and Eating Disorders Association. Copyright © 2016 John Wiley & Sons, Ltd and Eating Disorders Association.
Full Text Available This research aims to contribute to the literature on the ability to recognize anger, happiness, fear, surprise, sadness, disgust and neutral emotions from facial information. By investigating children’s performance in detecting these emotions from a specific face region, we were interested to know whether children would show differences in recognizing these expressions from the upper or lower face, and if any difference between specific facial regions depended on the emotion in question. For this purpose, a group of 6-7 year-old children was selected. Participants were asked to recognize emotions by using a labeling task with three stimulus types (region of the eyes, of the mouth, and full face. The findings seem to indicate that children correctly recognize basic facial expressions when pictures represent the whole face, except for a neutral expression, which was recognized from the mouth, and sadness, which was recognized from the eyes. Children are also able to identify anger from the eyes as well as from the whole face. With respect to gender differences, there is no female advantage in emotional recognition. The results indicate a significant interaction ‘gender x face region’ only for anger and neutral emotions.
CARLOS FELIPE PARDO-VÉLEZ
Full Text Available Gender differences in the recognition of facial expressions of anger, happiness and sadness wereresearched in students 18-25 years of age. A reaction time procedure was used, and the percentage ofcorrect answers when recognizing was also measured. Though the work hypothesis expected genderdifferences in facial expression recognition, results suggest that these differences are not significant at alevel of 0.05%. Statistical analysis shows a greater easiness (at a non-significant level for women torecognize happiness expressions, and for men to recognize anger expressions. The implications ofthese data are discussed, and possible extensions of this investigation in terms of sample size andcollege major of the participants.
Mjaaland, Trond A; Finset, Arnstein; Jensen, Bård Fossli; Gulbrandsen, Pål
Patients express their negative emotions in medical consultations either implicitly as cue to an underlying unpleasant emotion or explicitly as a clear, unambiguous concern. The health provider's response to such cues and concerns is important for the outcome of consultations. Yet, physicians often neglect patient's negative emotions. Most studies of this subject are from primary health care. We aimed to describe how physicians in a hospital respond to negative emotions in an outpatient setting. Ninety six consultations were videotaped in a general teaching hospital. The Verona Coding Definitions of Emotional Sequences was used to identify patients' expression of negative emotions in terms of cue and concern and to code physicians' subsequent responses. Cohen's kappa was used as interrater reliability measure. Acceptable kappa level was set to .60. We observed 163 expressions of negative emotions. In general, the physician responses to patients' cues and concerns did not include follow up or exploration. Concerns more often than cues led to lack of emotional exploration. When patients expressed negative emotions or cues to such, hospital physicians tended to move away from emotional communication, particularly if the emotion was expressed as an explicit concern. Medical training should enable physicians' to explore the patients' emotions in situations where it will improve the medical treatment. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.
Giordano, Bruno L; Egermann, Hauke; Bresin, Roberto
Several studies have investigated the encoding and perception of emotional expressivity in music performance. A relevant question concerns how the ability to communicate emotions in music performance is acquired. In accordance with recent theories on the embodiment of emotion, we suggest here that both the expression and recognition of emotion in music might at least in part rely on knowledge about the sounds of expressive body movements. We test this hypothesis by drawing parallels between musical expression of emotions and expression of emotions in sounds associated with a non-musical motor activity: walking. In a combined production-perception design, two experiments were conducted, and expressive acoustical features were compared across modalities. An initial performance experiment tested for similar feature use in walking sounds and music performance, and revealed that strong similarities exist. Features related to sound intensity, tempo and tempo regularity were identified as been used similarly in both domains. Participants in a subsequent perception experiment were able to recognize both non-emotional and emotional properties of the sound-generating walkers. An analysis of the acoustical correlates of behavioral data revealed that variations in sound intensity, tempo, and tempo regularity were likely used to recognize expressed emotions. Taken together, these results lend support the motor origin hypothesis for the musical expression of emotions.
Bruno L Giordano
Full Text Available Several studies have investigated the encoding and perception of emotional expressivity in music performance. A relevant question concerns how the ability to communicate emotions in music performance is acquired. In accordance with recent theories on the embodiment of emotion, we suggest here that both the expression and recognition of emotion in music might at least in part rely on knowledge about the sounds of expressive body movements. We test this hypothesis by drawing parallels between musical expression of emotions and expression of emotions in sounds associated with a non-musical motor activity: walking. In a combined production-perception design, two experiments were conducted, and expressive acoustical features were compared across modalities. An initial performance experiment tested for similar feature use in walking sounds and music performance, and revealed that strong similarities exist. Features related to sound intensity, tempo and tempo regularity were identified as been used similarly in both domains. Participants in a subsequent perception experiment were able to recognize both non-emotional and emotional properties of the sound-generating walkers. An analysis of the acoustical correlates of behavioral data revealed that variations in sound intensity, tempo, and tempo regularity were likely used to recognize expressed emotions. Taken together, these results lend support the motor origin hypothesis for the musical expression of emotions.
Giordano, Bruno L.; Egermann, Hauke; Bresin, Roberto
Several studies have investigated the encoding and perception of emotional expressivity in music performance. A relevant question concerns how the ability to communicate emotions in music performance is acquired. In accordance with recent theories on the embodiment of emotion, we suggest here that both the expression and recognition of emotion in music might at least in part rely on knowledge about the sounds of expressive body movements. We test this hypothesis by drawing parallels between musical expression of emotions and expression of emotions in sounds associated with a non-musical motor activity: walking. In a combined production-perception design, two experiments were conducted, and expressive acoustical features were compared across modalities. An initial performance experiment tested for similar feature use in walking sounds and music performance, and revealed that strong similarities exist. Features related to sound intensity, tempo and tempo regularity were identified as been used similarly in both domains. Participants in a subsequent perception experiment were able to recognize both non-emotional and emotional properties of the sound-generating walkers. An analysis of the acoustical correlates of behavioral data revealed that variations in sound intensity, tempo, and tempo regularity were likely used to recognize expressed emotions. Taken together, these results lend support the motor origin hypothesis for the musical expression of emotions. PMID:25551392
Canamero, Dolores; Fredslund, Jacob
We report work on a LEGO robot that displays different emotional expressions in response to physical stimulation, for the purpose of social interaction with humans. This is a first step toward our longer-term goal of exploring believable emotional exchanges to achieve plausible interaction...... with a simple robot. Drawing inspiration from theories of human basic emotions, we implemented several prototypical expressions in the robot's caricatured face and conducted experiments to assess the recognizability of these expressions...
Schneevogt, Daniela; Paggio, Patrizia
Recent studies have demonstrated gender and cultural differences in the recognition of emotions in facial expressions. However, most studies were conducted on American subjects. In this pa- per, we explore the generalizability of several findings to a non-American culture in the form of Danish...... subjects. We conduct an emotion recognition task followed by two stereotype question- naires with different genders and age groups. While recent findings (Krems et al., 2015) suggest that women are biased to see anger in neutral facial expressions posed by females, in our sample both genders assign higher...... ratings of anger to all emotions expressed by females. Furthermore, we demonstrate an effect of gender on the fear-surprise-confusion observed by Tomkins and McCarter (1964); females overpredict fear, while males overpredict surprise....
Cañadas, Elena; Lupiáñez, Juan; Kawakami, Kerry; Niedenthal, Paula M; Rodríguez-Bailón, Rosa
Individuals spontaneously categorise other people on the basis of their gender, ethnicity and age. But what about the emotions they express? In two studies we tested the hypothesis that facial expressions are similar to other social categories in that they can function as contextual cues to control attention. In Experiment 1 we associated expressions of anger and happiness with specific proportions of congruent/incongruent flanker trials. We also created consistent and inconsistent category members within each of these two general contexts. The results demonstrated that participants exhibited a larger congruency effect when presented with faces in the emotional group associated with a high proportion of congruent trials. Notably, this effect transferred to inconsistent members of the group. In Experiment 2 we replicated the effects with faces depicting true and false smiles. Together these findings provide consistent evidence that individuals spontaneously utilise emotions to categorise others and that such categories determine the allocation of attentional control.
Maran, Thomas; Sachse, Pierre; Furtner, Marco
Previous findings suggest that visual working memory (VWM) preferentially remembers angry looking faces. However, the meaning of facial actions is construed in relation to context. To date, there are no studies investigating the role of perceiver-based context when processing emotional cues in VWM. To explore the influence of affective context on VWM for faces, we conducted two experiments using both a VWM task for emotionally expressive faces and a mood induction procedure. Affective context was manipulated by unpleasant (Experiment 1) and pleasant (Experiment 2) IAPS pictures in order to induce an affect high in motivational intensity (defensive or appetitive, respectively) compared to a low arousal control condition. Results indicated specifically increased sensitivity of VWM for angry looking faces in the neutral condition. Enhanced VWM for angry faces was prevented by inducing affects of high motivational intensity. In both experiments, affective states led to a switch from specific enhancement of angry expressions in VWM to an equally sensitive representation of all emotional expressions. Our findings demonstrate that emotional expressions are of different behavioral relevance for the receiver depending on the affective context, supporting a functional organization of VWM along with flexible resource allocation. In VWM, stimulus processing adjusts to situational requirements and transitions from a specifically prioritizing default mode in predictable environments to a sensitive, hypervigilant mode in exposure to emotional events.
Full Text Available Previous findings suggest that visual working memory preferentially remembers angry looking faces. However, the meaning of facial actions is construed in relation to context. To date, there are no studies investigating the role of perceiver-based context when processing emotional cues in visual working memory. To explore the influence of affective context on visual working memory for faces, we conducted two experiments using both a visual working memory task for emotionally expressive faces and a mood induction procedure. Affective context was manipulated by unpleasant (Experiment 1 and pleasant (Experiment 2 IAPS pictures in order to induce an affect high in motivational intensity (defensive or appetitive, respectively compared to a low arousal control condition. Results indicated specifically increased sensitivity of visual working memory for angry looking faces in the neutral condition. Enhanced visual working memory for angry faces was prevented by inducing affects of high motivational intensity. In both experiments, affective states led to a switch from specific enhancement of angry expressions in visual working memory to an equally sensitive representation of all emotional expressions. Our findings demonstrate that emotional expressions are of different behavioral relevance for the receiver depending on the affective context, supporting a functional organization of visual working memory along with flexible resource allocation. In visual working memory, stimulus processing adjusts to situational requirements and transitions from a specifically prioritizing default mode in predictable environments to a sensitive, hypervigilant mode in exposure to emotional events.
Saha, Priya; Bhattacharjee, Debotosh; De, Barin K.; Nasipuri, Mita
Facial expressions in infrared imaging have been introduced to solve the problem of illumination, which is an integral constituent of visual imagery. The paper investigates facial skin temperature distribution on mixed thermal facial expressions of our created face database where six are basic expressions and rest 12 are a mixture of those basic expressions. Temperature analysis has been performed on three facial regions of interest (ROIs); periorbital, supraorbital and mouth. Temperature variability of the ROIs in different expressions has been measured using statistical parameters. The temperature variation measurement in ROIs of a particular expression corresponds to a vector, which is later used in recognition of mixed facial expressions. Investigations show that facial features in mixed facial expressions can be characterized by positive emotion induced facial features and negative emotion induced facial features. Supraorbital is a useful facial region that can differentiate basic expressions from mixed expressions. Analysis and interpretation of mixed expressions have been conducted with the help of box and whisker plot. Facial region containing mixture of two expressions is generally less temperature inducing than corresponding facial region containing basic expressions.
Jack, Rachael E; Caldara, Roberto; Schyns, Philippe G
Facial expressions have long been considered the "universal language of emotion." Yet consistent cultural differences in the recognition of facial expressions contradict such notions (e.g., R. E. Jack, C. Blais, C. Scheepers, P. G. Schyns, & R. Caldara, 2009). Rather, culture--as an intricate system of social concepts and beliefs--could generate different expectations (i.e., internal representations) of facial expression signals. To investigate, they used a powerful psychophysical technique (reverse correlation) to estimate the observer-specific internal representations of the 6 basic facial expressions of emotion (i.e., happy, surprise, fear, disgust, anger, and sad) in two culturally distinct groups (i.e., Western Caucasian [WC] and East Asian [EA]). Using complementary statistical image analyses, cultural specificity was directly revealed in these representations. Specifically, whereas WC internal representations predominantly featured the eyebrows and mouth, EA internal representations showed a preference for expressive information in the eye region. Closer inspection of the EA observer preference revealed a surprising feature: changes of gaze direction, shown primarily among the EA group. For the first time, it is revealed directly that culture can finely shape the internal representations of common facial expressions of emotion, challenging notions of a biologically hardwired "universal language of emotion."
Ono, Makiko; Fujita, Mizuho; Yamada, Shigeyuki
The effects of communicating during and after expressing emotions and receiving empathy after exposure to stress were investigated for 18 female students (9 pairs). After mental and physical tasks, a subject spoke to a listener about the stress task. In Experiment 1, responses to speaking about negative emotions aroused by the task (the "with emotion" condition) were compared to speaking about only objective facts about the task (the control). In Experiment 2, responses to empathetic reactions from the listener (the "with empathy" condition) were compared to no reaction (the control). Electroencephalograms were recorded, and heart rate variability (HRV) was calculated from electrocardiogram data. Subjective stress was estimated by a visual analog scale. Experiment 1 demonstrated that expressing emotions activated the left temporal region (T3) in the "with emotion" condition. In Experiment 2, physiological responses depended on cognition of different elements of empathy. During communication, feeling that the listener had the same emotion decreased the subject's T3 activity and sympathetic activity balance indicated by HRV. After communication, feeling that the listener understood her emotions decreased bilateral frontal and temporal activity. On the other hand, subjective stress did not differ between conditions in both experiments. These findings indicate that the comfort of having shared a message reduced physiological activity, especially in the "with empathy" condition. Conversely, even in the "with empathy" condition, not sharing a message can result in more discomfort or stress than the control. Sharing might be associated with cognition of the degree of success of communication, which reflected in the physiological responses. In communication, therefore, expressing emotions and receiving empathy did not in themselves reduce stress, and the level of cognition of having shared a message is a key factor in reducing stress.
Kristensen, Liv Kondrup; Otrel-Cass, Kathrin
Observing science classroom activities presents an opportunity to observe the emotional aspect of interactions, and this chapter presents how this can be done and why. Drawing on ideas proposed by French philosopher Maurice Merleau-Ponty, emotions are theorized as publicly embodied enactments......, where differences in behavior between people shape emotional responses. Merleau-Ponty’s theorization of the body and feelings is connected to embodiment while examining central concepts such as consciousness and perception. Merleau-Ponty describes what he calls the emotional atmosphere and how it shapes...... the ways we experience events and activities. We use our interpretation of his understanding of emotions to examine an example of a group of year 8 science students who were engaged in a physics activity. Using the analytical framework of analyzing bodily stance by Goodwin, Cekaite, and Goodwin...
Han, Kyung-Hun; Zaytseva, Yuliya; Bao, Yan; Pöppel, Ernst; Chung, Sun Yong; Kim, Jong Woo; Kim, Hyun Taek
Vocal expression of emotions (EE) in retrieval of events from autobiographical memory was investigated in patients in early stages of Alzheimer's disease (AD). Twenty-one AD patients and 19 controls were interviewed, and EE of the reported memories was rated by 8 independent evaluators. The AD group had lower EE of both recent and remote memory than controls, although EE in remote memories was better preserved in both groups. We observed positive correlations between EE and indicators of cognitive competence in AD patients. AD Patients are impaired in the ability to express emotions already at early stages of the disease, and EE seems to deteriorate along with the progression of cognitive impairment.
Sawyer, Alyssa C. P.; Williamson, Paul; Young, Robyn L.
Research has shown that individuals with Autism Spectrum Disorders (ASD) have difficulties recognising emotions from facial expressions. Since eye contact is important for accurate emotion recognition, and individuals with ASD tend to avoid eye contact, this tendency for gaze aversion has been proposed as an explanation for the emotion recognition…
Wingenbach, Tanja S H; Ashwin, Chris; Brosnan, Mark
Most of the existing sets of facial expressions of emotion contain static photographs. While increasing demand for stimuli with enhanced ecological validity in facial emotion recognition research has led to the development of video stimuli, these typically involve full-blown (apex) expressions. However, variations of intensity in emotional facial expressions occur in real life social interactions, with low intensity expressions of emotions frequently occurring. The current study therefore developed and validated a set of video stimuli portraying three levels of intensity of emotional expressions, from low to high intensity. The videos were adapted from the Amsterdam Dynamic Facial Expression Set (ADFES) and termed the Bath Intensity Variations (ADFES-BIV). A healthy sample of 92 people recruited from the University of Bath community (41 male, 51 female) completed a facial emotion recognition task including expressions of 6 basic emotions (anger, happiness, disgust, fear, surprise, sadness) and 3 complex emotions (contempt, embarrassment, pride) that were expressed at three different intensities of expression and neutral. Accuracy scores (raw and unbiased (Hu) hit rates) were calculated, as well as response times. Accuracy rates above chance level of responding were found for all emotion categories, producing an overall raw hit rate of 69% for the ADFES-BIV. The three intensity levels were validated as distinct categories, with higher accuracies and faster responses to high intensity expressions than intermediate intensity expressions, which had higher accuracies and faster responses than low intensity expressions. To further validate the intensities, a second study with standardised display times was conducted replicating this pattern. The ADFES-BIV has greater ecological validity than many other emotion stimulus sets and allows for versatile applications in emotion research. It can be retrieved free of charge for research purposes from the corresponding author.
Wingenbach, Tanja S. H.
Most of the existing sets of facial expressions of emotion contain static photographs. While increasing demand for stimuli with enhanced ecological validity in facial emotion recognition research has led to the development of video stimuli, these typically involve full-blown (apex) expressions. However, variations of intensity in emotional facial expressions occur in real life social interactions, with low intensity expressions of emotions frequently occurring. The current study therefore developed and validated a set of video stimuli portraying three levels of intensity of emotional expressions, from low to high intensity. The videos were adapted from the Amsterdam Dynamic Facial Expression Set (ADFES) and termed the Bath Intensity Variations (ADFES-BIV). A healthy sample of 92 people recruited from the University of Bath community (41 male, 51 female) completed a facial emotion recognition task including expressions of 6 basic emotions (anger, happiness, disgust, fear, surprise, sadness) and 3 complex emotions (contempt, embarrassment, pride) that were expressed at three different intensities of expression and neutral. Accuracy scores (raw and unbiased (Hu) hit rates) were calculated, as well as response times. Accuracy rates above chance level of responding were found for all emotion categories, producing an overall raw hit rate of 69% for the ADFES-BIV. The three intensity levels were validated as distinct categories, with higher accuracies and faster responses to high intensity expressions than intermediate intensity expressions, which had higher accuracies and faster responses than low intensity expressions. To further validate the intensities, a second study with standardised display times was conducted replicating this pattern. The ADFES-BIV has greater ecological validity than many other emotion stimulus sets and allows for versatile applications in emotion research. It can be retrieved free of charge for research purposes from the corresponding author
Tanja S H Wingenbach
Full Text Available Most of the existing sets of facial expressions of emotion contain static photographs. While increasing demand for stimuli with enhanced ecological validity in facial emotion recognition research has led to the development of video stimuli, these typically involve full-blown (apex expressions. However, variations of intensity in emotional facial expressions occur in real life social interactions, with low intensity expressions of emotions frequently occurring. The current study therefore developed and validated a set of video stimuli portraying three levels of intensity of emotional expressions, from low to high intensity. The videos were adapted from the Amsterdam Dynamic Facial Expression Set (ADFES and termed the Bath Intensity Variations (ADFES-BIV. A healthy sample of 92 people recruited from the University of Bath community (41 male, 51 female completed a facial emotion recognition task including expressions of 6 basic emotions (anger, happiness, disgust, fear, surprise, sadness and 3 complex emotions (contempt, embarrassment, pride that were expressed at three different intensities of expression and neutral. Accuracy scores (raw and unbiased (Hu hit rates were calculated, as well as response times. Accuracy rates above chance level of responding were found for all emotion categories, producing an overall raw hit rate of 69% for the ADFES-BIV. The three intensity levels were validated as distinct categories, with higher accuracies and faster responses to high intensity expressions than intermediate intensity expressions, which had higher accuracies and faster responses than low intensity expressions. To further validate the intensities, a second study with standardised display times was conducted replicating this pattern. The ADFES-BIV has greater ecological validity than many other emotion stimulus sets and allows for versatile applications in emotion research. It can be retrieved free of charge for research purposes from the
Sandberg, Kristian; Bibby, Bo Martin; Overgaard, Morten
Comparison of behavioural measures of consciousness has attracted much attention recently. In a recent article, Szczepanowski et al. conclude that confidence ratings (CR) predict accuracy better than both the perceptual awareness scale (PAS) and post-decision wagering (PDW) when using stimuli with emotional content (fearful vs. neutral faces). Although we find the study interesting, we disagree with the conclusion that CR is superior to PAS because of two methodological issues. First, the conclusion is not based on a formal test. We performed this test and found no evidence that CR predicted accuracy better than PAS (p=.4). Second, Szczepanowski et al. used the present version of PAS in a manner somewhat different from how it was originally intended, and the participants may not have been adequately instructed. We end our commentary with a set of recommendations for future studies using PAS. Copyright © 2013 Elsevier Inc. All rights reserved.
van Osch, Yvette; Zeelenberg, Marcel; Breugelmans, Seger M
In spite of various claims for cross-cultural differences in the experience of pride, studies on the expression of pride have revealed few cross-cultural differences. Five studies using archival data from Olympic and national championships do show cross-cultural differences in the expression of pride and other positive emotions in pride-eliciting contexts, contingent on the social context of the expression, notably the in-group or out-group status of the audience. Chinese gold medalists were perceived to express less pride than American medalists when outperforming in-group competitors; when outperforming out-group members, however, no or smaller cross-cultural differences were observed. These findings are important because they indicate that cultural norms about emotion expression may be activated only in situations in which they serve a function in coordinating people's behaviour.
Durán, Auxiliadora; Extremera, Natalio; Rey, Lourdes
This study examined the relationship among dimensions of self-reported Emotional Intelligence, Engagement and Burnout, using the Trait Meta-Mood Scale, Maslach Burnout Inventory and Utrecht Work Engagement Scale in a sample of Spanish professionals who work at institutions for people with intellectual disabilities. The results showed that Emotional Clarity was significantly associated with Personal Accomplishment (r=.25) and Dedication (r=.25). Further, Repair to moods was significantly correlated with all Engagement dimensions (.20 Vigor, .30 Dedication, .36 Absorption) and with Personal Accomplishment (.31). These findings extend previous research with college students in which Clarity and Repair to moods subscales were relevant predictors of well-being indexes and interpersonal functioning and suggest that the Trait Meta-Mood Scale subscales also show significant relationships with emotional functioning and work-related variables in a professional sample.
Full Text Available Neuropsychological studies have underlined the significant presence of distinct brain correlates deputed to analyze facial expression of emotion. It was observed that some cerebral circuits were considered as specific for emotional face comprehension as a function of conscious vs. unconscious processing of emotional information. Moreover, the emotional content of faces (i.e. positive vs. negative; more or less arousing may have an effect in activating specific cortical networks. Between the others, recent studies have explained the contribution of hemispheres in comprehending face, as a function of type of emotions (mainly related to the distinction positive vs. negative and of specific tasks (comprehending vs. producing facial expressions. Specifically, ERPs (event-related potentials analysis overview is proposed in order to comprehend how face may be processed by an observer and how he can make face a meaningful construct even in absence of awareness. Finally, brain oscillations is considered in order to explain the synchronization of neural populations in response to emotional faces when a conscious vs. unconscious processing is activated
Full Text Available The production of facial expressions (FEs is an important skill that allows children to share and adapt emotions with their relatives and peers during social interactions. These skills are impaired in children with Autism Spectrum Disorder. However, the way in which typical children develop and master their production of FEs has still not been clearly assessed. This study aimed to explore factors that could influence the production of FEs in childhood such as age, gender, emotion subtype (sadness, anger, joy, and neutral, elicitation task (on request, imitation, area of recruitment (French Riviera and Parisian and emotion multimodality. A total of one hundred fifty-seven children aged 6–11 years were enrolled in Nice and Paris, France. We asked them to produce FEs in two different tasks: imitation with an avatar model and production on request without a model. Results from a multivariate analysis revealed that: (1 children performed better with age. (2 Positive emotions were easier to produce than negative emotions. (3 Children produced better FE on request (as opposed to imitation; and (4 Riviera children performed better than Parisian children suggesting regional influences on emotion production. We conclude that facial emotion production is a complex developmental process influenced by several factors that needs to be acknowledged in future research.
Moed, Anat; Dix, Theodore; Anderson, Edward R; Greene, Shannon M
Research is unclear about when expressing negative emotions to children performs valuable socialization and regulatory functions and when, instead, it undermines children's adjustment. In this study, we isolated 1 kind of negative expression to test the aversion sensitivity hypothesis: that rapid increases in mothers' negativity as a function of increases in the aversiveness of children's behavior are uniquely problematic for children. During multiple assessments of a divorcing sample over 2 years (N = 284), 12-min interactions between mothers and their 4- to 11-year-old children were recorded. Forty-seven observed child behaviors were ranked from low to high aversive. Within-dyad changes demonstrated that mothers' general negativity-their tendency to express negative emotion at high rates-was unrelated to children's adjustment. In contrast, mothers' aversion-focused negativity-their tendency to increase negative emotional expression rapidly as the aversiveness of children's behavior increased-predicted children's poor social competence, poor emotion regulation, and externalizing behavior problems at the next assessment. The findings suggest that negative expression that reflects mothers' affective sensitivity to aversive child behavior may promote interaction patterns and adaptations in children that are particularly likely to place children at risk for adjustment problems. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Full Text Available Infant emotional expressions, such as distress cries, evoke maternal physiological reactions. Most of which involve accelerated sympathetic nervous activity. Comparatively little is known about effects of positive infant expressions, such as happy smiles, on maternal physiological responses. This study investigated how physiological and psychological maternal states change in response to infants' emotional expressions. Thirty first-time mothers viewed films of their own 6- to 7-month-old infants' affective behavior. Each observed a video of a distress cry followed by a video showing one of two expressions (randomly assigned: a happy smiling face (smile condition or a calm neutral face (neutral condition. Both before and after the session, participants completed a self-report inventory assessing their emotional states. The results of the self-report inventory revealed no effects of exposure to the infant videos. However, the mothers in the smile condition, but not in the neutral condition, showed deceleration of skin conductance. These findings demonstrate that the mothers who observed their infants smiling showed decreased sympathetic activity. We propose that an infant's positive emotional expression may affect the branch of the maternal stress-response system that modulates the homeostatic balance of the sympathetic and parasympathetic nervous systems.
Pamela M. Pallett
Full Text Available To distinguish between high-level visual processing mechanisms, the degree to which holistic processing is involved in facial identity, facial expression, and object perception is often examined through measuring inversion effects. However, participants may be biased by different experimental paradigms to use more or less holistic processing. Here we take a novel psychophysical approach to directly compare human face and object processing in the same experiment, with face processing broken into two categories: variant properties and invariant properties as they were tested using facial expressions of emotion and gender, respectively. Specifically, participants completed two different perceptual discrimination tasks. One involved making judgments of stimulus similarity and the other tested the ability to detect differences between stimuli. Each task was completed for both upright and inverted stimuli. Results show significant inversion effects for the detection of differences in facial expressions of emotion and gender, but not for objects. More interestingly, participants exhibited a selective inversion deficit when making similarity judgments between different facial expressions of emotion, but not for gender or objects. These results suggest a 3-way dissociation between facial expression of emotion, gender, and object processing.
Alonso-Recio, Laura; Martín, Pilar; Rubio, Sandra; Serrano, Juan M
Our objective was to compare the ability to discriminate and categorize emotional facial expressions (EFEs) and facial identity characteristics (age and/or gender) in a group of 53 individuals with Parkinson's disease (PD) and another group of 53 healthy subjects. On the one hand, by means of discrimination and identification tasks, we compared two stages in the visual recognition process that could be selectively affected in individuals with PD. On the other hand, facial expression versus gender and age comparison permits us to contrast whether the emotional or non-emotional content influences the configural perception of faces. In Experiment I, we did not find differences between groups, either with facial expression or age, in discrimination tasks. Conversely, in Experiment II, we found differences between the groups, but only in the EFE identification task. Taken together, our results indicate that configural perception of faces does not seem to be globally impaired in PD. However, this ability is selectively altered when the categorization of emotional faces is required. A deeper assessment of the PD group indicated that decline in facial expression categorization is more evident in a subgroup of patients with higher global impairment (motor and cognitive). Taken together, these results suggest that the problems found in facial expression recognition may be associated with the progressive neuronal loss in frontostriatal and mesolimbic circuits, which characterizes PD. © 2013 The British Psychological Society.
Schouten, Barbara C; Schinkel, Sanne
This study compared native-Dutch and Turkish-Dutch patients' expressions of emotional cues/concerns and GPs' responses to these cues/concerns. Relations between patient's cues/concerns and GPs' perceptions of the patient's health complaint were examined too. 82 audiotaped encounters with native-Dutch and 38 with Turkish-Dutch GP patients were coded using the VR-CoDES and VR-CoDES-P. Patients filled out a survey before each consultation to assess their cultural identification, Dutch language proficiency and health-related variables. GPs filled out a survey after each consultation to assess their perceptions of the patient's health complaint. Turkish-Dutch patients expressed more cues than native-Dutch patients, which was explained by higher worries about their health and worse perceived general health. GPs responded more often with space-providing responses to Turkish-Dutch patients compared to native-Dutch patients. Turkish-Dutch patients' cue expression strongly influenced GPs' perceptions about the presence of psychosocial problems. Migrant patient-related factors influence the amount of emotional cue expression in primary care. GPs perceive these cues as indicating the presence of psychosocial problems and provide space for patients to elaborate on their emotional distress. GPs should be trained in using more affective communication techniques to enhance elicitation of the underlying reasons for migrant patients' enhanced emotional cue expression. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Full Text Available Emotional facial expressions are immediate indicators of the affective dispositions of others. Recently it has been shown that early stages of social perception can already be influenced by (implicit attributions made by the observer about the agent's mental state and intentions. In the current study possible mechanisms underpinning distortions in the perception of dynamic, ecologically-valid, facial expressions were explored. In four experiments we examined to what extent basic perceptual processes such as contrast/context effects, adaptation and representational momentum underpinned the perceptual distortions, and to what extent 'emotional anticipation', i.e. the involuntary anticipation of the other's emotional state of mind on the basis of the immediate perceptual history, might have played a role. Neutral facial expressions displayed at the end of short video-clips, in which an initial facial expression of joy or anger gradually morphed into a neutral expression, were misjudged as being slightly angry or slightly happy, respectively (Experiment 1. This response bias disappeared when the actor's identity changed in the final neutral expression (Experiment 2. Videos depicting neutral-to-joy-to-neutral and neutral-to-anger-to-neutral sequences again produced biases but in opposite direction (Experiment 3. The bias survived insertion of a 400 ms blank (Experiment 4. These results suggested that the perceptual distortions were not caused by any of the low-level perceptual mechanisms (adaptation, representational momentum and contrast effects. We speculate that especially when presented with dynamic, facial expressions, perceptual distortions occur that reflect 'emotional anticipation' (a low-level mindreading mechanism, which overrules low-level visual mechanisms. Underpinning neural mechanisms are discussed in relation to the current debate on action and emotion understanding.
Partin, Alexander C; Hosek, Matthew P; Luong, Jonathan A; Lella, Srihari K; Sharma, Sachein A R; Ploski, Jonathan E
The amygdala is a heterogeneous, medial temporal lobe structure that has been implicated in the formation, expression and extinction of emotional memories. This structure is composed of numerous nuclei that vary in cytoarchitectonics and neural connections. In particular the lateral nucleus of the amygdala (LA), central nucleus of the amygdala (CeA), and the basal (B) nucleus contribute an essential role to emotional learning. However, to date it is still unclear to what extent these nuclei differ at the molecular level. Therefore we have performed whole genome gene expression analysis on these nuclei to gain a better understanding of the molecular differences and similarities among these nuclei. Specifically the LA, CeA and B nuclei were laser microdissected from the rat brain, and total RNA was isolated from these nuclei and subjected to RNA amplification. Amplified RNA was analyzed by whole genome microarray analysis which revealed that 129 genes are differentially expressed among these nuclei. Notably gene expression patterns differed between the CeA nucleus and the LA and B nuclei. However gene expression differences were not considerably different between the LA and B nuclei. Secondary confirmation of numerous genes was performed by in situ hybridization to validate the microarray findings, which also revealed that for many genes, expression differences among these nuclei were consistent with the embryological origins of these nuclei. Knowing the stable gene expression differences among these nuclei will provide novel avenues of investigation into how these nuclei contribute to emotional arousal and emotional learning, and potentially offer new genetic targets to manipulate emotional learning and memory. Copyright © 2013 Elsevier Inc. All rights reserved.
Palumbo, Letizia; Jellema, Tjeerd
Emotional facial expressions are immediate indicators of the affective dispositions of others. Recently it has been shown that early stages of social perception can already be influenced by (implicit) attributions made by the observer about the agent's mental state and intentions. In the current study possible mechanisms underpinning distortions in the perception of dynamic, ecologically-valid, facial expressions were explored. In four experiments we examined to what extent basic perceptual processes such as contrast/context effects, adaptation and representational momentum underpinned the perceptual distortions, and to what extent 'emotional anticipation', i.e. the involuntary anticipation of the other's emotional state of mind on the basis of the immediate perceptual history, might have played a role. Neutral facial expressions displayed at the end of short video-clips, in which an initial facial expression of joy or anger gradually morphed into a neutral expression, were misjudged as being slightly angry or slightly happy, respectively (Experiment 1). This response bias disappeared when the actor's identity changed in the final neutral expression (Experiment 2). Videos depicting neutral-to-joy-to-neutral and neutral-to-anger-to-neutral sequences again produced biases but in opposite direction (Experiment 3). The bias survived insertion of a 400 ms blank (Experiment 4). These results suggested that the perceptual distortions were not caused by any of the low-level perceptual mechanisms (adaptation, representational momentum and contrast effects). We speculate that especially when presented with dynamic, facial expressions, perceptual distortions occur that reflect 'emotional anticipation' (a low-level mindreading mechanism), which overrules low-level visual mechanisms. Underpinning neural mechanisms are discussed in relation to the current debate on action and emotion understanding.
Kashdan, Todd B; Volkmann, Jeffrey R; Breen, William E; Han, Susan
In general, expressing emotions is beneficial and withholding emotions has personal and social costs. Yet, to serve social functions there are situations when emotions are withheld strategically. We examined whether social anxiety influenced when and how emotion expressiveness influences interpersonal closeness in existing romantic relationships. For people with greater social anxiety, withholding the expression of negative emotions was proposed to preserve romantic relationships and their benefits. We examined whether social anxiety and emotion expressiveness interacted to predict prospective changes in romantic relationship closeness over a 12-week period. For people with less social anxiety, relationship closeness was enhanced over time when negative emotions were openly expressed whereas relationship deterioration was found for those more likely to withhold emotions. The reverse pattern was found for people with greater social anxiety such that relationship closeness was enhanced over time for those more likely to withhold negative emotions. Related social anxiety findings were found for discrepancies between desired and actual feelings of closeness over time. Findings were not attributable to depressive symptoms. These results suggest that the costs and benefits of emotion expression are influenced by a person's degree of social anxiety.
Hildebrandt, A; Kiy, A; Reuter, M; Sommer, W; Wilhelm, O
Face cognition, including face identity and facial expression processing, is a crucial component of socio-emotional abilities, characterizing humans as highest developed social beings. However, for these trait domains molecular genetic studies investigating gene-behavior associations based on well-founded phenotype definitions are still rare. We examined the relationship between 5-HTTLPR/rs25531 polymorphisms - related to serotonin-reuptake - and the ability to perceive and recognize faces and emotional expressions in human faces. For this aim we conducted structural equation modeling on data from 230 young adults, obtained by using a comprehensive, multivariate task battery with maximal effort tasks. By additionally modeling fluid intelligence and immediate and delayed memory factors, we aimed to address the discriminant relationships of the 5-HTTLPR/rs25531 polymorphisms with socio-emotional abilities. We found a robust association between the 5-HTTLPR/rs25531 polymorphism and facial emotion perception. Carriers of two long (L) alleles outperformed carriers of one or two S alleles. Weaker associations were present for face identity perception and memory for emotional facial expressions. There was no association between the 5-HTTLPR/rs25531 polymorphism and non-social abilities, demonstrating discriminant validity of the relationships. We discuss the implications and possible neural mechanisms underlying these novel findings. © 2016 John Wiley & Sons Ltd and International Behavioural and Neural Genetics Society.
Nordström, Henrik; Laukka, Petri; Thingujam, Nutankumar S.; Schubert, Emery; Elfenbein, Hillary Anger
This study explored the perception of emotion appraisal dimensions on the basis of speech prosody in a cross-cultural setting. Professional actors from Australia and India vocally portrayed different emotions (anger, fear, happiness, pride, relief, sadness, serenity and shame) by enacting emotion-eliciting situations. In a balanced design, participants from Australia and India then inferred aspects of the emotion-eliciting situation from the vocal expressions, described in terms of appraisal ...
Lena Rachel Quinto
Full Text Available Singing involves vocal production accompanied by a dynamic and meaningful use of facial expressions, which may serve as ancillary gestures that complement, disambiguate, or reinforce the acoustic signal. In this investigation, we examined the use of facial movements to communicate emotion, focusing on movements arising in three epochs: before vocalisation (pre-production, during vocalisation (production, and immediately after vocalisation (post-production. The stimuli were recordings of seven vocalists’ facial movements as they sang short (14 syllable melodic phrases with the intention of communicating happiness, sadness, irritation, or no emotion. Facial movements were presented as point-light displays to 16 observers who judged the emotion conveyed. Experiment 1 revealed that the accuracy of emotional judgement varied with singer, emotion and epoch. Accuracy was highest in the production epoch, however, happiness was well communicated in the pre-production epoch. In Experiment 2, observers judged point-light displays of exaggerated movements. The ratings suggested that the extent of facial and head movements is largely perceived as a gauge of emotional arousal. In Experiment 3, observers rated point-light displays of scrambled movements. Configural information was removed in these stimuli but velocity and acceleration were retained. Exaggerated scrambled movements were likely to be associated with happiness or irritation whereas unexaggerated scrambled movements were more likely to be identified as neutral. An analysis of the motions of singers revealed systematic changes in facial movement as a function of the emotional intentions of singers. The findings confirm the central role of facial expressions in vocal emotional communication, and highlight individual differences between singers in the amount and intelligibility of facial movements made before, during, and after vocalization.
Green, Beth L.; Malsch, Anna M.; Kothari, Brianne Hood; Busse, Jessica; Brennan, Eileen
This article describes the development, implementation, and outcomes of a pilot intervention designed to enhance preschool programs' ability to support children's social-emotional development. Working with two Head Start programs, the intervention included (1) restructuring existing early childhood mental health consultation services; (2) engaging…
Narrative and emotion integration in psychotherapy: investigating the relationship between autobiographical memory specificity and expressed emotional arousal in brief emotion-focused and client-centred treatments of depression.
Boritz, Tali Zweig; Angus, Lynne; Monette, Georges; Hollis-Walker, Laurie; Warwar, Serine
Clinically depressed individuals have consistently been shown to demonstrate a bias for overgeneral autobiographical memory (ABM) disclosure, a strategy used to protect against the access of intense, primary emotions that may accompany specific memories. The present study examined how ABM specificity in client narratives was related to expressed emotional arousal in brief emotion-focused and client-centred psychotherapy for depression. Emotion episodes identified in two early-, two middle-, and two late-therapy transcripts drawn from 34 clients from the York I Depression Study were rated for degree of ABM specificity and expressed emotional arousal. A hierarchical linear modelling analysis demonstrated that greater ABM specificity was associated with higher expressed emotional arousal for clients who were no longer depressed at therapy termination.
Silvio José Lemos Vasconcellos
Full Text Available Recent studies have investigated the ability of adult psychopaths and children with psychopathy traits to identify specific facial expressions of emotion. Conclusive results have not yet been found regarding whether psychopathic traits are associated with a specific deficit in the ability of identifying negative emotions such as fear and sadness. This study compared 20 adolescents with psychopathic traits and 21 adolescents without these traits in terms of their ability to recognize facial expressions of emotion using facial stimuli presented during 200 milliseconds, 500 milliseconds, and 1 second expositions. Analyses indicated significant differences between the two groups' performances only for fear and when displayed for 200 ms. This finding is consistent with findings from other studies in the field and suggests that controlling the duration of exposure to affective stimuli in future studies may help to clarify the mechanisms underlying the facial affect recognition deficits of individuals with psychopathic traits.
Di Paola, Fiammetta; Faravelli, Carlo; Ricca, Valdo
The aim of this study was to verify the level of expressed emotion (EE) as perceived from patients with an eating disorder (ED). The Italian translation of the Level of Expressed Emotion Scale was administered to 63 female patients with Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition, diagnosis of anorexia nervosa, bulimia nervosa, and binge-eating disorder and 63 control subjects, according to a case-control procedure. Patients with ED showed higher level of perceived EE than controls, whereas no significant differences were observed when comparing the 3 patient subgroups. The level of perceived EE was found to be independent of age, person who has been most influential in the patient's life, amount of contacts, and duration of treatment. Different associations between eating disorder psychopathology and EE were found, suggesting a close relationship between the emotional response and tolerance of influential person and the dysfunctional attitudes regarding eating, weight, and body shape. Copyright 2010 Elsevier Inc. All rights reserved.
Sievers, Beau; Polansky, Larry; Casey, Michael; Wheatley, Thalia
Music moves us. Its kinetic power is the foundation of human behaviors as diverse as dance, romance, lullabies, and the military march. Despite its significance, the music-movement relationship is poorly understood. We present an empirical method for testing whether music and movement share a common structure that affords equivalent and universal emotional expressions. Our method uses a computer program that can generate matching examples of music and movement from a single set of features: rate, jitter (regularity of rate), direction, step size, and dissonance/visual spikiness. We applied our method in two experiments, one in the United States and another in an isolated tribal village in Cambodia. These experiments revealed three things: (i) each emotion was represented by a unique combination of features, (ii) each combination expressed the same emotion in both music and movement, and (iii) this common structure between music and movement was evident within and across cultures. PMID:23248314
Sievers, Beau; Polansky, Larry; Casey, Michael; Wheatley, Thalia
Music moves us. Its kinetic power is the foundation of human behaviors as diverse as dance, romance, lullabies, and the military march. Despite its significance, the music-movement relationship is poorly understood. We present an empirical method for testing whether music and movement share a common structure that affords equivalent and universal emotional expressions. Our method uses a computer program that can generate matching examples of music and movement from a single set of features: rate, jitter (regularity of rate), direction, step size, and dissonance/visual spikiness. We applied our method in two experiments, one in the United States and another in an isolated tribal village in Cambodia. These experiments revealed three things: (i) each emotion was represented by a unique combination of features, (ii) each combination expressed the same emotion in both music and movement, and (iii) this common structure between music and movement was evident within and across cultures.
Full Text Available Individuals regulate their emotions in a wide variety of ways. In the present review it has been addressed the issue of whether some forms of emotion regulation are healthier than others by focusing on two commonly used emotion regulation strategies: cognitive reappraisal (changing the way one thinks about potentially emotion-eliciting events and expressive suppression (changing the way one behaviorally responds to emotion-eliciting events. In the first section, experimental findings showing that cognitive reappraisal has a healthier profile of short-term affective, cognitive, and social consequences than expressive suppression are briefly reported. In the second section, individual-difference findings are reviewed showing that using cognitive reappraisal to regulate emotions is associated with healthier patterns of affect, social functioning, and well-being than is using expressive suppression. Finally, brain structural basis and functional activation linked to the habitual usage of cognitive reappraisal and expressive suppression are discussed in detail.
Individuals regulate their emotions in a wide variety of ways. In the present review it has been addressed the issue of whether some forms of emotion regulation are healthier than others by focusing on two commonly used emotion regulation strategies: cognitive reappraisal (changing the way one thinks about potentially emotion-eliciting events) and expressive suppression (changing the way one behaviorally responds to emotion-eliciting events). In the first section, experimental findings showing that cognitive reappraisal has a healthier profile of short-term affective, cognitive, and social consequences than expressive suppression are briefly reported. In the second section, individual-difference findings are reviewed showing that using cognitive reappraisal to regulate emotions is associated with healthier patterns of affect, social functioning, and well-being than is using expressive suppression. Finally, brain structural basis and functional activation linked to the habitual usage of cognitive reappraisal and expressive suppression are discussed in detail.
Mamonto, N. E.; Maulana, H.; Liliana, D. Y.; Basaruddin, T.
Datasets that have been developed before contain facial expression from foreign people. The development of multimedia content aims to answer the problems experienced by the research team and other researchers who will conduct similar research. The method used in the development of multimedia content as facial expression datasets for human emotion recognition is the Villamil-Molina version of the multimedia development method. Multimedia content developed with 10 subjects or talents with each talent performing 3 shots with each capturing talent having to demonstrate 19 facial expressions. After the process of editing and rendering, tests are carried out with the conclusion that the multimedia content can be used as a facial expression dataset for recognition of human emotions.
Liess, Anna; Simon, Wendy; Yutsis, Maya; Owen, Jason E.; Piemme, Karen Altree; Golant, Mitch; Giese-Davis, Janine
Accurately detecting emotional expression in women with primary breast cancer participating in support groups may be important for therapists and researchers. In 2 small studies (N = 20 and N = 16), the authors examined whether video coding, human text coding, and automated text analysis provided consistent estimates of the level of emotional…
Dagan, Meirav; Sanderman, Robbert; Hoff, Christiaan; Meijerink, W. J. H. Jeroen; Baas, Peter C.; van Haastert, Michiel; Hagedoorn, Mariët
The central aim of this longitudinal observational study was to test whether patients with a high need for emotional expression are especially sensitive to their partners' responsive behavior, and therefore at risk for depressive symptoms when responsiveness is withheld. Patients with colorectal
Jeon, Hana; Moulson, Margaret C.; Fox, Nathan; Zeanah, Charles; Nelson, Charles A., III
The current study examined the effects of institutionalization on the discrimination of facial expressions of emotion in three groups of 42-month-old children. One group consisted of children abandoned at birth who were randomly assigned to Care-as-Usual (institutional care) following a baseline assessment. Another group consisted of children…
Carvajal, Fernando; Fernandez-Alcaraz, Camino; Rueda, Maria; Sarrion, Louise
The processing of facial expressions of emotions by 23 adults with Down syndrome and moderate intellectual disability was compared with that of adults with intellectual disability of other etiologies (24 matched in cognitive level and 26 with mild intellectual disability). Each participant performed 4 tasks of the Florida Affect Battery and an…
Daou, Nidal; Hady, Ryma T.; Poulson, Claire L.
The developmental literature has focused extensively on deficits in the expression and recognition of emotion in people with autism, and has reported on the use of interactive tools to address the problems of affect. The behavioral literature has offered interventions to teach children with autism to engage in appropriate affective displays, and…
Jabbi, Mbemba; Keysers, Christian
The observation of movies of facial expressions of others has been shown to recruit similar areas involved in experiencing one's own emotions: the inferior frontal gyrus (IFG). the anterior insula and adjacent frontal operculum (IFO). The Causal link bet between activity in these 2 regions,
Eisner, Lori R.; Johnson, Sheri L.
Expressed emotion (EE) is a robust predictor of outcome in bipolar disorder. Despite decades of research, interventions to reduce EE levels have had only modest effects. This study used an expanded model of EE to develop an intervention. Research has demonstrated a strong link between attributions and EE in families of patients with psychiatric…
Falkmer, Marita; Bjallmark, Anna; Larsson, Matilda; Falkmer, Torbjorn
Can the disadvantages persons with Asperger syndrome frequently experience with reading facially expressed emotions be attributed to a different visual perception, affecting their scanning patterns? Visual search strategies, particularly regarding the importance of information from the eye area, and the ability to recognise facially expressed…
Alvarez-Bolado, Gonzalo; Celio, Marco R
The lateral hypothalamus has been long suspected of triggering the expression of positive emotions, because stimulations of its tuberal portion provoke bursts of laughter. Electrophysiological studies in various species have indeed confirmed that the lateral hypothalamus contributes to reward mechanisms. However, only the rudiments of the neural circuit underlying the expression of positive emotions are known. The prefrontal cortex, the lateral hypothalamus, and the periaqueductal gray matter (PAG) are involved in these circuits; so, too, are the brainstem nuclei that control the laryngeal muscles and subserve mimicry, as well as the cardiovascular and respiratory systems. The implicated populations of hypothalamic neurons have not been defined either anatomically or molecularly. One promising candidate is the novel parvafox nucleus, which we recently described, in the murine medial forebrain bundle (mfb), which specifically expresses parvalbumin and Foxb1. With the molecularly defined parvafox nucleus as a centerpiece, the inputs from the prefrontal cortex and the projections to the PAG and brainstem can be studied with precision. By drawing on genetic approaches, it will be possible to manipulate the circuitry selectively with spatial and temporal exactitude and to evaluate the concomitant autonomic changes. These data will serve as a basis for imaging studies in humans using various paradigms to provoke the expression of positive emotions. In conclusion, studies of the hypothalamic parvafox nucleus will reveal whether this entity represents the fulcrum for positive emotions, as is the amygdala for fear and the insula for disgust. © 2015 Wiley Periodicals, Inc.
Spangler, Sibylle M.; Schwarzer, Gudrun; Korell, Monika; Maier-Karius, Johanna
Four experiments were conducted with 5- to 11-year-olds and adults to investigate whether facial identity, facial speech, emotional expression, and gaze direction are processed independently of or in interaction with one another. In a computer-based, speeded sorting task, participants sorted faces according to facial identity while disregarding…
Garcia-Lopez, Luis-Joaquin; Muela, Jose M.; Espinosa-Fernandez, Lourdes; Diaz-Castela, Mar
The role that the involvement of parents may play in the treatment outcome of their children with anxiety disorders is still under debate. Some studies dealing with other disorders have examined the role that the expressed emotion (EE) construct (parental overinvolvement, criticism and hostility) may play in treatment outcome and relapse. Given…
The study investigated the recognition of standardized facial expressions of emotion (anger, fear, disgust, happiness, sadness, surprise) at a perceptual level (experiment 1) and at a semantic level (experiments 2 and 3) in children with autism (N= 20) and normally developing children (N= 20). Results revealed that children with autism were as…
Lacroix, Agnes; Guidetti, Michele; Roge, Bernadette; Reilly, Judy
The aim of our study was to compare two neurodevelopmental disorders (Williams syndrome and autism) in terms of the ability to recognize emotional and nonemotional facial expressions. The comparison of these two disorders is particularly relevant to the investigation of face processing and should contribute to a better understanding of social…
Full Text Available Although current emotion theories emphasize the importance of contextual factors for emotional expressive behavior, developmental studies that examine such factors are currently thin on the ground. In this research, we studied the course of emotional expressions of 8- and 11-year-old children after winning a (large first prize or a (substantially smaller consolation prize, while playing a game competing against the computer or a physically co-present peer. We analyzed their emotional reactions by conducting two perception tests in which participants rated children’s level of happiness. Results showed that co-presence positively affected children’s happiness only when receiving the first prize. Moreover, for children who were in the presence of a peer, we found that eye contact affected children’s expressions of happiness, but that the effect was different for different age groups: 8-year-old children were negatively affected, and 11-year-old children positively. Overall, we can conclude that as children grow older and their social awareness increases, the presence of a peer affects their nonverbal expressions, regardless of their appreciation of their prize.
Nugter, M. A.; Dingemans, P. M.; Linszen, D. H.; van der Does, A. J.; Gersons, B. P.
The relationships between expressed emotion (EE), affective style (AS) and communication deviance (CD) were studied during hospitalization and after discharge. EE was measured with both the Camberwell Family Interview (CFI) and the Five-Minutes Speech Sample (FMSS). The study subjects were patients
Haskett, Mary E.; Stelter, Rebecca; Proffit, Katie; Nice, Rachel
Objective: Identifying factors associated with school functioning of abused children is important in prevention of long-term negative outcomes associated with school failure. The purpose of this study was to examine the degree to which parent emotional expressiveness and children's self-regulation predicted early school behavior of abused…
Miiamaaria V Kujala
Full Text Available Facial expressions are important for humans in communicating emotions to the conspecifics and enhancing interpersonal understanding. Many muscles producing facial expressions in humans are also found in domestic dogs, but little is known about how humans perceive dog facial expressions, and which psychological factors influence people's perceptions. Here, we asked 34 observers to rate the valence, arousal, and the six basic emotions (happiness, sadness, surprise, disgust, fear, and anger/aggressiveness from images of human and dog faces with Pleasant, Neutral and Threatening expressions. We investigated how the subjects' personality (the Big Five Inventory, empathy (Interpersonal Reactivity Index and experience of dog behavior affect the ratings of dog and human faces. Ratings of both species followed similar general patterns: human subjects classified dog facial expressions from pleasant to threatening very similarly to human facial expressions. Subjects with higher emotional empathy evaluated Threatening faces of both species as more negative in valence and higher in anger/aggressiveness. More empathetic subjects also rated the happiness of Pleasant humans but not dogs higher, and they were quicker in their valence judgments of Pleasant human, Threatening human and Threatening dog faces. Experience with dogs correlated positively with ratings of Pleasant and Neutral dog faces. Personality also had a minor effect on the ratings of Pleasant and Neutral faces in both species. The results imply that humans perceive human and dog facial expression in a similar manner, and the perception of both species is influenced by psychological factors of the evaluators. Especially empathy affects both the speed and intensity of rating dogs' emotional facial expressions.
Kujala, Miiamaaria V.; Somppi, Sanni; Jokela, Markus; Vainio, Outi; Parkkonen, Lauri
Facial expressions are important for humans in communicating emotions to the conspecifics and enhancing interpersonal understanding. Many muscles producing facial expressions in humans are also found in domestic dogs, but little is known about how humans perceive dog facial expressions, and which psychological factors influence people’s perceptions. Here, we asked 34 observers to rate the valence, arousal, and the six basic emotions (happiness, sadness, surprise, disgust, fear, and anger/aggressiveness) from images of human and dog faces with Pleasant, Neutral and Threatening expressions. We investigated how the subjects’ personality (the Big Five Inventory), empathy (Interpersonal Reactivity Index) and experience of dog behavior affect the ratings of dog and human faces. Ratings of both species followed similar general patterns: human subjects classified dog facial expressions from pleasant to threatening very similarly to human facial expressions. Subjects with higher emotional empathy evaluated Threatening faces of both species as more negative in valence and higher in anger/aggressiveness. More empathetic subjects also rated the happiness of Pleasant humans but not dogs higher, and they were quicker in their valence judgments of Pleasant human, Threatening human and Threatening dog faces. Experience with dogs correlated positively with ratings of Pleasant and Neutral dog faces. Personality also had a minor effect on the ratings of Pleasant and Neutral faces in both species. The results imply that humans perceive human and dog facial expression in a similar manner, and the perception of both species is influenced by psychological factors of the evaluators. Especially empathy affects both the speed and intensity of rating dogs’ emotional facial expressions. PMID:28114335
Poole, Ernest Andre
In this review I explore and discuss the use of micro and subtle expression training in the social sciences. These trainings, offered commercially, are designed and endorsed by noted psychologist Paul Ekman, co-author of the Facial Action Coding System, a comprehensive system of measuring muscular movement in the face and its relationship to the…
Bouhuys, Antoinette L.; Bloem, Gerda M.; Groothuis, Ton G.G.
The judgement of healthy subject rating the emotional expressions of a set of schematic drawn faces is validated (study 1) to examine the relationship between mood (depressed/elated) and judgement of emotional expressions of these faces (study 2). Study 1: 30 healthy subjects judged 12 faces with
Klapwijk, Eduard T.; Aghajani, Moji; Lelieveld, Gert-Jan; van Lang, Natasja D. J.; Popma, Arne; van der Wee, Nic J. A.; Colins, Olivier F.; Vermeiren, Robert R. J. M.
Little is known about how emotions expressed by others influence social decisions and associated brain responses in autism spectrum disorders (ASD). We investigated the neural mechanisms underlying fairness decisions in response to explicitly expressed emotions of others in boys with ASD and typically developing (TD) boys. Participants with ASD…
Widen, Sherri C.; Russell, James A.
Lay people and scientists alike assume that, especially for young children, facial expressions are a strong cue to another's emotion. We report a study in which children (N=120; 3-4 years) described events that would cause basic emotions (surprise, fear, anger, disgust, sadness) presented as its facial expression, as its label, or as its…
Meffert, Harma; Brislin, Sarah J; White, Stuart F; Blair, James R
Social referencing paradigms in humans and observational learning paradigms in animals suggest that emotional expressions are important for communicating valence. It has been proposed that these expressions initiate stimulus-reinforcement learning. Relatively little is known about the role of emotional expressions in reinforcement learning, particularly in the context of social referencing. In this study, we examined object valence learning in the context of a social referencing paradigm. Participants viewed objects and faces that turned toward the objects and displayed a fearful, happy or neutral reaction to them, while judging the gender of these faces. Notably, amygdala activation was larger when the expressions following an object were less expected. Moreover, when asked, participants were both more likely to want to approach, and showed stronger amygdala responses to, objects associated with happy relative to objects associated with fearful expressions. This suggests that the amygdala plays two roles in social referencing: (i) initiating learning regarding the valence of an object as a function of prediction errors to expressions displayed toward this object and (ii) orchestrating an emotional response to the object when value judgments are being made regarding this object. Published by Oxford University Press 2014. This work is written by US Government employees and is in the public domain in the US.
There are currently scarce research regarding further training programs and employees’ attitudes toward them. This present work examined 95 preschool employees from one municipal community in matters of emotional intelligence, need for cognition, and cognitive reflective ability and how these influenced their attitudes toward a further training program called International Child Development Programme, ICDP (study 1). Six participants were also interviewed in regards to more organizational asp...
Delis, Ioannis; Chen, Chaona; Jack, Rachael E.; Garrod, Oliver G. B.; Panzeri, Stefano; Schyns, Philippe G.
Visual categorization is the brain computation that reduces high-dimensional information in the visual environment into a smaller set of meaningful categories. An important problem in visual neuroscience is to identify the visual information that the brain must represent and then use to categorize visual inputs. Here we introduce a new mathematical formalism—termed space-by-time manifold decomposition—that describes this information as a low-dimensional manifold separable in space and time. We use this decomposition to characterize the representations used by observers to categorize the six classic facial expressions of emotion (happy, surprise, fear, disgust, anger, and sad). By means of a Generative Face Grammar, we presented random dynamic facial movements on each experimental trial and used subjective human perception to identify the facial movements that correlate with each emotion category. When the random movements projected onto the categorization manifold region corresponding to one of the emotion categories, observers categorized the stimulus accordingly; otherwise they selected “other.” Using this information, we determined both the Action Unit and temporal components whose linear combinations lead to reliable categorization of each emotion. In a validation experiment, we confirmed the psychological validity of the resulting space-by-time manifold representation. Finally, we demonstrated the importance of temporal sequencing for accurate emotion categorization and identified the temporal dynamics of Action Unit components that cause typical confusions between specific emotions (e.g., fear and surprise) as well as those resolving these confusions. PMID:27305521
Drolet, Matthis; Schubotz, Ricarda I; Fischer, Julia
Authenticity of vocal emotion expression affects emotion recognition and brain activity in the so-called Theory of Mind (ToM) network, which is implied in the ability to explain and predict behavior by attributing mental states to other individuals. Exploiting the variability of the fundamental frequency (F0 contour), which varies more (higher contour) in play-acted expressions than authentic ones, we examined whether contour biases explicit categorization toward a particular authenticity or emotion category. Moreover, we tested whether contour modulates blood-oxygen-level dependent (BOLD) response in the ToM network and explored the role of task as a top-down modulator. The effects of contour on BOLD signal were analyzed by contrasting high and low contour stimuli within two previous fMRI studies that implemented emotion and authenticity rating tasks. Participants preferentially categorized higher contour stimuli as play-acted and lower contour stimuli as sad. Higher contour was found to up-regulate activation task-independently in the primary auditory cortex. Stimulus contour and task were found to interact in a network including medial prefrontal cortex, with an increase in BOLD signal for low-contour stimuli during explicit perception of authenticity and an increase for high-contour stimuli during explicit perception of emotion. Contour-induced BOLD effects appear to be purely stimulus-driven in early auditory and intonation perception, while being strongly task-dependent in regions involved in higher cognition.
Eerola, Tuomas; Friberg, Anders; Bresin, Roberto
The aim of this study is to manipulate musical cues systematically to determine the aspects of music that contribute to emotional expression, and whether these cues operate in additive or interactive fashion, and whether the cue levels can be characterized as linear or non-linear. An optimized factorial design was used with six primary musical cues (mode, tempo, dynamics, articulation, timbre, and register) across four different music examples. Listeners rated 200 musical examples according to four perceived emotional characters (happy, sad, peaceful, and scary). The results exhibited robust effects for all cues and the ranked importance of these was established by multiple regression. The most important cue was mode followed by tempo, register, dynamics, articulation, and timbre, although the ranking varied across the emotions. The second main result suggested that most cue levels contributed to the emotions in a linear fashion, explaining 77–89% of variance in ratings. Quadratic encoding of cues did lead to minor but significant increases of the models (0–8%). Finally, the interactions between the cues were non-existent suggesting that the cues operate mostly in an additive fashion, corroborating recent findings on emotional expression in music (Juslin and Lindström, 2010). PMID:23908642
Malatesta, C Z; Culver, C; Tesman, J R; Shepard, B
This study examines the course of emotion expression development over the first 2 years of life in a sample of full-term and preterm children. 58 mother/infant pairs were videotaped at infant ages of 2 1/2, 5, 7 1/2, and 22 months, recording face-to-face interaction involving play and separation/reunion sessions. The tapes were coded on a second-to-second basis using Izard's facial affect coding system. Data analysis focused on (1) differences in expressive behavior at 22 months as a function of risk status, gender, attachment status, and patterns of earlier maternal contingency behavior; (2) stability of specific emotional expressive patterns across assessment periods; and (3) the relation of expressive behavior and security of attachment at 2 years to qualities of earlier affective interchange. Mother's contingency behavior (both general level and specific contingency patterns) appeared to have a material effect on the course of emotional development, as did birth status and gender. Prematurity was associated with differential socioemotional development well into the second year, much in contrast to the "catch-up effect" observed in linguistic and cognitive functioning. Discrete emotions analysis of attachment groups yielded differentiation along a broad negative/positive dimension, but it also showed that insecurely attached children can be characterized as showing inhibited anger expression. The results of this study are discussed within the framework of organizational models of infant affective development; attachment theory and discrete emotions approaches were found to yield different yet equally informative data on the course of socioemotional development.
Mill, Aire; Kööts-Ausmees, Liisi; Allik, Jüri; Realo, Anu
The main aim of the current study was to examine the role of co-occurring emotions and their interactive effects with the Big Five personality traits in anger expression. Everyday anger expression (“anger-in” and “anger-out” behavior) was studied with the experience-sampling method in a group of 110 participants for 14 consecutive days on 7 random occasions per day. Our results showed that the simultaneously co-occurring emotions that buffer against anger expression are sadness, surprise, disgust, disappointment, and irritation for anger-in behavior, and fear, sadness and disappointment for anger-out reactions. While previous studies have shown that differentiating one's current affect into discrete emotion categories buffers against anger expression (Pond et al., 2012), our study further demonstrated the existence of specific interactive effects between the experience of momentary emotions and personality traits that lead to higher levels of either suppression or expression of anger behavior (or both). For example, the interaction between the trait Openness and co-occurring surprise, in predicting anger-in behavior, indicates that less open people hold their anger back more, and more open people use less anger-in behavior. Co-occurring disgust increases anger-out reactions in people low in Conscientiousness, but decreases anger-out reactions in people high in Conscientiousness. People high in Neuroticism are less likely to engage in anger-in behavior when experiencing disgust, surprise, or irritation alongside anger, but show more anger out in the case of co-occurring contempt. The results of the current study help to further clarify the interactions between the basic personality traits and the experience of momentary co-occurring emotions in determining anger behavior. PMID:29479333
It was widely accepted that emotion such as fear, anger and pleasure could not be studied using a modern scientific tools. During the very early periods of emotion researches, psychologists, but not biologist, dominated in studying emotion and its disorders. Intuitively, one may think that emotion arises from brain first and then bodily responses follow. For example, we are sad first, and then cry. However, groups of psychologists suggested a proposal that our feeling follows bodily responses; that is, we feel sad because we cry! This proposal seems counterintuitive but became a popular hypothesis for emotion. Another example for this hypothesis is as follows. When you accidentally confront a large bear in a mountain, what would be your responses?; you may feel terrified first, and then run, or you may run first, and then feel terrified later on. In fact, the latter explanation is correct! You feel fear after you run (even because you run?). Or, you can imagine that you date with your girl friend who you love so much. Your heart must be beating fast and your body temperature must be elevated! In this situation, if you take a very cold bath, what would you expect? Your hot feeling is usually calmed down after this cold bath; that is, you feel hot because your heart and bodily temperature change. While some evidence supported this hypothesis, others do not. In the case of patients whose cervical vertebrae were severed with an accident, they still retained significant amount of emotion (feelings!) in some cases (but other patients lost most of emotional experience). In addition, one can imagine that there would be a specific set of physical responses for specific emotion if the original hypothesis is correct (e.g. fasten heart beating and redden face for anger etc.). However, some psychologists failed to find any specific set of physical responses for specific emotion, though others insisted that there existed such specific responses. Based on these controversial
Nentjes, Lieke; Bernstein, David P; Meijer, Ewout; Arntz, Arnoud; Wiers, Reinout W
This study investigated the physiological, self-reported, and facial correlates of emotion regulation in psychopathy. Specifically, we compared psychopathic offenders (n = 42), nonpsychopathic offenders (n = 42), and nonoffender controls (n = 26) in their ability to inhibit and express emotion while watching affective films (fear, happy, and sad). Results showed that all participants were capable of drastically diminishing facial emotions under inhibition instructions. Contrary to expectation, psychopaths were not superior in adopting such a "poker face." Further, the inhibition of emotion was associated with cardiovascular changes, an effect that was also not dependent on psychopathy (or its factors), suggesting emotion inhibition to be an effortful process in psychopaths as well. Interestingly, psychopathic offenders did not differ from nonpsychopaths in the capacity to show content-appropriate facial emotions during the expression condition. Taken together, these data challenge the view that psychopathy is associated with either superior emotional inhibitory capacities or a generalized impairment in showing facial affect.
Belyk, Michel; Brown, Steven; Lim, Jessica; Kotz, Sonja A
Humans communicate through a combination of linguistic and emotional channels, including propositional speech, writing, sign language, music, but also prosodic, facial, and gestural expression. These channels can be interpreted separately or they can be integrated to multimodally convey complex meanings. Neural models of the perception of semantics and emotion include nodes for both functions in the inferior frontal gyrus pars orbitalis (IFGorb). However, it is not known whether this convergence involves a common functional zone or instead specialized subregions that process semantics and emotion separately. To address this, we performed Kernel Density Estimation meta-analyses of published neuroimaging studies of the perception of semantics or emotion that reported activation in the IFGorb. The results demonstrated that the IFGorb contains two zones with distinct functional profiles. A lateral zone, situated immediately ventral to Broca's area, was implicated in both semantics and emotion. Another zone, deep within the ventral frontal operculum, was engaged almost exclusively by studies of emotion. Follow-up analysis using Meta-Analytic Connectivity Modeling demonstrated that both zones were frequently co-activated with a common network of sensory, motor, and limbic structures, although the lateral zone had a greater association with prefrontal cortical areas involved in executive function. The status of the lateral IFGorb as a point of convergence between the networks for processing semantic and emotional content across modalities of communication is intriguing since this structure is preserved across primates with limited semantic abilities. Hence, the IFGorb may have initially evolved to support the comprehension of emotional signals, being later co-opted to support semantic communication in humans by forming new connections with brain regions that formed the human semantic network. Copyright © 2017 Elsevier Inc. All rights reserved.
Song, Yongning; Hakoda, Yuji
Autism spectrum disorders (ASD) are characterized by early onset qualitative impairments in reciprocal social development. However, whether individuals with ASD exhibit impaired recognition of facial expressions corresponding to basic emotions is debatable. To investigate subtle deficits in facial emotion recognition, we asked 14 children diagnosed with high-functioning autism (HFA)/AS and 17 typically developing peers to complete a new highly sensitive test of facial emotion recognition. The test stimuli comprised faces expressing increasing degrees of emotional intensity that slowly changed from a neutral to a full-intensity happiness, sadness, surprise, anger, disgust, or fear expression. We assessed individual differences in the intensity of stimuli required to make accurate judgments about emotional expressions. We found that, different emotions had different identification thresholds and the two groups were generally similar in terms of the sequence of discrimination threshold of six basic expressions. It was easier for individuals in both groups to identify emotions that were relatively fully expressed (e.g., intensity > 50%). Compared with control participants, children with ASD generally required stimuli with significantly greater intensity for the correct identification of anger, disgust, and fear expressions. These results suggest that individuals with ASD do not have a general but rather a selective impairment in basic emotion recognition.
Kalokerinos, Elise K; Greenaway, Katharine H; Pedder, David J; Margetts, Elise A
People who express positive emotion usually have better social outcomes than people who do not, and suppressing the expression of emotions can have interpersonal costs. Nevertheless, social convention suggests that there are situations in which people should suppress the expression of positive emotions, such as when trying to appear humble in victory. The present research tested whether there are interpersonal costs to expressing positive emotions when winning. In Experiment 1, inexpressive winners were evaluated more positively and rated as lower in hubristic-but not authentic-pride compared with expressive winners. Experiment 2 confirmed that inexpressive winners were perceived as using expressive suppression to downregulate their positive emotion expression. Experiment 3 replicated the findings of Experiment 1, and also found that people were more interested in forming a friendship with inexpressive winners than expressive winners. The effects were mediated by the perception that the inexpressive winner tried to protect the loser's feelings. This research is the first to identify social costs of expressing positive emotion, and highlights the importance of understanding the situational context when determining optimal emotion regulation strategies. PsycINFO Database Record (c) 2014 APA, all rights reserved.
Wasserman, Stephanie; Weisman, Amy; Suro, Giulia
Expressed emotion (EE) is a measure of the amount of criticism and emotional over involvement expressed by a key relative towards a relative with a disorder or illness (Hooley, 2007). Research has established that living in a high EE environment, which is characterized by increased levels of critical and emotionally exaggerated communication, leads to a poorer prognosis for patients with a mental illness when compared to low EE environments. Despite evidence that EE is a strong predictor of course of illness, there continue to be questions concerning why some family members express excessive levels of high EE attitudes about their mentally ill relatives while others do not. Based on indirect evidence from previous research, the current study tested whether religious and nonreligious coping serve as predictors of EE. A sample of 72 family members of patients with schizophrenia completed an EE interview, along with questionnaires assessing situational nonreligious coping and religious coping. In line with hypotheses, results indicated that nonreligious coping predicted EE. Specifically, less use of adaptive emotion-focused coping predicted high EE. Also consistent with predictions, maladaptive religious coping predicted high EE above and beyond nonreligious coping. PMID:23393424
Mioni, Giovanna; Meligrana, Lucia; Grondin, Simon; Perini, Francesco; Bartolomei, Luigi; Stablum, Franca
Previous studies have demonstrated that emotional facial expressions alter temporal judgments. Moreover, while some studies conducted with Parkinson's disease (PD) patients suggest dysfunction in the recognition of emotional facial expression, others have shown a dysfunction in time perception. In the present study, we investigate the magnitude of temporal distortions caused by the presentation of emotional facial expressions (anger, shame, and neutral) in PD patients and controls. Twenty-five older adults with PD and 17 healthy older adults took part in the present study. PD patients were divided into two sub-groups, with and without mild cognitive impairment (MCI), based on their neuropsychological performance. Participants were tested with a time bisection task with standard intervals lasting 400 ms and 1600 ms. The effect of facial emotional stimuli on time perception was evident in all participants, yet the effect was greater for PD-MCI patients. Furthermore, PD-MCI patients were more likely to underestimate long and overestimate short temporal intervals than PD-non-MCI patients and controls. Temporal impairment in PD-MCI patients seem to be mainly caused by a memory dysfunction. (JINS, 2016, 22, 890-899).
Strauss, Gregory P; Keller, William R; Koenig, James I; Sullivan, Sara K; Gold, James M; Buchanan, Robert W
Lower endogenous oxytocin levels have been associated with impaired social cognition in schizophrenia, particularly facial affect identification. Little is known about the relationship between oxytocin and other forms of emotion perception. In the current study, 41 individuals with schizophrenia (SZ) and 22 demographically matched healthy controls (CN) completed a forced-choice affective body expression classification task. Stimuli included dynamic videos of male and female actors portraying 4 discrete emotions: happiness, sadness, anger, and neutral. Plasma oxytocin levels were determined via radioimmunoassay. Results indicated that SZ had significantly higher plasma oxytocin concentrations than CN. SZ were also less accurate at identifying expressions of happiness and sadness; however, there were no group differences for anger or neutral stimuli. A group×sex interaction was also present, such that female CN were more accurate than male CN, whereas male SZ were more accurate than female SZ. Higher endogenous oxytocin levels were associated with better total recognition in both SZ and CN; this association was specific to females in SZ. Findings indicate that sex plays an important role in identifying emotional expressions in body gestures in SZ, and that individual differences in endogenous oxytocin predict emotion perception accuracy. Copyright © 2015 Elsevier B.V. All rights reserved.
Brescoll, Victoria L; Uhlmann, Eric Luis
Three studies examined the relationships among anger, gender, and status conferral. As in prior research, men who expressed anger in a professional context were conferred higher status than men who expressed sadness. However, both male and female evaluators conferred lower status on angry female professionals than on angry male professionals. This was the case regardless of the actual occupational rank of the target, such that both a female trainee and a female CEO were given lower status if they expressed anger than if they did not. Whereas women's emotional reactions were attributed to internal characteristics (e.g., "she is an angry person,"she is out of control"), men's emotional reactions were attributed to external circumstances. Providing an external attribution for the target person's anger eliminated the gender bias. Theoretical implications and practical applications are discussed.