Emotion perception

Source: Wikipedia, the free encyclopedia.

Emotion perception refers to the capacities and abilities of recognizing and identifying emotions in others, in addition to biological and physiological processes involved. Emotions are typically viewed as having three components: subjective experience, physical changes, and cognitive appraisal; emotion perception is the ability to make accurate decisions about another's subjective experience by interpreting their physical changes through sensory systems responsible for converting these observed changes into mental representations. The ability to perceive emotion is believed to be both innate and subject to environmental influence and is also a critical component in social interactions. How emotion is experienced and interpreted depends on how it is perceived. Likewise, how emotion is perceived is dependent on past experiences and interpretations. Emotion can be accurately perceived in humans. Emotions can be perceived visually, audibly, through smell and also through bodily sensations and this process is believed to be different from the perception of non-emotional material.

Modes of perception

Emotions can be perceived through visual, auditory, olfactory, taste and physiological sensory processes. Nonverbal actions can provide social partners with information about subjective and emotional states. This nonverbal information is believed to hold special importance and sensory systems and certain brain regions are suspected to specialize in decoding emotional information for rapid and efficient processing.

Visual

The visual system is the primary mode of perception for the way people receive emotional information. People use emotional cues displayed by social partners to make decisions regarding their affective state. Emotional cues can be in the form of facial expressions, which are actually a combination of many distinct muscle groups within the face, or bodily postures (alone or in relation to others), or found through the interpretation of a situation or environment known to have particular emotional properties (i.e., a funeral, a wedding, a war zone, a scary alley, etc.). While the visual system is the means by which emotional information is gathered, it is the cognitive interpretation and evaluation of this information that assigns it emotional value, garners the appropriate cognitive resources, and then initiates a physiological response. This process is by no means exclusive to visual perception and in fact may overlap considerably with other modes of perception, suggesting an emotional sensory system comprising multiple perceptual processes all of which are processed through similar channels.

Facial perception

A great deal of research conducted on emotion perception revolves around how people perceive emotion in others' facial expressions. Whether the emotion contained in someone's face is classified categorically or along dimensions of valence and arousal, the face provides reliable cues to one's subjective emotional state. As efficient as humans are in identifying and recognizing emotion in another's face, accuracy goes down considerably for most emotions, with the exception of happiness, when facial features are inverted (i.e., mouth placed above eyes and nose), suggesting that a primary means of facial perception includes the identification of spatial features that resemble a prototypical face, such that two eyes are placed above a nose which is above a mouth; any other formation of features does not immediately constitute a face and requires extra spatial manipulation to identify such features as resembling a face.

Discrete versus dimensional views

Research on the classification of perceived emotions has centered around the debate between two fundamentally distinct viewpoints. One side of the debate posits that emotions are separate and discrete entities whereas the other side suggests that emotions can be classified as values on the dimensions of valence (positive versus negative) and arousal (calm/soothing versus exciting/agitating). Psychologist Paul Ekman supported the discrete emotion perspective with his groundbreaking work comparing emotion perception and expression between literate and preliterate cultures.[1] Ekman concluded that the ability to produce and perceive emotions is universal and innate and that emotions manifest categorically as basic emotions (anger, disgust, fear, happiness, sadness, contempt, surprise, and possibly contempt). The alternative dimensional view garnered support from psychologist James Russell, who is best known for his contributions toward the circumplex of emotion. Russell described emotions as constructs which lie on the dimensions of valence and arousal and it is the combination of these values which delineate emotion.[2] Psychologist Robert Plutchik sought to reconcile these views and proposed that certain emotions be considered "primary emotions" which are grouped either positively or negatively and can then be combined to form more complex emotions, sometimes considered "secondary emotions", such as remorse, guilt, submission, and anticipation. Plutchik created the "wheel of emotions" to outline his theory.[3]

Culture

Culture plays a significant role in emotion perception, most notably in facial perception. Although the features of the face convey important information, the upper (eyes/brow) and lower (mouth/nose) regions of the face have distinct qualities that can provide both consistent and conflicting information. As values, etiquette, and quality of social interactions vary across cultures, facial perception is believed to be moderated accordingly. In western cultures, where overt emotion is ubiquitous, emotional information is primarily obtained from viewing the features of the mouth, which is the most expressive part of the face. However, in eastern cultures, where overt emotional expression is less common and therefore the mouth plays a lesser role in emotional expression, emotional information is more often obtained from viewing the upper region of the face, primarily the eyes.[4] These cultural differences suggest a strong environmental and learned component in emotion expression and emotion perception.

Context

Although facial expressions convey key emotional information, context also plays an important role in both providing additional emotional information and modulating what emotion is actually perceived in a facial expression. Contexts come in three categories: stimulus-based context, in which a face is physically presented with other sensory input that has informational value; perceiver-based context, in which processes within the brain or body of a perceiver can shape emotion perception; and cultural contexts that affect either the encoding or the understanding of facial actions.[5]

Auditory

The auditory system can provide important emotional information about the environment. Voices, screams, murmurs, and music can convey emotional information. Emotional interpretations of sounds tend to be quite consistent. Traditionally, emotion perception in the voice has been determined through research studies analyzing, via prosodic parameters such as pitch and duration, the way in which a speaker expresses an emotion, known as encoding. Alternatively, a listener who attempts to identify a particular emotion as intended by a speaker can decode emotion. More sophisticated methods include manipulating or synthesizing important prosodic parameters in speech signal (e.g., pitch, duration, loudness, voice quality) in both natural and simulated affective speech.[6] Pitch and duration tend to contribute more to emotional recognition than loudness.[7] Music has long been known to have emotional qualities and is a popular strategy in emotion regulation. When asked to rate emotions present in classical music, music professionals could identify all six basic emotions with happiness and sadness the most represented, and in decreasing order of importance, anger, fear, surprise, and disgust.[8] The emotions of happiness, sadness, fear, and peacefulness can be perceived in a short length of exposure, as little as 9–16 seconds[9] including instrumental-only music selections.[10]

Olfactory

Aromas and scents also influence mood, for example through aromatherapy,[11] and humans can extract emotional information from scents just as they can from facial expressions and emotional music. Odors may be able to exert their effects through learning and conscious perception, such that responses typically associated with particular odors are learned through association with their matched emotional experiences. In-depth research has documented that emotion elicited by odors, both pleasant and unpleasant, affects the same physiological correlates of emotion seen with other sensory mechanisms.[12]

Somatic

Theories on emotion have focused on perception, subjective experience, and appraisal. Predominant theories of emotion and emotion perception include what type of emotion is perceived, how emotion is perceived somatically, and at what stage of an event emotion is perceived and translated into subjective, physical experience.

James–Lange theory

Following the influence of René Descartes and his ideas regarding the split between body and mind, in 1884 William James proposed the theory that it is not that the human body acts in response to a person's emotional state, as common sense might suggest, but rather, people interpret their emotions on the basis of their already present bodily state. In the words of James, "we feel sad because we cry, angry because we strike, afraid because we tremble, and neither we cry, strike, nor tremble because we are sorry, angry, or fearful, as the case may be." James believed it was particular and distinct physical patterns that map onto specific experienced emotions. Contemporaneously, psychologist Carl Lange arrived at the same conclusion about the experience of emotions. Thus, the idea that felt emotion is the result of perceiving specific patterns of bodily responses is called the James–Lange theory of emotion.[13] In support of the James–Lange theory of emotion, Silvan Tomkins proposed the facial feedback hypothesis in 1963; he suggested that facial expressions actually trigger the experience of emotions and not the other way around. This theory was tested in 1974 by James Laird in an experiment where Laird asked participants to hold a pencil either between their teeth (artificially producing a smile) or between their upper lip and their nose (artificially producing a frown) and then rate cartoons. Laird found that these cartoons were rated as being funnier by those participants holding a pencil in between their teeth. In addition, Paul Ekman recorded extensive physiological data while participants posed his basic emotional facial expressions and found that heart rate raised for sadness, fear, and anger yet did not change at al for happiness, surprise, or disgust, and skin temperature raised when participants posed anger but not other emotions. While contemporary psychologists still agree with the James–Lange theory of emotion, human subjective emotion is complex and physical reactions or antecedents do not fully explain the subjective emotional experience.

Cannon–Bard theory of emotion

Walter Bradford Cannon and his doctoral student Philip Bard agreed that physiological responses played a crucial role in emotions, but did not believe that physiological responses alone could explain subjective emotional experiences. They argued that physiological responses were too slow relative to the relatively rapid and intense subjective awareness of emotion and that often these emotions are similar and imperceptible to people at such a short timescale. Cannon proposed that the mind and body operate independently in the experience of emotions such that differing brain regions (cortex versus subcortex) process information from an emotion-producing stimulus independently and simultaneously resulting in both an emotional and a physical response. This is best illustrated by imagining an encounter with a grizzly bear; you would simultaneously experience fear, begin to sweat, experience an elevated heart rate, and attempt to run. All of these things would happen at the same time.[14]

Two-factor theory

Stanley Schachter and his doctoral student Jerome Singer formulated their theory of emotion based on evidence that without an actual emotion-producing stimulus, people are unable to attribute specific emotions to their bodily states. They believed that there must be a cognitive component to emotion perception beyond that of just physical changes and subjective feelings. Schachter and Singer suggested that when someone encounters such an emotion-producing stimulus, they would immediately recognize their bodily symptoms (sweating and elevated heart rate in the case of the grizzly bear) as the emotion fear. Their theory was devised as a result of a study in which participants were injected with either a stimulant (adrenaline) that causes elevated heart rate, sweaty palms and shaking, or a placebo. Participants were then either told what the effects of the drug were or were told nothing, and were then placed in a room with a person they did not know who, according to the research plan, would either play with a hula hoop and make paper airplanes (euphoric condition) or ask the participant intimate, personal questions (angry condition). They found that participants who knew what the effects of the drug were attributed their physical state to the effects of the drug; however, those who had no knowledge of the drug they received attributed their physical state to the situation with the other person in the room. These results led to the conclusion that physiological reactions contributed to emotional experience by facilitating a focused cognitive appraisal of a given physiologically arousing event and that this appraisal was what defined the subjective emotional experience. Emotions were thus a result of a two-stage process: first, physiological arousal in a response to an evoking stimulus, and second, cognitive elaboration of the context in which the stimulus occurred.[15]

Neural bases

Emotion perception is primarily a cognitive process driven by particular brain systems believed to specialize in identifying emotional information and subsequently allocating appropriate cognitive resources to prepare the body to respond. The relationship between various regions is still unclear, but a few key regions have been implicated in particular aspects of emotion perception and processing including areas suspected of being involved in the processing of faces and emotional information.

Fusiform face area

The fusiform face area, part of the fusiform gyrus is an area some believe to specialize in the identification and processing of human faces, although others suspect it is responsible for distinguishing between well known objects such as cars and animals[citation needed]. Neuroimaging studies have found activation in this area in response to participants viewing images of prototypical faces, but not scrambled or inverted faces, suggesting that this region is specialized for processing human faces but not other material. This area has been an area of increasing debate and while some psychologists may approach the fusiform face area in a simplistic manner, in that it specializes in the processing of human faces, more likely this area is implicated in the visual processing of many objects, particularly those familiar and prevalent in the environment. Impairments in the ability to recognize subtle differences in faces would greatly inhibit emotion perception and processing and have significant implications involving social interactions and appropriate biological responses to emotional information.

HPA axis

The hypothalamic-pituitary-adrenal (HPA) axis plays a role in emotion perception through its mediation of the physiological stress response. This occurs through the release of hypothalamic corticotropin-releasing factor, also known as corticotropin-releasing hormone (CRH), from nerve terminals in the median eminence arising in the paraventricular nucleus, which stimulates adrenocorticotropin release from the anterior pituitary that in turn induces the release of cortisol from the adrenal cortex. This progressive process culminating in the release of glucocorticoids to environmental stimuli is believed to be initiated by the amygdala, which evaluates the emotional significance of observed phenomena. Released glucocorticoids provide negative feedback on the system and also the hippocampus, which in turn regulates the shutting off of this biological stress response. It is through this response that information is encoded as emotional and bodily response is initiated, making the HPA axis an important component in emotion perception.

Amygdala

The amygdala appears to have a specific role in attention to emotional stimuli.[16] The amygdala is a small, almond-shaped region within the anterior part of the temporal lobe. Several studies of non-human primates and of patients with amygdala lesions, in addition to studies employing functional neuroimaging techniques, have demonstrated the importance of the amygdala in face and eye-gaze identification.[16] Other studies have emphasized the importance of the amygdala for the identification of emotional expressions displayed by others, in particular threat-related emotions such as fear, but also sadness and happiness. In addition, the amygdala is involved in the response to non-facial displays of emotion, including unpleasant auditory, olfactory and gustatory stimuli, and in memory for emotional information.[17] The amygdala receives information from both the thalamus and the cortex; information from the thalamus is rough in detail and the amygdala receives this very quickly, while information from the cortex is much more detailed but is received more slowly.[18] In addition, the amygdala's role in attention modulation toward emotion-specific stimuli may occur via projections from the central nucleus of the amygdala to cholinergic neurons, which lower cortical neuronal activation thresholds and potentiate cortical information processing.[19]

Disordered emotion perception

There is great individual difference in emotion perception and certain groups of people display abnormal processes. Some disorders are in part classified by maladaptive and abnormal emotion perception while others, such as mood disorders, exhibit mood-congruent emotional processing. Whether abnormal processing leads to the exacerbation of certain disorders or is the result of these disorders is yet unclear, however, difficulties or deficits in emotion perception are common among various disorders.

Research investigating face and emotion perception in autistic individuals is inconclusive. Past research has found atypical, piecemeal face-processing strategies among autistic individuals[20] and a better memory for lower versus upper regions of the face and increased abilities to identify partly obscured faces.[21][22] Autistic individuals tend to display deficits in social motivation and experience that may decrease overall experience with faces and this in turn may lead to abnormal cortical specialization for faces and decreased processing efficiency.[23] However, these results have not been adequately replicated and meta-analyses have found little to no differential face processing between typically-developing and autistic individuals although autistic people reliably display worse face memory and eye perception which could mediate face and possibly emotion perception.[24] Individuals with schizophrenia also have difficulties with all types of facial emotion expression perception,[25] incorporating contextual information in making affective decisions,[26] and indeed, facial perception more generally.[27] Neuropathological and structural neuroimaging studies of these patients have demonstrated abnormal neuronal cell integrity and volume reductions in the amygdala, insula, thalamus and hippocampus [28] and studies employing functional neuro-imaging techniques have demonstrated a failure to activate limbic regions in response to emotive stimuli,[29] all of which may contribute to impaired psychosocial functioning.

In patients with major depressive disorder, studies have demonstrated either generalized or specific impairments in the identification of emotional facial expressions, or a bias towards the identification of expressions as sad.[30] Neuro-pathological and structural neuroimaging studies in patients with major depressive disorder have indicated abnormalities within the subgenual anterior cingulate gyrus and volume reductions within the hippocampus, ventral striatal regions and amygdala.[31]

Similarly, anxiety has been commonly associated with individuals being able to perceive threat when in fact none is present,[32] and orient more quickly to threatening cues than other cues.[33] Anxiety has been associated with an enhanced orienting toward threat,[34] a late-stage attention maintenance toward threat,[35] or possibly vigilance-avoidance, or early-stage enhanced orienting and later-stage avoidance.[36] As a form of anxiety, post-traumatic stress disorder (PTSD) has also been linked with abnormal attention toward threatening information, in particular, threatening stimuli which relates to the personally relevant trauma, making such a bias in that context appropriate, but out of context, maladaptive.[37] Such processing of emotion can alter an individual's ability to accurately assess others' emotions as well. Mothers with violence-related PTSD have been noted to show decreased medial prefrontal cortical activation in response to seeing their own and unfamiliar toddlers in helpless or distressed states of mind that is also associated with maternal PTSD symptom severity, self-reported parenting stress, and difficulty in identifying emotions and which, in turn, impacts sensitive caregiving.[38] Moreover, child maltreatment and child abuse have been associated with emotion processing biases as well, most notably toward the experience-specific emotion of anger.[39] Research has found that abused children exhibit attention biases toward angry faces [40][41] such that they tend to interpret even ambiguous faces as angry versus other emotions [42] and have a difficulty disengaging from such expressions [43] while other research has found abused children to demonstrate an attentional avoidance of angry faces.[40] It is believed to be adaptive to attend to angry emotion as this may be a precursor to danger and harm and quick identification of even mild anger cues can facilitate the ability for a child to escape the situation,[43] however, such biases are considered maladaptive when anger is over-identified in inappropriate contexts and this may result in the development of psychopathology.

Research methods

Researchers employ several methods designed to examine biases toward emotional stimuli to determine the salience of particular emotional stimuli, population differences in emotion perception, and also attentional biases toward or away from emotional stimuli. Tasks commonly utilized include the modified Stroop task, the dot probe task, visual search tasks, and spatial cuing tasks. The Stroop task, or modified Stroop task, displays different types of words (e.g., threatening and neutral) in varying colors. The participant is then asked to identify the color of the word while ignoring the actual semantic content. Increased response time to indicate the color of threat words relative to neutral words suggests an attentional bias toward such threat.[44] The Stroop task, however, has some interpretational difficulties[45] in addition to the lack of allowance for the measurement of spatial attention allocation.[46] To address some of the limitations of the Stroop task, the dot probe task displays two words or pictures on a computer screen (either one at the top or left and the other on the bottom or right, respectively) and after a brief stimuli presentation, often less than 1000ms, a probe appears in the location of one of the two stimuli and participants are asked to press a button indicating the location of the probe. Different response times between target (e.g., threat) and neutral stimuli infer attentional biases to the target information with shorter response times for when the probe is in the place of the target stimuli indicating an attention bias for that type of information.[46] In another task that examines spatial attentional allocation, the visual search task asks participants to detect a target stimulus embedded in a matrix of distractors (e.g., an angry face among several neutral or other emotional faces or vice versa). Faster detection times to find emotional stimuli among neutral stimuli or slower detection times to find neutral stimuli among emotional distractors infer an attentional bias for such stimuli.[47][48] The spatial cuing task asks participants to focus on a point located between two rectangles at which point a cue is presented, either in the form of one of the rectangles lighting up or some emotional stimuli appearing within one of the rectangles and this cue either directs attention toward or away from the actual location of the target stimuli. Participants then press a button indicating the location of the target stimuli with faster response times indicating an attention bias toward such stimuli.[49][50] In the morph task participants gradually scroll a facial photograph from the neutral expression to an emotion or from one emotion to another and should indicate at what frame each emotion appears on the face.[51] A recently introduced method consists of presenting dynamic faces (videoclips) and measuring verbal reaction time (into a microphone); it is more precise than previous solutions: verbal responses to six basic emotions differ in hit rates and reaction times.[52]

See also

References

  1. ^ Ekman, P (1993). "Facial expression of emotion". American Psychologist. 48 (4): 384–392. doi:10.1037/0003-066x.48.4.384. PMID 8512154.
  2. ^ Russell, J.A. (1980). "A circumplex model of affect". Journal of Personality and Social Psychology. 39 (6): 1161–1178. doi:10.1037/h0077714. hdl:10983/22919.
  3. ^ Plutchik, R (2002). "Nature of emotions". American Scientist. 89: 349.
  4. ^ Yuki, M.; Maddux, W. W.; Masuda, T. (2007). "Are the windows to the soul the same in the East and West? Cultural differences in using the eyes and mouth as cues to recognize emotions in Japan and the United States". Journal of Experimental Social Psychology. 43 (2): 303–311. doi:10.1016/j.jesp.2006.02.004. hdl:2115/22527.
  5. ^ Barrett, L. F.; Mesquita, B.; Gendron, M. (2011). "Context in Emotion Perception". Current Directions in Psychological Science. 20 (5): 286–290. doi:10.1177/0963721411422522. S2CID 35713636.
  6. ^ Cummings, K.E.; Clements, M.A. (1995). "Analysis of the glottal excitation of emotionally styled and stressed speech". Journal of the Acoustical Society of America. 98 (1): 88–98. Bibcode:1995ASAJ...98...88C. doi:10.1121/1.413664. PMID 7608410.
  7. ^ Frick, R.W. (1985). Communicating emotion: the role of prosodic features. Psychological Bulletin, 97, 412± 429.
  8. ^ Kallinen, K (2005). "Emotional ratings of music excerpts in the Western art music repertoire and their self-organization in the Kohonen Neural Network". Psychology of Music. 33 (4): 373–393. doi:10.1177/0305735605056147. S2CID 145571252.
  9. ^ Vieillard, S.; Peretz, I.; Gosselin, N.; Khalfa, S.; Gagnon, L.; Bouchard, B. (2008). "Happy, sad, scary and peaceful musical excerpts for research on emotions". Cognition & Emotion. 22 (4): 720–752. doi:10.1080/02699930701503567. S2CID 15310907.
  10. ^ Mohn, C.; Argstatter, H.; Wilker, F. W. (2011). "Perception of six basic emotions in music". Psychology of Music. 39 (4): 503–517. doi:10.1177/0305735610378183. S2CID 146141346.
  11. ^ Herz, Rachel S. (2009). "Aromatherapy Facts and Fictions: A Scientific Analysis of Olfactory Effects on Mood, Physiology and Behavior". International Journal of Neuroscience. 119 (2). Informa UK Limited: 263–290. doi:10.1080/00207450802333953. ISSN 0020-7454. PMID 19125379. S2CID 205422999.
  12. ^ Alaoui-Ismaïli, O.; Robin, O.; Rada, H.; Dittmar, A.; Vernet-Maury, E. (1997). "Basic Emotions Evoked by Odorants". Physiology & Behavior. 62 (4). Elsevier BV: 713–720. doi:10.1016/s0031-9384(97)90016-0. ISSN 0031-9384. PMID 284489. S2CID 23785949.
  13. ^ Cannon, Walter B. (1927). "The James-Lange theory of emotion: A critical examination and an alternative theory". The American Journal of Psychology. 39 (1/4): 106–124. doi:10.2307/1415404. JSTOR 1415404. S2CID 27900216.
  14. ^ Cannon, Walter B. (1929). "Organization for Physiological Homeostasis". Physiological Reviews. 9 (3): 399–421. doi:10.1152/physrev.1929.9.3.399.
  15. ^ Schachter, S.; Singer, J. (1962). "Cognitive, Social, and Physiological Determinants of Emotional State". Psychological Review. 69 (5): 379–399. doi:10.1037/h0046234. PMID 14497895.
  16. ^ a b Davis, M; Whalen, PJ (2001). "The amygdala: Vigilance and emotion". Molecular Psychiatry. 6 (1): 3–34. doi:10.1038/sj.mp.4000812. PMID 11244481.
  17. ^ Calder, AJ; Lawrence, AD; Young, AW (2001). "Neuropsychology of fear and loathing". Nature Reviews Neuroscience. 2 (5): 352–363. doi:10.1038/35072584. PMID 11331919. S2CID 195368223.
  18. ^ LeDoux, J. E. (1995). "Emotion: Clues from the brain". Annual Review of Psychology. 46 (1): 209–235. doi:10.1146/annurev.ps.46.020195.001233. PMID 7872730.
  19. ^ Whalen, PJ (1998). "Fear, vigilance, and ambiguity: Initial neuro-imaging studies of the human amygdala". Current Directions in Psychological Science. 7 (6): 177–188. doi:10.1111/1467-8721.ep10836912. S2CID 145438024.
  20. ^ Dawson, G.; Webb, S. J.; McPartland, J. (2005). "Understanding the nature of face processing impairment in autism: insights from behavioral and electrophysiological studies". Developmental Neuropsychology. 27 (3): 403–424. doi:10.1207/s15326942dn2703_6. PMID 15843104. S2CID 2566676.
  21. ^ Tantam, D.; Monoghan, L.; Nicholson, H.; Stirling, J. (1989). "Autistic children's ability to interpret faces: A research note". Journal of Child Psychology and Psychiatry. 30 (4): 623–630. doi:10.1111/j.1469-7610.1989.tb00274.x. PMID 2768363.
  22. ^ Langdell, T (1978). "Recognition of faces: An approach to the study of autism". Journal of Child Psychology and Psychiatry and Allied Disciplines. 19 (3): 255–268. doi:10.1111/j.1469-7610.1978.tb00468.x. PMID 681468.
  23. ^ Dawson, G.; Carver, L.; Meltzoff, A.; Panagiotides, H.; McPartland, J.; Webb, S. (2002). "Neural correlates of face and object recognition in young children with autism spectrum disorder, developmental delay, and typical development". Child Development. 73 (3): 700–717. doi:10.1111/1467-8624.00433. PMC 3651041. PMID 12038546.
  24. ^ Weigelt, S.; Koldewyn, K.; Kanwisher, N. (2012). "Face identity recognition in autism spectrum disorders: A review of behavioral studies". Neuroscience & Biobehavioral Reviews. 36 (3): 1060–1084. doi:10.1016/j.neubiorev.2011.12.008. PMID 22212588. S2CID 13909935.
  25. ^ Kohler, C. G.; Walker, J. B.; Martin, E. A.; Healey, K. M.; Moberg, P. J. (2010). "Facial Emotion Perception in Schizophrenia: A Meta-analytic Review". Schizophrenia Bulletin. 36 (5): 1009–1019. doi:10.1093/schbul/sbn192. PMC 2930336. PMID 19329561.
  26. ^ Kring, A. M.; Campellone, T. R. (2012). "Emotion Perception in Schizophrenia: Context Matters". Emotion Review. 4 (2): 182–186. doi:10.1177/1754073911430140. S2CID 15608777.
  27. ^ Kerr, S. L.; Neale, J. M. (1993). "Emotion perception in schizophrenia: Specific deficit or further evidence of generalized poor performance?". Journal of Abnormal Psychology. 102 (2): 312–318. doi:10.1037/0021-843x.102.2.312. PMID 8315144.
  28. ^ Wright, I. C., Rabe-Hesketh, S., Woodruff, P.W. R., et al (2000) Regional brain structure in schizophrenia: a meta-analysis of volumetric MRI studies. American Journal of Psychiatry, 157, 16^25.
  29. ^ Crespo-Facorro, B.; Paradiso, S.; Andreasen, N. C.; et al. (2001). "Neural mechanisms of anhedonia in schizophrenia". JAMA. 286 (4): 427–435. doi:10.1001/jama.286.4.427. PMID 11466121.
  30. ^ Gur, R. C.; Erwin, R. J.; Gur, R. E.; et al. (1992). "Facial emotion discrimination. II: Behavioral findings in depression". Psychiatry Research. 42 (3): 241–251. doi:10.1016/0165-1781(92)90116-k. PMID 1496056. S2CID 22074921.
  31. ^ Phillips, M. L. (2003). "Understanding the neurobiology of emotion perception: implications for psychiatry". The British Journal of Psychiatry. 182 (3): 190–192. doi:10.1192/bjp.182.3.190. PMID 12611778.
  32. ^ Richards, A.; French, C. C.; Calder, A. J.; Webb, B.; Fox, R.; Young, A. W. (2002). "Anxiety-related bias in the classification of emotionally ambiguous facial expressions". Emotion. 2 (3): 273–287. doi:10.1037/1528-3542.2.3.273. PMID 12899360.
  33. ^ Bar-Haim, Y.; Lamy, D.; Pergamin, L.; Bakermans-Kranenburg, M. J.; van IJzendoorn, M. H. (2007). "Threat-related attentional bias in anxious and nonanxious individuals: A meta-analytic study". Psychological Bulletin. 133 (1): 1–24. doi:10.1037/0033-2909.133.1.1. PMID 17201568.
  34. ^ Williams, J. M. G., Watts, F. N., MacLeod, C., & Matthews, A. (1988). Cognitive psychology and emotional disorders. Chichester, England:Wiley.
  35. ^ Foa, E. B.; Kozak, M. J. (1986). "Emotional processing of fear: Exposure to corrective information". Psychological Bulletin. 99 (1): 20–35. doi:10.1037/0033-2909.99.1.20. PMID 2871574.
  36. ^ Amir, N.; Foa, E. B.; Coles, M. E. (1998). "Automatic activation and strategic avoidance of threat-relevant information in social phobia". Journal of Abnormal Psychology. 107 (2): 285–290. doi:10.1037/0021-843x.107.2.285. PMID 9604557.
  37. ^ Buckley, T. C.; Blanchard, E. B.; Neill, W. T. (2000). "Information processing and PTSD: A review of the empirical literature". Clinical Psychology Review. 28 (8): 1041–1065. doi:10.1016/S0272-7358(99)00030-6. PMID 11098399.
  38. ^ Moser, DA; Aue, T; Suardi, F; Manini, A; Sancho Rossignol, A; Cordero, MI; Merminod, G; Ansermet, F; Rusconi Serpa, S; Favez, N; Schechter, DS (2015). "The relation of general socio-emotional processing to parenting specific behavior: A study of mothers with and without posttraumatic stress disorder". Frontiers in Psychology. 6: 1575. doi:10.3389/fpsyg.2015.01575. PMC 4625041. PMID 26578996.
  39. ^ Cicchetti, D., Toth, S. L., & Maughan, A. (2000). An ecological- transactional model of child maltreatment. In A. J. Sameroff, M. Lewis, & S. M. Miller (Eds.), Handbook of developmental psychopathology (2nd ed., pp. 689-722). New York: Kluwer Academic/Plenum.
  40. ^ a b Pine, D. S.; Mogg, K.; Bradley, B. P.; Montgomery, L.; Monk, C. S.; McClure, E.; et al. (2005). "Attention bias to threat in maltreated children: Implications for vulnerability to stress-related psychopathology". American Journal of Psychiatry. 162 (2): 291–296. doi:10.1176/appi.ajp.162.2.291. PMID 15677593.
  41. ^ Pollak, S. D.; Tolley-Schell, S. A. (2003). "Selective attention to facial emotion in physically abused children". Journal of Abnormal Psychology. 112 (3): 323–338. doi:10.1037/0021-843x.112.3.323. PMID 12943012.
  42. ^ Pollak, S. D.; Kistler, D. J. (2002-06-18). "Early experience is associated with the development of categorical representations for facial expressions of emotion". Proceedings of the National Academy of Sciences. 99 (13): 9072–9076. Bibcode:2002PNAS...99.9072P. doi:10.1073/pnas.142165999. ISSN 0027-8424. PMC 124425. PMID 12072570.
  43. ^ a b Pollak, S. D. (2003). Experience-dependent affective learning and risk for psychopathology in children. In J. A. King, C. F. Ferris, & I. I. Lederhendler (Eds.), Roots of mental illness in children (pp. 102-111). New York: Annals of the New York Academy of Sciences.
  44. ^ Stroop, J. R. (1935). "Studies of interference in serial verbal reactions". Journal of Experimental Psychology. 18 (6): 643–662. doi:10.1037/h0054651. hdl:11858/00-001M-0000-002C-5ADB-7. (Reprinted in 1992 in the Journal of Experimental Psychology: General, 121, 15–23).
  45. ^ Algom, D.; Chajut, E.; Lev, S. (2004). "A rational look at the emotional Stroop paradigm: A generic slowdown, not a Stroop effect". Journal of Experimental Psychology: General. 133 (3): 323–338. doi:10.1037/0096-3445.133.3.323. PMID 15355142.
  46. ^ a b MacLeod, C.; Mathews, A.; Tata, P. (1986). "Attentional bias in the emotional disorders". Journal of Abnormal Psychology. 95 (1): 15–20. doi:10.1037/0021-843x.95.1.15. PMID 3700842.
  47. ^ Öhman, A.; Flykt, A.; Esteves, F. (2001). "Emotion drives attention: Detecting the snake in the grass". Journal of Experimental Psychology: General. 130 (3): 466–478. doi:10.1037/0096-3445.130.3.466. PMID 11561921.
  48. ^ Rinck, M.; Becker, E. S.; Kellermann, J.; Roth, W. T. (2003). "Selective attention in anxiety: Distraction and enhancement in visual search". Depression and Anxiety. 18 (1): 18–28. doi:10.1002/da.10105. PMID 12900949. S2CID 40092068.
  49. ^ Fox, E.; Russo, R.; Bowles, R.; Dutton, K. (2001). "Do threatening stimuli draw or hold visual attention in subclinical anxiety?" (PDF). Journal of Experimental Psychology: General. 130 (4): 681–700. doi:10.1037/0096-3445.130.4.681. PMC 1924776. PMID 11757875.
  50. ^ Posner, M. I. (1980). "Orienting of attention". Quarterly Journal of Experimental Psychology. 32 (1): 3–25. doi:10.1080/00335558008248231. PMID 7367577. S2CID 2842391.
  51. ^ Niedenthal, Paula M.; Halberstadt, Jamin B.; Margolin, Jonathan; Innes‐Ker, Åse H. (2000). "Emotional state and the detection of change in facial expression of emotion". European Journal of Social Psychology. 30 (2): 211–222. doi:10.1002/(sici)1099-0992(200003/04)30:2<211::aid-ejsp988>3.0.co;2-3. ISSN 1099-0992.
  52. ^ Kosonogov, Vladimir; Titova, Alisa (2019). "Recognition of all basic emotions varies in accuracy and reaction time: A new verbal method of measurement" (PDF). International Journal of Psychology. 54 (5): 582–588. doi:10.1002/ijop.12512. ISSN 1464-066X. PMID 30010190. S2CID 51627393.

External links