Emotional prosody

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Emotional prosody is characterized as an individual's tone of voice in speech that is conveyed through changes in pitch, loudness, timbre, speech rate, and pauses which is different from linguistic and semantic information. It can be isolated from linguistics and interacts with verbal content (e.g. sarcasm). It is perceived or decoded slightly worse than facial expressions but accuracy varies with emotions. Anger and sadness are perceived most easily, followed by fear and happiness, with disgust being the most poorly perceived.[1][self-published source?]

Prosody itself has multiple aspects: lexical prosody, phrasal prosody, and clausal prosody.

Production of vocal emotion[edit]

Studies have found that some emotions, such as fear, joy and anger, are portrayed at a higher frequency than emotions such as sadness.[2]

  • Anger: Anger can be divided into two types: "anger" and "hot anger". In comparison to neutral speech, anger is produced with a lower pitch, higher intensity, more energy (500 Hz) across the vocalization, higher first formant (first sound produced) and faster attack times at voice onset (the start of speech). "Hot anger", in contrast, is produced with a higher, more varied pitch, and even greater energy (2000 Hz).[3]
  • Disgust: In comparison to neutral speech, disgust is produced with a lower, downward directed pitch, with energy (500 Hz), lower first formant, and fast attack times similar to anger. Less variation and shorter durations are also characteristics of disgust.[3]
  • Fear: Fear can be divided into two types: "panic" and "anxiety". In comparison to neutral speech, fearful emotions have a higher pitch, little variation, lower energy, and a faster speech rate with more pauses.[3]
  • Sadness: In comparison to neutral speech, sad emotions are produced with a higher pitch, less intensity but more vocal energy (2000 Hz), longer duration with more pauses, and a lower first formant.[3]

Perception of vocal emotion[edit]

Decoding emotions in speech includes three (3) stages: determining acoustic features, creating meaningful connections with these features, and processing the acoustic patterns in relation to the connections established. In the processing stage, connections with basic emotional knowledge is stored separately in memory network specific to associations. These associations can be used to form a baseline for emotional expressions encountered in the future. Emotional meanings of speech are implicitly and automatically registered after the circumstances, importance and other surrounding details of an event have been analyzed.[4]

On average, listeners are able to perceive intended emotions exhibited to them at a rate significantly better than chance (chance=approximately 10%).[3] However, error rates are also high. This is partly due to the observation that listeners are more accurate at emotional inference from particular voices and perceive some emotions better than others.[2] Vocal expressions of anger and sadness are perceived most easily, fear and happiness are only moderately well-perceived, and disgust has low perceptibility.[1][self-published source?]

The brain in vocal emotions[edit]

Language can be split into two components: the verbal and vocal channels. The verbal channel is the semantic content made by the speaker's chosen words. In the verbal channel, the semantic content of the speakers words determines the meaning of the sentence. The way a sentence is spoken however, can change its meaning which is the vocal channel. This channel of language conveys emotions felt by the speaker and gives us as listeners a better idea of the intended meaning. Nuances in this channel are expressed through intonation, intensity, a rhythm which combined for prosody. Usually these channels convey the same emotion, but sometimes they differ. Sarcasm and irony are two forms of humor based on this incongruent style.[5]

Neurological processes integrating verbal and vocal (prosodic) components are relatively unclear. However, it is assumed that verbal content and vocal are processed in different hemispheres of the brain. Verbal content composed of syntactic and semantic information is processed the left hemisphere.Syntactic information is processed primarily in the frontal regions and a small part of the temporal lobe of the brain while semantic information is processed primarily in the temporal regions with a smaller part of the frontal lobes incorporated. In contrast, prosody is processed primarily in the same pathway as verbal content, but in the right hemisphere. Neuroimaging studies using functional magnetic resonance imaging (fMRI) machines provide further support for this hemisphere lateralization and temporo-frontal activation. Some studies however show evidence that prosody perception is not exclusively lateralized to the right hemisphere and may be more bilateral. There is some evidence that the basal ganglia may also play an important role in the perception of prosody.[5]

Impairment of emotion recognition[edit]

Deficits in expressing and understanding prosody, caused by right hemisphere lesions, are known as aprosodias. These can manifest in different forms and in various mental illnesses or diseases. Aprosodia can be caused by stroke and alcohol abuse as well.The types of aprosodia include: motor (the inability to produce vocal inflection, expressive (when brain limitations and not motor functions are the cause of this inability), and receptive (when a person cannot decipher the emotional speech).[6]

It has been found that it gets increasingly difficult to recognize vocal expressions of emotion with increasing age. Older adults have slightly more difficulty labeling vocal expressions of emotion, particularly sadness and anger) than young adults but have a much greater difficulty integrating vocal emotions and corresponding facial expressions. A possible explanation for this difficulty is that combining two sources of emotion requires greater activation of emotion areas of the brain, in which adults show decreased volume and activity. Another possible explanation is that hearing loss could have led to a mishearing of vocal expressions. High frequency hearing loss is known to begin occurring around the age of 50, particularly in men.[7]

Because the right hemisphere of the brain is associated with prosody, patients with right hemisphere lesions have difficulty varying speech patterns to convey emotion. Their speech may therefore sound monotonous. In addition, people with right-hemisphere damage have been studied to be impaired when it comes to identifying the emotion in intoned sentences.

Difficulty in decoding both syntactic and affective prosody is also found in people with AUTISM SPECTRUM DISORDER AND SCHIZOPHRENia, where "patients have deficits in a large number of functional domains, including social skills and social cognition. These social impairments consist of difficulties in perceiving, understanding, anticipating and reacting to social cues that are crucial for normal social interaction." This has been determined in multiple studies, such as Hoekert et al.'s 2017 study on emotional prosody in schizophrenia, which illustrated that more research must be done to fully confirm the correlation between the illness and emotional prosody. However, people with schizophrenia have no problem deciphering non-emotional prosody.[8]

Non-linguistic emotional prosody[edit]

Emotional states such as happiness, sadness, anger, and disgust can be determined solely based on the acoustic structure of a non-linguistic speech act. These acts can be grunts, sighs, exclamations, etc. There is some research the supports the notion that these non-linguistic acts are universal, eliciting the same assumptions even from speakers of different languages.

In addition, it has been proven that emotion can be expressed in non-linguistic vocalizations differently than in speech. As Lauka et al. state: Speech requires highly precise and coordinated movement of the articulators (e.g., lips, tongue, and larynx) in order to transmit linguistic information, whereas non-linguistic vocalizations are not constrained by linguistic codes and thus do not require such precise articulations. This entails that non-linguistic vocalizations can exhibit larger ranges for many acoustic features than prosodic expressions.

In their study, actors were instructed to vocalize an array of different emotions without words. The study showed that listeners could identify a wide range of positive and negative emotions above chance. However, emotions like guilt and pride were less easily recognized.[9]

In a 2015 study by Verena Kersken, Klaus Zuberbühler and Juan-Carlos Gomez, non-linguistic vocalizations of infants were presented to adults to see if the adults could distinguish from infant vocalizations indicating requests for help, pointing to an object, or indicating an event. Infants show different prosodic elements in crying, depending on what they are crying for. They also have differing outbursts for positive and negative emotional states. Decipherment ability of this information was determined to be applicable across cultures and independent of the adult’s level of experience with infants.

Gender differences[edit]

Men and women differ in both how they use language and also how they understand it. It is known that there is a different in the rate of speech, the range of pitch, and the duration of speech, and pitch slope (Fitzsimmons et al). For example, "In a study of relationship of spectral and prosodic signs, it was established that the dependence of pitch and duration differed in men and women uttering the sentences in affirmative and inquisitive intonation. Tempo of speech, pitch range, and pitch steepness differ between the genders" (Nesic et al). One such illustration is how women are more likely to speak faster, elongate the ends of words, and raise their pitch at the end of sentences.

Women and men are also different in how they neurologically process emotional prosody. In an fMRI study, men showed a stronger activation in more cortical areas than female subjects when processing the meaning or manner of an emotional phrase. In the manner task, men had more activation in the bilateral middle temporal gyri For women, the only area of significance was the right posterior cerebellar lobe. Male subjects in this study showed stronger activation in the frontomedian cortex, and on average needed a longer response time than female subjects. This result was interpreted to mean that men need to make conscious inferences about the acts and intentions of the speaker, while women may do this sub-consciously. Therefore, men needed to integrate linguistic semantics and emotional intent "at a higher stage than the semantic processing stage."[10]

Considerations[edit]

Most research regarding vocal expression of emotion has been studied through the use of synthetic speech or portrayals of emotion by professional actors. Little research has been done with spontaneous, "natural" speech samples. These artificial speech samples have been considered to be close to natural speech but specifically portrayals by actors may be influenced stereotypes of emotional vocal expression and may exhibit intensified characteristics of speech skewing listeners perceptions. Another consideration lies in listeners individual perceptions. Studies typically take the average of responses but few examine individual differences in great depth. This may provide a better insight into the vocal expressions of emotions.[3]

References[edit]

  1. ^ a b "The Social and Emotional Voice" (PDF). Retrieved 29 March 2012. 
  2. ^ a b Bachorowski, Jo-Anne (1999). "Vocal Expression and Perception of Emotion". Current Directions in Psychological Science. 8 (2): 53–57. doi:10.1111/1467-8721.00013. 
  3. ^ a b c d e f Sauter, Disa A.; Eisner, Frank; Calder, Andrew J.; Scott, Sophie K. (2010). "Perceptual cues in nonverbal vocal expressions of emotion". The Quarterly Journal of Experimental Psychology. 63 (11): 2251–2272. doi:10.1080/17470211003721642.  closed access publication – behind paywall
  4. ^ Pell, Marc D.; Kotz, Sonja A. (2011). "On the Time Course of Vocal Emotion Recognition". PLoS ONE. 6 (11): e27256. doi:10.1371/journal.pone.0027256. Retrieved 1 May 2012.  open access publication – free to read
  5. ^ a b Berckmoes, Celine; Guy Vingerhoets (2004). "Neural Foundations of Emotional Speech Processing". Current Directions in Psychological Science. 13 (5): 182–185. doi:10.1111/j.0963-7214.2004.00303.x. 
  6. ^ Gorelick PB, Ross ED (May 1987). "The aprosodias: further functional-anatomical evidence for the organisation of affective language in the right hemisphere" (PDF). Journal of Neurology, Neurosurgery, and Psychiatry. 50 (5): 553–560. doi:10.1136/jnnp.50.5.553. PMC 1031966Freely accessible. PMID 2438386. 
  7. ^ Ryan, Melissa; Murray, Janice; Ruffman, Ted (2010). "Aging and the perception of emotion: Processing vocal emotions alone and with faces". Experimental Aging Research. 36 (1): 1–22. doi:10.1080/03610730903418372. 
  8. ^ Hoekert, L. M. (2009). "Impaired recognition and expression of emotional prosody in schizophrenia: review and meta-analysis". Beyond what is being said: emotional prosody (PDF). 
  9. ^ Laukka P, Elfenbein HA, Söder N, Nordström H, Althoff J, Chui W, Iraki FK, Rockstuhl T, Thingujam NS (July 2013). "Cross-cultural decoding of positive and negative non-linguistic emotion vocalizations". Frontiers in Psychology. 4. doi:10.3389/fpsyg.2013.00353. PMC 3728469Freely accessible. PMID 23914178. 
  10. ^ Fitzsimons M, Sheahan N, Staunton H (July 2001). "Gender and the integration of acoustic dimensions of prosody: implications for clinical studies". Brain and Language. 78 (1): 94–108. doi:10.1006/brln.2000.2448. PMID 11412018.