Not all members of a category are equally good members; for example, a robin is generally considered to be a more typical member of the category ‘birds’ than an ostrich. Similarly, one vocal expression of, for example, happiness can be more typical than another expression of the same emotion, though both expressions clearly are perceived as ‘happy’. This chapter presents ongoing studies investigating the determinants of the typicality (graded structure) of vocal emotion expressions. In two experiments, separate groups of judges rated expressive speech stimuli (both acted and spontaneous expressions) with regard to typicality, ideal (suitability to express the respective emotion), and frequency of instantiation. A measure of similarity to central tendency was also obtained from listener judgments. Partial correlations and multiple regression analyses revealed that similarity to ideal, and not frequency of instantiation or similarity to central tendency, explained most variance in judged typicality. In other words, the typicality of vocal expressions was mainly determined by their similarity to ideal category members. Because ideals depend on the goals that people have, they can be independent of the particular category members that a person usually encounters. Thus it is argued that these results may indicate that prototypical vocal expressions are best characterized as goal-derived categories, rather than common taxonomic categories. This could explain how prototypical expressions can be acoustically distinct and highly recognizable, while at the same time occur relatively rarely in everyday speech.
The ability of young and old adults to recognize emotions from vocal expressions and music performances was compared. The stimuli consisted of a) acted speech (anger, disgust, fear, happiness, and sadness; each posed with both weak and strong emotion intensity), b) synthesized speech (anger, fear, happiness, and sadness), and c) short melodies played on the electric guitar (anger, fear, happiness, and sadness; each played with both weak and strong emotion intensity). Both groups of listeners rated the stimuli using forced-choice and also rated the emotion intensity of each stimulus. Results showed emotion-specific age-related differences in recognition accuracy. Old adults consistently received significantly lower recognition accuracy for negative, but not for positive, emotions across all types of stimuli. Age-related differences in recognition of emotion intensity were also found. The results show the importance of considering individual emotions in studies on age-related differences in emotion recognition.
This study investigated the effects of anxiety on nonverbal aspects of speech using data collected in the framework of a large study of social phobia treatment. The speech of social phobics (N = 71) was recorded during an anxiogenic public speaking task both before and after treatment. The speech samples were analyzed with respect to various acoustic parameters related to pitch, loudness, voice quality, and temporal apsects of speech. The samples were further content-masked by low-pass filtering (which obscures the linguistic content of the speech but preserves nonverbal affective cues) and subjected to listening tests. Results showed that a decrease in experienced state anxiety after treatment was accompanied by corresponding decreases in a) several acoustic parameters (i.e., mean and maximum voice pitch, high-frequency components in the energy spectrum, and proportion of silent pauses), and b) listeners' perceived level of nervousness. Both speakers' self-ratings of state anxiety and listeners' ratings of perceived nervousness were further correlated with similar acoustic parameters. The results complement earlier studies on vocal affect expression which have been conducted on posed, rather than authentic, emotional speech.
The majority of previous studies on vocal expression have been conducted on posed expressions. In contrast, we utilized a large corpus of authentic affective speech recorded from real-life voice-controlled telephone services. Listeners rated a selection of 200 utterances from this corpus with regard to level of perceived irritation, resignation, neutrality, and emotion intensity. The selected utterances came from 64 different speakers who each provided both neutral and affective stimuli. All utterances were further automatically analyzed regarding a comprehensive set of acoustic measures related to F0, intensity, formants, voice source, and temporal characteristics of speech. Results first showed that several significant acoustic differences were found between utterances classified as neutral and utterances classified as irritated or resigned using a within-persons design. Second, listeners’ ratings on each scale were associated with several acoustic measures. In general the acoustic correlates of irritation, resignation, and emotion intensity were similar to previous findings obtained with posed expressions, though the effect sizes were smaller for the authentic expressions. Third, automatic classification (using LDA classifiers both with and without speaker adaptation) of irritation, resignation, and neutral performed at a level comparable to human performance, though human listeners and machines did not necessarily classify individual utterances similarly. Fourth, clearly perceived exemplars of irritation and resignation were rare in our corpus. These findings were discussed in relation to future research.
Music is present in many sport and exercise situations, but empirical investigations on the motives for listening to music in sports remain scarce. In this study, Swedish elite athletes (N = 252) answered a questionnaire that focused on the emotional and motivational uses of music in sports and exercise. The questionnaire contained both quantitative items that assessed the prevalence of various uses of music, and open-ended items that targeted specific emotional episodes in relation to music in sports. Results showed that the athletes most often reported listening to music during pre-event preparations, warm-up, and training sessions; and the most common motives for listening to music were to increase pre-event activation, positive affect, motivation, performance levels and to experience flow. The athletes further reported that they mainly experienced positive affective states (e.g., happiness, alertness, confidence, relaxation) in relation to music in sports, and also reported on their beliefs about the causes of the musical emotion episodes in sports. In general, the results suggest that the athletes used music in purposeful ways in order to facilitate their training and performance.
Individuals with schizophrenia show reliable deficits in the ability to recognize emotions from vocal expressions. Here, we examined emotion recognition ability in 23 schizophrenia patients relative to 17 healthy controls using a stimulus battery with well-characterized acoustic features. We further evaluated performance deficits relative to ancillary assessments of underlying pitch perception abilities. As predicted, patients showed reduced emotion recognition ability across a range of emotions, which correlated with impaired basic tone matching abilities. Emotion identification deficits were strongly related to pitch-based acoustic cues such as mean and variability of fundamental frequency. Whereas healthy subjects' performance varied as a function of the relative presence or absence of these cues, with higher cue levels leading to enhanced performance, schizophrenia patients showed significantly less variation in performance as a function of cue level. In contrast to pitch-based cues, both groups showed equivalent variation in performance as a function of intensity-based cues. Finally, patients were less able than controls to differentiate between expressions with high and low emotion intensity, and this deficit was also correlated with impaired tone matching ability. Both emotion identification and intensity rating deficits were unrelated to valence of intended emotions. Deficits in both auditory emotion identification and more basic perceptual abilities correlated with impaired functional outcome. Overall, these findings support the concept that auditory emotion identification deficits in schizophrenia reflect, at least in part, a relative inability to process critical acoustic characteristics of prosodic stimuli and that such deficits contribute to poor global outcome.