Mirror neurons are theorized to serve as a neural substrate for spoken language in humans, but the existence and functions of auditory-vocal mirror neurons in the human brain remain largely matters of speculation. Songbirds resemble humans in their capacity for vocal learning and depend on their learned songs to facilitate courtship and individual recognition. Recent neurophysiological studies have detected putative auditory-vocal mirror neurons in a sensorimotor region of the songbird's brain that plays an important role in expressive and receptive aspects of vocal communication. This review discusses the auditory and motor-related properties of these cells, considers their potential role on song learning and communication in relation to classical studies of birdsong, and points to the circuit and developmental mechanisms that may give rise to auditory-vocal mirroring in the songbird's brain.
Lewandowski, Brian; Vyssotski, Alexei; Hahnloser, Richard H R; Schmidt, Marc
Communication between auditory and vocal motor nuclei is essential for vocal learning. In songbirds, the nucleus interfacialis of the nidopallium (NIf) is part of a sensorimotor loop, along with auditory nucleus avalanche (Av) and song system nucleus HVC, that links the auditory and song systems. Most of the auditory information comes through this sensorimotor loop, with the projection from NIf to HVC representing the largest single source of auditory information to the song system. In addition to providing the majority of HVC's auditory input, NIf is also the primary driver of spontaneous activity and premotor-like bursting during sleep in HVC. Like HVC and RA, two nuclei critical for song learning and production, NIf exhibits behavioral-state dependent auditory responses and strong motor bursts that precede song output. NIf also exhibits extended periods of fast gamma oscillations following vocal production. Based on the converging evidence from studies of physiology and functional connectivity it would be reasonable to expect NIf to play an important role in the learning, maintenance, and production of song. Surprisingly, however, lesions of NIf in adult zebra finches have no effect on song production or maintenance. Only the plastic song produced by juvenile zebra finches during the sensorimotor phase of song learning is affected by NIf lesions. In this review, we carefully examine what is known about NIf at the anatomical, physiological, and behavioral levels. We reexamine conclusions drawn from previous studies in the light of our current understanding of the song system, and establish what can be said with certainty about NIf's involvement in song learning, maintenance, and production. Finally, we review recent theories of song learning integrating possible roles for NIf within these frameworks and suggest possible parallels between NIf and sensorimotor areas that form part of the neural circuitry for speech processing in humans. Copyright © 2013 Elsevier
Prather, J F; Peters, S; Nowicki, S; Mooney, R
Brain mechanisms for communication must establish a correspondence between sensory and motor codes used to represent the signal. One idea is that this correspondence is established at the level of single neurons that are active when the individual performs a particular gesture or observes a similar gesture performed by another individual. Although neurons that display a precise auditory-vocal correspondence could facilitate vocal communication, they have yet to be identified. Here we report that a certain class of neurons in the swamp sparrow forebrain displays a precise auditory-vocal correspondence. We show that these neurons respond in a temporally precise fashion to auditory presentation of certain note sequences in this songbird's repertoire and to similar note sequences in other birds' songs. These neurons display nearly identical patterns of activity when the bird sings the same sequence, and disrupting auditory feedback does not alter this singing-related activity, indicating it is motor in nature. Furthermore, these neurons innervate striatal structures important for song learning, raising the possibility that singing-related activity in these cells is compared to auditory feedback to guide vocal learning.
Arie, Miri; Henkin, Yael; Lamy, Dominique; Tetin-Schneider, Simona; Apter, Alan; Sadeh, Avi; Bar-Haim, Yair
Because abnormal Auditory Efferent Activity (AEA) is associated with auditory distortions during vocalization, we tested whether auditory processing is impaired during vocalization in children with Selective Mutism (SM). Participants were children with SM and abnormal AEA, children with SM and normal AEA, and normally speaking controls, who had to detect aurally presented target words embedded within word lists under two conditions: silence (single task), and while vocalizing (dual task). To ascertain specificity of auditory-vocal deficit, effects of concurrent vocalizing were also examined during a visual task. Children with SM and abnormal AEA showed impaired auditory processing during vocalization relative to children with SM and normal AEA, and relative to control children. This impairment is specific to the auditory modality and does not reflect difficulties in dual task per se. The data extends previous findings suggesting that deficient auditory processing is involved in speech selectivity in SM.
Cardin, Jessica A; Raksin, Jonathan N; Schmidt, Marc F
Sensorimotor integration in the avian song system is crucial for both learning and maintenance of song, a vocal motor behavior. Although a number of song system areas demonstrate both sensory and motor characteristics, their exact roles in auditory and premotor processing are unclear. In particular, it is unknown whether input from the forebrain nucleus interface of the nidopallium (NIf), which exhibits both sensory and premotor activity, is necessary for both auditory and premotor processing in its target, HVC. Here we show that bilateral NIf lesions result in long-term loss of HVC auditory activity but do not impair song production. NIf is thus a major source of auditory input to HVC, but an intact NIf is not necessary for motor output in adult zebra finches.
Bartels-Velthuis, A.A.; Jenner, J.A.; van de Willige, G.; van Os, J.; Wiersma, D.
Background Hearing voices occurs in middle childhood, but little is known about prevalence, aetiology and immediate consequences. Aims To investigate prevalence, developmental risk factors and behavioural correlates of auditory vocal hallucinations in 7- and 8-year-olds. Method Auditory vocal
Full Text Available Introduction Enhanced auditory perception in musicians is likely to result from auditory perceptual learning during several years of training and practice. Many studies have focused on biological processing of auditory stimuli among musicians. However, there is a lack of literature on temporal resolution and active auditory discrimination skills in vocal musicians. Objective The aim of the present study is to assess temporal resolution and active auditory discrimination skill in vocal musicians. Method The study participants included 15 vocal musicians with a minimum professional experience of 5 years of music exposure, within the age range of 20 to 30 years old, as the experimental group, while 15 age-matched non-musicians served as the control group. We used duration discrimination using pure-tones, pulse-train duration discrimination, and gap detection threshold tasks to assess temporal processing skills in both groups. Similarly, we assessed active auditory discrimination skill in both groups using Differential Limen of Frequency (DLF. All tasks were done using MATLab software installed in a personal computer at 40dBSL with maximum likelihood procedure. The collected data were analyzed using SPSS (version 17.0. Result Descriptive statistics showed better threshold for vocal musicians compared with non-musicians for all tasks. Further, independent t-test showed that vocal musicians performed significantly better compared with non-musicians on duration discrimination using pure tone, pulse train duration discrimination, gap detection threshold, and differential limen of frequency. Conclusion The present study showed enhanced temporal resolution ability and better (lower active discrimination threshold in vocal musicians in comparison to non-musicians.
Fukushima, Makoto; Saunders, Richard C; Leopold, David A; Mishkin, Mortimer; Averbeck, Bruno B
The mammalian auditory cortex integrates spectral and temporal acoustic features to support the perception of complex sounds, including conspecific vocalizations. Here we investigate coding of vocal stimuli in different subfields in macaque auditory cortex. We simultaneously measured auditory evoked potentials over a large swath of primary and higher order auditory cortex along the supratemporal plane in three animals chronically using high-density microelectrocorticographic arrays. To evaluate the capacity of neural activity to discriminate individual stimuli in these high-dimensional datasets, we applied a regularized multivariate classifier to evoked potentials to conspecific vocalizations. We found a gradual decrease in the level of overall classification performance along the caudal to rostral axis. Furthermore, the performance in the caudal sectors was similar across individual stimuli, whereas the performance in the rostral sectors significantly differed for different stimuli. Moreover, the information about vocalizations in the caudal sectors was similar to the information about synthetic stimuli that contained only the spectral or temporal features of the original vocalizations. In the rostral sectors, however, the classification for vocalizations was significantly better than that for the synthetic stimuli, suggesting that conjoined spectral and temporal features were necessary to explain differential coding of vocalizations in the rostral areas. We also found that this coding in the rostral sector was carried primarily in the theta frequency band of the response. These findings illustrate a progression in neural coding of conspecific vocalizations along the ventral auditory pathway.
Full Text Available BACKGROUND: We ordinarily perceive our voice sound as occurring simultaneously with vocal production, but the sense of simultaneity in vocalization can be easily interrupted by delayed auditory feedback (DAF. DAF causes normal people to have difficulty speaking fluently but helps people with stuttering to improve speech fluency. However, the underlying temporal mechanism for integrating the motor production of voice and the auditory perception of vocal sound remains unclear. In this study, we investigated the temporal tuning mechanism integrating vocal sensory and voice sounds under DAF with an adaptation technique. METHODS AND FINDINGS: Participants produced a single voice sound repeatedly with specific delay times of DAF (0, 66, 133 ms during three minutes to induce 'Lag Adaptation'. They then judged the simultaneity between motor sensation and vocal sound given feedback. We found that lag adaptation induced a shift in simultaneity responses toward the adapted auditory delays. This indicates that the temporal tuning mechanism in vocalization can be temporally recalibrated after prolonged exposure to delayed vocal sounds. Furthermore, we found that the temporal recalibration in vocalization can be affected by averaging delay times in the adaptation phase. CONCLUSIONS: These findings suggest vocalization is finely tuned by the temporal recalibration mechanism, which acutely monitors the integration of temporal delays between motor sensation and vocal sound.
Neuhoff, John G; Hamilton, Grace R; Gittleson, Amanda L; Mejia, Adolfo
Infant vocalizations and "looming sounds" are classes of environmental stimuli that are critically important to survival but can have dramatically different emotional valences. Here, we simultaneously presented listeners with a stationary infant vocalization and a 3D virtual looming tone for which listeners made auditory time-to-arrival judgments. Negatively valenced infant cries produced more cautious (anticipatory) estimates of auditory arrival time of the tone over a no-vocalization control. Positively valenced laughs had the opposite effect, and across all conditions, men showed smaller anticipatory biases than women. In Experiment 2, vocalization-matched vocoded noise stimuli did not influence concurrent auditory time-to-arrival estimates compared with a control condition. In Experiment 3, listeners estimated the egocentric distance of a looming tone that stopped before arriving. For distant stopping points, women estimated the stopping point as closer when the tone was presented with an infant cry than when it was presented with a laugh. For near stopping points, women showed no differential effect of vocalization type. Men did not show differential effects of vocalization type at either distance. Our results support the idea that both the sex of the listener and the emotional valence of infant vocalizations can influence auditory motion perception and can modulate motor responses to other behaviorally relevant environmental sounds. We also find support for previous work that shows sex differences in emotion processing are diminished under conditions of higher stress.
Liu, Ying; Fan, Hao; Li, Jingting; Jones, Jeffery A; Liu, Peng; Zhang, Baofeng; Liu, Hanjun
When people hear unexpected perturbations in auditory feedback, they produce rapid compensatory adjustments of their vocal behavior. Recent evidence has shown enhanced vocal compensations and cortical event-related potentials (ERPs) in response to attended pitch feedback perturbations, suggesting that this reflex-like behavior is influenced by selective attention. Less is known, however, about auditory-motor integration for voice control during divided attention. The present cross-modal study investigated the behavioral and ERP correlates of auditory feedback control of vocal pitch production during divided attention. During the production of sustained vowels, 32 young adults were instructed to simultaneously attend to both pitch feedback perturbations they heard and flashing red lights they saw. The presentation rate of the visual stimuli was varied to produce a low, intermediate, and high attentional load. The behavioral results showed that the low-load condition elicited significantly smaller vocal compensations for pitch perturbations than the intermediate-load and high-load conditions. As well, the cortical processing of vocal pitch feedback was also modulated as a function of divided attention. When compared to the low-load and intermediate-load conditions, the high-load condition elicited significantly larger N1 responses and smaller P2 responses to pitch perturbations. These findings provide the first neurobehavioral evidence that divided attention can modulate auditory feedback control of vocal pitch production.
Liu, Ying; Hu, Huijing; Jones, Jeffery A; Guo, Zhiqiang; Li, Weifeng; Chen, Xi; Liu, Peng; Liu, Hanjun
Speakers rapidly adjust their ongoing vocal productions to compensate for errors they hear in their auditory feedback. It is currently unclear what role attention plays in these vocal compensations. This event-related potential (ERP) study examined the influence of selective and divided attention on the vocal and cortical responses to pitch errors heard in auditory feedback regarding ongoing vocalisations. During the production of a sustained vowel, participants briefly heard their vocal pitch shifted up two semitones while they actively attended to auditory or visual events (selective attention), or both auditory and visual events (divided attention), or were not told to attend to either modality (control condition). The behavioral results showed that attending to the pitch perturbations elicited larger vocal compensations than attending to the visual stimuli. Moreover, ERPs were likewise sensitive to the attentional manipulations: P2 responses to pitch perturbations were larger when participants attended to the auditory stimuli compared to when they attended to the visual stimuli, and compared to when they were not explicitly told to attend to either the visual or auditory stimuli. By contrast, dividing attention between the auditory and visual modalities caused suppressed P2 responses relative to all the other conditions and caused enhanced N1 responses relative to the control condition. These findings provide strong evidence for the influence of attention on the mechanisms underlying the auditory-vocal integration in the processing of pitch feedback errors. In addition, selective attention and divided attention appear to modulate the neurobehavioral processing of pitch feedback errors in different ways. © 2015 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.
Full Text Available We ordinarily perceive our voice sound as occurring simultaneously with vocal production, but the sense of simultaneity in vocalization can be easily interrupted by delayed auditory feedback (DAF. DAF causes normal people to have difficulty speaking fluently but helps people with stuttering to improve speech fluency. However, the underlying temporal mechanism for integrating the motor production of voice and the auditory perception of vocal sound remains unclear. In this study, we investigated the temporal tuning mechanism integrating vocal sensory and voice sounds under DAF with an adaptation technique. Participants read some sentences with specific delay times of DAF (0, 30, 75, 120 ms during three minutes to induce ‘Lag Adaptation’. After the adaptation, they then judged the simultaneity between motor sensation and vocal sound given feedback in producing simple voice but not speech. We found that speech production with lag adaptation induced a shift in simultaneity responses toward the adapted auditory delays. This indicates that the temporal tuning mechanism in vocalization can be temporally recalibrated after prolonged exposure to delayed vocal sounds. These findings suggest vocalization is finely tuned by the temporal recalibration mechanism, which acutely monitors the integration of temporal delays between motor sensation and vocal sound.
Houix, Olivier; Voisin, Frédéric; Misdariis, Nicolas; Susini, Patrick
Imitative behaviors are widespread in humans, in particular whenever two persons communicate and interact. Several tokens of spoken languages (onomatopoeias, ideophones, and phonesthemes) also display different degrees of iconicity between the sound of a word and what it refers to. Thus, it probably comes at no surprise that human speakers use a lot of imitative vocalizations and gestures when they communicate about sounds, as sounds are notably difficult to describe. What is more surprising is that vocal imitations of non-vocal everyday sounds (e.g. the sound of a car passing by) are in practice very effective: listeners identify sounds better with vocal imitations than with verbal descriptions, despite the fact that vocal imitations are inaccurate reproductions of a sound created by a particular mechanical system (e.g. a car driving by) through a different system (the voice apparatus). The present study investigated the semantic representations evoked by vocal imitations of sounds by experimentally quantifying how well listeners could match sounds to category labels. The experiment used three different types of sounds: recordings of easily identifiable sounds (sounds of human actions and manufactured products), human vocal imitations, and computational “auditory sketches” (created by algorithmic computations). The results show that performance with the best vocal imitations was similar to the best auditory sketches for most categories of sounds, and even to the referent sounds themselves in some cases. More detailed analyses showed that the acoustic distance between a vocal imitation and a referent sound is not sufficient to account for such performance. Analyses suggested that instead of trying to reproduce the referent sound as accurately as vocally possible, vocal imitations focus on a few important features, which depend on each particular sound category. These results offer perspectives for understanding how human listeners store and access long
Vasconcelos, Raquel O.; Fonseca, Paulo J.; Amorim, M. Clara P.; Ladich, Friedrich
Many fishes rely on their auditory skills to interpret crucial information about predators and prey, and to communicate intraspecifically. Few studies, however, have examined how complex natural sounds are perceived in fishes. We investigated the representation of conspecific mating and agonistic calls in the auditory system of the Lusitanian toadfish Halobatrachus didactylus, and analysed auditory responses to heterospecific signals from ecologically relevant species: a sympatric vocal fish (meagre Argyrosomus regius) and a potential predator (dolphin Tursiops truncatus). Using auditory evoked potential (AEP) recordings, we showed that both sexes can resolve fine features of conspecific calls. The toadfish auditory system was most sensitive to frequencies well represented in the conspecific vocalizations (namely the mating boatwhistle), and revealed a fine representation of duration and pulsed structure of agonistic and mating calls. Stimuli and corresponding AEP amplitudes were highly correlated, indicating an accurate encoding of amplitude modulation. Moreover, Lusitanian toadfish were able to detect T. truncatus foraging sounds and A. regius calls, although at higher amplitudes. We provide strong evidence that the auditory system of a vocal fish, lacking accessory hearing structures, is capable of resolving fine features of complex vocalizations that are probably important for intraspecific communication and other relevant stimuli from the auditory scene. PMID:20861044
Schneider, David M.; Woolley, Sarah M. N.
Many social animals including songbirds use communication vocalizations for individual recognition. The perception of vocalizations depends on the encoding of complex sounds by neurons in the ascending auditory system, each of which is tuned to a particular subset of acoustic features. Here, we examined how well the responses of single auditory neurons could be used to discriminate among bird songs and we compared discriminability to spectrotemporal tuning. We then used biologically realistic...
Chen, Zhaocong; Wong, Francis C K; Jones, Jeffery A; Li, Weifeng; Liu, Peng; Chen, Xi; Liu, Hanjun
Speech perception and production are intimately linked. There is evidence that speech motor learning results in changes to auditory processing of speech. Whether speech motor control benefits from perceptual learning in speech, however, remains unclear. This event-related potential study investigated whether speech-sound learning can modulate the processing of feedback errors during vocal pitch regulation. Mandarin speakers were trained to perceive five Thai lexical tones while learning to associate pictures with spoken words over 5 days. Before and after training, participants produced sustained vowel sounds while they heard their vocal pitch feedback unexpectedly perturbed. As compared to the pre-training session, the magnitude of vocal compensation significantly decreased for the control group, but remained consistent for the trained group at the post-training session. However, the trained group had smaller and faster N1 responses to pitch perturbations and exhibited enhanced P2 responses that correlated significantly with their learning performance. These findings indicate that the cortical processing of vocal pitch regulation can be shaped by learning new speech-sound associations, suggesting that perceptual learning in speech can produce transfer effects to facilitating the neural mechanisms underlying the online monitoring of auditory feedback regarding vocal production.
Schneider, David M; Woolley, Sarah M N
Many social animals including songbirds use communication vocalizations for individual recognition. The perception of vocalizations depends on the encoding of complex sounds by neurons in the ascending auditory system, each of which is tuned to a particular subset of acoustic features. Here, we examined how well the responses of single auditory neurons could be used to discriminate among bird songs and we compared discriminability to spectrotemporal tuning. We then used biologically realistic models of pooled neural responses to test whether the responses of groups of neurons discriminated among songs better than the responses of single neurons and whether discrimination by groups of neurons was related to spectrotemporal tuning and trial-to-trial response variability. The responses of single auditory midbrain neurons could be used to discriminate among vocalizations with a wide range of abilities, ranging from chance to 100%. The ability to discriminate among songs using single neuron responses was not correlated with spectrotemporal tuning. Pooling the responses of pairs of neurons generally led to better discrimination than the average of the two inputs and the most discriminating input. Pooling the responses of three to five single neurons continued to improve neural discrimination. The increase in discriminability was largest for groups of neurons with similar spectrotemporal tuning. Further, we found that groups of neurons with correlated spike trains achieved the largest gains in discriminability. We simulated neurons with varying levels of temporal precision and measured the discriminability of responses from single simulated neurons and groups of simulated neurons. Simulated neurons with biologically observed levels of temporal precision benefited more from pooling correlated inputs than did neurons with highly precise or imprecise spike trains. These findings suggest that pooling correlated neural responses with the levels of precision observed in the
Colburn, H. Steven
A brief introduction to the basic auditory abilities of the human perceiver with particular attention toward issues that may be important for the design of auditory interfaces is presented. The importance of appropriate auditory inputs to observers with normal hearing is probably related to the role of hearing as an omnidirectional, early warning system and to its role as the primary vehicle for communication of strong personal feelings.
Ryals, Brenda M.; Dooling, Robert J.
A bird study found that with sufficient time and training after hair cell and hearing loss and hair cell regeneration, the mature avian auditory system can accommodate input from a newly regenerated periphery sufficiently to allow for recognition of previously familiar vocalizations and the learning of new complex acoustic classifications.…
Alexei L Vyssotski
Full Text Available What cortical inputs are provided to motor control areas while they drive complex learned behaviors? We study this question in the nucleus interface of the nidopallium (NIf, which is required for normal birdsong production and provides the main source of auditory input to HVC, the driver of adult song. In juvenile and adult zebra finches, we find that spikes in NIf projection neurons precede vocalizations by several tens of milliseconds and are insensitive to distortions of auditory feedback. We identify a local isometry between NIf output and vocalizations: quasi-identical notes produced in different syllables are preceded by highly similar NIf spike patterns. NIf multiunit firing during song precedes responses in auditory cortical neurons by about 50 ms, revealing delayed congruence between NIf spiking and a neural representation of auditory feedback. Our findings suggest that NIf codes for imminent acoustic events within vocal performance.
de Ceballos, Albanita Gomes da Costa; Carvalho, Fernando Martins; de Araújo, Tânia Maria; Dos Reis, Eduardo José Farias Borges
Teachers are professionals who demand much of their voices and, consequently, present a high risk of developing vocal disorders during the course of employment. To identify factors associated with vocal disorders among teachers. An exploratory cross-sectional study, which investigated 476 teachers in primary and secondary schools in the city of Salvador, Bahia. Teachers answered a questionnaire and were submitted to auditory vocal analysis. The GRBAS was used for the diagnosis of vocal disorders. The study population comprised 82.8% women, teachers with an average age of 40.7 years, teachers with higher education (88.4%), with an average workday of 38 hours per week, average 11.5 years of professional practice and average monthly income of R$1.817.18. The prevalence of voice disorders was 53.6%. (255 teachers). The bivariate analysis showed statistically significant associations between vocal disorders and age above 40 years (PR = 1.83; 95% CI; 1.27-2.64), family history of dysphonia (PR = 1.72; 95% CI; 1.06-2.80), over 20 hours of weekly working hours (PR = 1.66; 95% CI; 1.09-2.52) and presence of chalk dust in the classroom (PR = 1.70; 95% CI; 1.14-2.53). The study concluded that teachers, 40 years old and over, with a family history of dysphonia, working over 20 hours weekly, and teaching in classrooms with chalk dust are more likely to develop voice disorders than others.
Gadziola, Marie A.
The underlying goal of this dissertation is to understand how the amygdala, a brain region involved in establishing the emotional significance of sensory input, contributes to the processing of complex sounds. The general hypothesis is that communication calls of big brown bats (Eptesicus fuscus) transmit relevant information about social context that is reflected in the activity of amygdalar neurons. The first specific aim analyzed social vocalizations emitted under a variety of behavioral contexts, and related vocalizations to an objective measure of internal physiological state by monitoring the heart rate of vocalizing bats. These experiments revealed a complex acoustic communication system among big brown bats in which acoustic cues and call structure signal the emotional state of a sender. The second specific aim characterized the responsiveness of single neurons in the basolateral amygdala to a range of social syllables. Neurons typically respond to the majority of tested syllables, but effectively discriminate among vocalizations by varying the response duration. This novel coding strategy underscores the importance of persistent firing in the general functioning of the amygdala. The third specific aim examined the influence of acoustic context by characterizing both the behavioral and neurophysiological responses to natural vocal sequences. Vocal sequences differentially modify the internal affective state of a listening bat, with lower aggression vocalizations evoking the greatest change in heart rate. Amygdalar neurons employ two different coding strategies: low background neurons respond selectively to very few stimuli, whereas high background neurons respond broadly to stimuli but demonstrate variation in response magnitude and timing. Neurons appear to discriminate the valence of stimuli, with aggression sequences evoking robust population-level responses across all sound levels. Further, vocal sequences show improved discrimination among stimuli
THIS ARTICLE DISCUSSES THE POSSIBLE HOMOLOGIES BETWEEN THE HUMAN LANGUAGE NETWORKS AND COMPARABLE AUDITORY PROJECTION SYSTEMS IN THE MACAQUE BRAIN, IN AN ATTEMPT TO RECONCILE TWO EXISTING VIEWS ON LANGUAGE EVOLUTION: one that emphasizes hand control and gestures, and the other that emphasizes auditory-vocal mechanisms. The capacity for language is based on relatively well defined neural substrates whose rudiments have been traced in the non-human primate brain. At its core, this circuit constitutes an auditory-vocal sensorimotor circuit with two main components, a "ventral pathway" connecting anterior auditory regions with anterior ventrolateral prefrontal areas, and a "dorsal pathway" connecting auditory areas with parietal areas and with posterior ventrolateral prefrontal areas via the arcuate fasciculus and the superior longitudinal fasciculus. In humans, the dorsal circuit is especially important for phonological processing and phonological working memory, capacities that are critical for language acquisition and for complex syntax processing. In the macaque, the homolog of the dorsal circuit overlaps with an inferior parietal-premotor network for hand and gesture selection that is under voluntary control, while vocalizations are largely fixed and involuntary. The recruitment of the dorsal component for vocalization behavior in the human lineage, together with a direct cortical control of the subcortical vocalizing system, are proposed to represent a fundamental innovation in human evolution, generating an inflection point that permitted the explosion of vocal language and human communication. In this context, vocal communication and gesturing have a common history in primate communication.
Bazilinskyy, P.; de Winter, J.C.F.
This study investigated peoples’ opinion on auditory interfaces in contemporary
cars and their willingness to be exposed to auditory feedback in automated driving. We used an Internet-based survey to collect 1,205 responses from 91 countries. The respondents stated their attitudes towards two
Begault, Durand R.; Anderson, Mark R.; Bittner, Rachael M.
The Western Electric Company produced a multi-line telephone during the 1940s-1970s using a six-button interface design that provided robust tactile, haptic and auditory cues regarding the "state" of the communication system. This multi-line telephone was used as a model for a trade study comparison of two interfaces: a touchscreen interface (iPad)) versus a pressure-sensitive strain gauge button interface (Phidget USB interface controllers). The experiment and its results are detailed in the authors' AES 133rd convention paper " Multimodal Information Management: Evaluation of Auditory and Haptic Cues for NextGen Communication Dispays". This Engineering Brief describes how the interface logic, visual indications, and auditory cues of the original telephone were synthesized using MAX/MSP, including the logic for line selection, line hold, and priority line activation.
Pannese, Alessia; Grandjean, Didier; Frühholz, Sascha
Discriminating between auditory signals of different affective value is critical to successful social interaction. It is commonly held that acoustic decoding of such signals occurs in the auditory system, whereas affective decoding occurs in the amygdala. However, given that the amygdala receives direct subcortical projections that bypass the auditory cortex, it is possible that some acoustic decoding occurs in the amygdala as well, when the acoustic features are relevant for affective discrimination. We tested this hypothesis by combining functional neuroimaging with the neurophysiological phenomena of repetition suppression (RS) and repetition enhancement (RE) in human listeners. Our results show that both amygdala and auditory cortex responded differentially to physical voice features, suggesting that the amygdala and auditory cortex decode the affective quality of the voice not only by processing the emotional content from previously processed acoustic features, but also by processing the acoustic features themselves, when these are relevant to the identification of the voice's affective value. Specifically, we found that the auditory cortex is sensitive to spectral high-frequency voice cues when discriminating vocal anger from vocal fear and joy, whereas the amygdala is sensitive to vocal pitch when discriminating between negative vocal emotions (i.e., anger and fear). Vocal pitch is an instantaneously recognized voice feature, which is potentially transferred to the amygdala by direct subcortical projections. These results together provide evidence that, besides the auditory cortex, the amygdala too processes acoustic information, when this is relevant to the discrimination of auditory emotions. Copyright Â© 2016 Elsevier Ltd. All rights reserved.
Fergus, Daniel J; Feng, Ni Y; Bass, Andrew H
Successful animal communication depends on a receiver's ability to detect a sender's signal. Exemplars of adaptive sender-receiver coupling include acoustic communication, often important in the context of seasonal reproduction. During the reproductive summer season, both male and female midshipman fish (Porichthys notatus) exhibit similar increases in the steroid-dependent frequency sensitivity of the saccule, the main auditory division of the inner ear. This form of auditory plasticity enhances detection of the higher frequency components of the multi-harmonic, long-duration advertisement calls produced repetitively by males during summer nights of peak vocal and spawning activity. The molecular basis of this seasonal auditory plasticity has not been fully resolved. Here, we utilize an unbiased transcriptomic RNA sequencing approach to identify differentially expressed transcripts within the saccule's hair cell epithelium of reproductive summer and non-reproductive winter fish. We assembled 74,027 unique transcripts from our saccular epithelial sequence reads. Of these, 6.4 % and 3.0 % were upregulated in the reproductive and non-reproductive saccular epithelium, respectively. Gene ontology (GO) term enrichment analyses of the differentially expressed transcripts showed that the reproductive saccular epithelium was transcriptionally, translationally, and metabolically more active than the non-reproductive epithelium. Furthermore, the expression of a specific suite of candidate genes, including ion channels and components of steroid-signaling pathways, was upregulated in the reproductive compared to the non-reproductive saccular epithelium. We found reported auditory functions for 14 candidate genes upregulated in the reproductive midshipman saccular epithelium, 8 of which are enriched in mouse hair cells, validating their hair cell-specific functions across vertebrates. We identified a suite of differentially expressed genes belonging to neurotransmission and
Beal, Deryk S; Cheyne, Douglas O; Gracco, Vincent L; Quraan, Maher A; Taylor, Margot J; De Nil, Luc F
We used magnetoencephalography to investigate auditory evoked responses to speech vocalizations and non-speech tones in adults who do and do not stutter. Neuromagnetic field patterns were recorded as participants listened to a 1 kHz tone, playback of their own productions of the vowel /i/ and vowel-initial words, and actively generated the vowel /i/ and vowel-initial words. Activation of the auditory cortex at approximately 50 and 100 ms was observed during all tasks. A reduction in the peak amplitudes of the M50 and M100 components was observed during the active generation versus passive listening tasks dependent on the stimuli. Adults who stutter did not differ in the amount of speech-induced auditory suppression relative to fluent speakers. Adults who stutter had shorter M100 latencies for the actively generated speaking tasks in the right hemisphere relative to the left hemisphere but the fluent speakers showed similar latencies across hemispheres. During passive listening tasks, adults who stutter had longer M50 and M100 latencies than fluent speakers. The results suggest that there are timing, rather than amplitude, differences in auditory processing during speech in adults who stutter and are discussed in relation to hypotheses of auditory-motor integration breakdown in stuttering. Copyright 2010 Elsevier Inc. All rights reserved.
Full Text Available Although clinical studies reported hyperactivation of the auditory system and amygdala in patients with auditory hallucinations (hearing others’ but not one’s own voice, independent of any external stimulus, neural mechanisms of self/other attribution is not well understood. We recorded neuronal responses in the dorsal amygdala including the lateral amygdaloid nucleus to ultrasonic vocalization (USVs emitted by subjects and conspecifics during free social interaction in 16 adult male rats. The animals emitting the USVs were identified by EMG recordings. One-quarter of the amygdalar neurons (15/60 responded to 50 kHz calls by the subject and/or conspecifics. Among the responsive neurons, most neurons (Type-Other neurons (73%, 11/15 responded only to calls by conspecifics but not subjects. Two Type-Self neurons (13%, 2/15 responded to calls by the subject but not those by conspecifics, although their response selectivity to subjects vs. conspecifics was lower than that of Type-Other neurons. The remaining two neurons (13% responded to calls by both the subject and conspecifics. Furthermore, population coding of the amygdalar neurons represented distinction of subject vs. conspecific calls. The present results provide the first neurophysiological evidence that the amygdala discriminately represents affective social calls by subject and conspecifics. These findings suggest that the amygdala is an important brain region for self/other attribution. Furthermore, pathological activation of the amygdala, where Type-Other neurons predominate, could induce external misattribution of percepts of vocalization.
Ng, Chi-Wing; Plakke, Bethany; Poremba, Amy
Neural correlates of auditory processing, including for species-specific vocalizations that convey biological and ethological significance (e.g., social status, kinship, environment), have been identified in a wide variety of areas including the temporal and frontal cortices. However, few studies elucidate how non-human primates interact with these vocalization signals when they are challenged by tasks requiring auditory discrimination, recognition and/or memory. The present study employs a delayed matching-to-sample task with auditory stimuli to examine auditory memory performance of rhesus macaques (Macaca mulatta), wherein two sounds are determined to be the same or different. Rhesus macaques seem to have relatively poor short-term memory with auditory stimuli, and we examine if particular sound types are more favorable for memory performance. Experiment 1 suggests memory performance with vocalization sound types (particularly monkey), are significantly better than when using non-vocalization sound types, and male monkeys outperform female monkeys overall. Experiment 2, controlling for number of sound exemplars and presentation pairings across types, replicates Experiment 1, demonstrating better performance or decreased response latencies, depending on trial type, to species-specific monkey vocalizations. The findings cannot be explained by acoustic differences between monkey vocalizations and the other sound types, suggesting the biological, and/or ethological meaning of these sounds are more effective for auditory memory. 2009 Elsevier B.V.
Full Text Available Direct social contact and social interaction affect speech development in human infants and are required in order to maintain perceptual abilities; however the processes involved are still poorly known. In the present study, we tested the hypothesis that social segregation during development would prevent the proper development of a central auditory area, using a "classical" animal model of vocal development, a songbird. Based on our knowledge of European starling, we raised young female starlings with peers and only adult male tutors. This ensured that female would show neither social bond with nor vocal copying from males. Electrophysiological recordings performed when these females were adult revealed perceptual abnormalities: they presented a larger auditory area, a lower proportion of specialized neurons and a larger proportion of generalist sites than wild-caught females, whereas these characteristics were similar to those observed in socially deprived (physically separated females. These results confirmed and added to earlier results for males, suggesting that the degree of perceptual deficiency reflects the degree of social separation. To our knowledge, this report constitutes the first evidence that social segregation can, as much as physical separation, alter the development of a central auditory area.
Hage, Steffen R; Nieder, Andreas
Complex audio-vocal integration systems depend on a strong interconnection between the auditory and the vocal motor system. To gain cognitive control over audio-vocal interaction during vocal motor control, the PFC needs to be involved. Neurons in the ventrolateral PFC (VLPFC) have been shown to separately encode the sensory perceptions and motor production of vocalizations. It is unknown, however, whether single neurons in the PFC reflect audio-vocal interactions. We therefore recorded single-unit activity in the VLPFC of rhesus monkeys (Macaca mulatta) while they produced vocalizations on command or passively listened to monkey calls. We found that 12% of randomly selected neurons in VLPFC modulated their discharge rate in response to acoustic stimulation with species-specific calls. Almost three-fourths of these auditory neurons showed an additional modulation of their discharge rates either before and/or during the monkeys' motor production of vocalization. Based on these audio-vocal interactions, the VLPFC might be well positioned to combine higher order auditory processing with cognitive control of the vocal motor output. Such audio-vocal integration processes in the VLPFC might constitute a precursor for the evolution of complex learned audio-vocal integration systems, ultimately giving rise to human speech. Copyright © 2015 the authors 0270-6474/15/357030-11$15.00/0.
Gao, Patrick P; Zhang, Jevin W; Fan, Shu-Juan; Sanes, Dan H; Wu, Ed X
The cortex contains extensive descending projections, yet the impact of cortical input on brainstem processing remains poorly understood. In the central auditory system, the auditory cortex contains direct and indirect pathways (via brainstem cholinergic cells) to nuclei of the auditory midbrain, called the inferior colliculus (IC). While these projections modulate auditory processing throughout the IC, single neuron recordings have samples from only a small fraction of cells during stimulation of the corticofugal pathway. Furthermore, assessments of cortical feedback have not been extended to sensory modalities other than audition. To address these issues, we devised blood-oxygen-level-dependent (BOLD) functional magnetic resonance imaging (fMRI) paradigms to measure the sound-evoked responses throughout the rat IC and investigated the effects of bilateral ablation of either auditory or visual cortices. Auditory cortex ablation increased the gain of IC responses to noise stimuli (primarily in the central nucleus of the IC) and decreased response selectivity to forward species-specific vocalizations (versus temporally reversed ones, most prominently in the external cortex of the IC). In contrast, visual cortex ablation decreased the gain and induced a much smaller effect on response selectivity. The results suggest that auditory cortical projections normally exert a large-scale and net suppressive influence on specific IC subnuclei, while visual cortical projections provide a facilitatory influence. Meanwhile, auditory cortical projections enhance the midbrain response selectivity to species-specific vocalizations. We also probed the role of the indirect cholinergic projections in the auditory system in the descending modulation process by pharmacologically blocking muscarinic cholinergic receptors. This manipulation did not affect the gain of IC responses but significantly reduced the response selectivity to vocalizations. The results imply that auditory cortical
Gaucher, Quentin; Huetz, Chloé; Gourévitch, Boris; Laudanski, Jonathan; Occelli, Florian; Edeline, Jean-Marc
A major goal in auditory neuroscience is to characterize how communication sounds are represented at the cortical level. The present review aims at investigating the role of auditory cortex in the processing of speech, bird songs and other vocalizations, which all are spectrally and temporally highly structured sounds. Whereas earlier studies have simply looked for neurons exhibiting higher firing rates to particular conspecific vocalizations over their modified, artificially synthesized versions, more recent studies determined the coding capacity of temporal spike patterns, which are prominent in primary and non-primary areas (and also in non-auditory cortical areas). In several cases, this information seems to be correlated with the behavioral performance of human or animal subjects, suggesting that spike-timing based coding strategies might set the foundations of our perceptive abilities. Also, it is now clear that the responses of auditory cortex neurons are highly nonlinear and that their responses to natural stimuli cannot be predicted from their responses to artificial stimuli such as moving ripples and broadband noises. Since auditory cortex neurons cannot follow rapid fluctuations of the vocalizations envelope, they only respond at specific time points during communication sounds, which can serve as temporal markers for integrating the temporal and spectral processing taking place at subcortical relays. Thus, the temporal sparse code of auditory cortex neurons can be considered as a first step for generating high level representations of communication sounds independent of the acoustic characteristic of these sounds. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives". Copyright © 2013 Elsevier B.V. All rights reserved.
Arruda, Polyanna; Diniz da Rosa, Marine Raquel; Almeida, Larissa Nadjara Alves; de Araujo Pernambuco, Leandro; Almeida, Anna Alice
Estradiol production varies cyclically, changes in levels are hypothesized to affect the voice. The main objective of this study was to investigate vocal acoustic and auditory-perceptual characteristics during fluctuations in the levels of the hormone estradiol during the menstrual cycle. A total of 44 volunteers aged between 18 and 45 were selected. Of these, 27 women with regular menstrual cycles comprised the test group (TG) and 17 combined oral contraceptive users comprised the control group (CG). The study was performed in two phases. In phase 1, anamnesis was performed. Subsequently, the TG underwent blood sample collection for measurement of estradiol levels and voice recording for later acoustic and auditory-perceptual analysis. The CG underwent only voice recording. Phase 2 involved the same measurements as phase 1 for each group. Variables were evaluated using descriptive and inferential analysis to compare groups and phases and to determine relationships between variables. Voice changes were found during the menstrual cycle, and such changes were determined to be related to variations in estradiol levels. Impaired voice quality was observed to be associated with decreased levels of estradiol. The CG did not demonstrate significant vocal changes during phases 1 and 2. The TG showed significant increases in vocal parameters of roughness, tension, and instability during phase 2 (the period of low estradiol levels) when compared with the CG. Low estradiol levels were also found to be negatively correlated with the parameters of tension, instability, and jitter and positively correlated with fundamental voice frequency. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Guo, Zhiqiang; Wu, Xiuqin; Li, Weifeng; Jones, Jeffery A; Yan, Nan; Sheft, Stanley; Liu, Peng; Liu, Hanjun
provides two lines of converging evidence, for the first time, that working memory cannot only enhance the perception of vocal feedback errors but also exert inhibitory control over vocal motor behavior. These findings represent a major advance in our understanding of the top-down modulatory mechanisms that support the detection and correction of prediction-feedback mismatches during sensorimotor control of speech production driven by working memory. Rather than being an exclusively bottom-up and automatic process, auditory-motor integration for voice control can be modulated by top-down influences arising from working memory. Copyright © 2017 the authors 0270-6474/17/3710324-11$15.00/0.
Full Text Available The inferior colliculus (IC is a major center for integration of auditory information as itreceives ascending projections from a variety of brainstem nuclei as well as descending projectionsfrom the thalamus and auditory cortex. The ascending projections are both excitatory andinhibitory and their convergence at the IC results in a microcircuitry that is important forshaping responses to simple, binaural, and modulated sounds in the IC. Here, we examined therole inhibition plays in shaping selectivity to vocalizations in the IC of awake, normal-hearingadult mice (CBA/CaJ strain. Neurons in the IC of mice show selectivity in their responses tovocalizations, and we hypothesized that this selectivity is created by inhibitory microcircuitryin the IC. We compared single unit responses in the IC to pure tones and a variety of ultrasonicmouse vocalizations before and after iontophoretic application of GABAA receptor (GABAARand glycine receptor (GlyR antagonists. The most pronounced effects of blocking GABAAR andGlyR on IC neurons were to increase spike rates and broaden excitatory frequency tuning curvesin response to pure tone stimuli, and to decrease selectivity to vocalizations. Thus, inhibitionplays an important role in creating selectivity to vocalizations in the inferior colliculus.
Fukushima, Makoto; Saunders, Richard C; Fujii, Naotaka; Averbeck, Bruno B; Mishkin, Mortimer
Vocal production is an example of controlled motor behavior with high temporal precision. Previous studies have decoded auditory evoked cortical activity while monkeys listened to vocalization sounds. On the other hand, there have been few attempts at decoding motor cortical activity during vocal production. Here we recorded cortical activity during vocal production in the macaque with a chronically implanted electrocorticographic (ECoG) electrode array. The array detected robust activity in motor cortex during vocal production. We used a nonlinear dynamical model of the vocal organ to reduce the dimensionality of `Coo' calls produced by the monkey. We then used linear regression to evaluate the information in motor cortical activity for this reduced representation of calls. This simple linear model accounted for circa 65% of the variance in the reduced sound representations, supporting the feasibility of using the dynamical model of the vocal organ for decoding motor cortical activity during vocal production.
Talkington, William J.; Rapuano, Kristina M.; Hitt, Laura; Frum, Chris A.; Lewis, James W.
Numerous species possess cortical regions that are most sensitive to vocalizations produced by their own kind (conspecifics). In humans, the superior temporal sulci (STS) putatively represent homologous voice-sensitive areas of cortex. However, STS regions have recently been reported to represent auditory experience or “expertise” in general rather than showing exclusive sensitivity to human vocalizations per se. Using functional magnetic resonance imaging and a unique non-stereotypical category of complex human non-verbal vocalizations – human-mimicked versions of animal vocalizations – we found a cortical hierarchy in humans optimized for processing meaningful conspecific utterances. This left-lateralized hierarchy originated near primary auditory cortices and progressed into traditional speech-sensitive areas. These results suggest that the cortical regions supporting vocalization perception are initially organized by sensitivity to the human vocal tract in stages prior to the STS. Additionally, these findings have implications for the developmental time course of conspecific vocalization processing in humans as well as its evolutionary origins. PMID:22674283
Fabron, Eliana Maria Gradim; Regaçone, Simone Fiuza; Marino, Viviane Cristina de Castro; Mastria, Marina Ludovico; Motonaga, Suely Mayumi; Sebastião, Luciana Tavares
To compare the vocal self-perception and vocal complaints reported by two groups of students of the pedagogy course (freshmen and graduates); to relate the vocal self-perception to the vocal complaints for these groups; and to compare the voice quality of the students from these groups through perceptual auditory assessment and acoustic analysis. Initially, 89 students from the pedagogy course answered a questionnaire about self-perceived voice quality and vocal complaints. In a second phase, auditory-perceptual evaluation and acoustic analyses of 48 participants were made through voice recordings of sustained vowel emission and poem reading. The most reported vocal complaints were fatigue while using the voice, sore throat, effort to speak, irritation or burning in the throat, hoarseness, tightness in the neck, and variations of voice throughout the day. There was a higher occurrence of complaints from graduates than from freshmen, with significant differences for four of the nine complaints. It was also possible to observe the relationship between vocal self-perception and complaints reported by these students. No significant differences were observed in the results of auditory-perceptual evaluation; however, some graduates had their voices evaluated with higher severity of deviation of normalcy. During acoustic analysis no difference was observed between groups. The increase in vocal demand by the graduates may have caused the greatest number and diversity of vocal complaints, and several of them are related to the self-assessment of voice quality. The auditory-perceptual evaluation and acoustic analysis showed no deviations in their voice.
Tobias, Martha L.; Corke, Anna; Korsh, Jeremy; Yin, David; Kelley, Darcy B.
Male Xenopus laevis frogs produce underwater advertisement calls that attract gravid females and suppress calling by male competitors. Here we explore whether groups of males establish vocal ranks and whether auditory cues alone suffice for vocal suppression. Tests of male–male pairs within assigned groups reveal linear vocal dominance relations, in which each male has a defined rank. Both the duration over which males interact, as well as the number of competitive opportunities, affect linea...
Mirror neurons are theorized to serve as a neural substrate for spoken language in humans, but the existence and functions of auditory–vocal mirror neurons in the human brain remain largely matters of speculation. Songbirds resemble humans in their capacity for vocal learning and depend on their learned songs to facilitate courtship and individual recognition. Recent neurophysiological studies have detected putative auditory–vocal mirror neurons in a sensorimotor region of the songbird's brain that plays an important role in expressive and receptive aspects of vocal communication. This review discusses the auditory and motor-related properties of these cells, considers their potential role on song learning and communication in relation to classical studies of birdsong, and points to the circuit and developmental mechanisms that may give rise to auditory–vocal mirroring in the songbird's brain. PMID:24778375
Papp, III, Albert Louis [Univ. of California, Davis, CA (United States)
This dissertation describes a methodology and example implementation for the dynamic regulation of temporally overlapping auditory messages in computer-user interfaces. The regulation mechanism exists to schedule numerous overlapping auditory messages in such a way that each individual message remains perceptually distinct from all others. The method is based on the research conducted in the area of auditory scene analysis. While numerous applications have been engineered to present the user with temporally overlapped auditory output, they have generally been designed without any structured method of controlling the perceptual aspects of the sound. The method of scheduling temporally overlapping sounds has been extended to function in an environment where numerous applications can present sound independently of each other. The Centralized Audio Presentation System is a global regulation mechanism that controls all audio output requests made from all currently running applications. The notion of multimodal objects is explored in this system as well. Each audio request that represents a particular message can include numerous auditory representations, such as musical motives and voice. The Presentation System scheduling algorithm selects the best representation according to the current global auditory system state, and presents it to the user within the request constraints of priority and maximum acceptable latency. The perceptual conflicts between temporally overlapping audio messages are examined in depth through the Computational Auditory Scene Synthesizer. At the heart of this system is a heuristic-based auditory scene synthesis scheduling method. Different schedules of overlapped sounds are evaluated and assigned penalty scores. High scores represent presentations that include perceptual conflicts between over-lapping sounds. Low scores indicate fewer and less serious conflicts. A user study was conducted to validate that the perceptual difficulties predicted by
Lee, Yune Sang; Peelle, Jonathan E; Kraemer, David; Lloyd, Samuel; Granger, Richard
Past neuroimaging studies have documented discrete regions of human temporal cortex that are more strongly activated by conspecific voice sounds than by nonvoice sounds. However, the mechanisms underlying this voice sensitivity remain unclear. In the present functional MRI study, we took a novel approach to examining voice sensitivity, in which we applied a signal detection paradigm to the assessment of multivariate pattern classification among several living and nonliving categories of auditory stimuli. Within this framework, voice sensitivity can be interpreted as a distinct neural representation of brain activity that correctly distinguishes human vocalizations from other auditory object categories. Across a series of auditory categorization tests, we found that bilateral superior and middle temporal cortex consistently exhibited robust sensitivity to human vocal sounds. Although the strongest categorization was in distinguishing human voice from other categories, subsets of these regions were also able to distinguish reliably between nonhuman categories, suggesting a general role in auditory object categorization. Our findings complement the current evidence of cortical sensitivity to human vocal sounds by revealing that the greatest sensitivity during categorization tasks is devoted to distinguishing voice from nonvoice categories within human temporal cortex. Copyright © 2015 the American Physiological Society.
Chi-Wing, Ng; Bethany, Plakke; Amy, Poremba
Neural correlates of auditory processing, including for species-specific vocalizations that convey biological and ethological significance (e.g. social status, kinship, environment),have been identified in a wide variety of areas including the temporal and frontal cortices. However, few studies elucidate how non-human primates interact with these vocalization signals when they are challenged by tasks requiring auditory discrimination, recognition, and/or memory. The present study employs a de...
Pinheiro, Ana P; Barros, Carla; Vasconcelos, Margarida; Obermeier, Christian; Kotz, Sonja A
The capacity to predict what should happen next and to minimize any discrepancy between an expected and an actual sensory input (prediction error) is a central aspect of perception. Particularly in vocal communication, the effective prediction of an auditory input that informs the listener about the emotionality of a speaker is critical. What is currently unknown is how the perceived valence of an emotional vocalization affects the capacity to predict and detect a change in the auditory input. This question was probed in a combined event-related potential (ERP) and time-frequency analysis approach. Specifically, we examined the brain response to standards (Repetition Positivity) and to deviants (Mismatch Negativity - MMN), as well as the anticipatory response to the vocal sounds (pre-stimulus beta oscillatory power). Short neutral, happy (laughter), and angry (growls) vocalizations were presented both as standard and deviant stimuli in a passive oddball listening task while participants watched a silent movie and were instructed to ignore the vocalizations. MMN amplitude was increased for happy compared to neutral and angry vocalizations. The Repetition Positivity was enhanced for happy standard vocalizations. Induced pre-stimulus upper beta power was increased for happy vocalizations, and predicted the modulation of the standard Repetition Positivity. These findings indicate enhanced sensory prediction for positive vocalizations such as laughter. Together, the results suggest that positive vocalizations are more effective predictors in social communication than angry and neutral ones, possibly due to their high social significance. Copyright © 2017 Elsevier Ltd. All rights reserved.
Full Text Available This study investigated peoples’ opinion on auditory interfaces in contemporary cars and their willingness to be exposed to auditory feedback in automated driving. We used an Internet-based survey to collect 1,205 responses from 91 countries. The respondents stated their attitudes towards two existing auditory driver assistance systems, a parking assistant (PA and a forward collision warning system (FCWS, as well as towards a futuristic augmented sound system (FS proposed for fully automated driving. The respondents were positive towards the PA and FCWS, and rated the willingness to have automated versions of these systems as 3.87 and 3.77, respectively (on a scale from 1 = disagree strongly to 5 = agree strongly. The respondents tolerated the FS (the mean willingness to use it was 3.00 on the same scale. The results showed that among the available response options, the female voice was the most preferred feedback type for takeover requests in highly automated driving, regardless of whether the respondents’ country was English speaking or not. The present results could be useful for designers of automated vehicles and other stakeholders.
Heffner, Henry E.; Heffner, Rickye S.
Japanese macaques were trained to discriminate two forms of their coo vocalization before and after unilateral and bilateral ablation of the temporal cortex. Unilateral ablation of the left superior temporal gyrus, including auditory cortex, resulted in an initial impairment in the discrimination, but similar unilateral ablation of the right superior temporal gyrus had no effect. Bilateral temporal lesions including auditory cortex completely abolished the ability of the animals to discriminate their coos. Neither unilateral nor bilateral ablation of cortex dorsal to and sparing the auditory cortex had any effect on the discrimination. The perception of species-specific vocalizations by Japanese macaques seems to be mediated by the temporal cortex, with the left hemisphere playing a predominant role.
Full Text Available Impaired self-monitoring and abnormalities of cognitive bias have been implicated as cognitive mechanisms of hallucination; regions fundamental to these processes including inferior frontal gyrus (IFG and superior temporal gyrus (STG are abnormally activated in individuals that hallucinate. A recent study showed activation in IFG-STG to be modulated by auditory attractiveness, but no study has investigated whether these IFG-STG activations are impaired in schizophrenia. We aimed to clarify the cerebral function underlying the perception of auditory attractiveness in schizophrenia patients. Cerebral activation was examined in 18 schizophrenia patients and 18 controls when performing Favourability Judgment Task (FJT and Gender Differentiation Task (GDT for pairs of greetings using event-related functional MRI. A full-factorial analysis revealed that the main effect of task was associated with activation of left IFG and STG. The main effect of Group revealed less activation of left STG in schizophrenia compared with controls, whereas significantly greater activation in schizophrenia than in controls was revealed at the left middle frontal gyrus (MFG, right temporo-parietal junction (TPJ, right occipital lobe, and right amygdala (p<0.05, FDR-corrected. A significant positive correlation was observed at the right TPJ and right MFG between cerebral activation under FJT minus GDT contrast and the score of hallucinatory behaviour on the Positive and Negative Symptom Scale. Findings of hypo-activation in the left STG could designate brain dysfunction in accessing vocal attractiveness in schizophrenia, whereas hyper-activation in the right TPJ and MFG may reflect the process of mentalizing other person’s behaviour by auditory hallucination by abnormality of cognitive bias.
Full Text Available Using functional magnetic resonance imaging in awake behaving monkeys we investigated how species-specific vocalizations are represented in auditory and auditory-related regions of the macaque brain. We found clusters of active voxels along the ascending auditory pathway that responded to various types of complex sounds: inferior colliculus (IC, medial geniculate nucleus (MGN, auditory core, belt, and parabelt cortex, and other parts of the superior temporal gyrus (STG and sulcus (STS. Regions sensitive to monkey calls were most prevalent in the anterior STG, but some clusters were also found in frontal and parietal cortex on the basis of comparisons between responses to calls and environmental sounds. Surprisingly, we found that spectrotemporal control sounds derived from the monkey calls (scrambled calls also activated the parietal and frontal regions. Taken together, our results demonstrate that species-specific vocalizations in rhesus monkeys activate preferentially the auditory ventral stream, and in particular areas of the antero-lateral belt and parabelt.
Halder, S; Käthner, I; Kübler, A
Auditory brain-computer interfaces are an assistive technology that can restore communication for motor impaired end-users. Such non-visual brain-computer interface paradigms are of particular importance for end-users that may lose or have lost gaze control. We attempted to show that motor impaired end-users can learn to control an auditory speller on the basis of event-related potentials. Five end-users with motor impairments, two of whom with additional visual impairments, participated in five sessions. We applied a newly developed auditory brain-computer interface paradigm with natural sounds and directional cues. Three of five end-users learned to select symbols using this method. Averaged over all five end-users the information transfer rate increased by more than 1800% from the first session (0.17 bits/min) to the last session (3.08 bits/min). The two best end-users achieved information transfer rates of 5.78 bits/min and accuracies of 92%. Our results show that an auditory BCI with a combination of natural sounds and directional cues, can be controlled by end-users with motor impairment. Training improves the performance of end-users to the level of healthy controls. To our knowledge, this is the first time end-users with motor impairments controlled an auditory brain-computer interface speller with such high accuracy and information transfer rates. Further, our results demonstrate that operating a BCI with event-related potentials benefits from training and specifically end-users may require more than one session to develop their full potential. Copyright © 2015 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.
Rojas, Gleidy Vannesa E; Ricz, Hilton; Tumas, Vitor; Rodrigues, Guilherme R; Toscano, Patrícia; Aguiar-Ricz, Lílian
The study aimed to compare and correlate perceptual-auditory analysis of vocal parameters and self-perception in individuals with adductor spasmodic dysphonia before and after the application of botulinum toxin. This is a prospective cohort study. Sixteen individuals with a diagnosis of adductor spasmodic dysphonia were submitted to the application of botulinum toxin in the thyroarytenoid muscle, to the recording of a voice signal, and to the Voice Handicap Index (VHI) questionnaire before the application and at two time points after application. Two judges performed a perceptual-auditory analysis of eight vocal parameters with the aid of the Praat software for the visualization of narrow band spectrography, pitch, and intensity contour. Comparison of the vocal parameters before toxin application and on the first return revealed a reduction of oscillation intensity (P = 0.002), voice breaks (P = 0.002), and vocal tremor (P = 0.002). The same parameters increased on the second return. The degree of severity, strained-strangled voice, roughness, breathiness, and asthenia was unchanged. The total score and the emotional domain score of the VHI were reduced on the first return. There was a moderate correlation between the degree of voice severity and the total VHI score before application and on the second return, and a weak correlation on the first return. Perceptual-auditory analysis and self-perception proved to be efficient in the recognition of vocal changes and of the vocal impact on individuals with adductor spasmodic dysphonia under treatment with botulinum toxin, permitting the quantitation of changes along time. Copyright © 2017. Published by Elsevier Inc.
Muchnik, Chava; Ari-Even Roth, Daphne; Hildesheimer, Minka; Arie, Miri; Bar-Haim, Yair; Henkin, Yael
Two efferent feedback pathways to the auditory periphery may play a role in monitoring self-vocalization: the middle-ear acoustic reflex (MEAR) and the medial olivocochlear bundle (MOCB) reflex. Since most studies regarding the role of auditory efferent activity during self-vocalization were conducted in animals, human data are scarce. The working premise of the current study was that selective mutism (SM), a rare psychiatric disorder characterized by consistent failure to speak in specific social situations despite the ability to speak normally in other situations, may serve as a human model for studying the potential involvement of auditory efferent activity during self-vocalization. For this purpose, auditory efferent function was assessed in a group of 31 children with SM and compared to that of a group of 31 normally developing control children (mean age 8.9 and 8.8 years, respectively). All children exhibited normal hearing thresholds and type A tympanograms. MEAR and MOCB functions were evaluated by means of acoustic reflex thresholds and decay functions and the suppression of transient-evoked otoacoustic emissions, respectively. Auditory afferent function was tested by means of auditory brainstem responses (ABR). Results indicated a significantly higher proportion of children with abnormal MEAR and MOCB function in the SM group (58.6 and 38%, respectively) compared to controls (9.7 and 8%, respectively). The prevalence of abnormal MEAR and/or MOCB function was significantly higher in the SM group (71%) compared to controls (16%). Intact afferent function manifested in normal absolute and interpeak latencies of ABR components in all children. The finding of aberrant efferent auditory function in a large proportion of children with SM provides further support for the notion that MEAR and MOCB may play a significant role in the process of self-vocalization. © 2013 S. Karger AG, Basel.
Bar-Haim, Yair; Henkin, Yael; Ari-Even-Roth, Daphne; Tetin-Schneider, Simona; Hildesheimer, Minka; Muchnik, Chava
Selective mutism is a psychiatric disorder of childhood characterized by consistent inability to speak in specific situations despite the ability to speak normally in others. The objective of this study was to test whether reduced auditory efferent activity, which may have direct bearings on speaking behavior, is compromised in selectively mute children. Participants were 16 children with selective mutism and 16 normally developing control children matched for age and gender. All children were tested for pure-tone audiometry, speech reception thresholds, speech discrimination, middle-ear acoustic reflex thresholds and decay function, transient evoked otoacoustic emission, suppression of transient evoked otoacoustic emission, and auditory brainstem response. Compared with control children, selectively mute children displayed specific deficiencies in auditory efferent activity. These aberrations in efferent activity appear along with normal pure-tone and speech audiometry and normal brainstem transmission as indicated by auditory brainstem response latencies. The diminished auditory efferent activity detected in some children with SM may result in desensitization of their auditory pathways by self-vocalization and in reduced control of masking and distortion of incoming speech sounds. These children may gradually learn to restrict vocalization to the minimal amount possible in contexts that require complex auditory processing.
Nikjeh, Dee A; Lister, Jennifer J; Frisch, Stefan A
Cortical auditory evoked potentials of instrumental musicians suggest that music expertise modifies pitch processing, yet less is known about vocal musicians. Mismatch negativity (MMN) to pitch deviances and difference limen for frequency (DLF) were examined among 61 young adult women, including 20 vocalists, 21 instrumentalists, and 20 nonmusicians. Stimuli were harmonic tone complexes from the mid-female vocal range (C4-G4). MMN was elicited by multideviant paradigm. DLF was obtained by an adaptive psychophysical paradigm. Musicians detected pitch changes earlier and DLFs were 50% smaller than nonmusicians. Both vocal and instrumental musicians possess superior sensory-memory representations for acoustic parameters. Vocal musicians with instrumental training appear to have an auditory neural advantage over instrumental or vocal only musicians. An incidental finding reveals P3a as a sensitive index of music expertise.
Raksin, Jonathan N; Glaze, Christopher M; Smith, Sarah; Schmidt, Marc F
Motor-related forebrain areas in higher vertebrates also show responses to passively presented sensory stimuli. However, sensory tuning properties in these areas, especially during wakefulness, and their relation to perception, are poorly understood. In the avian song system, HVC (proper name) is a vocal-motor structure with auditory responses well defined under anesthesia but poorly characterized during wakefulness. We used a large set of stimuli including the bird's own song (BOS) and many conspecific songs (CON) to characterize auditory tuning properties in putative interneurons (HVC(IN)) during wakefulness. Our findings suggest that HVC contains a diversity of responses that vary in overall excitability to auditory stimuli, as well as bias in spike rate increases to BOS over CON. We used statistical tests to classify cells in order to further probe auditory responses, yielding one-third of neurons that were either unresponsive or suppressed and two-thirds with excitatory responses to one or more stimuli. A subset of excitatory neurons were tuned exclusively to BOS and showed very low linearity as measured by spectrotemporal receptive field analysis (STRF). The remaining excitatory neurons responded well to CON stimuli, although many cells still expressed a bias toward BOS. These findings suggest the concurrent presence of a nonlinear and a linear component to responses in HVC, even within the same neuron. These characteristics are consistent with perceptual deficits in distinguishing BOS from CON stimuli following lesions of HVC and other song nuclei and suggest mirror neuronlike qualities in which "self" (here BOS) is used as a referent to judge "other" (here CON).
Pereira, Eny Regina Bóia Neves; Tavares, Elaine Lara Mendes; Martins, Regina Helena Garcia
Dysphonia is more prevalent in teachers than among the general population. The objective of this study was to analyze clinical, vocal, and videolaryngoscopical aspects in dysphonic teachers. Ninety dysphonic teachers were inquired about their voice, comorbidities, and work conditions. They underwent vocal auditory-perceptual evaluation (maximum phonation time and GRBASI scale), acoustic voice analysis, and videolaryngoscopy. The results were compared with a control group consisting of 90 dysphonic nonteachers, of similar gender and ages, and with professional activities excluding teaching and singing. In both groups, there were 85 women and five men (age range 31-50 years). In the controls, the majority of subjects worked in domestic activities, whereas the majority of teachers worked in primary (42.8%) and secondary school (37.7%). Teachers and controls reported, respectively: vocal abuse (76.7%; 37.8%), weekly hours of work between 21 and 40 years (72.2%; 80%), under 10 years of practice (36%; 23%), absenteeism (23%; 0%), sinonasal (66%; 20%) and gastroesophageal symptoms (44%; 22%), hoarseness (82%; 78%), throat clearing (70%; 62%), and phonatory effort (72%; 52%). In both groups, there were decreased values of maximum phonation time, impairment of the G parameter in the GRBASI scale (82%), decrease of F0 and increase of the rest of acoustic parameters. Nodules and laryngopharyngeal reflux were predominant in teachers; laryngopharyngeal reflux, polyps, and sulcus vocalis predominated in the controls. Vocal symptoms, comorbidities, and absenteeism were predominant among teachers. The vocal analyses were similar in both groups. Nodules and laryngopharyngeal reflux were predominant among teachers, whereas polyps, laryngopharyngeal reflux, and sulcus were predominant among controls. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Behroozmand, Roozbeh; Ibrahim, Nadine; Korzyukov, Oleg; Robin, Donald A.; Larson, Charles R.
The ability to process auditory feedback for vocal pitch control is crucial during speaking and singing. Previous studies have suggested that musicians with absolute pitch (AP) develop specialized left-hemisphere mechanisms for pitch processing. The present study adopted an auditory feedback pitch perturbation paradigm combined with ERP recordings to test the hypothesis whether the neural mechanisms of the left-hemisphere enhance vocal pitch error detection and control in AP musicians compared with relative pitch (RP) musicians and non-musicians (NM). Results showed a stronger N1 response to pitch-shifted voice feedback in the right-hemisphere for both AP and RP musicians compared with the NM group. However, the left-hemisphere P2 component activation was greater in AP and RP musicians compared with NMs and also for the AP compared with RP musicians. The NM group was slower in generating compensatory vocal reactions to feedback pitch perturbation compared with musicians, and they failed to re-adjust their vocal pitch after the feedback perturbation was removed. These findings suggest that in the earlier stages of cortical neural processing, the right hemisphere is more active in musicians for detecting pitch changes in voice feedback. In the later stages, the left-hemisphere is more active during the processing of auditory feedback for vocal motor control and seems to involve specialized mechanisms that facilitate pitch processing in the AP compared with RP musicians. These findings indicate that the left hemisphere mechanisms of AP ability are associated with improved auditory feedback pitch processing during vocal pitch control in tasks such as speaking or singing. PMID:24355545
Estudo do comportamento vocal no ciclo menstrual: avaliação perceptivo-auditiva, acústica e auto-perceptiva Vocal behavior during menstrual cycle: perceptual-auditory, acoustic and self-perception analysis
Luciane C. de Figueiredo
Full Text Available Durante o período pré-menstrual é comum a ocorrência de disfonia, e são poucas as mulheres que se dão conta dessa variação da voz dentro do ciclo menstrual (Quinteiro, 1989. OBJETIVO: Verificar se há diferença no padrão vocal de mulheres no período de ovulação em relação ao primeiro dia do ciclo menstrual, utilizando-se da análise perceptivo-auditiva, da espectrografia, dos parâmetros acústicos e quando esta diferença está presente, se é percebida pelas mulheres. FORMA DE ESTUDO: Caso-controle. MATERIAL E MÉTODO: A amostra coletada foi de 30 estudantes de Fonoaudiologia, na faixa etária de 18 a 25 anos, não-fumantes, com ciclo menstrual regular e sem o uso de contraceptivo oral. As vozes foram gravadas no primeiro dia de menstruação e no décimo-terceiro dia pós-menstruação (ovulação, para posterior comparação. RESULTADOS: Observou-se durante o período menstrual que as vozes estão rouco-soprosa de grau leve a moderado, instáveis, sem a presença de quebra de sonoridade, com pitch e loudness adequados e ressonância equilibrada. Há pior qualidade de definição dos harmônicos, maior quantidade de ruído entre eles e menor extensão dos harmônicos superiores. Encontramos uma f0 mais aguda, jitter e shimmer aumentados e PHR diminuída. CONCLUSÃO: No período menstrual há mudanças na qualidade vocal, no comportamento dos harmônicos e nos parâmetros vocais (f0,jitter, shimmer e PHR. Além disso, a maioria das estudantes de Fonoaudiologia não percebeu a variação da voz durante o ciclo menstrual.During the premenstruation period dysphonia often can be observed and only few women are aware of this voice variation (Quinteiro, 1989. AIM: To verify if there are vocal quality variations between the ovulation period and the first day of the menstrual cycle, by using perceptual-auditory and acoustic analysis, including spectrography, and the self perception of the vocal changes when it occurs. STUDY DESIGN: Case
Albanita Gomes da Costa de Ceballos
the city of Salvador, Bahia. Teachers answered a questionnaire and were submitted to auditory vocal analysis. The GRBAS was used for the diagnosis of vocal disorders. RESULTS: The study population comprised 82.8% women, teachers with an average age of 40.7 years, teachers with higher education (88.4%, with an average workday of 38 hours per week, average 11.5 years of professional practice and average monthly income of R$1.817.18. The prevalence of voice disorders was 53.6%. (255 teachers. The bivariate analysis showed statistically significant associations between vocal disorders and age above 40 years (PR = 1.83; 95% CI; 1.27-2.64, family history of dysphonia (PR = 1.72; 95% CI; 1.06-2.80, over 20 hours of weekly working hours (PR = 1.66; 95% CI; 1.09-2.52 and presence of chalk dust in the classroom (PR = 1.70; 95% CI; 1.14-2.53. CONCLUSION: The study concluded that teachers, 40 years old and over, with a family history of dysphonia, working over 20 hours weekly, and teaching in classrooms with chalk dust are more likely to develop voice disorders than others.
Roberta Werlang Isolan-Cury
Full Text Available OBJETIVO: Caracterizar a qualidade vocal, por meio de análise computadorizada e perceptivo-auditiva, de pacientes com hipertireoidismo (grupo A e hipotireoidismo (grupo B. MÉTODOS: Vinte mulheres não fumantes, com idades entre 18 e 55 anos, atendidas no Ambulatório de Endocrinologia da instituição, foram avaliadas após o diagnóstico clínico e laboratorial de hipertireoidismo ou hipotireoidismo. Os parâmetros investigados foram: tempo da doença, presença de queixa vocal, tempos máximos de fonação /a/, /s/ e /z/, freqüência fundamental (F0, ruído glótico (GNE. Os aspectos avaliados na análise perceptivo-auditiva, foram: coordenação pneumo-fonoarticulatória (coordenada ou incoordenada, pitch, loudness, ataque vocal, ressonância, velocidade de fala e qualidade vocal, que poderia ter até duas das seguintes classificações: neutra, rouca, soprosa, áspera ou tensa, e grau: leve, moderado ou severo. Os dados foram tabulados e analisados estatisticamente através do programa EPI-INFO 6.04b, método qualitativo Fisher, com nível de significância menor do que 0.05. RESULTADOS: A análise perceptivo-auditiva mostrou que sete pacientes hipotireoideos e nove pacientes hipertireoideos apresentaram alteração na qualidade vocal. Oito pacientes em ambos os grupos apresentaram incoordenação pneumo-fonoarticulatória. Oito pacientes do grupo A e seis pacientes do grupo B referiam queixas vocais como rouquidão e voz grossa, respectivamente. Na análise acústica, nove pacientes apresentaram o ruído glótico alterado. CONCLUSÃO: Os resultados evidenciaram grande incidência de alteração vocal nos grupos estudados (grupos dos pacientes com hipertireoidismo e com hipotireoidismo, o que demonstra a relação entre disfonia e disfunções tireoideanas.PURPOSE: To characterize the vocal quality of subjects with hyperthyroidism (group A, and hypothyroidism (group B through a computer-aided and auditory-perceptive analysis. METHODS
Prickett, Carol A.; Bridges, Madeline S.
Explores whether the basic song repertoire of vocal/choral music education majors is significantly better than instrumental music education majors. Participants attempted to identify 25 standard songs. Reveals no significant difference between the two groups, indicating that neither had developed a strong repertoire of songs. (CMK)
Behroozmand, Roozbeh; Korzyukov, Oleg; Larson, Charles R
The present study investigated the neural mechanisms of voice pitch control for different levels of harmonic complexity in the auditory feedback. Event-related potentials (ERPs) were recorded in response to+200 cents pitch perturbations in the auditory feedback of self-produced natural human vocalizations, complex and pure tone stimuli during active vocalization and passive listening conditions. During active vocal production, ERP amplitudes were largest in response to pitch shifts in the natural voice, moderately large for non-voice complex stimuli and smallest for the pure tones. However, during passive listening, neural responses were equally large for pitch shifts in voice and non-voice complex stimuli but still larger than that for pure tones. These findings suggest that pitch change detection is facilitated for spectrally rich sounds such as natural human voice and non-voice complex stimuli compared with pure tones. Vocalization-induced increase in neural responses for voice feedback suggests that sensory processing of naturally-produced complex sounds such as human voice is enhanced by means of motor-driven mechanisms (e.g. efference copies) during vocal production. This enhancement may enable the audio-vocal system to more effectively detect and correct for vocal errors in the feedback of natural human vocalizations to maintain an intended vocal output for speaking. Copyright Â© 2011 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.
Moss, Cynthia F; Chiu, Chen; Surlykke, Annemarie
Echolocation operates through adaptive sensorimotor systems that collectively enable the bat to localize and track sonar objects as it flies. The features of sonar signals used by a bat to probe its surroundings determine the information available to its acoustic imaging system. In turn, the bat......'s perception of a complex scene guides its active adjustments in the features of subsequent sonar vocalizations. Here, we propose that the bat's active vocal-motor behaviors play directly into its representation of a dynamic auditory scene....
Pinaud, R.; Terleph, T. A.; Wynne, R. D.; Tremere, L. A.
Songbirds have emerged as powerful experimental models for the study of auditory processing of complex natural communication signals. Intact hearing is necessary for several behaviors in developing and adult animals including vocal learning, territorial defense, mate selection and individual recognition. These behaviors are thought to require the processing, discrimination and memorization of songs. Although much is known about the brain circuits that participate in sensorimotor (auditory-vocal) integration, especially the ``song-control" system, less is known about the anatomical and functional organization of central auditory pathways. Here we discuss findings associated with a telencephalic auditory area known as the caudomedial nidopallium (NCM). NCM has attracted significant interest as it exhibits functional properties that may support higher order auditory functions such as stimulus discrimination and the formation of auditory memories. NCM neurons are vigorously dr iven by auditory stimuli. Interestingly, these responses are selective to conspecific, relative to heterospecific songs and artificial stimuli. In addition, forms of experience-dependent plasticity occur in NCM and are song-specific. Finally, recent experiments employing high-throughput quantitative proteomics suggest that complex protein regulatory pathways are engaged in NCM as a result of auditory experience. These molecular cascades are likely central to experience-associated plasticity of NCM circuitry and may be part of a network of calcium-driven molecular events that support the formation of auditory memory traces.
Lohr, Bernard; Dooling, Robert J.; Gill, Douglas E.
Some grassland bird species, in particular grasshopper sparrows (Ammodramus savannarum), sing songs with especially high mean frequencies (7.0-8.0 kHz). Acoustic interference is one potential explanation for the evolution of high frequency vocalizations, particularly in open habitats. We tested predictions from a model of effective auditory communication distances to understand the potential effects of vocal production and environmental auditory masking on vocal behavior and territoriality. Variation in the spectral structure of songs and the size and shape of territories was measured for grasshopper sparrows in typical grassland habitats. Median territory areas were 1629 m2 at a site in the center of the species range in Nebraska, and 1466 m2 at our study site in Maryland, with average territory diameters measuring 20.2 m. Species densities and sound pressure levels also were determined for stridulating insects and other noise sources in the habitat. Based on current models of effective communication distances, known noise levels, and information on hearing abilities, our results suggest that auditory sensitivity and environmental noise could be factors influencing the mean frequency and spatial dynamics of territorial behavior in grassland birds. [Work supported by NIH and the CRFRC.
Forlano, Paul M; Sisneros, Joseph A
The plainfin midshipman fish (Porichthys notatus) is a well-studied model to understand the neural and endocrine mechanisms underlying vocal-acoustic communication across vertebrates. It is well established that steroid hormones such as estrogen drive seasonal peripheral auditory plasticity in female Porichthys in order to better encode the male's advertisement call. However, little is known of the neural substrates that underlie the motivation and coordinated behavioral response to auditory social signals. Catecholamines, which include dopamine and noradrenaline, are good candidates for this function, as they are thought to modulate the salience of and reinforce appropriate behavior to socially relevant stimuli. This chapter summarizes our recent studies which aimed to characterize catecholamine innervation in the central and peripheral auditory system of Porichthys as well as test the hypotheses that innervation of the auditory system is seasonally plastic and catecholaminergic neurons are activated in response to conspecific vocalizations. Of particular significance is the discovery of direct dopaminergic innervation of the saccule, the main hearing end organ, by neurons in the diencephalon, which also robustly innervate the cholinergic auditory efferent nucleus in the hindbrain. Seasonal changes in dopamine innervation in both these areas appear dependent on reproductive state in females and may ultimately function to modulate the sensitivity of the peripheral auditory system as an adaptation to the seasonally changing soundscape. Diencephalic dopaminergic neurons are indeed active in response to exposure to midshipman vocalizations and are in a perfect position to integrate the detection and appropriate motor response to conspecific acoustic signals for successful reproduction.
Larson Charles R
Full Text Available Abstract Background The motor-driven predictions about expected sensory feedback (efference copies have been proposed to play an important role in recognition of sensory consequences of self-produced motor actions. In the auditory system, this effect was suggested to result in suppression of sensory neural responses to self-produced voices that are predicted by the efference copies during vocal production in comparison with passive listening to the playback of the identical self-vocalizations. In the present study, event-related potentials (ERPs were recorded in response to upward pitch shift stimuli (PSS with five different magnitudes (0, +50, +100, +200 and +400 cents at voice onset during active vocal production and passive listening to the playback. Results Results indicated that the suppression of the N1 component during vocal production was largest for unaltered voice feedback (PSS: 0 cents, became smaller as the magnitude of PSS increased to 200 cents, and was almost completely eliminated in response to 400 cents stimuli. Conclusions Findings of the present study suggest that the brain utilizes the motor predictions (efference copies to determine the source of incoming stimuli and maximally suppresses the auditory responses to unaltered feedback of self-vocalizations. The reduction of suppression for 50, 100 and 200 cents and its elimination for 400 cents pitch-shifted voice auditory feedback support the idea that motor-driven suppression of voice feedback leads to distinctly different sensory neural processing of self vs. non-self vocalizations. This characteristic may enable the audio-vocal system to more effectively detect and correct for unexpected errors in the feedback of self-produced voice pitch compared with externally-generated sounds.
Behroozmand, Roozbeh; Larson, Charles R
The motor-driven predictions about expected sensory feedback (efference copies) have been proposed to play an important role in recognition of sensory consequences of self-produced motor actions. In the auditory system, this effect was suggested to result in suppression of sensory neural responses to self-produced voices that are predicted by the efference copies during vocal production in comparison with passive listening to the playback of the identical self-vocalizations. In the present study, event-related potentials (ERPs) were recorded in response to upward pitch shift stimuli (PSS) with five different magnitudes (0, +50, +100, +200 and +400 cents) at voice onset during active vocal production and passive listening to the playback. Results indicated that the suppression of the N1 component during vocal production was largest for unaltered voice feedback (PSS: 0 cents), became smaller as the magnitude of PSS increased to 200 cents, and was almost completely eliminated in response to 400 cents stimuli. Findings of the present study suggest that the brain utilizes the motor predictions (efference copies) to determine the source of incoming stimuli and maximally suppresses the auditory responses to unaltered feedback of self-vocalizations. The reduction of suppression for 50, 100 and 200 cents and its elimination for 400 cents pitch-shifted voice auditory feedback support the idea that motor-driven suppression of voice feedback leads to distinctly different sensory neural processing of self vs. non-self vocalizations. This characteristic may enable the audio-vocal system to more effectively detect and correct for unexpected errors in the feedback of self-produced voice pitch compared with externally-generated sounds.
Pepperberg, Irene M.
This chapter briefly reviews what is known-and what remains to be understood--about Grey parrot vocal learning. I review Greys' physical capacities--issues of auditory perception and production--then discuss how these capacities are used in vocal learning and can be recruited for referential communication with humans. I discuss cross-species…
Yoder, Kathleen M; Lu, Kai; Vicario, David S
Estradiol (E2) has recently been shown to modulate sensory processing in an auditory area of the songbird forebrain, the caudomedial nidopallium (NCM). When a bird hears conspecific song, E2 increases locally in NCM, where neurons express both the aromatase enzyme that synthesizes E2 from precursors and estrogen receptors. Auditory responses in NCM show a form of neuronal memory: repeated playback of the unique learned vocalizations of conspecific individuals induces long-lasting stimulus-specific adaptation of neural responses to each vocalization. To test the role of E2 in this auditory memory, we treated adult male zebra finches (n=16) with either the aromatase inhibitor fadrozole (FAD) or saline for 8 days. We then exposed them to 'training' songs and, 6 h later, recorded multiunit auditory responses with an array of 16 microelectrodes in NCM. Adaptation rates (a measure of stimulus-specific adaptation) to playbacks of training and novel songs were computed, using established methods, to provide a measure of neuronal memory. Recordings from the FAD-treated birds showed a significantly reduced memory for the training songs compared with saline-treated controls, whereas auditory processing for novel songs did not differ between treatment groups. In addition, FAD did not change the response bias in favor of conspecific over heterospecific song stimuli. Our results show that E2 depletion affects the neuronal memory for vocalizations in songbird NCM, and suggest that E2 plays a necessary role in auditory processing and memory for communication signals.
Charles R Larson; Donald A Robin
The pitch-shift paradigm has become a widely used method for studying the role of voice pitch auditory feedback in voice control. This paradigm introduces small, brief pitch shifts in voice auditory feedback to vocalizing subjects. The perturbations trigger a reflexive mechanism that counteracts the change in pitch. The underlying mechanisms of the vocal responses are thought to reflect a negative feedback control system that is similar to constructs developed to explain other forms of motor ...
Sisneros, Joseph A
The plainfin midshipman fish (Porichthys notatus Girard, 1854) is a vocal species of batrachoidid fish that generates acoustic signals for intraspecific communication during social and reproductive activity and has become a good model for investigating the neural and endocrine mechanisms of vocal-acoustic communication. Reproductively active female plainfin midshipman fish use their auditory sense to detect and locate "singing" males, which produce a multiharmonic advertisement call to attract females for spawning. The seasonal onset of male advertisement calling in the midshipman fish coincides with an increase in the range of frequency sensitivity of the female's inner ear saccule, the main organ of hearing, thus leading to enhanced encoding of the dominant frequency components of male advertisement calls. Non-reproductive females treated with either testosterone or 17β-estradiol exhibit a dramatic increase in the inner ear's frequency sensitivity that mimics the reproductive female's auditory phenotype and leads to an increased detection of the male's advertisement call. This novel form of auditory plasticity provides an adaptable mechanism that enhances coupling between sender and receiver in vocal communication. This review focuses on recent evidence for seasonal reproductive-state and steroid-dependent plasticity of auditory frequency sensitivity in the peripheral auditory system of the midshipman fish. The potential steroid-dependent mechanism(s) that lead to this novel form of auditory and behavioral plasticity are also discussed. © 2009 ISZS, Blackwell Publishing and IOZ/CAS.
Full Text Available OBJETIVO: descrever a qualidade vocal de personagens idosos dos filmes de Hollywood. MÉTODOS: foram colhidas 50 amostras de fala de personagens idosos, 11 do sexo feminino e 39 do masculino, de 38 filmes hollywoodianos dos anos de 1993 a 2001. Através da análise perceptivo-auditiva das amostras de fala, 20 fonoaudiólogos treinados classificaram cada personagem em idoso e não idoso, além de avaliarem as vozes quanto aos seguintes parâmetros citados pela literatura como mais alterados: rouquidão, crepitação, soprosidade, tensão, aspereza, astenia, nasalidade, tremor, modulação, pitch e estabilidade da frequência fundamental. RESULTADOS: após a análise perceptivo-auditiva, foi observado que a grande maioria dos atores (82% utilizou voz de idoso para representar seus papéis. O marcador mais evidente nas vozes foi alteração na qualidade vocal (92%, demonstrada por crepitação (80%, soprosidade (54%, tensão (38%, rouquidão (30% e astenia (28%. O segundo marcador mais utilizado pelos atores nas suas representações foi a modulação vocal ampla e variada (44%. Também foram observadas alterações no controle da voz (36% e instabilidade da frequência fundamental (38%. CONCLUSÃO: a partir dos resultados obtidos pode-se concluir que os filmes de Hollywood caracterizam o idoso através de desvios evidentes na qualidade e modulação da voz, utilizando tipos de vozes alteradas e modulação vocal ampla e instável.PURPOSE: to describe the vocal quality of Hollywood movies characters playing elderly people roles. METHODS: a total of 50 aged character voice samples were used, 11 female and 39 male, from 38 Hollywood movies from the period between 1993 and 2001. Twenty speech therapists performed a perceptual auditory analysis. The listener's task required classifying each character either as elderly or as adult by their speech features, and also assessing their voices following the parameters that are most frequently addressed in the
Giret, Nicolas; Menardy, Fabien; Del Negro, Catherine
Understanding how communication sounds are encoded in the central auditory system is critical to deciphering the neural bases of acoustic communication. Songbirds use learned or unlearned vocalizations in a variety of social interactions. They have telencephalic auditory areas specialized for processing natural sounds and considered as playing a critical role in the discrimination of behaviorally relevant vocal sounds. The zebra finch, a highly social songbird species, forms lifelong pair bonds. Only male zebra finches sing. However, both sexes produce the distance call when placed in visual isolation. This call is sexually dimorphic, is learned only in males and provides support for individual recognition in both sexes. Here, we assessed whether auditory processing of distance calls differs between paired males and females by recording spiking activity in a secondary auditory area, the caudolateral mesopallium (CLM), while presenting the distance calls of a variety of individuals, including the bird itself, the mate, familiar and unfamiliar males and females. In males, the CLM is potentially involved in auditory feedback processing important for vocal learning. Based on both the analyses of spike rates and temporal aspects of discharges, our results clearly indicate that call-evoked responses of CLM neurons are sexually dimorphic, being stronger, lasting longer, and conveying more information about calls in males than in females. In addition, how auditory responses vary among call types differ between sexes. In females, response strength differs between familiar male and female calls. In males, temporal features of responses reveal a sensitivity to the bird's own call. These findings provide evidence that sexual dimorphism occurs in higher-order processing areas within the auditory system. They suggest a sexual dimorphism in the function of the CLM, contributing to transmit information about the self-generated calls in males and to storage of information about the
Full Text Available Understanding how communication sounds are encoded in the central auditory system is critical to deciphering the neural bases of acoustic communication. Songbirds use learned or unlearned vocalizations in a variety of social interactions. They have telencephalic auditory areas specialized for processing natural sounds and considered as playing a critical role in the discrimination of behaviorally relevant vocal sounds. The zebra finch, a highly social songbird species, forms lifelong pair bonds. Only male zebra finches sing. However, both sexes produce the distance call when placed in visual isolation. This call is sexually dimorphic, is learned only in males and provides support for individual recognition in both sexes. Here, we assessed whether auditory processing of distance calls differs between paired males and females by recording spiking activity in a secondary auditory area, the caudolateral mesopallium (CLM, while presenting the distance calls of a variety of individuals, including the bird itself, the mate, familiar and unfamiliar males and females. In males, the CLM is potentially involved in auditory feedback processing important for vocal learning. Based on both the analyses of spike rates and temporal aspects of discharges, our results clearly indicate that call-evoked responses of CLM neurons are sexually dimorphic, being stronger, lasting longer and conveying more information about calls in males than in females. In addition, how auditory responses vary among call types differ between sexes. In females, response strength differs between familiar male and female calls. In males, temporal features of responses reveal a sensitivity to the bird’s own call. These findings provide evidence that sexual dimorphism occurs in higher-order processing areas within the auditory system. They suggest a sexual dimorphism in the function of the CLM, contributing to transmit information about the self-generated calls in males and to storage of
Kari L Hoffman
Full Text Available Auditory and visual signals often occur together, and the two sensory channels are known to infl uence each other to facilitate perception. The neural basis of this integration is not well understood, although other forms of multisensory infl uences have been shown to occur at surprisingly early stages of processing in cortex. Primary visual cortex neurons can show frequency-tuning to auditory stimuli, and auditory cortex responds selectively to certain somatosensory stimuli, supporting the possibility that complex visual signals may modulate early stages of auditory processing. To elucidate which auditory regions, if any, are responsive to complex visual stimuli, we recorded from auditory cortex and the superior temporal sulcus while presenting visual stimuli consisting of various objects, neutral faces, and facial expressions generated during vocalization. Both objects and conspecifi c faces elicited robust fi eld potential responses in auditory cortex sites, but the responses varied by category: both neutral and vocalizing faces had a highly consistent negative component (N100 followed by a broader positive component (P180 whereas object responses were more variable in time and shape, but could be discriminated consistently from the responses to faces. The face response did not vary within the face category, i.e., for expressive vs. neutral face stimuli. The presence of responses for both objects and neutral faces suggests that auditory cortex receives highly informative visual input that is not restricted to those stimuli associated with auditory components. These results reveal selectivity for complex visual stimuli in a brain region conventionally described as non-visual unisensory cortex.
Höhne, Johannes; Tangermann, Michael
Realizing the decoding of brain signals into control commands, brain-computer interfaces (BCI) aim to establish an alternative communication pathway for locked-in patients. In contrast to most visual BCI approaches which use event-related potentials (ERP) of the electroencephalogram, auditory BCI systems are challenged with ERP responses, which are less class-discriminant between attended and unattended stimuli. Furthermore, these auditory approaches have more complex interfaces which imposes a substantial workload on their users. Aiming for a maximally user-friendly spelling interface, this study introduces a novel auditory paradigm: “CharStreamer”. The speller can be used with an instruction as simple as “please attend to what you want to spell”. The stimuli of CharStreamer comprise 30 spoken sounds of letters and actions. As each of them is represented by the sound of itself and not by an artificial substitute, it can be selected in a one-step procedure. The mental mapping effort (sound stimuli to actions) is thus minimized. Usability is further accounted for by an alphabetical stimulus presentation: contrary to random presentation orders, the user can foresee the presentation time of the target letter sound. Healthy, normal hearing users (n = 10) of the CharStreamer paradigm displayed ERP responses that systematically differed between target and non-target sounds. Class-discriminant features, however, varied individually from the typical N1-P2 complex and P3 ERP components found in control conditions with random sequences. To fully exploit the sequential presentation structure of CharStreamer, novel data analysis approaches and classification methods were introduced. The results of online spelling tests showed that a competitive spelling speed can be achieved with CharStreamer. With respect to user rating, it clearly outperforms a control setup with random presentation sequences. PMID:24886978
Elisângela Barros Soares
, observational and crossed-nature study. A perceptual-auditive analysis was carried out for the sampling RESULTS: we noted that the majority of the guides demonstrated adequate loudness, normal pitch and modified voice. Moreover, the averages of the maximum times of phonation of the vowels and the affricative and isocronic vocal attack were reduced. The resonance, in the majority of the guides, was balanced, but it had an incidence of laryngeal - pharyngeal resonance. Articulation was precise, with mixed and nasal respiratory mode and type, respectively. As for GRBAS scale, the alterations appeared in light form in G (degree of vocal alteration in 68%. CONCLUSION: in the studied sample, the majority was of the feminine gender with average age of 46 years, and vocal profile characterized by reduced maximum phonation time, adequate s/z relation, adequate isocronic vocal attack, normal pitch, loudness, modified vocal quality, with presence of roughness, breathiness, tension. The resonance of the majority was balanced and articulation precise, with mixed and nasal respiratory mode and type, respectively. As for GRBAS scale, the alterations appeared in light form in G (degree of vocal alteration in 68% and S (tension in 78% of the subjects.
Schneider, David M; Mooney, Richard
In the auditory system, corollary discharge signals are theorized to facilitate normal hearing and the learning of acoustic behaviors, including speech and music. Despite clear evidence of corollary discharge signals in the auditory cortex and their presumed importance for hearing and auditory-guided motor learning, the circuitry and function of corollary discharge signals in the auditory cortex are not well described. In this review, we focus on recent developments in the mouse and songbird that provide insights into the circuitry that transmits corollary discharge signals to the auditory system and the function of these signals in the context of hearing and vocal learning. Copyright © 2015 Elsevier Ltd. All rights reserved.
Plakke, Bethany; Hwang, Jaewon; Romanski, Lizabeth M
The prefrontal cortex is associated with cognitive functions that include planning, reasoning, decision-making, working memory, and communication. Neurophysiology and neuropsychology studies have established that dorsolateral prefrontal cortex is essential in spatial working memory while the ventral frontal lobe processes language and communication signals. Single-unit recordings in nonhuman primates has shown that ventral prefrontal (VLPFC) neurons integrate face and vocal information and are active during audiovisual working memory. However, whether VLPFC is essential in remembering face and voice information is unknown. We therefore trained nonhuman primates in an audiovisual working memory paradigm using naturalistic face-vocalization movies as memoranda. We inactivated VLPFC, with reversible cortical cooling, and examined performance when faces, vocalizations or both faces and vocalization had to be remembered. We found that VLPFC inactivation impaired subjects' performance in audiovisual and auditory-alone versions of the task. In contrast, VLPFC inactivation did not disrupt visual working memory. Our studies demonstrate the importance of VLPFC in auditory and audiovisual working memory for social stimuli but suggest a different role for VLPFC in unimodal visual processing. The ventral frontal lobe, or inferior frontal gyrus, plays an important role in audiovisual communication in the human brain. Studies with nonhuman primates have found that neurons within ventral prefrontal cortex (VLPFC) encode both faces and vocalizations and that VLPFC is active when animals need to remember these social stimuli. In the present study, we temporarily inactivated VLPFC by cooling the cortex while nonhuman primates performed a working memory task. This impaired the ability of subjects to remember a face and vocalization pair or just the vocalization alone. Our work highlights the importance of the primate VLPFC in the processing of faces and vocalizations in a manner that
Wronkiewicz, Mark; Larson, Eric; Lee, Adrian Kc
Brain-computer interface (BCI) technology allows users to generate actions based solely on their brain signals. However, current non-invasive BCIs generally classify brain activity recorded from surface electroencephalography (EEG) electrodes, which can hinder the application of findings from modern neuroscience research. In this study, we use source imaging-a neuroimaging technique that projects EEG signals onto the surface of the brain-in a BCI classification framework. This allowed us to incorporate prior research from functional neuroimaging to target activity from a cortical region involved in auditory attention. Classifiers trained to detect attention switches performed better with source imaging projections than with EEG sensor signals. Within source imaging, including subject-specific anatomical MRI information (instead of using a generic head model) further improved classification performance. This source-based strategy also reduced accuracy variability across three dimensionality reduction techniques-a major design choice in most BCIs. Our work shows that source imaging provides clear quantitative and qualitative advantages to BCIs and highlights the value of incorporating modern neuroscience knowledge and methods into BCI systems.
Baykara, E; Ruf, C A; Fioravanti, C; Käthner, I; Simon, N; Kleih, S C; Kübler, A; Halder, S
Brain-computer interface (BCI) technology aims at helping end-users with severe motor paralysis to communicate with their environment without using the natural output pathways of the brain. For end-users in complete paralysis, loss of gaze control may necessitate non-visual BCI systems. The present study investigated the effect of training on performance with an auditory P300 multi-class speller paradigm. For half of the participants, spatial cues were added to the auditory stimuli to see whether performance can be further optimized. The influence of motivation, mood and workload on performance and P300 component was also examined. In five sessions, 16 healthy participants were instructed to spell several words by attending to animal sounds representing the rows and columns of a 5 × 5 letter matrix. 81% of the participants achieved an average online accuracy of ⩾ 70%. From the first to the fifth session information transfer rates increased from 3.72 bits/min to 5.63 bits/min. Motivation significantly influenced P300 amplitude and online ITR. No significant facilitative effect of spatial cues on performance was observed. Training improves performance in an auditory BCI paradigm. Motivation influences performance and P300 amplitude. The described auditory BCI system may help end-users to communicate independently of gaze control with their environment. Copyright © 2015 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.
Dansereau Richard M
Full Text Available We present a new technique for separating two speech signals from a single recording. The proposed method bridges the gap between underdetermined blind source separation techniques and those techniques that model the human auditory system, that is, computational auditory scene analysis (CASA. For this purpose, we decompose the speech signal into the excitation signal and the vocal-tract-related filter and then estimate the components from the mixed speech using a hybrid model. We first express the probability density function (PDF of the mixed speech's log spectral vectors in terms of the PDFs of the underlying speech signal's vocal-tract-related filters. Then, the mean vectors of PDFs of the vocal-tract-related filters are obtained using a maximum likelihood estimator given the mixed signal. Finally, the estimated vocal-tract-related filters along with the extracted fundamental frequencies are used to reconstruct estimates of the individual speech signals. The proposed technique effectively adds vocal-tract-related filter characteristics as a new cue to CASA models using a new grouping technique based on an underdetermined blind source separation. We compare our model with both an underdetermined blind source separation and a CASA method. The experimental results show that our model outperforms both techniques in terms of SNR improvement and the percentage of crosstalk suppression.
Mohammad H. Radfar
Full Text Available We present a new technique for separating two speech signals from a single recording. The proposed method bridges the gap between underdetermined blind source separation techniques and those techniques that model the human auditory system, that is, computational auditory scene analysis (CASA. For this purpose, we decompose the speech signal into the excitation signal and the vocal-tract-related filter and then estimate the components from the mixed speech using a hybrid model. We first express the probability density function (PDF of the mixed speech's log spectral vectors in terms of the PDFs of the underlying speech signal's vocal-tract-related filters. Then, the mean vectors of PDFs of the vocal-tract-related filters are obtained using a maximum likelihood estimator given the mixed signal. Finally, the estimated vocal-tract-related filters along with the extracted fundamental frequencies are used to reconstruct estimates of the individual speech signals. The proposed technique effectively adds vocal-tract-related filter characteristics as a new cue to CASA models using a new grouping technique based on an underdetermined blind source separation. We compare our model with both an underdetermined blind source separation and a CASA method. The experimental results show that our model outperforms both techniques in terms of SNR improvement and the percentage of crosstalk suppression.
Goutte, Sandra; Mason, Matthew J; Christensen-Dalsgaard, Jakob
The emergence and maintenance of animal communication systems requires the co-evolution of signal and receiver. Frogs and toads rely heavily on acoustic communication for coordinating reproduction and typically have ears tuned to the dominant frequency of their vocalizations, allowing discriminat...
The present study focused on gender differences in emotion identification from auditory and visual stimuli produced by two male and two female actors. Differences in emotion identification from nonsense samples, language samples and prolonged vowels were investigated. It was also studied whether auditory stimuli can convey the emotional content of speech without visual stimuli, and whether visual stimuli can convey the emotional content of speech without auditory stimuli. The aim was to get a better knowledge of vocal attributes and a more holistic understanding of the nonverbal communication of emotion. Females tended to be more accurate in emotion identification than males. Voice quality parameters played a role in emotion identification in both genders. The emotional content of the samples was best conveyed by nonsense sentences, better than by prolonged vowels or shared native language of the speakers and participants. Thus, vocal non-verbal communication tends to affect the interpretation of emotion even in the absence of language. The emotional stimuli were better recognized from visual stimuli than auditory stimuli by both genders. Visual information about speech may not be connected to the language; instead, it may be based on the human ability to understand the kinetic movements in speech production more readily than the characteristics of the acoustic cues.
Halder, S; Rea, M; Andreoni, R; Nijboer, F; Hammer, E M; Kleih, S C; Birbaumer, N; Kübler, A
Brain-computer interfaces (BCIs) provide non-muscular communication for individuals diagnosed with late-stage motoneuron disease (e.g., amyotrophic lateral sclerosis (ALS)). In the final stages of the disease, a BCI cannot rely on the visual modality. This study examined a method to achieve high accuracies using auditory stimuli only. We propose an auditory BCI based on a three-stimulus paradigm. This paradigm is similar to the standard oddball but includes an additional target (i.e. two target stimuli, one frequent stimulus). Three versions of the task were evaluated in which the target stimuli differed in loudness, pitch or direction. Twenty healthy participants achieved an average information transfer rate (ITR) of up to 2.46 bits/min and accuracies of 78.5%. Most subjects (14 of 20) achieved their best performance with targets differing in pitch. With this study, the viability of the paradigm was shown for healthy participants and will next be evaluated with individuals diagnosed with ALS or locked-in syndrome (LIS) after stroke. The here presented BCI offers communication with binary choices (yes/no) independent of vision. As it requires only little time per selection, it may constitute a reliable means of communication for patients who lost all motor function and have a short attention span. 2009 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.
Urbano, Catherine M; Peterson, Jennifer R; Cooper, Brenton G
Songbirds and humans use auditory feedback to acquire and maintain their vocalizations. The Bengalese finch (Lonchura striata domestica) is a songbird species that rapidly modifies its vocal output to adhere to an internal song memory. In this species, the left side of the bipartite vocal organ is specialized for producing louder, higher frequencies (≥2.2kHz) and denervation of the left vocal muscles eliminates these notes. Thus, the return of higher frequency notes after cranial nerve injury can be used as a measure of vocal recovery. Either the left or right side of the syrinx was denervated by resection of the tracheosyringeal portion of the hypoglossal nerve. Histologic analyses of syringeal muscle tissue showed significant muscle atrophy in the denervated side. After left nerve resection, songs were mainly composed of lower frequency syllables, but three out of five birds recovered higher frequency syllables. Right nerve resection minimally affected phonology, but it did change song syntax; syllable sequence became abnormally stereotyped after right nerve resection. Therefore, damage to the neuromuscular control of sound production resulted in reduced motor variability, and Bengalese finches are a potential model for functional vocal recovery following cranial nerve injury. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.
Zhu, Lin L; Beauchamp, Michael S
Cortex in and around the human posterior superior temporal sulcus (pSTS) is known to be critical for speech perception. The pSTS responds to both the visual modality (especially biological motion) and the auditory modality (especially human voices). Using fMRI in single subjects with no spatial smoothing, we show that visual and auditory selectivity are linked. Regions of the pSTS were identified that preferred visually presented moving mouths (presented in isolation or as part of a whole face) or moving eyes. Mouth-preferring regions responded strongly to voices and showed a significant preference for vocal compared with nonvocal sounds. In contrast, eye-preferring regions did not respond to either vocal or nonvocal sounds. The converse was also true: regions of the pSTS that showed a significant response to speech or preferred vocal to nonvocal sounds responded more strongly to visually presented mouths than eyes. These findings can be explained by environmental statistics. In natural environments, humans see visual mouth movements at the same time as they hear voices, while there is no auditory accompaniment to visual eye movements. The strength of a voxel's preference for visual mouth movements was strongly correlated with the magnitude of its auditory speech response and its preference for vocal sounds, suggesting that visual and auditory speech features are coded together in small populations of neurons within the pSTS. SIGNIFICANCE STATEMENT Humans interacting face to face make use of auditory cues from the talker's voice and visual cues from the talker's mouth to understand speech. The human posterior superior temporal sulcus (pSTS), a brain region known to be important for speech perception, is complex, with some regions responding to specific visual stimuli and others to specific auditory stimuli. Using BOLD fMRI, we show that the natural statistics of human speech, in which voices co-occur with mouth movements, are reflected in the neural architecture of
Rutkowski, Tomasz M; Mori, Hiromu
The paper presents a report on the recently developed BCI alternative for users suffering from impaired vision (lack of focus or eye-movements) or from the so-called "ear-blocking-syndrome" (limited hearing). We report on our recent studies of the extents to which vibrotactile stimuli delivered to the head of a user can serve as a platform for a brain computer interface (BCI) paradigm. In the proposed tactile and bone-conduction auditory BCI novel multiple head positions are used to evoke combined somatosensory and auditory (via the bone conduction effect) P300 brain responses, in order to define a multimodal tactile and bone-conduction auditory brain computer interface (tbcaBCI). In order to further remove EEG interferences and to improve P300 response classification synchrosqueezing transform (SST) is applied. SST outperforms the classical time-frequency analysis methods of the non-linear and non-stationary signals such as EEG. The proposed method is also computationally more effective comparing to the empirical mode decomposition. The SST filtering allows for online EEG preprocessing application which is essential in the case of BCI. Experimental results with healthy BCI-naive users performing online tbcaBCI, validate the paradigm, while the feasibility of the concept is illuminated through information transfer rate case studies. We present a comparison of the proposed SST-based preprocessing method, combined with a logistic regression (LR) classifier, together with classical preprocessing and LDA-based classification BCI techniques. The proposed tbcaBCI paradigm together with data-driven preprocessing methods are a step forward in robust BCI applications research. Copyright © 2014 Elsevier B.V. All rights reserved.
McMullen, Kyla A.
Although the concept of virtual spatial audio has existed for almost twenty-five years, only in the past fifteen years has modern computing technology enabled the real-time processing needed to deliver high-precision spatial audio. Furthermore, the concept of virtually walking through an auditory environment did not exist. The applications of such an interface have numerous potential uses. Spatial audio has the potential to be used in various manners ranging from enhancing sounds delivered in virtual gaming worlds to conveying spatial locations in real-time emergency response systems. To incorporate this technology in real-world systems, various concerns should be addressed. First, to widely incorporate spatial audio into real-world systems, head-related transfer functions (HRTFs) must be inexpensively created for each user. The present study further investigated an HRTF subjective selection procedure previously developed within our research group. Users discriminated auditory cues to subjectively select their preferred HRTF from a publicly available database. Next, the issue of training to find virtual sources was addressed. Listeners participated in a localization training experiment using their selected HRTFs. The training procedure was created from the characterization of successful search strategies in prior auditory search experiments. Search accuracy significantly improved after listeners performed the training procedure. Next, in the investigation of auditory spatial memory, listeners completed three search and recall tasks with differing recall methods. Recall accuracy significantly decreased in tasks that required the storage of sound source configurations in memory. To assess the impacts of practical scenarios, the present work assessed the performance effects of: signal uncertainty, visual augmentation, and different attenuation modeling. Fortunately, source uncertainty did not affect listeners' ability to recall or identify sound sources. The present
Plakke, B; Romanski, L M
Working memory is the ability to employ recently seen or heard stimuli and apply them to changing cognitive context. Although much is known about language processing and visual working memory, the neurobiological basis of auditory working memory is less clear. Historically, part of the problem has been the difficulty in obtaining a robust animal model to study auditory short-term memory. In recent years there has been neurophysiological and lesion studies indicating a cortical network involving both temporal and frontal cortices. Studies specifically targeting the role of the prefrontal cortex (PFC) in auditory working memory have suggested that dorsal and ventral prefrontal regions perform different roles during the processing of auditory mnemonic information, with the dorsolateral PFC performing similar functions for both auditory and visual working memory. In contrast, the ventrolateral PFC (VLPFC), which contains cells that respond robustly to auditory stimuli and that process both face and vocal stimuli may be an essential locus for both auditory and audiovisual working memory. These findings suggest a critical role for the VLPFC in the processing, integrating, and retaining of communication information. This article is part of a Special Issue entitled SI: Auditory working memory. Copyright © 2015 Elsevier B.V. All rights reserved.
Shi, Lucy L; Giraldez-Rodriguez, Laureano A; Johns, Michael M
The aim of this study was to illustrate the risk of vocal fold atrophy in patients who receive serial subepithelial steroid injections for vocal fold scar. This study is a retrospective case report of two patients who underwent a series of weekly subepithelial infusions of 10 mg/mL dexamethasone for benign vocal fold lesion. Shortly after the procedures, both patients developed a weak and breathy voice. The first patient was a 53-year-old man with radiation-induced vocal fold stiffness. Six injections were performed unilaterally, and 1 week later, he developed unilateral vocal fold atrophy with new glottal insufficiency. The second patient was a 67-year-old woman with severe vocal fold inflammation related to laryngitis and calcinosis, Raynaud's phenomenon, esophagean dysmotility, sclerodactyly, and telangiectasia (CREST) syndrome. Five injections were performed bilaterally, and 1 week later, she developed bilateral vocal fold atrophy with a large midline glottal gap during phonation. In both cases, the steroid-induced vocal atrophy resolved spontaneously after 4 months. Serial subepithelial steroid infusions of the vocal folds, although safe in the majority of patients, carry the risk of causing temporary vocal fold atrophy when given at short intervals. Copyright Â© 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Zhang, Dian; Cui, Jianguo; Tang, Yezhong
In anurans reproductive behavior is strongly seasonal. During the spring, frogs emerge from hibernation and males vocalize for mating or advertising territories. Female frogs have the ability to evaluate the quality of the males' resources on the basis of these vocalizations. Although studies revealed that central single torus semicircularis neurons in frogs exhibit season plasticity, the plasticity of peripheral auditory sensitivity in frog is unknown. In this study the seasonally plasticity of peripheral auditory sensitivity was test in the Emei music frog Babina daunchina, by comparing thresholds and latencies of auditory brainstem responses (ABRs) evoked by tone pips and clicks in the reproductive and non-reproductive seasons. The results show that both ABR thresholds and latency differ significantly between the reproductive and non-reproductive seasons. The thresholds of tone pip evoked ABRs in the non-reproductive season increased significantly about 10 dB than those in the reproductive season for frequencies from 1 KHz to 6 KHz. ABR latencies to waveform valley values for tone pips for the same frequencies using appropriate threshold stimulus levels are longer than those in the reproductive season for frequencies from 1.5 to 6 KHz range, although from 0.2 to 1.5 KHz range it is shorter in the non-reproductive season. These results demonstrated that peripheral auditory frequency sensitivity exhibits seasonal plasticity changes which may be adaptive to seasonal reproductive behavior in frogs.
Diehl, Maria M; Romanski, Lizabeth M
Social communication relies on the integration of auditory and visual information, which are present in faces and vocalizations. Evidence suggests that the integration of information from multiple sources enhances perception compared with the processing of a unimodal stimulus. Our previous studies demonstrated that single neurons in the ventrolateral prefrontal cortex (VLPFC) of the rhesus monkey (Macaca mulatta) respond to and integrate conspecific vocalizations and their accompanying facial gestures. We were therefore interested in how VLPFC neurons respond differentially to matching (congruent) and mismatching (incongruent) faces and vocalizations. We recorded VLPFC neurons during the presentation of movies with congruent or incongruent species-specific facial gestures and vocalizations as well as their unimodal components. Recordings showed that while many VLPFC units are multisensory and respond to faces, vocalizations, or their combination, a subset of neurons showed a significant change in neuronal activity in response to incongruent versus congruent vocalization movies. Among these neurons, we typically observed incongruent suppression during the early stimulus period and incongruent enhancement during the late stimulus period. Incongruent-responsive VLPFC neurons were both bimodal and nonlinear multisensory, fostering their ability to respond to changes in either modality of a face-vocalization stimulus. These results demonstrate that ventral prefrontal neurons respond to changes in either modality of an audiovisual stimulus, which is important in identity processing and for the integration of multisensory communication information. Copyright © 2014 the authors 0270-6474/14/3411233-11$15.00/0.
Chaves, Patrícia P; Valdoria, Ciara M C; Amorim, M Clara P; Vasconcelos, Raquel O
Studies addressing structure-function relationships of the fish auditory system during development are sparse compared to other taxa. The Batrachoididae has become an important group to investigate mechanisms of auditory plasticity and evolution of auditory-vocal systems. A recent study reported ontogenetic improvements in the inner ear saccule sensitivity of the Lusitanian toadfish, Halobatrachus didactylus, but whether this results from changes in the sensory morphology remains unknown. We investigated how the macula and organization of auditory receptors in the saccule and utricle change during growth in this species. Inner ear sensory epithelia were removed from the end organs of previously PFA-fixed specimens, from non-vocal posthatch fry (23 cm). Epithelia were phalloidin-stained and analysed for area, shape, number and orientation patterns of hair cells (HC), and number and size of saccular supporting cells (SC). Saccular macula area expanded 41x in total, and significantly more (relative to body length) among vocal juveniles (2.3-2.9 cm). Saccular HC number increased 25x but HC density decreased, suggesting that HC addition is slower relative to epithelial growth. While SC density decreased, SC apical area increased, contributing to the epithelial expansion. The utricule revealed increased HC density (striolar region) and less epithelial expansion (5x) with growth, contrasting with the saccule that may have a different developmental pattern due to its larger size and main auditory functions. Both macula shape and HC orientation patterns were already established in the posthatch fry and retained throughout growth in both end organs. We suggest that previously reported ontogenetic improvements in saccular sensitivity might be associated with changes in HC number (not density), size and/or molecular mechanisms controlling HC sensitivity. This is one of the first studies investigating the ontogenetic development of the saccule and utricle in a vocal fish and
Silva, Bárbara Gabriela; Chammas, Tiago Visacre; Zenari, Marcia Simões; Moreira, Renata Rodrigues; Samelli, Alessandra Giannella; Nemr, Kátia
To measure the risk of dysphonia in teachers, as well as investigate whether the perceptual-auditory and acoustic aspects of the voice of teachers in situations of silence and noise, the signal-to-noise ratio, and the noise levels in the classroom are associated with the presence of dysphonia. This is an observational cross-sectional research with 23 primary and secondary school teachers from a private school in the municipality of São Paulo, Brazil, divided into the groups without dysphonia and with dysphonia. We performed the following procedures: general Dysphonia Risk Screening Protocol (General-DRSP) and complementary to speaking voice - teacher (Specific-DRSP), voice recording during class and in an individual situation in a silent room, and measurement of the signal-to-noise ratio and noise levels of classrooms. We have found differences between groups regarding physical activity (General-DRSP) and particularities of the profession (Specific-DRSP), as well as in all aspects of the perceptual-auditory vocal analysis. We have found signs of voice wear in the group without dysphonia. Regarding the vocal resources in the situations of noise and silence, we have identified a difference for the production of abrupt vocal attack and the tendency of a more precise speech in the situation of noise. Both the signal-to-noise ratio and the room noise levels during class were high in both groups. Teachers in both groups are at high risk for developing dysphonia and have negative vocal signals to a greater or lesser extent. Signal-to-noise ratio was inadequate in most classrooms, considering the standards for both children with normal hearing and with hearing loss, as well as equivalent noise levels.
Levendoski, Elizabeth Erickson; Leydon, Ciara; Thibeault, Susan L.
Purpose: Vocal fold epithelium is composed of layers of individual epithelial cells joined by junctional complexes constituting a unique interface with the external environment. This barrier provides structural stability to the vocal folds and protects underlying connective tissue from injury while being nearly continuously exposed to potentially…
De Vos, Maarten; Gandras, Katharina; Debener, Stefan
In a previous study we presented a low-cost, small, and wireless 14-channel EEG system suitable for field recordings (Debener et al., 2012, psychophysiology). In the present follow-up study we investigated whether a single-trial P300 response can be reliably measured with this system, while subjects freely walk outdoors. Twenty healthy participants performed a three-class auditory oddball task, which included rare target and non-target distractor stimuli presented with equal probabilities of 16%. Data were recorded in a seated (control condition) and in a walking condition, both of which were realized outdoors. A significantly larger P300 event-related potential amplitude was evident for targets compared to distractors (pbrain-computer interface (BCI) study. This leads us to conclude that a truly mobile auditory BCI system is feasible. © 2013.
Piristine, Hande C; Choetso, Tenzin; Gobes, Sharon M H
Sensory feedback is essential for acquiring and maintaining complex motor behaviors, including birdsong. In zebra finches, auditory feedback reaches the song control circuits primarily through the nucleus interfacialis nidopalii (Nif), which provides excitatory input to HVC (proper name)-a premotor region essential for the production of learned vocalizations. Despite being one of the major inputs to the song control pathway, the role of Nif in generating vocalizations is not well understood. To address this, we transiently inactivated Nif in late juvenile zebra finches. Upon Nif inactivation (in both hemispheres or on one side only), birds went from singing stereotyped zebra finch song to uttering highly variable and unstructured vocalizations resembling sub-song, an early juvenile song form driven by a basal ganglia circuit. Simultaneously inactivating Nif and LMAN (lateral magnocellular nucleus of the anterior nidopallium), the output nucleus of a basal ganglia circuit, inhibited song production altogether. These results suggest that Nif is required for generating the premotor drive for song. Permanent Nif lesions, in contrast, have only transient effects on vocal production, with song recovering within a day. The sensorimotor nucleus Nif thus produces a premotor drive to the motor pathway that is acutely required for generating learned vocalizations, but once permanently removed, the song system can compensate for its absence. © 2016 Wiley Periodicals, Inc. Develop Neurobiol 76: 1213-1225, 2016. © 2016 Wiley Periodicals, Inc.
N. Jeremy Hill
Full Text Available Most brain-computer interface (BCI systems require users to modulate brain signals in response to visual stimuli. Thus, they may not be useful to people with limited vision, such as those with severe paralysis. One important approach for overcoming this issue is auditory streaming, an approach whereby a BCI system is driven by shifts of attention between two dichotically presented auditory stimulus streams. Motivated by the long-term goal of translating such a system into a reliable, simple yes-no interface for clinical usage, we aim to answer two main questions. First, we asked which of two previously-published variants provides superior performance: a fixed-phase (FP design in which the streams have equal period and opposite phase, or a drifting-phase (DP design where the periods are unequal. We found FP to be superior to DP (p = 0.002: average performance levels were 80% and 72% correct, respectively. We were also able to show, in a pilot with one subject, that auditory streaming can support continuous control and neurofeedback applications: by shifting attention between ongoing left and right auditory streams, the subject was able to control the position of a paddle in a computer game. Second, we examined whether the system is dependent on eye movements, since it is known that eye movements and auditory attention may influence each other, and any dependence on the ability to move one’s eyes would be a barrier to translation to paralyzed users. We discovered that, despite instructions, some subjects did make eye movements that were indicative of the direction of attention. However, there was no correlation, across subjects, between the reliability of the eye movement signal and the reliability of the BCI system, indicating that our system was configured to work independently of eye movement. Together, these findings are an encouraging step forward toward BCIs that provide practical communication and control options for the most severely
Hill, N Jeremy; Moinuddin, Aisha; Häuser, Ann-Katrin; Kienzle, Stephan; Schalk, Gerwin
Most brain-computer interface (BCI) systems require users to modulate brain signals in response to visual stimuli. Thus, they may not be useful to people with limited vision, such as those with severe paralysis. One important approach for overcoming this issue is auditory streaming, an approach whereby a BCI system is driven by shifts of attention between two simultaneously presented auditory stimulus streams. Motivated by the long-term goal of translating such a system into a reliable, simple yes-no interface for clinical usage, we aim to answer two main questions. First, we asked which of two previously published variants provides superior performance: a fixed-phase (FP) design in which the streams have equal period and opposite phase, or a drifting-phase (DP) design where the periods are unequal. We found FP to be superior to DP (p = 0.002): average performance levels were 80 and 72% correct, respectively. We were also able to show, in a pilot with one subject, that auditory streaming can support continuous control and neurofeedback applications: by shifting attention between ongoing left and right auditory streams, the subject was able to control the position of a paddle in a computer game. Second, we examined whether the system is dependent on eye movements, since it is known that eye movements and auditory attention may influence each other, and any dependence on the ability to move one's eyes would be a barrier to translation to paralyzed users. We discovered that, despite instructions, some subjects did make eye movements that were indicative of the direction of attention. However, there was no correlation, across subjects, between the reliability of the eye movement signal and the reliability of the BCI system, indicating that our system was configured to work independently of eye movement. Together, these findings are an encouraging step forward toward BCIs that provide practical communication and control options for the most severely paralyzed users.
Jennifer M. Gibson
Full Text Available In recent years, domestic dogs have been recognized for their ability to utilize human communicative gestures in choice tasks, as well as communicate with humans through visual and auditory means. A few dogs have even demonstrated the capacity to learn hundreds to thousands of human words and object labels with extensive training. However less is known about dogs‟ understanding or perception of human vocalizations in the absence of explicit training. This study was conducted to determine what aspects of human scolding vocalizations dogs would be most responsive to when presented with a choice to consume or avoid available food items. Variables included the gender, authenticity, word clarity and the human quality of the vocal commands. Our results suggest that dogs are generally cautious about novel sounds produced in the proximity of food. However they are most likely to avoid consumption when hearing a vocalization originally produced by a scolding human, suggesting awareness of vocal qualities common to human speech.
Rosen, Clark A.; Mau, Ted; Remacle, Marc; Hess, Markus; Eckel, Hans E.; Young, VyVy N.; Hantzakos, Anastasios; Yung, Katherine C.; Dikkers, Frederik G.
The terms used to describe vocal fold motion impairment are confusing and not standardized. This results in a failure to communicate accurately and to major limitations of interpreting research studies involving vocal fold impairment. We propose standard nomenclature for reporting vocal fold
Rosen, Clark A.; Mau, Ted; Remacle, Marc; Hess, Markus; Eckel, Hans E.; Young, VyVy N.; Hantzakos, Anastasios; Yung, Katherine C.; Dikkers, Frederik G.
The terms used to describe vocal fold motion impairment are confusing and not standardized. This results in a failure to communicate accurately and to major limitations of interpreting research studies involving vocal fold impairment. We propose standard nomenclature for reporting vocal fold
Full Text Available Darwin (1872 postulated that emotional expressions contain universals that are retained across species. We recently showed that human rating responses were strongly affected by a listener's familiarity with vocalization types, whereas evidence for universal cross-taxa emotion recognition was limited. To disentangle the impact of evolutionarily retained mechanisms (phylogeny and experience-driven cognitive processes (familiarity, we compared the temporal unfolding of event-related potentials (ERPs in response to agonistic and affiliative vocalizations expressed by humans and three animal species. Using an auditory oddball novelty paradigm, ERPs were recorded in response to task-irrelevant novel sounds, comprising vocalizations varying in their degree of phylogenetic relationship and familiarity to humans. Vocalizations were recorded in affiliative and agonistic contexts. Offline, participants rated the vocalizations for valence, arousal, and familiarity. Correlation analyses revealed a significant correlation between a posteriorly distributed early negativity and arousal ratings. More specifically, a contextual category effect of this negativity was observed for human infant and chimpanzee vocalizations but absent for other species vocalizations. Further, a significant correlation between the later and more posteriorly P3a and P3b responses and familiarity ratings indicates a link between familiarity and attentional processing. A contextual category effect of the P3b was observed for the less familiar chimpanzee and tree shrew vocalizations. Taken together, these findings suggest that early negative ERP responses to agonistic and affiliative vocalizations may be influenced by evolutionary retained mechanisms, whereas the later orienting of attention (positive ERPs may mainly be modulated by the prior experience.
Full Text Available In anurans reproductive behavior is strongly seasonal. During the spring, frogs emerge from hibernation and males vocalize for mating or advertising territories. Female frogs have the ability to evaluate the quality of the males' resources on the basis of these vocalizations. Although studies revealed that central single torus semicircularis neurons in frogs exhibit season plasticity, the plasticity of peripheral auditory sensitivity in frog is unknown. In this study the seasonally plasticity of peripheral auditory sensitivity was test in the Emei music frog Babina daunchina, by comparing thresholds and latencies of auditory brainstem responses (ABRs evoked by tone pips and clicks in the reproductive and non-reproductive seasons. The results show that both ABR thresholds and latency differ significantly between the reproductive and non-reproductive seasons. The thresholds of tone pip evoked ABRs in the non-reproductive season increased significantly about 10 dB than those in the reproductive season for frequencies from 1 KHz to 6 KHz. ABR latencies to waveform valley values for tone pips for the same frequencies using appropriate threshold stimulus levels are longer than those in the reproductive season for frequencies from 1.5 to 6 KHz range, although from 0.2 to 1.5 KHz range it is shorter in the non-reproductive season. These results demonstrated that peripheral auditory frequency sensitivity exhibits seasonal plasticity changes which may be adaptive to seasonal reproductive behavior in frogs.
Achey, Meredith A; He, Mike Z; Akst, Lee M
This study sought to assess classical singing students' compliance with vocal hygiene practices identified in the literature and to explore the relationship between self-reported vocal hygiene practice and self-reported singing voice handicap in this population. The primary hypothesis was that increased attention to commonly recommended vocal hygiene practices would correlate with reduced singing voice handicap. This is a cross-sectional, survey-based study. An anonymous survey assessing demographics, attention to 11 common vocal hygiene recommendations in both performance and nonperformance periods, and the Singing Voice Handicap Index 10 (SVHI-10) was distributed to classical singing teachers to be administered to their students at two major schools of music. Of the 215 surveys distributed, 108 were returned (50.2%), of which 4 were incomplete and discarded from analysis. Conservatory students of classical singing reported a moderate degree of vocal handicap (mean SVHI-10, 12; range, 0-29). Singers reported considering all 11 vocal hygiene factors more frequently when preparing for performances than when not preparing for performances. Of these, significant correlations with increased handicap were identified for consideration of stress reduction in nonperformance (P = 0.01) and performance periods (P = 0.02) and with decreased handicap for consideration of singing voice use in performance periods alone (P = 0.02). Conservatory students of classical singing report more assiduous attention to vocal hygiene practices when preparing for performances and report moderate degrees of vocal handicap overall. These students may have elevated risk for dysphonia and voice disorders which is not effectively addressed through common vocal hygiene recommendations alone. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Marquezin, Daniela Maria Santos Serrano; Viola, Izabel; Ghirardi, Ana Carolina de Assis Moura; Madureira, Sandra; Ferreira, Léslie Piccolotto
To analyze speech expressiveness in a group of executives based on perceptive and acoustic aspects of vocal dynamics. Four male subjects participated in the research study (S1, S2, S3, and S4). The assessments included the Kingdomality test to obtain the keywords of communicative attitudes; perceptive-auditory assessment to characterize vocal quality and dynamics, performed by three judges who are speech language pathologists; perceptiveauditory assessment to judge the chosen keywords; speech acoustics to assess prosodic elements (Praat software); and a statistical analysis. According to the perceptive-auditory analysis of vocal dynamics, S1, S2, S3, and S4 did not show vocal alterations and all of them were considered with lowered habitual pitch. S1: pointed out as insecure, nonobjective, nonempathetic, and unconvincing with inappropriate use of pauses that are mainly formed by hesitations; inadequate separation of prosodic groups with breaking of syntagmatic constituents. S2: regular use of pauses for respiratory reload, organization of sentences, and emphasis, which is considered secure, little objective, empathetic, and convincing. S3: pointed out as secure, objective, empathetic, and convincing with regular use of pauses for respiratory reload and organization of sentences and hesitations. S4: the most secure, objective, empathetic, and convincing, with proper use of pauses for respiratory reload, planning, and emphasis; prosodic groups agreed with the statement, without separating the syntagmatic constituents. The speech characteristics and communicative attitudes were highlighted in two subjects in a different manner, in such a way that the slow rate of speech and breaks of the prosodic groups transmitted insecurity, little objectivity, and nonpersuasion.
Chirathivat, Napim; Raja, Sahitya C; Gobes, Sharon M H
Many aspects of song learning in songbirds resemble characteristics of speech acquisition in humans. Genetic, anatomical and behavioural parallels have most recently been extended with demonstrated similarities in hemispheric dominance between humans and songbirds: the avian higher order auditory cortex is left-lateralized for processing song memories in juvenile zebra finches that already have formed a memory of their fathers' song, just like Wernicke's area in the left hemisphere of the human brain is dominant for speech perception. However, it is unclear if hemispheric specialization is due to pre-existing functional asymmetry or the result of learning itself. Here we show that in juvenile male and female zebra finches that had never heard an adult song before, neuronal activation after initial exposure to a conspecific song is bilateral. Thus, like in humans, hemispheric dominance develops with vocal proficiency. A left-lateralized functional system that develops through auditory-vocal learning may be an evolutionary adaptation that could increase the efficiency of transferring information within one hemisphere, benefiting the production and perception of learned communication signals.
Bárbara Gabriela Silva
Full Text Available ABSTRACT OBJECTIVE To measure the risk of dysphonia in teachers, as well as investigate whether the perceptual-auditory and acoustic aspects of the voice of teachers in situations of silence and noise, the signal-to-noise ratio, and the noise levels in the classroom are associated with the presence of dysphonia. METHODS This is an observational cross-sectional research with 23 primary and secondary school teachers from a private school in the municipality of São Paulo, Brazil, divided into the groups without dysphonia and with dysphonia. We performed the following procedures: general Dysphonia Risk Screening Protocol (General-DRSP and complementary to speaking voice - teacher (Specific-DRSP, voice recording during class and in an individual situation in a silent room, and measurement of the signal-to-noise ratio and noise levels of classrooms. RESULTS We have found differences between groups regarding physical activity (General-DRSP and particularities of the profession (Specific-DRSP, as well as in all aspects of the perceptual-auditory vocal analysis. We have found signs of voice wear in the group without dysphonia. Regarding the vocal resources in the situations of noise and silence, we have identified a difference for the production of abrupt vocal attack and the tendency of a more precise speech in the situation of noise. Both the signal-to-noise ratio and the room noise levels during class were high in both groups. CONCLUSIONS Teachers in both groups are at high risk for developing dysphonia and have negative vocal signals to a greater or lesser extent. Signal-to-noise ratio was inadequate in most classrooms, considering the standards for both children with normal hearing and with hearing loss, as well as equivalent noise levels.
Charles R Larson
Full Text Available The pitch-shift paradigm has become a widely used method for studying the role of voice pitch auditory feedback in voice control. This paradigm introduces small, brief pitch shifts in voice auditory feedback to vocalizing subjects. The perturbations trigger a reflexive mechanism that counteracts the change in pitch. The underlying mechanisms of the vocal responses are thought to reflect a negative feedback control system that is similar to constructs developed to explain other forms of motor control. Another use of this technique requires subjects to voluntarily change the pitch of their voice when they hear a pitch shift stimulus. Under these conditions, short latency responses are produced that change voice pitch to match that of the stimulus. The pitch-shift technique has been used with magnetoencephalography (MEG and electroencephalography (EEG recordings, and has shown that at vocal onset there is normally a suppression of neural activity related to vocalization. However, if a pitch-shift is also presented at voice onset, there is a cancellation of this suppression, which has been interpreted to mean that one way in which a person distinguishes self-vocalization from vocalization of others is by a comparison of the intended voice and the actual voice. Studies of the pitch shift reflex in the fMRI environment show that the superior temporal gyrus (STG plays an important role in the process of controlling voice F0 based on auditory feedback. Additional studies using fMRI for effective connectivity modeling show that the left and right STG play critical roles in correcting for an error in voice production. While both the left and right STG are involved in this process, a feedback loop develops between left and right STG during perturbations, in which the left to right connection becomes stronger, and a new negative right to left connection emerges along with the emergence of other feedback loops within the cortical network tested.
Full Text Available Gaze-independent brain computer interfaces (BCIs are a potential communication tool for persons with paralysis. This study applies affective auditory stimuli to investigate their effects using a P300 BCI. Fifteen able-bodied participants operated the P300 BCI, with positive and negative affective sounds (PA: a meowing cat sound, NA: a screaming cat sound. Permuted stimuli of the positive and negative affective sounds (permuted-PA, permuted-NA were also used for comparison. Electroencephalography data was collected, and offline classification accuracies were compared. We used a visual analog scale (VAS to measure positive and negative affective feelings in the participants. The mean classification accuracies were 84.7% for PA and 67.3% for permuted-PA, while the VAS scores were 58.5 for PA and −12.1 for permuted-PA. The positive affective stimulus showed significantly higher accuracy and VAS scores than the negative affective stimulus. In contrast, mean classification accuracies were 77.3% for NA and 76.0% for permuted-NA, while the VAS scores were −50.0 for NA and −39.2 for permuted NA, which are not significantly different. We determined that a positive affective stimulus with accompanying positive affective feelings significantly improved BCI accuracy. Additionally, an ALS patient achieved 90% online classification accuracy. These results suggest that affective stimuli may be useful for preparing a practical auditory BCI system for patients with disabilities.
Bhandiwad, Ashwin A; Whitchurch, Elizabeth A; Colleye, Orphal; Zeddies, David G; Sisneros, Joseph A
Adult female and nesting (type I) male midshipman fish (Porichthys notatus) exhibit an adaptive form of auditory plasticity for the enhanced detection of social acoustic signals. Whether this adaptive plasticity also occurs in "sneaker" type II males is unknown. Here, we characterize auditory-evoked potentials recorded from hair cells in the saccule of reproductive and non-reproductive "sneaker" type II male midshipman to determine whether this sexual phenotype exhibits seasonal, reproductive state-dependent changes in auditory sensitivity and frequency response to behaviorally relevant auditory stimuli. Saccular potentials were recorded from the middle and caudal region of the saccule while sound was presented via an underwater speaker. Our results indicate saccular hair cells from reproductive type II males had thresholds based on measures of sound pressure and acceleration (re. 1 µPa and 1 ms -2 , respectively) that were ~8-21 dB lower than non-reproductive type II males across a broad range of frequencies, which include the dominant higher frequencies in type I male vocalizations. This increase in type II auditory sensitivity may potentially facilitate eavesdropping by sneaker males and their assessment of vocal type I males for the selection of cuckoldry sites during the breeding season.
Howard, David M
The advent and now increasingly widespread availability of 3-D printers is transforming our understanding of the natural world by enabling observations to be made in a tangible manner. This paper describes the use of 3-D printed models of the vocal tract for different vowels that are used to create an acoustic output when stimulated with an appropriate sound source in a new musical instrument: the Vocal Tract Organ. The shape of each printed vocal tract is recovered from magnetic resonance imaging. It sits atop a loudspeaker to which is provided an acoustic L-F model larynx input signal that is controlled by the notes played on a musical instrument digital interface device such as a keyboard. The larynx input is subject to vibrato with extent and frequency adjustable as desired within the ranges usually found for human singing. Polyphonic inputs for choral singing textures can be applied via a single loudspeaker and vocal tract, invoking the approximation of linearity in the voice production system, thereby making multiple vowel stops a possibility while keeping the complexity of the instrument in reasonable check. The Vocal Tract Organ offers a much more human and natural sounding result than the traditional Vox Humana stops found in larger pipe organs, offering the possibility of enhancing pipe organs of the future as well as becoming the basis for a "multi-vowel" chamber organ in its own right. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Sarah A MacLean
Full Text Available The threat sensitivity hypothesis predicts that organisms will evaluate the relative danger of and respond differentially to varying degrees of predation threat. Doing so allows potential prey to balance the costs and benefits of anti-predator behaviors. Threat sensitivity has undergone limited testing in the auditory modality, and the relative threat level of auditory cues from different sources is difficult to infer across populations when variables such as background risk and experience are not properly controlled. We experimentally exposed a single population of two sympatric gull species to auditory stimuli representing a range of potential threats in order to compare the relative threat of heterospecific alarm calls, conspecific alarms calls, predator vocalizations, and novel auditory cues. Gulls were able to discriminate among a diverse set of threat indicators and respond in a graded manner commensurate with the level of threat. Vocalizations of two potential predators, the human voice and bald eagle call, differed in their threat level compared to each other and to alarm calls. Conspecific alarm calls were more threatening than heterospecfic alarm calls to the larger great black-backed gull, but the smaller herring gull weighed both equally. A novel cue elicited a response intermediate between known threats and a known non-threat in herring gulls, but not great black-backed gulls. Our results show that the relative threat level of auditory cues from different sources is highly species-dependent, and that caution should be exercised when comparing graded and threshold threat sensitive responses.
Zheng, Zane Z.; Vicente-Grabovetsky, Alejandro; MacDonald, Ewen N.
The everyday act of speaking involves the complex processes of speech motor control. An important component of control is monitoring, detection, and processing of errors when auditory feedback does not correspond to the intended motor gesture. Here we show, using fMRI and converging operations...... within a multivoxel pattern analysis framework, that this sensorimotor process is supported by functionally differentiated brain networks. During scanning, a real-time speech-tracking system was used to deliver two acoustically different types of distorted auditory feedback or unaltered feedback while...... human participants were vocalizing monosyllabic words, and to present the same auditory stimuli while participants were passively listening. Whole-brain analysis of neural-pattern similarity revealed three functional networks that were differentially sensitive to distorted auditory feedback during...
Petersen, Christopher L; Timothy, Miky; Kim, D Spencer; Bhandiwad, Ashwin A; Mohr, Robert A; Sisneros, Joseph A; Forlano, Paul M
While the neural circuitry and physiology of the auditory system is well studied among vertebrates, far less is known about how the auditory system interacts with other neural substrates to mediate behavioral responses to social acoustic signals. One species that has been the subject of intensive neuroethological investigation with regard to the production and perception of social acoustic signals is the plainfin midshipman fish, Porichthys notatus, in part because acoustic communication is essential to their reproductive behavior. Nesting male midshipman vocally court females by producing a long duration advertisement call. Females localize males by their advertisement call, spawn and deposit all their eggs in their mate's nest. As multiple courting males establish nests in close proximity to one another, the perception of another male's call may modulate individual calling behavior in competition for females. We tested the hypothesis that nesting males exposed to advertisement calls of other males would show elevated neural activity in auditory and vocal-acoustic brain centers as well as differential activation of catecholaminergic neurons compared to males exposed only to ambient noise. Experimental brains were then double labeled by immunofluorescence (-ir) for tyrosine hydroxylase (TH), an enzyme necessary for catecholamine synthesis, and cFos, an immediate-early gene product used as a marker for neural activation. Males exposed to other advertisement calls showed a significantly greater percentage of TH-ir cells colocalized with cFos-ir in the noradrenergic locus coeruleus and the dopaminergic periventricular posterior tuberculum, as well as increased numbers of cFos-ir neurons in several levels of the auditory and vocal-acoustic pathway. Increased activation of catecholaminergic neurons may serve to coordinate appropriate behavioral responses to male competitors. Additionally, these results implicate a role for specific catecholaminergic neuronal groups in
Ghika-Schmid, F; Ghika, J; Vuilleumier, P; Assal, G; Vuadens, P; Scherer, K; Maeder, P; Uske, A; Bogousslavsky, J
A right-handed man developed a sudden transient, amnestic syndrome associated with bilateral hemorrhage of the hippocampi, probably due to Urbach-Wiethe disease. In the 3rd month, despite significant hippocampal structural damage on imaging, only a milder degree of retrograde and anterograde amnesia persisted on detailed neuropsychological examination. On systematic testing of recognition of facial and vocal expression of emotion, we found an impairment of the vocal perception of fear, but not that of other emotions, such as joy, sadness and anger. Such selective impairment of fear perception was not present in the recognition of facial expression of emotion. Thus emotional perception varies according to the different aspects of emotions and the different modality of presentation (faces versus voices). This is consistent with the idea that there may be multiple emotion systems. The study of emotional perception in this unique case of bilateral involvement of hippocampus suggests that this structure may play a critical role in the recognition of fear in vocal expression, possibly dissociated from that of other emotions and from that of fear in facial expression. In regard of recent data suggesting that the amygdala is playing a role in the recognition of fear in the auditory as well as in the visual modality this could suggest that the hippocampus may be part of the auditory pathway of fear recognition.
Tarciso A F Velho
Full Text Available Norepinephrine (NE is thought to play important roles in the consolidation and retrieval of long-term memories, but its role in the processing and memorization of complex acoustic signals used for vocal communication has yet to be determined. We have used a combination of gene expression analysis, electrophysiological recordings and pharmacological manipulations in zebra finches to examine the role of noradrenergic transmission in the brain's response to birdsong, a learned vocal behavior that shares important features with human speech. We show that noradrenergic transmission is required for both the expression of activity-dependent genes and the long-term maintenance of stimulus-specific electrophysiological adaptation that are induced in central auditory neurons by stimulation with birdsong. Specifically, we show that the caudomedial nidopallium (NCM, an area directly involved in the auditory processing and memorization of birdsong, receives strong noradrenergic innervation. Song-responsive neurons in this area express α-adrenergic receptors and are in close proximity to noradrenergic terminals. We further show that local α-adrenergic antagonism interferes with song-induced gene expression, without affecting spontaneous or evoked electrophysiological activity, thus dissociating the molecular and electrophysiological responses to song. Moreover, α-adrenergic antagonism disrupts the maintenance but not the acquisition of the adapted physiological state. We suggest that the noradrenergic system regulates long-term changes in song-responsive neurons by modulating the gene expression response that is associated with the electrophysiological activation triggered by song. We also suggest that this mechanism may be an important contributor to long-term auditory memories of learned vocalizations.
Distúrbio de voz em professores: autorreferência, avaliação perceptiva da voz e das pregas vocais Voice disorders in teachers: self-report, auditory-perceptive assessment of voice and vocal fold assessment
Maria Fabiana Bonfim de Lima-Silva
Full Text Available OBJETIVO: Analisar a presença do distúrbio de voz em professores na concordância entre autorreferência, avaliação perceptiva da voz e das pregas vocais. MÉTODOS: Deste estudo transversal, participaram 60 professores de duas escolas públicas de ensino fundamental e médio. Após responderem questionário de autopercepção (Condição de Produção Vocal do Professor - CPV-P para caracterização da amostra e levantamento de dados sobre autorreferência ao distúrbio de voz, foram submetidos à coleta de amostra de fala e exame nasofibrolaringoscópico. Para classificar as vozes, três juízes fonoaudiólogos utilizaram à escala GRBASI e, para pregas vocais (PPVV, um otorrinolaringologista descreveu as alterações encontradas. Os dados foram analisados descritivamente, e a seguir submetidos a testes de associação. RESULTADOS: No questionário, 63,3% dos participantes referiram ter ou ter tido distúrbio de voz. Do total, 43,3% foram diagnosticados com alteração em voz e 46,7%, em prega vocal. Não houve associação entre autorreferência e avaliação da voz, nem entre autorreferência e avaliação de PPVV, com registro de concordância baixa entre as três avaliações. Porém, houve associação entre a avaliação da voz e de PPVV, com concordância intermediária entre elas. CONCLUSÃO: Há maior autorreferência a distúrbio de voz do que o constatado pela avaliação perceptiva da voz e das pregas vocais. A concordância intermediária entre as duas avaliações prediz a necessidade da realização de pelo menos uma delas por ocasião da triagem em professores.PURPOSE: To analyze the presence of voice disorders in teachers in agreement between self-report, auditory-perceptive assessment of voice quality and vocal fold assessment. METHODS: The subjects of this cross-sectional study were 60 public elementary, middle and high-school teachers. After answering a self-awareness questionnaire (Voice Production Conditions of
Crowell, Sara E.; Berlin, Alicia; Carr, Catherine E.; Olsen, Glenn H.; Therrien, Ronald E.; Yannuzzi, Sally E.; Ketten, Darlene R.
There is little biological data available for diving birds because many live in hard-to-study, remote habitats. Only one species of diving bird, the black-footed penguin (Spheniscus demersus), has been studied in respect to auditory capabilities (Wever et al., Proc Natl Acad Sci USA 63:676–680, 1969). We, therefore, measured in-air auditory threshold in ten species of diving birds, using the auditory brainstem response (ABR). The average audiogram obtained for each species followed the U-shape typical of birds and many other animals. All species tested shared a common region of the greatest sensitivity, from 1000 to 3000 Hz, although audiograms differed significantly across species. Thresholds of all duck species tested were more similar to each other than to the two non-duck species tested. The red-throated loon (Gavia stellata) and northern gannet (Morus bassanus) exhibited the highest thresholds while the lowest thresholds belonged to the duck species, specifically the lesser scaup (Aythya affinis) and ruddy duck (Oxyura jamaicensis). Vocalization parameters were also measured for each species, and showed that with the exception of the common eider (Somateria mollisima), the peak frequency, i.e., frequency at the greatest intensity, of all species' vocalizations measured here fell between 1000 and 3000 Hz, matching the bandwidth of the most sensitive hearing range.
Owren, Michael J; Amoss, R Toby; Rendall, Drew
Vocal communication in nonhuman primates receives considerable research attention, with many investigators arguing for similarities between this calling and speech in humans. Data from development and neural organization show a central role of affect in monkey and ape sounds, however, suggesting that their calls are homologous to spontaneous human emotional vocalizations while having little relation to spoken language. Based on this evidence, we propose two principles that can be useful in evaluating the many and disparate empirical findings that bear on the nature of vocal production in nonhuman and human primates. One principle distinguishes production-first from reception-first vocal development, referring to the markedly different role of auditory-motor experience in each case. The second highlights a phenomenon dubbed dual neural pathways, specifically that when a species with an existing vocal system evolves a new functionally distinct vocalization capability, it occurs through emergence of a second parallel neural pathway rather than through expansion of the extant circuitry. With these principles as a backdrop, we review evidence of acoustic modification of calling associated with background noise, conditioning effects, audience composition, and vocal convergence and divergence in nonhuman primates. Although each kind of evidence has been interpreted to show flexible cognitively mediated control over vocal production, we suggest that most are more consistent with affectively grounded mechanisms. The lone exception is production of simple, novel sounds in great apes, which is argued to reveal at least some degree of volitional vocal control. If also present in early hominins, the cortically based circuitry surmised to be associated with these rudimentary capabilities likely also provided the substrate for later emergence of the neural pathway allowing volitional production in modern humans. © 2010 Wiley-Liss, Inc.
Caffier, Philipp P; Ibrahim Nasr, Ahmed; Ropero Rendon, Maria Del Mar; Wienhausen, Sascha; Forbes, Eleanor; Seidner, Wolfram; Nawka, Tadeus
To multidimensionally investigate common vocal effects in experienced professional nonclassical singers, to examine their mechanism of production and reproducibility, to demonstrate the existence of partial glottal vibration, and to assess the potential of damage to the voice from nonclassical singing. Individual cohort study. Ten male singers aged between 25 and 46 years (34 ± 7 years [mean ± SD]) with different stylistic backgrounds were recruited (five pop/rock/metal, five musical theater). Participants repeatedly presented the usual nonclassical vocal effects and techniques in their repertoire. All performances were documented and analyzed using established instruments (eg, auditory-perceptual assessment, videolaryngostroboscopy, electroglottography, voice function diagnostics). The vocal apparatus of all singers was healthy and capable of high performance. Typical nonclassical vocal effects were breathy voice, creaky voice, vocal fry, grunting, distortion, rattle, belt, and twang. All effects could be easily differentiated from each other. They were intraindividually consistently repeatable and also interindividually produced in a similar manner. A special feature in one singer was the first evidence of partial glottal vibration when belting in the high register. The unintended transition to this reduced voice quality was accompanied by physical fatigue and inflexible respiratory support. The long-lasting use of the investigated nonclassical vocal effects had no negative impact on trained singers. The possibility of long-term damage depends on the individual constitution, specific use, duration, and extent of the hyperfunction. The incidence of partial glottal vibration and its consequences require continuing research to learn more about efficient and healthy vocal function in nonclassical singing. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Full Text Available OBJECTIVE: The aim of this study was to describe the auditory-perceptive evaluation and the psychodynamic aspects of voice samples among suicidal movie characters. METHOD: Voice samples of 48 characters (27 male, 21 female, extracted from 36 movies produced between 1968 and 2006, were analyzed. The samples were evaluated through a specific protocol focusing on the auditory-perceptive evaluation (voice quality, resonance, pitch, loudness, modulation, pauses, articulation and rhythm and the psychodynamic aspects of voice. RESULTS: 85.5% of the samples exhibited abnormal findings in at least five parameters of the auditory-perceptive analysis, such as breathiness (n = 42; 87.5% of the samples, hoarseness (n = 39; 81.2% and strain (n = 29; 60.4%, as well as laryngopharingeal resonance (n = 39; 81.2%, either high pitch (n = 14; 29.2%, or decreased loudness (n = 31; 64.6%. With respect to the psychodynamic aspects, dismay was detected in 50% (n = 24 of the samples, hopelessness in 47.9% (n = 23, resignation in 37.5% (n = 18, and sadness in 33.3% (n = 16. CONCLUSION: Our findings suggest the existence of specific patterns used by actors during the interpretation of suicidal characters. The replication of these findings among real patients may contribute to improvement in the evaluation of potential suicidal patients, as well as the implementation of preventive measures.OBJETIVO: O objetivo do presente estudo foi descrever a análise perceptivo-auditiva e de psicodinâmica vocal de amostras de fala de personagens suicidas em filmes de cinema. MÉTODO: Foram analisadas amostras de fala de 48 personagens suicidas (27 homens, 21 mulheres, extraídas de 36 filmes produzidos no período de 1968 a 2006. As amostras foram analisadas utilizando-se um protocolo especificamente produzido para o registro das características da voz por meio da análise perceptivo-auditiva (qualidade vocal, ressonância, pitch, loudness, modulação, pausas, articulação e ritmo
Joly, Olivier; Orban, Guy A.; Pallier, Christophe; Ramus, Franck; Pressnitzer, Daniel; Vanduffel, Wim
Humans and many other animals use acoustical signals to mediate social interactions with con-specifics. The evolution of sound-based communication is still poorly understood and its neural correlates have only recently begun to be investigated. In the present study, we applied functional MRI to humans and macaque monkeys listening to identical stimuli in order to compare the cortical networks involved in the processing of vocalizations. At the first stages of auditory processing, both species showed similar fMRI activity maps within and around the lateral sulcus (the Sylvian fissure in humans). Monkeys showed remarkably similar responses to monkey calls and to human vocal sounds (speech or otherwise), mainly in the lateral sulcus and the adjacent superior temporal gyrus (STG). In contrast, a preference for human vocalizations and especially for speech was observed in the human STG and superior temporal sulcus (STS). The STS and Broca's region were especially responsive to intelligible utterances. The evolution of the language faculty in humans appears to have recruited most of the STS. It may be that in monkeys, a much simpler repertoire of vocalizations requires less involvement of this temporal territory. (authors)
Full Text Available Gaze-independent brain-computer interfaces (BCIs are a possible communication channel for persons with paralysis. We investigated if it is possible to use auditory stimuli to create a BCI for the Japanese Hiragana syllabary, which has 46 Hiragana characters. Additionally, we investigated if training has an effect on accuracy despite the high amount of different stimuli involved. Able-bodied participants (N=6 were asked to select 25 syllables (out of fifty possible choices using a two step procedure: first the consonant (ten choices and then the vowel (five choices. This was repeated on three separate days. Additionally, a person with spinal cord injury (SCI participated in the experiment. Four out of six healthy participants reached Hiragana syllable accuracies above 70% and the information transfer rate increased from 1.7 bits/min in the first session to 3.2 bits/min in the third session. The accuracy of the participant with SCI increased from 12% (0.2 bits/min to 56% (2 bits/min in session three. Reliable selections from a 10×5 matrix using auditory stimuli were possible and performance is increased by training. We were able to show that auditory P300 BCIs can be used for communication with up to fifty symbols. This enables the use of the technology of auditory P300 BCIs with a variety of applications.
Halder, Sebastian; Takano, Kouji; Ora, Hiroki; Onishi, Akinari; Utsumi, Kota; Kansaku, Kenji
Gaze-independent brain-computer interfaces (BCIs) are a possible communication channel for persons with paralysis. We investigated if it is possible to use auditory stimuli to create a BCI for the Japanese Hiragana syllabary, which has 46 Hiragana characters. Additionally, we investigated if training has an effect on accuracy despite the high amount of different stimuli involved. Able-bodied participants ( N = 6) were asked to select 25 syllables (out of fifty possible choices) using a two step procedure: First the consonant (ten choices) and then the vowel (five choices). This was repeated on 3 separate days. Additionally, a person with spinal cord injury (SCI) participated in the experiment. Four out of six healthy participants reached Hiragana syllable accuracies above 70% and the information transfer rate increased from 1.7 bits/min in the first session to 3.2 bits/min in the third session. The accuracy of the participant with SCI increased from 12% (0.2 bits/min) to 56% (2 bits/min) in session three. Reliable selections from a 10 × 5 matrix using auditory stimuli were possible and performance is increased by training. We were able to show that auditory P300 BCIs can be used for communication with up to fifty symbols. This enables the use of the technology of auditory P300 BCIs with a variety of applications.
Mouterde, Solveig C; Elie, Julie E; Mathevon, Nicolas; Theunissen, Frédéric E
One of the most complex tasks performed by sensory systems is "scene analysis": the interpretation of complex signals as behaviorally relevant objects. The study of this problem, universal to species and sensory modalities, is particularly challenging in audition, where sounds from various sources and localizations, degraded by propagation through the environment, sum to form a single acoustical signal. Here we investigated in a songbird model, the zebra finch, the neural substrate for ranging and identifying a single source. We relied on ecologically and behaviorally relevant stimuli, contact calls, to investigate the neural discrimination of individual vocal signature as well as sound source distance when calls have been degraded through propagation in a natural environment. Performing electrophysiological recordings in anesthetized birds, we found neurons in the auditory forebrain that discriminate individual vocal signatures despite long-range degradation, as well as neurons discriminating propagation distance, with varying degrees of multiplexing between both information types. Moreover, the neural discrimination performance of individual identity was not affected by propagation-induced degradation beyond what was induced by the decreased intensity. For the first time, neurons with distance-invariant identity discrimination properties as well as distance-discriminant neurons are revealed in the avian auditory cortex. Because these neurons were recorded in animals that had prior experience neither with the vocalizers of the stimuli nor with long-range propagation of calls, we suggest that this neural population is part of a general-purpose system for vocalizer discrimination and ranging. SIGNIFICANCE STATEMENT Understanding how the brain makes sense of the multitude of stimuli that it continually receives in natural conditions is a challenge for scientists. Here we provide a new understanding of how the auditory system extracts behaviorally relevant information
Perinatal hypothyroidism causes serious damage to auditory functions that are essential for vocalization development. In rat pups, perinatal hypothyroidism potentially affects the development of ultrasonic vocalization (USV) as a result of hearing deficits. This study examined the effect of perinatal hypothyroidism on the development of USVs in rat pups. Twelve pregnant rats were divided into three groups and treated with the anti-thyroid drug methimazole (MMI) via drinking water, from gestational day 15 to postnatal day (PND) 21. The MMI concentration (w/v) was 0% (control group), 0.01% (low-dose group), or 0.015% (high-dose group). After birth, the pups were individually separated from the dam and littermates on PNDs 5, 10, 15, and 20, and their USVs were recorded for 5min. On PNDs 5 and 10, compared with the control group, the low- and high-dose groups exhibited reductions of both frequency-modulated and downward USVs. On PND 15, however, the low- and high-dose groups displayed increases in number, duration, and amplitude of USVs compared with those in the control group. Lower body weights were observed for the low- and high-dose groups than for the control group. Total thyroxine concentrations in plasma were dose-dependently reduced. The onset of auditory functions appeared on PNDs 11-14. Thus, the rat pups were unable to hear externally produced USVs before PND 11. USVs emitted on PNDs 5 and 10 might have been spontaneous and independent of the pups' own or littermate-emitted USVs. The developmental retardation of vocalization-related organs or muscles might underlie the acoustic alterations of USVs on PNDs 5 and 10. The greater number, duration, and amplitude of USVs on PND 15, after which the hearing onset occurred, suggested that the elevation of auditory thresholds occurred as a result of hearing deficits in the low- and high-dose groups. Perinatal hypothyroidism appears to have caused acoustic alterations in the USV development. Copyright © 2016 Elsevier
Järvinen, Anna; Ng, Rowena; Crivelli, Davide; Neumann, Dirk; Arnold, Andrew J; Woo-VonHoogenstyn, Nicholas; Lai, Philip; Trauner, Doris; Bellugi, Ursula
Both Williams syndrome (WS) and autism spectrum disorders (ASD) are associated with unusual auditory phenotypes with respect to processing vocal and musical stimuli, which may be shaped by the atypical social profiles that characterize the syndromes. Autonomic nervous system (ANS) reactivity to vocal and musical emotional stimuli was examined in 12 children with WS, 17 children with ASD, and 20 typically developing (TD) children, and related to their level of social functioning. The results of this small-scale study showed that after controlling for between-group differences in cognitive ability, all groups showed similar emotion identification performance across conditions. Additionally, in ASD, lower autonomic reactivity to human voice, and in TD, to musical emotion, was related to more normal social functioning. Compared to TD, both clinical groups showed increased arousal to vocalizations. A further result highlighted uniquely increased arousal to music in WS, contrasted with a decrease in arousal in ASD and TD. The ASD and WS groups exhibited arousal patterns suggestive of diminished habituation to the auditory stimuli. The results are discussed in the context of the clinical presentation of WS and ASD. © 2015 Wiley Periodicals, Inc.
Chang, Soo-Eun; Kenney, Mary Kay; Loucks, Torrey M J; Poletto, Christopher J; Ludlow, Christy L
The issue of whether speech is supported by the same neural substrates as non-speech vocal tract gestures has been contentious. In this fMRI study we tested whether producing non-speech vocal tract gestures in humans shares the same functional neuroanatomy as non-sense speech syllables. Production of non-speech vocal tract gestures, devoid of phonological content but similar to speech in that they had familiar acoustic and somatosensory targets, was compared to the production of speech syllables without meaning. Brain activation related to overt production was captured with BOLD fMRI using a sparse sampling design for both conditions. Speech and non-speech were compared using voxel-wise whole brain analyses, and ROI analyses focused on frontal and temporoparietal structures previously reported to support speech production. Results showed substantial activation overlap between speech and non-speech function in regions. Although non-speech gesture production showed greater extent and amplitude of activation in the regions examined, both speech and non-speech showed comparable left laterality in activation for both target perception and production. These findings posit a more general role of the previously proposed "auditory dorsal stream" in the left hemisphere--to support the production of vocal tract gestures that are not limited to speech processing.
Bruna Ferreira Valenzuela de Oliveira
Full Text Available OBJETIVO: Analisar parâmetros perceptivo-auditivos e acústicos da voz em indivíduos adultos gagos. MÉTODOS: Foram analisados 15 indivíduos gagos do gênero masculino na faixa etária de 21 a 41 anos (média 26,6 anos, atendidos no Centro Clínico de Fonoaudiologia da instituição no período de fevereiro de 2005 a julho de 2007. Os parâmetros perceptivo-auditivos analisados envolveram a qualidade vocal, tipo de voz, ressonância, tensão vocal, velocidade de fala, coordenação pneumofônica, ataque vocal e gama tonal; quanto aos parâmetros acústicos, foram analisadas a frequência fundamental e sua variabilidade durante a fala espontânea. RESULTADOS: A análise perceptivo-auditiva mostrou que as características mais frequentes nos indivíduos gagos foram: qualidade vocal normal (60%, ressonância alterada (66%, tensão vocal (86%, ataque vocal alterado (73%, velocidade de fala normal (54%, gama tonal alterada (80% e coordenação pneumofônica alterada (100%. No entanto, a análise estatística revelou que apenas a presença de tensão vocal, coordenação pneumofônica e a gama tonal alteradas apresentaram-se estatisticamente significativas nos indivíduos gagos estudados. Na análise acústica, a frequência fundamental variou de 125,54 a 149,59 Hz e a variabilidade da frequência fundamental foi de 16 a 21 semitons ou 112,50 a 172,40 Hz. CONCLUSÃO: Os parâmetros perceptivo-auditivos analisados que tiveram frequência significativa nos indivíduos gagos estudados foram: presença de tensão vocal, alteração da gama tonal e na coordenação pneumofônica. Desta forma, é importante avaliar os aspectos vocais nesses pacientes, pois a desordem da fluência pode comprometer alguns parâmetros vocais podendo ocasionar disfonia.PURPOSE: To analyze auditory-perceptual and acoustic parameters of the voices of adult stutterers. METHODS: Fifteen male stutterers in the age range from 21 to 41 years (mean 26.6 years, attended at the
Gabriël J L Beckers
Full Text Available BACKGROUND: Many situations involving animal communication are dominated by recurring, stereotyped signals. How do receivers optimally distinguish between frequently recurring signals and novel ones? Cortical auditory systems are known to be pre-attentively sensitive to short-term delivery statistics of artificial stimuli, but it is unknown if this phenomenon extends to the level of behaviorally relevant delivery patterns, such as those used during communication. METHODOLOGY/PRINCIPAL FINDINGS: We recorded and analyzed complete auditory scenes of spontaneously communicating zebra finch (Taeniopygia guttata pairs over a week-long period, and show that they can produce tens of thousands of short-range contact calls per day. Individual calls recur at time scales (median interval 1.5 s matching those at which mammalian sensory systems are sensitive to recent stimulus history. Next, we presented to anesthetized birds sequences of frequently recurring calls interspersed with rare ones, and recorded, in parallel, action and local field potential responses in the medio-caudal auditory forebrain at 32 unique sites. Variation in call recurrence rate over natural ranges leads to widespread and significant modulation in strength of neural responses. Such modulation is highly call-specific in secondary auditory areas, but not in the main thalamo-recipient, primary auditory area. CONCLUSIONS/SIGNIFICANCE: Our results support the hypothesis that pre-attentive neural sensitivity to short-term stimulus recurrence is involved in the analysis of auditory scenes at the level of delivery patterns of meaningful sounds. This may enable birds to efficiently and automatically distinguish frequently recurring vocalizations from other events in their auditory scene.
Full Text Available The auditory stimuli provide information about the objects and events around us. They can also carry biologically significant emotional information (such as unseen dangers and conspecific vocalizations, which provides cues for allocation of attention and mental resources. Here, we investigated whether task-irrelevant auditory emotional information can provide cues for orientation of auditory spatial attention. We employed a covert spatial orienting task: the dot-probe task. In each trial, two task irrelevant auditory cues were simultaneously presented at two separate locations (left-right or front-back. Environmental sounds were selected to form emotional vs. neutral, emotional vs. emotional, and neutral vs. neutral cue pairs. The participants’ task was to detect the location of an acoustic target that was presented immediately after the task-irrelevant auditory cues. The target was presented at the same location as one of the auditory cues. The results indicated that participants were significantly faster to locate the target when it replaced the negative cue compared to when it replaced the neutral cue. The positive cues did not produce a clear attentional bias. Further, same valence pairs (emotional-emotional or neutral-neutral did not modulate reaction times due to a lack of spatial attention capture by one cue in the pair. Taken together, the results indicate that negative affect can provide cues for the orientation of spatial attention in the auditory domain.
Pelaez, Martha; Virues-Ortega, Javier; Gewirtz, Jacob L.
Maternal vocal imitation of infant vocalizations is highly prevalent during face-to-face interactions of infants and their caregivers. Although maternal vocal imitation has been associated with later verbal development, its potentially reinforcing effect on infant vocalizations has not been explored experimentally. This study examined the…
Levendoski, Elizabeth Erickson; Leydon, Ciara; Thibeault, Susan L.
Purpose Vocal fold epithelium is composed of layers of individual epithelial cells joined by junctional complexes constituting a unique interface with the external environment. This barrier provides structural stability to the vocal folds and protects underlying connective tissue from injury while being nearly continuously exposed to potentially hazardous insults including environmental or systemic-based irritants such as pollutants and reflux, surgical procedures, and vibratory trauma. Small disruptions in the epithelial barrier may have a large impact on susceptibility to injury and overall vocal health. The purpose of this article is to provide a broad-based review of our current knowledge of the vocal fold epithelial barrier. Methods A comprehensive review of the literature was conducted. Details of the structure of the vocal fold epithelial barrier are presented and evaluated in the context of function in injury and pathology. The importance of the epithelial-associated vocal fold mucus barrier is also introduced. Results/Conclusions Information presented in this review is valuable for clinicians and researchers as it highlights the importance of this understudied portion of the vocal folds to overall vocal health and disease. Prevention and treatment of injury to the epithelial barrier is a significant area awaiting further investigation. PMID:24686981
Full Text Available The functional auditory system extends from the ears to the frontal lobes with successively more complex functions occurring as one ascends the hierarchy of the nervous system. Several areas of the frontal lobe receive afferents from both early and late auditory processing regions within the temporal lobe. Afferents from the early part of the cortical auditory system, the auditory belt cortex, which are presumed to carry information regarding auditory features of sounds, project to only a few prefrontal regions and are most dense in the ventrolateral prefrontal cortex (VLPFC. In contrast, projections from the parabelt and the rostral superior temporal gyrus (STG most likely convey more complex information and target a larger, widespread region of the prefrontal cortex. Neuronal responses reflect these anatomical projections as some prefrontal neurons exhibit responses to features in acoustic stimuli, while other neurons display task-related responses. For example, recording studies in non-human primates indicate that VLPFC is responsive to complex sounds including vocalizations and that VLPFC neurons in area 12/47 respond to sounds with similar acoustic morphology. In contrast, neuronal responses during auditory working memory involve a wider region of the prefrontal cortex. In humans, the frontal lobe is involved in auditory detection, discrimination, and working memory. Past research suggests that dorsal and ventral subregions of the prefrontal cortex process different types of information with dorsal cortex processing spatial/visual information and ventral cortex processing non-spatial/auditory information. While this is apparent in the non-human primate and in some neuroimaging studies, most research in humans indicates that specific task conditions, stimuli or previous experience may bias the recruitment of specific prefrontal regions, suggesting a more flexible role for the frontal lobe during auditory cognition.
Plakke, Bethany; Romanski, Lizabeth M.
The functional auditory system extends from the ears to the frontal lobes with successively more complex functions occurring as one ascends the hierarchy of the nervous system. Several areas of the frontal lobe receive afferents from both early and late auditory processing regions within the temporal lobe. Afferents from the early part of the cortical auditory system, the auditory belt cortex, which are presumed to carry information regarding auditory features of sounds, project to only a few prefrontal regions and are most dense in the ventrolateral prefrontal cortex (VLPFC). In contrast, projections from the parabelt and the rostral superior temporal gyrus (STG) most likely convey more complex information and target a larger, widespread region of the prefrontal cortex. Neuronal responses reflect these anatomical projections as some prefrontal neurons exhibit responses to features in acoustic stimuli, while other neurons display task-related responses. For example, recording studies in non-human primates indicate that VLPFC is responsive to complex sounds including vocalizations and that VLPFC neurons in area 12/47 respond to sounds with similar acoustic morphology. In contrast, neuronal responses during auditory working memory involve a wider region of the prefrontal cortex. In humans, the frontal lobe is involved in auditory detection, discrimination, and working memory. Past research suggests that dorsal and ventral subregions of the prefrontal cortex process different types of information with dorsal cortex processing spatial/visual information and ventral cortex processing non-spatial/auditory information. While this is apparent in the non-human primate and in some neuroimaging studies, most research in humans indicates that specific task conditions, stimuli or previous experience may bias the recruitment of specific prefrontal regions, suggesting a more flexible role for the frontal lobe during auditory cognition. PMID:25100931
Lee, Christopher S; Todd, Neil P McAngus
The world's languages display important differences in their rhythmic organization; most particularly, different languages seem to privilege different phonological units (mora, syllable, or stress foot) as their basic rhythmic unit. There is now considerable evidence that such differences have important consequences for crucial aspects of language acquisition and processing. Several questions remain, however, as to what exactly characterizes the rhythmic differences, how they are manifested at an auditory/acoustic level and how listeners, whether adult native speakers or young infants, process rhythmic information. In this paper it is proposed that the crucial determinant of rhythmic organization is the variability in the auditory prominence of phonetic events. In order to test this auditory prominence hypothesis, an auditory model is run on two multi-language data-sets, the first consisting of matched pairs of English and French sentences, and the second consisting of French, Italian, English and Dutch sentences. The model is based on a theory of the auditory primal sketch, and generates a primitive representation of an acoustic signal (the rhythmogram) which yields a crude segmentation of the speech signal and assigns prominence values to the obtained sequence of events. Its performance is compared with that of several recently proposed phonetic measures of vocalic and consonantal variability.
Songbirds, such as zebra finches, learn their songs from a ‘tutor’ (usually the father), early in life. There are strong parallels between the behavioural, cognitive and neural processes that underlie vocal learning in humans and songbirds. In both cases there is a sensitive period for auditory
Tüzüner, Arzu; Demirci, Sule; Yavanoglu, Ahmet; Kurkcuoglu, Melih; Arslan, Necmi
Reinke edema is one of the common cause of dysphonia middle-aged population, and severe thickening of vocal folds require surgical treatment. Smoking plays a major role on etiology. Vocal fold cysts are also benign lesions and vocal trauma blamed for acquired cysts. We would like to present 3 cases with vocal fold cyst related with Reinke edema. First case had a subepidermal epidermoid cyst with Reinke edema, which could be easily observed before surgery during laryngostroboscopy. Second case had a mucous retention cyst into the edematous Reinke tissue, which was detected during surgical intervention, and third case had a epidermoid cyst that occurred 2 months after before microlaryngeal operation regarding Reinke edema reduction. These 3 cases revealed that surgical management of Reinke edema needs a careful dissection and close follow-up after surgery for presence of vocal fold cysts.
Mullen, Stuart; Dixon, Mark R.; Belisle, Jordan; Stanley, Caleb
The current study sought to evaluate the efficacy of a stimulus equivalence training procedure in establishing auditory-tactile-visual stimulus classes with 2 children with autism and developmental delays. Participants were exposed to vocal-tactile (A-B) and tactile-picture (B-C) conditional discrimination training and were tested for the…
Bele, Irene Velsvik
It is common practice in vocal training to make use of vocal exercise techniques that involve partial occlusion of the vocal tract. Various techniques are used; some of them form an occlusion within the front part of the oral cavity or at the lips. Another vocal exercise technique involves lengthening the vocal tract; for example, the method of phonation into small tubes. This essay presents some studies made on the effects of various vocal training methods that involve an artificially lengthened and constricted vocal tract. The influence of sufficient acoustic impedance on vocal fold vibration and economical voice production is presented.
Jenny Alejandra Gutiérrez Calderón
Full Text Available This paper explores the most important techniques currently used to detect sub-vocal speech in people with cerebral palsy as well as for commercial purposes, (e.g. allow communication in very noisy places. The methodologies presented deal with speech-signal acquisition and processing. Signal detection and analysis methods are described throughout the whole speech process, from signal generation (as neural impulses in the brain to the production sound in the vocal apparatus (located in the throat. Acquisition and processing quality depends on several factors that will be presented in various sections. A brief explanation to the whole voice generation process is provided in the first part of the article. Subsequently, sub-speech signal acquisition and analysis techniques are presented. Finally, a section about the advantages and disadvantages of the various techniques is presented in order to illustrate different implementations in a sub-vocal speech or silent speech detection device. The results from research indicate that Non-audible Murmur Microphone (NAM is one of the choices that offer huge benefits, not only for signal acquisition and processing, but also for future Spanish language phoneme discrimination.
Ron-Angevin, Ricardo; Velasco-Álvarez, Francisco; Fernández-Rodríguez, Álvaro; Díaz-Estrella, Antonio; Blanca-Mena, María José; Vizcaíno-Martín, Francisco Javier
Certain diseases affect brain areas that control the movements of the patients' body, thereby limiting their autonomy and communication capacity. Research in the field of Brain-Computer Interfaces aims to provide patients with an alternative communication channel not based on muscular activity, but on the processing of brain signals. Through these systems, subjects can control external devices such as spellers to communicate, robotic prostheses to restore limb movements, or domotic systems. The present work focus on the non-muscular control of a robotic wheelchair. A proposal to control a wheelchair through a Brain-Computer Interface based on the discrimination of only two mental tasks is presented in this study. The wheelchair displacement is performed with discrete movements. The control signals used are sensorimotor rhythms modulated through a right-hand motor imagery task or mental idle state. The peculiarity of the control system is that it is based on a serial auditory interface that provides the user with four navigation commands. The use of two mental tasks to select commands may facilitate control and reduce error rates compared to other endogenous control systems for wheelchairs. Seventeen subjects initially participated in the study; nine of them completed the three sessions of the proposed protocol. After the first calibration session, seven subjects were discarded due to a low control of their electroencephalographic signals; nine out of ten subjects controlled a virtual wheelchair during the second session; these same nine subjects achieved a medium accuracy level above 0.83 on the real wheelchair control session. The results suggest that more extensive training with the proposed control system can be an effective and safe option that will allow the displacement of a wheelchair in a controlled environment for potential users suffering from some types of motor neuron diseases.
Full Text Available Studies of behavioral and neural responses to distorted auditory feedback can help shed light on the neural mechanisms of animal vocalizations. We describe an apparatus for generating real-time acoustic feedback. The system can very rapidly detect acoustic features in a song and output acoustic signals if the detected features match the desired acoustic template. The system uses spectrogram-based detection of acoustic elements. It is low-cost and can be programmed for a variety of behavioral experiments requiring acoustic feedback or neural stimulation. We use the system to study the effects of acoustic feedback on birds' vocalizations and demonstrate that such an acoustic feedback can cause both immediate and long-term changes to birds’ songs.
Greimel, E; Trinkl, M; Bartling, J; Bakos, S; Grossheinrich, N; Schulte-Körne, G
Major depression (MD) is associated with deficits in selective attention. Previous studies in adults with MD using event-related potentials (ERPs) reported abnormalities in the neurophysiological correlates of auditory selective attention. However, it is yet unclear whether these findings can be generalized to MD in adolescence. Thus, the aim of the present ERP study was to explore the neural mechanisms of auditory selective attention in adolescents with MD. 24 male and female unmedicated adolescents with MD and 21 control subjects were included in the study. ERPs were collected during an auditory oddball paradigm. Depressive adolescents tended to show a longer N100 latency to target and non-target tones. Moreover, MD subjects showed a prolonged latency of the P200 component to targets. Across groups, longer P200 latency was associated with a decreased tendency of disinhibited behavior as assessed by a behavioral questionnaire. To be able to draw more precise conclusions about differences between the neural bases of selective attention in adolescents vs. adults with MD, future studies should include both age groups and apply the same experimental setting across all subjects. The study provides strong support for abnormalities in the neurophysiolgical bases of selective attention in adolecents with MD at early stages of auditory information processing. Absent group differences in later ERP components reflecting voluntary attentional processes stand in contrast to results reported in adults with MD and may suggest that adolescents with MD possess mechanisms to compensate for abnormalities in the early stages of selective attention. Copyright © 2014 Elsevier B.V. All rights reserved.
Full Text Available For Brain-Computer Interface (BCI systems that are designed for users with severe impairments of the oculomotor system, an appropriate mode of presenting stimuli to the user is crucial. To investigate whether multi-sensory integration can be exploited in the gaze-independent event-related potentials (ERP speller and to enhance BCI performance, we designed a visual-auditory speller. We investigate the possibility to enhance stimulus presentation by combining visual and auditory stimuli within gaze-independent spellers. In this study with N = 15 healthy users, two different ways of combining the two sensory modalities are proposed: simultaneous redundant streams (Combined-Speller and interleaved independent streams (Parallel-Speller. Unimodal stimuli were applied as control conditions. The workload, ERP components, classification accuracy and resulting spelling speed were analyzed for each condition. The Combined-speller showed a lower workload than uni-modal paradigms, without the sacrifice of spelling performance. Besides, shorter latencies, lower amplitudes, as well as a shift of the temporal and spatial distribution of discriminative information were observed for Combined-speller. These results are important and are inspirations for future studies to search the reason for these differences. For the more innovative and demanding Parallel-Speller, where the auditory and visual domains are independent from each other, a proof of concept was obtained: fifteen users could spell online with a mean accuracy of 87.7% (chance level <3% showing a competitive average speed of 1.65 symbols per minute. The fact that it requires only one selection period per symbol makes it a good candidate for a fast communication channel. It brings a new insight into the true multisensory stimuli paradigms. Novel approaches for combining two sensory modalities were designed here, which are valuable for the development of ERP-based BCI paradigms.
Miller, S.; Harrison, L.B.; Solomon, B.; Sessions, R.B.
A prospective evaluation of vocal changes in patients receiving radiation therapy for T1 and T2 (AJC) glottic carcinoma was undertaken in January 1987. Vocal analysis was performed prior to radiotherapy and at specific intervals throughout the radiation treatment program. The voicing ratio was extrapolated from a sustained vowel phonation using the Visipitch interfaced with the IBM-PC. Preliminary observations suggested three distinct patterns of vocal behavior: 1. reduced voicing ratio with precipitous improvement within the course of treatment, 2. high initial voicing ratio with reduction secondary to radiation induced edema, with rapid improvement in the voicing component after the edema subsided, and 3. fluctuating voicing ratio during and following treatment. Enrollment of new patients and a 2-year follow-up of current patients was undertaken
Brittan-Powell, Elizabeth F; Christensen-Dalsgaard, Jakob; Tang, Yezhong
Although lizards have highly sensitive ears, it is difficult to condition them to sound, making standard psychophysical assays of hearing sensitivity impractical. This paper describes non-invasive measurements of the auditory brainstem response (ABR) in both Tokay geckos (Gekko gecko; nocturnal...... animals, known for their loud vocalizations) and the green anole (Anolis carolinensis, diurnal, non-vocal animals). Hearing sensitivity was measured in 5 geckos and 7 anoles. The lizards were sedated with isoflurane, and ABRs were measured at levels of 1 and 3% isoflurane. The typical ABR waveform......). Above 5 kHz, however, anoles were more than 20 dB more sensitive than geckos and showed a wider range of sensitivity (1-7 kHz). Generally, thresholds from ABR audiograms were comparable to those of small birds. Best hearing sensitivity, however, extended over a larger frequency range in lizards than...
Beckers, Gabriël J L; Gahr, Manfred
Auditory systems bias responses to sounds that are unexpected on the basis of recent stimulus history, a phenomenon that has been widely studied using sequences of unmodulated tones (mismatch negativity; stimulus-specific adaptation). Such a paradigm, however, does not directly reflect problems that neural systems normally solve for adaptive behavior. We recorded multiunit responses in the caudomedial auditory forebrain of anesthetized zebra finches (Taeniopygia guttata) at 32 sites simultaneously, to contact calls that recur probabilistically at a rate that is used in communication. Neurons in secondary, but not primary, auditory areas respond preferentially to calls when they are unexpected (deviant) compared with the same calls when they are expected (standard). This response bias is predominantly due to sites more often not responding to standard events than to deviant events. When two call stimuli alternate between standard and deviant roles, most sites exhibit a response bias to deviant events of both stimuli. This suggests that biases are not based on a use-dependent decrease in response strength but involve a more complex mechanism that is sensitive to auditory deviance per se. Furthermore, between many secondary sites, responses are tightly synchronized, a phenomenon that is driven by internal neuronal interactions rather than by the timing of stimulus acoustic features. We hypothesize that this deviance-sensitive, internally synchronized network of neurons is involved in the involuntary capturing of attention by unexpected and behaviorally potentially relevant events in natural auditory scenes.
containing questions about self-perception of singing and vocal practices and the protocol Modern Singing Handicap Index (MSHI, composed by 30 questions regarding disability, handicap, and defect. We performed a screening for perceptual classification of adapted or changed voices, and measured the degrees of change. RESULTS: the total average score was 23 points in the MSHI. The highest subscale scores obtained were "defect" (10.9, followed by "disability" (7.6 and "handicap" (4.5, with the difference between them (p = 0.001. Singers who have never passed through singing lesson had higher scores in the "handicap" (p = 0.003. The higher was the score of MSHI, the score given by singers in relation to their own voice sank (p = 0.046. Participants with altered voice quality had higher scores on handicap and disability subscales and on total domination of MSHI when compared with those who have adapted vocal quality (p = 0.012, p = 0.049 and p = 0.015, respectively. Moreover, the greater is the degree of voice alteration, increased scores were related to disability subscale (p = 0.022. CONCLUSION: church singers have major vocal handicap. When you have voice disorders, this handicap is even greater. The higher is the degree of voice alteration, the greater are the limitations regarding the singing voice. Vocal singing lessons seem to minimize the handicap.
Drapeau, Joanie; Gosselin, Nathalie; Peretz, Isabelle; McKerral, Michelle
To assess emotion recognition from dynamic facial, vocal and musical expressions in sub-groups of adults with traumatic brain injuries (TBI) of different severities and identify possible common underlying mechanisms across domains. Forty-one adults participated in this study: 10 with moderate-severe TBI, nine with complicated mild TBI, 11 with uncomplicated mild TBI and 11 healthy controls, who were administered experimental (emotional recognition, valence-arousal) and control tasks (emotional and structural discrimination) for each domain. Recognition of fearful faces was significantly impaired in moderate-severe and in complicated mild TBI sub-groups, as compared to those with uncomplicated mild TBI and controls. Effect sizes were medium-large. Participants with lower GCS scores performed more poorly when recognizing fearful dynamic facial expressions. Emotion recognition from auditory domains was preserved following TBI, irrespective of severity. All groups performed equally on control tasks, indicating no perceptual disorders. Although emotional recognition from vocal and musical expressions was preserved, no correlation was found across auditory domains. This preliminary study may contribute to improving comprehension of emotional recognition following TBI. Future studies of larger samples could usefully include measures of functional impacts of recognition deficits for fearful facial expressions. These could help refine interventions for emotional recognition following a brain injury.
Full Text Available Paresis or paralysis of one or both vocal cords affects some significant aspects of a human life: breathing, swallowing and speech. The major causes for reduced mobility or even immobility are innervation damage, less often fixation of vocal cord or impaired mobility of crycoarytenoid joint. An injury of the superior or/and inferior laryngeal nerve can be a consequence of different medical procedures, tumor growth, trauma, infection, neurological disorders, radiation exposure, toxic damage, impaired circulation of the area or it is idiopathic. The symptoms are different in the case of unilateral and bilateral paresis of the vocal folds. They also depend on the cause for the impaired mobility. In the patients with unilateral vocal fold paresis, hoarseness and aspiration during swallowing are the leading symptoms. In the bilateral vocal fold paralysis, dyspnea prevails.
Full Text Available Representing an intuitive spelling interface for Brain-Computer Interfaces (BCI in the auditory domain is not straightforward. In consequence, all existing approaches based on event-related potentials (ERP rely at least partially on a visual representation of the interface. This online study introduces an auditory spelling interface that eliminates the necessity for such a visualization. In up to two sessions, a group of healthy subjects (N=21 was asked to use a text entry application, utilizing the spatial cues of the AMUSE paradigm (Auditory Multiclass Spatial ERP. The speller relies on the auditory sense both for stimulation and the core feedback. Without prior BCI experience, 76% of the participants were able to write a full sentence during the first session. By exploiting the advantages of a newly introduced dynamic stopping method, a maximum writing speed of 1.41 characters/minute (7.55 bits/minute could be reached during the second session (average: .94 char/min, 5.26 bits/min. For the first time, the presented work shows that an auditory BCI can reach performances similar to state-of-the-art visual BCIs based on covert attention. These results represent an important step towards a purely auditory BCI.
Rosenthal, Laura H Swibel; Benninger, Michael S; Deeb, Robert H
To determine the current etiology of vocal fold immobility, identify changing trends over the last 20 years, and compare results to historical reports. The present study is a retrospective analysis of all patients seen within a tertiary care institution between 1996 and 2005 with vocal fold immobility. The results were combined with a previous study of patients within the same institution from 1985 through 1995. Results were compared to the literature. The medical records of all patients assigned a primary or additional diagnostic code for vocal cord paralysis were obtained from the electronic database. Eight hundred twenty-seven patients were available for analysis (435 from the most recent cohort), which is substantially larger than any reported series to date. Vocal fold immobility was most commonly associated with a surgical procedure (37%). Nonthyroid surgeries (66%), such as anterior cervical approaches to the spine and carotid endarterectomies, have surpassed thyroid surgery (33%) as the most common iatrogenic causes. These data represent a change from historical figures in which extralaryngeal malignancies were considered the major cause of unilateral immobility. Thyroidectomy continues to cause the majority (80%) of iatrogenic bilateral vocal fold immobility and 30% of all bilateral immobility. This 20-year longitudinal assessment revealed that the etiology of unilateral vocal fold immobility has changed such that there has been a shift from extralaryngeal malignancies to nonthyroid surgical procedures as the major cause. Thyroid surgery remains the most common cause of bilateral vocal fold immobility.
Peh, Wendy Y X; Roberts, Todd F; Mooney, Richard
Vocal communication depends on the coordinated activity of sensorimotor neurons important to vocal perception and production. How vocalizations are represented by spatiotemporal activity patterns in these neuronal populations remains poorly understood. Here we combined intracellular recordings and two-photon calcium imaging in anesthetized adult zebra finches (Taeniopygia guttata) to examine how learned birdsong and its component syllables are represented in identified projection neurons (PNs) within HVC, a sensorimotor region important for song perception and production. These experiments show that neighboring HVC PNs can respond at markedly different times to song playback and that different syllables activate spatially intermingled PNs within a local (~100 μm) region of HVC. Moreover, noise correlations were stronger between PNs that responded most strongly to the same syllable and were spatially graded within and between classes of PNs. These findings support a model in which syllabic and temporal features of song are represented by spatially intermingled PNs functionally organized into cell- and syllable-type networks within local spatial scales in HVC. Copyright © 2015 the authors 0270-6474/15/355589-17$15.00/0.
Goutte, Sandra; Mason, Matthew J; Christensen-Dalsgaard, Jakob
discrimination from background noise and heterospecific calls. However, we present here evidence that two anurans, Brachycephalus ephippium and B. pitanga, are insensitive to the sound of their own calls. Both species produce advertisement calls outside their hearing sensitivity range and their inner ears...... by their high toxicity might help to explain why calling has not yet disappeared, and that visual communication may have replaced auditory in these colourful, diurnal frogs....
Full Text Available Perceiving and producing vocal sounds are important functions of the auditory-motor system and are fundamental to communication. Prior studies have identified a network of brain regions involved in pitch production, specifically pitch matching. Here we reverse engineer the function of the auditory perception-production network by targeting specific cortical regions (e.g., right and left posterior superior temporal (pSTG and posterior inferior frontal gyri (pIFG with cathodal transcranial direct current stimulation (tDCS—commonly found to decrease excitability in the underlying cortical region—allowing us to causally test the role of particular nodes in this network. Performance on a pitch-matching task was determined before and after 20 min of cathodal stimulation. Acoustic analyses of pitch productions showed impaired accuracy after cathodal stimulation to the left pIFG and the right pSTG in comparison to sham stimulation. Both regions share particular roles in the feedback and feedforward motor control of pitched vocal production with a differential hemispheric dominance.
Gökcan, Kürşat Mustafa; Dursun, Gürsel
The aim of the study was to present symptoms, laryngological findings, clinical course, management modalities, and consequences of vascular lesions of vocal fold. This study examined 162 patients, the majority professional voice users, with vascular lesions regarding their presenting symptoms, laryngological findings, clinical courses and treatment results. The most common complaint was sudden hoarseness with hemorrhagic polyp. Microlaryngoscopic surgery was performed in 108 cases and the main indication of surgery was the presence of vocal fold mass or development of vocal polyp during clinical course. Cold microsurgery was utilized for removal of vocal fold masses and feeding vessels cauterized using low power, pulsed CO(2) laser. Acoustic analysis of patients revealed a significant improvement of jitter, shimmer and harmonics/noise ratio values after treatment. Depending on our clinical findings, we propose treatment algorithm where voice rest and behavioral therapy is the integral part and indications of surgery are individualized for each patient.
Oba, Sandra I.; Galvin, John J.; Fu, Qian-Jie
Auditory training has been shown to significantly improve cochlear implant (CI) users’ speech and music perception. However, it is unclear whether post-training gains in performance were due to improved auditory perception or to generally improved attention, memory and/or cognitive processing. In this study, speech and music perception, as well as auditory and visual memory were assessed in ten CI users before, during, and after training with a non-auditory task. A visual digit span (VDS) task was used for training, in which subjects recalled sequences of digits presented visually. After the VDS training, VDS performance significantly improved. However, there were no significant improvements for most auditory outcome measures (auditory digit span, phoneme recognition, sentence recognition in noise, digit recognition in noise), except for small (but significant) improvements in vocal emotion recognition and melodic contour identification. Post-training gains were much smaller with the non-auditory VDS training than observed in previous auditory training studies with CI users. The results suggest that post-training gains observed in previous studies were not solely attributable to improved attention or memory, and were more likely due to improved auditory perception. The results also suggest that CI users may require targeted auditory training to improve speech and music perception. PMID:23516087
Rosburg, Timm; Johansson, Mikael; Sprondel, Volker; Mecklinger, Axel
Retrieval orientation refers to a pre-retrieval process and conceptualizes the specific form of processing that is applied to a retrieval cue. In the current event-related potential (ERP) study, we sought to find evidence for an involvement of the auditory cortex when subjects attempt to retrieve vocalized information, and hypothesized that adopting retrieval orientation would be beneficial for retrieval accuracy. During study, participants saw object words that they subsequently vocalized or visually imagined. At test, participants had to identify object names of one study condition as targets and to reject object names of the second condition together with new items. Target category switched after half of the test trials. Behaviorally, participants responded less accurately and more slowly to targets of the vocalize condition than to targets of the imagine condition. ERPs to new items varied at a single left electrode (T7) between 500 and 800ms, indicating a moderate retrieval orientation effect in the subject group as a whole. However, whereas the effect was strongly pronounced in participants with high retrieval accuracy, it was absent in participants with low retrieval accuracy. A current source density (CSD) mapping of the retrieval orientation effect indicated a source over left temporal regions. Independently from retrieval accuracy, the ERP retrieval orientation effect was surprisingly also modulated by test order. Findings are suggestive for an involvement of the auditory cortex in retrieval attempts of vocalized information and confirm that adopting retrieval orientation is potentially beneficial for retrieval accuracy. The effects of test order on retrieval-related processes might reflect a stronger focus on the newness of items in the more difficult test condition when participants started with this condition. Copyright © 2014 Elsevier Inc. All rights reserved.
Seyyedeh Maryam khoddami
Full Text Available Background and Aim: Vocal abuse and misuse are the most frequent causes of voice disorders. Consequently some therapy is needed to stop or modify such behaviors. This research was performed to study the effectiveness of vocal hygiene program on perceptual signs of voice in people with dysphonia.Methods: A Vocal hygiene program was performed to 8 adults with dysphonia for 6 weeks. At first, Consensus Auditory- Perceptual Evaluation of Voice was used to assess perceptual signs. Then the program was delivered, Individuals were followed in second and forth weeks visits. In the last session, perceptual assessment was performed and individuals’ opinions were collected. Perceptual findings were compared before and after the therapy.Results: After the program, mean score of perceptual assessment decreased. Mean score of every perceptual sign revealed significant difference before and after the therapy (p≤0.0001. «Loudness» had maximum score and coordination between speech and respiration indicated minimum score. All participants confirmed efficiency of the therapy.Conclusion: The vocal hygiene program improves all perceptual signs of voice although not equally. This deduction is confirmed by both clinician-based and patient-based assessments. As a result, vocal hygiene program is necessary for a comprehensive voice therapy but is not solely effective to resolve all voice problems.
Kozlov, Andrei S; Gentner, Timothy Q
High-level neurons processing complex, behaviorally relevant signals are sensitive to conjunctions of features. Characterizing the receptive fields of such neurons is difficult with standard statistical tools, however, and the principles governing their organization remain poorly understood. Here, we demonstrate multiple distinct receptive-field features in individual high-level auditory neurons in a songbird, European starling, in response to natural vocal signals (songs). We then show that receptive fields with similar characteristics can be reproduced by an unsupervised neural network trained to represent starling songs with a single learning rule that enforces sparseness and divisive normalization. We conclude that central auditory neurons have composite receptive fields that can arise through a combination of sparseness and normalization in neural circuits. Our results, along with descriptions of random, discontinuous receptive fields in the central olfactory neurons in mammals and insects, suggest general principles of neural computation across sensory systems and animal classes.
Chen, Wenli; Woo, Peak; Murry, Thomas
High-speed videoendoscopy captures the cycle-to-cycle vibratory motion of each individual vocal fold in normal and severely disordered phonation. Therefore, it provides a direct method to examine the specific vibratory changes following vocal fold surgery. The purpose of this study was to examine the vocal fold vibratory pattern changes in the surgically treated pathologic vocal fold and the contralateral vocal fold in three vocal pathologies: vocal polyp (n = 3), paresis or paralysis (n = 3), and scar (n = 3). Digital kymography was used to extract high-speed kymographic vocal fold images at the mid-membranous region of the vocal fold. Spectral analysis was subsequently applied to the digital kymography to quantify the cycle-to-cycle movements of each vocal fold, expressed as a spectrum. Surgical modification resulted in significantly improved spectral power of the treated pathologic vocal fold. Furthermore, the contralateral vocal fold also presented with improved spectral power irrespective of vocal pathology. In comparison with normal vocal fold spectrum, postsurgical vocal fold vibrations continued to demonstrate decreased vibratory amplitude in both vocal folds. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Pembrook, Randall G.
Reports on a study which reinforces prior findings on melodic memory that show a majority of students do not sing accurately enough after only one hearing of a melody to benefit from vocalization memory techniques. Questions whether vocalization can be a memory reinforcer in melodies that are shorter and simpler than those used in this research.…
Oba, Sandra I; Galvin, John J; Fu, Qian-Jie
Auditory training has been shown to significantly improve cochlear implant (CI) users' speech and music perception. However, it is unclear whether posttraining gains in performance were due to improved auditory perception or to generally improved attention, memory, and/or cognitive processing. In this study, speech and music perception, as well as auditory and visual memory, were assessed in 10 CI users before, during, and after training with a nonauditory task. A visual digit span (VDS) task was used for training, in which subjects recalled sequences of digits presented visually. After the VDS training, VDS performance significantly improved. However, there were no significant improvements for most auditory outcome measures (auditory digit span, phoneme recognition, sentence recognition in noise, digit recognition in noise), except for small (but significant) improvements in vocal emotion recognition and melodic contour identification. Posttraining gains were much smaller with the nonauditory VDS training than observed in previous auditory training studies with CI users. The results suggest that posttraining gains observed in previous studies were not solely attributable to improved attention or memory and were more likely due to improved auditory perception. The results also suggest that CI users may require targeted auditory training to improve speech and music perception.
... digital filtering for noise cancellation which interfaces to speech recognition software. It uses auditory features in speech recognition training, and provides applications to multilingual spoken language translation...
Caffier, Philipp P; Salmen, Tatjana; Ermakova, Tatiana; Forbes, Eleanor; Ko, Seo-Rin; Song, Wen; Gross, Manfred; Nawka, Tadeus
There are few data demonstrating the specific extent to which surgical intervention for vocal fold nodules (VFN) improves vocal function in professional (PVU) and non-professional voice users (NVU). The objective of this study was to compare and quantify results after phonomicrosurgery for VFN in these patient groups. In a prospective clinical study, surgery was performed via microlaryngoscopy in 37 female patients with chronic VFN manifestations (38±12 yrs, mean±SD). Pre- and postoperative evaluations of treatment efficacy comprised videolaryngostroboscopy, auditory-perceptual voice assessment, voice range profile (VRP), acoustic-aerodynamic analysis, and voice handicap index (VHI-9i). The dysphonia severity index (DSI) was compared with the vocal extent measure (VEM). PVU (n=24) and NVU (n=13) showed comparable laryngeal findings and levels of suffering (VHI-9i 16±7 vs 17±8), but PVU had a better pretherapeutic vocal range (26.8±7.4 vs 17.7±5.1 semitones, p<0.001) and vocal capacity (VEM 106±18 vs 74±29, p<0.01). Three months postoperatively, all patients had straight vocal fold edges, complete glottal closure, and recovered mucosal wave propagation. The mean VHI-9i score decreased by 8±6 points. DSI increased from 4.0±2.4 to 5.5±2.4, and VEM from 95±27 to 108±23 (p<0.001). Both parameters correlated significantly (rs=0.82). The average vocal range increased by 4.1±5.3 semitones, and the mean speaking pitch lowered by 0.5±1.4 semitones. These results confirm that phonomicrosurgery for VFN is a safe therapy for voice improvement in both PVU and NVU who do not respond to voice therapy alone. Top-level artistic capabilities in PVU were restored, but numeric changes of most vocal parameters were considerably larger in NVU.
Fadel, Congeta Bruniere Xavier; Dassie-Leite, Ana Paula; Santos, Rosane Sampaio; Santos, Celso Gonçalves Dos; Dias, Cláudio Antônio Sorondo; Sartori, Denise Jussara
The purpose of this study was to analyze the immediate effects of the semi-occluded vocal tract exercise (SOVTE) using the LaxVox® tube in singers. Participants were 23 singers, classical singing students, aged 18 to 47 years (mean age = 27.2 years). First, data was collected through the application of a demographic questionnaire and the recording of sustained emission - vowel /ε/, counting 1-10, and a music section from the participants' current repertoire. After that, the participants were instructed and performed the SOVTE using the LaxVox® tube for three minutes. Finally, the same vocal samples were collected immediately after SOVTE performance and the singers responded to a questionnaire on their perception regarding vocal changes after the exercise. The vocal samples were analyzed by referees (speech-language pathologists and singing teachers) and by means of acoustic analysis. Most of the singers reported improved voice post-exercise in both tasks - speech and singing. Regarding the perceptual assessment (sustained vowel, speech, and singing), the referees found no difference between pre- and post-exercise emissions. The acoustic analysis of the sustained vowel showed increased Fundamental Frequency (F0) and reduction of the Glottal to Noise Excitation (GNE) ratio post-exercise. The semi-occluded vocal tract exercise with LaxVox® tube promotes immediate positive effects on the self-assessment and acoustic analysis of voice in professional singers without vocal complains. No immediate significant changes were observed with respect to auditory-perceptual evaluation of speech and singing.
Full Text Available Isolation calls produced by dependent young are a fundamental form of communication. For species in which vocal signals remain important to adult communication, the function and social context of vocal behavior changes dramatically with the onset of sexual maturity. The ontogenetic relationship between these distinct forms of acoustic communication is surprisingly under-studied. We conducted a detailed analysis of vocal development in sister species of Neotropical singing mice, Scotinomys teguina and S. xerampelinus. Adult singing mice are remarkable for their advertisement songs, rapidly articulated trills used in long-distance communication; the vocal behavior of pups was previously undescribed. We recorded 30 S. teguina and 15 S. xerampelinus pups daily, from birth to weaning; 23 S. teguina and 11 S. xerampelinus were recorded until sexual maturity. Like other rodent species with poikilothermic young, singing mice were highly vocal during the first weeks of life and stopped vocalizing before weaning. Production of first advertisement songs coincided with the onset of sexual maturity after a silent period of ≧2 weeks. Species differences in vocal behavior emerged early in ontogeny and notes that comprise adult song were produced from birth. However, the organization and relative abundance of distinct note types was very different between pups and adults. Notably, the structure, note repetition rate, and intra-individual repeatability of pup vocalizations did not become more adult-like with age; the highly stereotyped structure of adult song appeared de novo in the first songs of young adults. We conclude that, while the basic elements of adult song are available from birth, distinct selection pressures during maternal dependency, dispersal, and territorial establishment favor major shifts in the structure and prevalence of acoustic signals. This study provides insight into how an evolutionarily conserved form of acoustic signaling provides
Petkov, Christopher I; Jarvis, Erich D
Vocal learners such as humans and songbirds can learn to produce elaborate patterns of structurally organized vocalizations, whereas many other vertebrates such as non-human primates and most other bird groups either cannot or do so to a very limited degree. To explain the similarities among humans and vocal-learning birds and the differences with other species, various theories have been proposed. One set of theories are motor theories, which underscore the role of the motor system as an evolutionary substrate for vocal production learning. For instance, the motor theory of speech and song perception proposes enhanced auditory perceptual learning of speech in humans and song in birds, which suggests a considerable level of neurobiological specialization. Another, a motor theory of vocal learning origin, proposes that the brain pathways that control the learning and production of song and speech were derived from adjacent motor brain pathways. Another set of theories are cognitive theories, which address the interface between cognition and the auditory-vocal domains to support language learning in humans. Here we critically review the behavioral and neurobiological evidence for parallels and differences between the so-called vocal learners and vocal non-learners in the context of motor and cognitive theories. In doing so, we note that behaviorally vocal-production learning abilities are more distributed than categorical, as are the auditory-learning abilities of animals. We propose testable hypotheses on the extent of the specializations and cross-species correspondences suggested by motor and cognitive theories. We believe that determining how spoken language evolved is likely to become clearer with concerted efforts in testing comparative data from many non-human animal species.
Jason A Miranda
Full Text Available Subcortical auditory nuclei were traditionally viewed as non-plastic in adulthood so that acoustic information could be stably conveyed to higher auditory areas. Studies in a variety of species, including humans, now suggest that prolonged acoustic training can drive long-lasting brainstem plasticity. The neurobiological mechanisms for such changes are not well understood in natural behavioral contexts due to a relative dearth of in vivo animal models in which to study this. Here, we demonstrate in a mouse model that a natural life experience with increased demands on the auditory system - motherhood - is associated with improved temporal processing in the subcortical auditory pathway. We measured the auditory brainstem response to test whether mothers and pup-naïve virgin mice differed in temporal responses to both broadband and tone stimuli, including ultrasonic frequencies found in mouse pup vocalizations. Mothers had shorter latencies for early ABR peaks, indicating plasticity in the auditory nerve and the cochlear nucleus. Shorter interpeak latency between waves IV and V also suggest plasticity in the inferior colliculus. Hormone manipulations revealed that these cannot be explained solely by estrogen levels experienced during pregnancy and parturition in mothers. In contrast, we found that pup-care experience, independent of pregnancy and parturition, contributes to shortening auditory brainstem response latencies. These results suggest that acoustic experience in the maternal context imparts plasticity on early auditory processing that lasts beyond pup weaning. In addition to establishing an animal model for exploring adult auditory brainstem plasticity in a neuroethological context, our results have broader implications for models of perceptual, behavioral and neural changes that arise during maternity, where subcortical sensorineural plasticity has not previously been considered.
Miranda, Jason A; Shepard, Kathryn N; McClintock, Shannon K; Liu, Robert C
Subcortical auditory nuclei were traditionally viewed as non-plastic in adulthood so that acoustic information could be stably conveyed to higher auditory areas. Studies in a variety of species, including humans, now suggest that prolonged acoustic training can drive long-lasting brainstem plasticity. The neurobiological mechanisms for such changes are not well understood in natural behavioral contexts due to a relative dearth of in vivo animal models in which to study this. Here, we demonstrate in a mouse model that a natural life experience with increased demands on the auditory system - motherhood - is associated with improved temporal processing in the subcortical auditory pathway. We measured the auditory brainstem response to test whether mothers and pup-naïve virgin mice differed in temporal responses to both broadband and tone stimuli, including ultrasonic frequencies found in mouse pup vocalizations. Mothers had shorter latencies for early ABR peaks, indicating plasticity in the auditory nerve and the cochlear nucleus. Shorter interpeak latency between waves IV and V also suggest plasticity in the inferior colliculus. Hormone manipulations revealed that these cannot be explained solely by estrogen levels experienced during pregnancy and parturition in mothers. In contrast, we found that pup-care experience, independent of pregnancy and parturition, contributes to shortening auditory brainstem response latencies. These results suggest that acoustic experience in the maternal context imparts plasticity on early auditory processing that lasts beyond pup weaning. In addition to establishing an animal model for exploring adult auditory brainstem plasticity in a neuroethological context, our results have broader implications for models of perceptual, behavioral and neural changes that arise during maternity, where subcortical sensorineural plasticity has not previously been considered.
Brumm, Henrik; Zollinger, Sue Anne
Sophisticated vocal communication systems of birds and mammals, including human speech, are characterized by a high degree of plasticity in which signals are individually adjusted in response to changes in the environment. Here, we present, to our knowledge, the first evidence for vocal plasticity in a reptile. Like birds and mammals, tokay geckos ( Gekko gecko ) increased the duration of brief call notes in the presence of broadcast noise compared to quiet conditions, a behaviour that facilitates signal detection by receivers. By contrast, they did not adjust the amplitudes of their call syllables in noise (the Lombard effect), which is in line with the hypothesis that the Lombard effect has evolved independently in birds and mammals. However, the geckos used a different strategy to increase signal-to-noise ratios: instead of increasing the amplitude of a given call type when exposed to noise, the subjects produced more high-amplitude syllable types from their repertoire. Our findings demonstrate that reptile vocalizations are much more flexible than previously thought, including elaborate vocal plasticity that is also important for the complex signalling systems of birds and mammals. We suggest that signal detection constraints are one of the major forces driving the evolution of animal communication systems across different taxa. © 2017 The Author(s).
Full Text Available We have been engaged in the development of a brain-computer interface (BCI based on the cognitive P300 event-related potentials (ERPs evoked by simultaneous presentation of visual and auditory stimuli in order to assist with the communication in severe physical limitation persons. The purpose of the simultaneous presentation of these stimuli is to give the user more choices as commands. First, we extracted P300 ERPs by either visual oddball paradigm or auditory oddball paradigm. Then amplitude and latency of the P300 ERPs were measured. Second, visual and auditory stimuli were presented simultaneously, we measured the P300 ERPs varying the condition of combinations of these stimuli. In this report, we used 3 colors as visual stimuli and 3 types of MIDI sounds as auditory stimuli. Two types of simultaneous presentations were examined. The one was conducted with random combination. The other was called group stimulation, combining one color, such as red, and one MIDI sound, such as piano, in order to make a group; three groups were made. Each group was presented to users randomly. We evaluated the possibility of BCI using these stimuli from the amplitudes and the latencies of P300 ERPs.
Li, Nicole Y.K.; Heris, Hossein K.; Mongeau, Luc
The vocal folds, which are located in the larynx, are the main organ of voice production for human communication. The vocal folds are under continuous biomechanical stress similar to other mechanically active organs, such as the heart, lungs, tendons and muscles. During speech and singing, the vocal folds oscillate at frequencies ranging from 20 Hz to 3 kHz with amplitudes of a few millimeters. The biomechanical stress associated with accumulated phonation is believed to alter vocal fold cell activity and tissue structure in many ways. Excessive phonatory stress can damage tissue structure and induce a cell-mediated inflammatory response, resulting in a pathological vocal fold lesion. On the other hand, phonatory stress is one major factor in the maturation of the vocal folds into a specialized tri-layer structure. One specific form of vocal fold oscillation, which involves low impact and large amplitude excursion, is prescribed therapeutically for patients with mild vocal fold injuries. Although biomechanical forces affect vocal fold physiology and pathology, there is little understanding of how mechanical forces regulate these processes at the cellular and molecular level. Research into vocal fold mechanobiology has burgeoned over the past several years. Vocal fold bioreactors are being developed in several laboratories to provide a biomimic environment that allows the systematic manipulation of physical and biological factors on the cells of interest in vitro. Computer models have been used to simulate the integrated response of cells and proteins as a function of phonation stress. The purpose of this paper is to review current research on the mechanobiology of the vocal folds as it relates to growth, pathogenesis and treatment as well as to propose specific research directions that will advance our understanding of this subject. PMID:24812638
Tan, Melin; Pitman, Michael J
We present a patient with a novel finding of bilateral mucosal bridges, bilateral type III trans-vocal fold sulci vocales, and a vocal fold polyp. Although sulci and mucosal bridges occur in the vocal folds, it is rare to find multiples of these lesions in a single patient, and it is even more uncommon when they occur in conjunction with a vocal fold polyp. To our knowledge, this is the first description of a vocal fold polyp in combination with multiple vocal fold bridges and multiple type III sulci vocales in a single patient. To describe and visually present the diagnosis and treatment of a patient with an intracordal polyp, bilateral mucosal bridges, as well as bilateral type III trans-vocal fold sulci vocales. Presentation of a set of high definition intraoperative photos displaying the extent of the vocal fold lesions and the resection of the intracordal polyp. This patient presented with only 6 months of significant dysphonia. It was felt that the recent change in voice was because of the polyp and not the bridges or sulci vocales. Considering the patient's presentation and the possible morbidity of resection of mucosal bridges and sulci, only the polyp was excised. Postoperatively, the patient's voice returned to his acceptable mild baseline dysphonia, and the benefit has persisted 6 months postoperatively. The combination of bilateral mucosal bridges, bilateral type III sulcus vocalis, and an intracordal polyp in one patient is rare if not novel. Treatment of the polyp alone returned the patient's voice to his lifelong baseline of mild dysphonia. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Narins, Peter M
Frogs and toads are capable of producing calls at potentially damaging levels that exceed 110 dB SPL at 50 cm. Most frog species have internally coupled ears (ICE) in which the tympanic membranes (TyMs) communicate directly via the large, permanently open Eustachian tubes, resulting in an inherently directional asymmetrical pressure-difference receiver. One active mechanism for auditory sensitivity reduction involves the pressure increase during vocalization that distends the TyM, reducing its low-frequency airborne sound sensitivity. Moreover, if sounds generated by the vocal folds arrive at both surfaces of the TyM with nearly equal amplitudes and phases, the net motion of the eardrum would be greatly attenuated. Both of these processes appear to reduce the motion of the frog's TyM during vocalizations. The implications of ICE in amphibians with respect to sound localizations are discussed, and the particularly interesting case of frogs that use ultrasound for communication yet exhibit exquisitely small localization jump errors is brought to light.
... here Home » Health Info » Voice, Speech, and Language Vocal Fold Paralysis On this page: What is vocal fold ... Where can I get additional information? What is vocal fold paralysis? Structures involved in speech and voice production ...
Guzman, Marco; Laukkanen, Anne-Maria; Krupa, Petr; Horáček, Jaromir; Švec, Jan G; Geneid, Ahmed
The present study aimed to investigate the vocal tract and glottal function during and after phonation into a tube and a stirring straw. A male classically trained singer was assessed. Computerized tomography (CT) was performed when the subject produced [a:] at comfortable speaking pitch, phonated into the resonance tube and when repeating [a:] after the exercise. Similar procedure was performed with a narrow straw after 15 minutes silence. Anatomic distances and area measures were obtained from CT midsagittal and transversal images. Acoustic, perceptual, electroglottographic (EGG), and subglottic pressure measures were also obtained. During and after phonation into the tube or straw, the velum closed the nasal passage better, the larynx position lowered, and hypopharynx area widened. Moreover, the ratio between the inlet of the lower pharynx and the outlet of the epilaryngeal tube became larger during and after tube/straw phonation. Acoustic results revealed a stronger spectral prominence in the singer/speaker's formant cluster region after exercising. Listening test demonstrated better voice quality after straw/tube than before. Contact quotient derived from EGG decreased during both tube and straw and remained lower after exercising. Subglottic pressure increased during straw and remained somewhat higher after it. CT and acoustic results indicated that vocal exercises with increased vocal tract impedance lead to increased vocal efficiency and economy. One of the major changes was the more prominent singer's/speaker's formant cluster. Vocal tract and glottal modifications were more prominent during and after straw exercising compared with tube phonation. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Baker, Vicki D.; Cohen, Nicki
The purpose of this study was to describe the university vocal training and vocal health of music educators and music therapists. The participants (N = 426), music educators (n = 351) and music therapists (n = 75), completed a survey addressing demographics, vocal training, voice usage, and vocal health. Both groups reported singing at least 50%…
Heo, Jeong; Baek, Hyun Jae; Hong, Seunghyeok; Chang, Min Hye; Lee, Jeong Su; Park, Kwang Suk
Patients with total locked-in syndrome are conscious; however, they cannot express themselves because most of their voluntary muscles are paralyzed, and many of these patients have lost their eyesight. To improve the quality of life of these patients, there is an increasing need for communication-supporting technologies that leverage the remaining senses of the patient along with physiological signals. The auditory steady-state response (ASSR) is an electro-physiologic response to auditory stimulation that is amplitude-modulated by a specific frequency. By leveraging the phenomenon whereby ASSR is modulated by mind concentration, a brain-computer interface paradigm was proposed to classify the selective attention of the patient. In this paper, we propose an auditory stimulation method to minimize auditory stress by replacing the monotone carrier with familiar music and natural sounds for an ergonomic system. Piano and violin instrumentals were employed in the music sessions; the sounds of water streaming and cicadas singing were used in the natural sound sessions. Six healthy subjects participated in the experiment. Electroencephalograms were recorded using four electrodes (Cz, Oz, T7 and T8). Seven sessions were performed using different stimuli. The spectral power at 38 and 42Hz and their ratio for each electrode were extracted as features. Linear discriminant analysis was utilized to classify the selections for each subject. In offline analysis, the average classification accuracies with a modulation index of 1.0 were 89.67% and 87.67% using music and natural sounds, respectively. In online experiments, the average classification accuracies were 88.3% and 80.0% using music and natural sounds, respectively. Using the proposed method, we obtained significantly higher user-acceptance scores, while maintaining a high average classification accuracy. Copyright © 2017 Elsevier Ltd. All rights reserved.
Full Text Available O posicionamento da prega vocal paralisada e o grau de disfonia são fatores importantes para decidir as opções de tratamento na paralisia de prega vocal unilateral (PPVU. OBJETIVO: Verificar as características perceptivo-auditivas da voz e a posição da prega vocal paralisada, em homens, com PPVU. MATERIAIS E MÉTODOS: Estudo retrospectivo, coorte histórica, com corte transversal, com dados de 24 homens com PPVU, com média de 60,7 anos, submetidos à avaliação vocal perceptivo-auditiva da voz, por três juízas fonoaudiólogas e perceptivo-visual das imagens laríngeas, com a classificação da posição da prega vocal paralisada, por três juízes otorrinolaringologistas. RESULTADOS: A prega vocal paralisada em posição paramediana ocorreu em 45,83% dos casos; a intermediária, em 25%; a lateral, em 20,83%, e a mediana, em 4,16%; a disfonia resultante da PPVU foi caracterizada pela rouquidão, aspereza e tensão, de grau moderado; soprosidade (maior frequência do grau grave; astenia e instabilidade (maior frequência do grau leve; a posição da prega vocal paralisada influenciou significativamente o grau geral de desvio vocal. CONCLUSÃO: O grau geral de disfonia está relacionado com a posição da prega vocal paralisada; a disfonia é caracterizada pela presença de rouquidão, soprosidade, aspereza e tensão de grau moderado a grave.The paralyzed vocal fold positioning and the degree of dysphonia are important inputs when one is deciding upon treatment options for unilateral vocal fold paralysis (UVFP. OBJECTIVE: To check voice characteristics and paralyzed vocal fold position in men with UVFP. MATERIALS AND METHODS: This is a retrospective historical cross-sectional cohort study, with data from 24 men with UVFP with mean age of 60.7 years, submitted to voice assessment by three speech therapists and three ENT physicians used laryngeal images to classify the position of the paralyzed vocal fold. RESULTS: The paralyzed vocal fold
Angadi, Vrushali; Croake, Daniel; Stemple, Joseph
The purpose of the present review was to systematically analyze the evidence for the effectiveness of vocal function exercises (VFEs) in improving voice production. A systematic literature search was performed by two independent reviewers using PubMed and EBSCOHost to access relevant databases and to locate outcome studies that used VFEs as an intervention. Articles that met inclusion criteria were appraised based on the American Speech-Language and Hearing Association's levels of evidence. Effect sizes for outcomes were calculated using Hedge's g. Voice outcomes were categorized according to the five domains of voice assessment: visual perceptual analysis, acoustic analysis, aerodynamic analysis, auditory-perceptual analysis, and patient self-report measures. Twenty-one articles were included for the final appraisal. All studies demonstrated positive effects of VFEs as demonstrated by effect sizes across selected voice parameters. Effect sizes across parameters ranged from -0.59 to 1.55. None of the included studies reported adverse voice outcomes as a result of VFEs. Outcome studies demonstrate that VFEs are efficacious in enhancing vocal function in individuals with normal and disordered voices, presbylaryngeus, and professional voice users. The available research suggests moderate to strong evidence to support the use of VFEs for a variety of voice disorders. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Grøntved, Ågot Møller; Faber, Christian; Jakobsen, John
INTRODUCTION: Thyroplasty with silicone rubber implantation is a surgical procedure for treatment of patients with vocal fold paralysis. The aim of the present study was to evaluate the outcome of the operation and to monitor which of the analyses were the more beneficial. MATERIAL AND METHODS...... because it offers a quantitative measure of the voice capacity and intensity, which are the major problems experienced by patients with vocal fold paralysis. Used together, these tools are highly instrumental in guiding the patient's choice of surgery or no surgery. Udgivelsesdato: 2009-Jan-12...
Full Text Available Natural sounds, including vocal communication sounds, contain critical information at multiple time scales. Two essential temporal modulation rates in speech have been argued to be in the low gamma band (~20-80 ms duration information and the theta band (~150-300 ms, corresponding to segmental and syllabic modulation rates, respectively. On one hypothesis, auditory cortex implements temporal integration using time constants closely related to these values. The neural correlates of a proposed dual temporal window mechanism in human auditory cortex remain poorly understood. We recorded MEG responses from participants listening to non-speech auditory stimuli with different temporal structures, created by concatenating frequency-modulated segments of varied segment durations. We show that these non-speech stimuli with temporal structure matching speech-relevant scales (~25 ms and ~200 ms elicit reliable phase tracking in the corresponding associated oscillatory frequencies (low gamma and theta bands. In contrast, stimuli with non-matching temporal structure do not. Furthermore, the topography of theta band phase tracking shows rightward lateralization while gamma band phase tracking occurs bilaterally. The results support the hypothesis that there exists multi-time resolution processing in cortex on discontinuous scales and provide evidence for an asymmetric organization of temporal analysis (asymmetrical sampling in time, AST. The data argue for a macroscopic-level neural mechanism underlying multi-time resolution processing: the sliding and resetting of intrinsic temporal windows on privileged time scales.
Eisenberg, Laurie S; Johnson, Karen C; Martinez, Amy S; DesJardin, Jean L; Stika, Carren J; Dzubak, Danielle; Mahalak, Mandy Lutz; Rector, Emily P
We had an opportunity to evaluate an American child whose family traveled to Italy to receive an auditory brainstem implant (ABI). The goal of this evaluation was to obtain insight into possible benefits derived from the ABI and to begin developing assessment protocols for pediatric clinical trials. Case study. Tertiary referral center. Pediatric ABI Patient 1 was born with auditory nerve agenesis. Auditory brainstem implant surgery was performed in December, 2005, in Verona, Italy. The child was assessed at the House Ear Institute, Los Angeles, in July 2006 at the age of 3 years 11 months. Follow-up assessment has continued at the HEAR Center in Birmingham, Alabama. Auditory brainstem implant. Performance was assessed for the domains of audition, speech and language, intelligence and behavior, quality of life, and parental factors. Patient 1 demonstrated detection of sound, speech pattern perception with visual cues, and inconsistent auditory-only vowel discrimination. Language age with signs was approximately 2 years, and vocalizations were increasing. Of normal intelligence, he exhibited attention deficits with difficulty completing structured tasks. Twelve months later, this child was able to identify speech patterns consistently; closed-set word identification was emerging. These results were within the range of performance for a small sample of similarly aged pediatric cochlear implant users. Pediatric ABI assessment with a group of well-selected children is needed to examine risk versus benefit in this population and to analyze whether open-set speech recognition is achievable.
Smith, Simeon L; Titze, Ingo R
The fluid-structure interaction and energy transfer from respiratory airflow to self-sustained vocal fold oscillation continues to be a topic of interest in vocal fold research. Vocal fold vibration is driven by pressures on the vocal fold surface, which are determined by the shape of the glottis and the contact between vocal folds. Characterization of three-dimensional glottal shapes and contact patterns can lead to increased understanding of normal and abnormal physiology of the voice, as well as to development of improved vocal fold models, but a large inventory of shapes has not been directly studied previously. This study aimed to take an initial step toward characterizing vocal fold contact patterns systematically. Vocal fold motion and contact was modeled based on normal mode vibration, as it has been shown that vocal fold vibration can be almost entirely described by only the few lowest order vibrational modes. Symmetric and asymmetric combinations of the four lowest normal modes of vibration were superimposed on left and right vocal fold medial surfaces, for each of three prephonatory glottal configurations, according to a surface wave approach. Contact patterns were generated from the interaction of modal shapes at 16 normalized phases during the vibratory cycle. Eight major contact patterns were identified and characterized by the shape of the flow channel, with the following descriptors assigned: convergent, divergent, convergent-divergent, uniform, split, merged, island, and multichannel. Each of the contact patterns and its variation are described, and future work and applications are discussed. Copyright © 2018 Elsevier Ltd. All rights reserved.
Fernández-Caballero, Antonio; Navarro, Elena; Fernández-Sotos, Patricia; González, Pascual; Ricarte, Jorge J; Latorre, José M; Rodriguez-Jimenez, Roberto
This perspective paper faces the future of alternative treatments that take advantage of a social and cognitive approach with regards to pharmacological therapy of auditory verbal hallucinations (AVH) in patients with schizophrenia. AVH are the perception of voices in the absence of auditory stimulation and represents a severe mental health symptom. Virtual/augmented reality (VR/AR) and brain computer interfaces (BCI) are technologies that are growing more and more in different medical and psychological applications. Our position is that their combined use in computer-based therapies offers still unforeseen possibilities for the treatment of physical and mental disabilities. This is why, the paper expects that researchers and clinicians undergo a pathway toward human-avatar symbiosis for AVH by taking full advantage of new technologies. This outlook supposes to address challenging issues in the understanding of non-pharmacological treatment of schizophrenia-related disorders and the exploitation of VR/AR and BCI to achieve a real human-avatar symbiosis.
Syka, Josef; Šuta, Daniel; Popelář, Jiří
Roč. 206, - (2005), s. 177-184 ISSN 0378-5955 R&D Projects: GA ČR GA309/04/1074; GA MZd(CZ) NR8113 Institutional research plan: CEZ:AV0Z5039906; CEZ:AV0Z50390512 Keywords : anesthesia * ketamine * vocalization Subject RIV: FH - Neurology Impact factor: 1.674, year: 2005
Schwarz, Karine; Fontanari, Anna Martha Vaitses; Costa, Angelo Brandelli; Soll, Bianca Machado Borba; da Silva, Dhiordan Cardoso; de Sá Villas-Bôas, Anna Paula; Cielo, Carla Aparecida; Bastilha, Gabriele Rodrigues; Ribeiro, Vanessa Veis; Dorfman, Maria Elza Kazumi Yamaguti; Lobato, Maria Inês Rodrigues
Voice is an important gender marker in the transition process as a transgender individual accepts a new gender identity. The objectives of this study were to describe and relate aspects of a perceptual-auditory analysis and the fundamental frequency (F0) of male-to-female (MtF) transsexual individuals. A case-control study was carried out with individuals aged 19-52 years who attended the Gender Identity Program of the Hospital de Clínicas of Porto Alegre. Vocal recordings from the MtF transgender and cisgender individuals (vowel /a:/ and six phrases of Consensus Auditory Perceptual Evaluation Voice [CAPE-V]) were edited and randomly coded before storage in a Dropbox folder. The voices (vowel /a:/) were analyzed by consensus on the same day by two judge speech therapists who had more than 10 years of experience in the voice area using the GRBASI perceptual-auditory vocal evaluation scale. Acoustic analysis of the voices was performed using the advanced Multi-Dimensional Voice Program software. The resonance focus and the degrees of masculinity and femininity for each voice recording were determined by listening to the CAPE-V phrases, for the same judges. There were significant differences between the groups regarding a greater frequency of subjects with F0 between 80 and 150 Hz (P = 0.003), and a greater frequency of hypernasal resonant focus (P < 0.001) in the MtF cases and greater frequency of subjects with absence of roughness (P = 0.031) in the control group. The MtF group of individuals showed altered vertical resonant focus, more masculine voices, and lower fundamental frequencies. The control group showed a significant absence of roughness. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Croake, Daniel J.; Andreatta, Richard D.; Stemple, Joseph C.
Purpose: The purpose of this study is to quantify the interactions of the 3 vocalization subsystems of respiration, phonation, and resonance before, during, and after a perturbation to the larynx (temporarily induced unilateral vocal fold paralysis) in 10 vocally healthy participants. Using dynamic systems theory as a guide, we hypothesized that…
Roy, Arani; Mooney, Richard
In adult male zebra finches, transecting the vocal nerve causes previously stable (i.e., crystallized) song to slowly degrade, presumably because of the resulting distortion in auditory feedback. How and where distorted feedback interacts with song motor networks to induce this process of song decrystallization remains unknown. The song premotor nucleus HVC is a potential site where auditory feedback signals could interact with song motor commands. Although the forebrain nucleus interface of the nidopallium (NIf) appears to be the primary auditory input to HVC, NIf lesions made in adult zebra finches do not trigger song decrystallization. One possibility is that NIf lesions do not interfere with song maintenance, but do compromise the adult zebra finch's ability to express renewed vocal plasticity in response to feedback perturbations. To test this idea, we bilaterally lesioned NIf and then transected the vocal nerve in adult male zebra finches. We found that bilateral NIf lesions did not prevent nerve section-induced song decrystallization. To test the extent to which the NIf lesions disrupted auditory processing in the song system, we made in vivo extracellular recordings in HVC and a downstream anterior forebrain pathway (AFP) in NIf-lesioned birds. We found strong and selective auditory responses to the playback of the birds' own song persisted in HVC and the AFP following NIf lesions. These findings suggest that auditory inputs to the song system other than NIf, such as the caudal mesopallium, could act as a source of auditory feedback signals to the song motor network.
Owens, Jessica L; Olsen, Mariana; Fontaine, Amy; Kloth, Christopher; Kershenbaum, Arik; Waller, Sara
Cat vocal behavior, in particular, the vocal and social behavior of feral cats, is poorly understood, as are the differences between feral and fully domestic cats. The relationship between feral cat social and vocal behavior is important because of the markedly different ecology of feral and domestic cats, and enhanced comprehension of the repertoire and potential information content of feral cat calls can provide both better understanding of the domestication and socialization process, and improved welfare for feral cats undergoing adoption. Previous studies have used conflicting classification schemes for cat vocalizations, often relying on onomatopoeic or popular descriptions of call types (e.g., "miow"). We studied the vocalizations of 13 unaltered domestic cats that complied with our behavioral definition used to distinguish feral cats from domestic. A total of 71 acoustic units were extracted and visually analyzed for the construction of a hierarchical classification of vocal sounds, based on acoustic properties. We identified 3 major categories (tonal, pulse, and broadband) that further breakdown into 8 subcategories, and show a high degree of reliability when sounds are classified blindly by independent observers (Fleiss' Kappa K = 0.863). Due to the limited behavioral contexts in this study, additional subcategories of cat vocalizations may be identified in the future, but our hierarchical classification system allows for the addition of new categories and new subcategories as they are described. This study shows that cat vocalizations are diverse and complex, and provides an objective and reliable classification system that can be used in future studies.
Agerkvist, Finn T.; Selamtzis, Andreas
frontend is used to measure the electroglottograph signal which reflects the opening and closing pattern of the vocal folds. The measurements were carried out for all four modes (Neutral, Curbing, Overdrive and Edge) for the vowel [a] in three different pitches: C3(131 Hz), G3 (196 Hz) and C4 (262Hz......The importance of the interaction between the acoustic impedance of the vocal tract with the flow across the vocal cords is well established. In this paper we are investigating the changes in vocal tract impedance when using the different modes of phonation according to Sadolin , going from...... the soft levels of the Neutral mode to the high levels of the fully ‘metallic’ Edge mode. The acoustic impedance of vocal tract as seen from the mouth opening is measured via a microphone placed close to the mouth when exciting the system with a volume velocity source . At the same time a Laryngograph...
Taitelbaum-Swead, Riki; Icht, Michal; Mama, Yaniv
In recent years, the effect of cognitive abilities on the achievements of cochlear implant (CI) users has been evaluated. Some studies have suggested that gaps between CI users and normal-hearing (NH) peers in cognitive tasks are modality specific, and occur only in auditory tasks. The present study focused on the effect of learning modality (auditory, visual) and auditory feedback on word memory in young adults who were prelingually deafened and received CIs before the age of 5 yr, and their NH peers. A production effect (PE) paradigm was used, in which participants learned familiar study words by vocal production (saying aloud) or by no-production (silent reading or listening). Words were presented (1) in the visual modality (written) and (2) in the auditory modality (heard). CI users performed the visual condition twice-once with the implant ON and once with it OFF. All conditions were followed by free recall tests. Twelve young adults, long-term CI users, implanted between ages 1.7 and 4.5 yr, and who showed ≥50% in monosyllabic consonant-vowel-consonant open-set test with their implants were enrolled. A group of 14 age-matched NH young adults served as the comparison group. For each condition, we calculated the proportion of study words recalled. Mixed-measures analysis of variances were carried out with group (NH, CI) as a between-subjects variable, and learning condition (aloud or silent reading) as a within-subject variable. Following this, paired sample t tests were used to evaluate the PE size (differences between aloud and silent words) and overall recall ratios (aloud and silent words combined) in each of the learning conditions. With visual word presentation, young adults with CIs (regardless of implant status CI-ON or CI-OFF), showed comparable memory performance (and a similar PE) to NH peers. However, with auditory presentation, young adults with CIs showed poorer memory for nonproduced words (hence a larger PE) relative to their NH peers. The
Brefczynski-Lewis, Julie A; Lewis, James W
Interaction with the world is a multisensory experience, but most of what is known about the neural correlates of perception comes from studying vision. Auditory inputs enter cortex with its own set of unique qualities, and leads to use in oral communication, speech, music, and the understanding of emotional and intentional states of others, all of which are central to the human experience. To better understand how the auditory system develops, recovers after injury, and how it may have transitioned in its functions over the course of hominin evolution, advances are needed in models of how the human brain is organized to process real-world natural sounds and "auditory objects". This review presents a simple fundamental neurobiological model of hearing perception at a category level that incorporates principles of bottom-up signal processing together with top-down constraints of grounded cognition theories of knowledge representation. Though mostly derived from human neuroimaging literature, this theoretical framework highlights rudimentary principles of real-world sound processing that may apply to most if not all mammalian species with hearing and acoustic communication abilities. The model encompasses three basic categories of sound-source: (1) action sounds (non-vocalizations) produced by 'living things', with human (conspecific) and non-human animal sources representing two subcategories; (2) action sounds produced by 'non-living things', including environmental sources and human-made machinery; and (3) vocalizations ('living things'), with human versus non-human animals as two subcategories therein. The model is presented in the context of cognitive architectures relating to multisensory, sensory-motor, and spoken language organizations. The models' predictive values are further discussed in the context of anthropological theories of oral communication evolution and the neurodevelopment of spoken language proto-networks in infants/toddlers. These phylogenetic
Full Text Available Numerous animal species emit vocalizations in response to various social stimuli. The neural basis of vocal communication has been investigated in monkeys, songbirds, rats, bats and invertebrates resulting in deep insights into motor control, neural coding and learning. Mice, which recently became very popular as a model system for mammalian neuroscience, also utilize ultrasonic vocalizations (USVs during mating behavior. However, our knowledge is lacking of both the behavior and its underlying neural mechanism. We developed a novel method for head-restrained male mice (HRMM to interact with non-restrained female mice (NRFM and show that mice can emit USVs in this context. We first recorded USVs in free arena with non-restrained male mice (NRMM and NRFM. Of the NRMM, which vocalized in the free arena, the majority could be habituated to also vocalize while head-restrained but only when a female mouse was present in proximity. The USVs emitted by HRMM are similar to the USVs of NRMM in the presence of a female mouse in their spectral structure, inter syllable interval distribution and USV sequence length, and therefore are interpreted as social USVs. By analyzing vocalizations of NRMM, we established criteria to predict which individuals are likely to vocalize while head fixed based on the USV rate and average syllable duration. To characterize the USVs emitted by HRMM, we analyzed the syllable composition of HRMM and NRMM and found that USVs emitted by HRMM have higher proportions of USVs with complex spectral representation, supporting previous studies showing that mice social USVs are context dependent. Our results suggest a way to study the neural mechanisms of production and control of social vocalization in mice using advanced methods requiring head fixation.
Weiner, Benjamin; Hertz, Stav; Perets, Nisim; London, Michael
Numerous animal species emit vocalizations in response to various social stimuli. The neural basis of vocal communication has been investigated in monkeys, songbirds, rats, bats, and invertebrates resulting in deep insights into motor control, neural coding, and learning. Mice, which recently became very popular as a model system for mammalian neuroscience, also utilize ultrasonic vocalizations (USVs) during mating behavior. However, our knowledge is lacking of both the behavior and its underlying neural mechanism. We developed a novel method for head-restrained male mice (HRMM) to interact with non-restrained female mice (NRFM) and show that mice can emit USVs in this context. We first recorded USVs in a free arena with non-restrained male mice (NRMM) and NRFM. Of the NRMM, which vocalized in the free arena, the majority could be habituated to also vocalize while head-restrained but only when a female mouse was present in proximity. The USVs emitted by HRMM are similar to the USVs of NRMM in the presence of a female mouse in their spectral structure, inter-syllable interval distribution, and USV sequence length, and therefore are interpreted as social USVs. By analyzing the vocalizations of NRMM, we established criteria to predict which individuals are likely to vocalize while head fixed based on the USV rate and average syllable duration. To characterize the USVs emitted by HRMM, we analyzed the syllable composition of HRMM and NRMM and found that USVs emitted by HRMM have a higher proportion of USVs with complex spectral representation, supporting previous studies showing that mice social USVs are context dependent. Our results suggest a way to study the neural mechanisms of production and control of social vocalization in mice using advanced methods requiring head fixation.
The auditory cortex communicates with the frontal lobe via the middle temporal gyrus (auditory ventral stream; AVS) or the inferior parietal lobule (auditory dorsal stream; ADS). Whereas the AVS is ascribed only with sound recognition, the ADS is ascribed with sound localization, voice detection, prosodic perception/production, lip-speech integration, phoneme discrimination, articulation, repetition, phonological long-term memory and working memory. Previously, I interpreted the juxtaposition of sound localization, voice detection, audio-visual integration and prosodic analysis, as evidence that the behavioral precursor to human speech is the exchange of contact calls in non-human primates. Herein, I interpret the remaining ADS functions as evidence of additional stages in language evolution. According to this model, the role of the ADS in vocal control enabled early Homo (Hominans) to name objects using monosyllabic calls, and allowed children to learn their parents' calls by imitating their lip movements. Initially, the calls were forgotten quickly but gradually were remembered for longer periods. Once the representations of the calls became permanent, mimicry was limited to infancy, and older individuals encoded in the ADS a lexicon for the names of objects (phonological lexicon). Consequently, sound recognition in the AVS was sufficient for activating the phonological representations in the ADS and mimicry became independent of lip-reading. Later, by developing inhibitory connections between acoustic-syllabic representations in the AVS and phonological representations of subsequent syllables in the ADS, Hominans became capable of concatenating the monosyllabic calls for repeating polysyllabic words (i.e., developed working memory). Finally, due to strengthening of connections between phonological representations in the ADS, Hominans became capable of encoding several syllables as a single representation (chunking). Consequently, Hominans began vocalizing and
Ezequiel M Arneodo
Full Text Available Because of the parallels found with human language production and acquisition, birdsong is an ideal animal model to study general mechanisms underlying complex, learned motor behavior. The rich and diverse vocalizations of songbirds emerge as a result of the interaction between a pattern generator in the brain and a highly nontrivial nonlinear periphery. Much of the complexity of this vocal behavior has been understood by studying the physics of the avian vocal organ, particularly the syrinx. A mathematical model describing the complex periphery as a nonlinear dynamical system leads to the conclusion that nontrivial behavior emerges even when the organ is commanded by simple motor instructions: smooth paths in a low dimensional parameter space. An analysis of the model provides insight into which parameters are responsible for generating a rich variety of diverse vocalizations, and what the physiological meaning of these parameters is. By recording the physiological motor instructions elicited by a spontaneously singing muted bird and computing the model on a Digital Signal Processor in real-time, we produce realistic synthetic vocalizations that replace the bird's own auditory feedback. In this way, we build a bio-prosthetic avian vocal organ driven by a freely behaving bird via its physiologically coded motor commands. Since it is based on a low-dimensional nonlinear mathematical model of the peripheral effector, the emulation of the motor behavior requires light computation, in such a way that our bio-prosthetic device can be implemented on a portable platform.
Full Text Available The extent to which human speech perception evolved by taking advantage of predispositions and pre-existing features of vertebrate auditory and cognitive systems remains a central question in the evolution of speech. This paper reviews asymmetries in vowel perception, speaker voice recognition, and speaker normalization in non-human animals – topics that have not been thoroughly discussed in relation to the abilities of non-human animals, but are nonetheless important aspects of vocal perception. Throughout this paper we demonstrate that addressing these issues in non-human animals is relevant and worthwhile because many non-human animals must deal with similar issues in their natural environment. That is, they must also discriminate between similar-sounding vocalizations, determine signaler identity from vocalizations, and resolve signaler-dependent variation in vocalizations from conspecifics. Overall, we find that, although plausible, the current evidence is insufficiently strong to conclude that directional asymmetries in vowel perception are specific to humans, or that non-human animals can use voice characteristics to recognize human individuals. However, we do find some indication that non-human animals can normalize speaker differences. Accordingly, we identify avenues for future research that would greatly improve and advance our understanding of these topics.
Full Text Available OBJETIVO: analisar o impacto vocal nas atividades diárias em professores do ensino médio. Correlacionar os achado da auto-percepção do problema vocal com os aspectos: efeitos no trabalho, na comunicação diária, na comunicação social e na sua emoção. MÉTODOS: a amostra foi constituída por 107 professores, sendo 86 com queixa e 21 sem queixa, selecionados em escolas da rede particular de ensino de Maceió-AL. Cada professor respondeu individualmente o protocolo Perfil Participação em Atividades Vocais na presença da pesquisadora, assinalando suas respostas em uma escala visual que varia de 0 a 10. O protocolo é composto por 28 questões com a presença integrada em cinco aspectos englobados para avaliar a qualidade de vida e o resultado de tratamentos vocais. O protocolo oferece, ainda, dois escores adicionais: pontuação de limitação nas atividades (PLA e de restrição de participação (PRP. RESULTADOS: na comparação dos grupos com e sem queixa vocal foram verificados que todos os resultados foram estatisticamente significantes (pPURPOSE: to analyze the vocal impact in the daily activities on high-school teachers. Correlate the finding of the auto-perception on the vocal problem with the following aspects: effects in the work, daily communication, social communication and, its emotion METHODS: the sample consisted of 107 teachers, 86 with and 21 with no complaint, selected from private teaching schools in Maceió-AL. Each teacher answered individually the Protocol for Voice Activity Participation Profile in the presence of the researcher, noting their responses on a visual scale ranging from 0 to 10. The protocol is composed of 28 questions with the presence integrated in five aspects to evaluate the quality of life and the result of vocal treatments. The protocol offers, still, two additional scores: punctuation of limitation in the activities (PLA and restriction of participation (PRP. RESULTS: comparing the groups with
Riede, Tobias; Titze, Ingo R.
The vocal folds of male Rocky Mountain elk (Cervus elaphus nelsoni) are about 3 cm long. If fundamental frequency were to be predicted by a simple vibrating string formula, as is often done for the human larynx, such long vocal folds would bear enormous stress to produce the species-specific mating call with an average fundamental frequency of 1 kHz. Predictions would be closer to 50 Hz. Vocal fold histology revealed the presence of a large vocal ligament between the vocal fold epithelium and...
Kikuchi, Yukiko; Horwitz, Barry; Mishkin, Mortimer; Rauschecker, Josef P
Many speech sounds and animal vocalizations contain components, referred to as complex tones, that consist of a fundamental frequency (F0) and higher harmonics. In this study we examined single-unit activity recorded in the core (A1) and lateral belt (LB) areas of auditory cortex in two rhesus monkeys as they listened to pure tones and pitch-shifted conspecific vocalizations ("coos"). The latter consisted of complex-tone segments in which F0 was matched to a corresponding pure-tone stimulus. In both animals, neuronal latencies to pure-tone stimuli at the best frequency (BF) were ~10 to 15 ms longer in LB than in A1. This might be expected, since LB is considered to be at a hierarchically higher level than A1. On the other hand, the latency of LB responses to coos was ~10 to 20 ms shorter than to the corresponding pure-tone BF, suggesting facilitation in LB by the harmonics. This latency reduction by coos was not observed in A1, resulting in similar coo latencies in A1 and LB. Multi-peaked neurons were present in both A1 and LB; however, harmonically-related peaks were observed in LB for both early and late response components, whereas in A1 they were observed only for late components. Our results suggest that harmonic features, such as relationships between specific frequency intervals of communication calls, are processed at relatively early stages of the auditory cortical pathway, but preferentially in LB.
Latham, Katherine; Messing, Barbara; Bidlack, Melissa; Merritt, Samantha; Zhou, Xian; Akst, Lee M
Most agree that education about vocal health and physiology can help singers avoid the development of vocal disorders. However, little is known about how this kind of education is provided to singers as part of their formal training. This study describes the amount of instruction in these topics provided through graduate-level curricula, who provides this instruction, and the kinds of affiliations such graduate singing programs have with medical professionals. This is an online survey of music schools with graduate singing programs. Survey questions addressed demographics of the programs, general attitudes about vocal health instruction for singers, the amount of vocal health instruction provided and by whom it was taught, perceived barriers to including more vocal health instruction, and any affiliations the voice program might have with medical personnel. Eighty-one survey responses were received. Instruction on vocal health was provided in 95% of the schools. In 55% of the schools, none of this instruction was given by a medical professional. Limited time in the curriculum, lack of financial support, and lack of availability of medical professional were the most frequently reported barriers to providing more instruction. When programs offered more hours of instruction, they were more likely to have some of that instruction given by a medical professional (P = 0.008) and to assess the amount of instruction provided positively (P = 0.001). There are several perceived barriers to incorporating vocal health education into graduate singing programs. Opportunity exists for more collaboration between vocal pedagogues and medical professionals in the education of singers about vocal health. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
D'haeseleer, Evelien; Claeys, Sofie; Bettens, Kim; Leemans, Laura; Van Calster, Ann-Sophie; Van Damme, Nina; Thijs, Zoë; Daelman, Julie; Leyns, Clara; Van Lierde, Kristiane
The purpose of this study was to measure the objective and subjective vocal quality in women aged between 60 and 75 years. Secondly, the impact of a teaching or singing career on the vocal quality was investigated by comparing the vocal quality of retired women with different careers. This is a case-control study. Seventy-three retired women between 60 and 75 years (mean age: 67 years, standard deviation: 4.49) participated in the study and were divided into three groups: women with a teaching career (n = 21), choir singers with a singing career (n = 12), and women with a non-vocal career (n = 40). All subjects underwent the same assessment protocol consisting of objective (aerodynamic, maximum performance, vocal range, acoustic measurements, and the Dysphonia Severity Index) and subjective (the Voice Handicap Index, auditory-perceptual evaluations by three listeners) voice measurements. In all three groups, objective and perceptual voice analysis showed a mild dysphonia. No differences in the Dysphonia Severity Index were found between the three groups. The voices of choir singers with a singing career were perceived significantly less rough than voices of the women with a non-vocal career. Additionally, the lowest frequency of the frequency range was significantly lower in the retired teachers and choir singers than in the controls. The results of this study prudently suggest that a singing or a teaching career compared with a non-vocal career has a positive impact on the vocal frequency range, and that singing has a positive impact on the perceptual vocal quality of the older female voice. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
King, Stephanie L.; Sayigh, Laela S.; Wells, Randall S.; Fellner, Wendi; Janik, Vincent M.
Vocal learning is relatively common in birds but less so in mammals. Sexual selection and individual or group recognition have been identified as major forces in its evolution. While important in the development of vocal displays, vocal learning also allows signal copying in social interactions. Such copying can function in addressing or labelling selected conspecifics. Most examples of addressing in non-humans come from bird song, where matching occurs in an aggressive context. However, in other animals, addressing with learned signals is very much an affiliative signal. We studied the function of vocal copying in a mammal that shows vocal learning as well as complex cognitive and social behaviour, the bottlenose dolphin (Tursiops truncatus). Copying occurred almost exclusively between close associates such as mother–calf pairs and male alliances during separation and was not followed by aggression. All copies were clearly recognizable as such because copiers consistently modified some acoustic parameters of a signal when copying it. We found no evidence for the use of copying in aggression or deception. This use of vocal copying is similar to its use in human language, where the maintenance of social bonds appears to be more important than the immediate defence of resources. PMID:23427174
Barbosa, Sara; Pires, Gabriel; Nunes, Urbano
Brain computer interfaces (BCIs) are one of the last communication options for patients in the locked-in state (LIS). For complete LIS patients, interfaces must be gaze-independent due to their eye impairment. However, unimodal gaze-independent approaches typically present levels of performance substantially lower than gaze-dependent approaches. The combination of multimodal stimuli has been pointed as a viable way to increase users' performance. A hybrid visual and auditory (HVA) P300-based BCI combining simultaneously visual and auditory stimulation is proposed. Auditory stimuli are based on natural meaningful spoken words, increasing stimuli discrimination and decreasing user's mental effort in associating stimuli to the symbols. The visual part of the interface is covertly controlled ensuring gaze-independency. Four conditions were experimentally tested by 10 healthy participants: visual overt (VO), visual covert (VC), auditory (AU) and covert HVA. Average online accuracy for the hybrid approach was 85.3%, which is more than 32% over VC and AU approaches. Questionnaires' results indicate that the HVA approach was the less demanding gaze-independent interface. Interestingly, the P300 grand average for HVA approach coincides with an almost perfect sum of P300 evoked separately by VC and AU tasks. The proposed HVA-BCI is the first solution simultaneously embedding natural spoken words and visual words to provide a communication lexicon. Online accuracy and task demand of the approach compare favorably with state-of-the-art. The proposed approach shows that the simultaneous combination of visual covert control and auditory modalities can effectively improve the performance of gaze-independent BCIs. Copyright © 2015 Elsevier B.V. All rights reserved.
Mehler, Bruce; Kidd, David; Reimer, Bryan; Reagan, Ian; Dobres, Jonathan; McCartt, Anne
One purpose of integrating voice interfaces into embedded vehicle systems is to reduce drivers' visual and manual distractions with 'infotainment' technologies. However, there is scant research on actual benefits in production vehicles or how different interface designs affect attentional demands. Driving performance, visual engagement, and indices of workload (heart rate, skin conductance, subjective ratings) were assessed in 80 drivers randomly assigned to drive a 2013 Chevrolet Equinox or Volvo XC60. The Chevrolet MyLink system allowed completing tasks with one voice command, while the Volvo Sensus required multiple commands to navigate the menu structure. When calling a phone contact, both voice systems reduced visual demand relative to the visual-manual interfaces, with reductions for drivers in the Equinox being greater. The Equinox 'one-shot' voice command showed advantages during contact calling but had significantly higher error rates than Sensus during destination address entry. For both secondary tasks, neither voice interface entirely eliminated visual demand. Practitioner Summary: The findings reinforce the observation that most, if not all, automotive auditory-vocal interfaces are multi-modal interfaces in which the full range of potential demands (auditory, vocal, visual, manipulative, cognitive, tactile, etc.) need to be considered in developing optimal implementations and evaluating drivers' interaction with the systems. Social Media: In-vehicle voice-interfaces can reduce visual demand but do not eliminate it and all types of demand need to be taken into account in a comprehensive evaluation.
Wiegand, Susanne; Teymoortash, Afshin; Hanschmann, Holger
Bilateral vocal fold paralysis can result in shortness of breath and severe dyspnea which can be life-threatening. Thirty-five patients with bilateral vocal fold paralysis who underwent endo-extralaryngeal laterofixation according to Lichtenberger were retrospectively analyzed regarding etiology, symptoms, treatment and complications. In 27 patients, laterofixation of the vocal cord alone was performed. Eight patients underwent laterofixation and additional posterior chordectomy of the opposite vocal cord according to Dennis and Kashima. The time of intervention ranged from 1 day to 38 years after the onset of bilateral vocal cord immobility. The intraoperative course was uneventful in all patients. None of the patients had postoperative aspiration. Postoperative voice function was acceptable in all patients. Complications of suture laterofixation were laryngeal edema, formation of fibrin, and malposition of the suture. Laterofixation of the vocal cords according to Lichtenberger is a safe and easy method that can be used as a first-stage treatment of vocal cord paralysis. Copyright© 2017, International Institute of Anticancer Research (Dr. George J. Delinasios), All rights reserved.
Full Text Available Parrots and songbirds learn their vocalizations from a conspecific tutor, much like human infants acquire spoken language. Parrots can learn human words and it has been suggested that they can use them to communicate with humans. The caudomedial pallium in the parrot brain is homologous with that of songbirds, and analogous to the human auditory association cortex, involved in speech processing. Here we investigated neuronal activation, measured as expression of the protein product of the immediate early gene ZENK, in relation to auditory learning in the budgerigar (Melopsittacus undulatus, a parrot. Budgerigar males successfully learned to discriminate two Japanese words spoken by another male conspecific. Re-exposure to the two discriminanda led to increased neuronal activation in the caudomedial pallium, but not in the hippocampus, compared to untrained birds that were exposed to the same words, or were not exposed to words. Neuronal activation in the caudomedial pallium of the experimental birds was correlated significantly and positively with the percentage of correct responses in the discrimination task. These results suggest that in a parrot, the caudomedial pallium is involved in auditory learning. Thus, in parrots, songbirds and humans, analogous brain regions may contain the neural substrate for auditory learning and memory.
Wilbiks, Jonathan M P; Vuvan, Dominique T; Girard, Pier-Yves; Peretz, Isabelle; Russo, Frank A
Congenital amusia is a condition in which an individual suffers from a deficit of musical pitch perception and production. Individuals suffering from congenital amusia generally tend to abstain from musical activities. Here, we present the unique case of Tim Falconer, a self-described musicophile who also suffers from congenital amusia. We describe and assess Tim's attempts to train himself out of amusia through a self-imposed 18-month program of formal vocal training and practice. We tested Tim with respect to music perception and vocal production across seven sessions including pre- and post-training assessments. We also obtained diffusion-weighted images of his brain to assess connectivity between auditory and motor planning areas via the arcuate fasciculus (AF). Tim's behavioral and brain data were compared to that of normal and amusic controls. While Tim showed temporary gains in his singing ability, he did not reach normal levels, and these gains faded when he was not engaged in regular lessons and practice. Tim did show some sustained gains with respect to the perception of musical rhythm and meter. We propose that Tim's lack of improvement in pitch perception and production tasks is due to long-standing and likely irreversible reduction in connectivity along the AF fiber tract.
Robbins, Lindsey; Margulis, Susan W
With the increase of mixed species exhibits in zoos, targeting enrichment for individual species may be problematic. Often, mammals may be the primary targets of enrichment, yet other species that share their environment (such as birds) will unavoidably be exposed to the enrichment as well. The purpose of this study was to determine if (1) auditory stimuli designed for enrichment of primates influenced the behavior of captive birds in the zoo setting, and (2) if the specific type of auditory enrichment impacted bird behavior. Three different African bird species were observed at the Buffalo Zoo during exposure to natural sounds, classical music and rock music. The results revealed that the average frequency of flying in all three bird species increased with naturalistic sounds and decreased with rock music (F = 7.63, df = 3,6, P = 0.018); vocalizations for two of the three species (Superb Starlings and Mousebirds) increased (F = 18.61, df = 2,6, P = 0.0027) in response to all auditory stimuli, however one species (Lady Ross's Turacos) increased frequency of duetting only in response to rock music (X(2) = 18.5, df = 2, P < 0.0001). Auditory enrichment implemented for large mammals may influence behavior in non-target species as well, in this case leading to increased activity by birds. © 2016 Wiley Periodicals, Inc.
Kátia de Freitas Alvarenga
Full Text Available Estudos voltados aos potenciais evocados auditivos com estímulos de fala em indivíduos ouvintes são importantes para compreender como a complexidade do estímulo influencia nas características do potencial cognitivo auditivo gerado. OBJETIVO: Caracterizar o potencial evocado auditivo cortical e o potencial cognitivo auditivo P3 com estímulos de contrastes vocálico e consonantal em indivíduos com audição normal. MÉTODO: Participaram deste estudo 31 indivíduos sem alterações auditivas, neurológicas e de linguagem na faixa etária de 7 a 30 anos. Os potenciais evocados auditivos corticais e cognitivo auditivo P3 foram registrados nos canais ativos Fz e Cz utilizando-se os contrastes de fala consonantal (/ba/-/da/ e vocálico (/i/-/a/. Desenho: Estudo de coorte, transversal e prospectivo. RESULTADOS: Houve diferença entre o contraste de fala utilizado e as latências dos componentes N2 (p = 0,00 e P3 (p = 0,00, assim como entre o canal ativo considerado (Fz/Cz e os valores de latência e amplitude de P3. Estas diferenças não ocorreram para os componentes exógenos N1 e P2. CONCLUSÃO: O contraste do estímulo de fala, vocálico ou consonantal, deve ser considerado na análise do potencial evocado cortical, componente N2, e do potencial cognitivo auditivo P3.Studies about cortical auditory evoked potentials using the speech stimuli in normal hearing individuals are important for understanding how the complexity of the stimulus influences the characteristics of the cortical potential generated. OBJECTIVE: To characterize the cortical auditory evoked potential and the P3 auditory cognitive potential with the vocalic and consonantal contrast stimuli in normally hearing individuals. METHOD: 31 individuals with no risk for hearing, neurologic and language alterations, in the age range between 7 and 30 years, participated in this study. The cortical auditory evoked potentials and the P3 auditory cognitive one were recorded in the Fz and Cz
Full Text Available OBJETIVO: investigar aspectos do histórico, hábitos e comportamentos vocais de cantores populares, conforme o sexo e as categorias profissional e amador. MÉTODO: entrevista com 47 cantores, 25 homens e 22 mulheres. RESULTADOS: significância estatística nos seguintes achados: MASCULINO - microfone nos ensaios, ausência de problemas vocais diagnosticados, ausência de orientações sobre higiene vocal, dor ou desconforto após cantar, ausência de alergias e problemas respiratórios; FEMININO - aulas de canto e conhecimento sobre postura; AMADOR - não cantar dançando, não imitar vozes, ausência de avaliação otorrinolaringológica, ausência de problemas vocais diagnosticados, ausência de terapia fonoaudiológica, ausência de orientações de anatomofisiologia vocal e não utilização de álcool nos ensaios; PROFISSIONAL - rouquidão, conhecimento sobre articulação, álcool durante os shows, "garganta suja" ou pigarro, dor após cantar. CONCLUSÕES: a comparação entre os sexos evidenciou que os homens utilizavam microfone no ensaio, não apresentavam problemas alérgicos ou respiratórios, nem problemas vocais diagnosticados, mas apresentavam sensação de dor ou desconforto após o canto e não possuíam noções sobre higiene vocal; e que as mulheres realizavam aulas de canto e possuíam orientações de postura. A comparação entre amadores e profissionais mostrou que os amadores não cantavam dançando, não imitavam vozes, não utilizavam álcool nos ensaios, e não apresentavam problemas vocais diagnosticados, mas não possuíam avaliação otorrinolaringológica, não realizavam terapia fonoaudiológica, e não possuíam conhecimento sobre anatomofisiologia vocal; e os profissionais apresentavam queixa de rouquidão, de "garganta suja" ou pigarro e de dor após cantar, e usavam álcool durante os shows, apesar de possuir conhecimento sobre articulação.PURPOSE: to investigate aspects of vocal history, vocal habits and
Full Text Available Paresis or paralysis of one or both vocal cords affects phonation, swallowing and breathing. The major cause for reduced mobility or even immobility is innervation damage, less often mechanical disorder.The main procedures in the diagnostics of disordered vocal fold mobility are indirect laryngoscopy and videoendostroboscopy. Different imaging techniques (especially computerized tomography are of great value in searching for a cause of the impaired mobility.In unilateral vocal fold immobility, the treatment is focused on the improvement of voice quality and the prevention of aspiration during swallowing. In bilateral paralysis, it is crucial to find a balance between effective breathing and sufficient voice quality. The treatment of unilateral paralysis is started with voice therapy and swallowing rehabilitation. If these procedures are not enough surgical treatment for the medialization of the paralyzed vocal fold is applied. In the case of breathing difficulties in bilateral vocal fold immobility there is a possibility of surgical lateralization of one or both folds or a surgical excision of a part or the entire vocal fold. Surgical reinnervation, functional electrostimulation of certain laryngeal muscles and gene therapy are developing treatment modalities.
Ogawa, Makoto; Inohara, Hidenori
To update our knowledge regarding the effectiveness of voice therapy for the treatment of vocal disturbance associated with benign vocal fold lesions, including vocal polyps, nodules and cysts, and for determining the utility of voice therapy in treating organic voice disorders, while highlighting problems for the future development of this clinical field. We conducted a review of the most recent literature on the therapeutic effects of voice therapy, vocal hygiene education or direct vocal training on vocal quality, the lesion appearance and discomfort felt by patients due to the clinical entity of benign vocal fold mass lesions. Although voice therapy is principally indicated for the treatment of functional dysphonia without any organic abnormalities in the vocal folds, a number of clinicians have attempted to perform voice therapy even in dysphonic patients with benign mass lesions in the vocal folds. The two major possible reasons for the effectiveness of voice therapy on vocal disturbance associated with benign vocal fold lesions are hypothesized to be the regression of lesions and the correction of excessive/inappropriate muscle contraction of the phonatory organs. According to the current literature, a substantial proportion of vocal polyps certainly tend to shrink after voice therapy, but whether or not the regression results from voice therapy, vocal hygiene education or a natural cure is unclear at present due to the lack of controlled studies comparing two groups with and without interventions. Regarding vocal nodules, no studies have investigated the effectiveness of voice therapy using proper experimental methodology. Vocal cysts are difficult to cure by voice therapy without surgical excision according to previous studies. Evidences remains insufficient to support the use of voice therapy against benign vocal fold lesions. Evidences at present is therefore still insufficient to support the use of voice therapy for the treatment of benign vocal fold
El Amine Abderrahim, Med; Breksi Reguig, Fethi
This research has been to show the realization of a morphological analyzer of the Arabic language (vocalized or not vocalized). This analyzer is based upon our object model for the Arabic Natural Language Processing (NLP) and can be exploited by NLP applications such as translation machine, orthographical correction and the search for information.
Simon, Nadine; Käthner, Ivo; Ruf, Carolin A; Pasqualotto, Emanuele; Kübler, Andrea; Halder, Sebastian
Brain-computer interfaces (BCIs) can serve as muscle independent communication aids. Persons, who are unable to control their eye muscles (e.g., in the completely locked-in state) or have severe visual impairments for other reasons, need BCI systems that do not rely on the visual modality. For this reason, BCIs that employ auditory stimuli were suggested. In this study, a multiclass BCI spelling system was implemented that uses animal voices with directional cues to code rows and columns of a letter matrix. To reveal possible training effects with the system, 11 healthy participants performed spelling tasks on 2 consecutive days. In a second step, the system was tested by a participant with amyotrophic lateral sclerosis (ALS) in two sessions. In the first session, healthy participants spelled with an average accuracy of 76% (3.29 bits/min) that increased to 90% (4.23 bits/min) on the second day. Spelling accuracy by the participant with ALS was 20% in the first and 47% in the second session. The results indicate a strong training effect for both the healthy participants and the participant with ALS. While healthy participants reached high accuracies in the first session and second session, accuracies for the participant with ALS were not sufficient for satisfactory communication in both sessions. More training sessions might be needed to improve spelling accuracies. The study demonstrated the feasibility of the auditory BCI with healthy users and stresses the importance of training with auditory multiclass BCIs, especially for potential end-users of BCI with disease.
Full Text Available Brain-computer interfaces (BCIs can serve as muscle independent communication aids. Persons, who are unable to control their eye muscles (e.g. in the completely locked-in state or have severe visual impairments for other reasons, need BCI systems that do not rely on the visual modality. For this reason, BCIs that employ auditory stimuli were suggested. In this study, a multiclass BCI spelling system was implemented that uses animal voices with directional cues to code rows and columns of a letter matrix. To reveal possible training effects with the system, 11 healthy participants performed spelling tasks on two consecutive days. In a second step, the system was tested by a participant with amyotrophic lateral sclerosis (ALS in two sessions. In the first session, healthy participants spelled with an average accuracy of 76% (3.29 bits/min that increased to 90% (4.23 bits/min on the second day. Spelling accuracy by the participant with ALS was 20% in the first and 47% in the second session. The results indicate a strong training effect for both the healthy participants and the participant with ALS. While healthy participants reached high accuracies in the first session and second session, accuracies for the participant with ALS were not sufficient for satisfactory communication in both sessions. More training sessions might be needed to improve spelling accuracies. The study demonstrated the feasibility of the auditory BCI with healthy users and stresses the importance of training with auditory multiclass BCIs, especially for potential end-users of BCI with disease.
Kraus, Nina; Strait, Dana; Parbery-Clark, Alexandra
Musicians benefit from real-life advantages such as a greater ability to hear speech in noise and to remember sounds, although the biological mechanisms driving such advantages remain undetermined. Furthermore, the extent to which these advantages are a consequence of musical training or innate characteristics that predispose a given individual to pursue music training is often debated. Here, we examine biological underpinnings of musicians’ auditory advantages and the mediating role of auditory working memory. Results from our laboratory are presented within a framework that emphasizes auditory working memory as a major factor in the neural processing of sound. Within this framework, we provide evidence for music training as a contributing source of these abilities. PMID:22524346
technology for recording and stimulating from the auditory and olfactory sensory nervous systems of the awake, swimming nurse shark , G. cirratum (Figures...overlay of the central nervous system of the nurse shark on a horizontal MR image. Implantable Neural Interfaces for Sharks ...Neural Interfaces for Characterizing Population Responses to Odorants and Electrical Stimuli in the Nurse Shark , Ginglymostoma cirratum.” AChemS Abs
Noyes, Blakeslee E; Kemp, James S
Vocal cord dysfunction is characterised by paradoxical vocal cord adduction that occurs during inspiration, resulting in symptoms of dyspnoea, wheeze, chest or throat tightness and cough. Although the condition is well described in children and adults, confusion with asthma often triggers the use of an aggressive treatment regimen directed against asthma. The laryngoscopic demonstration of vocal cord adduction during inspiration has been considered the gold standard for the diagnosis of vocal cord dysfunction, but historical factors and pulmonary function findings may provide adequate clues to the correct diagnosis. Speech therapy, and in some cases psychological counselling, is often beneficial in this disorder. The natural course and prognosis of vocal cord dysfunction are still not well described in adults or children.
Demirci, Sule; Tuzuner, Arzu; Callıoglu, Elif Ersoy; Yumusak, Nihat; Arslan, Necmi; Baltacı, Bülent
The aim of this study was to investigate the use of glass ionomer cement (GIC) as an injection material for vocal fold augmentation and to evaluate the biocompatibility of the material. Ten adult New Zealand rabbits were used. Under general anesthesia, 0.1-cc GIC was injected to one vocal fold and the augmentation of vocal fold was observed. No injection was applied to the opposite side, which was accepted as the control group. The animals were sacrificed after 3 months and the laryngeal specimens were histopathologically evaluated. The injected and the noninjected control vocal folds were analyzed. The GIC particles were observed in histological sections on the injected side, and no foreign body giant cells, granulomatous inflammation, necrosis, or marked chronic inflammation were detected around the glass ionomer particles. Mild inflammatory reactions were noticed in only two specimens. The noninjected sides of vocal folds were completely normal. The findings of this study suggest that GIC is biocompatible and may be further investigated as an alternative injection material for augmentation of the vocal fold. Further studies are required to examine the viscoelastic properties of GIC and the long-term effects in experimental studies. NA. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.
Brittan-Powell, Elizabeth F; Christensen-Dalsgaard, Jakob; Tang, Yezhong; Carr, Catherine; Dooling, Robert J
Although lizards have highly sensitive ears, it is difficult to condition them to sound, making standard psychophysical assays of hearing sensitivity impractical. This paper describes non-invasive measurements of the auditory brainstem response (ABR) in both Tokay geckos (Gekko gecko; nocturnal animals, known for their loud vocalizations) and the green anole (Anolis carolinensis, diurnal, non-vocal animals). Hearing sensitivity was measured in 5 geckos and 7 anoles. The lizards were sedated with isoflurane, and ABRs were measured at levels of 1 and 3% isoflurane. The typical ABR waveform in response to click stimulation showed one prominent and several smaller peaks occurring within 10 ms of the stimulus onset. ABRs to brief tone bursts revealed that geckos and anoles were most sensitive between 1.6-2 kHz and had similar hearing sensitivity up to about 5 kHz (thresholds typically 20-50 dB SPL). Above 5 kHz, however, anoles were more than 20 dB more sensitive than geckos and showed a wider range of sensitivity (1-7 kHz). Generally, thresholds from ABR audiograms were comparable to those of small birds. Best hearing sensitivity, however, extended over a larger frequency range in lizards than in most bird species.
Tomé, David; Sampaio, Mafalda; Mendes-Ribeiro, José; Barbosa, Fernando; Marques-Teixeira, João
Benign focal epilepsy in childhood with centro-temporal spikes (BECTS) is one of the most common forms of idiopathic epilepsy, with onset from age 3 to 14 years. Although the prognosis for children with BECTS is excellent, some studies have revealed neuropsychological deficits in many domains, including language. Auditory event-related potentials (AERPs) reflect activation of different neuronal populations and are suggested to contribute to the evaluation of auditory discrimination (N1), attention allocation and phonological categorization (N2), and echoic memory (mismatch negativity--MMN). The scarce existing literature about this theme motivated the present study, which aims to investigate and document the existing AERP changes in a group of children with BECTS. AERPs were recorded, during the day, to pure and vocal tones and in a conventional auditory oddball paradigm in five children with BECTS (aged 8-12; mean=10 years; male=5) and in six gender and age-matched controls. Results revealed high amplitude of AERPs for the group of children with BECTS with a slight latency delay more pronounced in fronto-central electrodes. Children with BECTS may have abnormal central auditory processing, reflected by electrophysiological measures such as AERPs. In advance, AERPs seem a good tool to detect and reliably reveal cortical excitability in children with typical BECTS. Copyright © 2014 Elsevier B.V. All rights reserved.
Brown, Rachel M; Palmer, Caroline
In two experiments, we investigated how auditory-motor learning influences performers' memory for music. Skilled pianists learned novel melodies in four conditions: auditory only (listening), motor only (performing without sound), strongly coupled auditory-motor (normal performance), and weakly coupled auditory-motor (performing along with auditory recordings). Pianists' recognition of the learned melodies was better following auditory-only or auditory-motor (weakly coupled and strongly coupled) learning than following motor-only learning, and better following strongly coupled auditory-motor learning than following auditory-only learning. Auditory and motor imagery abilities modulated the learning effects: Pianists with high auditory imagery scores had better recognition following motor-only learning, suggesting that auditory imagery compensated for missing auditory feedback at the learning stage. Experiment 2 replicated the findings of Experiment 1 with melodies that contained greater variation in acoustic features. Melodies that were slower and less variable in tempo and intensity were remembered better following weakly coupled auditory-motor learning. These findings suggest that motor learning can aid performers' auditory recognition of music beyond auditory learning alone, and that motor learning is influenced by individual abilities in mental imagery and by variation in acoustic features.
Eliane Cristina Pereira
Full Text Available OBJETIVO: verificar o efeito imediato das técnicas vocais vibração, som nasal e sobrearticulação na voz e na laringe de mulheres sem queixas vocais. MÉTODO: participaram da pesquisa 32 sujeitos do sexo feminino, com idades entre 20 e 45 anos, sem queixas vocais, com qualidade vocal avaliada entre normal e alteração de grau leve Os sujeitos foram submetidos à análise perceptivo-auditiva pela escala visual analógica da vogal /ε/ e fala espontânea, análise acústica e laringoestroboscopia antes e após a realização das técnicas. RESULTADOS: a análise perceptivo-auditiva revelou melhora significante dos parâmetros impressão global da voz, rouquidão e estabilidade na vogal /ε/ e articulação na fala espontânea. A análise acústica evidenciou melhora significante do jitter e shimmer. A laringoestroboscopia evidenciou significante melhora no fechamento glótico e melhora na movimentação muco-ondulatória das pregas vocais. CONCLUSÃO: as técnicas vocais estudadas são capazes de proporcionar melhora imediata significante da qualidade vocal e da configuração laríngea.PURPOSE: to check the immediate effect of vocal techniques: vibration, nasal sound and overarticulation. METHOD: 32 female subjects with normal to mild dysphonia took part in the study, with ages from 20 to 45 years. Subjects were submitted to perceptual analysis and laryngostroboscopic exams before and after the use of vocal techniques. RESULTS: subjects' vocal classification in perceptual analysis after accomplishing the vocal techniques showed significant improvement on parameters voice global impression, hoarseness and stability; and, in spontaneous speech, one showed a significant improvement on the parameter articulation. The acoustic analysis evidenced significant improvement of the jitter and shimmer. Laryngostroboscopic examination evidenced a significant increase in the glottic closing and an increase in the mucondulatory movement of the vocal folds
Hunter, Eric J; Banks, Russell E
Occupational voice users report higher instances of vocal health problems. Women, who are more likely than men to report voice problems, are the largest members of some occupational voice users, such as teachers. While a common complaint among this population is vocal fatigue, it has been difficult to quantify. Therefore, the goal of this study is to quantify vocal fatigue generally in school teachers and investigate any related gender differences. Six hundred forty (518 female, 122 male) teachers were surveyed using an online questionnaire consisting in part of the Vocal Fatigue Index (VFI), an index specifically designed to quantify vocal fatigue. Compared to vocally healthy adults, the teachers surveyed were 3 times as likely to report vocal tiredness or vocal avoidance and over 3 times as likely to report physical voice discomfort. Additionally, female teachers were more likely to have scores approaching those with dysphonia. The VFI quantified elevated levels of vocal fatigue in teachers, with a significant prevalence of symptoms reported among females compared to males. Further, because the VFI indicated elevated complaints (between normal and dysphonic) in a population likely to be elevated, the VFI might be used to identify early indications of voice problems and/or track recovery.
Full Text Available Introduction. Hyperkinetic (hyperfunctional dysphonia is a common pathology. The disorder is often found in vocal professionals faced with high vocal requirements. Objective. The objective of this study was to evaluate the effects of vocal therapy on voice condition characterized by hyperkinetic dysphonia with prenodular lesions and soft nodules. Methods. The study included 100 adult patients and 27 children aged 4-16 years with prenodular lesions and soft nodules. A subjective acoustic analysis using the GIRBAS scale was performed prior to and after vocal therapy. Twenty adult patients and 10 children underwent objective acoustic analysis including several acoustic parameters. Pathological vocal qualities (hoarse, harsh and breathy voice were also obtained by computer analysis. Results. The subjective acoustic analysis revealed a significant (p<0.01 reduction in all dysphonia parameters after vocal treatment in adults and children. After treatment, all levels of dysphonia were lowered in 85% (85/100 of adult patients and 29% (29/100 had a normal voice. Before vocal therapy 9 children had severe, 13 had moderate and 8 slight dysphonia. After vocal therapy only 1 child had severe dysphonia, 7 had moderate, 10 had slight levels of dysphonia and 9 were without voice disorder. The objective acoustic analysis in adults revealed a significant improvement (p≤0.025 in all dysphonia parameters except SD F0 and jitter %. In children, the acoustic parameters SD F0, jitter % and NNE (normal noise energy were significantly improved (p=0.003-0.03. Pathological voice qualities were also improved in adults and children (p<0.05. Conclusion. Vocal therapy effectively improves the voice in hyperkinetic dysphonia with prenodular lesions and soft nodules in both adults and children, affecting diverse acoustic parameters.
Uetsuki, Shizuka; Kinoshita, Hiroshi; Takahashi, Ryuichi; Obata, Satoshi; Kakigi, Tatsuya; Wada, Yoshiko; Yokoyama, Kazumasa
A 53-year-old right-handed woman had an extensive lesion in the left hemisphere due to an infarction caused by vasospasm secondary to subarachnoid bleeding. She exhibited persistent expressive-vocal amusia with no symptoms of aphasia. Evaluation of the patient's musical competence using the Montreal Battery for Evaluation of Amusia, rhythm reproduction tests, acoustic analysis of pitch upon singing familiar music, Japanese standard language tests, and other detailed clinical examinations revealed that her amusia was more dominantly related to pitch production. The intactness of her speech provided strong evidence that the right hemisphere played a major role in her linguistic processing. Data from functional magnetic resonance imaging while she was singing a familiar song, a scale, and reciting lyrics indicated that perilesional residual activation in the left hemisphere was associated with poor pitch production, while right hemispheric activation was involved in linguistic processing. The localization of infarction more anterior to the left Sylvian fissure might be related to the dominant deficits in expressive aspects of the singing of the patient. Compromised motor programming producing a single tone may have made a major contribution to her poor singing. Imperfect auditory feedback due to borderline perceptual ability or improper audio-motor associations might also have played a role. Copyright © 2016 Elsevier Inc. All rights reserved.
Begault, Durand R.; Wenzel, Elizabeth M.
The implementation of binaural sound to speech and auditory sound cues (auditory icons) is addressed from both an applications and technical standpoint. Techniques overviewed include processing by means of filtering with head-related transfer functions. Application to advanced cockpit human interface systems is discussed, although the techniques are extendable to any human-machine interface. Research issues pertaining to three-dimensional sound displays under investigation at the Aerospace Human Factors Division at NASA Ames Research Center are described.
Mumović, Gordana; Veselinović, Mila; Arbutina, Tanja; Škrbić, Renata
Hyperkinetic (hyperfunctional) dysphonia is a common pathology. The disorder is often found in vocal professionals faced with high vocal requirements. The objective of this study was to evaluate the effects of vocal therapy on voice condition characterized by hyperkinetic dysphonia with prenodular lesions and soft nodules. The study included 100 adult patients and 27 children aged 4-16 years with prenodular lesions and soft nodules. A subjective acoustic analysis using the GIRBAS scale was performed prior to and after vocal therapy. Twenty adult patients and 10 children underwent objective acoustic analysis including several acoustic parameters. Pathological vocal qualities (hoarse, harsh and breathy voice) were also obtained by computer analysis. The subjective acoustic analysis revealed a significant (pvocal treatment in adults and children. After treatment, all levels of dysphonia were lowered in 85% (85/100) of adult patients and 29% (29/100) had a normal voice. Before vocal therapy 9 children had severe, 13 had moderate and 8 slight dysphonia. After vocal therapy only 1 child had severe dysphonia, 7 had moderate, 10 had slight levels of dysphonia and 9 were without voice disorder. The objective acoustic analysis in adults revealed a significant improvement (p≤0.025) in all dysphonia parameters except SD FO and jitter %. In children, the acoustic parameters SD FO, jitter % and NNE (normal noise energy) were significantly improved (p=0.003-0.03). Pathological voice qualities were also improved in adults and children (pVocal therapy effectively improves the voice in hyperkinetic dysphonia with prenodular lesions and soft nodules in both adults and children, affectinq diverse acoustic parameters.
Liszkowski, Ulf; Albrecht, Konstanze; Carpenter, Malinda; Tomasello, Michael
In the current study we investigated infants' communication in the visual and auditory modalities as a function of the recipient's visual attention. We elicited pointing at interesting events from thirty-two 12-month olds and thirty-two 18-month olds in two conditions: when the recipient either was or was not visually attending to them before and during the point. The main result was that infants initiated more pointing when the recipient's visual attention was on them than when it was not. In addition, when the recipient did not respond by sharing interest in the designated event, infants initiated more repairs (repeated pointing) than when she did, again, especially when the recipient was visually attending to them. Interestingly, accompanying vocalizations were used intentionally and increased in both experimental conditions when the recipient did not share attention and interest. However, there was little evidence that infants used their vocalizations to direct attention to their gestures when the recipient was not attending to them.
Análise de características vocais e de aspectos psicológicos em indivíduos com transtorno obsessivo-compulsivo Analysis of vocal characteristics and psychological aspects in individuals with obsessive-compulsive disorder
Full Text Available OBJETIVO: Avaliar a auto-imagem vocal e caracterizar auditiva e acusticamente as vozes de sujeitos com Transtorno Obsessivo Compulsivo (TOC, comparadas a um grupo controle sem queixas psiquiátricas e vocais, além de analisar aspectos psicológicos que possam estar envolvidos nas questões vocais avaliadas. MÉTODOS: Constituiu-se uma amostra de 35 indivíduos, homens e mulheres, com idades entre 16 e 74 anos, sendo 17 com TOC e 18 casos controle. Todos os pesquisados foram submetidos aos seguintes protocolos de pesquisa: Escala de Depressão de Beck, Escala de Ansiedade de Beck e o protocolo para a caracterização da voz de sujeitos com manifestações psiquiátricas. Os sujeitos responderam ainda ao questionário de análise psicodinâmica da voz com enfoque na auto-imagem vocal e foram submetidos à avaliação perceptivo-auditiva e análise acústica da voz. RESULTADOS: Na análise da auto-imagem vocal, os aspectos significativos descritos pelo grupo clínico foram as características vocais triste e ruim. Na avaliação perceptivo-auditiva, houve predomínio do tipo voz rouco-soprosa em grau leve, alterações na ressonância, velocidade de fala e na modulação e entonação. Houve diferenças entre os valores de jitter e shimmer. Não foram observados valores fora do padrão de normalidade com relação ao tremor, nem diferenças entre os grupos quanto aos valores de frequência fundamental. CONCLUSÃO: Foi possível compreender a percepção do indivíduo com TOC sobre sua própria voz e os desvios na emissão vocal. Assim, o fonoaudiólogo pode obter informações que permitam melhorar a qualidade de vida destes por meio de uma intervenção fonoaudiológica, visando, também, a interdisciplinaridade.PURPOSE: To assess the vocal self-image and to characterize auditory and acoustic aspects of the voices of individuals with Obsessive-Compulsive Disorder (OCD, compared to a control group without psychiatric and vocal complaints, in
Anikin, Andrey; Bååth, Rasmus; Persson, Tomas
Recent research on human nonverbal vocalizations has led to considerable progress in our understanding of vocal communication of emotion. However, in contrast to studies of animal vocalizations, this research has focused mainly on the emotional interpretation of such signals. The repertoire of human nonverbal vocalizations as acoustic types, and the mapping between acoustic and emotional categories, thus remain underexplored. In a cross-linguistic naming task (Experiment 1), verbal categorization of 132 authentic (non-acted) human vocalizations by English-, Swedish- and Russian-speaking participants revealed the same major acoustic types: laugh, cry, scream, moan, and possibly roar and sigh. The association between call type and perceived emotion was systematic but non-redundant: listeners associated every call type with a limited, but in some cases relatively wide, range of emotions. The speed and consistency of naming the call type predicted the speed and consistency of inferring the caller's emotion, suggesting that acoustic and emotional categorizations are closely related. However, participants preferred to name the call type before naming the emotion. Furthermore, nonverbal categorization of the same stimuli in a triad classification task (Experiment 2) was more compatible with classification by call type than by emotion, indicating the former's greater perceptual salience. These results suggest that acoustic categorization may precede attribution of emotion, highlighting the need to distinguish between the overt form of nonverbal signals and their interpretation by the perceiver. Both within- and between-call acoustic variation can then be modeled explicitly, bringing research on human nonverbal vocalizations more in line with the work on animal communication.
Full Text Available The aim of this study is to survey and compare the development of auditory skills in young children with Mondini dysplasia and profoundly-deaf young children with radiologically normal inner ears over a period of 3 years after cochlear implantation. A total of 545 young children (age 7 to 36 months with prelingual, severe to profound hearing loss participated in this study. All children received cochlear implantation. Based on whether or not there was a Mondini dysplasia as diagnosed with CT scanning, the subjects were divided into 2 groups: (A 514 young children with radiologically normal inner ears and (B 31 young children with Mondini dysplasia. The Infant-Toddler Meaningful Auditory Integration Scale (IT-MAIS was used to assess the children's auditory skills that include vocalization changes, spontaneous alerting to sounds in everyday living environments, and the ability to derive meaning from sounds. The assessment was performed prior to surgery and at 1, 3, 6, 9, 12, 24, and 36 months after implant device switch-on. The mean scores for overall auditory skills were not significantly different between groups A and B at pre-surgery, 1, 12, 24, and 36 months post-surgery, but were significantly different at 3, 6, and 9 months post-surgery. The mean scores for all auditory skills in children with Mondini dysplasia showed significant improvement over time. The mean scores for the three subcategories of auditory skills in children with Mondini dysplasia also showed significant differences at pre-surgery, 1, 3, 6, and 9 months, however, there were no significant differences at 12, 24, and 36 months. Overall, the auditory skills of young children with Mondini dysplasia developed rapidly after cochlear implantation, in a similar manner to that of young children with radiologically normal inner ears. Cochlear implantation is an effective intervention for young children with Mondini dysplasia.
Rigoulot, S; Pell, M D; Armony, J L
Previous functional magnetic resonance imaging (fMRI) studies have suggested that different cerebral regions preferentially process human voice and music. Yet, little is known on the temporal course of the brain processes that decode the category of sounds and how the expertise in one sound category can impact these processes. To address this question, we recorded the electroencephalogram (EEG) of 15 musicians and 18 non-musicians while they were listening to short musical excerpts (piano and violin) and vocal stimuli (speech and non-linguistic vocalizations). The task of the participants was to detect noise targets embedded within the stream of sounds. Event-related potentials revealed an early differentiation of sound category, within the first 100 ms after the onset of the sound, with mostly increased responses to musical sounds. Importantly, this effect was modulated by the musical background of participants, as musicians were more responsive to music sounds than non-musicians, consistent with the notion that musical training increases sensitivity to music. In late temporal windows, brain responses were enhanced in response to vocal stimuli, but musicians were still more responsive to music. These results shed new light on the temporal course of neural dynamics of auditory processing and reveal how it is impacted by the stimulus category and the expertise of participants. Copyright © 2015 IBRO. Published by Elsevier Ltd. All rights reserved.
Stevens, Kimberly A; Thomson, Scott L; Jetté, Marie E; Thibeault, Susan L
The aim of this study was to quantify porcine vocal fold medial surface geometry and three-dimensional geometric distortion induced by freezing the larynx, especially in the region of the vocal folds. The medial surface geometries of five excised porcine larynges were quantified and reported. Five porcine larynges were imaged in a micro-CT scanner, frozen, and rescanned. Segmentations and three-dimensional reconstructions were used to quantify and characterize geometric features. Comparisons were made with geometry data previously obtained using canine and human vocal folds as well as geometries of selected synthetic vocal fold models. Freezing induced an overall expansion of approximately 5% in the transverse plane and comparable levels of nonuniform distortion in sagittal and coronal planes. The medial surface of the porcine vocal folds was found to compare reasonably well with other geometries, although the compared geometries exhibited a notable discrepancy with one set of published human female vocal fold geometry. Porcine vocal folds are qualitatively geometrically similar to data available for canine and human vocal folds, as well as commonly used models. Freezing of tissue in the larynx causes distortion of around 5%. The data can provide direction in estimating uncertainty due to bulk distortion of tissue caused by freezing, as well as quantitative geometric data that can be directly used in developing vocal fold models. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Plakke, Bethany; Diltz, Mark D; Romanski, Lizabeth M
Neuronal activity in single prefrontal neurons has been correlated with behavioral responses, rules, task variables and stimulus features. In the non-human primate, neurons recorded in ventrolateral prefrontal cortex (VLPFC) have been found to respond to species-specific vocalizations. Previous studies have found multisensory neurons which respond to simultaneously presented faces and vocalizations in this region. Behavioral data suggests that face and vocal information are inextricably linked in animals and humans and therefore may also be tightly linked in the coding of communication calls in prefrontal neurons. In this study we therefore examined the role of VLPFC in encoding vocalization call type information. Specifically, we examined previously recorded single unit responses from the VLPFC in awake, behaving rhesus macaques in response to 3 types of species-specific vocalizations made by 3 individual callers. Analysis of responses by vocalization call type and caller identity showed that ∼19% of cells had a main effect of call type with fewer cells encoding caller. Classification performance of VLPFC neurons was ∼42% averaged across the population. When assessed at discrete time bins, classification performance reached 70 percent for coos in the first 300 ms and remained above chance for the duration of the response period, though performance was lower for other call types. In light of the sub-optimal classification performance of the majority of VLPFC neurons when only vocal information is present, and the recent evidence that most VLPFC neurons are multisensory, the potential enhancement of classification with the addition of accompanying face information is discussed and additional studies recommended. Behavioral and neuronal evidence has shown a considerable benefit in recognition and memory performance when faces and voices are presented simultaneously. In the natural environment both facial and vocalization information is present simultaneously and
Saidi, Hiba; Erath, Byron D.
The vocal folds play a major role in human communication by initiating voiced sound production. During voiced speech, the vocal folds are set into sustained vibrations. Synthetic self-oscillating vocal fold models are regularly employed to gain insight into flow-structure interactions governing the phonation process. Commonly, a fixed boundary condition is applied to the lateral, anterior, and posterior sides of the synthetic vocal fold models. However, physiological observations reveal the presence of adipose tissue on the lateral surface between the thyroid cartilage and the vocal folds. The goal of this study is to investigate the influence of including this substrate layer of adipose tissue on the dynamics of phonation. For a more realistic representation of the human vocal folds, synthetic multi-layer vocal fold models have been fabricated and tested while including a soft lateral layer representative of adipose tissue. Phonation parameters have been collected and are compared to those of the standard vocal fold models. Results show that vocal fold kinematics are affected by adding the adipose tissue layer as a new boundary condition.
Pell, Marc D.; Kotz, Sonja A.
How quickly do listeners recognize emotions from a speaker's voice, and does the time course for recognition vary by emotion type? To address these questions, we adapted the auditory gating paradigm to estimate how much vocal information is needed for listeners to categorize five basic emotions (anger, disgust, fear, sadness, happiness) and neutral utterances produced by male and female speakers of English. Semantically-anomalous pseudo-utterances (e.g., The rivix jolled the silling) conveying each emotion were divided into seven gate intervals according to the number of syllables that listeners heard from sentence onset. Participants (n = 48) judged the emotional meaning of stimuli presented at each gate duration interval, in a successive, blocked presentation format. Analyses looked at how recognition of each emotion evolves as an utterance unfolds and estimated the “identification point” for each emotion. Results showed that anger, sadness, fear, and neutral expressions are recognized more accurately at short gate intervals than happiness, and particularly disgust; however, as speech unfolds, recognition of happiness improves significantly towards the end of the utterance (and fear is recognized more accurately than other emotions). When the gate associated with the emotion identification point of each stimulus was calculated, data indicated that fear (M = 517 ms), sadness (M = 576 ms), and neutral (M = 510 ms) expressions were identified from shorter acoustic events than the other emotions. These data reveal differences in the underlying time course for conscious recognition of basic emotions from vocal expressions, which should be accounted for in studies of emotional speech processing. PMID:22087275
Klein, Travis A L; Gaziano, Joy E; Ridley, Marion B
A unique case of acute onset vocal fold paralysis secondary to phonotrauma is presented. The cause was forceful vocalization by a drill instructor on a firearm range. Imaging studies revealed extensive intralaryngeal and retropharyngeal hemorrhage. Laryngoscopy showed a complete left vocal fold paralysis. Relative voice rest was recommended, and the patient regained normal vocal fold mobility and function after approximately 12 weeks. Copyright © 2014 The Voice Foundation. All rights reserved.
Perlman, Marcus; Dale, Rick; Lupyan, Gary
Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative 'vocal' charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic 'meaning template' we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems.
Meeuwisse, Marieke; de Meijer, Lonneke A.; Born, Marise Ph.; Severiens, Sabine E.
Given the poorer academic outcomes of non-Western ethnic minority students compared to ethnic majority students, we investigated whether differences exist in work-study interface between ethnic groups. We tested a work-study interface model, in which the work-related factors work-study congruence, job control, job demands, work hours, job…
Maamary, Joel A; Cole, Ian; Darveniza, Paul; Pemberton, Cecilia; Brake, Helen Mary; Tisch, Stephen
The objective of this study was to better define the relationship of laryngeal electromyography and video laryngostroboscopy in the diagnosis of vocal fold paralysis. Retrospective diagnostic cohort study with cross-sectional data analysis METHODS: Data were obtained from 57 patients with unilateral vocal fold paralysis who attended a large tertiary voice referral center. Electromyographic findings were classified according to recurrent laryngeal nerve, superior laryngeal nerve, and high vagal/combined lesions. Video laryngostroboscopy recordings were classified according to the position of the immobile fold into median, paramedian, lateral, and a foreshortened/hooded vocal fold. The position of the paralyzed vocal fold was then analyzed according to the lesion as determined by electromyography. The recurrent laryngeal nerve was affected in the majority of cases with left-sided lesions more common than right. Vocal fold position differed between recurrent laryngeal and combined vagal lesions. Recurrent laryngeal nerve lesions were more commonly associated with a laterally displaced immobile fold. No fold position was suggestive of a combined vagal lesion. The inter-rater reliability for determining fold position was high. Laryngeal electromyography is useful in diagnosing neuromuscular dysfunction of the larynx and best practice recommends its continued implementation along with laryngostroboscopy. While recurrent laryngeal nerve lesions are more likely to present with a lateral vocal fold, this does not occur in all cases. Such findings indicate that further unknown mechanisms contribute to fold position in unilateral paralysis. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Mumović Gordana; Veselinović Mila; Arbutina Tanja; Škrbić Renata
Introduction. Hyperkinetic (hyperfunctional) dysphonia is a common pathology. The disorder is often found in vocal professionals faced with high vocal requirements. Objective. The objective of this study was to evaluate the effects of vocal therapy on voice condition characterized by hyperkinetic dysphonia with prenodular lesions and soft nodules. Methods. The study included 100 adult patients and 27 children aged 4-16 years with prenodular lesions and soft...
Valentino, Amber L.; Shillingsburg, M. Alice; Call, Nathan A.; Burton, Britney; Bowen, Crystal N.
Children with autism have significant communication delays. Although some children develop vocalizations through shaping and differential reinforcement, others rarely exhibit vocalizations, and alternative methods are targeted in intervention. However, vocal language often remains a goal for caregivers and clinicians. Thus, strategies to increase…
Anastasia H Dalziell
Full Text Available Some of the most striking vocalizations in birds are made by males that incorporate vocal mimicry in their sexual displays. Mimetic vocalization in females is largely undescribed, but it is unclear whether this is because of a lack of selection for vocal mimicry in females, or whether the phenomenon has simply been overlooked. These issues are thrown into sharp relief in the superb lyrebird, Menura novaehollandiae, a basal oscine passerine with a lek-like mating system and female uniparental care. The spectacular mimetic song display produced by courting male lyrebirds is a textbook example of a sexually selected trait, but the vocalizations of female lyrebirds are largely unknown. Here, we provide the first analysis of the structure and context of the vocalizations of female lyrebirds. Female lyrebirds were completely silent during courtship; however, females regularly produced sophisticated vocal displays incorporating both lyrebird-specific vocalizations and imitations of sounds within their environment. The structure of female vocalizations varied significantly with context. While foraging, females mostly produced a complex lyrebird-specific song, whereas they gave lyrebird-specific alarm calls most often during nest defense. Within their vocal displays females also included a variety of mimetic vocalizations, including imitations of the calls of dangerous predators, and of alarm calls and song of harmless heterospecifics. Females gave more mimetic vocalizations during nest defense than while foraging, and the types of sounds they imitated varied between these contexts, suggesting that mimetic vocalizations have more than one function. These results are inconsistent with previous portrayals of vocalizations by female lyrebirds as rare, functionless by-products of sexual selection on males. Instead, our results support the hypotheses that complex female vocalizations play a role in nest defense and mediate female-female competition for
Ratanasit, Dan; Moore, Melody M.
The inability of computer users who are visually impaired to access graphical user interfaces (GUIs) has led researchers to propose approaches for adapting GUIs to auditory interfaces, with the goal of providing access for visually impaired people. This article outlines the issues involved in nonvisual access to graphical user interfaces, reviews…
Boltežar, Lučka; Šereg Bahar, Maja
The aim of this paper is to compare the prevalence of voice disorders and the risk factors for them in different occupations with a vocal load in Slovenia. A meta-analysis of six different Slovenian studies involving teachers, physicians, salespeople, catholic priests, nurses and speech-and-language therapists (SLTs) was performed. In all six studies, similar questions about the prevalence of voice disorders and the causes for them were included. The comparison of the six studies showed that more than 82% of the 2347 included subjects had voice problems at some time during their career. The teachers were the most affected by voice problems. The prevalent cause of voice problems was the vocal load in teachers and salespeople and respiratory-tract infections in all the other occupational groups. When the occupational groups were compared, it was stated that the teachers had more voice problems and showed less care for their voices than the priests. The physicians had more voice problems and showed better consideration of vocal hygiene rules than the SLTs. The majority of all the included subjects did not receive instructions about voice care during education. In order to decrease the prevalence of voice disorders in vocal professionals, a screening program is recommended before the beginning of their studies. Regular courses on voice care and proper vocal technique should be obligatory for all professional voice users during their career. The inclusion of dysphonia in the list of occupational diseases should be considered in Slovenia as it is in some European countries.
Kobayashi, Noriko; Hirose, Hajime; Nishiyama, Koichiro
For the treatment of vocal nodules, educational programs for vocal hygiene and voice training for acquisition of correct phonation are essential. In the case of children, special considerations are necessary as some of their vocal behaviors and reaction to voice disorders are different from those of adults. In this study, a voice therapy program for child vocal nodules were developed and good results were obtained for six children. They were four boys and two girls (Age: 4-11 yr) and bilateral nodules were found for all of them. In addition to a conventional vocal hygiene program for children, correct production of loud voice (so-called gBeltingh) was the major focus of the voice therapy as the visual inspection of the larynges and perceptual evaluations of the voice revealed inappropriate loud voice production with laryngeal constriction in all children. After 5-24 voice therapy sessions, disappearance of the nodules was found in five children and the reduction of the nodule sizes was found in one child. Improvement of the GRBAS scores, longer maximum phonation time, and extension of vocal ranges were found after the completion of the therapy programs.
Martins, Regina Helena Garcia; Santana, Marcela Ferreira; Tavares, Elaine Lara Mendes
Vocal cysts are benign laryngeal lesions, which affect children and adults. They can be classified as epidermic or mucous-retention cyst. The objective was to study the clinical, endoscopic, and surgical aspects of vocal cysts. We reviewed the medical charts of 72 patients with vocal cysts, considering age, gender, occupation, time of vocal symptoms, nasosinusal and gastroesophageal symptoms, vocal abuse, tabagism, alcoholism, associated lesions, treatment, and histological details. Of the 72 cases, 46 were adults (36 females and 10 male) and 26 were children (eight girls and 18 boys). As far as occupation is concerned, there was a higher incidence of students and teachers. All the patients had symptoms of chronic hoarseness. Nasosinusal (27.77%) and gastroesophageal (32%) symptoms were not relevant. Vocal abuse was reported by 45.83%, smoking by 18%, and alcoholism by 8.4% of the patients. Unilateral cysts were seen in 93% of the cases, 22 patients had associated lesions, such as bridge, sulcus vocalis, and microweb. Surgical treatment was performed in 46 cases. Histological analysis of the epidermic cysts revealed a cavity with caseous content, covered by stratified squamous epithelium, often keratinized. Mucous cysts presented mucous content, and the walls were coated by a cylindrical ciliated epithelium. Vocal cysts are benign vocal fold lesions that affect children and adults, being often associated with vocal overuse, which frequently affects people who use their voices professionally. Vocal symptoms are chronic in course, often times since childhood, and the treatment of choice is surgical removal. A careful examination of the vocal folds is necessary during surgery, because other laryngeal lesions may be associated with vocal cysts. Copyright Â© 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Full Text Available To date, most speech synthesis techniques have relied upon the representation of the vocal tract by some form of filter, a typical example being linear predictive coding (LPC. This paper describes the development of a physiologically realistic model of the vocal tract using the well-established technique of transmission line modelling (TLM. This technique is based on the principle of wave scattering at transmission line segment boundaries and may be used in one, two, or three dimensions. This work uses this technique to model the vocal tract using a one-dimensional transmission line. A six-port scattering node is applied in the region separating the pharyngeal, oral, and the nasal parts of the vocal tract.
Martins, H R; Romao, M; Placido, D; Provenzano, F; Tierra-Criollo, C J
The technological improvement helps many medical areas. The audiometric exams involving the auditory evoked potentials can make better diagnoses of auditory disorders. This paper proposes the development of a stimulator based on Digital Signal Processor. This stimulator is the first step of an auditory evoked potential system based on the ADSP-BF533 EZ KIT LITE (Analog Devices Company - USA). The stimulator can generate arbitrary waveform like Sine Waves, Modulated Amplitude, Pulses, Bursts and Pips. The waveforms are generated through a graphical interface programmed in C++ in which the user can define the parameters of the waveform. Furthermore, the user can set the exam parameters as number of stimuli, time with stimulation (Time ON) and time without stimulus (Time OFF). In future works will be implemented another parts of the system that includes the acquirement of electroencephalogram and signal processing to estimate and analyze the evoked potential
Martins, H R; Romao, M; Placido, D; Provenzano, F; Tierra-Criollo, C J [Universidade Federal de Minas Gerais (UFMG), Departamento de Engenharia Eletrica (DEE), Nucleo de Estudos e Pesquisa em Engenharia Biomedica NEPEB, Av. Ant. Carlos, 6627, sala 2206, Pampulha, Belo Horizonte, MG, 31.270-901 (Brazil)
The technological improvement helps many medical areas. The audiometric exams involving the auditory evoked potentials can make better diagnoses of auditory disorders. This paper proposes the development of a stimulator based on Digital Signal Processor. This stimulator is the first step of an auditory evoked potential system based on the ADSP-BF533 EZ KIT LITE (Analog Devices Company - USA). The stimulator can generate arbitrary waveform like Sine Waves, Modulated Amplitude, Pulses, Bursts and Pips. The waveforms are generated through a graphical interface programmed in C++ in which the user can define the parameters of the waveform. Furthermore, the user can set the exam parameters as number of stimuli, time with stimulation (Time ON) and time without stimulus (Time OFF). In future works will be implemented another parts of the system that includes the acquirement of electroencephalogram and signal processing to estimate and analyze the evoked potential.
Full Text Available Background/Aim. An excessive use or misuse of voice by vocal professionals may result in symptoms such are husky voice, hoarse voice, total loss of voice, or even organic changes taking place on vocal folds - minimal pathological lesions - MAPLs. The purpose of this study was to identify the type of MAPLs which affects vocal professionals, as well as to identify the risk factors that bring about these changes. Methods. There were 94 vocal professionals who were examined altogether, out of whom 46 were affected by MAPLs, whereas 48 of them were diagnosed with no MAPLs, so that they served as the control group. All these patients were clinically examined (anamnesis, clinical examination, bacteoriological examination of nose and pharynx, radiography of paranasal cavities, allergological processing, phoniatric examination, endo-video-stroboscopic examination, as well as gastroenterologic examination, and finally endocrinological and pulmological analyses. Results. The changes that occurred most often were identified as nodules (50%; n = 23/46 and polyps (24%; n = 11/46. Risk factors causing MAPLs in vocal professionals were as follows: age, which reduced the risk by 23.9% [OR 0.861 (0.786-0.942] whereas the years of career increase the risk [OR 1.114 (1.000-1.241], as well as the presence of a chronic respiratory disease [OR 7.310 (1.712- 31.218], and the presence of gastro-oesophageal reflux disease [OR 4.542 (1.263-16.334]. The following factors did not contribute to development of MAPLs in vocal professionals: sex, a place of residence, irritation, smoking, endocrinologic disease and the presence of poly-sinusitis. Conclusion. It is necessary to introduce comprehensive procedures for prevention of MAPLs, particularly in high-risk groups. Identification of the risk factors for MAPLs and prevention of their influence on vocal professionals (given that their income depends on their vocal ability is of the highest importance.
Schneider, Berit; Denk, Doris-Maria; Bigenzahn, Wolfgang
A persistent insufficiency of glottal closure is mostly a consequence of a unilateral vocal fold movement impairment. It can also be caused by vocal fold atrophy or scarring processes with regular bilateral respiratory vocal fold function. Because of consequential voice, breathing, and swallowing impairments, a functional surgical treatment is required. The goal of the study was to outline the functional results after medialization thyroplasty with the titanium vocal fold medialization implant according to Friedrich. In the period of 1999 to 2001, an external vocal fold medialization using the titanium implant was performed on 28 patients (12 women and 16 men). The patients were in the age range of 19 to 84 years. Twenty-two patients had a paralysis of the left-side vocal fold, and six patients, of the right-side vocal fold. Detailed functional examinations were executed on all patients before and after the surgery: perceptive voice sound analysis according to the "roughness, breathiness, and hoarseness" method, judgment of the s/z ratio and voice dysfunction index, voice range profile measurements, videostroboscopy, and pulmonary function tests. In case of dysphagia/aspiration, videofluoroscopy of swallowing was also performed. The respective data were statistically analyzed (paired t test, Wilcoxon-test). All patients reported on improvement of voice, swallowing, and breathing functions postoperatively. Videostroboscopy revealed an almost complete glottal closure after surgery in all of the patients. All voice-related parameters showed a significant improvement. An increase of the laryngeal resistance by the medialization procedure could be excluded by analysis of the pulmonary function test. The results confirm the external medialization of the vocal folds as an adequate method in the therapy of voice, swallowing, and breathing impairment attributable to an insufficient glottal closure. The titanium implant offers, apart from good tissue tolerability, the
Modi, Vikash K
Unilateral vocal fold paralysis (UVFP) can cause glottic insufficiency that can result in hoarseness, chronic cough, dysphagia, and/or aspiration. In rare circumstances, UVFP can cause airway obstruction necessitating a tracheostomy. The treatment options for UVFP include observation, speech therapy, vocal fold injection medialization laryngoplasty, thyroplasty, and laryngeal reinnervation. In this chapter, the author will discuss the technique of vocal fold injection for medialization of a UVFP. Copyright © 2012 S. Karger AG, Basel.
Garcia, Elisângela Zacanti; Yamashita, Hélio Kiitiro; Garcia, Davi Sousa; Padovani, Marina Martins Pereira; Azevedo, Renata Rangel; Chiari, Brasília Maria
Cone beam computed tomography (CBCT), which represents an alternative to traditional computed tomography and magnetic resonance imaging, may be a useful instrument to study vocal tract physiology related to vocal exercises. This study aims to evaluate the applicability of CBCT to the assessment of variations in the vocal tract of healthy individuals before and after vocal exercises. Voice recordings and CBCT images before and after vocal exercises performed by 3 speech-language pathologists without vocal complaints were collected and compared. Each participant performed 1 type of exercise, i.e., Finnish resonance tube technique, prolonged consonant "b" technique, or chewing technique. The analysis consisted of an acoustic analysis and tomographic imaging. Modifications of the vocal tract settings following vocal exercises were properly detected by CBCT, and changes in the acoustic parameters were, for the most part, compatible with the variations detected in image measurements. CBCT was shown to be capable of properly assessing the changes in vocal tract settings promoted by vocal exercises. © 2017 S. Karger AG, Basel.
Poremba, Amy; Mishkin, Mortimer
Just as cortical visual processing continues far beyond the boundaries of early visual areas, so too does cortical auditory processing continue far beyond the limits of early auditory areas. In passively listening rhesus monkeys examined with metabolic mapping techniques, cortical areas reactive to auditory stimulation were found to include the entire length of the superior temporal gyrus (STG) as well as several other regions within the temporal, parietal, and frontal lobes. Comparison of these widespread activations with those from an analogous study in vision supports the notion that audition, like vision, is served by several cortical processing streams, each specialized for analyzing a different aspect of sensory input, such as stimulus quality, location, or motion. Exploration with different classes of acoustic stimuli demonstrated that most portions of STG show greater activation on the right than on the left regardless of stimulus class. However, there is a striking shift to left-hemisphere "dominance" during passive listening to species-specific vocalizations, though this reverse asymmetry is observed only in the region of temporal pole. The mechanism for this left temporal pole "dominance" appears to be suppression of the right temporal pole by the left hemisphere, as demonstrated by a comparison of the results in normal monkeys with those in split-brain monkeys.
Wapnick, Joel; Darrow, Alice Ann; Kovacs, Jolan; Dalrymple, Lucinda
Studies whether physical attractiveness of singers affects judges' ratings of their vocal performances. Reveals that physical attractiveness does impact evaluation, that male raters were more severe than female raters, and that the rating of undergraduate majors versus graduate students and professors combined were not differently affected by…
Schaadt, Gesa; Pannekamp, Ann; van der Meer, Elke
These days, illiteracy is still a major problem. There is empirical evidence that auditory phoneme discrimination is one of the factors contributing to written language acquisition. The current study investigated auditory phoneme discrimination in participants who did not acquire written language sufficiently. Auditory phoneme discrimination was…
Harris, G; O'Meara, C; Pemberton, C; Rough, J; Darveniza, P; Tisch, S; Cole, I
To review the clinical signs of vocal fold paresis on laryngeal videostroboscopy, to quantify its impact on patients' quality of life and to confirm the benefit of laryngeal electromyography in its diagnosis. Twenty-nine vocal fold paresis patients were referred for laryngeal electromyography. Voice Handicap Index 10 results were compared to 43 patients diagnosed with vocal fold paralysis. Laryngeal videostroboscopy analysis was conducted to determine side of paresis. Blinded laryngeal electromyography confirmed vocal fold paresis in 92.6 per cent of cases, with vocal fold lag being the most common diagnostic sign. The laryngology team accurately predicted side of paresis in 76 per cent of cases. Total Voice Handicap Index 10 responses were not significantly different between vocal fold paralysis and vocal fold paresis groups (26.08 ± 0.21 and 22.93 ± 0.17, respectively). Vocal fold paresis has a significant impact on quality of life. This study shows that laryngeal electromyography is an important diagnostic tool. Patients with persisting dysphonia and apparently normal vocal fold movement, who fail to respond to appropriate speech therapy, should be investigated for a diagnosis of vocal fold paresis.
Full Text Available Shrews have rich vocal repertoires that include vocalizations within the human audible frequency range and ultrasonic vocalizations. Here, we recorded and analyzed in detail the acoustic structure of a vocalization with unclear functional significance that was spontaneously produced by 15 adult, captive Asian house shrews (Suncus murinus while they were lying motionless and resting in their nests. This vocalization was usually emitted repeatedly in a long series with regular intervals. It showed some structural variability; however, the shrews most frequently emitted a tonal, low-frequency vocalization with minimal frequency modulation and a low, non-vocal click that was clearly noticeable at its beginning. There was no effect of sex, but the acoustic structure of the analyzed vocalizations differed significantly between individual shrews. The encoded individuality was low, but it cannot be excluded that this individuality would allow discrimination of family members, i.e., a male and female with their young, collectively resting in a common nest. The question remains whether the Asian house shrews indeed perceive the presence of their mates, parents or young resting in a common nest via the resting-associated vocalization and whether they use it to discriminate among their family members. Additional studies are needed to explain the possible functional significance of resting-associated vocalizations emitted by captive Asian house shrews. Our study highlights that the acoustic communication of shrews is a relatively understudied topic, particularly considering that they are highly vocal mammals.
Chen, Wenli; Woo, Peak; Murry, Thomas
High-speed videoendoscopy (HSV) captures direct cycle-to-cycle visualization of vocal fold movement in real time. This ultrafast recording rate is capable of visualizing the vibratory motion of the vocal folds in severely disordered phonation and provides a direct method for examining vibratory changes after vocal fold surgery. The purpose of this study was to examine the vibratory motion before and after surgical intervention. HSV was captured from two subjects with identifiable midvocal fold benign lesions and six subjects with highly aperiodic vocal fold vibration before and after phonosurgery. Digital kymography (DKG) was used to extract high-speed kymographic vocal fold images sampled at the midmembranous, anterior 1/3, and posterior 1/3 region. Spectral analysis was subsequently applied to the DKG to quantify the cycle-to-cycle movements of the left and the right vocal fold, expressed as a spectrum. Before intervention, the vibratory spectrum consisted of decreased and flat-like spectral peaks with robust power asymmetry. After intervention, increases in spectral power and decreases in power symmetry were noted. Spectral power increases were most remarkable in the midmembranous region of the vocal fold. Surgical modification resulted in improved lateral excursion of the vocal folds, vibratory function, and perceptual measures of Voice Handicap Index-10. These changes in vibratory behavior trended toward normal vocal fold vibration. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Long, Jennifer L
A vibratory vocal fold replacement would introduce a new treatment paradigm for structural vocal fold diseases such as scarring and lamina propria loss. This work implants a tissue-engineered replacement for vocal fold lamina propria and epithelium in rabbits and compares histology and function to injured controls and orthotopic transplants. Hypotheses were that the cell-based implant would engraft and control the wound response, reducing fibrosis and restoring vibration. Translational research. Rabbit adipose-derived mesenchymal stem cells (ASC) were embedded within a three-dimensional fibrin gel, forming the cell-based outer vocal fold replacement (COVR). Sixteen rabbits underwent unilateral resection of vocal fold epithelium and lamina propria, as well as reconstruction with one of three treatments: fibrin glue alone with healing by secondary intention, replantation of autologous resected vocal fold cover, or COVR implantation. After 4 weeks, larynges were examined histologically and with phonation. Fifteen rabbits survived. All tissues incorporated well after implantation. After 1 month, both graft types improved histology and vibration relative to injured controls. Extracellular matrix (ECM) of the replanted mucosa was disrupted, and ECM of the COVR implants remained immature. Immune reaction was evident when male cells were implanted into female rabbits. Best histologic and short-term vibratory outcomes were achieved with COVR implants containing male cells implanted into male rabbits. Vocal fold cover replacement with a stem cell-based tissue-engineered construct is feasible and beneficial in acute rabbit implantation. Wound-modifying behavior of the COVR implant is judged to be an important factor in preventing fibrosis. NA. Laryngoscope, 128:153-159, 2018. © 2017 The American Laryngological, Rhinological and Otological Society, Inc.
Pelegrin Garcia, David; Brunskog, Jonas
acoustic conditions, artificially generated by electroacoustic means. The vocal intensity decreased with the objective parameter support, which quantifies the amount of sound reflections provided by the room at the talker‟s ears,relative to the direct sound, at a rate of -0.21 dB/dB. The reading pace......Many teachers suffer from voice problems related to the use of their voices in the working environment. The noise generated by students and external sound sources (like traffic noise or neighboring classrooms) is a major problem, as it leads to an increased vocal effort. In the absence of high...... levels of background noise, the room has also an effect on the talker‟s voice. In order to quantify the relative importance of the acoustic environment on the vocal demands for teachers, a laboratory investigation was carried out. Thirteen teachers had to read a text aloud under ten different room...
Myers, Alexander McNaughton
A series of five experiments was conducted to determine whether operant or respondent factors controlled the emission of a particular vocalization ( "Q" ) by human infants 16 to 18 months old. Experiment 1 consisted of a pilot investigation of the effects of an autoshaping procedure on three infants' vocal behavior. All three subjects demonstrated increased emission of the target sound during the CR period. Experiments 2 through 4 attempted to replicate the findings of Experiment 1 under cont...
Hill, N. J.; Schölkopf, B.
We report on the development and online testing of an electroencephalogram-based brain-computer interface (BCI) that aims to be usable by completely paralysed users—for whom visual or motor-system-based BCIs may not be suitable, and among whom reports of successful BCI use have so far been very rare. The current approach exploits covert shifts of attention to auditory stimuli in a dichotic-listening stimulus design. To compare the efficacy of event-related potentials (ERPs) and steady-state auditory evoked potentials (SSAEPs), the stimuli were designed such that they elicited both ERPs and SSAEPs simultaneously. Trial-by-trial feedback was provided online, based on subjects' modulation of N1 and P3 ERP components measured during single 5 s stimulation intervals. All 13 healthy subjects were able to use the BCI, with performance in a binary left/right choice task ranging from 75% to 96% correct across subjects (mean 85%). BCI classification was based on the contrast between stimuli in the attended stream and stimuli in the unattended stream, making use of every stimulus, rather than contrasting frequent standard and rare ‘oddball’ stimuli. SSAEPs were assessed offline: for all subjects, spectral components at the two exactly known modulation frequencies allowed discrimination of pre-stimulus from stimulus intervals, and of left-only stimuli from right-only stimuli when one side of the dichotic stimulus pair was muted. However, attention modulation of SSAEPs was not sufficient for single-trial BCI communication, even when the subject's attention was clearly focused well enough to allow classification of the same trials via ERPs. ERPs clearly provided a superior basis for BCI. The ERP results are a promising step towards the development of a simple-to-use, reliable yes/no communication system for users in the most severely paralysed states, as well as potential attention-monitoring and -training applications outside the context of assistive technology.
Hill, N J; Schölkopf, B
We report on the development and online testing of an EEG-based brain-computer interface (BCI) that aims to be usable by completely paralysed users—for whom visual or motor-system-based BCIs may not be suitable, and among whom reports of successful BCI use have so far been very rare. The current approach exploits covert shifts of attention to auditory stimuli in a dichotic-listening stimulus design. To compare the efficacy of event-related potentials (ERPs) and steady-state auditory evoked potentials (SSAEPs), the stimuli were designed such that they elicited both ERPs and SSAEPs simultaneously. Trial-by-trial feedback was provided online, based on subjects’ modulation of N1 and P3 ERP components measured during single 5-second stimulation intervals. All 13 healthy subjects were able to use the BCI, with performance in a binary left/right choice task ranging from 75% to 96% correct across subjects (mean 85%). BCI classification was based on the contrast between stimuli in the attended stream and stimuli in the unattended stream, making use of every stimulus, rather than contrasting frequent standard and rare “oddball” stimuli. SSAEPs were assessed offline: for all subjects, spectral components at the two exactly-known modulation frequencies allowed discrimination of pre-stimulus from stimulus intervals, and of left-only stimuli from right-only stimuli when one side of the dichotic stimulus pair was muted. However, attention-modulation of SSAEPs was not sufficient for single-trial BCI communication, even when the subject’s attention was clearly focused well enough to allow classification of the same trials via ERPs. ERPs clearly provided a superior basis for BCI. The ERP results are a promising step towards the development of a simple-to-use, reliable yes/no communication system for users in the most severely paralysed states, as well as potential attention-monitoring and -training applications outside the context of assistive technology. PMID:22333135
Full Text Available The quality of the prosthetic-neural interface is a critical point for cochlear implant efficiency. It depends not only on technical and anatomical factors such as electrode position into the cochlea (depth and scalar placement, electrode impedance, and distance between the electrode and the stimulated auditory neurons, but also on the number of functional auditory neurons. The efficiency of electrical stimulation can be assessed by the measurement of e-CAP in cochlear implant users. In the present study, we modeled the activation of auditory neurons in cochlear implant recipients (nucleus device. The electrical response, measured using auto-NRT (neural responses telemetry algorithm, has been analyzed using multivariate regression with cubic splines in order to take into account the variations of insertion depth of electrodes amongst subjects as well as the other technical and anatomical factors listed above. NRT thresholds depend on the electrode squared impedance (β = −0.11 ± 0.02, P<0.01, the scalar placement of the electrodes (β = −8.50 ± 1.97, P<0.01, and the depth of insertion calculated as the characteristic frequency of auditory neurons (CNF. Distribution of NRT residues according to CNF could provide a proxy of auditory neurons functioning in implanted cochleas.
Silva, Wégina Jordâna Nascimento da; Lopes, Leonardo Wanderley; Macedo, Anny Elizabety Ramalho de; Costa, Denise Batista da; Almeida, Anna Alice Figueiredo de
The origin and development of dysphonia, particularly behavioral dysphonia, is associated with several risk factors. Here, we verified the effectiveness of group therapy in reducing the risk factors, and established the association between risk factors and sex, age, profession, and diagnosis of laryngeal disorders in patients with behavioral dysphonia. This is a descriptive, quantitative, field intervention study. Participants (n = 26, adult patients of both sexes), with a diagnosis of behavioral dysphonia, received group therapy intervention. Data for risk factors were collected pre- and posttherapy using the Vocal Screening Protocol. The data were analyzed using descriptive and inferential statistics (Student t test, chi-squared test or Spearman correlation test). The majority (80.8%, n = 21) of patients were female, 65.4% (n = 17) were not in a vocal profession, and 42.3% (n = 11) presented with a lesion in the membranous portion of the vocal fold. The number of personal risk factors decreased after group therapy (P = 0.04). In addition, age was correlated with total (P = 0.001), environmental (P = 0.002), and personal (P = 0.003) vocal risk factors posttherapy. This study revealed an association between the reduction of personal risk factors and vocal group therapy, and a correlation between age and total, environmental, and personal vocal risk factors posttherapy. Thus, maintenance and origins of the behaviors that modify the behavioral aspects of the participants directly influence the production of individual vocal habits. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Lennon, Christen J; Murry, Thomas; Sulica, Lucian
Vocal fold hemorrhage is an acute phonotraumatic injury treated with voice rest; recurrence is a generally accepted indication for surgical intervention. This study aims to identify factors predictive of recurrence based on outcomes of a large clinical series. Retrospective cohort. Retrospective review of cases of vocal fold hemorrhage presenting to a university laryngology service. Demographic information was compiled. Videostroboscopic exams were evaluated for hemorrhage extent, presence of varix, mucosal lesion, and/or vocal fold paresis. Vocal fold hemorrhage recurrence was the main outcome measure. Follow-up telephone survey was used to complement clinical data. Forty-seven instances of vocal fold hemorrhage were evaluated (25M:22F; 32 professional voice users). Twelve of the 47 (26%) patients experienced recurrence. Only the presence of varix demonstrated significant association with recurrence (P = 0.0089) on multivariate logistic regression. Vocal fold hemorrhage recurred in approximately 26% of patients. Varix was a predictor of recurrence, with 48% of those with varix experiencing recurrence. Monitoring, behavioral management and/or surgical intervention may be indicated to treat patients with such characteristics. © 2013 The American Laryngological, Rhinological and Otological Society, Inc.
Elie, Julie Estelle; Soula, Hédi Antoine; Trouvé, Colette; Mathevon, Nicolas; Vignal, Clémentine
Individual cages represent a widely used housing condition in laboratories. This isolation represents an impoverished physical and social environment in gregarious animals. It prevents animals from socializing, even when auditory and visual contact is maintained. Zebra finches are colonial songbirds that are widely used as laboratory animals for the study of vocal communication from brain to behavior. In this study, we investigated the effect of single housing on the vocal behavior and the brain activity of male zebra finches (Taeniopygia guttata): male birds housed in individual cages were compared to freely interacting male birds housed as a social group in a communal cage. We focused on the activity of septo-hypothalamic regions of the "social behavior network" (SBN), a set of limbic regions involved in several social behaviors in vertebrates. The activity of four structures of the SBN (BSTm, medial bed nucleus of the stria terminalis; POM, medial preoptic area; lateral septum; ventromedial hypothalamus) and one associated region (paraventricular nucleus of the hypothalamus) was assessed using immunoreactive nuclei density of the immediate early gene Zenk (egr-1). We further assessed the identity of active cell populations by labeling vasotocin (VT). Brain activity was related to behavioral activities of birds like physical and vocal interactions. We showed that individual housing modifies vocal exchanges between birds compared to communal housing. This is of particular importance in the zebra finch, a model species for the study of vocal communication. In addition, a protocol that daily removes one or two birds from the group affects differently male zebra finches depending of their housing conditions: while communally-housed males changed their vocal output, brains of individually housed males show increased Zenk labeling in non-VT cells of the BSTm and enhanced correlation of Zenk-revealed activity between the studied structures. These results show that
Hintze, Justin M; Gnagi, Sharon H; Lott, David G
Bilateral true vocal fold paralysis is rarely attributable to inflammatory diseases. Sarcoidosis is a rare but important etiology of bilateral true vocal fold paralysis by compressive lymphadenopathy, granulomatous infiltration, and neural involvement. We describe the first reported case of sarcoidosis presenting as bilateral vocal fold immobility caused by direct fixation by granulomatous infiltration severe enough to necessitate tracheostomy insertion. In addition, we discuss the presentation, the pathophysiology, and the treatment of this disease with a review of the literature of previously reported cases of sarcoidosis-related vocal fold immobility. Sarcoidosis should therefore be an important consideration for the otolaryngologist's differential diagnosis of true vocal fold immobility. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Qiao, Zhengxue; Yang, Aiying; Qiu, Xiaohui; Yang, Xiuxian; Zhang, Congpei; Zhu, Xiongzhao; He, Jincai; Wang, Lin; Bai, Bing; Sun, Hailian; Zhao, Lun; Yang, Yanjie
Gender differences in rates of major depressive disorder (MDD) are well established, but gender differences in cognitive function have been little studied. Auditory mismatch negativity (MMN) was used to investigate gender differences in pre-attentive information processing in first episode MDD. In the deviant-standard reverse oddball paradigm, duration auditory MMN was obtained in 30 patients (15 males) and 30 age-/education-matched controls. Over frontal-central areas, mean amplitude of increment MMN (to a 150-ms deviant tone) was smaller in female than male patients; there was no sex difference in decrement MMN (to a 50-ms deviant tone). Neither increment nor decrement MMN differed between female and male patients over temporal areas. Frontal-central MMN and temporal MMN did not differ between male and female controls in any condition. Over frontal-central areas, mean amplitude of increment MMN was smaller in female patients than female controls; there was no difference in decrement MMN. Neither increment nor decrement MMN differed between female patients and female controls over temporal areas. Frontal-central MMN and temporal MMN did not differ between male patients and male controls. Mean amplitude of increment MMN in female patients did not correlate with symptoms, suggesting this sex-specific deficit is a trait- not a state-dependent phenomenon. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Tatiana Fernandes Rocha
Full Text Available OBJETIVO: comparar a extensão vocal de idosos coralistas e não coralistas e analisar a influência da prática do canto-coral amador na extensão vocal dos mesmos. MÉTODOS: extração dos valores da extensão vocal em semitons por meio de um teclado musical e análise comparativa do número de semitons entre 40 idosos coralistas e 40 não coralistas. RESULTADOS: o número de semitons atingido pelos coralistas é significativamente maior que o atingido pelos não coralistas. O perfil de extensão vocal dos idosos coralistas foi de 27 a 39 semitons, perfazendo um total de 3 oitavas, 1 tom e 1 semitom. O perfil de extensão vocal dos idosos não coralistas foi de 18 a 35 semitons, perfazendo um total de 2 oitavas, 5 tons e 1 semitom. CONCLUSÃO: a prática do canto coral amador aumenta a extensão vocal de idosos coralistas.PURPOSE: compare the vocal extension of senior choristers and non-choristers and analyze the influence of the practice of the amateur coral-song in the vocal extension of the aforementioned subjects. METHODS: extracting the vocal extension through a musical keyboard and comparative analysis of the number of half-notes among 40 senior choristers and 40 non-choristers. RESULTS: the number of half-notes achieved by the choristers is significantly higher than the one achieved by the non-choristers. The vocal extension profile of the seniors choristers was from 27 to 39 half-notes, totalizing a sum of 3 octaves, 1 tone and 1 half-note. The profile of the no-choristers seniors' vocal extension was from 18 to 35 half-notes, totalizing a sum of 2 octaves, 5 tones and 1 half-note. CONCLUSION: The practice of the amateur coral song increases the choristers seniors' vocal extension.
D'haeseleer, E; Claeys, S; Wuyts, F; Van Lierde, K M
The main purpose of this study was to determine the vocal quality of 20 male and 9 female university teachers using a multi-parameter approach. Secondly, the effect of an academic lecture on the voice profiles of the university teachers was measured. All groups underwent subjective voice evaluations (perceptual evaluation, Voice Handicap Index, anamnesis of vocal complaints and vocal abuse) and objective voice evaluations (aerodynamic and acoustic parameters, vocal performance, and the Dysphonia Severity Index). The same voice assessment was performed after an academic lecture with a mean length of one and a half hours. The mean DSI score was + 2.2 for the male teachers and + 4.0 for the female teachers. The mean VHI score was 13. Perceptually, all voice parameters were rated as normal. The questionnaire revealed a relatively high amount of vocal abuse. No changes in the objective vocal parameters were found after the lecture. Perceptually, however, the voices of the university teachers were significantly less instable after the lecture. Although no negative changes in objective vocal quality were observed, 48% of the university teachers experienced subjective vocal changes. The authors concluded that university teachers are professional voice users with good vocal quality who suffer no handicapping effect from possible voice disorders. No important changes in the vocal profile after a teaching activity of one and a half hours were found, despite the high prevalence of voice complaints.
Mizuta, Masanobu; Kurita, Takashi; Dillon, Neal P; Kimball, Emily E; Garrett, C Gaelyn; Sivasankar, M Preeti; Webster, Robert J; Rousseau, Bernard
A custom-designed probe was developed to measure vocal fold surface resistance in vivo. The purpose of this study was to demonstrate proof of concept of using vocal fold surface resistance as a proxy of functional tissue integrity after acute phonotrauma using an animal model. Prospective animal study. New Zealand White breeder rabbits received 120 minutes of airflow without vocal fold approximation (control) or 120 minutes of raised intensity phonation (experimental). The probe was inserted via laryngoscope and placed on the left vocal fold under endoscopic visualization. Vocal fold surface resistance of the middle one-third of the vocal fold was measured after 0 (baseline), 60, and 120 minutes of phonation. After the phonation procedure, the larynx was harvested and prepared for transmission electron microscopy. In the control group, vocal fold surface resistance values remained stable across time points. In the experimental group, surface resistance (X% ± Y% relative to baseline) was significantly decreased after 120 minutes of raised intensity phonation. This was associated with structural changes using transmission electron microscopy, which revealed damage to the vocal fold epithelium after phonotrauma, including disruption of the epithelium and basement membrane, dilated paracellular spaces, and alterations to epithelial microprojections. In contrast, control vocal fold specimens showed well-preserved stratified squamous epithelia. These data demonstrate the feasibility of measuring vocal fold surface resistance in vivo as a means of evaluating functional vocal fold epithelial barrier integrity. Device prototypes are in development for additional testing, validation, and for clinical applications in laryngology. NA Laryngoscope, 127:E364-E370, 2017. © 2017 The American Laryngological, Rhinological and Otological Society, Inc.
Chen, Hao; Sun, Jing Wu; Wan, Guang Lun; Hu, Yan Ming
To explore the character of laryngoscopy finding, voice, and therapy of vocal fold fibrous mass. Clinical data, morphology, voice character, surgery and pathology of 15 cases with vocal fold fibrous mass were analyzed. The morbidity of vocal fold fibrous mass might be related to overuse of voice and laryngopharyngeal reflex. Laryngoscopy revealed shuttle line appearance, smoothness and decreased mucosal wave of vocal fold. These patients were invalid for voice training and might be improved by surgery, but recovery is slow. The morbidity of vocal fold fibrous mass might be related to overuse of voice and laryngopharyngeal reflex. Conservative treatment is ineffective for this disease, and surgery might improve. Copyright© by the Editorial Department of Journal of Clinical Otorhinolaryngology Head and Neck Surgery.
Kim, Dong Wook; Kim, Eun Ji; Kim, Eun Na; Sung, Myung Whun; Kwon, Tack-Kyun; Cho, Yong Woo; Kwon, Seong Keun
Vocal fold paralysis results from various etiologies and can induce voice changes, swallowing complications, and issues with aspiration. Vocal fold paralysis is typically managed using injection laryngoplasty with fat or synthetic polymers. Injection with autologous fat has shown excellent biocompatibility. However, it has several disadvantages such as unpredictable resorption rate, morbidities associated with liposuction procedure which has to be done in operating room under general anesthesia. Human adipose-derived extracellular matrix (ECM) grafts have been reported to form new adipose tissue and have greater biostability than autologous fat graft. Here, we present an injectable hydrogel that is constructed from adipose tissue derived soluble extracellular matrix (sECM) and methylcellulose (MC) for use in vocal fold augmentation. Human sECM derived from adipose tissue was extracted using two major steps-ECM was isolated from human adipose tissue and was subsequently solubilized. Injectable sECM/MC hydrogels were prepared by blending of sECM and MC. Sustained vocal fold augmentation and symmetric vocal fold vibration were accomplished by the sECM/MC hydrogel in paralyzed vocal fold which were confirmed by laryngoscope, histology and a high-speed imaging system. There were increased number of collagen fibers and fatty granules at the injection site without significant inflammation or fibrosis. Overall, these results indicate that the sECM/MC hydrogel can enhance vocal function in paralyzed vocal folds without early resorption and has potential as a promising material for injection laryngoplasty for stable vocal fold augmentation which can overcome the shortcomings of autologous fat such as unpredictable duration and morbidity associated with the fat harvest.
Tsuji, Domingos Hiroshi; Hachiya, Adriana; Dajer, Maria Eugenia; Ishikawa, Camila Cristina; Takahashi, Marystella Tomoe; Montagnoli, Arlindo Neto
Introduction The study of the dynamic properties of vocal fold vibration is important for understanding the vocal production mechanism and the impact of organic and functional changes. The advent of high-speed videolaryngoscopy (HSV) has provided the possibility of seeing the real cycle of vocal fold vibration in detail through high sampling rate of successive frames and adequate spatial resolution. Objective To describe the technique, advantages, and limitations of using HSV and digital videokymography in the diagnosis of vocal pathologies. Methods We used HSV and digital videokymography to evaluate one normophonic individual and four patients with vocal fold pathologies (nodules, unilateral paralysis of the left vocal fold, intracordal cyst, and adductor spasmodic dysphonia). The vocal fold vibration parameters (glottic closure, vibrational symmetry, periodicity, mucosal wave, amplitude, and glottal cycle phases) were assessed. Results Differences in the vocal vibration parameters were observed and correlated with the pathophysiology. Conclusion HSV is the latest diagnostic tool in visual examination of vocal behavior and has considerable potential to refine our knowledge regarding the vocal fold vibration and voice production, as well as regarding the impact of pathologic conditions have on the mechanism of phonation. PMID:25992109
Tsuji, Domingos Hiroshi
Full Text Available Introduction The study of the dynamic properties of vocal fold vibration is important for understanding the vocal production mechanism and the impact of organic and functional changes. The advent of high-speed videolaryngoscopy (HSV has provided the possibility of seeing the real cycle of vocal fold vibration in detail through high sampling rate of successive frames and adequate spatial resolution. Objective To describe the technique, advantages, and limitations of using HSV and digital videokymography in the diagnosis of vocal pathologies. Methods We used HSV and digital videokymography to evaluate one normophonic individual and four patients with vocal fold pathologies (nodules, unilateral paralysis of the left vocal fold, intracordal cyst, and adductor spasmodic dysphonia. The vocal fold vibration parameters (glottic closure, vibrational symmetry, periodicity, mucosal wave, amplitude, and glottal cycle phases were assessed. Results Differences in the vocal vibration parameters were observed and correlated with the pathophysiology. Conclusion HSV is the latest diagnostic tool in visual examination of vocal behavior and has considerable potential to refine our knowledge regarding the vocal fold vibration and voice production, as well as regarding the impact of pathologic conditions have on the mechanism of phonation.
Koelsch, S; Schröger, E; Tervaniemi, M
The present study focuses on influences of long-term experience on auditory processing, providing the first evidence for pre-attentively superior auditory processing in musicians. This was revealed by the brain's automatic change-detection response, which is reflected electrically as the mismatch negativity (MMN) and generated by the operation of sensoric (echoic) memory, the earliest cognitive memory system. Major chords and single tones were presented to both professional violinists and non-musicians under ignore and attend conditions. Slightly impure chords, presented among perfect major chords elicited a distinct MMN in professional musicians, but not in non-musicians. This demonstrates that compared to non-musicians, musicians are superior in pre-attentively extracting more information out of musically relevant stimuli. Since effects of long-term experience on pre-attentive auditory processing have so far been reported for language-specific phonemes only, results indicate that sensory memory mechanisms can be modulated by training on a more general level.
Slevc, L Robert; Shell, Alison R
Auditory agnosia refers to impairments in sound perception and identification despite intact hearing, cognitive functioning, and language abilities (reading, writing, and speaking). Auditory agnosia can be general, affecting all types of sound perception, or can be (relatively) specific to a particular domain. Verbal auditory agnosia (also known as (pure) word deafness) refers to deficits specific to speech processing, environmental sound agnosia refers to difficulties confined to non-speech environmental sounds, and amusia refers to deficits confined to music. These deficits can be apperceptive, affecting basic perceptual processes, or associative, affecting the relation of a perceived auditory object to its meaning. This chapter discusses what is known about the behavioral symptoms and lesion correlates of these different types of auditory agnosia (focusing especially on verbal auditory agnosia), evidence for the role of a rapid temporal processing deficit in some aspects of auditory agnosia, and the few attempts to treat the perceptual deficits associated with auditory agnosia. A clear picture of auditory agnosia has been slow to emerge, hampered by the considerable heterogeneity in behavioral deficits, associated brain damage, and variable assessments across cases. Despite this lack of clarity, these striking deficits in complex sound processing continue to inform our understanding of auditory perception and cognition. © 2015 Elsevier B.V. All rights reserved.
There are four phocids in waters around Antarctica: Weddell, leopard, crabeater, and Ross seals. These four species provide a unique opportunity to examine underwater vocal behavior in species sharing the same ecosystem. Some species live in pack ice, others in factice, but all are restricted to the Antarctic or sub-Antarctic islands. All breed and produce vocalizations under water. Social systems range from polygyny in large breeding colonies, to serial monogamy, to solitary species. The type of mating system influences the number of underwater vocalizations in the repertoire, with monogamous seals producing only a single call, polygynous species producing up to 35 calls, and solitary species an intermediate number of about 10 calls. Breeding occurs during the austral spring and each species carves-out an acoustic niche for communicating, with species using different frequency ranges, temporal patterns, and amplitude changes to convey their species-specific calls and presumably reduce acoustic competition. Some species exhibit geographic variations in their vocalizations around the continent, which may reflect discrete breeding populations. Some seals become silent during a vulnerable time of predation by killer whales, perhaps to avoid detection. Overall, vocalizations of these seals exhibit adaptive characteristics that reflect the co-evolution among species in the same ecosystem.
Full Text Available The principal symptoms of unilateral vocal fold paralysis are hoarseness and difficulty in swallowing. Dyspnea is comparatively rare (Laccourreye et al., 2003. The extent to which unilateral vocal fold paralysis may lead to respiratory problems at all - in contrast to bilateral vocal fold paralysis- has not yet well been determined. On the one hand, inspiration is impaired with unilateral vocal fold paralysis; on the other hand, neither the position of the vocal fold paralysis nor the degree of breathiness correlates with respiratory parameters (Cantarella et al., 2003; 2005. The question of what respiratory stress a patient with a vocal fold paresis can endure has not yet been dealt with.A 43 year-old female patient was suffering from recurrent unspecific respiratory complaints for four months after physical activity. During training for a marathon, she experienced no difficulty in breathing. These unspecific respiratory complaints occurred only after athletic activity and persisted for hours. The patient observed neither an increased coughing nor a stridor. Her voice remained unaltered during the attacks, nor were there any signs of a symptomatic gastroesophageal reflux or infectious disease. A cardio-pulmonary and a radiological examination by means of an X-ray of the thorax also revealed no pathological phenomena. As antiallergic and antiobstructive therapy remained unsuccessful, a laryngological examination was performed in order to exclude a vocal cord dysfunction.Surprisingly enough, the laryngostroboscopy showed, as an initial description, a vocal fold paralysis of the left vocal fold in median position (Figure 1. The anamnestic background for the cause was unclear. The only clue was a thoracotomy on the left side due to a pleuritis in childhood. A subsequent laryngoscopic examination had never been performed. Good mucosa waves and amplitudes were shown bilateral with complete glottal closure. Neither in the acoustic analysis, nor in the
Grosso, A; Cambiaghi, M; Concina, G; Sacco, T; Sacchetti, B
Emotional memories represent the core of human and animal life and drive future choices and behaviors. Early research involving brain lesion studies in animals lead to the idea that the auditory cortex participates in emotional learning by processing the sensory features of auditory stimuli paired with emotional consequences and by transmitting this information to the amygdala. Nevertheless, electrophysiological and imaging studies revealed that, following emotional experiences, the auditory cortex undergoes learning-induced changes that are highly specific, associative and long lasting. These studies suggested that the role played by the auditory cortex goes beyond stimulus elaboration and transmission. Here, we discuss three major perspectives created by these data. In particular, we analyze the possible roles of the auditory cortex in emotional learning, we examine the recruitment of the auditory cortex during early and late memory trace encoding, and finally we consider the functional interplay between the auditory cortex and subcortical nuclei, such as the amygdala, that process affective information. We conclude that, starting from the early phase of memory encoding, the auditory cortex has a more prominent role in emotional learning, through its connections with subcortical nuclei, than is typically acknowledged. Copyright © 2015 IBRO. Published by Elsevier Ltd. All rights reserved.
Brandt, Christian; Malmkvist, Jens; Nielsen, Rasmus L.
-tested (P=0.004). Large numbers of mink are kept in fur industry farms, and our results are important to the understanding of sound communication, which is part of their natural behaviour. Our results also suggest mink as an interesting model for studying the development of mammalian hearing and its......American mink (Neovison vison) kits are born altricial and fully dependent on maternal care, for which the kits’ vocalizations appear essential. We used auditory brainstem responses (ABRs) to determine: (1) hearing sensitivity of adult females from two breeding lines known to differ in maternal...... behaviour and (2) development of hearing in kits 8–52days of age. We also studied sound production in 20 kits throughout postnatal days 1 to 44. Adult female mink had a broad hearing range from 1kHz to above 70kHz, with peak sensitivity (threshold of 20dB SPL) at 8–10kHz, and no difference...
Michael H Graber
Full Text Available Motor functions are often guided by sensory experience, most convincingly illustrated by complex learned behaviors. Key to sensory guidance in motor areas may be the structural and functional organization of sensory inputs and their evoked responses. We study sensory responses in large populations of neurons and neuron-assistive cells in the songbird motor area HVC, an auditory-vocal brain area involved in sensory learning and in adult song production. HVC spike responses to auditory stimulation display remarkable preference for the bird's own song (BOS compared to other stimuli. Using two-photon calcium imaging in anesthetized zebra finches we measure the spatio-temporal structure of baseline activity and of auditory evoked responses in identified populations of HVC cells. We find strong correlations between calcium signal fluctuations in nearby cells of a given type, both in identified neurons and in astroglia. In identified HVC neurons only, auditory stimulation decorrelates ongoing calcium signals, less for BOS than for other sound stimuli. Overall, calcium transients show strong preference for BOS in identified HVC neurons but not in astroglia, showing diversity in local functional organization among identified neuron and astroglia populations.
Full Text Available Vocal folds are used as sound sources in various species, but it is unknown how vocal fold morphologies are optimized for different acoustic objectives. Here we identify two main variables affecting range of vocal fold vibration frequency, namely vocal fold elongation and tissue fiber stress. A simple vibrating string model is used to predict fundamental frequency ranges across species of different vocal fold sizes. While average fundamental frequency is predominantly determined by vocal fold length (larynx size, range of fundamental frequency is facilitated by (1 laryngeal muscles that control elongation and by (2 nonlinearity in tissue fiber tension. One adaptation that would increase fundamental frequency range is greater freedom in joint rotation or gliding of two cartilages (thyroid and cricoid, so that vocal fold length change is maximized. Alternatively, tissue layers can develop to bear a disproportionate fiber tension (i.e., a ligament with high density collagen fibers, increasing the fundamental frequency range and thereby vocal versatility. The range of fundamental frequency across species is thus not simply one-dimensional, but can be conceptualized as the dependent variable in a multi-dimensional morphospace. In humans, this could allow for variations that could be clinically important for voice therapy and vocal fold repair. Alternative solutions could also have importance in vocal training for singing and other highly-skilled vocalizations.
Kanazawa, Takeharu; Komazawa, Daigo; Indo, Kanako; Akagi, Yusuke; Lee, Yogaku; Nakamura, Kazuhiro; Matsushima, Koji; Kunieda, Chikako; Misawa, Kiyoshi; Nishino, Hiroshi; Watanabe, Yusuke
Severe vocal fold lesions such as vocal fold sulcus, scars, and atrophy induce a communication disorder due to severe hoarseness, but a treatment has not been established. Basic fibroblast growth factor (bFGF) therapies by either four-time repeated local injections or regenerative surgery for vocal fold scar and sulcus have previously been reported, and favorable outcomes have been observed. In this study, we modified bFGF therapy using a single of bFGF injection, which may potentially be used in office procedures. Retrospective chart review. Five cases of vocal fold sulcus, six cases of scars, seven cases of paralysis, and 17 cases of atrophy were treated by a local injection of bFGF. The injection regimen involved injecting 50 µg of bFGF dissolved in 0.5 mL saline only once into the superficial lamina propria using a 23-gauge injection needle. Two months to 3 months after the injection, phonological outcomes were evaluated. The maximum phonation time (MPT), mean airflow rate, pitch range, speech fundamental frequency, jitter, and voice handicap index improved significantly after the bFGF injection. Furthermore, improvement in the MPT was significantly greater in patients with (in increasing order) vocal fold atrophy, scar, and paralysis. The improvement in the MPT among all patients was significantly correlated with age; the MPT improved more greatly in younger patients. Regenerative treatments by bFGF injection—even a single injection—effectively improve vocal function in vocal fold lesions. 4 © 2015 The American Laryngological, Rhinological and Otological Society, Inc.
Zambon, Fabiana; Moreti, Felipe; Behlau, Mara
To understand the coping strategies used by teachers with vocal complaints, compare the differences between those who seek and those who do not seek voice therapy, and investigate the relationships among coping and voice perceptual analysis, coping and signs and symptoms of voice, and coping and participation restrictions and limitations in vocal activities. Cross-sectional nonrandomized prospective study with control group. Ninety female teachers participated in the study, of similar ages, divided into three groups: group 1 (G1) comprised 30 teachers with vocal complaints who sought voice therapy, group 2 (G2) comprised 30 teachers with vocal complaints who never sought voice therapy, and group 3 (G3) comprised 30 teachers without vocal complaints. The following analysis were conducted: identification and characterization questionnaire, addressing personal and occupational description, recording speech material for voice perceptual analysis, Voice Signs and Symptoms Questionnaire, Voice Activity and Participation Profile (VAPP), and Voice Disability Coping Questionnaire (VDCQ)-Brazilian Version. In relation to the voice perceptual analysis, there was statistically significant difference between the groups with vocal complaint (G1+G2), which had showed voices with mild-to-moderate deviation, and the group without vocal complaint (G1), which showed voices within the normal variability of voice quality (mean for G1 = 49.9, G2 = 43.7, and G3 = 32.3, P Teachers with vocal complaints who looked for voice therapy use more coping strategies. Moreover, they present a tendency to use more problem-focused coping strategies. Voice symptoms prompt the teachers into seeking treatment; however, they are not correlated with the coping itself. In general, the higher the perception of limitation and restriction of participating in vocal activities, the greater the use of coping strategies. Copyright © 2014 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Chang, Joseph; Yung, Katherine C
This case report is the first documentation of dysphonia and vocal fold telangiectasia as a complication of hereditary hemorrhagic telangiectasia (HHT). Case report of a 40-year-old man with HHT presenting with 2 years of worsening hoarseness. Hoarseness corresponded with a period of anticoagulation. Endoscopy revealed vocal fold scarring, vocal fold telangiectasias, and plica ventricular is suggestive of previous submucosal vocal fold hemorrhage and subsequent counterproductive compensation with ventricular phonation. Hereditary hemorrhagic telangiectasia may present as dysphonia with vocal fold telangiectasias and place patients at risk of vocal fold hemorrhage. © The Author(s) 2014.
Dalziell, Anastasia H; Welbergen, Justin A; Igic, Branislav; Magrath, Robert D
Mimicry is a classical example of adaptive signal design. Here, we review the current state of research into vocal mimicry in birds. Avian vocal mimicry is a conspicuous and often spectacular form of animal communication, occurring in many distantly related species. However, the proximate and ultimate causes of vocal mimicry are poorly understood. In the first part of this review, we argue that progress has been impeded by conceptual confusion over what constitutes vocal mimicry. We propose a modified version of Vane-Wright's (1980) widely used definition of mimicry. According to our definition, a vocalisation is mimetic if the behaviour of the receiver changes after perceiving the acoustic resemblance between the mimic and the model, and the behavioural change confers a selective advantage on the mimic. Mimicry is therefore specifically a functional concept where the resemblance between heterospecific sounds is a target of selection. It is distinct from other forms of vocal resemblance including those that are the result of chance or common ancestry, and those that have emerged as a by-product of other processes such as ecological convergence and selection for large song-type repertoires. Thus, our definition provides a general and functionally coherent framework for determining what constitutes vocal mimicry, and takes account of the diversity of vocalisations that incorporate heterospecific sounds. In the second part we assess and revise hypotheses for the evolution of avian vocal mimicry in the light of our new definition. Most of the current evidence is anecdotal, but the diverse contexts and acoustic structures of putative vocal mimicry suggest that mimicry has multiple functions across and within species. There is strong experimental evidence that vocal mimicry can be deceptive, and can facilitate parasitic interactions. There is also increasing support for the use of vocal mimicry in predator defence, although the mechanisms are unclear. Less progress has
Wasim Elhendi Halawa
Full Text Available Con el objetivo de analizar el grado de discapacidad que suponen los nódulos vocales para los pacientes,presentamos los resultados de la valoración subjetiva (el índice de discapacidad vocal (V.H.I.-30adaptado al español y valoración de la sintomatología asociada a la disfonía en 97 pacientesdiagnosticados de nódulos vocales, encontrando un grado importante de discapacidad reflejado por unosvalores elevados del V.H.I.-30 (61,18, por sus tres subescalas (orgánica -26,48, funcional -21,75 yemocional -12,94 y por un importante grado de afectación por los síntomas asociados. Se comparannuestros resultados con los del grupo control de nuestro entorno y se estratifican los resultados según laprofesión de los pacientes. Concluimos que la presencia de nódulos vocales supone una discapacidadimportante a nivel de las actividades sociales y laborales del paciente y un impacto emocionalconsiderable.
Hardison, Debra M.
The majority of studies in second-language (L2) speech processing have involved unimodal (i.e., auditory) input; however, in many instances, speech communication involves both visual and auditory sources of information. Some researchers have argued that multimodal speech is the primary mode of speech perception (e.g., Rosenblum 2005). Research on…
Li, Jin-rang; Sun, Jian-jun
To study the diagnosis and treatment of varices of the vocal cord. The clinical data of 21 cases with varix of vocal cord were analyzed. All the patients presented hoarseness. There were 15 female and 6 male cases with their ages ranged from 23 to 68 years (median 44 years old). The varix was found on the right vocal cord in 12 cases, on the left vocal cord in 9 cases. Isolated varix existed on the vocal cord in 10 cases, varix with vocal cord polyps or nodules in 10 cases, varix with vocal cord paralysis in 1 case. All the patients were diagnosed under the laryngovideoscopy. The lesions appeared on the superior surface of the vocal cord. Varices manifested as abnormally dilated capillary running in the anterior to posterior direction in 6 cases, as clusters of capillary in 3 cases, as a dot or small sheet or short line of capillary in 12 cases. The varices were disappeared in 2 of 8 cases with vocal cord varices and polyps after removed the polyps. The varices of others patients had no change after following up for more than 6 months, but one patient happened hemorrhage of the contralateral vocal cord. Varices are most commonly seen in female. Laryngovideoscopy is the key in determining the vocal fold varices. Management of patients with a varix includes medical therapy, speech therapy, and occasionally surgical vaporization.
Tavares, Elaine L M; Martins, Regina H G
The aim of this study was to perform voice evaluation in teachers with and without vocal symptoms, identifying etiologic factors of dysphonia, voice symptoms, vocal qualities, and laryngeal lesions. Eighty teachers were divided into two groups: GI (without or sporadic symptoms, 40) and GII (with frequent vocal symptoms, 40). They answered a specific questionnaire, and were subject to a perceptual vocal assessment (maximum phonation time, glottal attack, resonance, coordination of breathing and voicing, pitch, and loudness), GIRBAS scale, and to videolaryngoscopy. Females were predominant in both groups, and the age range was from 36 to 50 years. Elementary teachers predominated, working in classes with 31-40 students. Voice symptoms and alterations in the perceptual vocal analysis and in the GIRBAS scale were more frequent in GII. In 46 teachers (GI-16; GII-30), videolaryngoscopy exams were abnormal with the vocal nodules being the most frequent lesions. These results indicate that a teacher's voice is compromised, and requires more attention including control of environmental factors and associated diseases, preventive vocal hygiene, periodic laryngeal examinations, and access to adequate specialist treatment.
Keskinen, Anu Katri
This study is focused on the discipline of higher education contemporary commercial music (CCM) vocal pedagogy through the experiences of two vocal pedagogy teachers, the other in the USA and the other in Finland. The aim of this study has been to find out how the discipline presently looks from a vocal pedagogy teacher's viewpoint, what has the process of building higher education CCM vocal pedagogy courses been like, and where is the field headed. The discussion on CCM pedagogy, also kn...
Cates, Daniel J; Venkatesan, Naren N; Strong, Brandon; Kuhn, Maggie A; Belafsky, Peter C
The effect of vocal fold medialization (VFM) on vocal improvement in persons with unilateral vocal fold immobility (UVFI) is well established. The effect of VFM on the symptom of dysphagia is uncertain. The purpose of this study is to evaluate dysphagia symptoms in patients with UVFI pre- and post-VFM. Case series with chart review. Academic tertiary care medical center. The charts of 44 persons with UVFI who underwent VFM between June 1, 2013, and December 31, 2014, were abstracted from a prospectively maintained database at the University of California, Davis, Voice and Swallowing Center. Patient demographics, indications, and type of surgical procedure were recorded. Self-reported swallowing impairment was assessed with the validated 10-item Eating Assessment Tool (EAT-10) before and after surgery. A paired samples t test was used to compare pre- and postmedialization EAT-10 scores. Forty-four patients met criteria and underwent either vocal fold injection (73%) or thyroplasty (27%). Etiologies of vocal fold paralysis were iatrogenic (55%), idiopathic (29%), benign or malignant neoplastic (9%), traumatic (5%), or related to the late effects of radiation (2%). EAT-10 (mean ± SD) scores improved from 12.2 ± 11.1 to 7.7 ± 7.2 after medialization (P dysphagia and report significant improvement in swallowing symptoms following VFM. The symptomatic improvement appears to be durable over time. © American Academy of Otolaryngology—Head and Neck Surgery Foundation 2016.
Dong Wook Kim
Full Text Available Vocal fold paralysis results from various etiologies and can induce voice changes, swallowing complications, and issues with aspiration. Vocal fold paralysis is typically managed using injection laryngoplasty with fat or synthetic polymers. Injection with autologous fat has shown excellent biocompatibility. However, it has several disadvantages such as unpredictable resorption rate, morbidities associated with liposuction procedure which has to be done in operating room under general anesthesia. Human adipose-derived extracellular matrix (ECM grafts have been reported to form new adipose tissue and have greater biostability than autologous fat graft. Here, we present an injectable hydrogel that is constructed from adipose tissue derived soluble extracellular matrix (sECM and methylcellulose (MC for use in vocal fold augmentation. Human sECM derived from adipose tissue was extracted using two major steps-ECM was isolated from human adipose tissue and was subsequently solubilized. Injectable sECM/MC hydrogels were prepared by blending of sECM and MC. Sustained vocal fold augmentation and symmetric vocal fold vibration were accomplished by the sECM/MC hydrogel in paralyzed vocal fold which were confirmed by laryngoscope, histology and a high-speed imaging system. There were increased number of collagen fibers and fatty granules at the injection site without significant inflammation or fibrosis. Overall, these results indicate that the sECM/MC hydrogel can enhance vocal function in paralyzed vocal folds without early resorption and has potential as a promising material for injection laryngoplasty for stable vocal fold augmentation which can overcome the shortcomings of autologous fat such as unpredictable duration and morbidity associated with the fat harvest.
Full Text Available The operating theatre is a noisy place with many uninformative and redundant alarms. Using data from a recent observational study, we demonstrate that anaesthetists actively respond to only 3.4% of all audible alarms. We outline a range of possible solutions to the alarm problem. Ecological Interface Design (EID helps to outline the requirements for an information environment for anaesthetists and to indicate the possible benefits of continuous auditory signals. Our observational data are then "reworked" to give an indication of possible benefits of a continuous auditory display. Finally we indicate steps we are taking to test these ideas empirically
Louzada,Talita; Beraldinelle,Roberta; Berretin-Felix,Giédre; Brasolotto,Alcione Ghedini
The evaluation of oral and vocal fold diadochokinesis (DDK) in individuals with voice disorders may contribute to the understanding of factors that affect the balanced vocal production. Scientific studies that make use of this assessment tool support the knowledge advance of this area, reflecting the development of more appropriate therapeutic planning. Objective: To compare the results of oral and vocal fold DDK in dysphonic women and in women without vocal disorders. Material and methods: F...
Rutkowski, Tomasz M.
The paper reviews nine robotic and virtual reality (VR) brain–computer interface (BCI) projects developed by the author, in collaboration with his graduate students, within the BCI–lab research group during its association with University of Tsukuba, Japan. The nine novel approaches are discussed in applications to direct brain-robot and brain-virtual-reality-agent control interfaces using tactile and auditory BCI technologies. The BCI user intentions are decoded from the brainwaves in realti...
Bastian, Robert W; Thomas, James P
Assess the correlation between self-rating scales of talkativeness and loudness with various types of voice disorders. This is a retrospective study. A total of 974 patients were analyzed. The cohort study included 430 consecutive patients presenting to the senior author with voice complaints from December 1995 to December 1998. The case-control study added 544 consecutive patients referred to the same examiner from January 1988 to December 1998 for vocal fold examination before thyroid, parathyroid, and carotid surgery. Patient responses on seven-point Likert self-rating scales of talkativeness and loudness were compared with laryngeal disease. Mucosal lesions clearly associated with vibratory trauma are strongly associated with a high self-rating of talkativeness. Laryngeal deconditioning disorders were associated with a low self-rating of talkativeness. Use of a simple self-rating scale of vocal loudness and talkativeness during history taking can reliably orient the examiner to the types of voice disorders likely to be diagnosed subsequently during vocal capability testing and visual laryngeal examination. The high degree of talkativeness and loudness seen in vocal overdoers correlates well with mucosal disorders such as nodules, polyps, capillary ectasia, epidermoid inclusion cysts, and hemorrhage. A lower degree of talkativeness correlates with muscle deconditioning disorders such as vocal fold bowing, atrophy, presbyphonia, and vocal fatigue syndrome. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Chang, Wei-Han; Fang, Tuan-Jen; Li, Hsueh-Yu; Jaw, Fu-Shan; Wong, Alice M K; Pei, Yu-Cheng
Unilateral vocal fold paralysis with no preceding causes is diagnosed as idiopathic unilateral vocal fold paralysis. However, comprehensive guidelines for evaluating the defining characteristics of idiopathic unilateral vocal fold paralysis are still lacking. In the present study, we hypothesized that idiopathic unilateral vocal fold paralysis may have different clinical and neurologic characteristics from unilateral vocal fold paralysis caused by surgical trauma. Retrospective, case series study. Patients with unilateral vocal fold paralysis were evaluated using quantitative laryngeal electromyography, videolaryngostroboscopy, voice acoustic analysis, the Voice Outcome Survey, and the Short Form-36 Health Survey quality-of-life questionnaire. Patients with idiopathic and iatrogenic vocal fold paralysis were compared. A total of 124 patients were recruited. Of those, 17 with no definite identified causes after evaluation and follow-up were assigned to the idiopathic group. The remaining 107 patients with surgery-induced vocal fold paralysis were assigned to the iatrogenic group. Patients in the idiopathic group had higher recruitment of the thyroarytenoid-lateral cricoarytenoid muscle complex and better quality of life compared with the iatrogenic group. Idiopathic unilateral vocal fold paralysis has a distinct clinical presentation, with relatively minor denervation changes in the involved laryngeal muscles, and less impact on quality of life compared with iatrogenic vocal fold paralysis. 4. Laryngoscope, 126:E362-E368, 2016. © 2016 The American Laryngological, Rhinological and Otological Society, Inc.
Schierholz, Irina; Finke, Mareike; Kral, Andrej; Büchner, Andreas; Rach, Stefan; Lenarz, Thomas; Dengler, Reinhard; Sandmann, Pascale
There is substantial variability in speech recognition ability across patients with cochlear implants (CIs), auditory brainstem implants (ABIs), and auditory midbrain implants (AMIs). To better understand how this variability is related to central processing differences, the current electroencephalography (EEG) study compared hearing abilities and auditory-cortex activation in patients with electrical stimulation at different sites of the auditory pathway. Three different groups of patients with auditory implants (Hannover Medical School; ABI: n = 6, CI: n = 6; AMI: n = 2) performed a speeded response task and a speech recognition test with auditory, visual, and audio-visual stimuli. Behavioral performance and cortical processing of auditory and audio-visual stimuli were compared between groups. ABI and AMI patients showed prolonged response times on auditory and audio-visual stimuli compared with NH listeners and CI patients. This was confirmed by prolonged N1 latencies and reduced N1 amplitudes in ABI and AMI patients. However, patients with central auditory implants showed a remarkable gain in performance when visual and auditory input was combined, in both speech and non-speech conditions, which was reflected by a strong visual modulation of auditory-cortex activation in these individuals. In sum, the results suggest that the behavioral improvement for audio-visual conditions in central auditory implant patients is based on enhanced audio-visual interactions in the auditory cortex. Their findings may provide important implications for the optimization of electrical stimulation and rehabilitation strategies in patients with central auditory prostheses. Hum Brain Mapp 38:2206-2225, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.
Condições de produção vocal de vendedores de móveis e eletrodomésticos: correlação entre questões de saúde, hábitos e sintomas vocais Vocal production conditions for furniture and household appliance salespeople: correlation between health issues and vocal symptoms and habits
Léslie Piccolotto Ferreira
own work environment, to answer a questionnaire composed of 16 questions concerning personal information, issues on general health, vocal habits and symptoms. Data were digitized and analyzed using SPSS (Statistical Package for Social Sciences version 13.0. RESULTS: the sample was composed of 66 men and 34 women, with an average age of 24 years of age, working in this occupation for an mean time of 8.5 years. The subjects mostly reported general health disorders, related to emotional (31%, sleep (27%, and digestive (25% issues; in auditory disorders, such as ear-itch (36%, intolerance to loud sounds (29%, and hearing difficulty (20%. As far as vocal habits, the aspects mentioned were excessive talking (83%, fluid intake in cold temperatures (73% and drinking coffee (68%. The symptoms that were most reported were dry throat and mouth (30%, vocal fatigue (22% and phlegm (18%. The possible causes associated to these symptoms by the participants in the study were intense vocal usage (49%, presence of dust (27% and air-conditioning (23% in work environment. Only four subjects reported having a voice disorder. CONCLUSION: the salespeople at issue note the occurrence of vocal symptoms, but not those related to voice disorders. They are also unaware of vocal care habits. Therefore, they need health promotion and vocal disorder prevention actions, aiming at making them aware on these health issues.
Postma, G N; Courey, M S; Ossoff, R H
Microvascular lesions, also called varices or capillary ectasias, in contrast to vocal fold polyps with telangiectatic vessels, are relatively small lesions arising from the microcirculation of the vocal fold. Varices are most commonly seen in female professional vocalists and may be secondary to repetitive trauma, hormonal variations, or repeated inflammation. Microvascular lesions may either be asymptomatic or cause frank dysphonia by interrupting the normal vibratory pattern, mass, or closure of the vocal folds. They may also lead to vocal fold hemorrhage, scarring, or polyp formation. Laryngovideostroboscopy is the key in determining the functional significance of vocal fold varices. Management of patients with a varix includes medical therapy, speech therapy, and occasionally surgical vaporization. Indications for surgery are recurrent hemorrhage, enlargement of the varix, development of a mass in conjunction with the varix or hemorrhage, and unacceptable dysphonia after maximal medical and speech therapy due to a functionally significant varix.
Full Text Available Musical training during childhood has been linked to more robust encoding of sound later in life. We take this as evidence for an auditory reserve: a mechanism by which individuals capitalize on earlier life experiences to promote auditory processing. We assert that early auditory experiences guide how the reserve develops and is maintained over the lifetime. Experiences that occur after childhood, or which are limited in nature, are theorized to affect the reserve, although their influence on sensory processing may be less long-lasting and may potentially fade over time if not repeated. This auditory reserve may help to explain individual differences in how individuals cope with auditory impoverishment or loss of sensorineural function.
Baltus, Alina; Herrmann, Christoph Siegfried
Oscillatory EEG activity in the human brain with frequencies in the gamma range (approx. 30-80Hz) is known to be relevant for a large number of cognitive processes. Interestingly, each subject reveals an individual frequency of the auditory gamma-band response (GBR) that coincides with the peak in the auditory steady state response (ASSR). A common resonance frequency of auditory cortex seems to underlie both the individual frequency of the GBR and the peak of the ASSR. This review sheds light on the functional role of oscillatory gamma activity for auditory processing. For successful processing, the auditory system has to track changes in auditory input over time and store information about past events in memory which allows the construction of auditory objects. Recent findings support the idea of gamma oscillations being involved in the partitioning of auditory input into discrete samples to facilitate higher order processing. We review experiments that seem to suggest that inter-individual differences in the resonance frequency are behaviorally relevant for gap detection and speech processing. A possible application of these resonance frequencies for brain computer interfaces is illustrated with regard to optimized individual presentation rates for auditory input to correspond with endogenous oscillatory activity. This article is part of a Special Issue entitled SI: Auditory working memory. Copyright © 2015 Elsevier B.V. All rights reserved.
Full Text Available No other modality is more frequently represented in the prefrontal cortex than the auditory, but the role of auditory information in prefrontal functions is not well understood. Pathways from auditory association cortices reach distinct sites in the lateral, orbital, and medial surfaces of the prefrontal cortex in rhesus monkeys. Among prefrontal areas, frontopolar area 10 has the densest interconnections with auditory association areas, spanning a large antero-posterior extent of the superior temporal gyrus from the temporal pole to auditory parabelt and belt regions. Moreover, auditory pathways make up the largest component of the extrinsic connections of area 10, suggesting a special relationship with the auditory modality. Here we review anatomic evidence showing that frontopolar area 10 is indeed the main frontal auditory field as the major recipient of auditory input in the frontal lobe and chief source of output to auditory cortices. Area 10 is thought to be the functional node for the most complex cognitive tasks of multitasking and keeping track of information for future decisions. These patterns suggest that the auditory association links of area 10 are critical for complex cognition. The first part of this review focuses on the organization of prefrontal-auditory pathways at the level of the system and the synapse, with a particular emphasis on area 10. Then we explore ideas on how the elusive role of area 10 in complex cognition may be related to the specialized relationship with auditory association cortices.
Miller, Cory T; Thomas, A Wren; Nummela, Samuel U; de la Mothe, Lisa A
The role of primate frontal cortex in vocal communication and its significance in language evolution have a controversial history. While evidence indicates that vocalization processing occurs in ventrolateral prefrontal cortex neurons, vocal-motor activity has been conjectured to be primarily subcortical and suggestive of a distinctly different neural architecture from humans. Direct evidence of neural activity during natural vocal communication is limited, as previous studies were performed in chair-restrained animals. Here we recorded the activity of single neurons across multiple regions of prefrontal and premotor cortex while freely moving marmosets engaged in a natural vocal behavior known as antiphonal calling. Our aim was to test whether neurons in marmoset frontal cortex exhibited responses during vocal-signal processing and/or vocal-motor production in the context of active, natural communication. We observed motor-related changes in single neuron activity during vocal production, but relatively weak sensory responses for vocalization processing during this natural behavior. Vocal-motor responses occurred both prior to and during call production and were typically coupled to the timing of each vocalization pulse. Despite the relatively weak sensory responses a population classifier was able to distinguish between neural activity that occurred during presentations of vocalization stimuli that elicited an antiphonal response and those that did not. These findings are suggestive of the role that nonhuman primate frontal cortex neurons play in natural communication and provide an important foundation for more explicit tests of the functional contributions of these neocortical areas during vocal behaviors. Copyright © 2015 the American Physiological Society.
Welham, Nathan V.; Montequin, Douglas W.; Tateya, Ichiro; Tateya, Tomoko; Choi, Seong Hee; Bless, Diane M.
Purpose: To develop and evaluate a rat excised larynx model for the measurement of acoustic, aerodynamic, and vocal fold vibratory changes resulting from vocal fold scar. Method: Twenty-four 4-month-old male Sprague-Dawley rats were assigned to 1 of 4 experimental groups: chronic vocal fold scar, chronic vocal fold scar treated with 100-ng basic…
Mallur, Pavan S.; Rosen, Clark A.
Vocal fold injection is a procedure that has over a 100 year history but was rarely done as short as 20 years ago. A renaissance has occurred with respect to vocal fold injection due to new technologies (visualization and materials) and new injection approaches. Awake, un-sedated vocal fold injection offers many distinct advantages for the treatment of glottal insufficiency (vocal fold paralysis, vocal fold paresis, vocal fold atrophy and vocal fold scar). A review of materials available and ...
Leder, Steven B; Ross, Douglas A
This study prospectively investigated the incidence of vocal fold immobility, unilateral and bilateral, and its influence on aspiration status in a referred population of 1452 patients for a dysphagia evaluation from a large, urban, tertiary-care, teaching hospital. Main outcome measures included overall incidence of vocal fold immobility and aspiration status, with specific emphasis on age, etiology, and side of vocal fold immobility, i.e., right, left, or bilateral. Overall incidence of vocal fold immobility was 5.6% (81 of 1452 patients), including 47 males (mean age 55.7 yr) and 34 females (mean age 59.7 yr). In the subgroup of patients with vocal fold immobility, 31% (25 of 81) exhibited unilateral right, 60% (49 of 81) unilateral left, and 9% (7 of 81) bilateral impairment. Overall incidence of aspiration was found to be 29% (426 of 1452) of all patients referred for a swallow evaluation. Aspiration was observed in 44% (36 of 81) of patients presenting with vocal fold immobility, i.e., 44% (11 of 25) unilateral right, 43% (21 of 49) unilateral left, and 57% (4 of 7) bilateral vocal fold immobility. Left vocal fold immobility occurred most frequently due to surgical trauma. A liquid bolus was aspirated more often than a puree bolus. Side of vocal fold immobility and age were not factors that increased incidence of aspiration. In conclusion, vocal fold immobility, with an incidence of 5.6%, is not an uncommon finding in patients referred for a dysphagia evaluation in the acute-care setting, and vocal fold immobility, when present, was associated with a 15% increased incidence of aspiration when compared with a population already being evaluated for dysphagia.
Effect of neonatal asphyxia on the impairment of the auditory pathway by recording auditory brainstem responses in newborn piglets: a new experimentation model to study the perinatal hypoxic-ischemic damage on the auditory system.
Francisco Jose Alvarez
Full Text Available Hypoxia-ischemia (HI is a major perinatal problem that results in severe damage to the brain impairing the normal development of the auditory system. The purpose of the present study is to study the effect of perinatal asphyxia on the auditory pathway by recording auditory brain responses in a novel animal experimentation model in newborn piglets.Hypoxia-ischemia was induced to 1.3 day-old piglets by clamping 30 minutes both carotid arteries by vascular occluders and lowering the fraction of inspired oxygen. We compared the Auditory Brain Responses (ABRs of newborn piglets exposed to acute hypoxia/ischemia (n = 6 and a control group with no such exposure (n = 10. ABRs were recorded for both ears before the start of the experiment (baseline, after 30 minutes of HI injury, and every 30 minutes during 6 h after the HI injury.Auditory brain responses were altered during the hypoxic-ischemic insult but recovered 30-60 minutes later. Hypoxia/ischemia seemed to induce auditory functional damage by increasing I-V latencies and decreasing wave I, III and V amplitudes, although differences were not significant.The described experimental model of hypoxia-ischemia in newborn piglets may be useful for studying the effect of perinatal asphyxia on the impairment of the auditory pathway.
Full Text Available Background. The aim of the study was to assess the histological effects of autologous infusion of adipose-derived stem cells (ADSC on a chronic vocal fold scar in a rabbit model as compared to an untreated scar as well as in injection of hyaluronic acid. Study Design. Animal experiment. Method. We used 74 New Zealand rabbits. Sixteen of them were used as control/normal group. We created a bilateral vocal fold wound in the remaining 58 rabbits. After 18 months we separated our population into three groups. The first group served as control/scarred group. The second one was injected with hyaluronic acid in the vocal folds, and the third received an autologous adipose-derived stem cell infusion in the scarred vocal folds (ADSC group. We measured the variation of thickness of the lamina propria of the vocal folds and analyzed histopathologic changes in each group after three months. Results. The thickness of the lamina propria was significantly reduced in the group that received the ADSC injection, as compared to the normal/scarred group. The collagen deposition, the hyaluronic acid, the elastin levels, and the organization of elastic fibers tend to return to normal after the injection of ADSC. Conclusions. Autologous injection of adipose-derived stem cells on a vocal fold chronic scar enhanced the healing of the vocal folds and the reduction of the scar tissue, even when compared to other treatments.
Vassiliki, Kalodimou; Irini, Messini; Nikolaos, Psychalakis; Karampela, Eleftheria; Apostolos, Papalois
Background. The aim of the study was to assess the histological effects of autologous infusion of adipose-derived stem cells (ADSC) on a chronic vocal fold scar in a rabbit model as compared to an untreated scar as well as in injection of hyaluronic acid. Study Design. Animal experiment. Method. We used 74 New Zealand rabbits. Sixteen of them were used as control/normal group. We created a bilateral vocal fold wound in the remaining 58 rabbits. After 18 months we separated our population into three groups. The first group served as control/scarred group. The second one was injected with hyaluronic acid in the vocal folds, and the third received an autologous adipose-derived stem cell infusion in the scarred vocal folds (ADSC group). We measured the variation of thickness of the lamina propria of the vocal folds and analyzed histopathologic changes in each group after three months. Results. The thickness of the lamina propria was significantly reduced in the group that received the ADSC injection, as compared to the normal/scarred group. The collagen deposition, the hyaluronic acid, the elastin levels, and the organization of elastic fibers tend to return to normal after the injection of ADSC. Conclusions. Autologous injection of adipose-derived stem cells on a vocal fold chronic scar enhanced the healing of the vocal folds and the reduction of the scar tissue, even when compared to other treatments. PMID:26933440
Full Text Available Age-related hearing loss or presbycusis is a complex phenomenon consisting of elevation of hearing levels as well as changes in the auditory processing. It is commonly classified into four categories depending on the cause. Auditory brainstem responses (ABRs are a type of early evoked potentials recorded within the first 10 ms of stimulation. They represent the synchronized activity of the auditory nerve and the brainstem. Some of the changes that occur in the aging auditory system may significantly influence the interpretation of the ABRs in comparison with the ABRs of the young adults. The waves of ABRs are described in terms of amplitude, latencies and interpeak latency of the different waves. There is a tendency of the amplitude to decrease and the absolute latencies to increase with advancing age but these trends are not always clear due to increase in threshold with advancing age that act a major confounding factor in the interpretation of ABRs.
Jessica L Hanson
Full Text Available The laboratory mouse is an emerging model for context-dependent vocal signaling and reception. Mouse ultrasonic vocalizations are robustly produced in social contexts. In adults, male vocalization during courtship has become a model of interest for signal-receiver interactions. These vocalizations can be grouped into syllable types that are consistently produced by different subspecies and strains of mice. Vocalizations are unique to individuals, vary across development, and depend on social housing conditions. The behavioral significance of different syllable types, including the contexts in which different vocalizations are made and the responses listeners have to different types of vocalizations, is not well understood. We examined the effect of female presence and estrous state on male vocalizations by exploring the use of syllable types and the parameters of syllables during courtship. We also explored correlations between vocalizations and other behaviors. These experimental manipulations produced four main findings: 1 vocalizations varied among males, 2 the production of USVs and an increase in the use of a specific syllable type were temporally related to mounting behavior, 3 the frequency (kHz, bandwidth, and duration of syllables produced by males were influenced by the estrous phase of female partners, and 4 syllable types changed when females were removed. These findings show that mouse ultrasonic courtship vocalizations are sensitive to changes in female phase and presence, further demonstrating the context-sensitivity of these calls.
Krauter, K. G. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
Medical professionals can better serve their patients through continual update of their imaging tools. A wide range of pathologies and disease may afflict human vocal cords or, as they’re also known, vocal folds. These diseases can affect human speech hampering the ability of the patient to communicate. Vocal folds must be opened for breathing and the closed to produce speech. Currently methodologies to image markers of potential pathologies are difficult to use and often fail to detect early signs of disease. These current methodologies rely on a strobe light and slower frame rate camera in an attempt to obtain images as the vocal folds travel over the full extent of their motion.
Full Text Available This paper investigates the effectiveness of measures related to vocal tract characteristics in classifying normal and pathological speech. Unlike conventional approaches that mainly focus on features related to the vocal source, vocal tract characteristics are examined to determine if interaction effects between vocal folds and the vocal tract can be used to detect pathological speech. Especially, this paper examines features related to formant frequencies to see if vocal tract characteristics are affected by the nature of the vocal fold-related pathology. To test this hypothesis, stationary fragments of vowel /aa/ produced by 223 normal subjects, 472 vocal fold polyp subjects, and 195 unilateral vocal cord paralysis subjects are analyzed. Based on the acoustic-articulatory relationships, phonation for pathological subjects is found to be associated with measures correlated with a raised tongue body or an advanced tongue root. Vocal tract-related features are also found to be statistically significant from the Kruskal-Wallis test in distinguishing normal and pathological speech. Classification results demonstrate that combining the formant measurements with vocal fold-related features results in improved performance in differentiating vocal pathologies including vocal polyps and unilateral vocal cord paralysis, which suggests that measures related to vocal tract characteristics may provide additional information in diagnosing vocal disorders.
Ramos, Lorena de Almeida; Gama, Ana Cristina Côrtes
This study aimed to verify the effects of execution time on auditory-perceptual and acoustic responses in children with dysphonia completing straw phonation exercises. A randomized, prospective, comparative intra-subject study design was used. Twenty-seven children, ranging from 5 to 10 years of age, diagnosed with vocal cord nodules or cysts, were enrolled in the study. All subjects included in the Experimental Group were also included in the Control Group which involved complete voice rest. Sustained vowels (/a/e/ε/e/) counting from 1 to 10 were recorded before the exercises (m0) and then again after the first (m1), third (m3), fifth (m5), and seventh (m7) minutes of straw phonation exercises. The recordings were randomized and presented to five speech therapists, who evaluated vocal quality based on the Grade Roughness Breathiness Asthenia/Strain Instability scale. For acoustic analysis, fundamental frequency, jitter, shimmer, glottal to noise excitation ratio, and noise parameters were analyzed. Reduced roughness, breathiness, and noise measurements as well as increased glottal to noise excitation ratio were observed in the Experimental Group after 3 minutes of exercise. Reduced grade of dysphonia and breathiness were noted after 5 minutes. The ideal duration of straw phonation in children with dysphonia is from 3 to 5 minutes. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
An, Xingwei; Tang, Jiabei; Liu, Shuang; He, Feng; Qi, Hongzhi; Wan, Baikun; Ming, Dong
Combining visual and auditory stimuli in event-related potential (ERP)-based spellers gained more attention in recent years. Few of these studies notice the difference of ERP components and system efficiency caused by the shifting of visual and auditory onset. Here, we aim to study the effect of temporal congruity of auditory and visual stimuli onset on bimodal brain-computer interface (BCI) speller. We designed five visual and auditory combined paradigms with different visual-to-auditory delays (-33 to +100 ms). Eleven participants attended in this study. ERPs were acquired and aligned according to visual and auditory stimuli onset, respectively. ERPs of Fz, Cz, and PO7 channels were studied through the statistical analysis of different conditions both from visual-aligned ERPs and audio-aligned ERPs. Based on the visual-aligned ERPs, classification accuracy was also analyzed to seek the effects of visual-to-auditory delays. The latencies of ERP components depended mainly on the visual stimuli onset. Auditory stimuli onsets influenced mainly on early component accuracies, whereas visual stimuli onset determined later component accuracies. The latter, however, played a dominate role in overall classification. This study is important for further studies to achieve better explanations and ultimately determine the way to optimize the bimodal BCI application.
Schultz, Benjamin G
The Musical Instrument Digital Interface (MIDI) was readily adopted for auditory sensorimotor synchronization experiments. These experiments typically use MIDI percussion pads to collect responses, a MIDI-USB converter (or MIDI-PCI interface) to record responses on a PC and manipulate feedback, and
Stojmenova, Kristina; Sodnik, Jaka
There are 3 standardized versions of the Detection Response Task (DRT), 2 using visual stimuli (remote DRT and head-mounted DRT) and one using tactile stimuli. In this article, we present a study that proposes and validates a type of auditory signal to be used as DRT stimulus and evaluate the proposed auditory version of this method by comparing it with the standardized visual and tactile version. This was a within-subject design study performed in a driving simulator with 24 participants. Each participant performed 8 2-min-long driving sessions in which they had to perform 3 different tasks: driving, answering to DRT stimuli, and performing a cognitive task (n-back task). Presence of additional cognitive load and type of DRT stimuli were defined as independent variables. DRT response times and hit rates, n-back task performance, and pupil size were observed as dependent variables. Significant changes in pupil size for trials with a cognitive task compared to trials without showed that cognitive load was induced properly. Each DRT version showed a significant increase in response times and a decrease in hit rates for trials with a secondary cognitive task compared to trials without. Similar and significantly better results in differences in response times and hit rates were obtained for the auditory and tactile version compared to the visual version. There were no significant differences in performance rate between the trials without DRT stimuli compared to trials with and among the trials with different DRT stimuli modalities. The results from this study show that the auditory DRT version, using the signal implementation suggested in this article, is sensitive to the effects of cognitive load on driver's attention and is significantly better than the remote visual and tactile version for auditory-vocal cognitive (n-back) secondary tasks.
Full Text Available BACKGROUND: Although some molecules have been identified as responsible for human language disorders, there is still little information about what molecular mechanisms establish the faculty of human language. Since mice, like songbirds, produce complex ultrasonic vocalizations for intraspecific communication in several social contexts, they can be good mammalian models for studying the molecular basis of human language. Having found that cadherins are involved in the vocal development of the Bengalese finch, a songbird, we expected cadherins to also be involved in mouse vocalizations. METHODOLOGY/PRINCIPAL FINDINGS: To examine whether similar molecular mechanisms underlie the vocalizations of songbirds and mammals, we categorized behavioral deficits including vocalization in cadherin-6 knockout mice. Comparing the ultrasonic vocalizations of cadherin-6 knockout mice with those of wild-type controls, we found that the peak frequency and variations of syllables were differed between the mutant and wild-type mice in both pup-isolation and adult-courtship contexts. Vocalizations during male-male aggression behavior, in contrast, did not differ between mutant and wild-type mice. Open-field tests revealed differences in locomotors activity in both heterozygote and homozygote animals and no difference in anxiety behavior. CONCLUSIONS/SIGNIFICANCE: Our results suggest that cadherin-6 plays essential roles in locomotor activity and ultrasonic vocalization. These findings also support the idea that different species share some of the molecular mechanisms underlying vocal behavior.
Reddy, A Anita; Subramanian, Uma
Vibrato, a small, nevertheless an important component in the singing voice is known to enrich the overall singing voice quality. However, in the perception of overall performance, it is often neglected. Singing performance is often appreciated by a mixed audience of those who love music, but not necessarily sing and other singers who may or may not be teachers of singing. The objectives of the present study were aimed at investigating singers' and nonsingers' perception of vocal vibrato and its effect on the ratings of singer's overall performance. Prerecorded audio samples of the chorus of a hymn (How Great Thou Art) as sung by 10 singers (both men and women) were played via a speaker to two groups of judges which consisted of three experienced singers and three experienced nonsingers. The singer judges (SJs) were vocal instructors in Western classical, music theater, pop, and contemporary styles. Seven parameters (presence of vibrato, rate, extent, conspicuousness, quality, periodicity, and type) related to vibrato were evaluated through auditory perception by these two groups of judges on a rating scale developed specifically for the study, and one parameter evaluated singer's overall performance. Cohen's Kappa statistical analysis was used for inter-rater reliability within groups. Nonsinger judges (NSJs) within the group showed varied ratings as did SJs, yet SJs did have higher agreement than NSJs. Chi-square analysis was used across groups. Both groups were distinct from each other in their perception of vibrato. Ratings of singer's overall performance were not affected for NSJs, but certainly affected for SJ. It could not be concluded that ratings on singer's overall performance was affected as a result of vibrato. Since vibrato is often over-ridden by the singer's voice. But a rare occasion can arise where a vibrato may not sound pleasant and can affect the listener's perception of the singer's performance. Often a feedback from listeners would help monitor
Nemr, Kátia; Amar, Ali; Abrahão, Marcio; Leite, Grazielle Capatto de Almeida; Köhle, Juliana; Santos, Alexandra de O; Correa, Luiz Artur Costa
As a result of technology evolution and development, methods of voice evaluation have changed both in medical and speech and language pathology practice. To relate the results of perceptual evaluation, acoustic analysis and medical evaluation in the diagnosis of vocal and/or laryngeal affections of the population with vocal complaint. Clinical prospective. 29 people that attended vocal health protection campaign were evaluated. They were submitted to perceptual evaluation (AFPA), acoustic analysis (AA), indirect laryngoscopy (LI) and telelaryngoscopy (TL). Correlations between medical and speech language pathology evaluation methods were established, verifying possible statistical signification with the application of Fischer Exact Test. There were statistically significant results in the correlation between AFPA and LI, AFPA and TL, LI and TL. This research study conducted in a vocal health protection campaign presented correlations between speech language pathology evaluation and perceptual evaluation and clinical evaluation, as well as between vocal affection and/or laryngeal medical exams.
Full Text Available Several studies have explored brain computer interface (BCI systems based on auditory stimuli, which could help patients with visual impairments. Usability and user satisfaction are important considerations in any BCI. Although background music can influence emotion and performance in other task environments, and many users may wish to listen to music while using a BCI, auditory and other BCIs are typically studied without background music. Some work has explored the possibility of using polyphonic music in auditory BCI systems. However, this approach requires users with good musical skills, and has not been explored in online experiments. Our hypothesis was that an auditory BCI with background music would be preferred by subjects over a similar BCI without background music, without any difference in BCI performance. We introduce a simple paradigm (which does not require musical skill using percussion instrument sound stimuli and background music, and evaluated it in both offline and online experiments. The result showed that subjects preferred the auditory BCI with background music. Different performance measures did not reveal any significant performance effect when comparing background music vs. no background. Since the addition of background music does not impair BCI performance but is preferred by users, auditory (and perhaps other BCIs should consider including it. Our study also indicates that auditory BCIs can be effective even if the auditory channel is simultaneously otherwise engaged.
Sato, Kiminori; Umeno, Hirohito; Ono, Takeharu; Nakashima, Tadashi
Mechanotransduction caused by vocal fold vibration could possibly be an important factor in the maintenance of extracellular matrices and layered structure of the human adult vocal fold mucosa as a vibrating tissue after the layered structure has been completed. Vocal fold stellate cells (VFSCs) in the human maculae flavae of the vocal fold mucosa are inferred to be involved in the metabolism of extracellular matrices of the vocal fold mucosa. Maculae flavae are also considered to be an important structure in the growth and development of the human vocal fold mucosa. Tension caused by phonation (vocal fold vibration) is hypothesized to stimulate the VFSCs to accelerate production of extracellular matrices. A human adult vocal fold mucosa unphonated over a decade was investigated histopathologically. Vocal fold mucosa unphonated for 11 years and 2 months of a 64-year-old male with cerebral hemorrhage was investigated by light and electron microscopy. The vocal fold mucosae (including maculae flavae) were atrophic. The vocal fold mucosa did not have a vocal ligament, Reinke's space or a layered structure. The lamina propria appeared as a uniform structure. Morphologically, the VFSCs synthesized fewer extracellular matrices, such as fibrous protein and glycosaminoglycan. Consequently, VFSCs appeared to decrease their level of activity.
Guzman, M.; Laukkanen, A. M.; Krupa, P.; Horáček, Jaromír; Švec, J.G.; Geneid, A.
Roč. 27, č. 4 (2013), "523.e19"-"523.e34" ISSN 0892-1997 R&D Projects: GA ČR GAP101/12/1306 Institutional support: RVO:61388998 Keywords : vocal exercises * resonance tube * vocal tract impedance * computerized tomography * singer’s/speaker’s formant cluster Subject RIV: BI - Acoustics Impact factor: 0.944, year: 2013 http://www.sciencedirect.com/science/journal/08921997
Rubin, F; Villeneuve, A; Alciato, L; Slaïm, L; Bonfils, P; Laccourreye, O
To analyze the characteristics of adult idiopathic unilateral vocal-fold paralysis. Retrospective study of diagnostic problems, clinical data and recovery in an inception cohort of 100 adult patients with idiopathic unilateral vocal-fold paralysis (Group A) and comparison with a cohort of 211 patients with isolated non-idiopathic non-traumatic unilateral vocal-fold paralysis (Group B). Diagnostic problems were noted in 24% of cases in Group A: eight patients with concomitant common upper aerodigestive tract infection, five patients with a concomitant condition liable to induce immunodepression and 11 patients in whom a malignant tumor occurred along the path of the ipsilateral vagus and inferior laryngeal nerves or in the ipsilateral paralyzed larynx. There was no recovery of vocal-fold motion beyond 51 months after onset of paralysis. The 5-year actuarial estimate for recovery differed significantly (Pvocal-fold paralysis. In non-traumatic vocal-fold paralysis in adult patients, without recovery of vocal-fold motion, a minimum three years' regular follow-up is recommended. Copyright © 2018 Elsevier Masson SAS. All rights reserved.
Hackworth, Rhonda S.
This study examined the effects of vocal hygiene and behavior modification instruction on self-reported behaviors of music teachers. Subjects (N = 76) reported daily behaviors for eight weeks: water consumption, warm-up, talking over music/noise, vocal rest, nonverbal commands, and vocal problems. Subjects were in experimental group 1 or 2, or the…
The Central Auditory Processing Kit[TM]. Book 1: Auditory Memory [and] Book 2: Auditory Discrimination, Auditory Closure, and Auditory Synthesis [and] Book 3: Auditory Figure-Ground, Auditory Cohesion, Auditory Binaural Integration, and Compensatory Strategies.
Mokhemar, Mary Ann
This kit for assessing central auditory processing disorders (CAPD), in children in grades 1 through 8 includes 3 books, 14 full-color cards with picture scenes, and a card depicting a phone key pad, all contained in a sturdy carrying case. The units in each of the three books correspond with auditory skill areas most commonly addressed in…
Feng, Ling; Nielsen, Andreas Brinch; Hansen, Lars Kai
This paper explores the vocal and non-vocal music classification problem within popular songs. A newly built labeled database covering 147 popular songs is announced. It is designed for classifying signals from 1sec time windows. Features are selected for this particular task, in order to capture...
Wang, Dongcui; Mo, Fongming; Zhang, Yangde; Yang, Chao; Liu, Jun; Chen, Zhencheng; Zhao, Jinfeng
In a previous study (unpublished), Emotiv headset was validated for capturing event-related potentials (ERPs) from normal subjects. In the present follow-up study, the signal quality of Emotiv headset was tested by the accuracy rate of discriminating Major Depressive Disorder (MDD) patients from the normal subjects. ERPs of 22 MDD patients and 15 normal subjects were induced by an auditory oddball task and the amplitude of N1, N2 and P3 of ERP components were specifically analyzed. The features of ERPs were statistically investigated. It is found that Emotiv headset is capable of discriminating the abnormal N1, N2 and P3 components in MDD patients. Relief-F algorithm was applied to all features for feature selection. The selected features were then input to a linear discriminant analysis (LDA) classifier with leave-one-out cross-validation to characterize the ERP features of MDD. 127 possible combinations out of the selected 7 ERP features were classified using LDA. The best classification accuracy was achieved to be 89.66%. These results suggest that MDD patients are identifiable from normal subjects by ERPs measured by Emotiv headset.
Haas, Michael W.
The term Fusion Interface is defined as a class of interface which integrally incorporates both virtual and nonvirtual concepts and devices across the visual, auditory, and haptic sensory modalities. A fusion interface is a multisensory virtually-augmented synthetic environment. A new facility has been developed within the Human Engineering Division of the Armstrong Laboratory dedicated to exploratory development of fusion interface concepts. This new facility, the Fusion Interfaces for Tactical Environments (FITE) Facility is a specialized flight simulator enabling efficient concept development through rapid prototyping and direct experience of new fusion concepts. The FITE Facility also supports evaluation of fusion concepts by operation fighter pilots in an air combat environment. The facility is utilized by a multidisciplinary design team composed of human factors engineers, electronics engineers, computer scientists, experimental psychologists, and oeprational pilots. The FITE computational architecture is composed of twenty-five 80486-based microcomputers operating in real-time. The microcomputers generate out-the-window visuals, in-cockpit and head-mounted visuals, localized auditory presentations, haptic displays on the stick and rudder pedals, as well as executing weapons models, aerodynamic models, and threat models.
Full Text Available The extent to which auditory experience can shape general auditory perceptual abilities is still under constant debate. Some studies show that specific auditory expertise may have a general effect on auditory perceptual abilities, while others show a more limited influence, exhibited only in a relatively narrow range associated with the area of expertise. The current study addresses this issue by examining experience-dependent enhancement in perceptual abilities in the auditory domain. Three experiments were performed. In the first experiment, 12 pop and rock musicians and 15 non-musicians were tested in frequency discrimination (DLF, intensity discrimination, spectrum discrimination (DLS, and time discrimination (DLT. Results showed significant superiority of the musician group only for the DLF and DLT tasks, illuminating enhanced perceptual skills in the key features of pop music, in which miniscule changes in amplitude and spectrum are not critical to performance. The next two experiments attempted to differentiate between generalization and specificity in the influence of auditory experience, by comparing subgroups of specialists. First, seven guitar players and eight percussionists were tested in the DLF and DLT tasks that were found superior for musicians. Results showed superior abilities on the DLF task for guitar players, though no difference between the groups in DLT, demonstrating some dependency of auditory learning on the specific area of expertise. Subsequently, a third experiment was conducted, testing a possible influence of vowel density in native language on auditory perceptual abilities. Ten native speakers of German (a language characterized by a dense vowel system of 14 vowels, and 10 native speakers of Hebrew (characterized by a sparse vowel system of five vowels, were tested in a formant discrimination task. This is the linguistic equivalent of a DLS task. Results showed that German speakers had superior formant
Neumann-Werth, Yael; Levy, Erika S; Obler, Loraine K
Vocal emblems, such as shh and brr, are speech sounds that have linguistic and nonlinguistic features; thus, it is unclear how they are processed in the brain. Five adult dextral individuals with left-brain damage and moderate-severe Wernicke's aphasia, five adult dextral individuals with right-brain damage, and five Controls participated in two tasks: (1) matching vocal emblems to photographs ('picture task') and (2) matching vocal emblems to verbal translations ('phrase task'). Cross-group statistical analyses on items on which the Controls performed at ceiling revealed lower accuracy by the group with left-brain damage (than by Controls) on both tasks, and lower accuracy by the group with right-brain damage (than by Controls) on the picture task. Additionally, the group with left-brain damage performed significantly less accurately than the group with right-brain damage on the phrase task only. Findings suggest that comprehension of vocal emblems recruits more left- than right-hemisphere processing.
Schultz, E.E. Jr.; Johnson, G.L.
The extensive installation of computerized safety Parameter Display Systems (SPDSs) in nuclear power plants since the Three-Mile Island accident has enhanced plant safety. It has also raised new issues of how best to ensure an effective interface between human operators and the plant via computer systems. New developments in interface technologies since the current generation of SPDSs was installed can contribute to improving display interfaces. These technologies include new input devices, three-dimensional displays, delay indicators, and auditory displays. Examples of how they might be applied to improve current SPDSs are given. These examples illustrate how the new use interface technology could be applied to future nuclear plant displays
Geberzahn, Nicole; Aubin, Thierry
Vocal performance refers to the ability to produce vocal signals close to physical limits. Such motor skills can be used by conspecifics to assess a signaller's competitive potential. For example it is difficult for birds to produce repeated syllables both rapidly and with a broad frequency bandwidth. Deviation from an upper-bound regression of frequency bandwidth on trill rate has been widely used to assess vocal performance. This approach is, however, only applicable to simple trilled songs, and even then may be affected by differences in syllable complexity. Using skylarks (Alauda arvensis) as a birdsong model with a very complex song structure, we detected another performance trade-off: minimum gap duration between syllables was longer when the frequency ratio between the end of one syllable and the start of the next syllable (inter-syllable frequency shift) was large. This allowed us to apply a novel measure of vocal performance ¿ vocal gap deviation: the deviation from a lower-bound regression of gap duration on inter-syllable frequency shift. We show that skylarks increase vocal performance in an aggressive context suggesting that this trait might serve as a signal for competitive potential. We suggest using vocal gap deviation in future studies to assess vocal performance in songbird species with complex structure.
Full Text Available The physical shape of vocal tract and its formant (resonant frequencies are directly related. The study of this functional connectivity is essential in speech therapy practice with children. Most of the perceived children’s speech anomalies can be explained on a physical level: malfunctioning movement of articulation organs. The current problem is that there is no enough data on the anatomical shape of children’s vocal tract to create its acoustic model. Classical techniques for vocal tract shape imaging (X-ray, magnetic resonance, etc. are not appropriate for children. One possibility is to start from the shape of the adult vocal tract and correct it based on anatomical, morphological and articulatory differences between children and adults. This paper presents a method for vocal tract shape estimation of the child aged one year. The initial shapes of the vocal tract refer to the Russian vowels spoken by an adult male. All the relevant anatomical and articulation parameters, that influence the formant frequencies, are analyzed. Finally, the hypothetical configurations of the children’s vocal tract, for the five vowels, are presented.
Erath, Byron; Pirnia, Alireza; Peterson, Sean
During voiced speech a critical pressure is produced in the lungs that separates the vocal folds and creates a passage (the glottis) for airflow. As air passes through the vocal folds the resulting aerodynamic loading, coupled with the tissue properties of the vocal folds, produces self-sustained oscillations. Throughout each cycle a complex flow field develops, characterized by a plethora of viscous flow phenomena. Air passing through the glottis creates a jet, with periodically-shed vortices developing due to flow separation and the Kelvin-Helmholtz instability in the shear layer. These vortices have been hypothesized to be a crucial mechanism for producing vocal fold vibrations. In this study the effect of vortices on the vocal fold dynamics is investigated experimentally by passing a vortex ring over a flexible beam with the same non-dimensional mechanical properties as the vocal folds. Synchronized particle image velocimetry data are acquired in tandem with the beam dynamics. The resulting impact of the vortex ring loading on vocal fold dynamics is discussed in detail. This work was supported by the National Science Foundation Grant CBET #1511761.
Begault, Durand R.; Bittner, Rachel M.; Anderson, Mark R.
Auditory communication displays within the NextGen data link system may use multiple synthetic speech messages replacing traditional ATC and company communications. The design of an interface for selecting amongst multiple incoming messages can impact both performance (time to select, audit and release a message) and preference. Two design factors were evaluated: physical pressure-sensitive switches versus flat panel "virtual switches", and the presence or absence of auditory feedback from switch contact. Performance with stimuli using physical switches was 1.2 s faster than virtual switches (2.0 s vs. 3.2 s); auditory feedback provided a 0.54 s performance advantage (2.33 s vs. 2.87 s). There was no interaction between these variables. Preference data were highly correlated with performance.
Sarah A Klemuk
Full Text Available Vocal production requires active control of the respiratory system, larynx and vocal tract. Vocal sounds in mammals are produced by flow-induced vocal fold oscillation, which requires vocal fold tissue that can sustain the mechanical stress during phonation. Our understanding of the relationship between morphology and vocal function of vocal folds is very limited. Here we tested the hypothesis that vocal fold morphology and viscoelastic properties allow a prediction of fundamental frequency range of sounds that can be produced, and minimal lung pressure necessary to initiate phonation. We tested the hypothesis in lions and tigers who are well-known for producing low frequency and very loud roaring sounds that expose vocal folds to large stresses. In histological sections, we found that the Panthera vocal fold lamina propria consists of a lateral region with adipocytes embedded in a network of collagen and elastin fibers and hyaluronan. There is also a medial region that contains only fibrous proteins and hyaluronan but no fat cells. Young's moduli range between 10 and 2000 kPa for strains up to 60%. Shear moduli ranged between 0.1 and 2 kPa and differed between layers. Biomechanical and morphological data were used to make predictions of fundamental frequency and subglottal pressure ranges. Such predictions agreed well with measurements from natural phonation and phonation of excised larynges, respectively. We assume that fat shapes Panthera vocal folds into an advantageous geometry for phonation and it protects vocal folds. Its primary function is probably not to increase vocal fold mass as suggested previously. The large square-shaped Panthera vocal fold eases phonation onset and thereby extends the dynamic range of the voice.
Shen, Yi; Sivakumar, Rajeswari; Richards, Virginia M.
A Bayesian adaptive procedure, the quick-auditory-filter (qAF) procedure, was used to estimate auditory-filter shapes that were asymmetric about their peaks. In three experiments, listeners who were naive to psychoacoustic experiments detected a fixed-level, pure-tone target presented with a spectrally notched noise masker. The qAF procedure adaptively manipulated the masker spectrum level and the position of the masker notch, which was optimized for the efficient estimation of the five parameters of an auditory-filter model. Experiment I demonstrated that the qAF procedure provided a convergent estimate of the auditory-filter shape at 2 kHz within 150 to 200 trials (approximately 15 min to complete) and, for a majority of listeners, excellent test-retest reliability. In experiment II, asymmetric auditory filters were estimated for target frequencies of 1 and 4 kHz and target levels of 30 and 50 dB sound pressure level. The estimated filter shapes were generally consistent with published norms, especially at the low target level. It is known that the auditory-filter estimates are narrower for forward masking than simultaneous masking due to peripheral suppression, a result replicated in experiment III using fewer than 200 qAF trials. PMID:25324086
Allen, Jacqui E; Belafsky, Peter C
Promising new techniques in the management of vocal fold nodules have been developed in the past 2 years. Simultaneously, the therapeutic use of botulinum toxin has rapidly expanded. This review explores the use of botulinum toxin in treatment of vocal nodules and summarizes current therapeutic concepts. New microsurgical instruments and techniques, refinements in laser technology, radiosurgical excision and steroid intralesional injections are all promising new techniques in the management of vocal nodules. Botulinum toxin-induced 'voice rest' is a new technique we have employed in patients with recalcitrant nodules. Successful resolution of nodules is possible with this technique, without the risk of vocal fold scarring inherent in dissection/excision techniques. Botulinum toxin usage is exponentially increasing, and large-scale, long-term studies demonstrate its safety profile. Targeted vocal fold temporary paralysis induced by botulinum toxin injection is a new, well tolerated and efficacious treatment in patients with persistent vocal fold nodules.
Christopher T Noto
Full Text Available A major problem facing behavioral neuroscientists is a lack of unified, vendor-distributed data acquisition systems that allow stimulus presentation and behavioral monitoring while recording neural activity. Numerous systems perform one of these tasks well independently, but to our knowledge, a useful package with a straightforward user interface does not exist. Here we describe the development of a flexible, script-based user interface that enables customization for real-time stimulus presentation, behavioral monitoring and data acquisition. The experimental design can also incorporate neural microstimulation paradigms. We used this interface to deliver multimodal, auditory and visual (images or video stimuli to a nonhuman primate and acquire single-unit data. Our design is cost-effective and works well with commercially available hardware and software. Our design incorporates a script, providing high-level control of data acquisition via a sequencer running on a digital signal processor to enable behaviorally triggered control of the presentation of visual and auditory stimuli. Our experiments were conducted in combination with eye-tracking hardware. The script, however, is designed to be broadly useful to neuroscientists who may want to deliver stimuli of different modalities using any animal model.
Hackworth, Rhonda S.
The current study sought to determine the relationship among music teachers' length of teaching experience, specialty (vocal or instrumental), and ratings of behaviors and teaching activities related to vocal health. Participants (N = 379) were experienced (n = 208) and preservice (n = 171) music teachers, further categorized by specialty, either…
Nathan V Welham
Full Text Available The vocal fold mucosa is a biomechanically unique tissue comprised of a densely cellular epithelium, superficial to an extracellular matrix (ECM-rich lamina propria. Such ECM-rich tissues are challenging to analyze using proteomic assays, primarily due to extensive crosslinking and glycosylation of the majority of high M(r ECM proteins. In this study, we implemented an LC-MS/MS-based strategy to characterize the rat vocal fold mucosa proteome. Our sample preparation protocol successfully solubilized both proteins and certain high M(r glycoconjugates and resulted in the identification of hundreds of mucosal proteins. A straightforward approach to the treatment of protein identifications attributed to single peptide hits allowed the retention of potentially important low abundance identifications (validated by a cross-sample match and de novo interpretation of relevant spectra while still eliminating potentially spurious identifications (global single peptide hits with no cross-sample match. The resulting vocal fold mucosa proteome was characterized by a wide range of cellular and extracellular proteins spanning 12 functional categories.
Rosen, Clark A; Gartner-Schmidt, Jackie; Hathaway, Bridget; Simpson, C Blake; Postma, Gregory N; Courey, Mark; Sataloff, Robert T
There is a significant lack of uniform agreement regarding nomenclature for benign vocal fold lesions (BVFLs). This confusion results in difficulty for clinicians communicating with their patients and with each other. In addition, BVFL research and comparison of treatment methods are hampered by the lack of a detailed and uniform BVFL nomenclature. Clinical consensus conferences were held to develop an initial BVFL nomenclature paradigm. Perceptual video analysis was performed to validate the stroboscopy component of the paradigm. The culmination of the consensus conferences and the video-perceptual analysis was used to evaluate the BVFL nomenclature paradigm using a retrospective review of patients with BVFL. An initial BVFL nomenclature paradigm was proposed utilizing detailed definitions relating to vocal fold lesion morphology, stroboscopy, response to voice therapy and intraoperative findings. Video-perceptual analysis of stroboscopy demonstrated that the proposed binary stroboscopy system used in the BVFL nomenclature paradigm was valid and widely applicable. Retrospective review of 45 patients with BVFL followed to the conclusion of treatment demonstrated that slight modifications of the initial BVFL nomenclature paradigm were required. With the modified BVFL nomenclature paradigm, 96% of the patients fit into the predicted pattern and definitions of the BVFL nomenclature system. This study has validated a multidimensional BVFL nomenclature paradigm. This vocal fold nomenclature paradigm includes nine distinct vocal fold lesions: vocal fold nodules, vocal fold polyp, pseudocyst, vocal fold cyst (subepithelial or ligament), nonspecific vocal fold lesion, vocal fold fibrous mass (subepithelial or ligament), and reactive lesion. Copyright © 2011 The American Laryngological, Rhinological, and Otological Society, Inc.
Full Text Available Introduction: -The larynx is an air passage and a sphincteric device used in respiration and phonation. The larynx, from inside outwards has a framework of mucosa surrounded by fibro-elastic membrane which in turn is surrounded by cartilages and then a layer of muscles. Vocal folds are intrinsic ligament of larynx covered by mucosal folds. Larynx generates sound through rhythmic opening and closing of the vocal folds. The perceived pitch of human voice mainly depends upon fundamental frequency of sound generated by larynx. Aim: - The aim of present study is to measure various dimensions of vocal folds in Indian cadavers. Material & Methods: - 50 larynx were obtained from embalmed cadavers, of which 10 larynx were of females. Vocal cords were dissected from the larynx and morphometric analysis was done. Results and Conclusions: - The average total length of the vocal folds was found to be 16.11 mm. ± 2.62 mm. in male and 14.10 mm. ± 1.54 mm. in female cadavers. The average width of the vocal folds was found to be 4.38 mm. ± 0.74 mm. in male and 3.60 mm. ± 0.64 mm. in female cadavers. The average total length of the membranous part of the vocal folds was found to be 11.90 mm. ± 1.86 mm. in male and 10.45 mm. ± 1.81 mm. in female cadavers. The average ratio of the length of the membranous and the cartilaginous parts of the vocal folds was calculated to be 3.10 ± 0.96in male and 2.85 ± 0.73in female cadavers.
Paul, Nilanjan; Kumar, Suman; Chatterjee, Indranil; Mukherjee, Biswarup
In-depth study on laryngeal biomechanics and vocal fold vibratory patterns reveal that a single vibratory cycle can be divided into two major phases, the closed and open phase, which is subdivided into opening and closing phases. Studies reveal that the relative time course of abduction and adduction, which in turn is dependent on the relative relaxing and tensing of the vocal fold cover and body, to be the determining factor in production of a particular vocal register like the modal (or chest), falsetto, glottal fry registers. Studies further point out Electroglottography to be particularly suitable for the study of vocal vibratory patterns during register changes. However, to date, there has been limited study on quantitative parameterization of EGG wave form in vocal fry register. Moreover, contradictory findings abound in literature regarding effects of gender and vowel types on vocal vibratory patterns, especially during phonation at different registers. The present study endeavors to find out the effects of vowel and gender differences on the vocal fold vibratory patterns in different registers and how these would be reflected in standard EGG parameters of Contact Quotient (CQ) and Contact Index (CI), taking into consideration the Indian sociolinguistic context. Electroglottographic recordings of 10 young adults (5 males and 5 females) were taken while the subjects phonated the three vowels /a/,/i/,/u/ each in two vocal registers, modal and vocal fry. Obtained raw EGG were further normalized using the Derived EGG algorithm and theCQ and CI values were derived. Obtained data were subject to statistical analysis using the 3-way ANOVA with gender, vowel and vocal register as the three variables. Post-hoc Dunnett C multiple comparison analysis were also performed. Results reveal that CQ values are significantly higher in vocal fry than modal phonation for both males and females, indicating a relatively hyperconstricted vocal system during vocal fry. The males
More than half of patients presenting with hoarseness show benign vocal fold changes. The clinician should be familiar with the anatomy, physiology and functional aspects of voice disorders and also the modern diagnostic and therapeutic possibilities in order to ensure an optimal and patient specific management. This review article focuses on the diagnostic and therapeutic limitations and difficulties of treatment of benign vocal fold tumors, the management and prevention of scarred vocal folds and the issue of unilateral vocal fold paresis. PMID:24403969
Lin, R Jun; Smith, Libby J; Munin, Michael C; Sridharan, Shaum; Rosen, Clark A
Treatment options for symptomatic unilateral vocal fold paralysis (VFP) include vocal fold augmentation, laryngeal framework surgery, and laryngeal reinnervation. Laryngeal reinnervation (LR) has been suggested to provide "tone" to the paralyzed VF. This implies a loss of tone as a result of denervation without reinnervation. We performed laryngeal electromyography (LEMG) in patients with chronic VFP to understand the innervation status associated with a chronically paralyzed vocal fold. Retrospective review of LEMG data in adult patients with chronic VFP from January 2009 to December 2014. LEMG was performed at least 6 months after-onset of VFP. Qualitative LEMG, quantitative LEMG, and adductory synkinesis testing were performed, and the parameters were collected. Twenty-seven vocal folds were studied (23 unilateral VFP and 2 bilateral VFP). Average age was 59 ± 17 years. The median duration from recurrent laryngeal nerve injury to LEMG was 8.5 months (range 6-90 months). The majority of patients, 24 of 27 (89%), had motor unit potentials during phonation tasks on LEMG, and only 3 of 27 (11%) patients were electrically silent. Quantitative LEMG showed 287.8 mean turns per second (normal ≥ 400). Motor unit configuration was normal in 12 of 27 (44%), polyphasic in 12 of 27 (44%), and absent in the electrically silent patients. Adductory synkinesis was found in 6 of 20 (30%) patients. Chronic vocal fold paralysis is infrequently associated with absent motor-unit recruitment, indicating some degree of preserved innervation and/or reinnervation in these patients. LEMG should be part of the routine workup for chronic VFP prior to consideration of LR. 4. Laryngoscope, 2018. © 2018 The American Laryngological, Rhinological and Otological Society, Inc.
Karin Zazo Ortiz
Full Text Available OBJETIVO: Comparar os dados da análise perceptivo-auditiva (subjetiva com os dados da análise acústica (objetiva. MÉTODOS: Quarenta e dois pacientes disártricos, com diagnósticos neurológicos definidos, 21 do sexo masculino e 21 do sexo feminino foram submetidos à análise perceptual-auditiva e acústica. Todos os pacientes foram submetidos à gravação da voz, tendo sido avaliados, na análise auditiva, tipo de voz, ressonância (equilibrada, hipernasal ou laringo-faríngea, loudness (adequado, diminuído ou aumentado, pitch (adequado, grave, agudo ataque vocal (isocrônico, brusco ou soproso, e estabilidade (estável ou instável. Para a análise acústica foram utilizados os programas GRAM 5.1.7; para a análise da qualidade vocal e comportamento dos harmônicos na espectrografia e o Programa Vox Metria, para a obtenção das medidas objetivas. RESULTADOS: A comparação entre os achados das análises auditiva e acústica em sua maioria não foi significante, ou seja, não houve uma relação direta entre os achados subjetivos e os dados objetivos. Houve diferença estatisticamente significante apenas entre voz soprosa e Shimmer alterado (p=0,048 e entre a definição dos harmônicos e voz soprosa (p=0,040, sendo assim, observou-se correlação entre a presença de ruído à emissão e soprosidade. CONCLUSÕES: As análises perceptual-auditiva e acústica forneceram dados diferentes, porém complementares, auxiliando, de forma conjunta, no diagnóstico clínico das disartrias.PURPOSE: To compare data found in auditory-perceptual analyses (subjective and acoustic analyses (objective in dysarthric patients. METHODS: Forty-two patients with well defined neurological diagnosis, 21 male and 21 female, were evaluated in auditory-perceptual parameters and acoustic measures. All patients had their voices recorded. Auditory-perceptual voice analyses were made considering type of voice, resonance (balanced, hipernasal or laryngopharyngeal
Maria Cláudia Mendes Caminha Muniz
Full Text Available Objective: To present genres and styles currently running on western music scene, focusing on the practice of singing voice. Methods: An observational and documental study for which were selected sound sources presenting musical genres and styles that are part of the experience of the researchers, which were analyzed considering origins, formative elements and vocal features. Alongside we carried out a review of literature grounded in databases research and free review of websites and classical books of the area. Results: The selected styles (Rock and Roll, Heavy Metal, Trash Metal, Grunge, Gothic Metal, Rap, Funk, Blues, R&B – Rhythm and Blues, Soul, Gospel, MPB, Samba, Forro, Sertanejo, Bossa Nova, Opera and Chamber Music were described, pointing the reasons for the speech therapist to be informed about them and about singing voice aspects. His guidance may minimize possible vocal damage caused by each style, since each of them carries its own patterns to which the interpreter must submit. Conclusions: We conclude that the singer will use a specific vocal pattern that resembles the musical style he intends to sing, regardless of any harm it may or may not cause to vocal health. When choosing a musical style, it is important that the singer has the knowledge and understanding of how the use of his vocal apparatus will cause or not cause injury to his voice. Also be aware that the technique in singing is necessary for vocal longevity.
Full Text Available Hemangioma is one of the most common benign tumorsin the head and neck region. Laryngeal hemangiomasare benign vascular tumors of unknown etiology thatarise from subglottic region with stridor in infants. Thistype also known as congenital laryngeal hemangioma, isthe more common. Congenital hemangiomas occur usuallyin subglottic region and more frequent in girls. Laryngealhemangioma in adults is a very rare conditionand main symptom is hoarseness and breathing difficulties.Adult hemangiomas can be seen in different locationssuch as the epiglottis, aryepiglottic folds, arytenoidsand false and true vocal cords. They are more oftenof cavernous form and cause hoarseness. In this reportwe present an adult patient with hemangioma ofthe left vocal fold and review the literature. Diagnosticinvestigation revealed a pink-purple mass which was extendedfrom the anterior comissure to the posterior partof true vocal cord and false vocal cord, filling the ventriculeand extending to supraglottic region. Directlaryngoscopy was performed, but the lesion was not excisedbecause of its widespread extension in the larynx. JClin Exp Invest 2010; 2(1: 91-94
Hapner, Edie; Gilman, Marina
Jewish cantors comprise a subset of vocal professionals that is not well understood by vocal health professionals. This study aimed to document the vocal demands, vocal training, reported incidence of voice problems, and treatment-seeking behavior of Reform Jewish cantors. The study used a prospective observational design to anonymously query Reform Jewish cantors using a 35-item multiple-choice survey distributed online. Demographic information, medical history, vocal music training, cantorial duties, history of voice problems, and treatment-seeking behavior were addressed. Results indicated that many of the commonly associated risk factors for developing voice disorders were present in this population, including high vocal demands, reduced vocal downtime, allergies, and acid reflux. Greater than 65% of the respondents reported having had a voice problem that interfered with their ability to perform their duties at some time during their careers. Reform Jewish cantors are a population of occupational voice users who may be currently unidentified and underserved by vocal health professionals. The results of the survey suggest that Reform Jewish cantors are occupational voice users and are at high risk for developing voice disorders. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Humbert, Ianessa A; Poletto, Christopher J; Saxon, Keith G; Kearney, Pamela R; Ludlow, Christy L
Closure of the true and false vocal folds is a normal part of airway protection during swallowing. Individuals with reduced or delayed true vocal fold closure can be at risk for aspiration and may benefit from intervention to ameliorate the problem. Surface electrical stimulation is currently used during therapy for dysphagia, despite limited knowledge of its physiological effects. Prospective single effects study. The immediate physiological effect of surface stimulation on true vocal fold angle was examined at rest in 27 healthy adults using 10 different electrode placements on the submental and neck regions. Fiberoptic nasolaryngoscopic recordings during passive inspiration were used to measure change in true vocal fold angle with stimulation. Vocal fold angles changed only to a small extent during two electrode placements (P vocal fold abduction was 2.4 degrees; while horizontal placements of electrodes in the submental region produced a mean adduction of 2.8 degrees (P = .03). Surface electrical stimulation to the submental and neck regions does not produce immediate true vocal fold adduction adequate for airway protection during swallowing, and one position may produce a slight increase in true vocal fold opening.
Yin, Jun; Zhang, Zhaoyan
The influence of the thyroarytenoid (TA) and cricothyroid (CT) muscle activation on vocal fold stiffness and eigenfrequencies was investigated in a muscularly controlled continuum model of the vocal folds. Unlike the general understanding that vocal fold fundamental frequency was determined by vocal fold tension, this study showed that vocal fold eigenfrequencies were primarily determined by vocal fold stiffness. This study further showed that, with reference to the resting state of zero strain, vocal fold stiffness in both body and cover layers increased with either vocal fold elongation or shortening. As a result, whether vocal fold eigenfrequencies increased or decreased with CT/TA activation depended on how the CT/TA interaction influenced vocal fold deformation. For conditions of strong CT activation and thus an elongated vocal fold, increasing TA contraction reduced the degree of vocal fold elongation and thus reduced vocal fold eigenfrequencies. For conditions of no CT activation and thus a resting or slightly shortened vocal fold, increasing TA contraction increased the degree of vocal fold shortening and thus increased vocal fold eigenfrequencies. In the transition region of a slightly elongated vocal fold, increasing TA contraction first decreased and then increased vocal fold eigenfrequencies. PMID:23654401
Tibbetts, Kathleen M; Dominguez, Laura M; Simpson, C Blake
Vocal fold cysts are benign mid-membranous lesions of the true vocal fold, classified as mucus retention or epidermal inclusion cysts. Treatment is surgical excision with or without postoperative voice therapy. A retrospective review was performed of the demographics, treatment approach, and outcomes of patients treated for vocal fold cysts between 2009 and 2014. Voice Handicap Index (VHI)-10 scores before and after treatment were compared using the Wilcoxon Rank-Sum test and the two-tailed Student's t test. Videostroboscopy examinations were reviewed for posttreatment changes in vibratory characteristics of the vocal folds. Twenty-five patients were identified, and one was excluded for incomplete records. Mean age was 41.9 years (66.7% female), and mean follow-up time was 5.58 months. Microflap excision was pursued by 21/24 (87.5%) patients, with 14 patients (58.3%) undergoing perioperative voice therapy. One cyst recurred. Two patients elected for observation, and their cysts persisted. VHI-10 decreased from 23.8 to 6.6 (P Vocal fold cysts impact mucosal wave and glottic closure. Surgical excision resulted in low rates of recurrence, and in improvement in the mucosal wave and VHI-10. Perioperative voice therapy did not offer a significant benefit. Mucus retention cysts were the majority, in contrast to other published studies. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Colombo, Michael; D'Amato, Michael R.; Rodman, Hillary R.; Gross, Charles G.
Monkeys that were trained to perform auditory and visual short-term memory tasks (delayed matching-to-sample) received lesions of the auditory association cortex in the superior temporal gyrus. Although visual memory was completely unaffected by the lesions, auditory memory was severely impaired. Despite this impairment, all monkeys could discriminate sounds closer in frequency than those used in the auditory memory task. This result suggests that the superior temporal cortex plays a role in auditory processing and retention similar to the role the inferior temporal cortex plays in visual processing and retention.
Belyk, Michel; Pfordresher, Peter Q; Liotti, Mario; Brown, Steven
Vocal imitation is a phenotype that is unique to humans among all primate species, and so an understanding of its neural basis is critical in explaining the emergence of both speech and song in human evolution. Two principal neural models of vocal imitation have emerged from a consideration of nonhuman animals. One hypothesis suggests that putative mirror neurons in the inferior frontal gyrus pars opercularis of Broca's area may be important for imitation. An alternative hypothesis derived from the study of songbirds suggests that the corticostriate motor pathway performs sensorimotor processes that are specific to vocal imitation. Using fMRI with a sparse event-related sampling design, we investigated the neural basis of vocal imitation in humans by comparing imitative vocal production of pitch sequences with both nonimitative vocal production and pitch discrimination. The strongest difference between these tasks was found in the putamen bilaterally, providing a striking parallel to the role of the analogous region in songbirds. Other areas preferentially activated during imitation included the orofacial motor cortex, Rolandic operculum, and SMA, which together outline the corticostriate motor loop. No differences were seen in the inferior frontal gyrus. The corticostriate system thus appears to be the central pathway for vocal imitation in humans, as predicted from an analogy with songbirds.
Blom, Jan Dirk
Auditory hallucinations constitute a phenomenologically rich group of endogenously mediated percepts which are associated with psychiatric, neurologic, otologic, and other medical conditions, but which are also experienced by 10-15% of all healthy individuals in the general population. The group of phenomena is probably best known for its verbal auditory subtype, but it also includes musical hallucinations, echo of reading, exploding-head syndrome, and many other types. The subgroup of verbal auditory hallucinations has been studied extensively with the aid of neuroimaging techniques, and from those studies emerges an outline of a functional as well as a structural network of widely distributed brain areas involved in their mediation. The present chapter provides an overview of the various types of auditory hallucination described in the literature, summarizes our current knowledge of the auditory networks involved in their mediation, and draws on ideas from the philosophy of science and network science to reconceptualize the auditory hallucinatory experience, and point out directions for future research into its neurobiologic substrates. In addition, it provides an overview of known associations with various clinical conditions and of the existing evidence for pharmacologic and non-pharmacologic treatments. © 2015 Elsevier B.V. All rights reserved.
Villaume, William A.; Brown, Mary Helen
Notes that presbycusis, hearing loss associated with aging, may be marked by a second dimension of hearing loss, a loss in vocalic sensitivity. Reports on the development of the Vocalic Sensitivity Test, which controls for the verbal elements in speech while also allowing for the vocalics to exercise their normal metacommunicative function of…
Harbison, Amy L; Woynaroski, Tiffany G; Tapp, Jon; Wade, Joshua W; Warlaumont, Anne S; Yoder, Paul J
Children's vocal development occurs in the context of reciprocal exchanges with a communication partner who models "speechlike" productions. We propose a new measure of child vocal reciprocity, which we define as the degree to which an adult vocal response increases the probability of an immediately following child vocal response. Vocal reciprocity is likely to be associated with the speechlikeness of vocal communication in young children with autism spectrum disorder (ASD). Two studies were conducted to test the utility of the new measure. The first used simulated vocal samples with randomly sequenced child and adult vocalizations to test the accuracy of the proposed index of child vocal reciprocity. The second was an empirical study of 21 children with ASD who were preverbal or in the early stages of language development. Daylong vocal samples collected in the natural environment were computer analyzed to derive the proposed index of child vocal reciprocity, which was highly stable when derived from two daylong vocal samples and was associated with speechlikeness of vocal communication. This association was significant even when controlling for chance probability of child vocalizations to adult vocal responses, probability of adult vocalizations, or probability of child vocalizations. A valid measure of children's vocal reciprocity might eventually improve our ability to predict which children are on track to develop useful speech and/or are most likely to respond to language intervention. A link to a free, publicly-available software program to derive the new measure of child vocal reciprocity is provided. Autism Res 2018. © 2018 International Society for Autism Research, Wiley Periodicals, Inc. Children and adults often engage in back-and-forth vocal exchanges. The extent to which they do so is believed to support children's early speech and language development. Two studies tested a new measure of child vocal reciprocity using computer-generated and real
Hartog, Paula Maria den
Avian vocalizations function in mate attraction and territorial defence. Vocalizations can act as behavioural barriers and play an important role in speciation processes. Hybrid zones illustrate behavioural barriers are not always impermeable and provide a natural laboratory to examine the role of
Full Text Available Abstract Calcium-calmodulin dependent protein kinase IV (CaMKIV is a protein kinase that activates the transcription factor CREB. Our previous work demonstrated that mice lacking CaMKIV had a defect in fear memory while behavioral responses to noxious stimuli were unchanged. Here, we measured ultrasonic vocalizations (USVs before and after fear conditioning and in response to a noxious injection of capsaicin to measure behavioral responses to emotional stimuli. Consistent with previous findings, behavioral nociceptive responses to capsaicin were undistinguishable between wild-type and CaMKIV-/- mice. Wild-type animals showed a selective increase in 50 kHz USVs in response to capsaicin while such an increase was absent in CaMKIV-/- mice. The foot shock given during fear conditioning caused an increase in 30 kHz USVs in both wild-type and CaMKIV-/- mice. When returned to the context one hour later, USVs from the wild-type were significantly decreased. Additionally, the onset of a tone, which had previously been paired with the foot shock, caused a significant decrease in USVs during auditory conditioning. CaMKIV-/- mice showed significantly less reduction in USVs when placed in the same context three days after receiving the shock, consistent with the decrease in freezing reported previously. Our results provide a new approach for investigating the molecular mechanism for emotional vocalization in mice and suggest that CaMKIV dependent signaling pathways play an important role in the emotional response to pain and fear.
Millan, Jorge; Yunda, Leonardo
A portable, Internet-based EEG/Auditory Evoked Potential (AEP) monitoring system was developed for remote electrophysiological studies during sleep. The system records EEG/AEP simultaneously at the subject?s home for increased comfort and flexibility. The system provides simultaneous recording and remote viewing of EEG, EMG and EOG waves and allows on-line averaging of auditory evoked potentials. The design allows the recording of all major AEP components (brainstem, middle and late latency E...
Zhang, Qing; Kaga, Kimitaka; Hayashi, Akimasa
A 27-year-old female showed auditory agnosia after long-term severe hydrocephalus due to congenital spina bifida. After years of hydrocephalus, she gradually suffered from hearing loss in her right ear at 19 years of age, followed by her left ear. During the time when she retained some ability to hear, she experienced severe difficulty in distinguishing verbal, environmental, and musical instrumental sounds. However, her auditory brainstem response and distortion product otoacoustic emissions were largely intact in the left ear. Her bilateral auditory cortices were preserved, as shown by neuroimaging, whereas her auditory radiations were severely damaged owing to progressive hydrocephalus. Although she had a complete bilateral hearing loss, she felt great pleasure when exposed to music. After years of self-training to read lips, she regained fluent ability to communicate. Clinical manifestations of this patient indicate that auditory agnosia can occur after long-term hydrocephalus due to spina bifida; the secondary auditory pathway may play a role in both auditory perception and hearing rehabilitation.
Chen, Min; Chen, Jian; Yang, Yue; Cheng, Lei; Wu, Hai-Tao
Several studies have indicated the larynx as possible Helicobacter pylori (H. pylori) reservoirs. This study explored the association between H. pylori and vocal fold leukoplakia. The case-control study involved 51 patients with vocal fold leukoplakia and 35 control patients with vocal polyps. Helicobacter pylori was detected in tissues by the rapid urease test, nested polymerase chain reaction (PCR), and single-step PCR. The H. pylori-specific immunoglobulin antibodies were detected in plasma by enzyme-linked immunosorbent assay (ELISA). Helicobacter pylori-positive rate of vocal fold leukoplakia and vocal polyps was 23.5% versus 11.4% (P = .157), 37.2% versus 14.3% (P = .020), 27.5% versus 8.6% (P = .031), and 70.6% versus 68.6% (P = .841) detected by rapid urease test, nested PCR, single-step PCR, and ELISA, respectively. Regression analysis indicated that H. pylori infection (P = .044) was the independent risk factor for vocal fold leukoplakia. Helicobacter pylori infection exists in the larynx and may be associated with vocal fold leukoplakia. © 2018 Wiley Periodicals, Inc.
Scott, Brian H; Mishkin, Mortimer
Sounds are fleeting, and assembling the sequence of inputs at the ear into a coherent percept requires auditory memory across various time scales. Auditory short-term memory comprises at least two components: an active ׳working memory' bolstered by rehearsal, and a sensory trace that may be passively retained. Working memory relies on representations recalled from long-term memory, and their rehearsal may require phonological mechanisms unique to humans. The sensory component, passive short-term memory (pSTM), is tractable to study in nonhuman primates, whose brain architecture and behavioral repertoire are comparable to our own. This review discusses recent advances in the behavioral and neurophysiological study of auditory memory with a focus on single-unit recordings from macaque monkeys performing delayed-match-to-sample (DMS) tasks. Monkeys appear to employ pSTM to solve these tasks, as evidenced by the impact of interfering stimuli on memory performance. In several regards, pSTM in monkeys resembles pitch memory in humans, and may engage similar neural mechanisms. Neural correlates of DMS performance have been observed throughout the auditory and prefrontal cortex, defining a network of areas supporting auditory STM with parallels to that supporting visual STM. These correlates include persistent neural firing, or a suppression of firing, during the delay period of the memory task, as well as suppression or (less commonly) enhancement of sensory responses when a sound is repeated as a ׳match' stimulus. Auditory STM is supported by a distributed temporo-frontal network in which sensitivity to stimulus history is an intrinsic feature of auditory processing. This article is part of a Special Issue entitled SI: Auditory working memory. Published by Elsevier B.V.
Husain, Solomon; Sadoughi, Babak; Mor, Niv; Levin, Ariana M; Sulica, Lucian
To clarify the time course of recovery in patients with idiopathic vocal fold paralysis. Retrospective chart review. Medical records for all patients with idiopathic vocal fold paralysis over a 10-year period were reviewed to obtain demographic and clinical information, including onset of disease and recovery of vocal function. Stroboscopic exams of patients who recovered voice were reviewed blindly to assess return of vocal fold motion. Thirty-eight of 55 patients (69%) recovered vocal function. Time course of recovery could be assessed in 34 patients who did not undergo injection augmentation. The mean time to recovery was 152.8 ± 109.3 days (left, 179.8 ± 111.3 days; right, 105.3 ± 93.7 days; P = .088). Two-thirds of patients recovered within 6 months. Probability of recovery declined over time. Five of 22 patients who recovered voice had return of vocal fold motion; 17 did not. The mean time to recovery did not differ between these groups (return of motion, 127.4 ± 132.3 days; no return of motion, 160.1 ± 105.1 days; P = .290). Sixty-nine percent of patients with idiopathic vocal fold paralysis recovered vocal function, two-thirds doing so within 6 months of onset. Age, gender, laterality, use of injection augmentation did not influence recovery rate. Declining probability of recovery over time leads us to consider framework surgery after 6 months in patients with idiopathic paralysis. 4. Laryngoscope, 128:148-152, 2018. © 2017 The American Laryngological, Rhinological and Otological Society, Inc.
Kass, E S; Hillman, R E; Zeitels, S M
Phonomicrosurgery is optimized by maximally preserving the vocal fold's layered microstructure (laminae propriae). The technique of submucosal infusion of saline and epinephrine into the superficial lamina propria (SLP) was examined to delineate how, when, and why it was helpful toward this surgical goal. A retrospective review revealed that the submucosal infusion technique was used to enhance the surgery in 75 of 152 vocal fold procedures that were performed over the last 2 years. The vocal fold epithelium was noted to be adherent to the vocal ligament in 29 of the 75 cases: 19 from previous surgical scarring, 4 from cancer, 3 from sulcus vocalis, 2 from chronic hemorrhage, and 1 from radiotherapy. The submucosal infusion technique was most helpful when the vocal fold epithelium required resection and/or when extensive dissection in the SLP was necessary. The infusion enhanced the surgery by vasoconstriction of the microvasculature in the SLP, which improved visualization during cold-instrument tangential dissection. Improved visualization facilitated maximal preservation of the SLP, which is necessary for optimal pliability of the overlying epithelium. The infusion also improved the placement of incisions at the perimeter of benign, premalignant, and malignant lesions, and thereby helped preserve epithelium uninvolved by the disorder.
Full Text Available Spoken language and learned song are complex communication behaviors found in only a few species, including humans and three groups of distantly related birds--songbirds, parrots, and hummingbirds. Despite their large phylogenetic distances, these vocal learners show convergent behaviors and associated brain pathways for vocal communication. However, it is not clear whether this behavioral and anatomical convergence is associated with molecular convergence. Here we used oligo microarrays to screen for genes differentially regulated in brain nuclei necessary for producing learned vocalizations relative to adjacent brain areas that control other behaviors in avian vocal learners versus vocal non-learners. A top candidate gene in our screen was a calcium-binding protein, parvalbumin (PV. In situ hybridization verification revealed that PV was expressed significantly higher throughout the song motor pathway, including brainstem vocal motor neurons relative to the surrounding brain regions of all distantly related avian vocal learners. This differential expression was specific to PV and vocal learners, as it was not found in avian vocal non-learners nor for control genes in learners and non-learners. Similar to the vocal learning birds, higher PV up-regulation was found in the brainstem tongue motor neurons used for speech production in humans relative to a non-human primate, macaques. These results suggest repeated convergent evolution of differential PV up-regulation in the brains of vocal learners separated by more than 65-300 million years from a common ancestor and that the specialized behaviors of learned song and speech may require extra calcium buffering and signaling.
Woo, Peak; Parasher, Arjun K; Isseroff, Tova; Richards, Amanda; Sivak, Mark
The diagnosis of paresis in patients with vocal fold motion impairment remains a challenge. More than 27 clinical parameters have been cited that may signify paresis. We hypothesize that some features are more significant than others. Prospective case series. Two laryngologists rated laryngoscopy findings in 19 patients suspected of paresis. The diagnosis was confirmed with laryngeal electromyography. A standard set of 27 ratings was used for each examination that included movement, laryngeal configuration, and stroboscopy signs. A Fisher exact test was completed for each measure. A kappa coefficient was calculated for effectiveness in predicting the laterality of paresis. Left-sided vocal fold paresis (n = 13) was significantly associated with ipsilateral axis deviation, thinner vocal fold, bowing, reduced movement, reduced kinesis, and phase lag (P vocal fold paresis (n = 6) was significantly associated with ipsilateral shorter vocal fold, axis deviation, reduced movement, and reduced kinesis (P vocal fold, vocal fold bowing, reduced movement, reduced kinesis, and phase lag were more likely to be associated with vocal fold paresis. 4 Laryngoscope, 126:1831-1836, 2016. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.
Mitsuya, Takashi; MacDonald, Ewen N; Munhall, Kevin G
Past studies have shown that speakers spontaneously adjust their speech acoustics in response to their auditory feedback perturbed in real time. In the case of formant perturbation, the majority of studies have examined speaker's compensatory production using the English vowel /ɛ/ as in the word...... "head." Consistent behavioral observations have been reported, and there is lively discussion as to how the production system integrates auditory versus somatosensory feedback to control vowel production. However, different vowels have different oral sensation and proprioceptive information due...... to differences in the degree of lingual contact or jaw openness. This may in turn influence the ways in which speakers compensate for auditory feedback. The aim of the current study was to examine speakers' compensatory behavior with six English monophthongs. Specifically, the current study tested to see...
Matheson, Laura E; Sakata, Jon T
Social context affects behavioral displays across a variety of species. For example, social context acutely influences the acoustic and temporal structure of vocal communication signals such as speech and birdsong. Despite the prevalence and importance of such social influences, little is known about the neural mechanisms underlying the social modulation of communication. Catecholamines are implicated in the regulation of social behavior and motor control, but the degree to which catecholamines influence vocal communication signals remains largely unknown. Using a songbird, the Bengalese finch, we examined the extent to which the social context in which song is produced affected immediate early gene expression (EGR-1) in catecholamine-synthesising neurons in the midbrain. Further, we assessed the degree to which administration of amphetamine, which increases catecholamine concentrations in the brain, mimicked the effect of social context on vocal signals. We found that significantly more catecholaminergic neurons in the ventral tegmental area and substantia nigra (but not the central grey, locus coeruleus or subcoeruleus) expressed EGR-1 in birds that were exposed to females and produced courtship song than in birds that produced non-courtship song in isolation. Furthermore, we found that amphetamine administration mimicked the effects of social context and caused many aspects of non-courtship song to resemble courtship song. Specifically, amphetamine increased the stereotypy of syllable structure and sequencing, the repetition of vocal elements and the degree of sequence completions. Taken together, these data highlight the conserved role of catecholamines in vocal communication across species, including songbirds and humans. © 2015 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.
Pitman, Michael J; Kurita, Takashi; Powell, Maria E; Kimball, Emily E; Mizuta, Masanobu; Chang, Siyuan; Garrett, C Gaelyn; Rousseau, Bernard
Vocal fold scar is a major cause of dysphonia, and optimal treatments do not currently exist. Small intestinal submucosa (SIS) is a biomaterial developed for the treatment of a variety of pathologies. The purpose of this study was to investigate the effects of SIS implantation on tissue remodeling in scarred vocal folds using routine staining, immunohistochemistry, and high-speed videoendoscopy (HSV). Prospective, blinded group analysis. Thirteen New Zealand White rabbits underwent a vocal fold scarring procedure followed by microflap elevation with or without SIS implantation. Seven months later, they underwent a phonation procedure with HSV and laryngeal harvest. Alcian blue and elastica van Gieson staining and immunohistochemistry for collagen types I and III were used to evaluate histological healing outcomes. Dynamic functional remodeling of the scarred vocal fold in the presence of SIS implants was evaluated using HSV imaging to capture restoration of vibratory amplitude, amplitude ratio, and left-right phase symmetry. Density of collagen I was significantly decreased in SIS versus microflap-treated vocal folds. No differences were found between groups for hyaluronic acid, elastin, or collagen type III. Organization of elastin in the subepithelial region appeared to affect amplitude of vibration and the shape of the vocal fold edge. SIS implantation into chronic scar reduced the density of collagen I deposits. There was no evidence of a negative impact or complication from SIS implantation. Regardless of treatment type, organization of elastin in the subepithelial region may be important to vibratory outcomes. NA. Laryngoscope, 128:901-908, 2018. © 2017 The American Laryngological, Rhinological and Otological Society, Inc.
for studying the origins and neural basis of human language. Vocalizations belonging to the same species, or Conspecific Vocalizations (CVs), are...applications including automatic speech recognition , speech enhancement , voice activity detection , hyper-nasality detection , and emotion ...vocalizations. The feature sets chosen have the desirable property of capturing characteristics of the signals that are useful in both identifying and
Johns, Michael M.; Kolachala, Vasantha; Berg, Eric; Muller, Susan; Creighton, Frances X.; Branski, Ryan C.
Objectives To characterize fundamental late tissue effects in the human vocal fold following radiation therapy. To develop a murine model of radiation fibrosis to ultimately develop both treatment and prevention paradigms. Design Translational study using archived human and fresh murine irradiated vocal fold tissue. Methods 1) Irradiated vocal fold tissue from patients undergoing laryngectomy for loss of function from radiation fibrosis were identified from pathology archives. Histomorphometry, immunohistochemistry, and whole-genome microarray as well as real-time transcriptional analyses was performed. 2) Focused radiation to the head and neck was delivered to mice in a survival fashion. One month following radiation, vocal fold tissue was analyzed with histomorphometry, immunohistochemistry, and real-time PCR transcriptional analysis for selected markers of fibrosis. Results Human irradiated vocal folds demonstrated increased collagen transcription with increased deposition and disorganization of collagen in both the thyroarytenoid muscle and the superficial lamina propria. Fibronectin were increased in the superficial lamina propria. Laminin decreased in the thyroarytenoid muscle. Whole genome microarray analysis demonstrated increased transcription of markers for fibrosis, oxidative stress, inflammation, glycosaminoglycan production and apoptosis. Irradiated murine vocal folds demonstrated increases in collagen and fibronectin transcription and deposition in the lamina propria. Transforming growth factor (TGF)-β increased in the lamina propria. Conclusion Human irradiated vocal folds demonstrate molecular changes leading to fibrosis that underlie loss of vocal fold pliability that occurs in patients following laryngeal irradiation. Irradiated murine tissue demonstrates similar findings, and this mouse model may have utility in creating prevention and treatment strategies for vocal fold radiation fibrosis. PMID:23242839
Teresa, Garcia-Real; Díaz-Román, Tomás M
The objective of this study was to evaluate the presence of habits and symptoms of vocal hyperfunction in the parents of children with attention deficit hyperactivity disorder (ADHD). Parents of 24 children with ADHD and 30 children of a control group completed a specific questionnaire to detect the hyperfunctional use of the voice (excessive talking, excessive loudness, talking too fast, and shouting), hoarseness, vocal fatigue, mental and physical fatigue, and the degree of parental concern for the vocal health of their child. Parents of children with ADHD spoke more often, faster, and stronger than the parents of the control group; in addition, they also used a louder volume than they usually used when they spoke to their children. The parents manifested more vocal, mental, and physical fatigue than the parents of the control group. There was a significant correlation between the "concern" for the vocal health of their children with respect to vocal symptoms of the children, the habits of vocal hyperfunctioning, and the symptoms suffered by the parents. These results suggest that the parents of children with ADHD change their vocal attitude when communicating with their children. Most likely, the increased concern of parents with ADHD children and their respective level of stress lead to hyperfunctional vocal usage. This subsequently leads to symptoms of vocal, physical, and mental fatigue at the end of the day. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Nathani, Suneeti; Oller, D. Kimbrough; Cobo-Lewis, Alan B.
Sought to verify research findings that suggest there may be a U-shaped developmental trajectory for final syllable lengthening (FSL). Attempted to determine whether vocal maturity and deafness influence FSL . Eight normally hearing infants and eight deaf infants were examined at three levels of prelinguistic vocal development. (Author/VWL)
Full Text Available Abstract Background Current imaging techniques provide only limited information pertaining to the extent of infiltration of laryngeal carcinomas into vocal fold tissue layers. Therefore, it is needed to seek the contribute to the body of knowledge surrounding examination and characterization in laryngeal carcinoma infiltration. Methods Excised larynges were collected from 30 male laryngectomy patients with an average age of 43.5 years (ranging 36 to 55 years and history of smoking (≥10 years exhibiting T1, T2, or subglottal (normal vocal fold carcinomas. Vocal folds were preserved via freezing or immersion in paraffin. The depth of the mucosa, submucosa, and muscular layers in both normal vocal folds and tumor tissues of afflicted vocal folds was measured. Results The average depths of the mucosa, submucosa, and muscular layers in normal vocal folds were 0.15 ± 0.06 mm, 2.30 ± 0.59 mm, and 2.87 ± 0.88 mm, respectively. Infiltration measurements of T1 tumors showed a depth of 1.62 ± 0.51 mm and 1.32 ± 0.49 mm in frozen sections and paraffin-embedded samples, respectively. Similarly, T2 tumors showed a depth of 2.87 ± 0.68 mm and 2.58 ± 0.67 mm in frozen sections and paraffin-embedded samples, respectively. T1 and T2 tumors occupied 24.8 ± 10 and 48.5 ± 15 percent of the normal vocal fold depth, respectively. Conclusion This data provides a baseline for estimating infiltration of laryngeal carcinomas in vocal fold tissue layers, of particular interest to surgeons. This information may be used to assess typical depths of infiltration, thus allowing for more appropriate selection of surgical procedures based on individual patient assessment.
Fisher, Ryan A.; Scott, Julie K.
The purpose of this study was to examine the effects of vocal register use and age on the perceived vocal health of male elementary music teachers. Participants (N = 160) consisted of male elementary music teachers from two neighboring states in the south-central region of the United States. Participants responded to various demographic questions…
Maria Aparecida Coelho de Arruda Henry
Full Text Available CONTEXT: Gastroesophageal reflux disease is a chronic disease in which gastroduodenal contents reflux into the esophagus. The clinical picture of gastroesophageal reflux disease is usually composed by heartburn and regurgitation (typical manifestations. Atypical manifestations (vocal disturbances and asthma may also be complaint. OBJECTIVE: To analyse the clinical, endoscopic, manometric and pHmetric aspects of patients suffering from gastroesophageal reflux disease associated with vocal disturbances. METHODS: Fifty patients with gastroesophageal reflux disease were studied, including 25 with vocal disturbances (group 1 - G1 and 25 without these symptoms (group 2 - G2. All patients were submitted to endoscopy, manometry and esophageal pHmetry (2 probes. The group 1 patients were submitted to videolaryngoscopy. RESULTS: Endoscopic findings: non-erosive reflux disease was observed in 95% of G1 patients and 88% of G2. Videolaryngoscopy: vocal fold congestion, asymmetry, nodules and polyps were observed in G1 patients. Manometric findings: pressure in the lower esophageal sphincter (mm Hg: 11.6 ± 5.2 in G1 and 14.0 ± 6.2 in G2 (P = 0.14; pressure in the upper esophageal sphincter (mm Hg: 58.4 ± 15.9 in G1 and 69.5 ± 30.7 in the controls. pHmetric findings: De Meester index: 34.0 ± 20.9 in G1 and 15.4 ± 9.4 in G2 (P<0.001; number of reflux episodes in distal probe: 43.0 ± 20.4 in G1 and 26.4 ± 17.2 in G2 (P = 0.003; percentage of time with esophageal pH value lower than 4 units (distal sensor: 9.0% ± 6.4% in G1 and 3.4% ± 2.1% in G2 (P<0.001; number of reflux episodes in proximal probe: 7.5 ± 10.9 in G1 and 5.3 ± 5.7 in G2 (P = 0.38; percentage of time with esophageal pH values lower than 4 units (Proximal probe: 1.2 ± 2.7 in G1 and 0.5 ± 0.7 in G2 (P = 0.21. CONCLUSIONS: 1 The clinical, endoscopic, and manometric findings observed in patients with vocal disturbance do not differ from those without these symptoms; 2 gastroesophageal
Daliri, Ayoub; Max, Ludo
Auditory modulation during speech movement planning is limited in adults who stutter (AWS), but the functional relevance of the phenomenon itself remains unknown. We investigated for AWS and adults who do not stutter (AWNS) (a) a potential relationship between pre-speech auditory modulation and auditory feedback contributions to speech motor learning and (b) the effect on pre-speech auditory modulation of real-time versus delayed auditory feedback. Experiment I used a sensorimotor adaptation paradigm to estimate auditory-motor speech learning. Using acoustic speech recordings, we quantified subjects' formant frequency adjustments across trials when continually exposed to formant-shifted auditory feedback. In Experiment II, we used electroencephalography to determine the same subjects' extent of pre-speech auditory modulation (reductions in auditory evoked potential N1 amplitude) when probe tones were delivered prior to speaking versus not speaking. To manipulate subjects' ability to monitor real-time feedback, we included speaking conditions with non-altered auditory feedback (NAF) and delayed auditory feedback (DAF). Experiment I showed that auditory-motor learning was limited for AWS versus AWNS, and the extent of learning was negatively correlated with stuttering frequency. Experiment II yielded several key findings: (a) our prior finding of limited pre-speech auditory modulation in AWS was replicated; (b) DAF caused a decrease in auditory modulation for most AWNS but an increase for most AWS; and (c) for AWS, the amount of auditory modulation when speaking with DAF was positively correlated with stuttering frequency. Lastly, AWNS showed no correlation between pre-speech auditory modulation (Experiment II) and extent of auditory-motor learning (Experiment I) whereas AWS showed a negative correlation between these measures. Thus, findings suggest that AWS show deficits in both pre-speech auditory modulation and auditory-motor learning; however, limited pre
Picoloto, Luana Altran; Cardoso, Ana Cláudia Vieira; Cerqueira, Amanda Venuti; Oliveira, Cristiane Moço Canhetti de
To verify the effect of delayed auditory feedback on speech fluency of individuals who stutter with and without central auditory processing disorders. The participants were twenty individuals with stuttering from 7 to 17 years old and were divided into two groups: Stuttering Group with Auditory Processing Disorders (SGAPD): 10 individuals with central auditory processing disorders, and Stuttering Group (SG): 10 individuals without central auditory processing disorders. Procedures were: fluency assessment with non-altered auditory feedback (NAF) and delayed auditory feedback (DAF), assessment of the stuttering severity and central auditory processing (CAP). Phono Tools software was used to cause a delay of 100 milliseconds in the auditory feedback. The "Wilcoxon Signal Post" test was used in the intragroup analysis and "Mann-Whitney" test in the intergroup analysis. The DAF caused a statistically significant reduction in SG: in the frequency score of stuttering-like disfluencies in the analysis of the Stuttering Severity Instrument, in the amount of blocks and repetitions of monosyllabic words, and in the frequency of stuttering-like disfluencies of duration. Delayed auditory feedback did not cause statistically significant effects on SGAPD fluency, individuals with stuttering with auditory processing disorders. The effect of delayed auditory feedback in speech fluency of individuals who stutter was different in individuals of both groups, because there was an improvement in fluency only in individuals without auditory processing disorder.
Full Text Available Vocal cord palsy can have myriad causes. Unilateral vocal cord palsy is common and frequently asymptomatic. Trauma, head, neck and mediastinal tumors as well as cerebrovascular accidents have been implicated in causing unilateral vocal cord palsy. Viral neuronitis accounts for most idiopathic cases. Bilateral vocal cord palsy, on the other hand, is much less common and is a potentially life-threatening condition. Myasthenia gravis, an autoimmune disorder caused by antibodies targeting the post-synaptic acetylcholine receptor, has been infrequently implicated in its causation. We report here a case of bilateral vocal cord palsy developing in a 68-year-old man with no prior history of myasthenia gravis 2 months after he was operated on for diverticulitis of the large intestine. Delay in considering the diagnosis led to endotracheal intubation and prolonged mechanical ventilation with attendant complications. Our case adds to the existing literature implicating myasthenia gravis as an infrequent cause of bilateral vocal cord palsy. Our case is unusual as, in our patient, acute-onset respiratory distress and stridor due to bilateral vocal cord palsy was the first manifestation of a myasthenic syndrome.
Levendoski, Elizabeth Erickson; Sivasankar, M Preeti
The vocal fold epithelium is exposed to inhaled particulates including pollutants during breathing in everyday environments. Yet, our understanding of the effects of pollutants on vocal fold epithelial function is extremely limited. The objective of this study was to investigate the effect of the pollutant acrolein on two vocal fold epithelial mechanisms: ion transport and mucin (MUC) synthesis. These mechanisms were chosen as each plays a critical role in vocal defense and in maintaining surface hydration which is necessary for optimal voice production. Healthy, native porcine vocal folds (N = 85) were excised and exposed to an acrolein or sham challenge. A 60-min acrolein, but not sham challenge significantly reduced ion transport and inhibited cyclic adenosine monophosphate-dependent, increases in ion transport. Decreases in ion transport were associated with reduced sodium absorption. Within the same timeline, no significant acrolein-induced changes in MUC gene or protein expression were observed. These results improve our understanding of the effects of acrolein on key vocal fold epithelial functions and inform the development of future investigations that seek to elucidate the impact of a wide range of pollutant exposures on vocal fold health.
Henry U. Koishi
Full Text Available O aumento da resistência glótica é o principal mecanismo responsável pela elevação da intensidade vocal em baixas freqüências. Esse aumento da resistência, em situações normais, é determinado pela contração dos músculos adutores das pregas vocais que promovem o aumento da tensão e a aproximação das pregas vocais em direção à linha mediana. No entanto, essas mesmas alterações podem estar presentes em algumas doenças funcionais que envolvem a laringe, como a disfonia espasmódica em adução e a disfonia hipertônica, mesmo durante a emissão vocal em baixa intensidade. OBJETIVO: Avaliar o padrão de vibração das pregas vocais em indivíduos com vozes normais em condições distintas de intensidade vocal, na tentativa de estabelecer valores que expressem a normalidade, para as diferentes fases do ciclo vibratório, de acordo com o nível de intensidade. FORMA DE ESTUDO: Clínico prospectivo. CASUÍSTICA E MÉTODOS: Foram avaliados 58 indivíduos durante emissão vocal em intensidade habitual (suave de conversação e durante emissão vocal elevada. A análise da vibração das pregas vocais foi realizada com a videoquimografia e para a análise da intensidade vocal, foi utilizado um programa de análise acústica computadorizado. RESULTADOS: Os resultados mostraram aumento involuntário da freqüência fundamental (F0 e redução do quociente de abertura, com o aumento da intensidade. CONCLUSÃO: Foram estabelecidos os valores de intensidade sonora para a condição habitual (63,46dB e elevada (72,55dB de emissão vocal e seus respectivos valores de quociente de abertura (OQ- open quotient.The increase of glottal resistance is the main mechanism to increase vocal intensity at low fundamental frequency. This increase is due to adductory forces provoked by the contraction of intrinsic laryngeal muscles that increases tension and approximates the vocal folds to the midline. However, the same behavior can be observed in
Mirian Aratangy Arnaut
Full Text Available Contemporary cross-sectional cohort study. There is evidence of the auditory perception influence on the development of oral and written language, as well as on the self-perception of vocal conditions. The auditory system maturation can impact on this process. OBJECTIVE: To characterize the auditory skills of temporal ordering and localization in dysphonic children. MATERIALS AND METHODS: We assessed 42 children (4 to 8 years. Study group: 31 dysphonic children; Comparison group: 11 children without vocal change complaints. They all had normal auditory thresholds and also normal cochleo-eyelid reflexes. They were submitted to a Simplified assessment of the auditory process (Pereira, 1993. In order to compare the groups, we used the Mann-Whitney and Kruskal-Wallis statistical tests. Level of significance: 0.05 (5%. RESULTS: Upon simplified assessment, 100% of the Control Group and 61.29% of the Study Group had normal results. The groups were similar in the localization and verbal sequential memory tests. The nonverbal sequential memory showed worse results on dysphonic children. In this group, the performance was worse among the four to six years. CONCLUSION: The dysphonic children showed changes on the localization or temporal ordering skills, the skill of non-verbal temporal ordering differentiated the dysphonic group. In this group, the Sound Location improved with age.Estudo de coorte contemporânea com corte transversal. Há evidências da influência da percepção auditiva sobre o desenvolvimento da linguagem oral e escrita e da autopercepção das condições vocais. A maturação do sistema auditivo pode interferir nesse processo. OBJETIVO: Caracterizar habilidades auditivas de Localização e de Ordenação Temporal em crianças disfônicas. MATERIAL E MÉTODO: Avaliaram-se 42 crianças (4 a 8 anos. Grupo Pesquisa: 31 crianças disfônicas, Grupo de Comparação: 11 crianças sem queixas de alterações vocais. Todas apresentaram
Simon-Thomas, Emiliana R; Keltner, Dacher J; Sauter, Disa; Sinicropi-Yao, Lara; Abramson, Anna
Studies of emotion signaling inform claims about the taxonomic structure, evolutionary origins, and physiological correlates of emotions. Emotion vocalization research has tended to focus on a limited set of emotions: anger, disgust, fear, sadness, surprise, happiness, and for the voice, also tenderness. Here, we examine how well brief vocal bursts can communicate 22 different emotions: 9 negative (Study 1) and 13 positive (Study 2), and whether prototypical vocal bursts convey emotions more reliably than heterogeneous vocal bursts (Study 3). Results show that vocal bursts communicate emotions like anger, fear, and sadness, as well as seldom-studied states like awe, compassion, interest, and embarrassment. Ancillary analyses reveal family-wise patterns of vocal burst expression. Errors in classification were more common within emotion families (e.g., 'self-conscious,' 'pro-social') than between emotion families. The three studies reported highlight the voice as a rich modality for emotion display that can inform fundamental constructs about emotion.
Full Text Available OBJETIVO: analisar sintomas vocais de dois grupos de professores que foram avaliados em momentos distintos de um programa de saúde vocal. MÉTODOS: correlacionar condições de trabalho e hábitos com o número de sintomas vocais apresentados por 411 professores, agrupados em G1 (256 sujeitos a serem submetidos ao programa preventivo e G2 (155 sujeitos a serem submetidos ao programa preventivo e de tratamento. RESULTADOS: em ambos os grupos observou-se predomínio de mulheres (p = 0,550, entre 31 e 40 anos (p = 0,557, lecionando para mais de um grau de ensino (p = 0,345 com até 30 alunos/sala (p = 0,521, com presença de ruído no trabalho (p = 0,660, que relataram cuidados vocais (p = 0,231 e utilizavam voz extra-profissionalmente (p = 0,713, não tabagistas (p = 0,010 nem alcoolistas (p = 0,029. Em contrapartida, no G1 observou-se carga horária diária de trabalho de até 5 horas, enquanto a maior parte do G2 trabalhava de 6 a 10 horas (p 4 os grupos; sendo a média de 3,5 (57% para G1 e 5,8 (98,05% para G2 - (p PURPOSE: to analyze vocal symptoms from two groups of teachers assessed during two different instances of a vocal health program. METHODS: correlate work conditions and habits with the number of vocal symptoms submitted by 411 teachers divided into G1 (256 subjects to be submitted to a prevention program and G2 (155 subjects to be submitted to a prevention and treatment program. RESULTS: it was observed that in both groups there was a larger number of women (p=0.550, aged 31 to 40 years (p=0.557, teaching for more than one grade (p=0.345 and with up to 30 students per class (p=0.521; they related presence of noise in their work environment (p=0.660, used to take care of their voices (p=0.231, were non-smokers (p=0.010, used their voices in extra-professional activities and did not have the habit of drinking. On the other hand, both groups were different upon relating daily work hours; in G1 most teachers worked up to 5 hours a day
Olson, Christopher R; Fernández-Vargas, Marcela; Portfors, Christine V; Mello, Claudio V
Hummingbirds are a fascinating group of birds, but some aspects of their biology are poorly understood, such as their highly diverse vocal behaviors. We show here that the predominant vocalization of black jacobins (Florisuga fusca), a hummingbird prevalent in the mountains of the Brazilian Atlantic Forest, consists of a triplet of syllables with high fundamental frequency (mean F0 ∼11.8 kHz), rapid frequency oscillations and strong ultrasonic harmonics and no detectable elements below ∼10 kHz. These are the most common vocalizations of these birds, and their frequency range is above the known hearing range of any bird species recorded to date, including hearing specialists such as owls. These observations suggest that black jacobins either have an atypically high frequency hearing range, or alternatively their primary vocalization has a yet unknown function unrelated to vocal communication. Black jacobin vocalizations challenge current notions about vocal communication in birds. Copyright © 2018 Elsevier Ltd. All rights reserved.
Vocal fold vibration onset delay (VFVOD) is heard frequently in spasmodic dysphonia and in muscle tension dysphonia. VFVOD changes due to other vocal pathologies have not been investigated. VFVOD during sustained vowel production was estimated with high-speed video in 10 normal and 40 pathologic subjects (scars, vocal fold paralysis, vocal fold nodules, and polyps). Analysis of high-speed video was done using digital kymography. VFVOD can be divided into two portions. Pre-phonation delay (PPD) is the duration when the vocal folds are nearly approximated to the time of first observed oscillation. Steady state delay (SSD) is the time when vocal folds are observed to come into oscillation until steady state of oscillation is observed. Normal subjects have almost zero PPD with vocal fold oscillation observed before full vocal fold adduction. Pathologic cases showed prolonged PPD because of (1) false cord adduction, (2) prolonged true vocal fold adduction, and (3) delay to onset of vocal fold vibration. Normal subjects have SSD of three to five cycles before steady state. Pathologic states result in increased SSD. Causes for increased SSD include (1) slow ramping up to steady state, (2) partial vibration of vocal folds, and (3) diplophonia with alternating beats before achieving steady state. There are significant differences between normal and pathology groups in both PPD and SSD. VFVOD is elevated in pathologic states. This can be due to increase in PPD or SSD. VFVOD is an under-recognized phenomenon that may contribute to complaints of vocal fatigue and dysphonia. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Wu, Calvin; Stefanescu, Roxana A; Martel, David T; Shore, Susan E
Conventionally, sensory systems are viewed as separate entities, each with its own physiological process serving a different purpose. However, many functions require integrative inputs from multiple sensory systems and sensory intersection and convergence occur throughout the central nervous system. The neural processes for hearing perception undergo significant modulation by the two other major sensory systems, vision and somatosensation. This synthesis occurs at every level of the ascending auditory pathway: the cochlear nucleus, inferior colliculus, medial geniculate body and the auditory cortex. In this review, we explore the process of multisensory integration from (1) anatomical (inputs and connections), (2) physiological (cellular responses), (3) functional and (4) pathological aspects. We focus on the convergence between auditory and somatosensory inputs in each ascending auditory station. This review highlights the intricacy of sensory processing and offers a multisensory perspective regarding the understanding of sensory disorders.
Kokinous, Jenny; Kotz, Sonja A; Tavano, Alessandro; Schröger, Erich
We used human electroencephalogram to study early audiovisual integration of dynamic angry and neutral expressions. An auditory-only condition served as a baseline for the interpretation of integration effects. In the audiovisual conditions, the validity of visual information was manipulated using facial expressions that were either emotionally congruent or incongruent with the vocal expressions. First, we report an N1 suppression effect for angry compared with neutral vocalizations in the auditory-only condition. Second, we confirm early integration of congruent visual and auditory information as indexed by a suppression of the auditory N1 and P2 components in the audiovisual compared with the auditory-only condition. Third, audiovisual N1 suppression was modulated by audiovisual congruency in interaction with emotion: for neutral vocalizations, there was N1 suppression in both the congruent and the incongruent audiovisual conditions. For angry vocalizations, there was N1 suppression only in the congruent but not in the incongruent condition. Extending previous findings of dynamic audiovisual integration, the current results suggest that audiovisual N1 suppression is congruency- and emotion-specific and indicate that dynamic emotional expressions compared with non-emotional expressions are preferentially processed in early audiovisual integration. © The Author (2014). Published by Oxford University Press. For Permissions, please email: email@example.com.
Pillai, Roshni; Yathiraj, Asha
The study evaluated whether there exists a difference/relation in the way four different memory skills (memory score, sequencing score, memory span, & sequencing span) are processed through the auditory modality, visual modality and combined modalities. Four memory skills were evaluated on 30 typically developing children aged 7 years and 8 years across three modality conditions (auditory, visual, & auditory-visual). Analogous auditory and visual stimuli were presented to evaluate the three modality conditions across the two age groups. The children obtained significantly higher memory scores through the auditory modality compared to the visual modality. Likewise, their memory scores were significantly higher through the auditory-visual modality condition than through the visual modality. However, no effect of modality was observed on the sequencing scores as well as for the memory and the sequencing span. A good agreement was seen between the different modality conditions that were studied (auditory, visual, & auditory-visual) for the different memory skills measures (memory scores, sequencing scores, memory span, & sequencing span). A relatively lower agreement was noted only between the auditory and visual modalities as well as between the visual and auditory-visual modality conditions for the memory scores, measured using Bland-Altman plots. The study highlights the efficacy of using analogous stimuli to assess the auditory, visual as well as combined modalities. The study supports the view that the performance of children on different memory skills was better through the auditory modality compared to the visual modality. Copyright © 2017 Elsevier B.V. All rights reserved.
Smith, Libby J; Rosen, Clark A; Munin, Michael C
As laryngeal electromyography (LEMG) becomes more refined, accurate predictions of vocal fold motion recovery are possible. Focus has been on outcomes for patients with poor prognosis for vocal fold motion recovery. Limited information is available regarding the expected rate of purposeful vocal fold motion recovery when there is good to normal motor recruitment, no signs of denervation, and no signs of synkinetic activity with LEMG, termed excellent prognosis. The objective of this study is to determine the rate of vocal fold motion recovery with excellent prognosis findings on LEMG after acute recurrent laryngeal nerve injury. Retrospective review. Patients undergoing a standardized LEMG protocol, consisting of qualitative (evaluation of motor recruitment, motor unit configuration, detection of fibrillations, presence of synkinesis) and quantitative (turns analysis) measurements were evaluated for purposeful vocal-fold motion recovery, calculated after at least 6 months since onset of injury. Twenty-three patients who underwent LEMG for acute vocal fold paralysis met the inclusion criteria of excellent prognosis. Eighteen patients (78.3%) recovered vocal fold motion, as determined by flexible laryngoscopy. Nearly 80% of patients determined to have excellent prognosis for vocal fold motion recovery experienced return of vocal fold motion. This information will help clinicians not only counsel their patients on expectations but will also help guide treatment. 4. Laryngoscope, 126:2310-2314, 2016. © 2016 The American Laryngological, Rhinological and Otological Society, Inc.
Van Lierde, K M; Claeys, S; Dhaeseleer, E; Deley, S; Derde, K; Herregods, I; Strybol, I; Wuyts, F
The purpose of the present cross-sectional study was to determine the objective vocal quality and the vocal characteristics (vocal risk factors, vocal and corporal complaints) in 143 female student teachers during the 3 years of study. The objective vocal quality was measured by means of the Dysphonia Severity Index (DSI). Perceptual voice assessment, the Voice Handicap Index, questionnaires addressing vocal risks, and vocal and corporal complaints during and/or after voice usage were performed. Student teachers have a normal perceptual and objective vocal quality corresponding with a DSI% of 76. The analysis of variance revealed a significant improvement of the vocal quality between the first and the third year of study. No psychosocial handicapping effect of the voice was observed, though there are some vocal complaints and almost all student teachers reported the presence of corporal pain during and/or after speaking. Especially sore throat and headache were mentioned as the most present corporal pain symptoms. Due to the decreased awareness and the multifactorial genesis of the potential vocal risk factors, the student teachers are at risk for developing an occupational dysphonia during their teaching career. Because teaching is a high-risk profession for the development of voice problems, the incorporation of a direct vocal training technique to increase vocal endurance during teaching together with a vocal hygiene program, dietetics, and a stress management training program during the 3 years of study is needed to prevent occupational dysphonia. 2010 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Bourne, Tracy; Kenny, Dianna
To gather qualitative descriptions of music theater vocal qualities including belt, legit, and mix from expert pedagogues to better define this voice type. This is a prospective, semistructured interview. Twelve expert teachers from United States, United Kingdom, Asia, and Australia were interviewed by Skype and asked to identify characteristics of music theater vocal qualities including vocal production, physiology, esthetics, pitch range, and pedagogical techniques. Responses were compared with published studies on music theater voice. Belt and legit were generally described as distinct sounds with differing physiological and technical requirements. Teachers were concerned that belt should be taught "safely" to minimize vocal health risks. There was consensus between teachers and published research on the physiology of the glottis and vocal tract; however, teachers were not in agreement about breathing techniques. Neither were teachers in agreement about the meaning of "mix." Most participants described belt as heavily weighted, thick folds, thyroarytenoid-dominant, or chest register; however, there was no consensus on an appropriate term. Belt substyles were named and generally categorized by weightedness or tone color. Descriptions of male belt were less clear than for female belt. This survey provides an overview of expert pedagogical perspectives on the characteristics of belt, legit, and mix qualities in the music theater voice. Although teacher responses are generally in agreement with published research, there are still many controversial issues and gaps in knowledge and understanding of this vocal technique. Breathing techniques, vocal range, mix, male belt, and vocal registers require continuing investigation so that we can learn more about efficient and healthy vocal function in music theater singing. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Terry, Aaron D.; Ricks, Matthew T.; Thomson, Scott L.
Vocal fold vibration that is induced by air flowing from the lungs is believed to decrease blood flow through the vocal folds. This is important due to the critical role of blood flow in maintaining tissue health. However, the precise mechanical relationships between vocal fold vibration and blood perfusion remain understudied. A platform for studying liquid perfusion in a synthetic, life-size, self-oscillating vocal fold replica has recently been developed. The replicas are fabricated using molded silicone with material properties comparable to those of human vocal fold tissues and that include embedded microchannels through which liquid is perfused. The replicas are mounted on an air flow supply tube to initiate flow-induced vibration. A liquid reservoir is attached to the microchannel to cause liquid to perfuse through replica in the anterior-posterior direction. As replica vibration is initiated and amplitude increases, perfusion flow rate decreases. In this presentation, the replica design will be presented, along with data quantifying the relationships between parameters such as replica vibration amplitude, stiffness, microchannel diameter, and perfusion flow rate. This work was supported by Grant NIDCD R01DC005788 from the National Institutes of Health.
... Viral infections. Some viral infections, such as Lyme disease, Epstein-Barr and herpes, can cause inflammation and damage directly to the nerves in the larynx. Neurological conditions. If you have certain ... disease, you may experience vocal cord paralysis. Risk factors ...
Chen, Yining; Matheson, Laura E; Sakata, Jon T
Social processes profoundly influence speech and language acquisition. Despite the importance of social influences, little is known about how social interactions modulate vocal learning. Like humans, songbirds learn their vocalizations during development, and they provide an excellent opportunity to reveal mechanisms of social influences on vocal learning. Using yoked experimental designs, we demonstrate that social interactions with adult tutors for as little as 1 d significantly enhanced vocal learning. Social influences on attention to song seemed central to the social enhancement of learning because socially tutored birds were more attentive to the tutor's songs than passively tutored birds, and because variation in attentiveness and in the social modulation of attention significantly predicted variation in vocal learning. Attention to song was influenced by both the nature and amount of tutor song: Pupils paid more attention to songs that tutors directed at them and to tutors that produced fewer songs. Tutors altered their song structure when directing songs at pupils in a manner that resembled how humans alter their vocalizations when speaking to infants, that was distinct from how tutors changed their songs when singing to females, and that could influence attention and learning. Furthermore, social interactions that rapidly enhanced learning increased the activity of noradrenergic and dopaminergic midbrain neurons. These data highlight striking parallels between humans and songbirds in the social modulation of vocal learning and suggest that social influences on attention and midbrain circuitry could represent shared mechanisms underlying the social modulation of vocal learning.
Full Text Available Background: Vocal fold polyp is one of the most common causes for hoarseness. Many different etiological factors contribute to vocal fold polyp formation. The aim of the study was to find out whether the etiological factors for polyp formation have changed in the last 30 years.Methods: Eighty-one patients with unilateral vocal fold polyp were included in the study. A control group was composed of 50 volunteers without voice problems who matched the patients by age and gender. The data about etiological factors and the findings of phoniatric examination were obtained from the patients' medical documentation and from the questionnaires for the control group. The incidence of etiological factors was compared between the two groups. The program SPSS, Version 18 was used for statistical analysis.Results: The most frequent etiological factors were occupational voice load, GER, allergy and smoking. In 79% of patients 2 – 6 contemporary acting risk factors were found. Occupational voice load (p=0,018 and GER (p=0,004 were significantly more frequent in the patients than in the controls. The other factors did not significantly influence the polyp formation.Conclusions: There are several factors involved simultaneously in the formation of vocal fold polyps both nowadays and 30 years ago. Some of the most common factors remain the same (voice load, smoking, others are new (GER, allergy, which is probably due to the different lifestyle and working conditions than 30 years ago. Occupational voice load and GER were significantly more frequently present in the patients with polyp than in the control group. Regarding the given results it is important to instruct workers with professional vocal load about etiological factors for vocal fold polyp formation.
Niebuhr, Oliver; Lautenbacher, Stefan; Salinas-Ranneberg, Melissa
” (central vowel, sounding like a darker “e” as in hesitations like “ehm”)—as experimental approximations to natural vocalizations. Methods: In 50 students vowel production and self-report ratings were assessed during painful and nonpainful heat stimulation (hot water immersion) as well as during baseline......Introduction and Objectives: There have, yet, been only few attempts to phonetically characterize the vocalizations of pain, although there is wide agreement that moaning, groaning, or other nonverbal utterance can be indicative of pain. We studied the production of vowels “u,” “a,” “i”, and “schwa...... pain. Furthermore, changes from nonpainful to painful stimulations in these parameters also significantly predicted concurrent changes in pain ratings. Conclusion: Vocalization characteristics of pain seem to be best described by an increase in pitch and in loudness. Future studies using more specific...
Heckman, Jesse J; Proville, Rémi; Heckman, Gert J; Azarfar, Alireza; Celikel, Tansu; Englitz, Bernhard
Mice display a wide repertoire of vocalizations that varies with age, sex, and context. Especially during courtship, mice emit ultrasonic vocalizations (USVs) of high complexity, whose detailed structure is poorly understood. As animals of both sexes vocalize, the study of social vocalizations requires attributing single USVs to individuals. The state-of-the-art in sound localization for USVs allows spatial localization at centimeter resolution, however, animals interact at closer ranges, involving tactile, snout-snout exploration. Hence, improved algorithms are required to reliably assign USVs. We develop multiple solutions to USV localization, and derive an analytical solution for arbitrary vertical microphone positions. The algorithms are compared on wideband acoustic noise and single mouse vocalizations, and applied to social interactions with optically tracked mouse positions. A novel, (frequency) envelope weighted generalised cross-correlation outperforms classical cross-correlation techniques. It achieves a median error of ~1.4 mm for noise and ~4-8.5 mm for vocalizations. Using this algorithms in combination with a level criterion, we can improve the assignment for interacting mice. We report significant differences in mean USV properties between CBA mice of different sexes during social interaction. Hence, the improved USV attribution to individuals lays the basis for a deeper understanding of social vocalizations, in particular sequences of USVs.
Cross, N; Rogers, L J
Using a non-invasive method of sampling saliva followed by assay for cortisol levels, we found that common marmosets (Callithrix jacchus) show a decrease in cortisol levels after seeing a snake-model stimulus that reliably elicits mobbing (tsik) calls. In fact, there was a significant positive correlation between the number of tsik vocalizations made and the magnitude of the decrease in the cortisol concentrations. Furthermore, marmosets with higher levels of cortisol prior to being exposed to the stimulus produce more tsik calls than those with lower levels of cortisol. Subsequent experiments showed that, in response to 15 min of isolation with no visual or auditory contact with conspecifics (a traditional stressor), cortisol levels increased significantly. However, playback of the mobbing calls of a familiar conspecific to individual isolated marmosets not only prevented the rise in cortisol, but also actually caused a decrease in the levels of this hormone. This suggests that the mobbing calls serve to calm the marmoset after experiencing a stressful situation. This finding results in a greater understanding as to the role of physiological responses during communication in this species and could have implications for the welfare of marmosets in captivity.
Full Text Available Benign lesions of vocal folds are common disorders. Fifty percent of patients who have sound complaints are found to have these lesions after endoscopic and stroboscopic examinations. Benign vocal fold diseases are primarily caused by vibratory trauma. However they may also occur as a result of viral infections and congenital causes. These lesions are often presented with the complaints of dysphonia. [Archives Medical Review Journal 2013; 22(1.000: 86-95
Hechavarría, Julio C; Beetz, M Jerome; Macias, Silvio; Kössl, Manfred
Distress vocalizations (also known as alarm or screams) are an important component of the vocal repertoire of a number of animal species, including bats, humans, monkeys and birds, among others. Although the behavioral relevance of distress vocalizations is undeniable, at present, little is known about the rules that govern vocalization production when in alarmful situations. In this article, we show that when distressed, bats of the species Carollia perspicillata produce repetitive vocalization sequences in which consecutive syllables are likely to be similar to one another regarding their physical attributes. The uttered distress syllables are broadband (12-73 kHz) with most of their energy focussing at 23 kHz. Distress syllables are short (~4 ms), their average sound pressure level is close to 70 dB SPL, and they are produced at high repetition rates (every 14 ms). We discuss that, because of their physical attributes, bat distress vocalizations could serve a dual purpose: (1) advertising threatful situations to conspecifics, and (2) informing the threatener that the bats are ready to defend themselves. We also discuss possible advantages of advertising danger/discomfort using repetitive utterances, a calling strategy that appears to be ubiquitous across the animal kingdom.
Estes, Christine; Sulica, Lucian
To describe treatment results and identify predictors of the need for surgical intervention in patients with vocal fold pseudocyst. Retrospective cohort study with longitudinal followup via survey. Clinical records were reviewed for demographic information, VHI-10 score, and degree of severity of dysphonia. Videostroboscopic examinations were evaluated for presence of vocal fold pseudocyst, along with additional clinical variables, including laterality, reactive lesion, paresis, varix, and hemorrhage. Follow-up surveys were sent to all participants to evaluate current VHI-10 score and degree of vocal limitation. Results were analyzed to determine predictors of surgery and recurrence of pathology. Forty-six patients (41F:5M) with pseudocyst (40 unilateral: 6 bilateral) were reviewed. Twenty-three (50%) had reactive lesions, nineteen (41%) had paresis by clinical criteria, 10 (22%) had varices, and 6 (13%) had hemorrhage on examination. All underwent initial behavioral management (2-12 sessions of voice therapy; mean of 8 sessions). Seventeen (37%) eventually required surgical intervention. No demographic or clinical variables proved predictive of surgical intervention. Follow-up surveys were completed by 63% of patients, and 79% agreed with the statement that they were not professionally limited by their voices. This experience supports behavioral management as an initial intervention in patients with pseudocyst, sufficient by itself to restore vocal function in approximately two out of three patients. Neither initial severity nor any of the studied clinical findings predicted the need for surgery. The large majority of patients with pseudocyst are able to be treated effectively without impact in their professional function. © 2013 The American Laryngological, Rhinological and Otological Society, Inc.
Arnedo, Luisa F; Mendes, Francisco D C; Strier, Karen B
We investigated whether sex differences in spatial dynamics correlate with rates of staccato and neigh vocalizations in northern muriquis (Brachyteles hypoxanthus) at the Reserva Particular do Patrimônio Natural-Feliciano Miguel Abdala, Minas Gerais, Brazil. A total of 2,727 10 min focal subject samples were collected on 32 adult females and 31 adult males between April 2007 and March 2008. Compared with males, females spent a significantly lower proportion of their time in proximity to other group members and gave staccatos at significantly higher rates while feeding, resting, and traveling. Conversely, males emitted neigh vocalizations at significantly higher rates than females when feeding and resting only. Both sexes gave significantly more staccatos when feeding than when they were engaged in other activities, but their respective rates of neighs did not vary across activities. Both females and males emitted staccato vocalizations at significantly higher rates during times of the year when preferred foods were scarce, but no seasonal differences in the rates of neigh vocalizations were observed in either sex. Females and males showed a reduction in the number of neighbors following staccato vocalizations and an increase in the number of neighbors following neigh vocalizations. Our findings of sex differences in the rates of staccato and neigh vocalizations and the effects of these vocalizations on interindividual spacing are consistent with sex differences in spatial dynamics, and confirm the role of vocal communication in mediating spatial associations in this species.
Li, Linqing; Stiadle, Jeanna M.; Lau, Hang K.; Zerdoum, Aidan B.; Jia, Xinqiao; L.Thibeault, Susan; Kiick, Kristi L.
Vocal folds are soft laryngeal connective tissues with distinct layered structures and complex multicomponent matrix compositions that endow phonatory and respiratory functions. This delicate tissue is easily damaged by various environmental factors and pathological conditions, altering vocal biomechanics and causing debilitating vocal disorders that detrimentally affect the daily lives of suffering individuals. Modern techniques and advanced knowledge of regenerative medicine have led to a deeper understanding of the microstructure, microphysiology, and micropathophysiology of vocal fold tissues. State-of-the-art materials ranging from extracecullar-matrix (ECM)-derived biomaterials to synthetic polymer scaffolds have been proposed for the prevention and treatment of voice disorders including vocal fold scarring and fibrosis. This review intends to provide a thorough overview of current achievements in the field of vocal fold tissue engineering, including the fabrication of injectable biomaterials to mimic in vitro cell microenvironments, novel designs of bioreactors that capture in vivo tissue biomechanics, and establishment of various animal models to characterize the in vivo biocompatibility of these materials. The combination of polymeric scaffolds, cell transplantation, biomechanical stimulation, and delivery of antifibrotic growth factors will lead to successful restoration of functional vocal folds and improved vocal recovery in animal models, facilitating the application of these materials and related methodologies in clinical practice. PMID:27619243
Mehler, Bruce; Kidd, David; Reimer, Bryan; Reagan, Ian; Dobres, Jonathan; McCartt, Anne
Abstract One purpose of integrating voice interfaces into embedded vehicle systems is to reduce drivers’ visual and manual distractions with ‘infotainment’ technologies. However, there is scant research on actual benefits in production vehicles or how different interface designs affect attentional demands. Driving performance, visual engagement, and indices of workload (heart rate, skin conductance, subjective ratings) were assessed in 80 drivers randomly assigned to drive a 2013 Chevrolet Equinox or Volvo XC60. The Chevrolet MyLink system allowed completing tasks with one voice command, while the Volvo Sensus required multiple commands to navigate the menu structure. When calling a phone contact, both voice systems reduced visual demand relative to the visual–manual interfaces, with reductions for drivers in the Equinox being greater. The Equinox ‘one-shot’ voice command showed advantages during contact calling but had significantly higher error rates than Sensus during destination address entry. For both secondary tasks, neither voice interface entirely eliminated visual demand. Practitioner Summary: The findings reinforce the observation that most, if not all, automotive auditory–vocal interfaces are multi-modal interfaces in which the full range of potential demands (auditory, vocal, visual, manipulative, cognitive, tactile, etc.) need to be considered in developing optimal implementations and evaluating drivers’ interaction with the systems. Social Media: In-vehicle voice-interfaces can reduce visual demand but do not eliminate it and all types of demand need to be taken into account in a comprehensive evaluation. PMID:26269281
Wightman, Frederic L.; Jenison, Rick
All auditory sensory information is packaged in a pair of acoustical pressure waveforms, one at each ear. While there is obvious structure in these waveforms, that structure (temporal and spectral patterns) bears no simple relationship to the structure of the environmental objects that produced them. The properties of auditory objects and their layout in space must be derived completely from higher level processing of the peripheral input. This chapter begins with a discussion of the peculiarities of acoustical stimuli and how they are received by the human auditory system. A distinction is made between the ambient sound field and the effective stimulus to differentiate the perceptual distinctions among various simple classes of sound sources (ambient field) from the known perceptual consequences of the linear transformations of the sound wave from source to receiver (effective stimulus). Next, the definition of an auditory object is dealt with, specifically the question of how the various components of a sound stream become segregated into distinct auditory objects. The remainder of the chapter focuses on issues related to the spatial layout of auditory objects, both stationary and moving.
Zhang, Yu; Jiang, Jack J.
Nonlinear dynamic analysis and model simulations are used to study the nonlinear dynamic characteristics of vocal folds with vocal tremor, which can typically be characterized by low-frequency modulation and aperiodicity. Tremor voices from patients with disorders such as paresis, Parkinson's disease, hyperfunction, and adductor spasmodic dysphonia show low-dimensional characteristics, differing from random noise. Correlation dimension analysis statistically distinguishes tremor voices from normal voices. Furthermore, a nonlinear tremor model is proposed to study the vibrations of the vocal folds with vocal tremor. Fractal dimensions and positive Lyapunov exponents demonstrate the evidence of chaos in the tremor model, where amplitude and frequency play important roles in governing vocal fold dynamics. Nonlinear dynamic voice analysis and vocal fold modeling may provide a useful set of tools for understanding the dynamic mechanism of vocal tremor in patients with laryngeal diseases.
Chen, Min; Chen, Jian; Cheng, Lei; Wu, Haitao
This work aims to analyze the recurrence of vocal fold leukoplakia after carbon dioxide (CO 2 ) laser resection. In this retrospective study, all patients undergoing CO 2 laser resection of vocal fold leukoplakia were followed up for at least 2 years. Recurrence was diagnosed as any presence of leukoplakia in the vocal cord subsequent to previous successful complete resection. A total of 326 patients with complete resection of vocal fold leukoplakia and follow-up subsequent surveillance laryngoscopy were studied. The recurrence rate, the recurrence time, and risk factors were evaluated. Of these, 52 (16.0%) patients experienced recurrence with a mean follow-up time of 50.5 ± 15.4 months. The mean time to recurrence was 16.2 ± 14.1 months. Univariate analysis showed that the size of lesion (P vocal fold leukoplakia, long-term follow-up is required after CO 2 laser resection. In conclusion, the size of lesion combined with the pathological grade are important risk factors that predict vocal fold leukoplakia recurrence.
Lee, M D
Two experiments are presented that serve as a framework for exploring auditory information processing. The framework is referred to as polychotic listening or auditory search, and it requires a listener to scan multiple simultaneous auditory streams for the appearance of a target word (the name of a letter such as A or M). Participants' ability to scan between two and six simultaneous auditory streams of letter and digit names for the name of a target letter was examined using six loudspeakers. The main independent variable was auditory load, or the number of active audio streams on a given trial. The primary dependent variables were target localization accuracy and reaction time. Results showed that as load increased, performance decreased. The performance decrease was evident in reaction time, accuracy, and sensitivity measures. The second study required participants to practice the same task for 10 sessions, for a total of 1800 trials. Results indicated that even with extensive practice, performance was still affected by auditory load. The present results are compared with findings in the visual search literature. The implications for the use of multiple auditory displays are discussed. Potential applications include cockpit and automobile warning displays, virtual reality systems, and training systems.
Domanski, Mark C; Preciado, Diego A
Phrenic nerve pacing can be used to treat congenital central hypoventilation syndrome (CCHS). We report how the lack of normal vocal cord tone during phrenic paced respiration can result in passive vocal cord collapse and produce obstructive symptoms. We describe a case of passive vocal cord collapse during phrenic nerve paced respiration in a patient with CCHS. As far as we know, this is the first report of this etiology of airway obstruction. The patient, a 7-year-old with CCHS and normal waking vocal cord movement, continued to require nightly continuous positive airway pressure (CPAP) despite successful utilization of phrenic nerve pacers. On direct laryngoscopy, the patient's larynx was observed while the diaphragmatic pacers were sequentially engaged. No abnormal vocal cord stimulation was witnessed during engaging of either phrenic nerve stimulator. However, the lack of normal inspiratory vocal cord abduction during phrenic nerve-paced respiration resulted in vocal cord collapse and partial obstruction due to passive adduction of the vocal cords through the Bernoulli effect. Bilateral phrenic nerve stimulation resulted in more vocal cord collapse than unilateral stimulation. The lack of vocal cord abduction on inspiration presents a limit to phrenic nerve pacers.
Ulanovsky, Nachum; Las, Liora; Farkas, Dina; Nelken, Israel
Neurons in primary auditory cortex (A1) of cats show strong stimulus-specific adaptation (SSA). In probabilistic settings, in which one stimulus is common and another is rare, responses to common sounds adapt more strongly than responses to rare sounds. This SSA could be a correlate of auditory sensory memory at the level of single A1 neurons. Here we studied adaptation in A1 neurons, using three different probabilistic designs. We showed that SSA has several time scales concurrently, spanning many orders of magnitude, from hundreds of milliseconds to tens of seconds. Similar time scales are known for the auditory memory span of humans, as measured both psychophysically and using evoked potentials. A simple model, with linear dependence on both short-term and long-term stimulus history, provided a good fit to A1 responses. Auditory thalamus neurons did not show SSA, and their responses were poorly fitted by the same model. In addition, SSA increased the proportion of failures in the responses of A1 neurons to the adapting stimulus. Finally, SSA caused a bias in the neuronal responses to unbiased stimuli, enhancing the responses to eccentric stimuli. Therefore, we propose that a major function of SSA in A1 neurons is to encode auditory sensory memory on multiple time scales. This SSA might play a role in stream segregation and in binding of auditory objects over many time scales, a property that is crucial for processing of natural auditory scenes in cats and of speech and music in humans.
Rasmussen, Eva Rye; Mey, Kristianna
Ramsay Hunt syndrome is defined by herpes zoster oticus and peripheral facial nerve palsy which is often associated with otalgia. The syndrome is, in rare cases, associated with other cranial nerve paralyses including the vagal nerve causing unilateral vocal cord paralysis. Vocal cord paralysis...
Thoma, Robert J; Meier, Andrew; Houck, Jon; Clark, Vincent P; Lewine, Jeffrey D; Turner, Jessica; Calhoun, Vince; Stephen, Julia
Auditory sensory gating, assessed in a paired-click paradigm, indicates the extent to which incoming stimuli are filtered, or "gated", in auditory cortex. Gating is typically computed as the ratio of the peak amplitude of the event related potential (ERP) to a second click (S2) divided by the peak amplitude of the ERP to a first click (S1). Higher gating ratios are purportedly indicative of incomplete suppression of S2 and considered to represent sensory processing dysfunction. In schizophrenia, hallucination severity is positively correlated with gating ratios, and it was hypothesized that a failure of sensory control processes early in auditory sensation (gating) may represent a larger system failure within the auditory data stream; resulting in auditory verbal hallucinations (AVH). EEG data were collected while patients (N=12) with treatment-resistant AVH pressed a button to indicate the beginning (AVH-on) and end (AVH-off) of each AVH during a paired click protocol. For each participant, separate gating ratios were computed for the P50, N100, and P200 components for each of the AVH-off and AVH-on states. AVH trait severity was assessed using the Psychotic Symptoms Rating Scales AVH Total score (PSYRATS). The results of a mixed model ANOVA revealed an overall effect for AVH state, such that gating ratios were significantly higher during the AVH-on state than during AVH-off for all three components. PSYRATS score was significantly and negatively correlated with N100 gating ratio only in the AVH-off state. These findings link onset of AVH with a failure of an empirically-defined auditory inhibition system, auditory sensory gating, and pave the way for a sensory gating model of AVH. Copyright © 2017 Elsevier B.V. All rights reserved.
Vojnović, Milan; Bogavac, Ivana; Dobrijević, Ljiljana
The physical shape of vocal tract and its formant (resonant) frequencies are directly related. The study of this functional connectivity is essential in speech therapy practice with children. Most of the perceived children’s speech anomalies can be explained on a physical level: malfunctioning movement of articulation organs. The current problem is that there is no enough data on the anatomical shape of children’s vocal tract to create its acoustic model. Classical techniques for vocal tract...
Hopkins, Kevin; Kass, Steven J; Blalock, Lisa Durrance; Brill, J Christopher
In this study, we examined how spatially informative auditory and tactile cues affected participants' performance on a visual search task while they simultaneously performed a secondary auditory task. Visual search task performance was assessed via reaction time and accuracy. Tactile and auditory cues provided the approximate location of the visual target within the search display. The inclusion of tactile and auditory cues improved performance in comparison to the no-cue baseline conditions. In comparison to the no-cue conditions, both tactile and auditory cues resulted in faster response times in the visual search only (single task) and visual-auditory (dual-task) conditions. However, the effectiveness of auditory and tactile cueing for visual task accuracy was shown to be dependent on task-type condition. Crossmodal cueing remains a viable strategy for improving task performance without increasing attentional load within a singular sensory modality. Practitioner Summary: Crossmodal cueing with dual-task performance has not been widely explored, yet has practical applications. We examined the effects of auditory and tactile crossmodal cues on visual search performance, with and without a secondary auditory task. Tactile cues aided visual search accuracy when also engaged in a secondary auditory task, whereas auditory cues did not.
Marie A Gadziola
Full Text Available Bats are among the most gregarious and vocal mammals, with some species demonstrating a diverse repertoire of syllables under a variety of behavioral contexts. Despite extensive characterization of big brown bat (Eptesicus fuscus biosonar signals, there have been no detailed studies of adult social vocalizations. We recorded and analyzed social vocalizations and associated behaviors of captive big brown bats under four behavioral contexts: low aggression, medium aggression, high aggression, and appeasement. Even limited to these contexts, big brown bats possess a rich repertoire of social vocalizations, with 18 distinct syllable types automatically classified using a spectrogram cross-correlation procedure. For each behavioral context, we describe vocalizations in terms of syllable acoustics, temporal emission patterns, and typical syllable sequences. Emotion-related acoustic cues are evident within the call structure by context-specific syllable types or variations in the temporal emission pattern. We designed a paradigm that could evoke aggressive vocalizations while monitoring heart rate as an objective measure of internal physiological state. Changes in the magnitude and duration of elevated heart rate scaled to the level of evoked aggression, confirming the behavioral state classifications assessed by vocalizations and behavioral displays. These results reveal a complex acoustic communication system among big brown bats in which acoustic cues and call structure signal the emotional state of a caller.
Kayhan, Fatih; Uguz, Faruk; Kayhan, Ayşegül; Toktaş, Fikriye Ilay
Tics are stereotypical repetitive involuntary movements (motor tics) or sounds (vocal tics). Although the emergence of tics were reported in a few cases with the use of selective serotonin reuptake inhibitors, there was no case with bupropion extended-release (Bupropion XL). The current case report presents a male patient developing motor and vocal tics with the use of bupropion XL.
Full Text Available Voice, as a secondary sexual characteristic, is known to affect the perceived attractiveness of human individuals. But the underlying mechanism of vocal attractiveness has remained unclear. Here, we presented human listeners with acoustically altered natural sentences and fully synthetic sentences with systematically manipulated pitch, formants and voice quality based on a principle of body size projection reported for animal calls and emotional human vocal expressions. The results show that male listeners preferred a female voice that signals a small body size, with relatively high pitch, wide formant dispersion and breathy voice, while female listeners preferred a male voice that signals a large body size with low pitch and narrow formant dispersion. Interestingly, however, male vocal attractiveness was also enhanced by breathiness, which presumably softened the aggressiveness associated with a large body size. These results, together with the additional finding that the same vocal dimensions also affect emotion judgment, indicate that humans still employ a vocal interaction strategy used in animal calls despite the development of complex language.
Terry, Andrew Mark Ryder; Peake, Thomas More; McGregor, Peter Kenneth
Identifying the individuals within a population can generate information on life history parameters, generate input data for conservation models, and highlight behavioural traits that may affect management decisions and error or bias within census methods. Individual animals can be discriminated...... by features of their vocalisations. This vocal individuality can be utilised as an alternative marking technique in situations where the marks are difficult to detect or animals are sensitive to disturbance. Vocal individuality can also be used in cases were the capture and handling of an animal is either...... and techniques for using this to count and monitor populations over time. We present case studies in birds where vocal individuality has been applied to conservation and we discuss its role in mammals....
Hansen, Kirstin Anderson; Maxwell, Alyssa; Siebert, Ursula
In-air hearing in birds has been thoroughly investigated. Sound provides birds with auditory information for species and individual recognition from their complex vocalizations, as well as cues while foraging and for avoiding predators. Some 10% of existing species of birds obtain their food under...... the water surface. Whether some of these birds make use of acoustic cues while underwater is unknown. An interesting species in this respect is the great cormorant (Phalacrocorax carbo), being one of the most effective marine predators and relying on the aquatic environment for food year round. Here, its...... underwater hearing abilities were investigated using psychophysics, where the bird learned to detect the presence or absence of a tone while submerged. The greatest sensitivity was found at 2 kHz, with an underwater hearing threshold of 71 dB re 1 μPa rms. The great cormorant is better at hearing underwater...
Liu, Xinxin; Zheng, Wei; Sivasankar, M Preeti
Acrolein is a ubiquitous pollutant abundant in cigarette smoke, mobile exhaust, and industrial waste. There is limited literature on the effects of acrolein on vocal fold tissue, although there are clinical reports of voice changes after pollutant exposures. Vocal folds are responsible for voice production. The overall objective of this study was to investigate the effects of acrolein exposure on viable, excised vocal fold epithelial tissue and to characterize the mechanism underlying acrolein toxicity. Vocal fold epithelia were studied because they form the outermost layer of the vocal folds and are a primary recipient of inhaled pollutants. Porcine vocal fold epithelia were exposed to 0, 50, 100, 500, 900 or 1300 μM of acrolein for 3 hours; the metabolic activity, epithelial resistance, epithelial permeability, tight junction protein (occludin and claudin 3) expression, cell membrane integrity and lipid peroxidation were investigated. The data demonstrated that acrolein exposure at 500 μM significantly reduced vocal fold epithelial metabolic activity by 27.2% (p≤0.001). Incubation with 100 μM acrolein caused a marked increase in epithelial permeability by 130.5% (pacrolein-treated samples, the cell membrane integrity was significantly damaged with a 45.6% increase of lipid peroxidation as compared to controls (pacrolein exposure impairs vocal fold epithelial barrier integrity. Lipid peroxidation-induced cell membrane damage may play an important role in reducing the barrier function of the epithelium.
Erich S Tusch
Full Text Available The inhibitory deficit hypothesis of cognitive aging posits that older adults' inability to adequately suppress processing of irrelevant information is a major source of cognitive decline. Prior research has demonstrated that in response to task-irrelevant auditory stimuli there is an age-associated increase in the amplitude of the N1 wave, an ERP marker of early perceptual processing. Here, we tested predictions derived from the inhibitory deficit hypothesis that the age-related increase in N1 would be 1 observed under an auditory-ignore, but not auditory-attend condition, 2 attenuated in individuals with high executive capacity (EC, and 3 augmented by increasing cognitive load of the primary visual task. ERPs were measured in 114 well-matched young, middle-aged, young-old, and old-old adults, designated as having high or average EC based on neuropsychological testing. Under the auditory-ignore (visual-attend task, participants ignored auditory stimuli and responded to rare target letters under low and high load. Under the auditory-attend task, participants ignored visual stimuli and responded to rare target tones. Results confirmed an age-associated increase in N1 amplitude to auditory stimuli under the auditory-ignore but not auditory-attend task. Contrary to predictions, EC did not modulate the N1 response. The load effect was the opposite of expectation: the N1 to task-irrelevant auditory events was smaller under high load. Finally, older adults did not simply fail to suppress the N1 to auditory stimuli in the task-irrelevant modality; they generated a larger response than to identical stimuli in the task-relevant modality. In summary, several of the study's findings do not fit the inhibitory-deficit hypothesis of cognitive aging, which may need to be refined or supplemented by alternative accounts.
Yevgeniy B Sirotin
Full Text Available During rodent active behavior, multiple orofacial sensorimotor behaviors, including sniffing and whisking, display rhythmicity in the theta range (~5-10 Hz. During specific behaviors, these rhythmic patterns interlock, such that execution of individual motor programs becomes dependent on the state of the others. Here we performed simultaneous recordings of the respiratory cycle and ultrasonic vocalization emission by adult rats and mice in social settings. We used automated analysis to examine the relationship between breathing patterns and vocalization over long time periods. Rat ultrasonic vocalizations (USVs, ’50 kHz’ were emitted within stretches of active sniffing (5−10 Hz and were largely absent during periods of passive breathing (1-4 Hz. Because ultrasound was tightly linked to the exhalation phase, the sniffing cycle segmented vocal production into discrete calls and imposed its theta rhythmicity on their timing. In turn, calls briefly prolonged exhalations, causing an immediate drop in sniffing rate. Similar results were obtained in mice. Our results show that ultrasonic vocalizations are an integral part of the rhythmic orofacial behavioral ensemble. This complex behavioral program is thus involved not only in active sensing but also in the temporal structuring of social communication signals. Many other social signals of mammals, including monkey calls and human speech, show structure in the theta range. Our work points to a mechanism for such structuring in rodent ultrasonic vocalizations.
Rosow, David E
Office-based vocal fold injections have become increasingly popular over the past 15 years. Examination of trends in procedure coding for vocal fold injections in the United States from 2000 to 2012 was undertaken to see if they reflect this shift. The US Part B Medicare claims database was queried from 2000 through 2012 for multiple Current Procedural Terminology codes. Over the period studied, the number of nonoperative laryngoscopic injections (31513, 31570) and operative medialization laryngoplasties (31588) remained constant. Operative vocal fold injection (31571) demonstrated marked linear growth over the 12-year study period, from 744 procedures in 2000 to 4788 in 2012-an increase >640%. The dramatic increased incidence in the use of code 31571 reflects an increasing share of vocal fold injections being performed in the operating room and not in an office setting, running counter to the prevailing trend toward awake, office-based injection procedures. © American Academy of Otolaryngology—Head and Neck Surgery Foundation 2015.
Loiola-Barreiro, Camila Miranda; Silva, Marta Assumpção de Andrada E
To compare the voice handicap index of popular and erudite professional singers according to gender, age, professional experience time, and presence or absence of self-reported vocal complaints. One hundred thirty-two professional singers, 74 popular and 58 erudite, who responded to a questionnaire with regards to identification, age, gender, professional experience time in singing, musical genres (for popular singers), vocal classification (for erudite singers), presence of self-reported vocal complaints, and the specific protocols for popular (Modern Singing Handicap Index - MSHI) and erudite (Classical Singing Handicap Index - CSHI) singing. Higher proportion of women and higher incidence of vocal complaints were observed in the popular singers compared with the erudite singers. Most of the popular singers belonged to the genre of Brazilian Popular Music. Regarding the classification of erudite singers, there was greater participation of sopranos and tenors. No statistical differences were observed with respect to age and professional experience time between the groups. Comparison of the MSHI and CSHI scores showed no statistically significant difference between these scores and genre or age in both groups of singers. Professional experience time was related to the total score and the subscales disability and impairment in the MSHI, only for popular singers with vocal complaints. There was no correlation between these variables and the CSHI for erudite singers. The impact of vocal difficulty/problem interferes differently in these two musical genres when related to vocal complaint and professional experience time. The MSHI and CSHI protocols proved to be important tools not only for the identification of problems, but also for the understanding of how these individuals relate their voices with this occupational activity.
Yin, Jun; Zhang, Zhaoyan
The influence of the thyroarytenoid (TA) and cricothyroid (CT) muscle activation on vocal fold stiffness and eigenfrequencies was investigated in a muscularly controlled continuum model of the vocal folds. Unlike the general understanding that vocal fold fundamental frequency was determined by vocal fold tension, this study showed that vocal fold eigenfrequencies were primarily determined by vocal fold stiffness. This study further showed that, with reference to the resting state of zero stra...
Konopka, Genevieve; Roberts, Todd F
Disruptions in speech, language, and vocal communication are hallmarks of several neuropsychiatric disorders, most notably autism spectrum disorders. Historically, the use of animal models to dissect molecular pathways and connect them to behavioral endophenotypes in cognitive disorders has proven to be an effective approach for developing and testing disease-relevant therapeutics. The unique aspects of human language compared with vocal behaviors in other animals make such an approach potentially more challenging. However, the study of vocal learning in species with analogous brain circuits to humans may provide entry points for understanding this human-specific phenotype and diseases. We review animal models of vocal learning and vocal communication and specifically link phenotypes of psychiatric disorders to relevant model systems. Evolutionary constraints in the organization of neural circuits and synaptic plasticity result in similarities in the brain mechanisms for vocal learning and vocal communication. Comparative approaches and careful consideration of the behavioral limitations among different animal models can provide critical avenues for dissecting the molecular pathways underlying cognitive disorders that disrupt speech, language, and vocal communication. Copyright © 2016 Society of Biological Psychiatry. Published by Elsevier Inc. All rights reserved.
Seymour, N; Burkill, G; Harries, M
Positron emission tomography-computed tomography with fluorine-18 fluorodeoxy-D-glucose has a major role in the investigation of head and neck cancers. Fluorine-18 fluorodeoxy-D-glucose is not a tumour-specific tracer and can also accumulate in benign pathology. Therefore, positron emission tomography-computed tomography scan interpretation difficulties are common in the head and neck, which can produce false-positive results. This study aimed to investigate patients detected as having abnormal vocal fold uptake on fluorine-18 fluorodeoxy-D-glucose positron emission tomography-computed tomography. Positron emission tomography-computed tomography scans were identified over a 15-month period where reports contained evidence of unilateral vocal fold uptake or vocal fold pathology. Patients' notes and laryngoscopy results were analysed. Forty-six patients were identified as having abnormal vocal fold uptake on positron emission tomography-computed tomography. Twenty-three patients underwent positron emission tomography-computed tomography and flexible laryngoscopy: 61 per cent of patients had true-positive positron emission tomography-computed tomography scans and 39 per cent had false-positive scan results. Most patients referred to ENT for abnormal findings on positron emission tomography-computed tomography scans had true-positive findings. Asymmetrical fluorine-18 fluorodeoxy-D-glucose uptake should raise suspicion of vocal fold pathology, accepting a false-positive rate of approximately 40 per cent.
Taitz, Alan; Shalom, Diego E.; Trevisan, Marcos A.
Speech requires programming the sequence of vocal gestures that produce the sounds of words. Here we explored the timing of this program by asking our participants to pronounce, as quickly as possible, a sequence of consonant-consonant-vowel (CCV) structures appearing on screen. We measured the delay between visual presentation and voice onset. In the case of plosive consonants, produced by sharp and well defined movements of the vocal tract, we found that delays are positively correlated with the duration of the transition between consonants. We then used a battery of statistical tests and mathematical vocal models to show that delays reflect the motor planning of CCVs and transitions are proxy indicators of the vocal effort needed to produce them. These results support that the effort required to produce the sequence of movements of a vocal gesture modulates the onset of the motor plan.
Kirgezen, Tolga; Sunter, Ahmet Volkan; Yigit, Ozgur; Huq, Gulben Erdem
The study aimed to evaluate the existence of sex hormone receptors in the subunits of vocal fold. This is a cadaver study. The androgen, estrogen, and progesterone receptors were examined in the epithelium (EP), superficial layer of the lamina propria (SLP), vocal ligament (VL), and macula flava (MF) of the vocal folds from 42 human cadavers (21 male, 21 female) by immunohistochemical methods. Their staining ratios were scored and statistically compared. The androgen receptor score was significantly higher for the MF than for the EP and SLP (P vocal fold, mostly in the MF and VLs. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Arya, Divya D.
This article offers information that will allow music educators to incorporate North Indian classical vocal music into a multicultural music education curriculum. Obstacles to teaching North Indian classical vocal music are acknowledged, including lack of familiarity with the cultural/structural elements and challenges in teaching ear training and…
Full Text Available The neural substrate for the perception of vocalization is relatively well described, but we know much less about how the timing and specificity of vocalizations is tightly coupled with audiovocal communication behavior. In many vocal species, well-timed vocalizations accompany fear, vigilance and aggression. These emotive responses likely originate within the amygdala and other limbic structures, but the organization of motor outputs for triggering species-appropriate behaviors remains unclear. We performed electrical microstimulation at 461 highly restricted loci within the basal and central amygdala in awake mustached bats. At a subset of these sites, high frequency stimulation with weak constant current pulses presented at near-threshold levels triggered vocalization of either echolocation pulses or social calls. At the vast majority of locations, microstimulation produced a constellation of changes in autonomic and somatomotor outputs. These changes included widespread co-activation of significant tachycardia and hyperventilation and/or rhythmic ear pinna movements. In a few locations, responses were constrained to vocalization and/or pinna movements despite increases in the intensity of stimulation. The probability of eliciting echolocation pulses versus social calls decreased in a medial-posterior to anterolateral direction within the centrobasal amygdala. Microinjections of kainic acid at stimulation sites confirmed the contribution of cellular activity rather than fibers-of-passage in the control of multimodal outputs. The results suggest that multimodal clusters of neurons may simultaneously modulate the activity of multiple central pattern generators present within the brainstem.
Zimmermann, Jacqueline F; Moscovitch, Morris; Alain, Claude
Attention to memory describes the process of attending to memory traces when the object is no longer present. It has been studied primarily for representations of visual stimuli with only few studies examining attention to sound object representations in short-term memory. Here, we review the interplay of attention and auditory memory with an emphasis on 1) attending to auditory memory in the absence of related external stimuli (i.e., reflective attention) and 2) effects of existing memory on guiding attention. Attention to auditory memory is discussed in the context of change deafness, and we argue that failures to detect changes in our auditory environments are most likely the result of a faulty comparison system of incoming and stored information. Also, objects are the primary building blocks of auditory attention, but attention can also be directed to individual features (e.g., pitch). We review short-term and long-term memory guided modulation of attention based on characteristic features, location, and/or semantic properties of auditory objects, and propose that auditory attention to memory pathways emerge after sensory memory. A neural model for auditory attention to memory is developed, which comprises two separate pathways in the parietal cortex, one involved in attention to higher-order features and the other involved in attention to sensory information. This article is part of a Special Issue entitled SI: Auditory working memory. Copyright © 2015 Elsevier B.V. All rights reserved.
Nikkhah-Bahrami, Mansour; Ahmadi-Noubari, Hossein; Seyed Aghazadeh, Babak; Khadivi Heris, Hossein
This paper explores the use of hierarchical structure for diagnosis of vocal fold disorders. The hierarchical structure is initially used to train different second-level classifiers. At the first level normal and pathological signals have been distinguished. Next, pathological signals have been classified into neurogenic and organic vocal fold disorders. At the final level, vocal fold nodules have been distinguished from polyps in organic disorders category. For feature selection at each level of hierarchy, the reconstructed signal at each wavelet packet decomposition sub-band in 5 levels of decomposition with mother wavelet of (db10) is used to extract the nonlinear features of self-similarity and approximate entropy. Also, wavelet packet coefficients are used to measure energy and Shannon entropy features at different spectral sub-bands. Davies-Bouldin criterion has been employed to find the most discriminant features. Finally, support vector machines have been adopted as classifiers at each level of hierarchy resulting in the diagnosis accuracy of 92%.
Wingstrand, Vibe Lindeblad; Larsen, Christian Grønhøj; Jensen, David H
OBJECTIVES: Therapy with mesenchymal stem cells exhibits potential for the development of novel interventions for many diseases and injuries. The use of mesenchymal stem cells in regenerative therapy for vocal fold scarring exhibited promising results to reduce stiffness and enhance...... the biomechanical properties of injured vocal folds. This study evaluated the biomechanical effects of mesenchymal stem cell therapy for the treatment of vocal fold scarring. DATA SOURCES: PubMed, Embase, the Cochrane Library and Google Scholar were searched. METHODS: Controlled studies that assessed...... the biomechanical effects of mesenchymal stem cell therapy for the treatment of vocal fold scarring were included. Primary outcomes were viscoelastic properties and mucosal wave amplitude. RESULTS: Seven preclinical animal studies (n = 152 single vocal folds) were eligible for inclusion. Evaluation of viscoelastic...
Hamdan, Abdul-Latif; Khalifee, Elie; Tabet, Georges
The objective of this study was to report the first case of unilateral vocal fold paralysis in a patient with Parkinson disease (PD) and to review the literature. This is a case report and literature review following PubMed search using the keywords "Parkinson," "vocal fold paralysis," "vocal fold palsy," "vocal fold immobility," "vocal fold adductor palsy," "airway obstruction," and "stridor." A total of 18 subjects diagnosed with PD and vocal fold paralysis were described. In all cases, the vocal fold paralysis was bilateral and the main presenting symptoms were stridor and shortness of breath necessitating intubation and tracheostomy. This article describes the first case of PD presenting with dysphonia secondary to unilateral vocal fold paralysis (left). The management consisted of injection laryngoplasty for medialization of the paralyzed vocal fold. Patients with PD can present with unilateral vocal fold paralysis. Early treatment is advocated in view of the advent of injection laryngoplasty as a safe office procedure. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Yui K Matsumoto
Full Text Available Mice produce ultrasonic vocalizations featuring a variety of syllables. Vocalizations are observed during social interactions. In particular, males produce numerous syllables during courtship. Previous studies have shown that vocalizations change according to sexual behavior, suggesting that males vary their vocalizations depending on the phase of the courtship sequence. To examine this process, we recorded large sets of mouse vocalizations during male-female interactions and acoustically categorized these sounds into 12 vocal types. We found that males emitted predominantly short syllables during the first minute of interaction, more long syllables in the later phases, and mainly harmonic sounds during mounti