de Falco, Simona; Venuti, Paola; Esposito, Gianluca; Bornstein, Marc H
2011-02-01
The aim of this study was to compare functional features of maternal and paternal speech directed to children with Down syndrome and developmental age-matched typically developing children. Altogether 88 parents (44 mothers and 44 fathers) and their 44 young children (22 children with Down syndrome and 22 typically developing children) participated. Parents' speech directed to children was obtained through observation of naturalistic parent-child dyadic interactions. Verbatim transcripts of maternal and paternal language were categorized in terms of the primary function of each speech unit. Parents (both mothers and fathers) of children with Down syndrome used more affect-salient speech compared to parents of typically developing children. Although parents used the same amounts of information-salient speech, parents of children with Down syndrome used more direct statements and asked fewer questions than did parents of typically developing children. Concerning parent gender, in both groups mothers used more language than fathers and specifically more descriptions. These findings held controlling for child age and MLU and family SES. This study highlights strengths and weaknesses of parental communication to children with Down syndrome and helps to identify areas of potential improvement through intervention. Copyright © 2010 Elsevier Inc. All rights reserved.
Role of maternal gesture use in speech use by children with fragile X syndrome.
Hahn, Laura J; Zimmer, B Jean; Brady, Nancy C; Swinburne Romine, Rebecca E; Fleming, Kandace K
2014-05-01
The purpose of this study was to investigate how maternal gesture relates to speech production by children with fragile X syndrome (FXS). Participants were 27 young children with FXS (23 boys, 4 girls) and their mothers. Videotaped home observations were conducted between the ages of 25 and 37 months (toddler period) and again between the ages of 60 and 71 months (child period). The videos were later coded for types of maternal utterances and maternal gestures that preceded child speech productions. Children were also assessed with the Mullen Scales of Early Learning at both ages. Maternal gesture use in the toddler period was positively related to expressive language scores at both age periods and was related to receptive language scores in the child period. Maternal proximal pointing, in comparison to other gestures, evoked more speech responses from children during the mother-child interactions, particularly when combined with wh-questions. This study adds to the growing body of research on the importance of contextual variables, such as maternal gestures, in child language development. Parental gesture use may be an easily added ingredient to parent-focused early language intervention programs.
Directory of Open Access Journals (Sweden)
Zarei S
2015-05-01
Full Text Available Background: Nowadays, mothers are continuously exposed to different sources of electromagnetic fields before and even during pregnancy. It has recently been shown that exposure to mobile phone radiation during pregnancy may lead to adverse effects on the brain development in offspring and cause hyperactivity. Researchers have shown that behavioral problems in laboratory animals which have a similar appearance to ADHD are caused by intrauterine exposure to mobile phones. Objective: The purpose of this study was to investigate whether the maternal exposure to different sources of electromagnetic fields affect on the rate and severity of speech problems in their offspring. Methods: In this study, mothers of 35 healthy 3-5 year old children (control group and 77 children and diagnosed with speech problems who had been referred to a speech treatment center in Shiraz, Iran were interviewed. These mothers were asked whether they had exposure to different sources of electromagnetic fields such as mobile phones, mobile base stations, Wi-Fi, cordless phones, laptops and power lines. Results: We found a significant association between either the call time (P=0.002 or history of mobile phone use (months used and speech problems in the offspring (P=0.003. However, other exposures had no effect on the occurrence of speech problems. To the best of our knowledge, this is the first study to investigate a possible association between maternal exposure to electromagnetic field and speech problems in the offspring. Although a major limitation in our study is the relatively small sample size, this study indicates that the maternal exposure to common sources of electromagnetic fields such as mobile phones can affect the occurrence of speech problems in the offspring.
de Falco, Simona; Venuti, Paola; Esposito, Gianluca; Bornstein, Marc H.
2011-01-01
The aim of this study was to compare functional features of maternal and paternal speech directed to children with Down syndrome and developmental age-matched typically developing children. Altogether 88 parents (44 mothers and 44 fathers) and their 44 young children (22 children with Down syndrome and 22 typically developing children) participated. Parents’ speech directed to children was obtained through observation of naturalistic parent–child dyadic interactions. Verbatim transcripts of m...
Directory of Open Access Journals (Sweden)
Daniela Carvalho Neves
2012-01-01
Full Text Available This study had as its objective to investigate the knowledge of pregnant teenagers on Speech Therapy related to maternal-infant health care. A qualitative analysis was made based on a thematic investigation of the subject matter. Ten pregnant teenagers with chronological age between ten and nineteen years old joined in the survey. They found themselves around the fifth and ninth months of gestation age, being attended at the Center of Pregnant TeenagersAttention Care at Fortaleza General Hospital. The collection of data involved the application of a semi-structured interview broaching topics that could identify what the pregnant teenagers knew about Speech Therapy and maternal-infant care. Education interventions, related to Speech Therapy health promotion, were also accomplished. The results pointed out that the pregnant teenagers’ level of knowledge on aspects such as: food transition and utensils, oral habits, language stimulation and hearing loss detection, was still incipient and unsatisfactory. It is concluded that the knowledge of the pregnant teenagers on Speech Therapy related to maternal-infant health care was unsatisfactory, which demonstrates the importance of education interventions related to human communication health care for this studied sample.
International Nuclear Information System (INIS)
Kluin, K.J.; Gilman, S.; Markel, D.S.; Koeppe, R.A.; Rosenthal, G.; Junck, L.
1988-01-01
We compared the severity of ataxic and spastic dysarthria with local cerebral metabolic rates for glucose (lCMRGlc) in 30 patients with olivopontocerebellar atrophy (OPCA). Perceptual analysis was used to examine the speech disorders, and rating scales were devised to quantitate the degree of ataxia and spasticity in the speech of each patient. lCMRGlc was measured with 18 F-2-fluoro-2-deoxy-D-glucose and positron emission tomography (PET). PET studies revealed marked hypometabolism in the cerebellar hemispheres, cerebellar vermis, and brainstem of OPCA patients compared with 30 control subjects. With data normalized to the cerebral cortex, a significant inverse correlation was found between the severity of ataxia in speech and the lCMRGlc within the cerebellar vermis, cerebellar hemispheres, and brainstem, but not within the thalamus. No significant correlation was found between the severity of spasticity in speech and lCMRGlc in any of these structures. The findings support the view that the severity of ataxia in speech in OPCA is related to the functional activity of the cerebellum and its connections in the brainstem
Poor Speech Perception Is Not a Core Deficit of Childhood Apraxia of Speech: Preliminary Findings
Zuk, Jennifer; Iuzzini-Seigel, Jenya; Cabbage, Kathryn; Green, Jordan R.; Hogan, Tiffany P.
2018-01-01
Purpose: Childhood apraxia of speech (CAS) is hypothesized to arise from deficits in speech motor planning and programming, but the influence of abnormal speech perception in CAS on these processes is debated. This study examined speech perception abilities among children with CAS with and without language impairment compared to those with…
Abnormal maternal echocardiographic findings in triplet pregnancies presenting with dyspnoea.
Elhenicky, Marie; Distelmaier, Klaus; Mailath-Pokorny, Mariella; Worda, Christof; Langer, Martin; Worda, Katharina
2016-03-01
The objective of our study was to evaluate the prevalence of abnormal maternal echocardiographic findings in triplet pregnancies presenting with dyspnoea. Between 2003 and 2013, patients' records of 96 triplet pregnancies at our department were analysed including maternal and fetal outcome, echocardiographic parameters and N-terminal pro-B-type natriuretic peptide (NT-proBNP) levels. After exclusion of triplet pregnancies with fetal demise before 23 + 0 weeks, selective feticide or missing outcome data, the study population consisted of 60 triplet pregnancies. All women with dyspnoea underwent echocardiography and measurement of NT-proBNP. Dyspnoea towards the end of pregnancy was observed in 13.3% (8/60) of all women with triplet pregnancies, and all of these women underwent echocardiography. The prevalence of abnormal echocardiographic findings in women with dyspnoea was 37.5% (3/8) with peripartum cardiomyopathy in one woman. Median serum NT-proBNP was significantly higher in women with abnormal echocardiographic findings compared with those without (1779 ng/ml, range 1045-6076 ng/ml vs 172 ng/ml, range 50-311 ng/ml; p presenting with dyspnoea show a high prevalence of abnormal echocardiographic findings. Since dyspnoea is a common sign in triplet pregnancies and is associated with a high rate of cardiac involvement, echocardiography and evaluation of maternal NT-proBNP could be considered to improve early diagnosis and perinatal management.
Gravener, Julie A.; Rogosch, Fred A.; Oshri, Assaf; Narayan, Angela J.; Cicchetti, Dante; Toth, Sheree L.
2012-01-01
Direct and indirect relations among maternal depression, maternal Expressed Emotion (EE: Self- and Child-Criticism), child internalizing and externalizing symptoms, and child attachment were examined. Participants were mothers with depression (n = 130) and comparison mothers (n = 68) and their toddlers (M age = 20 mo.; 53% male). Assessments included the Diagnostic Interview Schedule (maternal depression); the Five Minute Speech Sample (EE); the Child Behavior Checklist (toddler behavior prob...
Applying Corpus-Based Findings to Form-Focused Instruction: The Case of Reported Speech
Barbieri, Federica; Eckhardt, Suzanne E. B.
2007-01-01
Arguing that the introduction of corpus linguistics in teaching materials and the language classroom should be informed by theories and principles of SLA, this paper presents a case study illustrating how corpus-based findings on reported speech can be integrated into a form-focused model of instruction. After overviewing previous work which…
New findings for maternal mortality age patterns: aggregated results for 38 countries.
Directory of Open Access Journals (Sweden)
Ann K Blanc
Full Text Available With recent results showing a global decline in overall maternal mortality during the last two decades and with the target date for achieving the Millennium Development Goals only four years away, the question of how to continue or even accelerate the decline has become more pressing. By knowing where the risk is highest as well as where the numbers of deaths are greatest, it may be possible to re-direct resources and fine-tune strategies for greater effectiveness in efforts to reduce maternal mortality.We aggregate data from 38 Demographic and Health Surveys that included a maternal mortality module and were conducted in 2000 or later to produce maternal mortality ratios, rates, and numbers of deaths by five year age groups, separately by residence, region, and overall mortality level.The age pattern of maternal mortality is broadly similar across regions, type of place of residence, and overall level of maternal mortality. A "J" shaped curve, with markedly higher risk after age 30, is evident in all groups. We find that the excess risk among adolescents is of a much lower magnitude than is generally assumed. The oldest age groups appear to be especially resistant to change. We also find evidence of extremely elevated risk among older mothers in countries with high levels of HIV prevalence.The largest number of deaths occurs in the age groups from 20-34, largely because those are the ages at which women are most likely to give birth so efforts directed at this group would most effectively reduce the number of deaths. Yet equity considerations suggest that efforts also be directed toward those most at risk, i.e., older women and adolescents. Because women are at risk each time they become pregnant, fulfilling the substantial unmet need for contraception is a cross-cutting strategy that can address both effectiveness and equity concerns.
The analysis of speech acts patterns in two Egyptian inaugural speeches
Directory of Open Access Journals (Sweden)
Imad Hayif Sameer
2017-09-01
Full Text Available The theory of speech acts, which clarifies what people do when they speak, is not about individual words or sentences that form the basic elements of human communication, but rather about particular speech acts that are performed when uttering words. A speech act is the attempt at doing something purely by speaking. Many things can be done by speaking. Speech acts are studied under what is called speech act theory, and belong to the domain of pragmatics. In this paper, two Egyptian inaugural speeches from El-Sadat and El-Sisi, belonging to different periods were analyzed to find out whether there were differences within this genre in the same culture or not. The study showed that there was a very small difference between these two speeches which were analyzed according to Searle’s theory of speech acts. In El Sadat’s speech, commissives came to occupy the first place. Meanwhile, in El–Sisi’s speech, assertives occupied the first place. Within the speeches of one culture, we can find that the differences depended on the circumstances that surrounded the elections of the Presidents at the time. Speech acts were tools they used to convey what they wanted and to obtain support from their audiences.
de Rauville, Ingrid; Chetty, Sandhya; Pahl, Jenny
2006-01-01
Word finding difficulties frequently found in learners with language learning difficulties (Casby, 1992) are an integral part of Speech-Language Therapists' management role when working with learning disabled children. This study investigated current management for word finding difficulties by 70 Speech-Language Therapists in South African remedial schools. A descriptive survey design using a quantitative and qualitative approach was used. A questionnaire and follow-up focus group discussion were used to collect data. Results highlighted the use of the Renfrew Word Finding Scale (Renfrew, 1972, 1995) as the most frequently used formal assessment tool. Language sample analysis and discourse analysis were the most frequently used informal assessment procedures. Formal intervention programmes were generally not used. Phonetic, phonemic or phonological cueing were the most frequently used therapeutic strategies. The authors note strengths and raise concerns about current management for word finding difficulties in South African remedial schools, particularly in terms of bilingualism. Opportunities are highlighted regarding the development of assessment and intervention measures relevant to the diverse learning disabled population in South Africa.
Childhood apraxia of speech: A survey of praxis and typical speech characteristics.
Malmenholt, Ann; Lohmander, Anette; McAllister, Anita
2017-07-01
The purpose of this study was to investigate current knowledge of the diagnosis childhood apraxia of speech (CAS) in Sweden and compare speech characteristics and symptoms to those of earlier survey findings in mainly English-speakers. In a web-based questionnaire 178 Swedish speech-language pathologists (SLPs) anonymously answered questions about their perception of typical speech characteristics for CAS. They graded own assessment skills and estimated clinical occurrence. The seven top speech characteristics reported as typical for children with CAS were: inconsistent speech production (85%), sequencing difficulties (71%), oro-motor deficits (63%), vowel errors (62%), voicing errors (61%), consonant cluster deletions (54%), and prosodic disturbance (53%). Motor-programming deficits described as lack of automatization of speech movements were perceived by 82%. All listed characteristics were consistent with the American Speech-Language-Hearing Association (ASHA) consensus-based features, Strand's 10-point checklist, and the diagnostic model proposed by Ozanne. The mode for clinical occurrence was 5%. Number of suspected cases of CAS in the clinical caseload was approximately one new patient/year and SLP. The results support and add to findings from studies of CAS in English-speaking children with similar speech characteristics regarded as typical. Possibly, these findings could contribute to cross-linguistic consensus on CAS characteristics.
Marchetto, Erika; Bonatti, Luca L.
2015-01-01
To achieve language proficiency, infants must find the building blocks of speech and master the rules governing their legal combinations. However, these problems are linked: words are also built according to rules. Here, we explored early morphosyntactic sensitivity by testing when and how infants could find either words or within-word structure…
Finding the music of speech: Musical knowledge influences pitch processing in speech.
Vanden Bosch der Nederlanden, Christina M; Hannon, Erin E; Snyder, Joel S
2015-10-01
Few studies comparing music and language processing have adequately controlled for low-level acoustical differences, making it unclear whether differences in music and language processing arise from domain-specific knowledge, acoustic characteristics, or both. We controlled acoustic characteristics by using the speech-to-song illusion, which often results in a perceptual transformation to song after several repetitions of an utterance. Participants performed a same-different pitch discrimination task for the initial repetition (heard as speech) and the final repetition (heard as song). Better detection was observed for pitch changes that violated rather than conformed to Western musical scale structure, but only when utterances transformed to song, indicating that music-specific pitch representations were activated and influenced perception. This shows that music-specific processes can be activated when an utterance is heard as song, suggesting that the high-level status of a stimulus as either language or music can be behaviorally dissociated from low-level acoustic factors. Copyright © 2015 Elsevier B.V. All rights reserved.
Yun, Chunfeng; Wang, Zhenjie; He, Ping; Guo, Chao; Chen, Gong; Zheng, Xiaoying
2016-01-01
Although the prevalence of oral clefts in China is among the highest worldwide, little is known about the prevalence of speech disability associated with cleft palate in Chinese children. The data for this study were collected from the Second China National Sample Survey on Disability, and identification of speech disability associated with cleft palate was based on consensus manuals. Logistic regression was used to estimate odds ratios (ORs) and 95% confidence intervals (CIs). A weighted number of 112,070 disabled children affected by cleft palate were identified, yielding a prevalence of 3.45 per 10,000 children (95% CI: 3.19–3.71). A history of speech disability in the mother (OR = 20.266, 95% CI 5.788–70.959, p cleft palate in the offspring. Our results showed that maternal speech disability, older paternal child-bearing age, and lower levels of parental education were independent risk factors for speech disability associated with cleft palate for children in China. These findings may have important implications for health disparities and prevention. PMID:27886104
Common neural substrates support speech and non-speech vocal tract gestures.
Chang, Soo-Eun; Kenney, Mary Kay; Loucks, Torrey M J; Poletto, Christopher J; Ludlow, Christy L
2009-08-01
The issue of whether speech is supported by the same neural substrates as non-speech vocal tract gestures has been contentious. In this fMRI study we tested whether producing non-speech vocal tract gestures in humans shares the same functional neuroanatomy as non-sense speech syllables. Production of non-speech vocal tract gestures, devoid of phonological content but similar to speech in that they had familiar acoustic and somatosensory targets, was compared to the production of speech syllables without meaning. Brain activation related to overt production was captured with BOLD fMRI using a sparse sampling design for both conditions. Speech and non-speech were compared using voxel-wise whole brain analyses, and ROI analyses focused on frontal and temporoparietal structures previously reported to support speech production. Results showed substantial activation overlap between speech and non-speech function in regions. Although non-speech gesture production showed greater extent and amplitude of activation in the regions examined, both speech and non-speech showed comparable left laterality in activation for both target perception and production. These findings posit a more general role of the previously proposed "auditory dorsal stream" in the left hemisphere--to support the production of vocal tract gestures that are not limited to speech processing.
Practical speech user interface design
Lewis, James R
2010-01-01
Although speech is the most natural form of communication between humans, most people find using speech to communicate with machines anything but natural. Drawing from psychology, human-computer interaction, linguistics, and communication theory, Practical Speech User Interface Design provides a comprehensive yet concise survey of practical speech user interface (SUI) design. It offers practice-based and research-based guidance on how to design effective, efficient, and pleasant speech applications that people can really use. Focusing on the design of speech user interfaces for IVR application
Maternal depression and child BMI: longitudinal findings from a US sample.
Duarte, C S; Shen, S; Wu, P; Must, A
2012-04-01
To examine the association between maternal depression and child body mass index (BMI) from Kindergarten (K) to fifth grade. Analysis of four waves of data from the Early Childhood Longitudinal Study - Kindergarten spanning K to fifth grade. Maternal depressive symptoms (MDSs) were measured by a brief version of the Center for Epidemiological Studies Depression scale. Data were analyzed using multiple regression analyses, adjusting for key covariates and potential confounders. The analytic sample was restricted to children of normal birth weight. The relationship between MDS and child BMI varies by child gender and age. Among girls, severe MDS at K was related to lower BMI at third grade (but not later at fifth grade) and to an increase in BMI from K to third and K to fifth grades. Among boys, severe MDS at K was related to higher boys' BMI at fifth grade. When severe MDS occurred at third grade, it was related to higher BMI at fifth grade among girls whereas no statistically significant relationship was found for boys. Low levels of physical activity in comparison to peers at fifth grade and more screen time on weekends at third grade are likely mediators of the relationship between MDS and child BMI among girls, while among boys the relationship appears to be mediated by unhealthy eating habits. Our findings, indicating developmental and gender differences in the relationship between maternal depression and child BMI, if confirmed, suggest that interventions addressing maternal depression may have concomitant impact on childhood obesity. © 2012 The Authors. Pediatric Obesity © 2012 International Association for the Study of Obesity.
Greene, Beth G; Logan, John S; Pisoni, David B
1986-03-01
We present the results of studies designed to measure the segmental intelligibility of eight text-to-speech systems and a natural speech control, using the Modified Rhyme Test (MRT). Results indicated that the voices tested could be grouped into four categories: natural speech, high-quality synthetic speech, moderate-quality synthetic speech, and low-quality synthetic speech. The overall performance of the best synthesis system, DECtalk-Paul, was equivalent to natural speech only in terms of performance on initial consonants. The findings are discussed in terms of recent work investigating the perception of synthetic speech under more severe conditions. Suggestions for future research on improving the quality of synthetic speech are also considered.
GREENE, BETH G.; LOGAN, JOHN S.; PISONI, DAVID B.
2012-01-01
We present the results of studies designed to measure the segmental intelligibility of eight text-to-speech systems and a natural speech control, using the Modified Rhyme Test (MRT). Results indicated that the voices tested could be grouped into four categories: natural speech, high-quality synthetic speech, moderate-quality synthetic speech, and low-quality synthetic speech. The overall performance of the best synthesis system, DECtalk-Paul, was equivalent to natural speech only in terms of performance on initial consonants. The findings are discussed in terms of recent work investigating the perception of synthetic speech under more severe conditions. Suggestions for future research on improving the quality of synthetic speech are also considered. PMID:23225916
Speech Perception as a Multimodal Phenomenon
Rosenblum, Lawrence D.
2008-01-01
Speech perception is inherently multimodal. Visual speech (lip-reading) information is used by all perceivers and readily integrates with auditory speech. Imaging research suggests that the brain treats auditory and visual speech similarly. These findings have led some researchers to consider that speech perception works by extracting amodal information that takes the same form across modalities. From this perspective, speech integration is a property of the input information itself. Amodal s...
Cycyk, Lauren M; Bitetti, Dana; Hammer, Carol Scheffner
2015-08-01
This study examined the impact of maternal depressive symptomatology and social support on the English and Spanish language growth of young bilingual children from low-income backgrounds. It was hypothesized that maternal depression would slow children's development in both languages but that social support would buffer the negative effect. Longitudinal data were collected from 83 mothers of Puerto Rican descent and their children who were attending Head Start preschool for 2 years. The effects of maternal depressive symptomatology and social support from family and friends on receptive vocabulary and oral comprehension development in both languages were examined. Growth curve modeling revealed that maternal depressive symptomatology negatively affected Spanish receptive vocabulary development only. Maternal depression did not affect children's English receptive vocabulary or their oral comprehension in either language. Social support was not related to maternal depressive symptomatology or child language. These findings suggest that maternal depression is 1 risk factor that contributes to less robust primary language development of bilingual children from low-income households. Speech-language pathologists must (a) increase their awareness of maternal depression in order to provide families with appropriate mental health referrals and (b) consider their roles as supportive adults for children whose mothers may be depressed.
Speech-specific audiovisual perception affects identification but not detection of speech
DEFF Research Database (Denmark)
Eskelund, Kasper; Andersen, Tobias
Speech perception is audiovisual as evidenced by the McGurk effect in which watching incongruent articulatory mouth movements can change the phonetic auditory speech percept. This type of audiovisual integration may be specific to speech or be applied to all stimuli in general. To investigate...... of audiovisual integration specific to speech perception. However, the results of Tuomainen et al. might have been influenced by another effect. When observers were naïve, they had little motivation to look at the face. When informed, they knew that the face was relevant for the task and this could increase...... visual detection task. In our first experiment, observers presented with congruent and incongruent audiovisual sine-wave speech stimuli did only show a McGurk effect when informed of the speech nature of the stimulus. Performance on the secondary visual task was very good, thus supporting the finding...
Medico-biological factors of speech and language development in young children (part 1
Directory of Open Access Journals (Sweden)
Chernov D.N.
2015-03-01
Full Text Available The article analyzed the main results of medico-biological directions in the study of the factors of children's speech and language. It shows that a variety of pre-, peri-and neonatal developmental factors (teratogenic effects, prematurity, low birth weight, maternal diseases during pregnancy, and chronic diseases of the child had a negative impact on the child-parent relationship that has a lasting influence on child speech and language development.
Rumbach, Anna F; Rose, Tanya A; Cheah, Mynn
2018-01-29
To explore Australian speech-language pathologists' use of non-speech oral motor exercises, and rationales for using/not using non-speech oral motor exercises in clinical practice. A total of 124 speech-language pathologists practising in Australia, working with paediatric and/or adult clients with speech sound difficulties, completed an online survey. The majority of speech-language pathologists reported that they did not use non-speech oral motor exercises when working with paediatric or adult clients with speech sound difficulties. However, more than half of the speech-language pathologists working with adult clients who have dysarthria reported using non-speech oral motor exercises with this population. The most frequently reported rationale for using non-speech oral motor exercises in speech sound difficulty management was to improve awareness/placement of articulators. The majority of speech-language pathologists agreed there is no clear clinical or research evidence base to support non-speech oral motor exercise use with clients who have speech sound difficulties. This study provides an overview of Australian speech-language pathologists' reported use and perceptions of non-speech oral motor exercises' applicability and efficacy in treating paediatric and adult clients who have speech sound difficulties. The research findings provide speech-language pathologists with insight into how and why non-speech oral motor exercises are currently used, and adds to the knowledge base regarding Australian speech-language pathology practice of non-speech oral motor exercises in the treatment of speech sound difficulties. Implications for Rehabilitation Non-speech oral motor exercises refer to oral motor activities which do not involve speech, but involve the manipulation or stimulation of oral structures including the lips, tongue, jaw, and soft palate. Non-speech oral motor exercises are intended to improve the function (e.g., movement, strength) of oral structures. The
2013-08-15
...-Speech Services for Individuals with Hearing and Speech Disabilities, Report and Order (Order), document...] Speech-to-Speech and Internet Protocol (IP) Speech-to-Speech Telecommunications Relay Services; Telecommunications Relay Services and Speech-to-Speech Services for Individuals With Hearing and Speech Disabilities...
[Improving speech comprehension using a new cochlear implant speech processor].
Müller-Deile, J; Kortmann, T; Hoppe, U; Hessel, H; Morsnowski, A
2009-06-01
sentences in the clinical setting S(0)N(CI), with speech signal at 0 degrees and noise lateral to the CI at 90 degrees . With the convincing findings from our evaluations of this multicenter study cohort, a trial with the Freedom 24 sound processor for all suitable CI users is recommended. For evaluating the benefits of a new processor, the comparative assessment paradigm used in our study design would be considered ideal for use with individual patients.
Visual speech information: a help or hindrance in perceptual processing of dysarthric speech.
Borrie, Stephanie A
2015-03-01
This study investigated the influence of visual speech information on perceptual processing of neurologically degraded speech. Fifty listeners identified spastic dysarthric speech under both audio (A) and audiovisual (AV) conditions. Condition comparisons revealed that the addition of visual speech information enhanced processing of the neurologically degraded input in terms of (a) acuity (percent phonemes correct) of vowels and consonants and (b) recognition (percent words correct) of predictive and nonpredictive phrases. Listeners exploited stress-based segmentation strategies more readily in AV conditions, suggesting that the perceptual benefit associated with adding visual speech information to the auditory signal-the AV advantage-has both segmental and suprasegmental origins. Results also revealed that the magnitude of the AV advantage can be predicted, to some degree, by the extent to which an individual utilizes syllabic stress cues to inform word recognition in AV conditions. Findings inform the development of a listener-specific model of speech perception that applies to processing of dysarthric speech in everyday communication contexts.
Suttora, Chiara; Salerni, Nicoletta; Zanchi, Paola; Zampini, Laura; Spinelli, Maria; Fasolo, Mirco
2017-01-01
This study aimed to investigate specific associations between structural and acoustic characteristics of infant-directed (ID) speech and word recognition. Thirty Italian-acquiring children and their mothers were tested when the children were 1;3. Children's word recognition was measured with the looking-while-listening task. Maternal ID speech was…
Use of Deixis in Donald Trump?s Campaign Speech
Hanim, Saidatul
2017-01-01
The aims of this study are (1) to find out the types of deixis in Donald Trump?s campaign speech, (2) to find out the reasons for the use of dominant type of deixis in Donald Trump?s campaign speech and (3) to find out whether or not the deixis is used appropriately in Donald Trump?s campaign speech. This research is conducted by using qualitative content analysis. The data of the study are the utterances from the script Donald Trump?s campaign speech. The data are analyzed by using Levinson ...
Speech-to-Speech Relay Service
Consumer Guide Speech to Speech Relay Service Speech-to-Speech (STS) is one form of Telecommunications Relay Service (TRS). TRS is a service that allows persons with hearing and speech disabilities ...
The speech perception skills of children with and without speech sound disorder.
Hearnshaw, Stephanie; Baker, Elise; Munro, Natalie
To investigate whether Australian-English speaking children with and without speech sound disorder (SSD) differ in their overall speech perception accuracy. Additionally, to investigate differences in the perception of specific phonemes and the association between speech perception and speech production skills. Twenty-five Australian-English speaking children aged 48-60 months participated in this study. The SSD group included 12 children and the typically developing (TD) group included 13 children. Children completed routine speech and language assessments in addition to an experimental Australian-English lexical and phonetic judgement task based on Rvachew's Speech Assessment and Interactive Learning System (SAILS) program (Rvachew, 2009). This task included eight words across four word-initial phonemes-/k, ɹ, ʃ, s/. Children with SSD showed significantly poorer perceptual accuracy on the lexical and phonetic judgement task compared with TD peers. The phonemes /ɹ/ and /s/ were most frequently perceived in error across both groups. Additionally, the phoneme /ɹ/ was most commonly produced in error. There was also a positive correlation between overall speech perception and speech production scores. Children with SSD perceived speech less accurately than their typically developing peers. The findings suggest that an Australian-English variation of a lexical and phonetic judgement task similar to the SAILS program is promising and worthy of a larger scale study. Copyright © 2017 Elsevier Inc. All rights reserved.
Conway, Laura J; Levickis, Penny A; Smith, Jodie; Mensah, Fiona; Wake, Melissa; Reilly, Sheena
2018-03-01
positively associated with 24-, 36- and 48-month language scores in adjusted models (e.g., receptive language at 24 months, ES = 0.21, at 48 months, ES = 0.18). Interaction analysis showed the negative association between successful directives and 24-month receptive language existed primarily in poorly connected dyads with low FC levels. These findings illustrate the effects of the combined interaction between different maternal communicative behaviours and features of the interaction itself on child language development, and the need to consider both in research and practice. Whilst more intrusive directives were associated with poorer language scores, this association attenuated when adjusting for co-occurring responsive expansions, and the association was strongest for children in lower quality interactions. This work may inform clinical practice by helping clinicians target the most appropriate communicative behaviours for specific mother-child dyads. © 2017 Royal College of Speech and Language Therapists.
Mermelshtine, Roni; Barnes, Jacqueline
2016-01-01
Maternal responsive-didactic caregiving (RDC) and infant advanced object play were investigated in a sample of 400 mothers and their 10-month-old infants during video-recorded semi-structured play interactions. Three maternal behaviours: contingent response, cognitively stimulating language and autonomy-promoting speech were coded and infant…
Social eye gaze modulates processing of speech and co-speech gesture.
Holler, Judith; Schubotz, Louise; Kelly, Spencer; Hagoort, Peter; Schuetze, Manuela; Özyürek, Aslı
2014-12-01
In human face-to-face communication, language comprehension is a multi-modal, situated activity. However, little is known about how we combine information from different modalities during comprehension, and how perceived communicative intentions, often signaled through visual signals, influence this process. We explored this question by simulating a multi-party communication context in which a speaker alternated her gaze between two recipients. Participants viewed speech-only or speech+gesture object-related messages when being addressed (direct gaze) or unaddressed (gaze averted to other participant). They were then asked to choose which of two object images matched the speaker's preceding message. Unaddressed recipients responded significantly more slowly than addressees for speech-only utterances. However, perceiving the same speech accompanied by gestures sped unaddressed recipients up to a level identical to that of addressees. That is, when unaddressed recipients' speech processing suffers, gestures can enhance the comprehension of a speaker's message. We discuss our findings with respect to two hypotheses attempting to account for how social eye gaze may modulate multi-modal language comprehension. Copyright © 2014 Elsevier B.V. All rights reserved.
Directory of Open Access Journals (Sweden)
Hwee Ling eLee
2014-08-01
Full Text Available This psychophysics study used musicians as a model to investigate whether musical expertise shapes the temporal integration window for audiovisual speech, sinewave speech or music. Musicians and non-musicians judged the audiovisual synchrony of speech, sinewave analogues of speech, and music stimuli at 13 audiovisual stimulus onset asynchronies (±360, ±300 ±240, ±180, ±120, ±60, and 0 ms. Further, we manipulated the duration of the stimuli by presenting sentences/melodies or syllables/tones. Critically, musicians relative to non-musicians exhibited significantly narrower temporal integration windows for both music and sinewave speech. Further, the temporal integration window for music decreased with the amount of music practice, but not with age of acquisition. In other words, the more musicians practiced piano in the past three years, the more sensitive they became to the temporal misalignment of visual and auditory signals. Collectively, our findings demonstrate that music practicing fine-tunes the audiovisual temporal integration window to various extents depending on the stimulus class. While the effect of piano practicing was most pronounced for music, it also generalized to other stimulus classes such as sinewave speech and to a marginally significant degree to natural speech.
Knollhoff, Stephanie; Barohn, Richard J.
2017-01-01
Purpose This preliminary study on lingual–alveolar contact pressures (LACP) in people with amyotrophic lateral sclerosis (ALS) had several aims: (a) to evaluate whether the protocol induced fatigue, (b) to compare LACP during speech (LACP-Sp) and during maximum isometric pressing (LACP-Max) in people with ALS (PALS) versus healthy controls, (c) to compare the percentage of LACP-Max utilized during speech (%Max) for PALS versus controls, and (d) to evaluate relationships between LACP-Sp and LACP-Max with word intelligibility. Method Thirteen PALS and 12 healthy volunteers produced /t, d, s, z, l, n/ sounds while LACP-Sp was recorded. LACP-Max was obtained before and after the speech protocol. Word intelligibility was obtained from auditory–perceptual judgments. Results LACP-Max values measured before and after completion of the speech protocol did not differ. LACP-Sp and LACP-Max were statistically lower in the ALS bulbar group compared with controls and PALS with only spinal symptoms. There was no statistical difference between groups for %Max. LACP-Sp and LACP-Max were correlated with word intelligibility. Conclusions It was feasible to obtain LACP-Sp measures without inducing fatigue. Reductions in LACP-Sp and LACP-Max for bulbar speakers might reflect tongue weakness. Although confirmation of results is needed, the data indicate that individuals with high word intelligibility maintained LACP-Sp at or above 2 kPa and LACP-Max at or above 50 kPa. PMID:28335033
Mermelshtine, Roni; Barnes, Jacqueline
2016-01-01
Maternal responsive-didactic caregiving (RDC) and infant advanced object play were investigated in a sample of 400 mothers and their 10-month-old infants during video-recorded semi-structured play interactions. Three maternal behaviours: contingent response, cognitively stimulating language and autonomy promoting speech were coded and infant object play. Factor analysis confirmed the three maternal behaviours loaded onto one underlying factor, labelled RDC. Based on ecological and transaction...
Schall, Sonja; von Kriegstein, Katharina
2014-01-01
It has been proposed that internal simulation of the talking face of visually-known speakers facilitates auditory speech recognition. One prediction of this view is that brain areas involved in auditory-only speech comprehension interact with visual face-movement sensitive areas, even under auditory-only listening conditions. Here, we test this hypothesis using connectivity analyses of functional magnetic resonance imaging (fMRI) data. Participants (17 normal participants, 17 developmental prosopagnosics) first learned six speakers via brief voice-face or voice-occupation training (comprehension. Overall, the present findings indicate that learned visual information is integrated into the analysis of auditory-only speech and that this integration results from the interaction of task-relevant face-movement and auditory speech-sensitive areas.
Maternal correlates of maternal child feeding practices: a systematic review.
McPhie, Skye; Skouteris, Helen; Daniels, Lynne; Jansen, Elena
2014-01-01
Establishing healthy eating habits early in life is one important strategy to combat childhood obesity. Given that early maternal child feeding practices have been linked to child food intake and weight, identifying the maternal correlates of maternal child feeding practices is important in order to understand the determinants of childhood obesity; this was the overall aim of the current review. Academic databases were searched for studies examining the relationship between maternal child feeding practices and parenting, personal characteristics and psychopathology of mothers with preschoolers. Papers were limited to those published in English, between January 2000 and June 2012. Only studies with mothers of normally developing children between the ages of 2 and 6 years were included. There were no restrictions regarding the inclusion of maternal nationality or socioeconomic status (SES). Seventeen eligible studies were sourced. Information on the aim, sample, measures and findings of these was summarised into tables. The findings of this review support a relationship between maternal controlling parenting, general and eating psychopathology, and SES and maternal child feeding practices. The main methodological issues of the studies reviewed included inconsistency in measures of maternal variables across studies and cross-sectional designs. We conclude that the maternal correlates associated with maternal child feeding practices are complex, and the pathways by which maternal correlates impact these feeding practices require further investigation. © 2012 John Wiley & Sons Ltd.
An analysis of machine translation and speech synthesis in speech-to-speech translation system
Hashimoto, K.; Yamagishi, J.; Byrne, W.; King, S.; Tokuda, K.
2011-01-01
This paper provides an analysis of the impacts of machine translation and speech synthesis on speech-to-speech translation systems. The speech-to-speech translation system consists of three components: speech recognition, machine translation and speech synthesis. Many techniques for integration of speech recognition and machine translation have been proposed. However, speech synthesis has not yet been considered. Therefore, in this paper, we focus on machine translation and speech synthesis, ...
Understanding the Linguistic Characteristics of the Great Speeches
Mouritzen, Kristian
2016-01-01
This dissertation attempts to find the common traits of great speeches. It does so by closely examining the language of some of the most well-known speeches in world. These speeches are presented in the book Speeches that Changed the World (2006) by Simon Sebag Montefiore. The dissertation specifically looks at four variables: The beginnings and endings of the speeches, the use of passive voice, the use of personal pronouns and the difficulty of the language. These four variables are based on...
Speech Adaptation to Kinematic Recording Sensors: Perceptual and Acoustic Findings
Dromey, Christopher; Hunter, Elise; Nissen, Shawn L.
2018-01-01
Purpose: This study used perceptual and acoustic measures to examine the time course of speech adaptation after the attachment of electromagnetic sensor coils to the tongue, lips, and jaw. Method: Twenty native English speakers read aloud stimulus sentences before the attachment of the sensors, immediately after attachment, and again 5, 10, 15,…
Maternal employment, breastfeeding, and health: evidence from maternity leave mandates.
Baker, Michael; Milligan, Kevin
2008-07-01
Public health agencies around the world have renewed efforts to increase the incidence and duration of breastfeeding. Maternity leave mandates present an economic policy that could help achieve these goals. We study their efficacy, focusing on a significant increase in maternity leave mandates in Canada. We find very large increases in mothers' time away from work post-birth and in the attainment of critical breastfeeding duration thresholds. We also look for impacts of the reform on self-reported indicators of maternal and child health captured in our data. For most indicators we find no effect.
DEFF Research Database (Denmark)
Niebuhr, Oliver; Brem, Alexander; Novák-Tót, Eszter
2016-01-01
to business speeches. Consistent with the public opinion, our findings are indicative of Steve Jobs being a more charismatic speaker than Mark Zuckerberg. Beyond previous studies, our data suggest that rhythm and emphatic accentuation are also involved in conveying charisma. Furthermore, the differences...... between Steve Jobs and Mark Zuckerberg and the investor- and customer-related sections of their speeches support the modern understanding of charisma as a gradual, multiparametric, and context-sensitive concept....
Speech Acquisition and Automatic Speech Recognition for Integrated Spacesuit Audio Systems
Huang, Yiteng; Chen, Jingdong; Chen, Shaoyan
2010-01-01
A voice-command human-machine interface system has been developed for spacesuit extravehicular activity (EVA) missions. A multichannel acoustic signal processing method has been created for distant speech acquisition in noisy and reverberant environments. This technology reduces noise by exploiting differences in the statistical nature of signal (i.e., speech) and noise that exists in the spatial and temporal domains. As a result, the automatic speech recognition (ASR) accuracy can be improved to the level at which crewmembers would find the speech interface useful. The developed speech human/machine interface will enable both crewmember usability and operational efficiency. It can enjoy a fast rate of data/text entry, small overall size, and can be lightweight. In addition, this design will free the hands and eyes of a suited crewmember. The system components and steps include beam forming/multi-channel noise reduction, single-channel noise reduction, speech feature extraction, feature transformation and normalization, feature compression, model adaption, ASR HMM (Hidden Markov Model) training, and ASR decoding. A state-of-the-art phoneme recognizer can obtain an accuracy rate of 65 percent when the training and testing data are free of noise. When it is used in spacesuits, the rate drops to about 33 percent. With the developed microphone array speech-processing technologies, the performance is improved and the phoneme recognition accuracy rate rises to 44 percent. The recognizer can be further improved by combining the microphone array and HMM model adaptation techniques and using speech samples collected from inside spacesuits. In addition, arithmetic complexity models for the major HMMbased ASR components were developed. They can help real-time ASR system designers select proper tasks when in the face of constraints in computational resources.
Directory of Open Access Journals (Sweden)
Katharine Noonan
2018-04-01
Full Text Available The association between low family income and socio-emotional behaviour problems in early childhood has been well-documented, and maternal psychological distress is highlighted as central in mediating this relationship. However, whether this relationship holds for older children, and the precise mechanisms by which income may influence child behaviour is unclear.This study investigated the relationship between family income and child socio-emotional behaviour at 11 years of age, and examined the mediating role of maternal psychological distress over time using the UK Millennium Cohort Study.The primary outcome was parent-reported behavioural problems, as captured by the Total Difficulties Score (TDS, derived from the Strengths and Difficulties Questionnaire (SDQ. Secondary outcomes were the emotional, peer-related, conduct, and hyperactivity/inattention problems subscales of the SDQ; and teacher-reported TDS. Permanent family income was the primary exposure variable; frequency of poverty up to age 11 years was the secondary exposure variable. Maternal psychological distress was operationalised to reflect the trajectory from child birth to age 11. Multivariable logistic regression models were used to estimate the effect of permanent family income on child behaviour at age 11, controlling for maternal psychological distress and other relevant covariates.Results showed a statistically significant protective effect of increased permanent family income on the likelihood of behavioural problems at age 11. This finding was consistent for all SDQ subscales apart from emotional problems, and was strongest for teacher-reported behavioural problems. Maternal distress was an important mediator in the income-child behaviour relationship for parent-reported, but not teacher-reported, behavioural problems.The results of this study strengthen empirical evidence that the child behaviour-income gradient is maintained in older childhood. Mother’s psychological
Sellers, Ruth; Harold, Gordon T; Elam, Kit; Rhoades, Kimberly A; Potter, Robert; Mars, Becky; Craddock, Nick; Thapar, Anita; Collishaw, Stephan
2014-01-01
Disruption in the parent-child relationship is a commonly hypothesized risk factor through which maternal depression may increase risk for offspring psychopathology. However, maternal depression is commonly accompanied by other psychopathology, including antisocial behaviour. Few studies have examined the role of co-occurring psychopathology in depressed mothers. Using a longitudinal study of offspring of mothers with recurrent depression, we aimed to test whether maternal warmth/hostility mediated links between maternal depression severity and child outcomes, and how far direct and indirect pathways were robust to controls for co-occurring maternal antisocial behaviour. Mothers with a history of recurrent major depressive disorder and their adolescent offspring (9-17 years at baseline) were assessed three times between 2007 and 2010. Mothers completed questionnaires assessing their own depression severity and antisocial behaviour at Time 1 (T1). The parent-child relationship was assessed using parent-rated questionnaire and interviewer-rated 5-min speech sample at Time 2 (T2). Offspring symptoms of depression and disruptive behaviours were assessed using the Child and Adolescent Psychiatric Assessment at Time 3 (T3). Maternal hostility and warmth, respectively, mediated the association between maternal depression severity and risk for offspring psychopathology. However, the effects were attenuated when maternal antisocial behaviour was included in the analysis. In tests of the full theoretical model, maternal antisocial behaviour predicted both maternal hostility and low warmth, maternal hostility predicted offspring disruptive behaviour disorder symptoms, but not depression, and maternal warmth was not associated with either child outcome. Parenting interventions aimed at reducing hostility may be beneficial for preventing or reducing adolescent disruptive behaviours in offspring of depressed mothers, especially when depressed mothers report co
Neural Entrainment to Speech Modulates Speech Intelligibility
Riecke, Lars; Formisano, Elia; Sorger, Bettina; Baskent, Deniz; Gaudrain, Etienne
2018-01-01
Speech is crucial for communication in everyday life. Speech-brain entrainment, the alignment of neural activity to the slow temporal fluctuations (envelope) of acoustic speech input, is a ubiquitous element of current theories of speech processing. Associations between speech-brain entrainment and
Castellanos, Irina; Kronenberger, William G; Beer, Jessica; Henning, Shirley C; Colson, Bethany G; Pisoni, David B
2014-07-01
Speech and language measures during grade school predict adolescent speech-language outcomes in children who receive cochlear implants (CIs), but no research has examined whether speech and language functioning at even younger ages is predictive of long-term outcomes in this population. The purpose of this study was to examine whether early preschool measures of speech and language performance predict speech-language functioning in long-term users of CIs. Early measures of speech intelligibility and receptive vocabulary (obtained during preschool ages of 3-6 years) in a sample of 35 prelingually deaf, early-implanted children predicted speech perception, language, and verbal working memory skills up to 18 years later. Age of onset of deafness and age at implantation added additional variance to preschool speech intelligibility in predicting some long-term outcome scores, but the relationship between preschool speech-language skills and later speech-language outcomes was not significantly attenuated by the addition of these hearing history variables. These findings suggest that speech and language development during the preschool years is predictive of long-term speech and language functioning in early-implanted, prelingually deaf children. As a result, measures of speech-language functioning at preschool ages can be used to identify and adjust interventions for very young CI users who may be at long-term risk for suboptimal speech and language outcomes.
Noonan, Katharine; Burns, Richéal; Violato, Mara
2018-04-01
The association between low family income and socio-emotional behaviour problems in early childhood has been well-documented, and maternal psychological distress is highlighted as central in mediating this relationship. However, whether this relationship holds for older children, and the precise mechanisms by which income may influence child behaviour is unclear. This study investigated the relationship between family income and child socio-emotional behaviour at 11 years of age, and examined the mediating role of maternal psychological distress over time using the UK Millennium Cohort Study. The primary outcome was parent-reported behavioural problems, as captured by the Total Difficulties Score (TDS), derived from the Strengths and Difficulties Questionnaire (SDQ). Secondary outcomes were the emotional, peer-related, conduct, and hyperactivity/inattention problems subscales of the SDQ; and teacher-reported TDS. Permanent family income was the primary exposure variable; frequency of poverty up to age 11 years was the secondary exposure variable. Maternal psychological distress was operationalised to reflect the trajectory from child birth to age 11. Multivariable logistic regression models were used to estimate the effect of permanent family income on child behaviour at age 11, controlling for maternal psychological distress and other relevant covariates. Results showed a statistically significant protective effect of increased permanent family income on the likelihood of behavioural problems at age 11. This finding was consistent for all SDQ subscales apart from emotional problems, and was strongest for teacher-reported behavioural problems. Maternal distress was an important mediator in the income-child behaviour relationship for parent-reported, but not teacher-reported, behavioural problems. The results of this study strengthen empirical evidence that the child behaviour-income gradient is maintained in older childhood. Mother's psychological distress
Acquired apraxia of speech: features, accounts, and treatment.
Peach, Richard K
2004-01-01
The features of apraxia of speech (AOS) are presented with regard to both traditional and contemporary descriptions of the disorder. Models of speech processing, including the neurological bases for apraxia of speech, are discussed. Recent findings concerning subcortical contributions to apraxia of speech and the role of the insula are presented. The key features to differentially diagnose AOS from related speech syndromes are identified. Treatment implications derived from motor accounts of AOS are presented along with a summary of current approaches designed to treat the various subcomponents of the disorder. Finally, guidelines are provided for treating the AOS patient with coexisting aphasia.
Eadie, Patricia; Morgan, Angela; Ukoumunne, Obioha C; Ttofari Eecen, Kyriaki; Wake, Melissa; Reilly, Sheena
2015-06-01
The epidemiology of preschool speech sound disorder is poorly understood. Our aims were to determine: the prevalence of idiopathic speech sound disorder; the comorbidity of speech sound disorder with language and pre-literacy difficulties; and the factors contributing to speech outcome at 4 years. One thousand four hundred and ninety-four participants from an Australian longitudinal cohort completed speech, language, and pre-literacy assessments at 4 years. Prevalence of speech sound disorder (SSD) was defined by standard score performance of ≤79 on a speech assessment. Logistic regression examined predictors of SSD within four domains: child and family; parent-reported speech; cognitive-linguistic; and parent-reported motor skills. At 4 years the prevalence of speech disorder in an Australian cohort was 3.4%. Comorbidity with SSD was 40.8% for language disorder and 20.8% for poor pre-literacy skills. Sex, maternal vocabulary, socio-economic status, and family history of speech and language difficulties predicted SSD, as did 2-year speech, language, and motor skills. Together these variables provided good discrimination of SSD (area under the curve=0.78). This is the first epidemiological study to demonstrate prevalence of SSD at 4 years of age that was consistent with previous clinical studies. Early detection of SSD at 4 years should focus on family variables and speech, language, and motor skills measured at 2 years. © 2014 Mac Keith Press.
ACOUSTIC SPEECH RECOGNITION FOR MARATHI LANGUAGE USING SPHINX
Directory of Open Access Journals (Sweden)
Aman Ankit
2016-09-01
Full Text Available Speech recognition or speech to text processing, is a process of recognizing human speech by the computer and converting into text. In speech recognition, transcripts are created by taking recordings of speech as audio and their text transcriptions. Speech based applications which include Natural Language Processing (NLP techniques are popular and an active area of research. Input to such applications is in natural language and output is obtained in natural language. Speech recognition mostly revolves around three approaches namely Acoustic phonetic approach, Pattern recognition approach and Artificial intelligence approach. Creation of acoustic model requires a large database of speech and training algorithms. The output of an ASR system is recognition and translation of spoken language into text by computers and computerized devices. ASR today finds enormous application in tasks that require human machine interfaces like, voice dialing, and etc. Our key contribution in this paper is to create corpora for Marathi language and explore the use of Sphinx engine for automatic speech recognition
Reliance on auditory feedback in children with childhood apraxia of speech.
Iuzzini-Seigel, Jenya; Hogan, Tiffany P; Guarino, Anthony J; Green, Jordan R
2015-01-01
Children with childhood apraxia of speech (CAS) have been hypothesized to continuously monitor their speech through auditory feedback to minimize speech errors. We used an auditory masking paradigm to determine the effect of attenuating auditory feedback on speech in 30 children: 9 with CAS, 10 with speech delay, and 11 with typical development. The masking only affected the speech of children with CAS as measured by voice onset time and vowel space area. These findings provide preliminary support for greater reliance on auditory feedback among children with CAS. Readers of this article should be able to (i) describe the motivation for investigating the role of auditory feedback in children with CAS; (ii) report the effects of feedback attenuation on speech production in children with CAS, speech delay, and typical development, and (iii) understand how the current findings may support a feedforward program deficit in children with CAS. Copyright © 2015 The Authors. Published by Elsevier Inc. All rights reserved.
Sapir, S; Canter, G J
1991-09-01
Using acoustic analysis techniques, Waldstein [J. Acoust. Soc. Am. 88, 2099-2114 (1990] reported abnormal speech findings in postlingual deaf speakers. She interpreted her findings to suggest that auditory feedback is important in motor speech control. However, it is argued here that Waldstein's interpretation may be unwarranted without addressing the possibility of neurologic deficits (e.g., dysarthria) as confounding (or even primary) causes of the abnormal speech in her subjects.
Directory of Open Access Journals (Sweden)
Farshid Tayari Ashtiani
2015-02-01
Full Text Available The present study was an attempt to investigate the impact of English verbal songs on connected speech aspects of adult English learners’ speech production. 40 participants were selected based on the results of their performance in a piloted and validated version of NELSON test given to 60 intermediate English learners in a language institute in Tehran. Then they were equally distributed in two control and experimental groups and received a validated pretest of reading aloud and speaking in English. Afterward, the treatment was performed in 18 sessions by singing preselected songs culled based on some criteria such as popularity, familiarity, amount, and speed of speech delivery, etc. In the end, the posttests of reading aloud and speaking in English were administered. The results revealed that the treatment had statistically positive effects on the connected speech aspects of English learners’ speech production at statistical .05 level of significance. Meanwhile, the results represented that there was not any significant difference between the experimental group’s mean scores on the posttests of reading aloud and speaking. It was thus concluded that providing the EFL learners with English verbal songs could positively affect connected speech aspects of both modes of speech production, reading aloud and speaking. The Findings of this study have pedagogical implications for language teachers to be more aware and knowledgeable of the benefits of verbal songs to promote speech production of language learners in terms of naturalness and fluency. Keywords: English Verbal Songs, Connected Speech, Speech Production, Reading Aloud, Speaking
Multistage audiovisual integration of speech: dissociating identification and detection
DEFF Research Database (Denmark)
Eskelund, Kasper; Tuomainen, Jyrki; Andersen, Tobias
2011-01-01
Speech perception integrates auditory and visual information. This is evidenced by the McGurk illusion where seeing the talking face influences the auditory phonetic percept and by the audiovisual detection advantage where seeing the talking face influences the detectability of the acoustic speech...... signal. Here we show that identification of phonetic content and detection can be dissociated as speech-specific and non-specific audiovisual integration effects. To this end, we employed synthetically modified stimuli, sine wave speech (SWS), which is an impoverished speech signal that only observers...... informed of its speech-like nature recognize as speech. While the McGurk illusion only occurred for informed observers the audiovisual detection advantage occurred for naïve observers as well. This finding supports a multi-stage account of audiovisual integration of speech in which the many attributes...
Inner Speech's Relationship With Overt Speech in Poststroke Aphasia.
Stark, Brielle C; Geva, Sharon; Warburton, Elizabeth A
2017-09-18
Relatively preserved inner speech alongside poor overt speech has been documented in some persons with aphasia (PWA), but the relationship of overt speech with inner speech is still largely unclear, as few studies have directly investigated these factors. The present study investigates the relationship of relatively preserved inner speech in aphasia with selected measures of language and cognition. Thirty-eight persons with chronic aphasia (27 men, 11 women; average age 64.53 ± 13.29 years, time since stroke 8-111 months) were classified as having relatively preserved inner and overt speech (n = 21), relatively preserved inner speech with poor overt speech (n = 8), or not classified due to insufficient measurements of inner and/or overt speech (n = 9). Inner speech scores (by group) were correlated with selected measures of language and cognition from the Comprehensive Aphasia Test (Swinburn, Porter, & Al, 2004). The group with poor overt speech showed a significant relationship of inner speech with overt naming (r = .95, p speech and language and cognition factors were not significant for the group with relatively good overt speech. As in previous research, we show that relatively preserved inner speech is found alongside otherwise severe production deficits in PWA. PWA with poor overt speech may rely more on preserved inner speech for overt picture naming (perhaps due to shared resources with verbal working memory) and for written picture description (perhaps due to reliance on inner speech due to perceived task difficulty). Assessments of inner speech may be useful as a standard component of aphasia screening, and therapy focused on improving and using inner speech may prove clinically worthwhile. https://doi.org/10.23641/asha.5303542.
Perception of Intersensory Synchrony in Audiovisual Speech: Not that Special
Vroomen, Jean; Stekelenburg, Jeroen J.
2011-01-01
Perception of intersensory temporal order is particularly difficult for (continuous) audiovisual speech, as perceivers may find it difficult to notice substantial timing differences between speech sounds and lip movements. Here we tested whether this occurs because audiovisual speech is strongly paired ("unity assumption"). Participants made…
Sixteen-Month-Old Infants' Segment Words from Infant- and Adult-Directed Speech
Mani, Nivedita; Pätzold, Wiebke
2016-01-01
One of the first challenges facing the young language learner is the task of segmenting words from a natural language speech stream, without prior knowledge of how these words sound. Studies with younger children find that children find it easier to segment words from fluent speech when the words are presented in infant-directed speech, i.e., the…
Mobile speech and advanced natural language solutions
Markowitz, Judith
2013-01-01
Mobile Speech and Advanced Natural Language Solutions provides a comprehensive and forward-looking treatment of natural speech in the mobile environment. This fourteen-chapter anthology brings together lead scientists from Apple, Google, IBM, AT&T, Yahoo! Research and other companies, along with academicians, technology developers and market analysts. They analyze the growing markets for mobile speech, new methodological approaches to the study of natural language, empirical research findings on natural language and mobility, and future trends in mobile speech. Mobile Speech opens with a challenge to the industry to broaden the discussion about speech in mobile environments beyond the smartphone, to consider natural language applications across different domains. Among the new natural language methods introduced in this book are Sequence Package Analysis, which locates and extracts valuable opinion-related data buried in online postings; microintonation as a way to make TTS truly human-like; and se...
Tuning Neural Phase Entrainment to Speech.
Falk, Simone; Lanzilotti, Cosima; Schön, Daniele
2017-08-01
Musical rhythm positively impacts on subsequent speech processing. However, the neural mechanisms underlying this phenomenon are so far unclear. We investigated whether carryover effects from a preceding musical cue to a speech stimulus result from a continuation of neural phase entrainment to periodicities that are present in both music and speech. Participants listened and memorized French metrical sentences that contained (quasi-)periodic recurrences of accents and syllables. Speech stimuli were preceded by a rhythmically regular or irregular musical cue. Our results show that the presence of a regular cue modulates neural response as estimated by EEG power spectral density, intertrial coherence, and source analyses at critical frequencies during speech processing compared with the irregular condition. Importantly, intertrial coherences for regular cues were indicative of the participants' success in memorizing the subsequent speech stimuli. These findings underscore the highly adaptive nature of neural phase entrainment across fundamentally different auditory stimuli. They also support current models of neural phase entrainment as a tool of predictive timing and attentional selection across cognitive domains.
Impairments of speech fluency in Lewy body spectrum disorder.
Ash, Sharon; McMillan, Corey; Gross, Rachel G; Cook, Philip; Gunawardena, Delani; Morgan, Brianna; Boller, Ashley; Siderowf, Andrew; Grossman, Murray
2012-03-01
Few studies have examined connected speech in demented and non-demented patients with Parkinson's disease (PD). We assessed the speech production of 35 patients with Lewy body spectrum disorder (LBSD), including non-demented PD patients, patients with PD dementia (PDD), and patients with dementia with Lewy bodies (DLB), in a semi-structured narrative speech sample in order to characterize impairments of speech fluency and to determine the factors contributing to reduced speech fluency in these patients. Both demented and non-demented PD patients exhibited reduced speech fluency, characterized by reduced overall speech rate and long pauses between sentences. Reduced speech rate in LBSD correlated with measures of between-utterance pauses, executive functioning, and grammatical comprehension. Regression analyses related non-fluent speech, grammatical difficulty, and executive difficulty to atrophy in frontal brain regions. These findings indicate that multiple factors contribute to slowed speech in LBSD, and this is mediated in part by disease in frontal brain regions. Copyright © 2011 Elsevier Inc. All rights reserved.
Use of digital speech recognition in diagnostics radiology
International Nuclear Information System (INIS)
Arndt, H.; Stockheim, D.; Mutze, S.; Petersein, J.; Gregor, P.; Hamm, B.
1999-01-01
Purpose: Applicability and benefits of digital speech recognition in diagnostic radiology were tested using the speech recognition system SP 6000. Methods: The speech recognition system SP 6000 was integrated into the network of the institute and connected to the existing Radiological Information System (RIS). Three subjects used this system for writing 2305 findings from dictation. After the recognition process the date, length of dictation, time required for checking/correction, kind of examination and error rate were recorded for every dictation. With the same subjects, a correlation was performed with 625 conventionally written finding. Results: After an 1-hour initial training the average error rates were 8.4 to 13.3%. The first adaptation of the speech recognition system (after nine days) decreased the average error rates to 2.4 to 10.7% due to the ability of the program to learn. The 2 nd and 3 rd adaptations resulted only in small changes of the error rate. An individual comparison of the error rate developments in the same kind of investigation showed the relative independence of the error rate on the individual user. Conclusion: The results show that the speech recognition system SP 6000 can be evaluated as an advantageous alternative for quickly recording radiological findings. A comparison between manually writing and dictating the findings verifies the individual differences of the writing speeds and shows the advantage of the application of voice recognition when faced with normal keyboard performance. (orig.) [de
Censored: Whistleblowers and impossible speech
Kenny, Kate
2017-01-01
What happens to a person who speaks out about corruption in their organization, and finds themselves excluded from their profession? In this article, I argue that whistleblowers experience exclusions because they have engaged in ‘impossible speech’, that is, a speech act considered to be unacceptable or illegitimate. Drawing on Butler’s theories of recognition and censorship, I show how norms of acceptable speech working through recruitment practices, alongside the actions of colleagues, can ...
Musician advantage for speech-on-speech perception
Başkent, Deniz; Gaudrain, Etienne
Evidence for transfer of musical training to better perception of speech in noise has been mixed. Unlike speech-in-noise, speech-on-speech perception utilizes many of the skills that musical training improves, such as better pitch perception and stream segregation, as well as use of higher-level
Multistage audiovisual integration of speech: dissociating identification and detection.
Eskelund, Kasper; Tuomainen, Jyrki; Andersen, Tobias S
2011-02-01
Speech perception integrates auditory and visual information. This is evidenced by the McGurk illusion where seeing the talking face influences the auditory phonetic percept and by the audiovisual detection advantage where seeing the talking face influences the detectability of the acoustic speech signal. Here, we show that identification of phonetic content and detection can be dissociated as speech-specific and non-specific audiovisual integration effects. To this end, we employed synthetically modified stimuli, sine wave speech (SWS), which is an impoverished speech signal that only observers informed of its speech-like nature recognize as speech. While the McGurk illusion only occurred for informed observers, the audiovisual detection advantage occurred for naïve observers as well. This finding supports a multistage account of audiovisual integration of speech in which the many attributes of the audiovisual speech signal are integrated by separate integration processes.
Ekström, Seth-Reino; Borg, Erik
2011-01-01
The masking effect of a piano composition, played at different speeds and in different octaves, on speech-perception thresholds was investigated in 15 normal-hearing and 14 moderately-hearing-impaired subjects. Running speech (just follow conversation, JFC) testing and use of hearing aids increased the everyday validity of the findings. A comparison was made with standard audiometric noises [International Collegium of Rehabilitative Audiology (ICRA) noise and speech spectrum-filtered noise (SPN)]. All masking sounds, music or noise, were presented at the same equivalent sound level (50 dBA). The results showed a significant effect of piano performance speed and octave (Ptempo had the largest effect; and high octave and slow tempo, the smallest. Music had a lower masking effect than did ICRA noise with two or six speakers at normal vocal effort (Pmusic offers an interesting opportunity for studying masking under realistic conditions, where spectral and temporal features can be varied independently. The results have implications for composing music with vocal parts, designing acoustic environments and creating a balance between speech perception and privacy in social settings.
Directory of Open Access Journals (Sweden)
J. SANGEETHA
2015-02-01
Full Text Available This paper provides an interface between the machine translation and speech synthesis system for converting English speech to Tamil text in English to Tamil speech to speech translation system. The speech translation system consists of three modules: automatic speech recognition, machine translation and text to speech synthesis. Many procedures for incorporation of speech recognition and machine translation have been projected. Still speech synthesis system has not yet been measured. In this paper, we focus on integration of machine translation and speech synthesis, and report a subjective evaluation to investigate the impact of speech synthesis, machine translation and the integration of machine translation and speech synthesis components. Here we implement a hybrid machine translation (combination of rule based and statistical machine translation and concatenative syllable based speech synthesis technique. In order to retain the naturalness and intelligibility of synthesized speech Auto Associative Neural Network (AANN prosody prediction is used in this work. The results of this system investigation demonstrate that the naturalness and intelligibility of the synthesized speech are strongly influenced by the fluency and correctness of the translated text.
Directory of Open Access Journals (Sweden)
Yong Chon Park
2018-02-01
Full Text Available Our study aimed to assess the psychometric validity of the conceptual disorganization item and other items of the Brief Psychiatric Rating Scale (BPRS for detecting disorganized speech in patients with schizophrenia. We included 357 schizophrenia patients with disorganized speech and 1082 without disorganized speech from the survey centers in India, Indonesia, Japan, Malaysia, and Taiwan, using the data from the Research on Asian Psychotropic Patterns for Antipsychotics (REAP-AP study. After adjusting the effects of confounding variables, a binary logistic regression model was fitted to identify BPRS items independently associated with disorganized speech. Receiver operating characteristic (ROC curves were used to identify optimum cut-off scores and their sensitivities and specificities for detecting disorganized speech. After adjusting the effects of confounding variables, the fitted binary logistic regression model indicated that conceptual disorganization (P < 0.0001, uncooperativeness (P = 0.010 and excitement (P = 0.001 were independently associated with disorganized speech. The ROC curve revealed that the conceptual disorganization item could accurately detect disorganized speech in patients with schizophrenia both separately and in combination with uncooperativeness and excitement. The subscale for conceptual disorganization, uncooperativeness and excitement items in the BPRS is a promising psychometric tool for detecting disorganized speech.
[Prosody, speech input and language acquisition].
Jungheim, M; Miller, S; Kühn, D; Ptok, M
2014-04-01
In order to acquire language, children require speech input. The prosody of the speech input plays an important role. In most cultures adults modify their code when communicating with children. Compared to normal speech this code differs especially with regard to prosody. For this review a selective literature search in PubMed and Scopus was performed. Prosodic characteristics are a key feature of spoken language. By analysing prosodic features, children gain knowledge about underlying grammatical structures. Child-directed speech (CDS) is modified in a way that meaningful sequences are highlighted acoustically so that important information can be extracted from the continuous speech flow more easily. CDS is said to enhance the representation of linguistic signs. Taking into consideration what has previously been described in the literature regarding the perception of suprasegmentals, CDS seems to be able to support language acquisition due to the correspondence of prosodic and syntactic units. However, no findings have been reported, stating that the linguistically reduced CDS could hinder first language acquisition.
Directory of Open Access Journals (Sweden)
Alena Galilee
Full Text Available Previous event-related potential (ERP research utilizing oddball stimulus paradigms suggests diminished processing of speech versus non-speech sounds in children with an Autism Spectrum Disorder (ASD. However, brain mechanisms underlying these speech processing abnormalities, and to what extent they are related to poor language abilities in this population remain unknown. In the current study, we utilized a novel paired repetition paradigm in order to investigate ERP responses associated with the detection and discrimination of speech and non-speech sounds in 4- to 6-year old children with ASD, compared with gender and verbal age matched controls. ERPs were recorded while children passively listened to pairs of stimuli that were either both speech sounds, both non-speech sounds, speech followed by non-speech, or non-speech followed by speech. Control participants exhibited N330 match/mismatch responses measured from temporal electrodes, reflecting speech versus non-speech detection, bilaterally, whereas children with ASD exhibited this effect only over temporal electrodes in the left hemisphere. Furthermore, while the control groups exhibited match/mismatch effects at approximately 600 ms (central N600, temporal P600 when a non-speech sound was followed by a speech sound, these effects were absent in the ASD group. These findings suggest that children with ASD fail to activate right hemisphere mechanisms, likely associated with social or emotional aspects of speech detection, when distinguishing non-speech from speech stimuli. Together, these results demonstrate the presence of atypical speech versus non-speech processing in children with ASD when compared with typically developing children matched on verbal age.
Galilee, Alena; Stefanidou, Chrysi; McCleery, Joseph P
2017-01-01
Previous event-related potential (ERP) research utilizing oddball stimulus paradigms suggests diminished processing of speech versus non-speech sounds in children with an Autism Spectrum Disorder (ASD). However, brain mechanisms underlying these speech processing abnormalities, and to what extent they are related to poor language abilities in this population remain unknown. In the current study, we utilized a novel paired repetition paradigm in order to investigate ERP responses associated with the detection and discrimination of speech and non-speech sounds in 4- to 6-year old children with ASD, compared with gender and verbal age matched controls. ERPs were recorded while children passively listened to pairs of stimuli that were either both speech sounds, both non-speech sounds, speech followed by non-speech, or non-speech followed by speech. Control participants exhibited N330 match/mismatch responses measured from temporal electrodes, reflecting speech versus non-speech detection, bilaterally, whereas children with ASD exhibited this effect only over temporal electrodes in the left hemisphere. Furthermore, while the control groups exhibited match/mismatch effects at approximately 600 ms (central N600, temporal P600) when a non-speech sound was followed by a speech sound, these effects were absent in the ASD group. These findings suggest that children with ASD fail to activate right hemisphere mechanisms, likely associated with social or emotional aspects of speech detection, when distinguishing non-speech from speech stimuli. Together, these results demonstrate the presence of atypical speech versus non-speech processing in children with ASD when compared with typically developing children matched on verbal age.
Speech-Language Dissociations, Distractibility, and Childhood Stuttering
Conture, Edward G.; Walden, Tedra A.; Lambert, Warren E.
2015-01-01
Purpose This study investigated the relation among speech-language dissociations, attentional distractibility, and childhood stuttering. Method Participants were 82 preschool-age children who stutter (CWS) and 120 who do not stutter (CWNS). Correlation-based statistics (Bates, Appelbaum, Salcedo, Saygin, & Pizzamiglio, 2003) identified dissociations across 5 norm-based speech-language subtests. The Behavioral Style Questionnaire Distractibility subscale measured attentional distractibility. Analyses addressed (a) between-groups differences in the number of children exhibiting speech-language dissociations; (b) between-groups distractibility differences; (c) the relation between distractibility and speech-language dissociations; and (d) whether interactions between distractibility and dissociations predicted the frequency of total, stuttered, and nonstuttered disfluencies. Results More preschool-age CWS exhibited speech-language dissociations compared with CWNS, and more boys exhibited dissociations compared with girls. In addition, male CWS were less distractible than female CWS and female CWNS. For CWS, but not CWNS, less distractibility (i.e., greater attention) was associated with more speech-language dissociations. Last, interactions between distractibility and dissociations did not predict speech disfluencies in CWS or CWNS. Conclusions The present findings suggest that for preschool-age CWS, attentional processes are associated with speech-language dissociations. Future investigations are warranted to better understand the directionality of effect of this association (e.g., inefficient attentional processes → speech-language dissociations vs. inefficient attentional processes ← speech-language dissociations). PMID:26126203
Primary progressive aphasia and apraxia of speech.
Jung, Youngsin; Duffy, Joseph R; Josephs, Keith A
2013-09-01
Primary progressive aphasia is a neurodegenerative syndrome characterized by progressive language dysfunction. The majority of primary progressive aphasia cases can be classified into three subtypes: nonfluent/agrammatic, semantic, and logopenic variants. Each variant presents with unique clinical features, and is associated with distinctive underlying pathology and neuroimaging findings. Unlike primary progressive aphasia, apraxia of speech is a disorder that involves inaccurate production of sounds secondary to impaired planning or programming of speech movements. Primary progressive apraxia of speech is a neurodegenerative form of apraxia of speech, and it should be distinguished from primary progressive aphasia given its discrete clinicopathological presentation. Recently, there have been substantial advances in our understanding of these speech and language disorders. The clinical, neuroimaging, and histopathological features of primary progressive aphasia and apraxia of speech are reviewed in this article. The distinctions among these disorders for accurate diagnosis are increasingly important from a prognostic and therapeutic standpoint. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.
Narayan, Angela; Cicchetti, Dante; Rogosch, Fred A; Toth, Sheree L
2015-02-01
Research has documented that maternal expressed emotion-criticism (EE-Crit) from the Five-Minute Speech Sample (FMSS) predicts family conflict and children's externalizing behavior in clinical and community samples. However, studies have not examined EE-Crit in maltreating or separated/divorced families, or whether these family risks exacerbate the links between EE-Crit and family conflict and externalizing behavior. The current study examined the associations between maternal EE-Crit, maltreatment, and separation/divorce, and whether maltreatment and separation/divorce moderated associations between EE-Crit and children's externalizing problems, and EE-Crit and family conflict. Participants included 123 children (M = 8.01 years, SD = 1.58; 64.2 % males) from maltreating (n = 83) or low-income, comparison (n = 40) families, and 123 mothers (n = 48 separated/divorced). Mothers completed the FMSS for EE-Crit and the Family Environment Scale for family conflict. Maltreatment was coded with the Maltreatment Classification System using information from official Child Protection Services (CPS) reports from the Department of Human Services (DHS). Trained summer camp counselors rated children's externalizing behavior. Maltreatment was directly associated with higher externalizing problems, and separation/divorce, but not maltreatment, moderated the association between EE-Crit and externalizing behavior. Analyses pertaining to family conflict were not significant. Findings indicate that maltreatment is a direct risk factor for children's externalizing behavior and separation/divorce is a vulnerability factor for externalizing behavior in family contexts with high maternal EE-Crit. Intervention, prevention, and policy efforts to promote resilience in high-risk families may be effective in targeting maltreating and critical parents, especially those with co-occurring separation/divorce. Key Words: expressed emotion, EE-Crit, Five-Minute Speech Sample; maltreatment, divorce
Dick, Anthony Steven; Mok, Eva H; Raja Beharelle, Anjali; Goldin-Meadow, Susan; Small, Steven L
2014-03-01
In everyday conversation, listeners often rely on a speaker's gestures to clarify any ambiguities in the verbal message. Using fMRI during naturalistic story comprehension, we examined which brain regions in the listener are sensitive to speakers' iconic gestures. We focused on iconic gestures that contribute information not found in the speaker's talk, compared with those that convey information redundant with the speaker's talk. We found that three regions-left inferior frontal gyrus triangular (IFGTr) and opercular (IFGOp) portions, and left posterior middle temporal gyrus (MTGp)--responded more strongly when gestures added information to nonspecific language, compared with when they conveyed the same information in more specific language; in other words, when gesture disambiguated speech as opposed to reinforced it. An increased BOLD response was not found in these regions when the nonspecific language was produced without gesture, suggesting that IFGTr, IFGOp, and MTGp are involved in integrating semantic information across gesture and speech. In addition, we found that activity in the posterior superior temporal sulcus (STSp), previously thought to be involved in gesture-speech integration, was not sensitive to the gesture-speech relation. Together, these findings clarify the neurobiology of gesture-speech integration and contribute to an emerging picture of how listeners glean meaning from gestures that accompany speech. Copyright © 2012 Wiley Periodicals, Inc.
National features of speech etiquette
Nacafova S.
2017-01-01
The article shows the differences between the speech etiquette of different peoples. The most important thing is to find a common language with this or that interlocutor. Knowledge of national etiquette, national character helps to learn the principles of speech of another nation. The article indicates in which cases certain forms of etiquette considered acceptable. At the same time, the rules of etiquette emphasized in the conduct of a dialogue in official meetings and for example, in the ex...
Evidence-based speech-language pathology practices in schools: findings from a national survey.
Hoffman, Lavae M; Ireland, Marie; Hall-Mills, Shannon; Flynn, Perry
2013-07-01
This study documented evidence-based practice (EBP) patterns as reported by speech-language pathologists (SLPs) employed in public schools during 2010-2011. Using an online survey, practioners reported their EBP training experiences, resources available in their workplaces, and the frequency with which they engage in specific EBP activities, as well as their resource needs and future training format preferences. A total of 2,762 SLPs in 28 states participated in the online survey, 85% of whom reported holding the Certificate of Clinical Competence in Speech-Language Pathology credential. Results revealed that one quarter of survey respondents had no formal training in EBP, 11% of SLPs worked in school districts with official EBP procedural guidelines, and 91% had no scheduled time to support EBP activities. The majority of SLPs posed and researched 0 to 2 EBP questions per year and read 0 to 4 American Speech-Language-Hearing Association (ASHA) journal articles per year on either assessment or intervention topics. Use of ASHA online resources and engagement in EBP activities were documented to be low. However, results also revealed that school-based SLPs have high interest in additional training and resources to support scientifically based practices. Suggestions for enhancing EBP support in public schools and augmenting knowledge transfer are provided.
Speech endpoint detection with non-language speech sounds for generic speech processing applications
McClain, Matthew; Romanowski, Brian
2009-05-01
Non-language speech sounds (NLSS) are sounds produced by humans that do not carry linguistic information. Examples of these sounds are coughs, clicks, breaths, and filled pauses such as "uh" and "um" in English. NLSS are prominent in conversational speech, but can be a significant source of errors in speech processing applications. Traditionally, these sounds are ignored by speech endpoint detection algorithms, where speech regions are identified in the audio signal prior to processing. The ability to filter NLSS as a pre-processing step can significantly enhance the performance of many speech processing applications, such as speaker identification, language identification, and automatic speech recognition. In order to be used in all such applications, NLSS detection must be performed without the use of language models that provide knowledge of the phonology and lexical structure of speech. This is especially relevant to situations where the languages used in the audio are not known apriori. We present the results of preliminary experiments using data from American and British English speakers, in which segments of audio are classified as language speech sounds (LSS) or NLSS using a set of acoustic features designed for language-agnostic NLSS detection and a hidden-Markov model (HMM) to model speech generation. The results of these experiments indicate that the features and model used are capable of detection certain types of NLSS, such as breaths and clicks, while detection of other types of NLSS such as filled pauses will require future research.
Segmental intelligibility of synthetic speech produced by rule.
Logan, J S; Greene, B G; Pisoni, D B
1989-08-01
This paper reports the results of an investigation that employed the modified rhyme test (MRT) to measure the segmental intelligibility of synthetic speech generated automatically by rule. Synthetic speech produced by ten text-to-speech systems was studied and compared to natural speech. A variation of the standard MRT was also used to study the effects of response set size on perceptual confusions. Results indicated that the segmental intelligibility scores formed a continuum. Several systems displayed very high levels of performance that were close to or equal to scores obtained with natural speech; other systems displayed substantially worse performance compared to natural speech. The overall performance of the best system, DECtalk--Paul, was equivalent to the data obtained with natural speech for consonants in syllable-initial position. The findings from this study are discussed in terms of the use of a set of standardized procedures for measuring intelligibility of synthetic speech under controlled laboratory conditions. Recent work investigating the perception of synthetic speech under more severe conditions in which greater demands are made on the listener's processing resources is also considered. The wide range of intelligibility scores obtained in the present study demonstrates important differences in perception and suggests that not all synthetic speech is perceptually equivalent to the listener.
Segmental intelligibility of synthetic speech produced by rule
Logan, John S.; Greene, Beth G.; Pisoni, David B.
2012-01-01
This paper reports the results of an investigation that employed the modified rhyme test (MRT) to measure the segmental intelligibility of synthetic speech generated automatically by rule. Synthetic speech produced by ten text-to-speech systems was studied and compared to natural speech. A variation of the standard MRT was also used to study the effects of response set size on perceptual confusions. Results indicated that the segmental intelligibility scores formed a continuum. Several systems displayed very high levels of performance that were close to or equal to scores obtained with natural speech; other systems displayed substantially worse performance compared to natural speech. The overall performance of the best system, DECtalk—Paul, was equivalent to the data obtained with natural speech for consonants in syllable-initial position. The findings from this study are discussed in terms of the use of a set of standardized procedures for measuring intelligibility of synthetic speech under controlled laboratory conditions. Recent work investigating the perception of synthetic speech under more severe conditions in which greater demands are made on the listener’s processing resources is also considered. The wide range of intelligibility scores obtained in the present study demonstrates important differences in perception and suggests that not all synthetic speech is perceptually equivalent to the listener. PMID:2527884
Giallo, Rebecca; Bahreinian, Salma; Brown, Stephanie; Cooklin, Amanda; Kingston, Dawn; Kozyrskyj, Anita
2015-01-01
There is a growing body of evidence attesting to links between early life exposure to stress and childhood asthma. However, available evidence is largely based on small, genetically high risk samples. The aim of this study was to explore the associations between the course of maternal depressive symptoms across early childhood and childhood asthma in a nationally representative longitudinal cohort study of Australian children. Participants were 4164 children and their biological mothers from the Longitudinal Study of Australian Children. Latent class analysis identified three trajectories of maternal depressive symptoms across four biennial waves from the first postnatal year to when children were 6-7 years: minimal symptoms (74.6%), sub-clinical symptoms (20.8%), and persistent and increasing high symptoms (4.6%). Logistic regression analyses revealed that childhood asthma at age 6-7 years was associated with persistent and increasing high depressive symptoms after accounting for known risk factors including smoking during pregnancy and maternal history of asthma (adjusted OR 2.36, 95% CI 1.61-3.45), p.001). Our findings from a nationally representative sample of Australian children provide empirical support for a relationship between maternal depressive symptoms across the early childhood period and childhood asthma. The burden of disease from childhood asthma may be reduced by strengthening efforts to promote maternal mental health in the early years of parenting.
Directory of Open Access Journals (Sweden)
Rebecca Giallo
Full Text Available There is a growing body of evidence attesting to links between early life exposure to stress and childhood asthma. However, available evidence is largely based on small, genetically high risk samples. The aim of this study was to explore the associations between the course of maternal depressive symptoms across early childhood and childhood asthma in a nationally representative longitudinal cohort study of Australian children. Participants were 4164 children and their biological mothers from the Longitudinal Study of Australian Children. Latent class analysis identified three trajectories of maternal depressive symptoms across four biennial waves from the first postnatal year to when children were 6-7 years: minimal symptoms (74.6%, sub-clinical symptoms (20.8%, and persistent and increasing high symptoms (4.6%. Logistic regression analyses revealed that childhood asthma at age 6-7 years was associated with persistent and increasing high depressive symptoms after accounting for known risk factors including smoking during pregnancy and maternal history of asthma (adjusted OR 2.36, 95% CI 1.61-3.45, p.001. Our findings from a nationally representative sample of Australian children provide empirical support for a relationship between maternal depressive symptoms across the early childhood period and childhood asthma. The burden of disease from childhood asthma may be reduced by strengthening efforts to promote maternal mental health in the early years of parenting.
Freedom of racist speech: Ego and expressive threats.
White, Mark H; Crandall, Christian S
2017-09-01
Do claims of "free speech" provide cover for prejudice? We investigate whether this defense of racist or hate speech serves as a justification for prejudice. In a series of 8 studies (N = 1,624), we found that explicit racial prejudice is a reliable predictor of the "free speech defense" of racist expression. Participants endorsed free speech values for singing racists songs or posting racist comments on social media; people high in prejudice endorsed free speech more than people low in prejudice (meta-analytic r = .43). This endorsement was not principled-high levels of prejudice did not predict endorsement of free speech values when identical speech was directed at coworkers or the police. Participants low in explicit racial prejudice actively avoided endorsing free speech values in racialized conditions compared to nonracial conditions, but participants high in racial prejudice increased their endorsement of free speech values in racialized conditions. Three experiments failed to find evidence that defense of racist speech by the highly prejudiced was based in self-relevant or self-protective motives. Two experiments found evidence that the free speech argument protected participants' own freedom to express their attitudes; the defense of other's racist speech seems motivated more by threats to autonomy than threats to self-regard. These studies serve as an elaboration of the Justification-Suppression Model (Crandall & Eshleman, 2003) of prejudice expression. The justification of racist speech by endorsing fundamental political values can serve to buffer racial and hate speech from normative disapproval. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Music and Speech Perception in Children Using Sung Speech.
Nie, Yingjiu; Galvin, John J; Morikawa, Michael; André, Victoria; Wheeler, Harley; Fu, Qian-Jie
2018-01-01
This study examined music and speech perception in normal-hearing children with some or no musical training. Thirty children (mean age = 11.3 years), 15 with and 15 without formal music training participated in the study. Music perception was measured using a melodic contour identification (MCI) task; stimuli were a piano sample or sung speech with a fixed timbre (same word for each note) or a mixed timbre (different words for each note). Speech perception was measured in quiet and in steady noise using a matrix-styled sentence recognition task; stimuli were naturally intonated speech or sung speech with a fixed pitch (same note for each word) or a mixed pitch (different notes for each word). Significant musician advantages were observed for MCI and speech in noise but not for speech in quiet. MCI performance was significantly poorer with the mixed timbre stimuli. Speech performance in noise was significantly poorer with the fixed or mixed pitch stimuli than with spoken speech. Across all subjects, age at testing and MCI performance were significantly correlated with speech performance in noise. MCI and speech performance in quiet was significantly poorer for children than for adults from a related study using the same stimuli and tasks; speech performance in noise was significantly poorer for young than for older children. Long-term music training appeared to benefit melodic pitch perception and speech understanding in noise in these pediatric listeners.
Directory of Open Access Journals (Sweden)
Isabela Crivellaro Gonçalves
2011-01-01
Full Text Available OBJECTIVE: To determine whether neurophysiological auditory brainstem responses to clicks and repeated speech stimuli differ between typically developing children and children with phonological disorders. INTRODUCTION: Phonological disorders are language impairments resulting from inadequate use of adult phonological language rules and are among the most common speech and language disorders in children (prevalence: 8 - 9%. Our hypothesis is that children with phonological disorders have basic differences in the way that their brains encode acoustic signals at brainstem level when compared to normal counterparts. METHODS: We recorded click and speech evoked auditory brainstem responses in 18 typically developing children (control group and in 18 children who were clinically diagnosed with phonological disorders (research group. The age range of the children was from 7-11 years. RESULTS: The research group exhibited significantly longer latency responses to click stimuli (waves I, III and V and speech stimuli (waves V and A when compared to the control group. DISCUSSION: These results suggest that the abnormal encoding of speech sounds may be a biological marker of phonological disorders. However, these results cannot define the biological origins of phonological problems. We also observed that speech-evoked auditory brainstem responses had a higher specificity/sensitivity for identifying phonological disorders than click-evoked auditory brainstem responses. CONCLUSIONS: Early stages of the auditory pathway processing of an acoustic stimulus are not similar in typically developing children and those with phonological disorders. These findings suggest that there are brainstem auditory pathway abnormalities in children with phonological disorders.
Miller, Suzanne M; Ciocci, Sandra R
2013-01-01
Undergraduate communication sciences and disorders students' attitudes toward speech-language pathology (SLP) clinical services to children and adults prior to and following community-based observations were examined. Participants (n=25) completed an online survey to elicit their opinions regarding their perceptions of their observation experiences. Findings revealed that after completion of community-based SLP clinical observations, 16 (64%) respondents reported a continued interest in a child-based clinical focus; 12 (48%) respondents continued to consider a clinical interest in adults, while 5 respondents (20%) changed career interests to an adult focus based on their observation experiences. Findings support the notion that observations of SLP appear to significantly influence students' career choices. Clinical observations typically occur at the junior/senior undergraduate levels; therefore, suggestions are offered for inclusion of gerontology education embedded throughout the undergraduate communication sciences and disorders curricula to foster and expand students' knowledge of aging, and to prepare our students to meet the healthcare challenges of elders in the 21st century.
... Health Info » Voice, Speech, and Language Apraxia of Speech On this page: What is apraxia of speech? ... about apraxia of speech? What is apraxia of speech? Apraxia of speech (AOS)—also known as acquired ...
The impact of exploiting spectro-temporal context in computational speech segregation
DEFF Research Database (Denmark)
Bentsen, Thomas; Kressner, Abigail Anne; Dau, Torsten
2018-01-01
Computational speech segregation aims to automatically segregate speech from interfering noise, often by employing ideal binary mask estimation. Several studies have tried to exploit contextual information in speech to improve mask estimation accuracy by using two frequently-used strategies that (1...... for measured intelligibility. The findings may have implications for the design of speech segregation systems, and for the selection of a cost function that correlates with intelligibility....
Using DEDICOM for completely unsupervised part-of-speech tagging.
Energy Technology Data Exchange (ETDEWEB)
Chew, Peter A.; Bader, Brett William; Rozovskaya, Alla (University of Illinois, Urbana, IL)
2009-02-01
A standard and widespread approach to part-of-speech tagging is based on Hidden Markov Models (HMMs). An alternative approach, pioneered by Schuetze (1993), induces parts of speech from scratch using singular value decomposition (SVD). We introduce DEDICOM as an alternative to SVD for part-of-speech induction. DEDICOM retains the advantages of SVD in that it is completely unsupervised: no prior knowledge is required to induce either the tagset or the associations of terms with tags. However, unlike SVD, it is also fully compatible with the HMM framework, in that it can be used to estimate emission- and transition-probability matrices which can then be used as the input for an HMM. We apply the DEDICOM method to the CONLL corpus (CONLL 2000) and compare the output of DEDICOM to the part-of-speech tags given in the corpus, and find that the correlation (almost 0.5) is quite high. Using DEDICOM, we also estimate part-of-speech ambiguity for each term, and find that these estimates correlate highly with part-of-speech ambiguity as measured in the original corpus (around 0.88). Finally, we show how the output of DEDICOM can be evaluated and compared against the more familiar output of supervised HMM-based tagging.
Iuzzini, Jenya
There is a lack of agreement on the features used to differentiate Childhood Apraxia of Speech (CAS) from Phonological Disorders (PD). One criterion which has gained consensus is lexical inconsistency of speech (ASHA, 2007); however, no accepted measure of this feature has been defined. Although lexical assessment provides information about consistency of an item across repeated trials, it may not capture the magnitude of inconsistency within an item. In contrast, segmental analysis provides more extensive information about consistency of phoneme usage across multiple contexts and word-positions. The current research compared segmental and lexical inconsistency metrics in preschool-aged children with PD, CAS, and typical development (TD) to determine how inconsistency varies with age in typical and disordered speakers, and whether CAS and PD were differentiated equally well by both assessment levels. Whereas lexical and segmental analyses may be influenced by listener characteristics or speaker intelligibility, the acoustic signal is less vulnerable to these factors. In addition, the acoustic signal may reveal information which is not evident in the perceptual signal. A second focus of the current research was motivated by Blumstein et al.'s (1980) classic study on voice onset time (VOT) in adults with acquired apraxia of speech (AOS) which demonstrated a motor impairment underlying AOS. In the current study, VOT analyses were conducted to determine the relationship between age and group with the voicing distribution for bilabial and alveolar plosives. Findings revealed that 3-year-olds evidenced significantly higher inconsistency than 5-year-olds; segmental inconsistency approached 0% in 5-year-olds with TD, whereas it persisted in children with PD and CAS suggesting that for child in this age-range, inconsistency is a feature of speech disorder rather than typical development (Holm et al., 2007). Likewise, whereas segmental and lexical inconsistency were
Prediction and constraint in audiovisual speech perception
Peelle, Jonathan E.; Sommers, Mitchell S.
2015-01-01
During face-to-face conversational speech listeners must efficiently process a rapid and complex stream of multisensory information. Visual speech can serve as a critical complement to auditory information because it provides cues to both the timing of the incoming acoustic signal (the amplitude envelope, influencing attention and perceptual sensitivity) and its content (place and manner of articulation, constraining lexical selection). Here we review behavioral and neurophysiological evidence regarding listeners' use of visual speech information. Multisensory integration of audiovisual speech cues improves recognition accuracy, particularly for speech in noise. Even when speech is intelligible based solely on auditory information, adding visual information may reduce the cognitive demands placed on listeners through increasing precision of prediction. Electrophysiological studies demonstrate oscillatory cortical entrainment to speech in auditory cortex is enhanced when visual speech is present, increasing sensitivity to important acoustic cues. Neuroimaging studies also suggest increased activity in auditory cortex when congruent visual information is available, but additionally emphasize the involvement of heteromodal regions of posterior superior temporal sulcus as playing a role in integrative processing. We interpret these findings in a framework of temporally-focused lexical competition in which visual speech information affects auditory processing to increase sensitivity to auditory information through an early integration mechanism, and a late integration stage that incorporates specific information about a speaker's articulators to constrain the number of possible candidates in a spoken utterance. Ultimately it is words compatible with both auditory and visual information that most strongly determine successful speech perception during everyday listening. Thus, audiovisual speech perception is accomplished through multiple stages of integration, supported
Prediction and constraint in audiovisual speech perception.
Peelle, Jonathan E; Sommers, Mitchell S
2015-07-01
During face-to-face conversational speech listeners must efficiently process a rapid and complex stream of multisensory information. Visual speech can serve as a critical complement to auditory information because it provides cues to both the timing of the incoming acoustic signal (the amplitude envelope, influencing attention and perceptual sensitivity) and its content (place and manner of articulation, constraining lexical selection). Here we review behavioral and neurophysiological evidence regarding listeners' use of visual speech information. Multisensory integration of audiovisual speech cues improves recognition accuracy, particularly for speech in noise. Even when speech is intelligible based solely on auditory information, adding visual information may reduce the cognitive demands placed on listeners through increasing the precision of prediction. Electrophysiological studies demonstrate that oscillatory cortical entrainment to speech in auditory cortex is enhanced when visual speech is present, increasing sensitivity to important acoustic cues. Neuroimaging studies also suggest increased activity in auditory cortex when congruent visual information is available, but additionally emphasize the involvement of heteromodal regions of posterior superior temporal sulcus as playing a role in integrative processing. We interpret these findings in a framework of temporally-focused lexical competition in which visual speech information affects auditory processing to increase sensitivity to acoustic information through an early integration mechanism, and a late integration stage that incorporates specific information about a speaker's articulators to constrain the number of possible candidates in a spoken utterance. Ultimately it is words compatible with both auditory and visual information that most strongly determine successful speech perception during everyday listening. Thus, audiovisual speech perception is accomplished through multiple stages of integration
Directory of Open Access Journals (Sweden)
Jerry D. Gibson
2016-06-01
Full Text Available Speech compression is a key technology underlying digital cellular communications, VoIP, voicemail, and voice response systems. We trace the evolution of speech coding based on the linear prediction model, highlight the key milestones in speech coding, and outline the structures of the most important speech coding standards. Current challenges, future research directions, fundamental limits on performance, and the critical open problem of speech coding for emergency first responders are all discussed.
Searl, Jeff; Knollhoff, Stephanie; Barohn, Richard J.
2017-01-01
Purpose: This preliminary study on lingual-alveolar contact pressures (LACP) in people with amyotrophic lateral sclerosis (ALS) had several aims: (a) to evaluate whether the protocol induced fatigue, (b) to compare LACP during speech (LACP-Sp) and during maximum isometric pressing (LACP-Max) in people with ALS (PALS) versus healthy controls, (c)…
THE USE OF EXPRESSIVE SPEECH ACTS IN HANNAH MONTANA SESSION 1
Directory of Open Access Journals (Sweden)
Nur Vita Handayani
2015-07-01
Full Text Available This study aims to describe kinds and forms of expressive speech act in Hannah Montana Session 1. It belongs to descriptive qualitative method. The research object was expressive speech act. The data source was utterances which contain expressive speech acts in the film Hannah Montana Session 1. The researcher used observation method and noting technique in collecting the data. In analyzing the data, descriptive qualitative method was used. The research findings show that there are ten kinds of expressive speech act found in Hannah Montana Session 1, namely expressing apology, expressing thanks, expressing sympathy, expressing attitudes, expressing greeting, expressing wishes, expressing joy, expressing pain, expressing likes, and expressing dislikes. The forms of expressive speech act are direct literal expressive speech act, direct non-literal expressive speech act, indirect literal expressive speech act, and indirect non-literal expressive speech act.
Commencement Speech as a Hybrid Polydiscursive Practice
Directory of Open Access Journals (Sweden)
Светлана Викторовна Иванова
2017-12-01
Full Text Available Discourse and media communication researchers pay attention to the fact that popular discursive and communicative practices have a tendency to hybridization and convergence. Discourse which is understood as language in use is flexible. Consequently, it turns out that one and the same text can represent several types of discourses. A vivid example of this tendency is revealed in American commencement speech / commencement address / graduation speech. A commencement speech is a speech university graduates are addressed with which in compliance with the modern trend is delivered by outstanding media personalities (politicians, athletes, actors, etc.. The objective of this study is to define the specificity of the realization of polydiscursive practices within commencement speech. The research involves discursive, contextual, stylistic and definitive analyses. Methodologically the study is based on the discourse analysis theory, in particular the notion of a discursive practice as a verbalized social practice makes up the conceptual basis of the research. This research draws upon a hundred commencement speeches delivered by prominent representatives of American society since 1980s till now. In brief, commencement speech belongs to institutional discourse public speech embodies. Commencement speech institutional parameters are well represented in speeches delivered by people in power like American and university presidents. Nevertheless, as the results of the research indicate commencement speech institutional character is not its only feature. Conceptual information analysis enables to refer commencement speech to didactic discourse as it is aimed at teaching university graduates how to deal with challenges life is rich in. Discursive practices of personal discourse are also actively integrated into the commencement speech discourse. More than that, existential discursive practices also find their way into the discourse under study. Commencement
Infants' brain responses to speech suggest analysis by synthesis.
Kuhl, Patricia K; Ramírez, Rey R; Bosseler, Alexis; Lin, Jo-Fu Lotus; Imada, Toshiaki
2014-08-05
Historic theories of speech perception (Motor Theory and Analysis by Synthesis) invoked listeners' knowledge of speech production to explain speech perception. Neuroimaging data show that adult listeners activate motor brain areas during speech perception. In two experiments using magnetoencephalography (MEG), we investigated motor brain activation, as well as auditory brain activation, during discrimination of native and nonnative syllables in infants at two ages that straddle the developmental transition from language-universal to language-specific speech perception. Adults are also tested in Exp. 1. MEG data revealed that 7-mo-old infants activate auditory (superior temporal) as well as motor brain areas (Broca's area, cerebellum) in response to speech, and equivalently for native and nonnative syllables. However, in 11- and 12-mo-old infants, native speech activates auditory brain areas to a greater degree than nonnative, whereas nonnative speech activates motor brain areas to a greater degree than native speech. This double dissociation in 11- to 12-mo-old infants matches the pattern of results obtained in adult listeners. Our infant data are consistent with Analysis by Synthesis: auditory analysis of speech is coupled with synthesis of the motor plans necessary to produce the speech signal. The findings have implications for: (i) perception-action theories of speech perception, (ii) the impact of "motherese" on early language learning, and (iii) the "social-gating" hypothesis and humans' development of social understanding.
Krieger-Redwood, Katya; Gaskell, M Gareth; Lindsay, Shane; Jefferies, Elizabeth
2013-12-01
Several accounts of speech perception propose that the areas involved in producing language are also involved in perceiving it. In line with this view, neuroimaging studies show activation of premotor cortex (PMC) during phoneme judgment tasks; however, there is debate about whether speech perception necessarily involves motor processes, across all task contexts, or whether the contribution of PMC is restricted to tasks requiring explicit phoneme awareness. Some aspects of speech processing, such as mapping sounds onto meaning, may proceed without the involvement of motor speech areas if PMC specifically contributes to the manipulation and categorical perception of phonemes. We applied TMS to three sites-PMC, posterior superior temporal gyrus, and occipital pole-and for the first time within the TMS literature, directly contrasted two speech perception tasks that required explicit phoneme decisions and mapping of speech sounds onto semantic categories, respectively. TMS to PMC disrupted explicit phonological judgments but not access to meaning for the same speech stimuli. TMS to two further sites confirmed that this pattern was site specific and did not reflect a generic difference in the susceptibility of our experimental tasks to TMS: stimulation of pSTG, a site involved in auditory processing, disrupted performance in both language tasks, whereas stimulation of occipital pole had no effect on performance in either task. These findings demonstrate that, although PMC is important for explicit phonological judgments, crucially, PMC is not necessary for mapping speech onto meanings.
Vogel, J P; Souza, J P; Mori, R; Morisaki, N; Lumbiganon, P; Laopaiboon, M; Ortiz-Panozo, E; Hernandez, B; Pérez-Cuevas, R; Roy, M; Mittal, S; Cecatti, J G; Tunçalp, Ö; Gülmezoglu, A M
2014-03-01
We aimed to determine the prevalence and risks of late fetal deaths (LFDs) and early neonatal deaths (ENDs) in women with medical and obstetric complications. Secondary analysis of the WHO Multicountry Survey on Maternal and Newborn Health (WHOMCS). A total of 359 participating facilities in 29 countries. A total of 308 392 singleton deliveries. We reported on perinatal indicators and determined risks of perinatal death in the presence of severe maternal complications (haemorrhagic, infectious, and hypertensive disorders, and other medical conditions). Fresh and macerated LFDs (defined as stillbirths ≥ 1000 g and/or ≥28 weeks of gestation) and ENDs. The LFD rate was 17.7 per 1000 births; 64.8% were fresh stillbirths. The END rate was 8.4 per 1000 liveborns; 67.1% occurred by day 3 of life. Maternal complications were present in 22.9, 27.7, and 21.2% [corrected] of macerated LFDs, fresh LFDs, and ENDs, respectively. The risks of all three perinatal mortality outcomes were significantly increased with placental abruption, ruptured uterus, systemic infections/sepsis, pre-eclampsia, eclampsia, and severe anaemia. Preventing intrapartum-related perinatal deaths requires a comprehensive approach to quality intrapartum care, beyond the provision of caesarean section. Early identification and management of women with complications could improve maternal and perinatal outcomes. © 2014 RCOG The World Health Organization retains copyright and all other rights in the manuscript of this article as submitted for publication.
Speech Production and Speech Discrimination by Hearing-Impaired Children.
Novelli-Olmstead, Tina; Ling, Daniel
1984-01-01
Seven hearing impaired children (five to seven years old) assigned to the Speakers group made highly significant gains in speech production and auditory discrimination of speech, while Listeners made only slight speech production gains and no gains in auditory discrimination. Combined speech and auditory training was more effective than auditory…
Methods of analysis speech rate: a pilot study.
Costa, Luanna Maria Oliveira; Martins-Reis, Vanessa de Oliveira; Celeste, Letícia Côrrea
2016-01-01
To describe the performance of fluent adults in different measures of speech rate. The study included 24 fluent adults, of both genders, speakers of Brazilian Portuguese, who were born and still living in the metropolitan region of Belo Horizonte, state of Minas Gerais, aged between 18 and 59 years. Participants were grouped by age: G1 (18-29 years), G2 (30-39 years), G3 (40-49 years), and G4 (50-59 years). The speech samples were obtained following the methodology of the Speech Fluency Assessment Protocol. In addition to the measures of speech rate proposed by the protocol (speech rate in words and syllables per minute), the rate of speech into phonemes per second and the articulation rate with and without the disfluencies were calculated. We used the nonparametric Friedman test and the Wilcoxon test for multiple comparisons. Groups were compared using the nonparametric Kruskal Wallis. The significance level was of 5%. There were significant differences between measures of speech rate involving syllables. The multiple comparisons showed that all the three measures were different. There was no effect of age for the studied measures. These findings corroborate previous studies. The inclusion of temporal acoustic measures such as speech rate in phonemes per second and articulation rates with and without disfluencies can be a complementary approach in the evaluation of speech rate.
Davidow, Jason H; Grossman, Heather L; Edge, Robin L
2018-05-01
Voluntary stuttering techniques involve persons who stutter purposefully interjecting disfluencies into their speech. Little research has been conducted on the impact of these techniques on the speech pattern of persons who stutter. The present study examined whether changes in the frequency of voluntary stuttering accompanied changes in stuttering frequency, articulation rate, speech naturalness, and speech effort. In total, 12 persons who stutter aged 16-34 years participated. Participants read four 300-syllable passages during a control condition, and three voluntary stuttering conditions that involved attempting to produce purposeful, tension-free repetitions of initial sounds or syllables of a word for two or more repetitions (i.e., bouncing). The three voluntary stuttering conditions included bouncing on 5%, 10%, and 15% of syllables read. Friedman tests and follow-up Wilcoxon signed ranks tests were conducted for the statistical analyses. Stuttering frequency, articulation rate, and speech naturalness were significantly different between the voluntary stuttering conditions. Speech effort did not differ between the voluntary stuttering conditions. Stuttering frequency was significantly lower during the three voluntary stuttering conditions compared to the control condition, and speech effort was significantly lower during two of the three voluntary stuttering conditions compared to the control condition. Due to changes in articulation rate across the voluntary stuttering conditions, it is difficult to conclude, as has been suggested previously, that voluntary stuttering is the reason for stuttering reductions found when using voluntary stuttering techniques. Additionally, future investigations should examine different types of voluntary stuttering over an extended period of time to determine their impact on stuttering frequency, speech rate, speech naturalness, and speech effort.
Comprehension of synthetic speech and digitized natural speech by adults with aphasia.
Hux, Karen; Knollman-Porter, Kelly; Brown, Jessica; Wallace, Sarah E
2017-09-01
Using text-to-speech technology to provide simultaneous written and auditory content presentation may help compensate for chronic reading challenges if people with aphasia can understand synthetic speech output; however, inherent auditory comprehension challenges experienced by people with aphasia may make understanding synthetic speech difficult. This study's purpose was to compare the preferences and auditory comprehension accuracy of people with aphasia when listening to sentences generated with digitized natural speech, Alex synthetic speech (i.e., Macintosh platform), or David synthetic speech (i.e., Windows platform). The methodology required each of 20 participants with aphasia to select one of four images corresponding in meaning to each of 60 sentences comprising three stimulus sets. Results revealed significantly better accuracy given digitized natural speech than either synthetic speech option; however, individual participant performance analyses revealed three patterns: (a) comparable accuracy regardless of speech condition for 30% of participants, (b) comparable accuracy between digitized natural speech and one, but not both, synthetic speech option for 45% of participants, and (c) greater accuracy with digitized natural speech than with either synthetic speech option for remaining participants. Ranking and Likert-scale rating data revealed a preference for digitized natural speech and David synthetic speech over Alex synthetic speech. Results suggest many individuals with aphasia can comprehend synthetic speech options available on popular operating systems. Further examination of synthetic speech use to support reading comprehension through text-to-speech technology is thus warranted. Copyright © 2017 Elsevier Inc. All rights reserved.
Common neural substrates support speech and non-speech vocal tract gestures
Chang, Soo-Eun; Kenney, Mary Kay; Loucks, Torrey M.J.; Poletto, Christopher J.; Ludlow, Christy L.
2009-01-01
The issue of whether speech is supported by the same neural substrates as non-speech vocal-tract gestures has been contentious. In this fMRI study we tested whether producing non-speech vocal tract gestures in humans shares the same functional neuroanatomy as non-sense speech syllables. Production of non-speech vocal tract gestures, devoid of phonological content but similar to speech in that they had familiar acoustic and somatosensory targets, were compared to the production of speech sylla...
International Nuclear Information System (INIS)
2001-01-01
This CD is multimedia presentation of programme safety upgrading of Bohunice V1 NPP. This chapter consist of introductory commentary and 4 introductory speeches (video records): (1) Introductory speech of Vincent Pillar, Board chairman and director general of Slovak electric, Plc. (SE); (2) Introductory speech of Stefan Schmidt, director of SE - Bohunice Nuclear power plants; (3) Introductory speech of Jan Korec, Board chairman and director general of VUJE Trnava, Inc. - Engineering, Design and Research Organisation, Trnava; Introductory speech of Dietrich Kuschel, Senior vice-president of FRAMATOME ANP Project and Engineering
Licata, Maria; Zietlow, Anna-Lena; Träuble, Birgit; Sodian, Beate; Reck, Corinna
High maternal emotional availability (EA) positively affects various domains of child development. However, the question of which factors promote or hinder maternal EA has not been investigated systematically. The present study investigated several maternal characteristics, namely maternal psychopathology, maternal attachment style insecurity, and theory of mind (ToM) as possible factors that influence maternal EA. The sample was comprised of 56 mothers and their preschool-aged children. Half of the mothers were diagnosed with postpartum depression and or anxiety disorders according to DSM-IV, and the other half were healthy controls. The results showed that both low maternal attachment style insecurity and high ToM skills significantly predicted maternal EA sensitivity, independently from maternal postpartum and concurrent psychopathology and education. Moreover, maternal attachment style insecurity fully mediated the link between maternal postpartum psychopathology and sensitivity. The findings suggest that maternal attachment style security can buffer negative effects of maternal psychopathology on maternal sensitivity in the mother-child interaction. © 2016 S. Karger AG, Basel.
Predicting speech intelligibility in conditions with nonlinearly processed noisy speech
DEFF Research Database (Denmark)
Jørgensen, Søren; Dau, Torsten
2013-01-01
The speech-based envelope power spectrum model (sEPSM; [1]) was proposed in order to overcome the limitations of the classical speech transmission index (STI) and speech intelligibility index (SII). The sEPSM applies the signal-tonoise ratio in the envelope domain (SNRenv), which was demonstrated...... to successfully predict speech intelligibility in conditions with nonlinearly processed noisy speech, such as processing with spectral subtraction. Moreover, a multiresolution version (mr-sEPSM) was demonstrated to account for speech intelligibility in various conditions with stationary and fluctuating...
Paavola-Ruotsalainen, Leila; Lehtosaari, Jaana; Palomäki, Josefina; Tervo, Immi
2018-01-01
Maternal responsive and directive speech to children at ages 0;10 and 2;0 was investigated by applying a procedure frst introduced by Flynn and Masur (2007) to a new language community (Finnish). The issues examined were consistency and stability over time, and also the role of responsiveness and directiveness in child linguistic development at…
Bradford, Billie; Maude, Robyn
2014-08-26
Maternal perception of decreased fetal movements is a specific indicator of fetal compromise, notably in the context of poor fetal growth. There is currently no agreed numerical definition of decreased fetal movements, with the subjective perception of a decrease on the part of the mother being the most significant definition clinically. Both qualitative and quantitative aspects of fetal activity may be important in identifying the compromised fetus.Yet, how pregnant women perceive and describe fetal activity is under-investigated by qualitative means. The aim of this study was to explore normal fetal activity, through first-hand descriptive accounts by pregnant women. Using qualitative descriptive methodology, interviews were conducted with 19 low-risk women experiencing their first pregnancy, at two timepoints in their third trimester. Interview transcripts were later analysed using qualitative content analysis and patterns of fetal activity identified were then considered along-side the characteristics of the women and their birth outcomes. This paper focuses on a novel finding; the description by pregnant women of fetal behaviour indicative of hunger and satiation. Full findings will be presented in later papers. Most participants (74% 14 of 19) indicated mealtimes were a time of increased fetal activity. Eight participants provided detailed descriptions of increased activity around meals, with seven (37% 7 of 19) of these specifying increased fetal activity prior to meals or in the context of their own hunger. These movements were interpreted as a fetal demand for food often prompting the mother to eat. Interestingly, the women who described increased fetal activity in the context of hunger subsequently gave birth to smaller infants (mean difference 364 gm) than those who did not describe a fetal response to hunger. Food seeking behaviour may have a pre-birth origin. Maternal-fetal interaction around mealtimes could constitute an endocrine mediated
Neural networks supporting audiovisual integration for speech: A large-scale lesion study.
Hickok, Gregory; Rogalsky, Corianne; Matchin, William; Basilakos, Alexandra; Cai, Julia; Pillay, Sara; Ferrill, Michelle; Mickelsen, Soren; Anderson, Steven W; Love, Tracy; Binder, Jeffrey; Fridriksson, Julius
2018-06-01
Auditory and visual speech information are often strongly integrated resulting in perceptual enhancements for audiovisual (AV) speech over audio alone and sometimes yielding compelling illusory fusion percepts when AV cues are mismatched, the McGurk-MacDonald effect. Previous research has identified three candidate regions thought to be critical for AV speech integration: the posterior superior temporal sulcus (STS), early auditory cortex, and the posterior inferior frontal gyrus. We assess the causal involvement of these regions (and others) in the first large-scale (N = 100) lesion-based study of AV speech integration. Two primary findings emerged. First, behavioral performance and lesion maps for AV enhancement and illusory fusion measures indicate that classic metrics of AV speech integration are not necessarily measuring the same process. Second, lesions involving superior temporal auditory, lateral occipital visual, and multisensory zones in the STS are the most disruptive to AV speech integration. Further, when AV speech integration fails, the nature of the failure-auditory vs visual capture-can be predicted from the location of the lesions. These findings show that AV speech processing is supported by unimodal auditory and visual cortices as well as multimodal regions such as the STS at their boundary. Motor related frontal regions do not appear to play a role in AV speech integration. Copyright © 2018 Elsevier Ltd. All rights reserved.
Energy Technology Data Exchange (ETDEWEB)
Ravishankar, C., Hughes Network Systems, Germantown, MD
1998-05-08
Speech is the predominant means of communication between human beings and since the invention of the telephone by Alexander Graham Bell in 1876, speech services have remained to be the core service in almost all telecommunication systems. Original analog methods of telephony had the disadvantage of speech signal getting corrupted by noise, cross-talk and distortion Long haul transmissions which use repeaters to compensate for the loss in signal strength on transmission links also increase the associated noise and distortion. On the other hand digital transmission is relatively immune to noise, cross-talk and distortion primarily because of the capability to faithfully regenerate digital signal at each repeater purely based on a binary decision. Hence end-to-end performance of the digital link essentially becomes independent of the length and operating frequency bands of the link Hence from a transmission point of view digital transmission has been the preferred approach due to its higher immunity to noise. The need to carry digital speech became extremely important from a service provision point of view as well. Modem requirements have introduced the need for robust, flexible and secure services that can carry a multitude of signal types (such as voice, data and video) without a fundamental change in infrastructure. Such a requirement could not have been easily met without the advent of digital transmission systems, thereby requiring speech to be coded digitally. The term Speech Coding is often referred to techniques that represent or code speech signals either directly as a waveform or as a set of parameters by analyzing the speech signal. In either case, the codes are transmitted to the distant end where speech is reconstructed or synthesized using the received set of codes. A more generic term that is applicable to these techniques that is often interchangeably used with speech coding is the term voice coding. This term is more generic in the sense that the
Prosodic differences between declaratives and interrogatives in infant-directed speech.
Geffen, Susan; Mintz, Toben H
2017-07-01
In many languages, declaratives and interrogatives differ in word order properties, and in syntactic organization more broadly. Thus, in order to learn the distinct syntactic properties of the two sentence types, learners must first be able to distinguish them using non-syntactic information. Prosodic information is often assumed to be a useful basis for this type of discrimination, although no systematic studies of the prosodic cues available to infants have been reported. Analysis of maternal speech in three Standard American English-speaking mother-infant dyads found that polar interrogatives differed from declaratives on the patterning of pitch and duration on the final two syllables, but wh-questions did not. Thus, while prosody is unlikely to aid discrimination of declaratives from wh-questions, infant-directed speech provides prosodic information that infants could use to distinguish declaratives and polar interrogatives. We discuss how learners could leverage this information to identify all question forms, in the context of syntax acquisition.
Haderlein, Tino; Döllinger, Michael; Matoušek, Václav; Nöth, Elmar
2016-10-01
Automatic voice assessment is often performed using sustained vowels. In contrast, speech analysis of read-out texts can be applied to voice and speech assessment. Automatic speech recognition and prosodic analysis were used to find regression formulae between automatic and perceptual assessment of four voice and four speech criteria. The regression was trained with 21 men and 62 women (average age 49.2 years) and tested with another set of 24 men and 49 women (48.3 years), all suffering from chronic hoarseness. They read the text 'Der Nordwind und die Sonne' ('The North Wind and the Sun'). Five voice and speech therapists evaluated the data on 5-point Likert scales. Ten prosodic and recognition accuracy measures (features) were identified which describe all the examined criteria. Inter-rater correlation within the expert group was between r = 0.63 for the criterion 'match of breath and sense units' and r = 0.87 for the overall voice quality. Human-machine correlation was between r = 0.40 for the match of breath and sense units and r = 0.82 for intelligibility. The perceptual ratings of different criteria were highly correlated with each other. Likewise, the feature sets modeling the criteria were very similar. The automatic method is suitable for assessing chronic hoarseness in general and for subgroups of functional and organic dysphonia. In its current version, it is almost as reliable as a randomly picked rater from a group of voice and speech therapists.
Free Speech as a Cultural Value in the United States
Directory of Open Access Journals (Sweden)
Mauricio J. Alvarez
2018-02-01
Full Text Available Political orientation influences support for free speech, with liberals often reporting greater support for free speech than conservatives. We hypothesized that this effect should be moderated by cultural context: individualist cultures value individual self-expression and self-determination, and collectivist cultures value group harmony and conformity. These different foci should differently influence liberals and conservatives’ support for free speech within these cultures. Two studies evaluated the joint influence of political orientation and cultural context on support for free speech. Study 1, using a multilevel analysis of data from 37 U.S. states (n = 1,001, showed that conservatives report stronger support for free speech in collectivist states, whereas there were no differences between conservatives and liberals in support for free speech in individualist states. Study 2 (n = 90 confirmed this pattern by priming independent and interdependent self-construals in liberals and conservatives. Results demonstrate the importance of cultural context for free speech. Findings suggest that in the U.S. support for free speech might be embraced for different reasons: conservatives’ support for free speech appears to be motivated by a focus on collectively held values favoring free speech, while liberals’ support for free speech might be motivated by a focus on individualist self-expression.
Thompson, Elaine C.; Carr, Kali Woodruff; White-Schwoch, Travis; Otto-Meyer, Sebastian; Kraus, Nina
2016-01-01
From bustling classrooms to unruly lunchrooms, school settings are noisy. To learn effectively in the unwelcome company of numerous distractions, children must clearly perceive speech in noise. In older children and adults, speech-in-noise perception is supported by sensory and cognitive processes, but the correlates underlying this critical listening skill in young children (3–5 year olds) remain undetermined. Employing a longitudinal design (two evaluations separated by ~12 months), we followed a cohort of 59 preschoolers, ages 3.0–4.9, assessing word-in-noise perception, cognitive abilities (intelligence, short-term memory, attention), and neural responses to speech. Results reveal changes in word-in-noise perception parallel changes in processing of the fundamental frequency (F0), an acoustic cue known for playing a role central to speaker identification and auditory scene analysis. Four unique developmental trajectories (speech-in-noise perception groups) confirm this relationship, in that improvements and declines in word-in-noise perception couple with enhancements and diminishments of F0 encoding, respectively. Improvements in word-in-noise perception also pair with gains in attention. Word-in-noise perception does not relate to strength of neural harmonic representation or short-term memory. These findings reinforce previously-reported roles of F0 and attention in hearing speech in noise in older children and adults, and extend this relationship to preschool children. PMID:27864051
Toward A Dual-Learning Systems Model of Speech Category Learning
Directory of Open Access Journals (Sweden)
Bharath eChandrasekaran
2014-07-01
Full Text Available More than two decades of work in vision posits the existence of dual-learning systems of category learning. The reflective system uses working memory to develop and test rules for classifying in an explicit fashion, while the reflexive system operates by implicitly associating perception with actions that lead to reinforcement. Dual-learning systems models hypothesize that in learning natural categories, learners initially use the reflective system and, with practice, transfer control to the reflexive system. The role of reflective and reflexive systems in auditory category learning and more specifically in speech category learning has not been systematically examined. In this article we describe a neurobiologically-constrained dual-learning systems theoretical framework that is currently being developed in speech category learning and review recent applications of this framework. Using behavioral and computational modeling approaches, we provide evidence that speech category learning is predominantly mediated by the reflexive learning system. In one application, we explore the effects of normal aging on non-speech and speech category learning. We find an age related deficit in reflective-optimal but not reflexive-optimal auditory category learning. Prominently, we find a large age-related deficit in speech learning. The computational modeling suggests that older adults are less likely to transition from simple, reflective, uni-dimensional rules to more complex, reflexive, multi-dimensional rules. In a second application we summarize a recent study examining auditory category learning in individuals with elevated depressive symptoms. We find a deficit in reflective-optimal and an enhancement in reflexive-optimal auditory category learning. Interestingly, individuals with elevated depressive symptoms also show an advantage in learning speech categories. We end with a brief summary and description of a number of future directions.
Integration of speech and gesture in aphasia.
Cocks, Naomi; Byrne, Suzanne; Pritchard, Madeleine; Morgan, Gary; Dipper, Lucy
2018-02-07
Information from speech and gesture is often integrated to comprehend a message. This integration process requires the appropriate allocation of cognitive resources to both the gesture and speech modalities. People with aphasia are likely to find integration of gesture and speech difficult. This is due to a reduction in cognitive resources, a difficulty with resource allocation or a combination of the two. Despite it being likely that people who have aphasia will have difficulty with integration, empirical evidence describing this difficulty is limited. Such a difficulty was found in a single case study by Cocks et al. in 2009, and is replicated here with a greater number of participants. To determine whether individuals with aphasia have difficulties understanding messages in which they have to integrate speech and gesture. Thirty-one participants with aphasia (PWA) and 30 control participants watched videos of an actor communicating a message in three different conditions: verbal only, gesture only, and verbal and gesture message combined. The message related to an action in which the name of the action (e.g., 'eat') was provided verbally and the manner of the action (e.g., hands in a position as though eating a burger) was provided gesturally. Participants then selected a picture that 'best matched' the message conveyed from a choice of four pictures which represented a gesture match only (G match), a verbal match only (V match), an integrated verbal-gesture match (Target) and an unrelated foil (UR). To determine the gain that participants obtained from integrating gesture and speech, a measure of multimodal gain (MMG) was calculated. The PWA were less able to integrate gesture and speech than the control participants and had significantly lower MMG scores. When the PWA had difficulty integrating, they more frequently selected the verbal match. The findings suggest that people with aphasia can have difficulty integrating speech and gesture in order to obtain
... Staying Safe Videos for Educators Search English Español Speech Problems KidsHealth / For Teens / Speech Problems What's in ... a person's ability to speak clearly. Some Common Speech and Language Disorders Stuttering is a problem that ...
O’Connor, Erin E.; Langer, David A.; Tompson, Martha C.
2017-01-01
Maternal depression is a well-documented risk factor for youth depression, and taking into account its severity and chronicity may provide important insight into the degree of risk conferred. This study explored the degree to which the severity/chronicity of maternal depression history explained variance in youth internalizing and externalizing symptoms above and beyond current maternal depressive symptoms among 171 youth (58% male) ages 8 to 12 over a span of three years. Severity and chronicity of past maternal depression and current maternal depressive symptoms were examined as predictors of parent-reported youth internalizing and externalizing symptomatology, as well as youth self-reported depressive symptoms. Severity and chronicity of past maternal depression did not account for additional variance in youth internalizing and externalizing symptoms at Time 1 beyond what was accounted for by maternal depressive symptoms at Time 1. Longitudinal growth curve modeling indicated that prior severity/chronicity of maternal depression predicted levels of youth internalizing and externalizing symptoms at each time point when controlling for current maternal depressive symptoms at each time point. Chronicity of maternal depression, apart from severity, also predicted rate of change in youth externalizing symptoms over time. These findings highlight the importance of screening and assessing for current maternal depressive symptoms, as well as the nature of past depressive episodes. Possible mechanisms underlying the association between severity/chronicity of maternal depression and youth outcomes, such as residual effects from depressive history on mother–child interactions, are discussed. PMID:27401880
The evolution of multivariate maternal effects.
Directory of Open Access Journals (Sweden)
Bram Kuijper
2014-04-01
Full Text Available There is a growing interest in predicting the social and ecological contexts that favor the evolution of maternal effects. Most predictions focus, however, on maternal effects that affect only a single character, whereas the evolution of maternal effects is poorly understood in the presence of suites of interacting traits. To overcome this, we simulate the evolution of multivariate maternal effects (captured by the matrix M in a fluctuating environment. We find that the rate of environmental fluctuations has a substantial effect on the properties of M: in slowly changing environments, offspring are selected to have a multivariate phenotype roughly similar to the maternal phenotype, so that M is characterized by positive dominant eigenvalues; by contrast, rapidly changing environments favor Ms with dominant eigenvalues that are negative, as offspring favor a phenotype which substantially differs from the maternal phenotype. Moreover, when fluctuating selection on one maternal character is temporally delayed relative to selection on other traits, we find a striking pattern of cross-trait maternal effects in which maternal characters influence not only the same character in offspring, but also other offspring characters. Additionally, when selection on one character contains more stochastic noise relative to selection on other traits, large cross-trait maternal effects evolve from those maternal traits that experience the smallest amounts of noise. The presence of these cross-trait maternal effects shows that individual maternal effects cannot be studied in isolation, and that their study in a multivariate context may provide important insights about the nature of past selection. Our results call for more studies that measure multivariate maternal effects in wild populations.
The evolution of multivariate maternal effects.
Kuijper, Bram; Johnstone, Rufus A; Townley, Stuart
2014-04-01
There is a growing interest in predicting the social and ecological contexts that favor the evolution of maternal effects. Most predictions focus, however, on maternal effects that affect only a single character, whereas the evolution of maternal effects is poorly understood in the presence of suites of interacting traits. To overcome this, we simulate the evolution of multivariate maternal effects (captured by the matrix M) in a fluctuating environment. We find that the rate of environmental fluctuations has a substantial effect on the properties of M: in slowly changing environments, offspring are selected to have a multivariate phenotype roughly similar to the maternal phenotype, so that M is characterized by positive dominant eigenvalues; by contrast, rapidly changing environments favor Ms with dominant eigenvalues that are negative, as offspring favor a phenotype which substantially differs from the maternal phenotype. Moreover, when fluctuating selection on one maternal character is temporally delayed relative to selection on other traits, we find a striking pattern of cross-trait maternal effects in which maternal characters influence not only the same character in offspring, but also other offspring characters. Additionally, when selection on one character contains more stochastic noise relative to selection on other traits, large cross-trait maternal effects evolve from those maternal traits that experience the smallest amounts of noise. The presence of these cross-trait maternal effects shows that individual maternal effects cannot be studied in isolation, and that their study in a multivariate context may provide important insights about the nature of past selection. Our results call for more studies that measure multivariate maternal effects in wild populations.
Alternative Speech Communication System for Persons with Severe Speech Disorders
Selouani, Sid-Ahmed; Sidi Yakoub, Mohammed; O'Shaughnessy, Douglas
2009-12-01
Assistive speech-enabled systems are proposed to help both French and English speaking persons with various speech disorders. The proposed assistive systems use automatic speech recognition (ASR) and speech synthesis in order to enhance the quality of communication. These systems aim at improving the intelligibility of pathologic speech making it as natural as possible and close to the original voice of the speaker. The resynthesized utterances use new basic units, a new concatenating algorithm and a grafting technique to correct the poorly pronounced phonemes. The ASR responses are uttered by the new speech synthesis system in order to convey an intelligible message to listeners. Experiments involving four American speakers with severe dysarthria and two Acadian French speakers with sound substitution disorders (SSDs) are carried out to demonstrate the efficiency of the proposed methods. An improvement of the Perceptual Evaluation of the Speech Quality (PESQ) value of 5% and more than 20% is achieved by the speech synthesis systems that deal with SSD and dysarthria, respectively.
Directory of Open Access Journals (Sweden)
Sunaina Seth
2015-11-01
Full Text Available High intrauterine cortisol exposure can inhibit fetal growth and have programming effects for the child’s subsequent stress reactivity. Placental 11beta-hydroxysteroid dehydrogenase (11β-HSD2 limits the amount of maternal cortisol transferred to the fetus. However, the relationship between maternal psychopathology and 11β-HSD2 remains poorly defined. This study examined the effect of maternal depressive disorder, antidepressant use and symptoms of depression and anxiety in pregnancy on placental 11β-HSD2 gene (HSD11B2 expression. Drawing on data from the Mercy Pregnancy and Emotional Wellbeing Study, placental HSD11B2 expression was compared among 33 pregnant women, who were selected based on membership of three groups; depressed (untreated, taking antidepressants and controls. Furthermore, associations between placental HSD11B2 and scores on the State-Trait Anxiety Inventory (STAI and Edinburgh Postnatal Depression Scale (EPDS during 12–18 and 28–34 weeks gestation were examined. Findings revealed negative correlations between HSD11B2 and both the EPDS and STAI (r = −0.11 to −0.28, with associations being particularly prominent during late gestation. Depressed and antidepressant exposed groups also displayed markedly lower placental HSD11B2 expression levels than controls. These findings suggest that maternal depression and anxiety may impact on fetal programming by down-regulating HSD11B2, and antidepressant treatment alone is unlikely to protect against this effect.
A Danish open-set speech corpus for competing-speech studies
DEFF Research Database (Denmark)
Nielsen, Jens Bo; Dau, Torsten; Neher, Tobias
2014-01-01
Studies investigating speech-on-speech masking effects commonly use closed-set speech materials such as the coordinate response measure [Bolia et al. (2000). J. Acoust. Soc. Am. 107, 1065-1066]. However, these studies typically result in very low (i.e., negative) speech recognition thresholds (SRTs......) when the competing speech signals are spatially separated. To achieve higher SRTs that correspond more closely to natural communication situations, an open-set, low-context, multi-talker speech corpus was developed. Three sets of 268 unique Danish sentences were created, and each set was recorded...... with one of three professional female talkers. The intelligibility of each sentence in the presence of speech-shaped noise was measured. For each talker, 200 approximately equally intelligible sentences were then selected and systematically distributed into 10 test lists. Test list homogeneity was assessed...
Directory of Open Access Journals (Sweden)
Stephen Gichuhi Kimotho
2016-06-01
Full Text Available Ethnicity in Kenya permeates all spheres of life. However, it is in politics that ethnicity is most visible. Election time in Kenya often leads to ethnic competition and hatred, often expressed through various media. Ethnic hate speech characterized the 2007 general elections in party rallies and through text messages, emails, posters and leaflets. This resulted in widespread skirmishes that left over 1200 people dead, and many displaced (KNHRC, 2008. In 2013, however, the new battle zone was the war of words on social media platform. More than any other time in Kenyan history, Kenyans poured vitriolic ethnic hate speech through digital media like Facebook, tweeter and blogs. Although scholars have studied the role and effects of the mainstream media like television and radio in proliferating the ethnic hate speech in Kenya (Michael Chege, 2008; Goldstein & Rotich, 2008a; Ismail & Deane, 2008; Jacqueline Klopp & Prisca Kamungi, 2007, little has been done in regard to social media. This paper investigated the nature of digitized hate speech by: describing the forms of ethnic hate speech on social media in Kenya; the effects of ethnic hate speech on Kenyan’s perception of ethnic entities; ethnic conflict and ethics of citizen journalism. This study adopted a descriptive interpretive design, and utilized Austin’s Speech Act Theory, which explains use of language to achieve desired purposes and direct behaviour (Tarhom & Miracle, 2013. Content published between January and April 2013 from six purposefully identified blogs was analysed. Questionnaires were used to collect data from university students as they form a good sample of Kenyan population, are most active on social media and are drawn from all parts of the country. Qualitative data were analysed using NVIVO 10 software, while responses from the questionnaire were analysed using IBM SPSS version 21. The findings indicated that Facebook and Twitter were the main platforms used to
Speech entrainment enables patients with Broca’s aphasia to produce fluent speech
Hubbard, H. Isabel; Hudspeth, Sarah Grace; Holland, Audrey L.; Bonilha, Leonardo; Fromm, Davida; Rorden, Chris
2012-01-01
A distinguishing feature of Broca’s aphasia is non-fluent halting speech typically involving one to three words per utterance. Yet, despite such profound impairments, some patients can mimic audio-visual speech stimuli enabling them to produce fluent speech in real time. We call this effect ‘speech entrainment’ and reveal its neural mechanism as well as explore its usefulness as a treatment for speech production in Broca’s aphasia. In Experiment 1, 13 patients with Broca’s aphasia were tested in three conditions: (i) speech entrainment with audio-visual feedback where they attempted to mimic a speaker whose mouth was seen on an iPod screen; (ii) speech entrainment with audio-only feedback where patients mimicked heard speech; and (iii) spontaneous speech where patients spoke freely about assigned topics. The patients produced a greater variety of words using audio-visual feedback compared with audio-only feedback and spontaneous speech. No difference was found between audio-only feedback and spontaneous speech. In Experiment 2, 10 of the 13 patients included in Experiment 1 and 20 control subjects underwent functional magnetic resonance imaging to determine the neural mechanism that supports speech entrainment. Group results with patients and controls revealed greater bilateral cortical activation for speech produced during speech entrainment compared with spontaneous speech at the junction of the anterior insula and Brodmann area 47, in Brodmann area 37, and unilaterally in the left middle temporal gyrus and the dorsal portion of Broca’s area. Probabilistic white matter tracts constructed for these regions in the normal subjects revealed a structural network connected via the corpus callosum and ventral fibres through the extreme capsule. Unilateral areas were connected via the arcuate fasciculus. In Experiment 3, all patients included in Experiment 1 participated in a 6-week treatment phase using speech entrainment to improve speech production
Visual feedback of tongue movement for novel speech sound learning
Directory of Open Access Journals (Sweden)
William F Katz
2015-11-01
Full Text Available Pronunciation training studies have yielded important information concerning the processing of audiovisual (AV information. Second language (L2 learners show increased reliance on bottom-up, multimodal input for speech perception (compared to monolingual individuals. However, little is known about the role of viewing one’s own speech articulation processes during speech training. The current study investigated whether real-time, visual feedback for tongue movement can improve a speaker’s learning of non-native speech sounds. An interactive 3D tongue visualization system based on electromagnetic articulography (EMA was used in a speech training experiment. Native speakers of American English produced a novel speech sound (/ɖ̠/; a voiced, coronal, palatal stop before, during, and after trials in which they viewed their own speech movements using the 3D model. Talkers’ productions were evaluated using kinematic (tongue-tip spatial positioning and acoustic (burst spectra measures. The results indicated a rapid gain in accuracy associated with visual feedback training. The findings are discussed with respect to neural models for multimodal speech processing.
Evidence-Based Speech-Language Pathology Practices in Schools: Findings from a National Survey
Hoffman, LaVae M.; Ireland, Marie; Hall-Mills, Shannon; Flynn, Perry
2013-01-01
Purpose: This study documented evidence-based practice (EBP) patterns as reported by speech-language pathologists (SLPs) employed in public schools during 2010-2011. Method: Using an online survey, practioners reported their EBP training experiences, resources available in their workplaces, and the frequency with which they engage in specific EBP…
Directory of Open Access Journals (Sweden)
Corianne Rogalsky
2014-04-01
Full Text Available The neural basis of speech perception has been debated for over a century. While it is generally agreed that the superior temporal lobes are critical for the perceptual analysis of speech, a major current topic is whether the motor system contributes to speech perception, with several conflicting findings attested. In a dorsal-ventral speech stream framework (Hickok & Poeppel 2007, this debate is essentially about the roles of the dorsal versus ventral speech processing streams. A major roadblock in characterizing the neuroanatomy of speech perception is task-specific effects. For example, much of the evidence for dorsal stream involvement comes from syllable discrimination type tasks, which have been found to behaviorally doubly dissociate from auditory comprehension tasks (Baker et al. 1981. Discrimination task deficits could be a result of difficulty perceiving the sounds themselves, which is the typical assumption, or it could be a result of failures in temporary maintenance of the sensory traces, or the comparison and/or the decision process. Similar complications arise in perceiving sentences: the extent of inferior frontal (i.e. dorsal stream activation during listening to sentences increases as a function of increased task demands (Love et al. 2006. Another complication is the stimulus: much evidence for dorsal stream involvement uses speech samples lacking semantic context (CVs, non-words. The present study addresses these issues in a large-scale lesion-symptom mapping study. 158 patients with focal cerebral lesions from the Mutli-site Aphasia Research Consortium underwent a structural MRI or CT scan, as well as an extensive psycholinguistic battery. Voxel-based lesion symptom mapping was used to compare the neuroanatomy involved in the following speech perception tasks with varying phonological, semantic, and task loads: (i two discrimination tasks of syllables (non-words and words, respectively, (ii two auditory comprehension tasks
Aging and Spectro-Temporal Integration of Speech
Directory of Open Access Journals (Sweden)
John H. Grose
2016-10-01
Full Text Available The purpose of this study was to determine the effects of age on the spectro-temporal integration of speech. The hypothesis was that the integration of speech fragments distributed over frequency, time, and ear of presentation is reduced in older listeners—even for those with good audiometric hearing. Younger, middle-aged, and older listeners (10 per group with good audiometric hearing participated. They were each tested under seven conditions that encompassed combinations of spectral, temporal, and binaural integration. Sentences were filtered into two bands centered at 500 Hz and 2500 Hz, with criterion bandwidth tailored for each participant. In some conditions, the speech bands were individually square wave interrupted at a rate of 10 Hz. Configurations of uninterrupted, synchronously interrupted, and asynchronously interrupted frequency bands were constructed that constituted speech fragments distributed across frequency, time, and ear of presentation. The over-arching finding was that, for most configurations, performance was not differentially affected by listener age. Although speech intelligibility varied across condition, there was no evidence of performance deficits in older listeners in any condition. This study indicates that age, per se, does not necessarily undermine the ability to integrate fragments of speech dispersed across frequency and time.
Electrophysiological assessment of audiovisual integration in speech perception
DEFF Research Database (Denmark)
Eskelund, Kasper; Dau, Torsten
Speech perception integrates signal from ear and eye. This is witnessed by a wide range of audiovisual integration effects, such as ventriloquism and the McGurk illusion. Some behavioral evidence suggest that audiovisual integration of specific aspects is special for speech perception. However, our...... knowledge of such bimodal integration would be strengthened if the phenomena could be investigated by objective, neutrally based methods. One key question of the present work is if perceptual processing of audiovisual speech can be gauged with a specific signature of neurophysiological activity...... on the auditory speech percept? In two experiments, which both combine behavioral and neurophysiological measures, an uncovering of the relation between perception of faces and of audiovisual integration is attempted. Behavioral findings suggest a strong effect of face perception, whereas the MMN results are less...
Directory of Open Access Journals (Sweden)
Seth-Reino Ekström
2011-01-01
Full Text Available The masking effect of a piano composition, played at different speeds and in different octaves, on speech-perception thresholds was investigated in 15 normal-hearing and 14 moderately-hearing-impaired subjects. Running speech (just follow conversation, JFC testing and use of hearing aids increased the everyday validity of the findings. A comparison was made with standard audiometric noises [International Collegium of Rehabilitative Audiology (ICRA noise and speech spectrum-filtered noise (SPN]. All masking sounds, music or noise, were presented at the same equivalent sound level (50 dBA. The results showed a significant effect of piano performance speed and octave (P<.01. Low octave and fast tempo had the largest effect; and high octave and slow tempo, the smallest. Music had a lower masking effect than did ICRA noise with two or six speakers at normal vocal effort (P<.01 and SPN (P<.05. Subjects with hearing loss had higher masked thresholds than the normal-hearing subjects (P<.01, but there were smaller differences between masking conditions (P<.01. It is pointed out that music offers an interesting opportunity for studying masking under realistic conditions, where spectral and temporal features can be varied independently. The results have implications for composing music with vocal parts, designing acoustic environments and creating a balance between speech perception and privacy in social settings.
Multimodal Speech Capture System for Speech Rehabilitation and Learning.
Sebkhi, Nordine; Desai, Dhyey; Islam, Mohammad; Lu, Jun; Wilson, Kimberly; Ghovanloo, Maysam
2017-11-01
Speech-language pathologists (SLPs) are trained to correct articulation of people diagnosed with motor speech disorders by analyzing articulators' motion and assessing speech outcome while patients speak. To assist SLPs in this task, we are presenting the multimodal speech capture system (MSCS) that records and displays kinematics of key speech articulators, the tongue and lips, along with voice, using unobtrusive methods. Collected speech modalities, tongue motion, lips gestures, and voice are visualized not only in real-time to provide patients with instant feedback but also offline to allow SLPs to perform post-analysis of articulators' motion, particularly the tongue, with its prominent but hardly visible role in articulation. We describe the MSCS hardware and software components, and demonstrate its basic visualization capabilities by a healthy individual repeating the words "Hello World." A proof-of-concept prototype has been successfully developed for this purpose, and will be used in future clinical studies to evaluate its potential impact on accelerating speech rehabilitation by enabling patients to speak naturally. Pattern matching algorithms to be applied to the collected data can provide patients with quantitative and objective feedback on their speech performance, unlike current methods that are mostly subjective, and may vary from one SLP to another.
van Lieshout, Pascal H. H. M.; Bose, Arpita; Square, Paula A.; Steele, Catriona M.
2007-01-01
Apraxia of speech (AOS) is typically described as a motor-speech disorder with clinically well-defined symptoms, but without a clear understanding of the underlying problems in motor control. A number of studies have compared the speech of subjects with AOS to the fluent speech of controls, but only a few have included speech movement data and if…
Nakayama, Masataka; Saito, Satoru
2015-08-01
The present study investigated principles of phonological planning, a common serial ordering mechanism for speech production and phonological short-term memory. Nakayama and Saito (2014) have investigated the principles by using a speech-error induction technique, in which participants were exposed to an auditory distracIor word immediately before an utterance of a target word. They demonstrated within-word adjacent mora exchanges and serial position effects on error rates. These findings support, respectively, the temporal distance and the edge principles at a within-word level. As this previous study induced errors using word distractors created by exchanging adjacent morae in the target words, it is possible that the speech errors are expressions of lexical intrusions reflecting interactive activation of phonological and lexical/semantic representations. To eliminate this possibility, the present study used nonword distractors that had no lexical or semantic representations. This approach successfully replicated the error patterns identified in the abovementioned study, further confirming that the temporal distance and edge principles are organizing precepts in phonological planning.
Maternal employment and childhood overweight in Germany.
Meyer, Sophie-Charlotte
2016-12-01
A widespread finding among studies from the US and the UK is that maternal employment is correlated with an increased risk of child overweight, even in a causal manner, whereas studies from other countries obtain less conclusive results. As evidence for Germany is still scarce, the purpose of this study is to identify the effect of maternal employment on childhood overweight in Germany using two sets of representative micro data. We further explore potential underlying mechanisms that might explain this relationship. In order to address the selection into maternal full-time employment, we use an instrumental variable strategy exploiting the number of younger siblings in the household as an instrument. While the OLS models suggest that maternal full-time employment is related to a 5 percentage point higher probability of the child to be overweight, IV estimates indicate a 25 percentage points higher overweight probability due to maternal full-time employment. Exploring various possible pathways, we find that maternal full-time employment promotes unhealthy dietary and activity behavior which might explain the positive effect of maternal employment on child overweight to some extent. Although there are limitations to our IV approach, several sensitivity analyses confirm the robustness of our findings. Copyright © 2016 Elsevier B.V. All rights reserved.
New tests of the distal speech rate effect: Examining cross-linguistic generalization
Directory of Open Access Journals (Sweden)
Laura eDilley
2013-12-01
Full Text Available Recent findings [Dilley and Pitt, 2010. Psych. Science. 21, 1664-1670] have shown that manipulating context speech rate in English can cause entire syllables to disappear or appear perceptually. The current studies tested two rate-based explanations of this phenomenon while attempting to replicate and extend these findings to another language, Russian. In Experiment 1, native Russian speakers listened to Russian sentences which had been subjected to rate manipulations and performed a lexical report task. Experiment 2 investigated speech rate effects in cross-language speech perception; non-native speakers of Russian of both high and low proficiency were tested on the same Russian sentences as in Experiment 1. They decided between two lexical interpretations of a critical portion of the sentence, where one choice contained more phonological material than the other (e.g., /stərʌ'na/ side vs. /strʌ'na/ country. In both experiments, with native and non-native speakers of Russian, context speech rate and the relative duration of the critical sentence portion were found to influence the amount of phonological material perceived. The results support the generalized rate normalization hypothesis, according to which the content perceived in a spectrally ambiguous stretch of speech depends on the duration of that content relative to the surrounding speech, while showing that the findings of Dilley and Pitt (2010 extend to a variety of morphosyntactic contexts and a new language, Russian. Findings indicate that relative timing cues across an utterance can be critical to accurate lexical perception by both native and non-native speakers.
Epidemiology of Maternal Mortality in Malawi
African Journals Online (AJOL)
live births. Causes and determinants of maternal mortal- ity. Global causes of maternal mortality. Across the globe the causes of maternal deaths are strik- ..... at home”. Findings from Thyolo, Mangochi and Chik- wawa were similar". Perceived qua/ity of care. Like anywhere in the world, the perceived quality of care in ...
Hullett, Patrick W; Hamilton, Liberty S; Mesgarani, Nima; Schreiner, Christoph E; Chang, Edward F
2016-02-10
The human superior temporal gyrus (STG) is critical for speech perception, yet the organization of spectrotemporal processing of speech within the STG is not well understood. Here, to characterize the spatial organization of spectrotemporal processing of speech across human STG, we use high-density cortical surface field potential recordings while participants listened to natural continuous speech. While synthetic broad-band stimuli did not yield sustained activation of the STG, spectrotemporal receptive fields could be reconstructed from vigorous responses to speech stimuli. We find that the human STG displays a robust anterior-posterior spatial distribution of spectrotemporal tuning in which the posterior STG is tuned for temporally fast varying speech sounds that have relatively constant energy across the frequency axis (low spectral modulation) while the anterior STG is tuned for temporally slow varying speech sounds that have a high degree of spectral variation across the frequency axis (high spectral modulation). This work illustrates organization of spectrotemporal processing in the human STG, and illuminates processing of ethologically relevant speech signals in a region of the brain specialized for speech perception. Considerable evidence has implicated the human superior temporal gyrus (STG) in speech processing. However, the gross organization of spectrotemporal processing of speech within the STG is not well characterized. Here we use natural speech stimuli and advanced receptive field characterization methods to show that spectrotemporal features within speech are well organized along the posterior-to-anterior axis of the human STG. These findings demonstrate robust functional organization based on spectrotemporal modulation content, and illustrate that much of the encoded information in the STG represents the physical acoustic properties of speech stimuli. Copyright © 2016 the authors 0270-6474/16/362014-13$15.00/0.
Out-of-synchrony speech entrainment in developmental dyslexia.
Molinaro, Nicola; Lizarazu, Mikel; Lallier, Marie; Bourguignon, Mathieu; Carreiras, Manuel
2016-08-01
Developmental dyslexia is a reading disorder often characterized by reduced awareness of speech units. Whether the neural source of this phonological disorder in dyslexic readers results from the malfunctioning of the primary auditory system or damaged feedback communication between higher-order phonological regions (i.e., left inferior frontal regions) and the auditory cortex is still under dispute. Here we recorded magnetoencephalographic (MEG) signals from 20 dyslexic readers and 20 age-matched controls while they were listening to ∼10-s-long spoken sentences. Compared to controls, dyslexic readers had (1) an impaired neural entrainment to speech in the delta band (0.5-1 Hz); (2) a reduced delta synchronization in both the right auditory cortex and the left inferior frontal gyrus; and (3) an impaired feedforward functional coupling between neural oscillations in the right auditory cortex and the left inferior frontal regions. This shows that during speech listening, individuals with developmental dyslexia present reduced neural synchrony to low-frequency speech oscillations in primary auditory regions that hinders higher-order speech processing steps. The present findings, thus, strengthen proposals assuming that improper low-frequency acoustic entrainment affects speech sampling. This low speech-brain synchronization has the strong potential to cause severe consequences for both phonological and reading skills. Interestingly, the reduced speech-brain synchronization in dyslexic readers compared to normal readers (and its higher-order consequences across the speech processing network) appears preserved through the development from childhood to adulthood. Thus, the evaluation of speech-brain synchronization could possibly serve as a diagnostic tool for early detection of children at risk of dyslexia. Hum Brain Mapp 37:2767-2783, 2016. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
Altvater-Mackensen, Nicole; Mani, Nivedita; Grossmann, Tobias
2016-01-01
Recent studies suggest that infants' audiovisual speech perception is influenced by articulatory experience (Mugitani et al., 2008; Yeung & Werker, 2013). The current study extends these findings by testing if infants' emerging ability to produce native sounds in babbling impacts their audiovisual speech perception. We tested 44 6-month-olds…
Causes of Speech Disorders in Primary School Students of Zahedan
Directory of Open Access Journals (Sweden)
Saeed Fakhrerahimi
2013-02-01
Full Text Available Background: Since making communication with others is the most important function of speech, undoubtedly, any type of disorder in speech will affect the human communicability with others. The objective of the study was to investigate reasons behind the [high] prevalence rate of stammer, producing disorders and aglossia.Materials and Methods: This descriptive-analytical study was conducted on 118 male and female students, who were studying in a primary school in Zahedan; they had referred to the Speech Therapy Centers of Zahedan University of Medical Sciences in a period of seven months. The speech therapist examinations, diagnosis tools common in speech therapy, Spielberg Children Trait and also patients' cases were used to find the reasons behind the [high] prevalence rate of speech disorders. Results: Psychological causes had the highest rate of correlation with the speech disorders among the other factors affecting the speech disorders. After psychological causes, family history and age of the subjects are the other factors which may bring about the speech disorders (P<0.05. Bilingualism and birth order has a negative relationship with the speech disorders. Likewise, another result of this study shows that only psychological causes, social causes, hereditary causes and age of subjects can predict the speech disorders (P<0.05.Conclusion: The present study shows that the speech disorders have a strong and close relationship with the psychological causes at the first step and also history of family and age of individuals at the next steps.
Fast Monaural Separation of Speech
DEFF Research Database (Denmark)
Pontoppidan, Niels Henrik; Dyrholm, Mads
2003-01-01
a Factorial Hidden Markov Model, with non-stationary assumptions on the source autocorrelations modelled through the Factorial Hidden Markov Model, leads to separation in the monaural case. By extending Hansens work we find that Roweis' assumptions are necessary for monaural speech separation. Furthermore we...
The Speech Act Theory between Linguistics and Language Philosophy
Directory of Open Access Journals (Sweden)
Liviu-Mihail MARINESCU
2006-10-01
Full Text Available Of all the issues in the general theory of language usage, speech act theory has probably aroused the widest interest. Psychologists, forexample, have suggested that the acquisition of the concepts underlying speech acts may be a prerequisite for the acquisition of language in general,literary critics have looked to speech act theory for an illumination of textual subtleties or for an understanding of the nature of literary genres,anthropologists have hoped to find in the theory some account of the nature of magical incantations, philosophers have seen potential applications to,amongst other things, the status of ethical statements, while linguists have seen the notions of speech act theory as variously applicable to problemsin syntax, semantics, second language learning, and elsewhere.
Private Speech Moderates the Effects of Effortful Control on Emotionality
Day, Kimberly L.; Smith, Cynthia L.; Neal, Amy; Dunsmore, Julie C.
2018-01-01
Research Findings: In addition to being a regulatory strategy, children's private speech may enhance or interfere with their effortful control used to regulate emotion. The goal of the current study was to investigate whether children's private speech during a selective attention task moderated the relations of their effortful control to their…
Inner Speech and Clarity of Self-Concept in Thought Disorder and Auditory-Verbal Hallucinations.
de Sousa, Paulo; Sellwood, William; Spray, Amy; Fernyhough, Charles; Bentall, Richard P
2016-12-01
Eighty patients and thirty controls were interviewed using one interview that promoted personal disclosure and another about everyday topics. Speech was scored using the Thought, Language and Communication scale (TLC). All participants completed the Self-Concept Clarity Scale (SCCS) and the Varieties of Inner Speech Questionnaire (VISQ). Patients scored lower than comparisons on the SCCS. Low scores were associated the disorganized dimension of TD. Patients also scored significantly higher on condensed and other people in inner speech, but not on dialogical or evaluative inner speech. The poverty of speech dimension of TD was associated with less dialogical inner speech, other people in inner speech, and less evaluative inner speech. Hallucinations were significantly associated with more other people in inner speech and evaluative inner speech. Clarity of self-concept and qualities of inner speech are differentially associated with dimensions of TD. The findings also support inner speech models of hallucinations.
Does prenatal care benefit maternal health? A study of post-partum maternal care use.
Liu, Tsai-Ching; Chen, Bradley; Chan, Yun-Shan; Chen, Chin-Shyan
2015-10-01
Most studies on prenatal care focus on its effects on infant health, while studying less about the effects on maternal health. Using the Longitudinal Health Insurance claims data in Taiwan in a recursive bivariate probit model, this study examines the impact of adequate prenatal care on the probability of post-partum maternal hospitalization during the first 6 months after birth. The results show that adequate prenatal care significantly reduces the probability of post-partum maternal hospitalization among women who have had vaginal delivery by 43.8%. This finding suggests that the benefits of prenatal care may have been underestimated among women with vaginal delivery. Timely and adequate prenatal care not only creates a positive impact on infant health, but also yields significant benefits for post-partum maternal health. However, we do not find similar benefits of prenatal care for women undergoing a cesarean section. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Enhancement of speech signals - with a focus on voiced speech models
DEFF Research Database (Denmark)
Nørholm, Sidsel Marie
This thesis deals with speech enhancement, i.e., noise reduction in speech signals. This has applications in, e.g., hearing aids and teleconference systems. We consider a signal-driven approach to speech enhancement where a model of the speech is assumed and filters are generated based...... on this model. The basic model used in this thesis is the harmonic model which is a commonly used model for describing the voiced part of the speech signal. We show that it can be beneficial to extend the model to take inharmonicities or the non-stationarity of speech into account. Extending the model...
Intelligibility for Binaural Speech with Discarded Low-SNR Speech Components.
Schoenmaker, Esther; van de Par, Steven
2016-01-01
Speech intelligibility in multitalker settings improves when the target speaker is spatially separated from the interfering speakers. A factor that may contribute to this improvement is the improved detectability of target-speech components due to binaural interaction in analogy to the Binaural Masking Level Difference (BMLD). This would allow listeners to hear target speech components within specific time-frequency intervals that have a negative SNR, similar to the improvement in the detectability of a tone in noise when these contain disparate interaural difference cues. To investigate whether these negative-SNR target-speech components indeed contribute to speech intelligibility, a stimulus manipulation was performed where all target components were removed when local SNRs were smaller than a certain criterion value. It can be expected that for sufficiently high criterion values target speech components will be removed that do contribute to speech intelligibility. For spatially separated speakers, assuming that a BMLD-like detection advantage contributes to intelligibility, degradation in intelligibility is expected already at criterion values below 0 dB SNR. However, for collocated speakers it is expected that higher criterion values can be applied without impairing speech intelligibility. Results show that degradation of intelligibility for separated speakers is only seen for criterion values of 0 dB and above, indicating a negligible contribution of a BMLD-like detection advantage in multitalker settings. These results show that the spatial benefit is related to a spatial separation of speech components at positive local SNRs rather than to a BMLD-like detection improvement for speech components at negative local SNRs.
An experimental Dutch keyboard-to-speech system for the speech impaired
Deliege, R.J.H.
1989-01-01
An experimental Dutch keyboard-to-speech system has been developed to explor the possibilities and limitations of Dutch speech synthesis in a communication aid for the speech impaired. The system uses diphones and a formant synthesizer chip for speech synthesis. Input to the system is in
Low-Frequency Cortical Entrainment to Speech Reflects Phoneme-Level Processing.
Di Liberto, Giovanni M; O'Sullivan, James A; Lalor, Edmund C
2015-10-05
The human ability to understand speech is underpinned by a hierarchical auditory system whose successive stages process increasingly complex attributes of the acoustic input. It has been suggested that to produce categorical speech perception, this system must elicit consistent neural responses to speech tokens (e.g., phonemes) despite variations in their acoustics. Here, using electroencephalography (EEG), we provide evidence for this categorical phoneme-level speech processing by showing that the relationship between continuous speech and neural activity is best described when that speech is represented using both low-level spectrotemporal information and categorical labeling of phonetic features. Furthermore, the mapping between phonemes and EEG becomes more discriminative for phonetic features at longer latencies, in line with what one might expect from a hierarchical system. Importantly, these effects are not seen for time-reversed speech. These findings may form the basis for future research on natural language processing in specific cohorts of interest and for broader insights into how brains transform acoustic input into meaning. Copyright © 2015 Elsevier Ltd. All rights reserved.
Maternal employment and early childhood overweight: findings from the UK Millennium Cohort Study
Hawkins, S. S.; Cole, T. J.; Law, C.; Millennium Cohort Study Child Health Group, The
2008-01-01
Background: In most developed countries, maternal employment has increased rapidly. Changing patterns of family life have been suggested to be contributing to the rising prevalence of childhood obesity. Objectives: Our primary objective was to examine the relationship between maternal and partner employment and overweight in children aged 3 years. Our secondary objective was to investigate factors related to early childhood overweight only among mothers in employment. Design: Cohort study. Su...
Maternal employment and early childhood overweight: findings from the UK Millennium Cohort Study
Hawkins, S. S.; Cole, T. J.; Law, C.; Millennium Cohort Study Child Hlth
2008-01-01
Background: In most developed countries, maternal employment has increased rapidly. Changing patterns of family life have been suggested to be contributing to the rising prevalence of childhood obesity.Objectives: Our primary objective was to examine the relationship between maternal and partner employment and overweight in children aged 3 years. Our secondary objective was to investigate factors related to early childhood overweight only among mothers in employment.Design: Cohort study.Subje...
Speech Function and Speech Role in Carl Fredricksen's Dialogue on Up Movie
Rehana, Ridha; Silitonga, Sortha
2013-01-01
One aim of this article is to show through a concrete example how speech function and speech role used in movie. The illustrative example is taken from the dialogue of Up movie. Central to the analysis proper form of dialogue on Up movie that contain of speech function and speech role; i.e. statement, offer, question, command, giving, and demanding. 269 dialogue were interpreted by actor, and it was found that the use of speech function and speech role.
Cross-Cultural Variation of Politeness Orientation & Speech Act Perception
Directory of Open Access Journals (Sweden)
Nisreen Naji Al-Khawaldeh
2013-05-01
Full Text Available This paper presents the findings of an empirical study which compares Jordanian and English native speakers’ perceptions about the speech act of thanking. The forty interviews conducted revealed some similarities but also of remarkable cross-cultural differences relating to the significance of thanking, the variables affecting it, and the appropriate linguistic and paralinguistic choices, as well as their impact on the interpretation of thanking behaviour. The most important theoretical finding is that the data, while consistent with many views found in the existing literature, do not support Brown and Levinson’s (1987 claim that thanking is a speech act which intrinsically threatens the speaker’s negative face because it involves overt acceptance of an imposition on the speaker. Rather, thanking should be viewed as a means of establishing and sustaining social relationships. The study findings suggest that cultural variation in thanking is due to the high degree of sensitivity of this speech act to the complex interplay of a range of social and contextual variables, and point to some promising directions for further research.
Optimizing acoustical conditions for speech intelligibility in classrooms
Yang, Wonyoung
High speech intelligibility is imperative in classrooms where verbal communication is critical. However, the optimal acoustical conditions to achieve a high degree of speech intelligibility have previously been investigated with inconsistent results, and practical room-acoustical solutions to optimize the acoustical conditions for speech intelligibility have not been developed. This experimental study validated auralization for speech-intelligibility testing, investigated the optimal reverberation for speech intelligibility for both normal and hearing-impaired listeners using more realistic room-acoustical models, and proposed an optimal sound-control design for speech intelligibility based on the findings. The auralization technique was used to perform subjective speech-intelligibility tests. The validation study, comparing auralization results with those of real classroom speech-intelligibility tests, found that if the room to be auralized is not very absorptive or noisy, speech-intelligibility tests using auralization are valid. The speech-intelligibility tests were done in two different auralized sound fields---approximately diffuse and non-diffuse---using the Modified Rhyme Test and both normal and hearing-impaired listeners. A hybrid room-acoustical prediction program was used throughout the work, and it and a 1/8 scale-model classroom were used to evaluate the effects of ceiling barriers and reflectors. For both subject groups, in approximately diffuse sound fields, when the speech source was closer to the listener than the noise source, the optimal reverberation time was zero. When the noise source was closer to the listener than the speech source, the optimal reverberation time was 0.4 s (with another peak at 0.0 s) with relative output power levels of the speech and noise sources SNS = 5 dB, and 0.8 s with SNS = 0 dB. In non-diffuse sound fields, when the noise source was between the speaker and the listener, the optimal reverberation time was 0.6 s with
Nooteboom, S.G.; Quené, H.
2008-01-01
This paper reports two experiments designed to investigate whether lexical bias in phonological speech errors is caused by immediate feedback of activation, by self-monitoring of inner speech, or by both. The experiments test a number of predictions derived from a model of self-monitoring of inner
Larm, Petra; Hongisto, Valtteri
2006-02-01
During the acoustical design of, e.g., auditoria or open-plan offices, it is important to know how speech can be perceived in various parts of the room. Different objective methods have been developed to measure and predict speech intelligibility, and these have been extensively used in various spaces. In this study, two such methods were compared, the speech transmission index (STI) and the speech intelligibility index (SII). Also the simplification of the STI, the room acoustics speech transmission index (RASTI), was considered. These quantities are all based on determining an apparent speech-to-noise ratio on selected frequency bands and summing them using a specific weighting. For comparison, some data were needed on the possible differences of these methods resulting from the calculation scheme and also measuring equipment. Their prediction accuracy was also of interest. Measurements were made in a laboratory having adjustable noise level and absorption, and in a real auditorium. It was found that the measurement equipment, especially the selection of the loudspeaker, can greatly affect the accuracy of the results. The prediction accuracy of the RASTI was found acceptable, if the input values for the prediction are accurately known, even though the studied space was not ideally diffuse.
Problems in Translating Figures of Speech: A Review of Persian Translations of Harry Potter Series
Directory of Open Access Journals (Sweden)
Fatemeh Masroor
2016-12-01
Full Text Available Due to the important role of figures of speech in prose, the present research tried to investigate the figures of speech in the novel, Harry Potter Series, and their Persian translations. The main goal of this research was to investigate the translators’ problems in translating figures of speech from English into Persian. To achieve this goal, the collected data were analyzed and compared with their Persian equivalents. Then, the theories of Newmark (1988 & 2001, Larson (1998, and Nolan (2005 were used in order to find the applied strategies for rendering the figures of speech by the translators. After identifying the applied translation strategies, the descriptive and inferential analyses were applied to answer the research question and test its related hypothesis. The results confirmed that the most common pitfalls in translating figures of speech from English into Persian based on Nolan (2005 were, not identifying of figures of speech, their related meanings and translating them literally. Overall, the research findings rejected the null hypothesis. The findings of present research can be useful for translators, especially beginners. They can be aware of the existing problems in translating figures of speech, so they can avoid committing the same mistakes in their works.
Long-term temporal tracking of speech rate affects spoken-word recognition.
Baese-Berk, Melissa M; Heffner, Christopher C; Dilley, Laura C; Pitt, Mark A; Morrill, Tuuli H; McAuley, J Devin
2014-08-01
Humans unconsciously track a wide array of distributional characteristics in their sensory environment. Recent research in spoken-language processing has demonstrated that the speech rate surrounding a target region within an utterance influences which words, and how many words, listeners hear later in that utterance. On the basis of hypotheses that listeners track timing information in speech over long timescales, we investigated the possibility that the perception of words is sensitive to speech rate over such a timescale (e.g., an extended conversation). Results demonstrated that listeners tracked variation in the overall pace of speech over an extended duration (analogous to that of a conversation that listeners might have outside the lab) and that this global speech rate influenced which words listeners reported hearing. The effects of speech rate became stronger over time. Our findings are consistent with the hypothesis that neural entrainment by speech occurs on multiple timescales, some lasting more than an hour. © The Author(s) 2014.
Shah, Rupal; Nahar, Quamrun; Gurley, Emily S
2016-03-01
We estimated the proportion of maternal deaths in Bangladesh associated with acute onset of jaundice. We used verbal autopsy data from a nationally representative maternal mortality survey to calculate the proportion of maternal deaths associated with jaundice and compared it to previously published estimates. Of all maternal deaths between 2008 and 2010, 23% were associated with jaundice, compared with 19% from 1998 to 2001. Approximately one of five maternal deaths was preceded by jaundice, unchanged in 10 years. Our findings highlight the need to better understand the etiology of these maternal deaths in Bangladesh. © The American Society of Tropical Medicine and Hygiene.
Perceived Speech Privacy in Computer Simulated Open-plan Offices
DEFF Research Database (Denmark)
Pop, Claudiu B.; Rindel, Jens Holger
2005-01-01
In open plan offices the lack of speech privacy between the workstations is one of the major acoustic problems. Improving the speech privacy in an open plan design is therefore the main concern for a successful open plan environment. The project described in this paper aimed to find an objective...... parameter that correlates well with the perceived degree of speech privacy and to derive a clear method for evaluating the acoustic conditions in open plan offices. Acoustic measurements were carried out in an open plan office, followed by data analysis at the Acoustic Department, DTU. A computer model...
Markers of Deception in Italian Speech
Directory of Open Access Journals (Sweden)
Katelyn eSpence
2012-10-01
Full Text Available Lying is a universal activity and the detection of lying a universal concern. Presently, there is great interest in determining objective measures of deception. The examination of speech, in particular, holds promise in this regard; yet, most of what we know about the relationship between speech and lying is based on the assessment of English-speaking participants. Few studies have examined indicators of deception in languages other than English. The world’s languages differ in significant ways, and cross-linguistic studies of deceptive communications are a research imperative. Here we review some of these differences amongst the world’s languages, and provide an overview of a number of recent studies demonstrating that cross-linguistic research is a worthwhile endeavour. In addition, we report the results of an empirical investigation of pitch, response latency, and speech rate as cues to deception in Italian speech. True and false opinions were elicited in an audio-taped interview. A within subjects analysis revealed no significant difference between the average pitch of the two conditions; however, speech rate was significantly slower, while response latency was longer, during deception compared with truth-telling. We explore the implications of these findings and propose directions for future research, with the aim of expanding the cross-linguistic branch of research on markers of deception.
Inner Speech and Clarity of Self-Concept in Thought Disorder and Auditory-Verbal Hallucinations
de Sousa, Paulo; Sellwood, William; Spray, Amy; Fernyhough, Charles; Bentall, Richard P.
2016-01-01
Abstract Eighty patients and thirty controls were interviewed using one interview that promoted personal disclosure and another about everyday topics. Speech was scored using the Thought, Language and Communication scale (TLC). All participants completed the Self-Concept Clarity Scale (SCCS) and the Varieties of Inner Speech Questionnaire (VISQ). Patients scored lower than comparisons on the SCCS. Low scores were associated the disorganized dimension of TD. Patients also scored significantly higher on condensed and other people in inner speech, but not on dialogical or evaluative inner speech. The poverty of speech dimension of TD was associated with less dialogical inner speech, other people in inner speech, and less evaluative inner speech. Hallucinations were significantly associated with more other people in inner speech and evaluative inner speech. Clarity of self-concept and qualities of inner speech are differentially associated with dimensions of TD. The findings also support inner speech models of hallucinations. PMID:27898489
Robust Speech/Non-Speech Classification in Heterogeneous Multimedia Content
Huijbregts, M.A.H.; de Jong, Franciska M.G.
In this paper we present a speech/non-speech classification method that allows high quality classification without the need to know in advance what kinds of audible non-speech events are present in an audio recording and that does not require a single parameter to be tuned on in-domain data. Because
Relationship between individual differences in speech processing and cognitive functions.
Ou, Jinghua; Law, Sam-Po; Fung, Roxana
2015-12-01
A growing body of research has suggested that cognitive abilities may play a role in individual differences in speech processing. The present study took advantage of a widespread linguistic phenomenon of sound change to systematically assess the relationships between speech processing and various components of attention and working memory in the auditory and visual modalities among typically developed Cantonese-speaking individuals. The individual variations in speech processing are captured in an ongoing sound change-tone merging in Hong Kong Cantonese, in which typically developed native speakers are reported to lose the distinctions between some tonal contrasts in perception and/or production. Three groups of participants were recruited, with a first group of good perception and production, a second group of good perception but poor production, and a third group of good production but poor perception. Our findings revealed that modality-independent abilities of attentional switching/control and working memory might contribute to individual differences in patterns of speech perception and production as well as discrimination latencies among typically developed speakers. The findings not only have the potential to generalize to speech processing in other languages, but also broaden our understanding of the omnipresent phenomenon of language change in all languages.
Beyond production: Brain responses during speech perception in adults who stutter
Directory of Open Access Journals (Sweden)
Tali Halag-Milo
2016-01-01
Full Text Available Developmental stuttering is a speech disorder that disrupts the ability to produce speech fluently. While stuttering is typically diagnosed based on one's behavior during speech production, some models suggest that it involves more central representations of language, and thus may affect language perception as well. Here we tested the hypothesis that developmental stuttering implicates neural systems involved in language perception, in a task that manipulates comprehensibility without an overt speech production component. We used functional magnetic resonance imaging to measure blood oxygenation level dependent (BOLD signals in adults who do and do not stutter, while they were engaged in an incidental speech perception task. We found that speech perception evokes stronger activation in adults who stutter (AWS compared to controls, specifically in the right inferior frontal gyrus (RIFG and in left Heschl's gyrus (LHG. Significant differences were additionally found in the lateralization of response in the inferior frontal cortex: AWS showed bilateral inferior frontal activity, while controls showed a left lateralized pattern of activation. These findings suggest that developmental stuttering is associated with an imbalanced neural network for speech processing, which is not limited to speech production, but also affects cortical responses during speech perception.
Association between Maternal Smoking during Pregnancy and Low Birthweight: Effects by Maternal Age.
Directory of Open Access Journals (Sweden)
Wei Zheng
Full Text Available Maternal smoking during pregnancy has been consistently related to low birthweight. However, older mothers, who are already at risk of giving birth to low birthweight infants, might be even more susceptible to the effects of maternal smoking. Therefore, this study aimed to examine the modified association between maternal smoking and low birthweight by maternal age.Data were obtained from a questionnaire survey of all mothers of children born between 2004 and 2010 in Okinawa, Japan who underwent medical check-ups at age 3 months. Variables assessed were maternal smoking during pregnancy, maternal age, gestational age, parity, birth year, and complications during pregnancy. Stratified analyses were performed using a logistic regression model.In total, 92641 participants provided complete information on all variables. Over the 7 years studied, the proportion of mothers smoking during pregnancy decreased from 10.6% to 5.0%, while the prevalence of low birthweight did not change remarkably (around 10%. Maternal smoking was significantly associated with low birthweight in all age groups. The strength of the association increased with maternal age, both in crude and adjusted models.Consistent with previous studies conducted in Western countries, this study demonstrates that maternal age has a modifying effect on the association between maternal smoking and birthweight. This finding suggests that specific education and health care programs for older smoking mothers are important to improve their foetal growth.
Neural Specialization for Speech in the First Months of Life
Shultz, Sarah; Vouloumanos, Athena; Bennett, Randi H.; Pelphrey, Kevin
2014-01-01
How does the brain's response to speech change over the first months of life? Although behavioral findings indicate that neonates' listening biases are sharpened over the first months of life, with a species-specific preference for speech emerging by 3 months, the neural substrates underlying this developmental change are unknown. We…
Intelligibility of speech of children with speech and sound disorders
Ivetac, Tina
2014-01-01
The purpose of this study is to examine speech intelligibility of children with primary speech and sound disorders aged 3 to 6 years in everyday life. The research problem is based on the degree to which parents or guardians, immediate family members (sister, brother, grandparents), extended family members (aunt, uncle, cousin), child's friends, other acquaintances, child's teachers and strangers understand the speech of children with speech sound disorders. We examined whether the level ...
Maternal Employment and Adolescent Achievement Revisited: An Ecological Perspective.
Paulson, Sharon E.
1996-01-01
Explores the relation of maternal employment and attitudes toward maternal employment to adolescent achievement. Examines parenting characteristics that mediate relations between maternal employment factors and achievement. Subjects were 240 ninth graders and their parents. Findings were that maternal employment did not influence adolescent…
... disorder; Voice disorders; Vocal disorders; Disfluency; Communication disorder - speech disorder; Speech disorder - stuttering ... evaluation tools that can help identify and diagnose speech disorders: Denver Articulation Screening Examination Goldman-Fristoe Test of ...
Dick, Anthony Steven; Goldin-Meadow, Susan; Hasson, Uri; Skipper, Jeremy I; Small, Steven L
2009-11-01
Everyday communication is accompanied by visual information from several sources, including co-speech gestures, which provide semantic information listeners use to help disambiguate the speaker's message. Using fMRI, we examined how gestures influence neural activity in brain regions associated with processing semantic information. The BOLD response was recorded while participants listened to stories under three audiovisual conditions and one auditory-only (speech alone) condition. In the first audiovisual condition, the storyteller produced gestures that naturally accompany speech. In the second, the storyteller made semantically unrelated hand movements. In the third, the storyteller kept her hands still. In addition to inferior parietal and posterior superior and middle temporal regions, bilateral posterior superior temporal sulcus and left anterior inferior frontal gyrus responded more strongly to speech when it was further accompanied by gesture, regardless of the semantic relation to speech. However, the right inferior frontal gyrus was sensitive to the semantic import of the hand movements, demonstrating more activity when hand movements were semantically unrelated to the accompanying speech. These findings show that perceiving hand movements during speech modulates the distributed pattern of neural activation involved in both biological motion perception and discourse comprehension, suggesting listeners attempt to find meaning, not only in the words speakers produce, but also in the hand movements that accompany speech.
Neurophysiology of speech differences in childhood apraxia of speech.
Preston, Jonathan L; Molfese, Peter J; Gumkowski, Nina; Sorcinelli, Andrea; Harwood, Vanessa; Irwin, Julia R; Landi, Nicole
2014-01-01
Event-related potentials (ERPs) were recorded during a picture naming task of simple and complex words in children with typical speech and with childhood apraxia of speech (CAS). Results reveal reduced amplitude prior to speaking complex (multisyllabic) words relative to simple (monosyllabic) words for the CAS group over the right hemisphere during a time window thought to reflect phonological encoding of word forms. Group differences were also observed prior to production of spoken tokens regardless of word complexity during a time window just prior to speech onset (thought to reflect motor planning/programming). Results suggest differences in pre-speech neurolinguistic processes.
Maternal nutrition and birth outcomes.
Abu-Saad, Kathleen; Fraser, Drora
2010-01-01
In this review, the authors summarize current knowledge on maternal nutritional requirements during pregnancy, with a focus on the nutrients that have been most commonly investigated in association with birth outcomes. Data sourcing and extraction included searches of the primary resources establishing maternal nutrient requirements during pregnancy (e.g., Dietary Reference Intakes), and searches of Medline for "maternal nutrition"/[specific nutrient of interest] and "birth/pregnancy outcomes," focusing mainly on the less extensively reviewed evidence from observational studies of maternal dietary intake and birth outcomes. The authors used a conceptual framework which took both primary and secondary factors (e.g., baseline maternal nutritional status, socioeconomic status of the study populations, timing and methods of assessing maternal nutritional variables) into account when interpreting study findings. The authors conclude that maternal nutrition is a modifiable risk factor of public health importance that can be integrated into efforts to prevent adverse birth outcomes, particularly among economically developing/low-income populations.
Intelligibility of clear speech: effect of instruction.
Lam, Jennifer; Tjaden, Kris
2013-10-01
The authors investigated how clear speech instructions influence sentence intelligibility. Twelve speakers produced sentences in habitual, clear, hearing impaired, and overenunciate conditions. Stimuli were amplitude normalized and mixed with multitalker babble for orthographic transcription by 40 listeners. The main analysis investigated percentage-correct intelligibility scores as a function of the 4 conditions and speaker sex. Additional analyses included listener response variability, individual speaker trends, and an alternate intelligibility measure: proportion of content words correct. Relative to the habitual condition, the overenunciate condition was associated with the greatest intelligibility benefit, followed by the hearing impaired and clear conditions. Ten speakers followed this trend. The results indicated different patterns of clear speech benefit for male and female speakers. Greater listener variability was observed for speakers with inherently low habitual intelligibility compared to speakers with inherently high habitual intelligibility. Stable proportions of content words were observed across conditions. Clear speech instructions affected the magnitude of the intelligibility benefit. The instruction to overenunciate may be most effective in clear speech training programs. The findings may help explain the range of clear speech intelligibility benefit previously reported. Listener variability analyses suggested the importance of obtaining multiple listener judgments of intelligibility, especially for speakers with inherently low habitual intelligibility.
Ludmer, Jaclyn A; Jamieson, Brittany; Gonzalez, Andrea; Levitan, Robert; Kennedy, James; Villani, Vanessa; Masellis, Mario; Basile, Vincenzo S; Atkinson, Leslie
2017-10-01
A mother's cortisol secretion is importantly associated with her own mental health and her infant's cortisol secretion. This study investigated the influences of maternal history of care and maternal DRD2, SLC6A3, and OXTR genotypes on maternal cortisol in the context of infant stress. A community sample of 296 mother-infant dyads completed a maternal separation at infant age 17 months. Maternal salivary cortisol, buccal cells, and self-reported history of care were collected. Multilevel models revealed that history of care had a greater influence on maternal baseline cortisol (but not cortisol trajectory) for mothers with more plasticity alleles of SLC6A3 (10R) and OXTR (G), relative to mothers with fewer or no plasticity alleles. Findings indicate that a mother's history of care is related to her cortisol secretion in anticipation of infant stress, but that this relation depends on her genetic characteristics. Findings are discussed in relation to the maternal protective system and anticipatory cortisol secretion. Copyright © 2017 Elsevier B.V. All rights reserved.
Infants' preference for native audiovisual speech dissociated from congruency preference.
Directory of Open Access Journals (Sweden)
Kathleen Shaw
Full Text Available Although infant speech perception in often studied in isolated modalities, infants' experience with speech is largely multimodal (i.e., speech sounds they hear are accompanied by articulating faces. Across two experiments, we tested infants' sensitivity to the relationship between the auditory and visual components of audiovisual speech in their native (English and non-native (Spanish language. In Experiment 1, infants' looking times were measured during a preferential looking task in which they saw two simultaneous visual speech streams articulating a story, one in English and the other in Spanish, while they heard either the English or the Spanish version of the story. In Experiment 2, looking times from another group of infants were measured as they watched single displays of congruent and incongruent combinations of English and Spanish audio and visual speech streams. Findings demonstrated an age-related increase in looking towards the native relative to non-native visual speech stream when accompanied by the corresponding (native auditory speech. This increase in native language preference did not appear to be driven by a difference in preference for native vs. non-native audiovisual congruence as we observed no difference in looking times at the audiovisual streams in Experiment 2.
Complaint Speech Act of Hotel and Restaurant Guests
Suryawan, I Nengah; Putra Yadnya, Ida Bagus; Puspani, Ida Ayu Made
2016-01-01
This paper is aimed at analyzing how complaint speech act of hotel and restaurant guests are performed and responded based on categories of speech acts and how they are performed considering the aspects of acts: locutionary, illocutionary, and perlocutionary. The method and technique of collecting data in this study is documentation method in which the data were collected using the technique of note taking and were qualitatively analyzed. The findings show that complaint of hotel and restaura...
Lip movements affect infants' audiovisual speech perception.
Yeung, H Henny; Werker, Janet F
2013-05-01
Speech is robustly audiovisual from early in infancy. Here we show that audiovisual speech perception in 4.5-month-old infants is influenced by sensorimotor information related to the lip movements they make while chewing or sucking. Experiment 1 consisted of a classic audiovisual matching procedure, in which two simultaneously displayed talking faces (visual [i] and [u]) were presented with a synchronous vowel sound (audio /i/ or /u/). Infants' looking patterns were selectively biased away from the audiovisual matching face when the infants were producing lip movements similar to those needed to produce the heard vowel. Infants' looking patterns returned to those of a baseline condition (no lip movements, looking longer at the audiovisual matching face) when they were producing lip movements that did not match the heard vowel. Experiment 2 confirmed that these sensorimotor effects interacted with the heard vowel, as looking patterns differed when infants produced these same lip movements while seeing and hearing a talking face producing an unrelated vowel (audio /a/). These findings suggest that the development of speech perception and speech production may be mutually informative.
Improvement of intelligibility of ideal binary-masked noisy speech by adding background noise.
Cao, Shuyang; Li, Liang; Wu, Xihong
2011-04-01
When a target-speech/masker mixture is processed with the signal-separation technique, ideal binary mask (IBM), intelligibility of target speech is remarkably improved in both normal-hearing listeners and hearing-impaired listeners. Intelligibility of speech can also be improved by filling in speech gaps with un-modulated broadband noise. This study investigated whether intelligibility of target speech in the IBM-treated target-speech/masker mixture can be further improved by adding a broadband-noise background. The results of this study show that following the IBM manipulation, which remarkably released target speech from speech-spectrum noise, foreign-speech, or native-speech masking (experiment 1), adding a broadband-noise background with the signal-to-noise ratio no less than 4 dB significantly improved intelligibility of target speech when the masker was either noise (experiment 2) or speech (experiment 3). The results suggest that since adding the noise background shallows the areas of silence in the time-frequency domain of the IBM-treated target-speech/masker mixture, the abruption of transient changes in the mixture is smoothed and the perceived continuity of target-speech components becomes enhanced, leading to improved target-speech intelligibility. The findings are useful for advancing computational auditory scene analysis, hearing-aid/cochlear-implant designs, and understanding of speech perception under "cocktail-party" conditions.
Variability and Intelligibility of Clarified Speech to Different Listener Groups
Silber, Ronnie F.
Two studies examined the modifications that adult speakers make in speech to disadvantaged listeners. Previous research that has focused on speech to the deaf individuals and to young children has shown that adults clarify speech when addressing these two populations. Acoustic measurements suggest that the signal undergoes similar changes for both populations. Perceptual tests corroborate these results for the deaf population, but are nonsystematic in developmental studies. The differences in the findings for these populations and the nonsystematic results in the developmental literature may be due to methodological factors. The present experiments addressed these methodological questions. Studies of speech to hearing impaired listeners have used read, nonsense, sentences, for which speakers received explicit clarification instructions and feedback, while in the child literature, excerpts of real-time conversations were used. Therefore, linguistic samples were not precisely matched. In this study, experiments used various linguistic materials. Experiment 1 used a children's story; experiment 2, nonsense sentences. Four mothers read both types of material in four ways: (1) in "normal" adult speech, (2) in "babytalk," (3) under the clarification instructions used in the "hearing impaired studies" (instructed clear speech) and (4) in (spontaneous) clear speech without instruction. No extra practice or feedback was given. Sentences were presented to 40 normal hearing college students with and without simultaneous masking noise. Results were separately tabulated for content and function words, and analyzed using standard statistical tests. The major finding in the study was individual variation in speaker intelligibility. "Real world" speakers vary in their baseline intelligibility. The four speakers also showed unique patterns of intelligibility as a function of each independent variable. Results were as follows. Nonsense sentences were less intelligible than story
Daliri, Ayoub; Max, Ludo
2018-02-01
Auditory modulation during speech movement planning is limited in adults who stutter (AWS), but the functional relevance of the phenomenon itself remains unknown. We investigated for AWS and adults who do not stutter (AWNS) (a) a potential relationship between pre-speech auditory modulation and auditory feedback contributions to speech motor learning and (b) the effect on pre-speech auditory modulation of real-time versus delayed auditory feedback. Experiment I used a sensorimotor adaptation paradigm to estimate auditory-motor speech learning. Using acoustic speech recordings, we quantified subjects' formant frequency adjustments across trials when continually exposed to formant-shifted auditory feedback. In Experiment II, we used electroencephalography to determine the same subjects' extent of pre-speech auditory modulation (reductions in auditory evoked potential N1 amplitude) when probe tones were delivered prior to speaking versus not speaking. To manipulate subjects' ability to monitor real-time feedback, we included speaking conditions with non-altered auditory feedback (NAF) and delayed auditory feedback (DAF). Experiment I showed that auditory-motor learning was limited for AWS versus AWNS, and the extent of learning was negatively correlated with stuttering frequency. Experiment II yielded several key findings: (a) our prior finding of limited pre-speech auditory modulation in AWS was replicated; (b) DAF caused a decrease in auditory modulation for most AWNS but an increase for most AWS; and (c) for AWS, the amount of auditory modulation when speaking with DAF was positively correlated with stuttering frequency. Lastly, AWNS showed no correlation between pre-speech auditory modulation (Experiment II) and extent of auditory-motor learning (Experiment I) whereas AWS showed a negative correlation between these measures. Thus, findings suggest that AWS show deficits in both pre-speech auditory modulation and auditory-motor learning; however, limited pre-speech
Aging affects hemispheric asymmetry in the neural representation of speech sounds.
Bellis, T J; Nicol, T; Kraus, N
2000-01-15
Hemispheric asymmetries in the processing of elemental speech sounds appear to be critical for normal speech perception. This study investigated the effects of age on hemispheric asymmetry observed in the neurophysiological responses to speech stimuli in three groups of normal hearing, right-handed subjects: children (ages, 8-11 years), young adults (ages, 20-25 years), and older adults (ages > 55 years). Peak-to-peak response amplitudes of the auditory cortical P1-N1 complex obtained over right and left temporal lobes were examined to determine the degree of left/right asymmetry in the neurophysiological responses elicited by synthetic speech syllables in each of the three subject groups. In addition, mismatch negativity (MMN) responses, which are elicited by acoustic change, were obtained. Whereas children and young adults demonstrated larger P1-N1-evoked response amplitudes over the left temporal lobe than over the right, responses from elderly subjects were symmetrical. In contrast, MMN responses, which reflect an echoic memory process, were symmetrical in all subject groups. The differences observed in the neurophysiological responses were accompanied by a finding of significantly poorer ability to discriminate speech syllables involving rapid spectrotemporal changes in the older adult group. This study demonstrates a biological, age-related change in the neural representation of basic speech sounds and suggests one possible underlying mechanism for the speech perception difficulties exhibited by aging adults. Furthermore, results of this study support previous findings suggesting a dissociation between neural mechanisms underlying those processes that reflect the basic representation of sound structure and those that represent auditory echoic memory and stimulus change.
Directory of Open Access Journals (Sweden)
Benjamin Abraham
2014-12-01
Full Text Available This article makes a case study of ‘flarfing’ (a creative Facebook user practice with roots in found-text poetry in order to contribute to an understanding of the potentials and limitations facing users of online social networking sites who wish to address the issue of online hate speech. The practice of ‘flarfing’ involves users posting ‘blue text’ hyperlinked Facebook page names into status updates and comment threads. Facebook flarf sends a visible, though often non-literal, message to offenders and onlookers about what kinds of speech the responding activist(s find (unacceptable in online discussion, belonging to a category of agonistic online activism that repurposes the tools of internet trolling for activist ends. I argue this practice represents users attempting to ‘take responsibility’ for the culture of online spaces they inhabit, promoting intolerance to hate speech online. Careful consideration of the limits of flarf's efficacy within Facebook’s specific regulatory environment shows the extent to which this practice and similar responses to online hate speech are constrained by the platforms on which they exist.
Whole-exome sequencing supports genetic heterogeneity in childhood apraxia of speech.
Worthey, Elizabeth A; Raca, Gordana; Laffin, Jennifer J; Wilk, Brandon M; Harris, Jeremy M; Jakielski, Kathy J; Dimmock, David P; Strand, Edythe A; Shriberg, Lawrence D
2013-10-02
Childhood apraxia of speech (CAS) is a rare, severe, persistent pediatric motor speech disorder with associated deficits in sensorimotor, cognitive, language, learning and affective processes. Among other neurogenetic origins, CAS is the disorder segregating with a mutation in FOXP2 in a widely studied, multigenerational London family. We report the first whole-exome sequencing (WES) findings from a cohort of 10 unrelated participants, ages 3 to 19 years, with well-characterized CAS. As part of a larger study of children and youth with motor speech sound disorders, 32 participants were classified as positive for CAS on the basis of a behavioral classification marker using auditory-perceptual and acoustic methods that quantify the competence, precision and stability of a speaker's speech, prosody and voice. WES of 10 randomly selected participants was completed using the Illumina Genome Analyzer IIx Sequencing System. Image analysis, base calling, demultiplexing, read mapping, and variant calling were performed using Illumina software. Software developed in-house was used for variant annotation, prioritization and interpretation to identify those variants likely to be deleterious to neurodevelopmental substrates of speech-language development. Among potentially deleterious variants, clinically reportable findings of interest occurred on a total of five chromosomes (Chr3, Chr6, Chr7, Chr9 and Chr17), which included six genes either strongly associated with CAS (FOXP1 and CNTNAP2) or associated with disorders with phenotypes overlapping CAS (ATP13A4, CNTNAP1, KIAA0319 and SETX). A total of 8 (80%) of the 10 participants had clinically reportable variants in one or two of the six genes, with variants in ATP13A4, KIAA0319 and CNTNAP2 being the most prevalent. Similar to the results reported in emerging WES studies of other complex neurodevelopmental disorders, our findings from this first WES study of CAS are interpreted as support for heterogeneous genetic origins of
Speech rhythm in Kannada speaking adults who stutter.
Maruthy, Santosh; Venugopal, Sahana; Parakh, Priyanka
2017-10-01
A longstanding hypothesis about the underlying mechanisms of stuttering suggests that speech disfluencies may be associated with problems in timing and temporal patterning of speech events. Fifteen adults who do and do not stutter read five sentences, and from these, the vocalic and consonantal durations were measured. Using these, pairwise variability index (raw PVI for consonantal intervals and normalised PVI for vocalic intervals) and interval based rhythm metrics (PercV, DeltaC, DeltaV, VarcoC and VarcoV) were calculated for all the participants. Findings suggested higher mean values in adults who stutter when compared to adults who do not stutter for all the rhythm metrics except for VarcoV. Further, statistically significant difference between the two groups was found for all the rhythm metrics except for VarcoV. Combining the present results with consistent prior findings based on rhythm deficits in children and adults who stutter, there appears to be strong empirical support for the hypothesis that individuals who stutter may have deficits in generation of rhythmic speech patterns.
The Effect of Maternal Employment on Children's Academic Performance
DEFF Research Database (Denmark)
Nielsen, Lisbeth Palmhøj; Hansen, Anne Toft
of household control variables, instrumenting for employment with the gender- and education-specific local unemployment rate, and by including maternal fixed effects. We find that maternal employment has a positive effect on children’s academic performance in all specifications, particularly when women work...... part-time. This is in contrast with the larger literature on maternal employment, much of which takes place in other contexts, and which finds no or a small negative effect of maternal employment on children’s cognitive development and academic performance. (JEL J13, J22)...
Maternal feeding controls fetal biological clock.
Directory of Open Access Journals (Sweden)
Hidenobu Ohta
Full Text Available BACKGROUND: It is widely accepted that circadian physiological rhythms of the fetus are affected by oscillators in the maternal brain that are coupled to the environmental light-dark (LD cycle. METHODOLOGY/PRINCIPAL FINDINGS: To study the link between fetal and maternal biological clocks, we investigated the effects of cycles of maternal food availability on the rhythms of Per1 gene expression in the fetal suprachiasmatic nucleus (SCN and liver using a transgenic rat model whose tissues express luciferase in vitro. Although the maternal SCN remained phase-locked to the LD cycle, maternal restricted feeding phase-advanced the fetal SCN and liver by 5 and 7 hours respectively within the 22-day pregnancy. CONCLUSIONS/SIGNIFICANCE: Our results demonstrate that maternal feeding entrains the fetal SCN and liver independently of both the maternal SCN and the LD cycle. This indicates that maternal-feeding signals can be more influential for the fetal SCN and particular organ oscillators than hormonal signals controlled by the maternal SCN, suggesting the importance of a regular maternal feeding schedule for appropriate fetal molecular clockwork during pregnancy.
Directory of Open Access Journals (Sweden)
Bulent Cakmak
2015-03-01
Results: AFP levels were detected over 2.5 MoM in 122 (1.14% of 8552 pregnant women. There were no pathologic findings in 85.2% (104/122 of cases, but 14.8% (18/122 of cases had pathologic results; neural tube defect (NTD 9% (11/122 and anhidroamnios 2.5% (3/122. Omphalocele, polycystic kidney, sacrococcygeal teratoma, placental chorangioma were detected in 0.8% ratio. NTDs were found in 11 (0.12% of 8552 pregnant women who evaluated for maternal serum AFP (MSAFP. Conclusion: The routine use of MSAFP may be queried in NTD screening due to common usage of sufficient fetal anatomic ultrasound scan. [J Contemp Med 2015; 5(1.000: 8-12
Technology assisted speech and language therapy.
Glykas, Michael; Chytas, Panagiotis
2004-06-30
Speech and language therapists (SLTs) are faced daily with a diversity of speech and language disabilities, which are associated with a variety of conditions ranging from client groups with overall cognitive deficits to those with more specific difficulties. It is desirable that those working with such a range of problems and with such a demanding workload, plan care efficiently. Therefore, the introduction of methodologies, reference models of work and tools, which significantly improve the effectiveness of therapy, are particularly welcome. This paper describes the first web-based tool for diagnosis, treatment and e-Learning in the field of language and speech therapy. The system allows SLTs to find the optimum treatment for each patient, it also allows any non-specialist user-SLT, patient or helper (relative etc.)-to explore their creativity, by designing their own communication aid in an interactive manner, with the use of editors such as: configuration and vocabulary. The system has been tested and piloted by potential users in Greece and the UK.
Listeners Experience Linguistic Masking Release in Noise-Vocoded Speech-in-Speech Recognition
Viswanathan, Navin; Kokkinakis, Kostas; Williams, Brittany T.
2018-01-01
Purpose: The purpose of this study was to evaluate whether listeners with normal hearing perceiving noise-vocoded speech-in-speech demonstrate better intelligibility of target speech when the background speech was mismatched in language (linguistic release from masking [LRM]) and/or location (spatial release from masking [SRM]) relative to the…
DELVING INTO SPEECH ACT A Case Of Indonesian EFL Young Learners
Directory of Open Access Journals (Sweden)
Swastika Septiani, S.Pd
2017-04-01
Full Text Available This study attempts to describe the use of speech acts applied in primary school. This study is intended to identify the speech acts performed in primary school, to find the most dominant speech acts performed in elementary school, to give brief description of how speech acts applied in primary school, and to know how to apply the result of the study in English teaching learning to young learners. The speech acts performed in primary school is classified based on Searle‘s theory of speech acts. The most dominant speech acts performed in primary school is Directive (41.17%, the second speech act mostly performed is Declarative (33.33%, the third speech act mostly performed is Representative and Expressive (each 11.76%, and the least speech act performed is Commisive (1.9%. The speech acts performed in elementary school is applied on the context of situation determined by the National Education Standards Agency (BSNP. The speech acts performed in fourth grade have to be applied in the context of classroom, and the speech acts performed in fifth grade have to be applied in the context of school, whereas the speech acts performed in sixth grade have to be applied in the context of the students‘ surroundings. The result of this study is highy expected to give significant contribution to English teaching learning to young learners. By acknowledging the characteristics of young learners, the way they learn English as a foreign language, the teachers are expected to have inventive strategies and various techniques to create a fun and condusive atmosphere in English class.
Speech Perception and Short-Term Memory Deficits in Persistent Developmental Speech Disorder
Kenney, Mary Kay; Barac-Cikoja, Dragana; Finnegan, Kimberly; Jeffries, Neal; Ludlow, Christy L.
2006-01-01
Children with developmental speech disorders may have additional deficits in speech perception and/or short-term memory. To determine whether these are only transient developmental delays that can accompany the disorder in childhood or persist as part of the speech disorder, adults with a persistent familial speech disorder were tested on speech…
A common functional neural network for overt production of speech and gesture.
Marstaller, L; Burianová, H
2015-01-22
The perception of co-speech gestures, i.e., hand movements that co-occur with speech, has been investigated by several studies. The results show that the perception of co-speech gestures engages a core set of frontal, temporal, and parietal areas. However, no study has yet investigated the neural processes underlying the production of co-speech gestures. Specifically, it remains an open question whether Broca's area is central to the coordination of speech and gestures as has been suggested previously. The objective of this study was to use functional magnetic resonance imaging to (i) investigate the regional activations underlying overt production of speech, gestures, and co-speech gestures, and (ii) examine functional connectivity with Broca's area. We hypothesized that co-speech gesture production would activate frontal, temporal, and parietal regions that are similar to areas previously found during co-speech gesture perception and that both speech and gesture as well as co-speech gesture production would engage a neural network connected to Broca's area. Whole-brain analysis confirmed our hypothesis and showed that co-speech gesturing did engage brain areas that form part of networks known to subserve language and gesture. Functional connectivity analysis further revealed a functional network connected to Broca's area that is common to speech, gesture, and co-speech gesture production. This network consists of brain areas that play essential roles in motor control, suggesting that the coordination of speech and gesture is mediated by a shared motor control network. Our findings thus lend support to the idea that speech can influence co-speech gesture production on a motoric level. Copyright © 2014 IBRO. Published by Elsevier Ltd. All rights reserved.
Automatic speech recognition (ASR) based approach for speech therapy of aphasic patients: A review
Jamal, Norezmi; Shanta, Shahnoor; Mahmud, Farhanahani; Sha'abani, MNAH
2017-09-01
This paper reviews the state-of-the-art an automatic speech recognition (ASR) based approach for speech therapy of aphasic patients. Aphasia is a condition in which the affected person suffers from speech and language disorder resulting from a stroke or brain injury. Since there is a growing body of evidence indicating the possibility of improving the symptoms at an early stage, ASR based solutions are increasingly being researched for speech and language therapy. ASR is a technology that transfers human speech into transcript text by matching with the system's library. This is particularly useful in speech rehabilitation therapies as they provide accurate, real-time evaluation for speech input from an individual with speech disorder. ASR based approaches for speech therapy recognize the speech input from the aphasic patient and provide real-time feedback response to their mistakes. However, the accuracy of ASR is dependent on many factors such as, phoneme recognition, speech continuity, speaker and environmental differences as well as our depth of knowledge on human language understanding. Hence, the review examines recent development of ASR technologies and its performance for individuals with speech and language disorders.
... OTC Relief for Diarrhea Home Diseases and Conditions Speech and Language Delay Condition Speech and Language Delay Share Print Table of Contents1. ... Treatment6. Everyday Life7. Questions8. Resources What is a speech and language delay? A speech and language delay ...
Kerns, Caroline E; Pincus, Donna B; McLaughlin, Katie A; Comer, Jonathan S
2017-08-01
Environmental contributions are thought to play a primary role in the familial aggregation of anxiety, but parenting influences remain poorly understood. We examined dynamic relations between maternal anxiety, maternal emotion regulation (ER) during child distress, maternal accommodation of child distress, and child anxiety. Mothers (N=45) of youth ages 3-8 years (M=4.8) participated in an experimental task during which they listened to a standardized audio recording of a child in anxious distress pleading for parental intervention. Measures of maternal and child anxiety, mothers' affective states, mothers' ER strategies during the child distress, and maternal accommodation of child anxiety were collected. Mothers' resting respiratory sinus arrhythmia (RSA) reactivity during the recording was also acquired. Higher maternal negative affect and greater maternal ER switching (i.e., using multiple ER strategies in a short time without positive regulatory results) during child distress were associated with child anxiety. Sequential mediation modeling showed that maternal anxiety predicted ineffective maternal ER during child distress exposure, which in turn predicted greater maternal accommodation, which in turn predicted higher child anxiety. Findings support the mediating roles of maternal ER and accommodation in linking maternal and child anxiety, and suggest that ineffective maternal ER and subsequent attempts to accommodate child distress may act as mechanisms underlying the familial aggregation of anxiety. Copyright © 2017 Elsevier Ltd. All rights reserved.
Plasticity in the Human Speech Motor System Drives Changes in Speech Perception
Lametti, Daniel R.; Rochet-Capellan, Amélie; Neufeld, Emily; Shiller, Douglas M.
2014-01-01
Recent studies of human speech motor learning suggest that learning is accompanied by changes in auditory perception. But what drives the perceptual change? Is it a consequence of changes in the motor system? Or is it a result of sensory inflow during learning? Here, subjects participated in a speech motor-learning task involving adaptation to altered auditory feedback and they were subsequently tested for perceptual change. In two separate experiments, involving two different auditory perceptual continua, we show that changes in the speech motor system that accompany learning drive changes in auditory speech perception. Specifically, we obtained changes in speech perception when adaptation to altered auditory feedback led to speech production that fell into the phonetic range of the speech perceptual tests. However, a similar change in perception was not observed when the auditory feedback that subjects' received during learning fell into the phonetic range of the perceptual tests. This indicates that the central motor outflow associated with vocal sensorimotor adaptation drives changes to the perceptual classification of speech sounds. PMID:25080594
[Time perception, maternal tasks, and maternal role behavior among pregnant Japanese women].
Yamamoto, A
1996-01-01
The relationship of time perception, maternal tasks, and maternal role behavior was examined in 140 pregnant Japanese women with a short-term longitudinal design. A model developed by Rubin provided the conceptual framework for this research. The Time Perception Scale. Time Production Method, and the Prefatory Maternal Response measured the study variables. Study results revealed significant differences in duration of time, time production, maternal-fetal attachment, and maternal role behavior before and after quickening(fetal movement)occurred. Medium to strong positive relationships among time orientation, maternal-fetal attachment, gratification, and maternal role behavior were found before and after movement. After quickening, a weak relationship between time orientation and duration was found. After controlling maternal-fetal attachment and gratification in pregnancy and maternal role, orientation in time perception accounted for significant amounts of variance in maternal role behavior before and after fetal movement. Results show that the process of becoming a mother, which started before quickening, increased in magnitude after fetal movement. The function of fetal movement is important in developing motherhood. In the process of becoming a mother, cognitive, emotional, and behavioral aspects in becoming a mother are inseparable from each other. Future orientation of time perception contributes to development of maternal role behavior. Having a future orientation during pregnancy may indicate hope or positive expectation. Based on these findings, several recommendations were proposed: (a)to study further the general process of becoming a mother and the role of time perception in developing motherhood, (b)to disseminate information to the general public about the process in development of motherhood, (c)to construct theory to explain the process of becoming a mother, and(d)to conduct future research to clarify the construct of time perception and attachment.
The evolution of primary progressive apraxia of speech.
Josephs, Keith A; Duffy, Joseph R; Strand, Edythe A; Machulda, Mary M; Senjem, Matthew L; Gunter, Jeffrey L; Schwarz, Christopher G; Reid, Robert I; Spychalla, Anthony J; Lowe, Val J; Jack, Clifford R; Whitwell, Jennifer L
2014-10-01
Primary progressive apraxia of speech is a recently described neurodegenerative disorder in which patients present with an isolated apraxia of speech and show focal degeneration of superior premotor cortex. Little is known about how these individuals progress over time, making it difficult to provide prognostic estimates. Thirteen subjects with primary progressive apraxia of speech underwent two serial comprehensive clinical and neuroimaging evaluations 2.4 years apart [median age of onset = 67 years (range: 49-76), seven females]. All underwent detailed speech and language, neurological and neuropsychological assessments, and magnetic resonance imaging, diffusion tensor imaging and (18)F-fluorodeoxyglucose positron emission tomography at both baseline and follow-up. Rates of change of whole brain, ventricle, and midbrain volumes were calculated using the boundary-shift integral and atlas-based parcellation, and rates of regional grey matter atrophy were assessed using tensor-based morphometry. White matter tract degeneration was assessed on diffusion-tensor imaging at each time-point. Patterns of hypometabolism were assessed at the single subject-level. Neuroimaging findings were compared with a cohort of 20 age, gender, and scan-interval matched healthy controls. All subjects developed extrapyramidal signs. In eight subjects the apraxia of speech remained the predominant feature. In the other five there was a striking progression of symptoms that had evolved into a progressive supranuclear palsy-like syndrome; they showed a combination of severe parkinsonism, near mutism, dysphagia with choking, vertical supranuclear gaze palsy or slowing, balance difficulties with falls and urinary incontinence, and one was wheelchair bound. Rates of whole brain atrophy (1.5% per year; controls = 0.4% per year), ventricular expansion (8.0% per year; controls = 3.3% per year) and midbrain atrophy (1.5% per year; controls = 0.1% per year) were elevated (P ≤ 0.001) in all 13
Gavin, Amelia R; Hill, Karl G; Hawkins, J David; Maas, Carl
2011-08-01
This study examined three research questions: (1) Is there an association between maternal early-life economic disadvantage and the birth weight of later-born offspring? (2) Is there an association between maternal abuse in childhood and the birth weight of later-born offspring? (3) To what extent are these early-life risks mediated through adolescent and adult substance use, mental and physical health status, and adult socioeconomic status (SES)? Analyses used structural equation modeling to examine data from two longitudinal studies, which included three generations. The first generation (G1) and the second generation (G2) were enrolled in the Seattle Social Development Project (SSDP), and the third generation (G3) was enrolled in the SSDP Intergenerational Project. Data for the study (N = 136) focused on (G2) mothers enrolled in the SSDP and their children (G3). Analyses revealed that G2 low childhood SES predicted G3 offspring birth weight. Early childhood abuse among G2 respondents predicted G3 offspring birth weight through a mediated pathway including G2 adolescent substance use and G2 prenatal substance use. Birth weight was unrelated to maternal adult SES, depression, or obesity. To our knowledge, this is the first study to identify the effect of maternal early-life risks of low childhood SES and child maltreatment on later-born offspring birth weight. These findings have far-reaching effects on the cumulative risk associated with early-life economic disadvantage and childhood maltreatment. Such findings encourage policies and interventions that enhance child health at birth by taking the mother's own early-life and development into account. Copyright © 2011 Society for Adolescent Health and Medicine. Published by Elsevier Inc. All rights reserved.
Neural Tuning to Low-Level Features of Speech throughout the Perisylvian Cortex.
Berezutskaya, Julia; Freudenburg, Zachary V; Güçlü, Umut; van Gerven, Marcel A J; Ramsey, Nick F
2017-08-16
Despite a large body of research, we continue to lack a detailed account of how auditory processing of continuous speech unfolds in the human brain. Previous research showed the propagation of low-level acoustic features of speech from posterior superior temporal gyrus toward anterior superior temporal gyrus in the human brain (Hullett et al., 2016). In this study, we investigate what happens to these neural representations past the superior temporal gyrus and how they engage higher-level language processing areas such as inferior frontal gyrus. We used low-level sound features to model neural responses to speech outside of the primary auditory cortex. Two complementary imaging techniques were used with human participants (both males and females): electrocorticography (ECoG) and fMRI. Both imaging techniques showed tuning of the perisylvian cortex to low-level speech features. With ECoG, we found evidence of propagation of the temporal features of speech sounds along the ventral pathway of language processing in the brain toward inferior frontal gyrus. Increasingly coarse temporal features of speech spreading from posterior superior temporal cortex toward inferior frontal gyrus were associated with linguistic features such as voice onset time, duration of the formant transitions, and phoneme, syllable, and word boundaries. The present findings provide the groundwork for a comprehensive bottom-up account of speech comprehension in the human brain. SIGNIFICANCE STATEMENT We know that, during natural speech comprehension, a broad network of perisylvian cortical regions is involved in sound and language processing. Here, we investigated the tuning to low-level sound features within these regions using neural responses to a short feature film. We also looked at whether the tuning organization along these brain regions showed any parallel to the hierarchy of language structures in continuous speech. Our results show that low-level speech features propagate throughout the
Speech-Language Therapy (For Parents)
... Staying Safe Videos for Educators Search English Español Speech-Language Therapy KidsHealth / For Parents / Speech-Language Therapy ... most kids with speech and/or language disorders. Speech Disorders, Language Disorders, and Feeding Disorders A speech ...
Digital speech processing using Matlab
Gopi, E S
2014-01-01
Digital Speech Processing Using Matlab deals with digital speech pattern recognition, speech production model, speech feature extraction, and speech compression. The book is written in a manner that is suitable for beginners pursuing basic research in digital speech processing. Matlab illustrations are provided for most topics to enable better understanding of concepts. This book also deals with the basic pattern recognition techniques (illustrated with speech signals using Matlab) such as PCA, LDA, ICA, SVM, HMM, GMM, BPN, and KSOM.
Intervention for Children with Severe Speech Disorder: A Comparison of Two Approaches
Crosbie, Sharon; Holm, Alison; Dodd, Barbara
2005-01-01
Background: Children with speech disorder are a heterogeneous group (e.g. in terms of severity, types of errors and underlying causal factors). Much research has ignored this heterogeneity, giving rise to contradictory intervention study findings. This situation provides clinical motivation to identify the deficits in the speech-processing chain…
Developmental apraxia of speech in children. Quantitive assessment of speech characteristics
Thoonen, G.H.J.
1998-01-01
Developmental apraxia of speech (DAS) in children is a speech disorder, supposed to have a neurological origin, which is commonly considered to result from particular deficits in speech processing (i.e., phonological planning, motor programming). However, the label DAS has often been used as
Directory of Open Access Journals (Sweden)
Basavana Gowda
2015-03-01
Full Text Available OBJECTIVES: A study of maternal death conducted to evaluate various factors responsible for maternal deaths. To identify complications in pregnancy, a childbirth which result in maternal death, and to identify opportunities for preventive intervention and understand the events leading to death; so that improving maternal health and reducing maternal mortality rate significantly. To analyze the causes and epidemiological amounts maternal mortality e.g. age parity, socioeconomic status and literacy. In order to reduce maternal mortality and to implement safe motherhood program and complications of pregnancy and to find out safe motherhood program. METHODS: The data collected was a retrograde by a proforma containing particulars of the diseased, detailed history and relatives were interviewed for additional information. The data collected was analysed. RESULTS: Maternal mortality rate in our own institution is 200/ 100,000 live births. Among 30 maternal deaths, 56% deaths (17 were among low socio - economic status, groups 60% deaths among unbooked 53.5% deaths more along illiterates evidenced by direct and indirect deaths about 25% of deaths were preventable. CONCLUSION: Maternal death is a great tragedy in the family life. It is crusade to know not just the medical cause of the death but the circumstances what makes these continued tragic death even more unacceptable is that deaths are largely preventable
Speech and Communication Changes Reported by People with Parkinson's Disease.
Schalling, Ellika; Johansson, Kerstin; Hartelius, Lena
2017-01-01
Changes in communicative functions are common in Parkinson's disease (PD), but there are only limited data provided by individuals with PD on how these changes are perceived, what their consequences are, and what type of intervention is provided. To present self-reported information about speech and communication, the impact on communicative participation, and the amount and type of speech-language pathology services received by people with PD. Respondents with PD recruited via the Swedish Parkinson's Disease Society filled out a questionnaire accessed via a Web link or provided in a paper version. Of 188 respondents, 92.5% reported at least one symptom related to communication; the most common symptoms were weak voice, word-finding difficulties, imprecise articulation, and getting off topic in conversation. The speech and communication problems resulted in restricted communicative participation for between a quarter and a third of the respondents, and their speech caused embarrassment sometimes or more often to more than half. Forty-five percent of the respondents had received speech-language pathology services. Most respondents reported both speech and language symptoms, and many experienced restricted communicative participation. Access to speech-language pathology services is still inadequate. Services should also address cognitive/linguistic aspects to meet the needs of people with PD. © 2018 S. Karger AG, Basel.
Maternal scaffolding behavior: links with parenting style and maternal education.
Carr, Amanda; Pike, Alison
2012-03-01
The purpose of this study was to specify the relationship between positive and harsh parenting and maternal scaffolding behavior. A 2nd aim was to disentangle the effects of maternal education and parenting quality, and a 3rd aim was to test whether parenting quality mediated the association between maternal education and scaffolding practices. We examined associations between positive and harsh parenting practices and contingent and noncontingent tutoring strategies. Ninety-six mother-child dyads (49 boys, 47 girls) from working- and middle-class English families participated. Mothers reported on parenting quality at Time 1 when children were 5 years old and again approximately 5 years later at Time 2. Mother-child pairs were observed working together on a block design task at Time 2, and interactions were coded for contingent (contingent shifting) and noncontingent (fixed failure feedback) dimensions of maternal scaffolding behavior. Positive and harsh parenting accounted for variance in contingent behavior over and above maternal education, whereas only harsh parenting accounted for unique variance in noncontingent scaffolding practices. Our findings provide new evidence for a more differentiated model of the relation between general parenting quality and specific scaffolding behaviors. PsycINFO Database Record (c) 2012 APA, all rights reserved.
A music perception disorder (congenital amusia) influences speech comprehension.
Liu, Fang; Jiang, Cunmei; Wang, Bei; Xu, Yi; Patel, Aniruddh D
2015-01-01
This study investigated the underlying link between speech and music by examining whether and to what extent congenital amusia, a musical disorder characterized by degraded pitch processing, would impact spoken sentence comprehension for speakers of Mandarin, a tone language. Sixteen Mandarin-speaking amusics and 16 matched controls were tested on the intelligibility of news-like Mandarin sentences with natural and flat fundamental frequency (F0) contours (created via speech resynthesis) under four signal-to-noise (SNR) conditions (no noise, +5, 0, and -5dB SNR). While speech intelligibility in quiet and extremely noisy conditions (SNR=-5dB) was not significantly compromised by flattened F0, both amusic and control groups achieved better performance with natural-F0 sentences than flat-F0 sentences under moderately noisy conditions (SNR=+5 and 0dB). Relative to normal listeners, amusics demonstrated reduced speech intelligibility in both quiet and noise, regardless of whether the F0 contours of the sentences were natural or flattened. This deficit in speech intelligibility was not associated with impaired pitch perception in amusia. These findings provide evidence for impaired speech comprehension in congenital amusia, suggesting that the deficit of amusics extends beyond pitch processing and includes segmental processing. Copyright © 2014 Elsevier Ltd. All rights reserved.
Newland, Rebecca P.; Parade, Stephanie H.; Dickstein, Susan; Seifer, Ronald
2016-01-01
The current study prospectively examined the ways in which goodness of fit between maternal and infant sleep contributes to maternal depressive symptoms and the mother-child relationship across the first years of life. In a sample of 173 mother-child dyads, maternal prenatal sleep, infant sleep, maternal depressive symptoms, and mother-child attachment security were assessed via self-report, actigraphy, and observational measures. Results suggested that a poor fit between mothers’ prenatal sleep and infants’ sleep at 8 months (measured by sleep diary and actigraphy) was associated with maternal depressive symptoms at 15 months. Additionally, maternal depression mediated the association between the interplay of mother and infant sleep (measured by sleep diary) and mother-child attachment security at 30 months. Findings emphasize the importance of the match between mother and infant sleep on maternal wellbeing and mother-child relationships and highlight the role of mothers’ perceptions of infant sleep. PMID:27448324
The Effect of Maternal Employment on Children’s Academic Performance
DEFF Research Database (Denmark)
Dunifon, Rachel; Toft Hansen, Anne; Nicholson, Sean
of household control variables, instrumenting for employment with the gender- and education-specific local unemployment rate, and by including maternal fixed effects. We find that maternal employment has a positive effect on children’s academic performance in all specifications, particularly when women work...... part-time. This is in contrast with the larger literature on maternal employment, much of which takes place in other contexts, and which finds no or a small negative effect of maternal employment on children’s cognitive development and academic performance....
Congdon, Eliza L; Novack, Miriam A; Brooks, Neon; Hemani-Lopez, Naureen; O'Keefe, Lucy; Goldin-Meadow, Susan
2017-08-01
When teachers gesture during instruction, children retain and generalize what they are taught (Goldin-Meadow, 2014). But why does gesture have such a powerful effect on learning? Previous research shows that children learn most from a math lesson when teachers present one problem-solving strategy in speech while simultaneously presenting a different, but complementary, strategy in gesture (Singer & Goldin-Meadow, 2005). One possibility is that gesture is powerful in this context because it presents information simultaneously with speech. Alternatively, gesture may be effective simply because it involves the body, in which case the timing of information presented in speech and gesture may be less important for learning. Here we find evidence for the importance of simultaneity: 3 rd grade children retain and generalize what they learn from a math lesson better when given instruction containing simultaneous speech and gesture than when given instruction containing sequential speech and gesture. Interpreting these results in the context of theories of multimodal learning, we find that gesture capitalizes on its synchrony with speech to promote learning that lasts and can be generalized.
DEFF Research Database (Denmark)
Hasse Jørgensen, Stina
2011-01-01
About Speech Matters - Katarina Gregos, the Greek curator's exhibition at the Danish Pavillion, the Venice Biannual 2011.......About Speech Matters - Katarina Gregos, the Greek curator's exhibition at the Danish Pavillion, the Venice Biannual 2011....
The normalities and abnormalities associated with speech in psychometrically-defined schizotypy.
Cohen, Alex S; Auster, Tracey L; McGovern, Jessica E; MacAulay, Rebecca K
2014-12-01
Speech deficits are thought to be an important feature of schizotypy--defined as the personality organization reflecting a putative liability for schizophrenia. There is reason to suspect that these deficits manifest as a function of limited cognitive resources. To evaluate this idea, we examined speech from individuals with psychometrically-defined schizotypy during a low cognitively-demanding task versus a relatively high cognitively-demanding task. A range of objective, computer-based measures of speech tapping speech production (silence, number and length of pauses, number and length of utterances), speech variability (global and local intonation and emphasis) and speech content (word fillers, idea density) were employed. Data for control (n=37) and schizotypy (n=39) groups were examined. Results did not confirm our hypotheses. While the cognitive-load task reduced speech expressivity for subjects as a group for most variables, the schizotypy group was not more pathological in speech characteristics compared to the control group. Interestingly, some aspects of speech in schizotypal versus control subjects were healthier under high cognitive load. Moreover, schizotypal subjects performed better, at a trend level, than controls on the cognitively demanding task. These findings hold important implications for our understanding of the neurocognitive architecture associated with the schizophrenia-spectrum. Of particular note concerns the apparent mismatch between self-reported schizotypal traits and objective performance, and the resiliency of speech under cognitive stress in persons with high levels of schizotypy. Copyright © 2014 Elsevier B.V. All rights reserved.
Speech-To-Text Conversion STT System Using Hidden Markov Model HMM
Directory of Open Access Journals (Sweden)
Su Myat Mon
2015-06-01
Full Text Available Abstract Speech is an easiest way to communicate with each other. Speech processing is widely used in many applications like security devices household appliances cellular phones ATM machines and computers. The human computer interface has been developed to communicate or interact conveniently for one who is suffering from some kind of disabilities. Speech-to-Text Conversion STT systems have a lot of benefits for the deaf or dumb people and find their applications in our daily lives. In the same way the aim of the system is to convert the input speech signals into the text output for the deaf or dumb students in the educational fields. This paper presents an approach to extract features by using Mel Frequency Cepstral Coefficients MFCC from the speech signals of isolated spoken words. And Hidden Markov Model HMM method is applied to train and test the audio files to get the recognized spoken word. The speech database is created by using MATLAB.Then the original speech signals are preprocessed and these speech samples are extracted to the feature vectors which are used as the observation sequences of the Hidden Markov Model HMM recognizer. The feature vectors are analyzed in the HMM depending on the number of states.
Directory of Open Access Journals (Sweden)
Anne Birgitta Nilsen
2014-12-01
Full Text Available The manifesto of the Norwegian terrorist Anders Behring Breivik is based on the “Eurabia” conspiracy theory. This theory is a key starting point for hate speech amongst many right-wing extremists in Europe, but also has ramifications beyond these environments. In brief, proponents of the Eurabia theory claim that Muslims are occupying Europe and destroying Western culture, with the assistance of the EU and European governments. By contrast, members of Al-Qaeda and other extreme Islamists promote the conspiracy theory “the Crusade” in their hate speech directed against the West. Proponents of the latter theory argue that the West is leading a crusade to eradicate Islam and Muslims, a crusade that is similarly facilitated by their governments. This article presents analyses of texts written by right-wing extremists and Muslim extremists in an effort to shed light on how hate speech promulgates conspiracy theories in order to spread hatred and intolerance.The aim of the article is to contribute to a more thorough understanding of hate speech’s nature by applying rhetorical analysis. Rhetorical analysis is chosen because it offers a means of understanding the persuasive power of speech. It is thus a suitable tool to describe how hate speech works to convince and persuade. The concepts from rhetorical theory used in this article are ethos, logos and pathos. The concept of ethos is used to pinpoint factors that contributed to Osama bin Laden's impact, namely factors that lent credibility to his promotion of the conspiracy theory of the Crusade. In particular, Bin Laden projected common sense, good morals and good will towards his audience. He seemed to have coherent and relevant arguments; he appeared to possess moral credibility; and his use of language demonstrated that he wanted the best for his audience.The concept of pathos is used to define hate speech, since hate speech targets its audience's emotions. In hate speech it is the
Zeytinoglu, Selin; Calkins, Susan D; Swingler, Margaret M; Leerkes, Esther M
2017-03-01
This study examined the direct and indirect pathways from maternal effortful control to 2 aspects of children's self-regulation-executive functioning and behavioral regulation-via maternal emotional support. Two hundred seventy-eight children and their primary caregivers (96% mothers) participated in laboratory visits when children were 4 and 5 years, and teachers reported on children's behavior at kindergarten. At the 4-year assessment, maternal effortful control was measured using the Adult Temperament Questionnaire (Evans & Rothbart, 2007) and maternal emotional support was observed during a semistructured mother-child problem-solving task. At the 5-year assessment, children's executive functioning was measured using laboratory tasks designed to assess updating/working memory, inhibitory control, and cognitive flexibility, whereas behavioral regulation was assessed via teacher-report questionnaires on children's attention control, discipline and persistence, and work habits. Results from structural equation modeling indicated that, after controlling for child gender and minority status, and maternal education, maternal effortful control was indirectly associated with both child executive functioning and behavioral regulation through maternal emotional support. Maternal effortful control had a direct association with children's teacher-reported behavioral regulation but not observed executive functioning. These findings suggest that maternal effortful control may be a key contributing factor to the development of children's self-regulatory competencies through its impact on maternal emotional support. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Nonverbal oral apraxia in primary progressive aphasia and apraxia of speech.
Botha, Hugo; Duffy, Joseph R; Strand, Edythe A; Machulda, Mary M; Whitwell, Jennifer L; Josephs, Keith A
2014-05-13
The goal of this study was to explore the prevalence of nonverbal oral apraxia (NVOA), its association with other forms of apraxia, and associated imaging findings in patients with primary progressive aphasia (PPA) and progressive apraxia of speech (PAOS). Patients with a degenerative speech or language disorder were prospectively recruited and diagnosed with a subtype of PPA or with PAOS. All patients had comprehensive speech and language examinations. Voxel-based morphometry was performed to determine whether atrophy of a specific region correlated with the presence of NVOA. Eighty-nine patients were identified, of which 34 had PAOS, 9 had agrammatic PPA, 41 had logopenic aphasia, and 5 had semantic dementia. NVOA was very common among patients with PAOS but was found in patients with PPA as well. Several patients exhibited only one of NVOA or apraxia of speech. Among patients with apraxia of speech, the severity of the apraxia of speech was predictive of NVOA, whereas ideomotor apraxia severity was predictive of the presence of NVOA in those without apraxia of speech. Bilateral atrophy of the prefrontal cortex anterior to the premotor area and supplementary motor area was associated with NVOA. Apraxia of speech, NVOA, and ideomotor apraxia are at least partially separable disorders. The association of NVOA and apraxia of speech likely results from the proximity of the area reported here and the premotor area, which has been implicated in apraxia of speech. The association of ideomotor apraxia and NVOA among patients without apraxia of speech could represent disruption of modules shared by nonverbal oral movements and limb movements.
Rapid, generalized adaptation to asynchronous audiovisual speech.
Van der Burg, Erik; Goodbourn, Patrick T
2015-04-07
The brain is adaptive. The speed of propagation through air, and of low-level sensory processing, differs markedly between auditory and visual stimuli; yet the brain can adapt to compensate for the resulting cross-modal delays. Studies investigating temporal recalibration to audiovisual speech have used prolonged adaptation procedures, suggesting that adaptation is sluggish. Here, we show that adaptation to asynchronous audiovisual speech occurs rapidly. Participants viewed a brief clip of an actor pronouncing a single syllable. The voice was either advanced or delayed relative to the corresponding lip movements, and participants were asked to make a synchrony judgement. Although we did not use an explicit adaptation procedure, we demonstrate rapid recalibration based on a single audiovisual event. We find that the point of subjective simultaneity on each trial is highly contingent upon the modality order of the preceding trial. We find compelling evidence that rapid recalibration generalizes across different stimuli, and different actors. Finally, we demonstrate that rapid recalibration occurs even when auditory and visual events clearly belong to different actors. These results suggest that rapid temporal recalibration to audiovisual speech is primarily mediated by basic temporal factors, rather than higher-order factors such as perceived simultaneity and source identity. © 2015 The Author(s) Published by the Royal Society. All rights reserved.
Measurement of speech levels in the presence of time varying background noise
Pearsons, K. S.; Horonjeff, R.
1982-01-01
Short-term speech level measurements which could be used to note changes in vocal effort in a time varying noise environment were studied. Knowing the changes in speech level would in turn allow prediction of intelligibility in the presence of aircraft flyover noise. Tests indicated that it is possible to use two second samples of speech to estimate long term root mean square speech levels. Other tests were also performed in which people read out loud during aircraft flyover noise. Results of these tests indicate that people do indeed raise their voice during flyovers at a rate of about 3-1/2 dB for each 10 dB increase in background level. This finding is in agreement with other tests of speech levels in the presence of steady state background noise.
Iuzzini-Seigel, Jenya; Hogan, Tiffany P.; Green, Jordan R.
2017-01-01
Purpose: The current research sought to determine (a) if speech inconsistency is a core feature of childhood apraxia of speech (CAS) or if it is driven by comorbid language impairment that affects a large subset of children with CAS and (b) if speech inconsistency is a sensitive and specific diagnostic marker that can differentiate between CAS and…
Prevalence and Phenotype of Childhood Apraxia of Speech in Youth with Galactosemia
Shriberg, Lawrence D.; Potter, Nancy L.; Strand, Edythe A.
2011-01-01
Purpose: In this article, the authors address the hypothesis that the severe and persistent speech disorder reported in persons with galactosemia meets contemporary diagnostic criteria for Childhood Apraxia of Speech (CAS). A positive finding for CAS in this rare metabolic disorder has the potential to impact treatment of persons with galactosemia…
THE INFLUENCE OF CULTURE ON THE STEPS OF PERSUASIVE KEYNOTE SPEECH
Directory of Open Access Journals (Sweden)
Alief Noor Farida
2017-04-01
Full Text Available Persuasion in business is important. It can be done by giving information about the products to the customers via advertisements or keynote speech when the company is launching the products. As the function of a keynote speech is important, the speaker of the event should be able to deliver the speech in a clear and concise manner. In this study, discourse analysis was done to find out the moves of persuasive keynote speech in mobile phone launching events. There were four keynote speeches analyzed. They have similar move structure: Introduction^Body^Conclusion, but the steps constructing the moves are different, especially in Move 1 and Move 3. In Move 2, the steps follow Monroe’s Motivated Sequence. The structure of the steps in the move 1 and move 2 of the keynote speeches under study varies. The variation of step structure is highly influenced by the cultural background of both the speakers and the audiences; eastern and western culture have some significant differences. This study shows that basic knowledge on the audience’s cultural background will help the keynote speaker to construct his/her speech to give better persuasive effect on the audience.
Boesch, Miriam C; Wendt, Oliver; Subramanian, Anu; Hsu, Ning
2013-09-01
The Picture Exchange Communication System (PECS) and a speech-generating device (SGD) were compared in a study with a multiple baseline, alternating treatment design. The effectiveness of these methods in increasing social-communicative behavior and natural speech production were assessed with three elementary school-aged children with severe autism who demonstrated extremely limited functional communication skills. Results for social-communicative behavior were mixed for all participants in both treatment conditions. Relatively little difference was observed between PECS and SGD conditions. Although findings were inconclusive, data patterns suggest that Phase II of the PECS training protocol is conducive to encouraging social-communicative behavior. Data for speech outcomes did not reveal any increases across participants, and no differences between treatment conditions were observed.
Pries, Lotta-Katrin; Guloksuz, Sinan; Menne-Lothmann, Claudia; Decoster, Jeroen; van Winkel, Ruud; Collip, Dina; Delespaul, Philippe; De Hert, Marc; Derom, Catherine; Thiery, Evert; Jacobs, Nele; Wichers, Marieke; Simons, Claudia J P; Rutten, Bart P F; van Os, Jim
2017-01-01
An association between white noise speech illusion and psychotic symptoms has been reported in patients and their relatives. This supports the theory that bottom-up and top-down perceptual processes are involved in the mechanisms underlying perceptual abnormalities. However, findings in nonclinical populations have been conflicting. The aim of this study was to examine the association between white noise speech illusion and subclinical expression of psychotic symptoms in a nonclinical sample. Findings were compared to previous results to investigate potential methodology dependent differences. In a general population adolescent and young adult twin sample (n = 704), the association between white noise speech illusion and subclinical psychotic experiences, using the Structured Interview for Schizotypy-Revised (SIS-R) and the Community Assessment of Psychic Experiences (CAPE), was analyzed using multilevel logistic regression analyses. Perception of any white noise speech illusion was not associated with either positive or negative schizotypy in the general population twin sample, using the method by Galdos et al. (2011) (positive: ORadjusted: 0.82, 95% CI: 0.6-1.12, p = 0.217; negative: ORadjusted: 0.75, 95% CI: 0.56-1.02, p = 0.065) and the method by Catalan et al. (2014) (positive: ORadjusted: 1.11, 95% CI: 0.79-1.57, p = 0.557). No association was found between CAPE scores and speech illusion (ORadjusted: 1.25, 95% CI: 0.88-1.79, p = 0.220). For the Catalan et al. (2014) but not the Galdos et al. (2011) method, a negative association was apparent between positive schizotypy and speech illusion with positive or negative affective valence (ORadjusted: 0.44, 95% CI: 0.24-0.81, p = 0.008). Contrary to findings in clinical populations, white noise speech illusion may not be associated with psychosis proneness in nonclinical populations.
Speech-like rhythm in a voiced and voiceless orangutan call.
Directory of Open Access Journals (Sweden)
Adriano R Lameira
Full Text Available The evolutionary origins of speech remain obscure. Recently, it was proposed that speech derived from monkey facial signals which exhibit a speech-like rhythm of ∼5 open-close lip cycles per second. In monkeys, these signals may also be vocalized, offering a plausible evolutionary stepping stone towards speech. Three essential predictions remain, however, to be tested to assess this hypothesis' validity; (i Great apes, our closest relatives, should likewise produce 5Hz-rhythm signals, (ii speech-like rhythm should involve calls articulatorily similar to consonants and vowels given that speech rhythm is the direct product of stringing together these two basic elements, and (iii speech-like rhythm should be experience-based. Via cinematic analyses we demonstrate that an ex-entertainment orangutan produces two calls at a speech-like rhythm, coined "clicks" and "faux-speech." Like voiceless consonants, clicks required no vocal fold action, but did involve independent manoeuvring over lips and tongue. In parallel to vowels, faux-speech showed harmonic and formant modulations, implying vocal fold and supralaryngeal action. This rhythm was several times faster than orangutan chewing rates, as observed in monkeys and humans. Critically, this rhythm was seven-fold faster, and contextually distinct, than any other known rhythmic calls described to date in the largest database of the orangutan repertoire ever assembled. The first two predictions advanced by this study are validated and, based on parsimony and exclusion of potential alternative explanations, initial support is given to the third prediction. Irrespectively of the putative origins of these calls and underlying mechanisms, our findings demonstrate irrevocably that great apes are not respiratorily, articulatorilly, or neurologically constrained for the production of consonant- and vowel-like calls at speech rhythm. Orangutan clicks and faux-speech confirm the importance of rhythmic speech
Variable Frame Rate and Length Analysis for Data Compression in Distributed Speech Recognition
DEFF Research Database (Denmark)
Kraljevski, Ivan; Tan, Zheng-Hua
2014-01-01
This paper addresses the issue of data compression in distributed speech recognition on the basis of a variable frame rate and length analysis method. The method first conducts frame selection by using a posteriori signal-to-noise ratio weighted energy distance to find the right time resolution...... length for steady regions. The method is applied to scalable source coding in distributed speech recognition where the target bitrate is met by adjusting the frame rate. Speech recognition results show that the proposed approach outperforms other compression methods in terms of recognition accuracy...... for noisy speech while achieving higher compression rates....
DEFF Research Database (Denmark)
Niebuhr, Oliver
2017-01-01
of reduction levels and perceived speaker attributes in which moderate reduction can make a better impression on listeners than no reduction. In addition to its relevance in reduction models and theories, this interplay is instructive for various fields of speech application from social robotics to charisma...... whether variation in the degree of reduction also has a systematic effect on the attributes we ascribe to the speaker who produces the speech signal. A perception experiment was carried out for German in which 46 listeners judged whether or not speakers showing 3 different combinations of segmental...... and prosodic reduction levels (unreduced, moderately reduced, strongly reduced) are appropriately described by 13 physical, social, and cognitive attributes. The experiment shows that clear speech is not mere speech, and less clear speech is not just reduced either. Rather, results revealed a complex interplay...
Dissociated Crossed Speech Areas in a Tumour Patient
Directory of Open Access Journals (Sweden)
Jörg Mauler
2017-05-01
Full Text Available In the past, the eloquent areas could be deliberately localised by the invasive Wada test. The very rare cases of dissociated crossed speech areas were accidentally found based on the clinical symptomatology. Today functional magnetic resonance imaging (fMRI-based imaging can be employed to non-invasively localise the eloquent areas in brain tumour patients for therapy planning. A 41-year-old, left-handed man with a low-grade glioma in the left frontal operculum extending to the insular cortex, tension headaches, and anomic aphasia over 5 months underwent a pre-operative speech area localisation fMRI measurement, which revealed the evidence of the transhemispheric disposition, where the dominant Wernicke speech area is located on the left and the Broca’s area is strongly lateralised to the right hemisphere. The outcome of the Wada test and the intraoperative cortico-subcortical stimulation mapping were congruent with this finding. After tumour removal, language area function was fully preserved. Upon the occurrence of brain tumours with a risk of impaired speech function, the rare dissociate crossed speech areas disposition may gain a clinically relevant meaning by allowing for more extended tumour removal. Hence, for its identification, diagnostics which take into account both brain hemispheres, such as fMRI, are recommended.
Audiovisual Temporal Recalibration for Speech in Synchrony Perception and Speech Identification
Asakawa, Kaori; Tanaka, Akihiro; Imai, Hisato
We investigated whether audiovisual synchrony perception for speech could change after observation of the audiovisual temporal mismatch. Previous studies have revealed that audiovisual synchrony perception is re-calibrated after exposure to a constant timing difference between auditory and visual signals in non-speech. In the present study, we examined whether this audiovisual temporal recalibration occurs at the perceptual level even for speech (monosyllables). In Experiment 1, participants performed an audiovisual simultaneity judgment task (i.e., a direct measurement of the audiovisual synchrony perception) in terms of the speech signal after observation of the speech stimuli which had a constant audiovisual lag. The results showed that the “simultaneous” responses (i.e., proportion of responses for which participants judged the auditory and visual stimuli to be synchronous) at least partly depended on exposure lag. In Experiment 2, we adopted the McGurk identification task (i.e., an indirect measurement of the audiovisual synchrony perception) to exclude the possibility that this modulation of synchrony perception was solely attributable to the response strategy using stimuli identical to those of Experiment 1. The characteristics of the McGurk effect reported by participants depended on exposure lag. Thus, it was shown that audiovisual synchrony perception for speech could be modulated following exposure to constant lag both in direct and indirect measurement. Our results suggest that temporal recalibration occurs not only in non-speech signals but also in monosyllabic speech at the perceptual level.
Under-resourced speech recognition based on the speech manifold
CSIR Research Space (South Africa)
Sahraeian, R
2015-09-01
Full Text Available Conventional acoustic modeling involves estimating many parameters to effectively model feature distributions. The sparseness of speech and text data, however, degrades the reliability of the estimation process and makes speech recognition a...
Gender, status and 'powerless' speech: interactions of students and lecturers.
McFadyen, R G
1996-09-01
The present study investigated whether the use of 'powerless' speech was affected by role status, speaker's gender and gender of another participant. Fifty-two university lecturers and 156 students participated. Students were paired with a lecturer or student of the same or opposite sex. The findings placed a question mark over the link between powerless speech and individuals of low role status. Moreover, against hypothesis, speaker's gender and gender of partner did not affect the use of qualifiers or fillers, although they affected the use of tag questions and some types of hesitation. A qualitative analysis was also conducted which suggested that the powerless features were, in fact, multi-functional with respect to power. In addition, the importance of a variety of interactional techniques, such as credibility techniques, in the creation or negotiation of relational power was documented. As a whole, these findings highlight problems with the concept of 'powerless' speech, at least with respect to relational power.
PRACTICING SPEECH THERAPY INTERVENTION FOR SOCIAL INTEGRATION OF CHILDREN WITH SPEECH DISORDERS
Directory of Open Access Journals (Sweden)
Martin Ofelia POPESCU
2016-11-01
Full Text Available The article presents a concise speech correction intervention program in of dyslalia in conjunction with capacity development of intra, interpersonal and social integration of children with speech disorders. The program main objectives represent: the potential increasing of individual social integration by correcting speech disorders in conjunction with intra- and interpersonal capacity, the potential growth of children and community groups for social integration by optimizing the socio-relational context of children with speech disorder. In the program were included 60 children / students with dyslalia speech disorders (monomorphic and polymorphic dyslalia, from 11 educational institutions - 6 kindergartens and 5 schools / secondary schools, joined with inter-school logopedic centre (CLI from Targu Jiu city and areas of Gorj district. The program was implemented under the assumption that therapeutic-formative intervention to correct speech disorders and facilitate the social integration will lead, in combination with correct pronunciation disorders, to social integration optimization of children with speech disorders. The results conirm the hypothesis and gives facts about the intervention program eficiency.
Newland, Rebecca P; Parade, Stephanie H; Dickstein, Susan; Seifer, Ronald
2016-08-01
The current study prospectively examined the ways in which goodness of fit between maternal and infant sleep contributes to maternal depressive symptoms and the mother-child relationship across the first years of life. In a sample of 173 mother-child dyads, maternal prenatal sleep, infant sleep, maternal depressive symptoms, and mother-child attachment security were assessed via self-report, actigraphy, and observational measures. Results suggested that a poor fit between mothers' prenatal sleep and infants' sleep at 8 months (measured by sleep diary and actigraphy) was associated with maternal depressive symptoms at 15 months. Additionally, maternal depression mediated the association between the interplay of mother and infant sleep (measured by sleep diary) and mother-child attachment security at 30 months. Findings emphasize the importance of the match between mother and infant sleep on maternal wellbeing and mother-child relationships and highlight the role of mothers' perceptions of infant sleep. Copyright © 2016 Elsevier Inc. All rights reserved.
Don't speak too fast! Processing of fast rate speech in children with specific language impairment.
Directory of Open Access Journals (Sweden)
Hélène Guiraud
Full Text Available Perception of speech rhythm requires the auditory system to track temporal envelope fluctuations, which carry syllabic and stress information. Reduced sensitivity to rhythmic acoustic cues has been evidenced in children with Specific Language Impairment (SLI, impeding syllabic parsing and speech decoding. Our study investigated whether these children experience specific difficulties processing fast rate speech as compared with typically developing (TD children.Sixteen French children with SLI (8-13 years old with mainly expressive phonological disorders and with preserved comprehension and 16 age-matched TD children performed a judgment task on sentences produced 1 at normal rate, 2 at fast rate or 3 time-compressed. Sensitivity index (d' to semantically incongruent sentence-final words was measured.Overall children with SLI perform significantly worse than TD children. Importantly, as revealed by the significant Group × Speech Rate interaction, children with SLI find it more challenging than TD children to process both naturally or artificially accelerated speech. The two groups do not significantly differ in normal rate speech processing.In agreement with rhythm-processing deficits in atypical language development, our results suggest that children with SLI face difficulties adjusting to rapid speech rate. These findings are interpreted in light of temporal sampling and prosodic phrasing frameworks and of oscillatory mechanisms underlying speech perception.
Don't speak too fast! Processing of fast rate speech in children with specific language impairment.
Guiraud, Hélène; Bedoin, Nathalie; Krifi-Papoz, Sonia; Herbillon, Vania; Caillot-Bascoul, Aurélia; Gonzalez-Monge, Sibylle; Boulenger, Véronique
2018-01-01
Perception of speech rhythm requires the auditory system to track temporal envelope fluctuations, which carry syllabic and stress information. Reduced sensitivity to rhythmic acoustic cues has been evidenced in children with Specific Language Impairment (SLI), impeding syllabic parsing and speech decoding. Our study investigated whether these children experience specific difficulties processing fast rate speech as compared with typically developing (TD) children. Sixteen French children with SLI (8-13 years old) with mainly expressive phonological disorders and with preserved comprehension and 16 age-matched TD children performed a judgment task on sentences produced 1) at normal rate, 2) at fast rate or 3) time-compressed. Sensitivity index (d') to semantically incongruent sentence-final words was measured. Overall children with SLI perform significantly worse than TD children. Importantly, as revealed by the significant Group × Speech Rate interaction, children with SLI find it more challenging than TD children to process both naturally or artificially accelerated speech. The two groups do not significantly differ in normal rate speech processing. In agreement with rhythm-processing deficits in atypical language development, our results suggest that children with SLI face difficulties adjusting to rapid speech rate. These findings are interpreted in light of temporal sampling and prosodic phrasing frameworks and of oscillatory mechanisms underlying speech perception.
Golfinopoulos, Elisa
Acoustic variability in fluent speech can arise at many stages in speech production planning and execution. For example, at the phonological encoding stage, the grouping of phonemes into syllables determines which segments are coarticulated and, by consequence, segment-level acoustic variation. Likewise phonetic encoding, which determines the spatiotemporal extent of articulatory gestures, will affect the acoustic detail of segments. Functional magnetic resonance imaging (fMRI) was used to measure brain activity of fluent adult speakers in four speaking conditions: fast, normal, clear, and emphatic (or stressed) speech. These speech manner changes typically result in acoustic variations that do not change the lexical or semantic identity of productions but do affect the acoustic saliency of phonemes, syllables and/or words. Acoustic responses recorded inside the scanner were assessed quantitatively using eight acoustic measures and sentence duration was used as a covariate of non-interest in the neuroimaging analysis. Compared to normal speech, emphatic speech was characterized acoustically by a greater difference between stressed and unstressed vowels in intensity, duration, and fundamental frequency, and neurally by increased activity in right middle premotor cortex and supplementary motor area, and bilateral primary sensorimotor cortex. These findings are consistent with right-lateralized motor planning of prosodic variation in emphatic speech. Clear speech involved an increase in average vowel and sentence durations and average vowel spacing, along with increased activity in left middle premotor cortex and bilateral primary sensorimotor cortex. These findings are consistent with an increased reliance on feedforward control, resulting in hyper-articulation, under clear as compared to normal speech. Fast speech was characterized acoustically by reduced sentence duration and average vowel spacing, and neurally by increased activity in left anterior frontal
Pearce, Anna; Whitehead, Margaret; Law, Catherine
2017-01-01
Background: Maternal employment has increased in European countries, but levels of employment are lower among mothers whose children have a limiting long-term illness or disability. However, we do not know whether having a child with a limiting illness prevents take-up or maintenance of paid employment or whether ‘common causes’, such as lack of qualifications or maternal disability lead to both maternal unemployment and childhood illness. Longitudinal data have the potential to distinguish between these. Methods: We analyzed four waves (3, 5, 7 and 11 years) of the Millennium Cohort Study (MCS) to examine the relationship between childhood limiting illness and maternal employment, unadjusted and adjusted for covariates. Multinomial regression models were used to test the association between child illness and trajectories of maternal employment. Fixed effects models assessed whether a new report of a child illness increased the odds of a mother exiting employment. Results: At every wave, maternal employment was more likely if the child did not have a limiting illness. After adjustment for covariates, childhood illness was associated with risks of continuous non-employment (adjusted Relative Risk Ratio = 1.46 [Confidence Interval: 1.21, 1.76]) or disrupted employment (aRRR = 1.26 [CI: 1.06, 1.49]), compared with entering or maintaining employment. If a child developed a limiting long-term illness, the likelihood of their mother exiting employment increased (adjusted Odds Ratio = 1.27 [CI: 1.05, 1.54]). Conclusions: ‘Common causes’ did not fully account for the association between child illness and maternal employment. Having a child with a limiting illness potentially reduces maternal employment opportunities. PMID:28177497
Zheng, Yingjun; Wu, Chao; Li, Juanhua; Li, Ruikeng; Peng, Hongjun; She, Shenglin; Ning, Yuping; Li, Liang
2018-04-04
Speech recognition under noisy "cocktail-party" environments involves multiple perceptual/cognitive processes, including target detection, selective attention, irrelevant signal inhibition, sensory/working memory, and speech production. Compared to health listeners, people with schizophrenia are more vulnerable to masking stimuli and perform worse in speech recognition under speech-on-speech masking conditions. Although the schizophrenia-related speech-recognition impairment under "cocktail-party" conditions is associated with deficits of various perceptual/cognitive processes, it is crucial to know whether the brain substrates critically underlying speech detection against informational speech masking are impaired in people with schizophrenia. Using functional magnetic resonance imaging (fMRI), this study investigated differences between people with schizophrenia (n = 19, mean age = 33 ± 10 years) and their matched healthy controls (n = 15, mean age = 30 ± 9 years) in intra-network functional connectivity (FC) specifically associated with target-speech detection under speech-on-speech-masking conditions. The target-speech detection performance under the speech-on-speech-masking condition in participants with schizophrenia was significantly worse than that in matched healthy participants (healthy controls). Moreover, in healthy controls, but not participants with schizophrenia, the strength of intra-network FC within the bilateral caudate was positively correlated with the speech-detection performance under the speech-masking conditions. Compared to controls, patients showed altered spatial activity pattern and decreased intra-network FC in the caudate. In people with schizophrenia, the declined speech-detection performance under speech-on-speech masking conditions is associated with reduced intra-caudate functional connectivity, which normally contributes to detecting target speech against speech masking via its functions of suppressing masking-speech signals.
How may the basal ganglia contribute to auditory categorization and speech perception?
Directory of Open Access Journals (Sweden)
Sung-Joo eLim
2014-08-01
Full Text Available Listeners must accomplish two complementary perceptual feats in extracting a message from speech. They must discriminate linguistically-relevant acoustic variability and generalize across irrelevant variability. Said another way, they must categorize speech. Since the mapping of acoustic variability is language-specific, these categories must be learned from experience. Thus, understanding how, in general, the auditory system acquires and represents categories can inform us about the toolbox of mechanisms available to speech perception. This perspective invites consideration of findings from cognitive neuroscience literatures outside of the speech domain as a means of constraining models of speech perception. Although neurobiological models of speech perception have mainly focused on cerebral cortex, research outside the speech domain is consistent with the possibility of significant subcortical contributions in category learning. Here, we review the functional role of one such structure, the basal ganglia. We examine research from animal electrophysiology, human neuroimaging, and behavior to consider characteristics of basal ganglia processing that may be advantageous for speech category learning. We also present emerging evidence for a direct role for basal ganglia in learning auditory categories in a complex, naturalistic task intended to model the incidental manner in which speech categories are acquired. To conclude, we highlight new research questions that arise in incorporating the broader neuroscience research literature in modeling speech perception, and suggest how understanding contributions of the basal ganglia can inform attempts to optimize training protocols for learning non-native speech categories in adulthood.
Detecting self-produced speech errors before and after articulation: An ERP investigation
Directory of Open Access Journals (Sweden)
Kevin Michael Trewartha
2013-11-01
Full Text Available It has been argued that speech production errors are monitored by the same neural system involved in monitoring other types of action errors. Behavioral evidence has shown that speech errors can be detected and corrected prior to articulation, yet the neural basis for such pre-articulatory speech error monitoring is poorly understood. The current study investigated speech error monitoring using a phoneme-substitution task known to elicit speech errors. Stimulus-locked event-related potential (ERP analyses comparing correct and incorrect utterances were used to assess pre-articulatory error monitoring and response-locked ERP analyses were used to assess post-articulatory monitoring. Our novel finding in the stimulus-locked analysis revealed that words that ultimately led to a speech error were associated with a larger P2 component at midline sites (FCz, Cz, and CPz. This early positivity may reflect the detection of an error in speech formulation, or a predictive mechanism to signal the potential for an upcoming speech error. The data also revealed that general conflict monitoring mechanisms are involved during this task as both correct and incorrect responses elicited an anterior N2 component typically associated with conflict monitoring. The response-locked analyses corroborated previous observations that self-produced speech errors led to a fronto-central ERN. These results demonstrate that speech errors can be detected prior to articulation, and that speech error monitoring relies on a central error monitoring mechanism.
Directory of Open Access Journals (Sweden)
Miladis Fornaris-Méndez
2017-04-01
Full Text Available Language therapy has trafficked from a medical focus until a preventive focus. However, difficulties are evidenced in the development of this last task, because he is devoted bigger space to the correction of the disorders of the language. Because the speech disorders is the dysfunction with more frequently appearance, acquires special importance the preventive work that is developed to avoid its appearance. Speech education since early age of the childhood makes work easier for prevent the appearance of speech disorders in the children. The present work has as objective to offer different activities for the prevention of the speech disorders.
Speech and Speech-Related Quality of Life After Late Palate Repair: A Patient's Perspective.
Schönmeyr, Björn; Wendby, Lisa; Sharma, Mitali; Jacobson, Lia; Restrepo, Carolina; Campbell, Alex
2015-07-01
Many patients with cleft palate deformities worldwide receive treatment at a later age than is recommended for normal speech to develop. The outcomes after late palate repairs in terms of speech and quality of life (QOL) still remain largely unstudied. In the current study, questionnaires were used to assess the patients' perception of speech and QOL before and after primary palate repair. All of the patients were operated at a cleft center in northeast India and had a cleft palate with a normal lip or with a cleft lip that had been previously repaired. A total of 134 patients (7-35 years) were interviewed preoperatively and 46 patients (7-32 years) were assessed in the postoperative survey. The survey showed that scores based on the speech handicap index, concerning speech and speech-related QOL, did not improve postoperatively. In fact, the questionnaires indicated that the speech became more unpredictable (P reported that their self-confidence had improved after the operation. Thus, the majority of interviewed patients who underwent late primary palate repair were satisfied with the surgery. At the same time, speech and speech-related QOL did not improve according to the speech handicap index-based survey. Speech predictability may even become worse and nasal regurgitation may increase after late palate repair, according to these results.
Jerger, Susan; Damian, Markus F.; McAlpine, Rachel P.; Abdi, Herve
2018-01-01
To communicate, children must discriminate and identify speech sounds. Because visual speech plays an important role in this process, we explored how visual speech influences phoneme discrimination and identification by children. Critical items had intact visual speech (e.g. baez) coupled to non-intact (excised onsets) auditory speech (signified…
Tackling the complexity in speech
DEFF Research Database (Denmark)
section includes four carefully selected chapters. They deal with facets of speech production, speech acoustics, and/or speech perception or recognition, place them in an integrated phonetic-phonological perspective, and relate them in more or less explicit ways to aspects of speech technology. Therefore......, we hope that this volume can help speech scientists with traditional training in phonetics and phonology to keep up with the latest developments in speech technology. In the opposite direction, speech researchers starting from a technological perspective will hopefully get inspired by reading about...... the questions, phenomena, and communicative functions that are currently addressed in phonetics and phonology. Either way, the future of speech research lies in international, interdisciplinary collaborations, and our volume is meant to reflect and facilitate such collaborations...
Automatic Speech Acquisition and Recognition for Spacesuit Audio Systems
Ye, Sherry
2015-01-01
NASA has a widely recognized but unmet need for novel human-machine interface technologies that can facilitate communication during astronaut extravehicular activities (EVAs), when loud noises and strong reverberations inside spacesuits make communication challenging. WeVoice, Inc., has developed a multichannel signal-processing method for speech acquisition in noisy and reverberant environments that enables automatic speech recognition (ASR) technology inside spacesuits. The technology reduces noise by exploiting differences between the statistical nature of signals (i.e., speech) and noise that exists in the spatial and temporal domains. As a result, ASR accuracy can be improved to the level at which crewmembers will find the speech interface useful. System components and features include beam forming/multichannel noise reduction, single-channel noise reduction, speech feature extraction, feature transformation and normalization, feature compression, and ASR decoding. Arithmetic complexity models were developed and will help designers of real-time ASR systems select proper tasks when confronted with constraints in computational resources. In Phase I of the project, WeVoice validated the technology. The company further refined the technology in Phase II and developed a prototype for testing and use by suited astronauts.
Directory of Open Access Journals (Sweden)
Behrooz Mahmoodi Bakhtiari
2012-10-01
Full Text Available Background and Aim: Recently, researchers have increasingly turned to study the relation between stuttering and syntactic complexity. This study investigates the effect of syntactic complexity on theamount of speech dysfluency in stuttering Persian-speaking children and adults in conversational speech. The obtained results can pave the way to a better understanding of stuttering in children andadults, and finding more appropriate treatments.Methods: In this cross-sectional study, the participants were 15 stuttering adult Persian-speakers, older than 15 years, and 15 stuttering child Persian-speakers of 4-6 years of age. In this study, first a 30 minute sample of the spontaneous speech of the participants was provided. Then the utterances of each person were studied in respect to the amount of dysfluency and syntactic complexity. The obtained information was analyzed using paired samples t-test.Results: In both groups of stuttering children and adults, there was a significant difference between the amount of dysfluency of simple and complex sentences (p<0.05.Conclusion: The results of this study showed that an increase in syntactic complexity in conversational speech, increased the amount of dysfluency in stuttering children and adults. Moreover,as a result of increase of syntactic complexity, dysfluency had a greater increase in stuttering children than stuttering adults.
Speech in spinocerebellar ataxia.
Schalling, Ellika; Hartelius, Lena
2013-12-01
Spinocerebellar ataxias (SCAs) are a heterogeneous group of autosomal dominant cerebellar ataxias clinically characterized by progressive ataxia, dysarthria and a range of other concomitant neurological symptoms. Only a few studies include detailed characterization of speech symptoms in SCA. Speech symptoms in SCA resemble ataxic dysarthria but symptoms related to phonation may be more prominent. One study to date has shown an association between differences in speech and voice symptoms related to genotype. More studies of speech and voice phenotypes are motivated, to possibly aid in clinical diagnosis. In addition, instrumental speech analysis has been demonstrated to be a reliable measure that may be used to monitor disease progression or therapy outcomes in possible future pharmacological treatments. Intervention by speech and language pathologists should go beyond assessment. Clinical guidelines for management of speech, communication and swallowing need to be developed for individuals with progressive cerebellar ataxia. Copyright © 2013 Elsevier Inc. All rights reserved.
Maeda, Yukihide; Takao, Soshi; Sugaya, Akiko; Kataoka, Yuko; Kariya, Shin; Tanaka, Satomi; Nagayasu, Rie; Nakagawa, Atsuko; Nishizaki, Kazunori
2018-02-01
To clarify how the pure-tone threshold (PTT) on the PTA predicts speech perception (SP) in elderly Japanese persons. Data on PTT and SP were cross-sectionally analyzed in Japanese persons (656 ears in 353 patients, aged ≥65 years). Correlations of SP and average PTT in all tested frequencies were evaluated by Pearson's correlation coefficient and simple linear regression. After adjusting for sex, laterality of ears, and age, the relationship of average and frequency-specific PTT with impaired SP ≤50% was estimated by logistic regression models. SP correlated well (r = -0.699) with the average PTT of all tested frequencies. On the other hand, the correlation between patient age and SP was weak, especially among ≤85-year-old persons (r = -0.092). Linear regression showed that the average PTT corresponding to SP of 50% was 76.4 dB nHL. Odds ratios for impaired SP were highest for PTT at 2000 Hz. Odds ratios were higher for middle (500, 1000, 2000 Hz) and high frequencies (4000, 8000 Hz) than low frequencies (125, 250 Hz). The PTT on the pure-tone audiogram (PTA) is a good predictor of SP by speech audiometry among older persons, which could provide clinically important information for hearing aid fitting and cochlear implantation.
Why the Left Hemisphere Is Dominant for Speech Production: Connecting the Dots
Directory of Open Access Journals (Sweden)
Harvey Martin Sussman
2015-12-01
Full Text Available Evidence from seemingly disparate areas of speech/language research is reviewed to form a unified theoretical account for why the left hemisphere is specialized for speech production. Research findings from studies investigating hemispheric lateralization of infant babbling, the primacy of the syllable in phonological structure, rhyming performance in split-brain patients, rhyming ability and phonetic categorization in children diagnosed with developmental apraxia of speech, rules governing exchange errors in spoonerisms, organizational principles of neocortical control of learned motor behaviors, and multi-electrode recordings of human neuronal responses to speech sounds are described and common threads highlighted. It is suggested that the emergence, in developmental neurogenesis, of a hard-wired, syllabically-organized, neural substrate representing the phonemic sound elements of one’s language, particularly the vocalic nucleus, is the crucial factor underlying the left hemisphere’s dominance for speech production.
Lee, Jimin; Hustad, Katherine C.; Weismer, Gary
2014-01-01
Purpose: Speech acoustic characteristics of children with cerebral palsy (CP) were examined with a multiple speech subsystems approach; speech intelligibility was evaluated using a prediction model in which acoustic measures were selected to represent three speech subsystems. Method: Nine acoustic variables reflecting different subsystems, and…
Stasenko, Alena; Bonn, Cory; Teghipco, Alex; Garcea, Frank E; Sweet, Catherine; Dombovy, Mary; McDonough, Joyce; Mahon, Bradford Z
2015-01-01
The debate about the causal role of the motor system in speech perception has been reignited by demonstrations that motor processes are engaged during the processing of speech sounds. Here, we evaluate which aspects of auditory speech processing are affected, and which are not, in a stroke patient with dysfunction of the speech motor system. We found that the patient showed a normal phonemic categorical boundary when discriminating two non-words that differ by a minimal pair (e.g., ADA-AGA). However, using the same stimuli, the patient was unable to identify or label the non-word stimuli (using a button-press response). A control task showed that he could identify speech sounds by speaker gender, ruling out a general labelling impairment. These data suggest that while the motor system is not causally involved in perception of the speech signal, it may be used when other cues (e.g., meaning, context) are not available.
Choice in maternity: rhetoric, reality and resistance.
Mander, Rosemary; Melender, Hanna-Leena
2009-12-01
to inform the organisation of the maternity services in Scotland, a phenomenological study was planned to examine maternity decision making in two similarly small countries. The aim was to examine the experience of contributing to decisions at clinical, organisational and policy-making levels. When examples were needed the informants were asked to use their experience of place of birth decisions. a hermeneutic phenomenological approach was employed. In-depth, semi-structured conversations were used. The fieldwork extended over a 4-month period in 2005. The data were analysed using Colaizzi's method. Finland and New Zealand were chosen because the parallels in their health care and maternity care systems would limit disparities. In one of the Finnish centres, the findings were particularly homogeneous and exemplified many of the issues arising in other settings. The findings of the fieldwork in this Finnish centre are the focus of this paper. the informants were mothers, midwife managers/policy makers, midwives and other maternity care providers. The findings of 12 conversations, including mothers and all groups of staff, are reported here. the background theme which emerged was 'trusting the system'. The informants were aware of the extent to which change is happening. One of the sub-themes contrasted the informants' perceptions of their lack of strength and courage with Finnish stereotypes. Being safe proved to be another crucial issue. The final sub-theme was 'playing the system'. trust in a well-respected health-care system was necessary for the informants to be able to subvert or resist that system. While such resistance has been documented in other disciplines, such as nursing, reference has not been found in relation to maternity. The resistance to the system was, at the time of the fieldwork, neither co-ordinated nor collaborative. the findings of this study carry important implications for women's and midwives' input into maternity care.
The Relationship between Speech Production and Speech Perception Deficits in Parkinson's Disease
De Keyser, Kim; Santens, Patrick; Bockstael, Annelies; Botteldooren, Dick; Talsma, Durk; De Vos, Stefanie; Van Cauwenberghe, Mieke; Verheugen, Femke; Corthals, Paul; De Letter, Miet
2016-01-01
Purpose: This study investigated the possible relationship between hypokinetic speech production and speech intensity perception in patients with Parkinson's disease (PD). Method: Participants included 14 patients with idiopathic PD and 14 matched healthy controls (HCs) with normal hearing and cognition. First, speech production was objectified…
The treatment of apraxia of speech : Speech and music therapy, an innovative joint effort
Hurkmans, Josephus Johannes Stephanus
2016-01-01
Apraxia of Speech (AoS) is a neurogenic speech disorder. A wide variety of behavioural methods have been developed to treat AoS. Various therapy programmes use musical elements to improve speech production. A unique therapy programme combining elements of speech therapy and music therapy is called
Verguet, Stéphane; Nandi, Arindam; Filippi, Véronique; Bundy, Donald A P
2016-09-26
High levels of maternal mortality and large associated inequalities exist in low-income and middle-income countries. Adolescent pregnancies remain common, and pregnant adolescent women face elevated risks of maternal mortality and poverty. We examined the distribution across socioeconomic groups of maternal deaths and impoverishment among adolescent girls (15-19 years old) in Niger, which has the highest total fertility rate globally, and India, which has the largest number of maternal deaths. In Niger and India, among adolescent girls, we estimated the distribution per income quintile of: the number of maternal deaths; and the impoverishment, measured by calculating the number of cases of catastrophic health expenditure incurred, caused by complicated pregnancies. We also examined the potential impact on maternal deaths and poverty of increasing adolescent girls' level of education by 1 year. We used epidemiological and cost inputs sourced from surveys and the literature. The number of maternal deaths would be larger among the poorer adolescents than among the richer adolescents in Niger and India. Impoverishment would largely incur among the richer adolescents in Niger and among the poorer adolescents in India. Increasing educational attainment of adolescent girls might avert both a large number of maternal deaths and a significant number of cases of catastrophic health expenditure in the 2 countries. Adolescent pregnancies can lead to large equity gaps and substantial impoverishment in low-income and middle-income countries. Increasing female education can reduce such inequalities and provide financial risk protection and poverty alleviation to adolescent girls. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/
Goffman, L
1999-12-01
It is often hypothesized that young children's difficulties with producing weak-strong (iambic) prosodic forms arise from perceptual or linguistically based production factors. A third possible contributor to errors in the iambic form may be biological constraints, or biases, of the motor system. In the present study, 7 children with specific language impairment (SLI) and speech deficits were matched to same age peers. Multiple levels of analysis, including kinematic (modulation and stability of movement), acoustic, and transcription, were applied to children's productions of iambic (weak-strong) and trochaic (strong-weak) prosodic forms. Findings suggest that a motor bias toward producing unmodulated rhythmic articulatory movements, similar to that observed in canonical babbling, contribute to children's acquisition of metrical forms. Children with SLI and speech deficits show less mature segmental and speech motor systems, as well as decreased modulation of movement in later developing iambic forms. Further, components of prosodic and segmental acquisition develop independently and at different rates.
Poole, Matthew L.; Brodtmann, Amy; Darby, David; Vogel, Adam P.
2017-01-01
Purpose: Our purpose was to create a comprehensive review of speech impairment in frontotemporal dementia (FTD), primary progressive aphasia (PPA), and progressive apraxia of speech in order to identify the most effective measures for diagnosis and monitoring, and to elucidate associations between speech and neuroimaging. Method: Speech and…
Transformation of Flaubert’s Free Indirect Speech in Film Adaptation Madame Bovary by Claude Chabrol
Florence Gacoin Marks
2013-01-01
The paper deals with the transformation of Flaubert’s free indirect speech in the film Madame Bovary by Claude Chabrol. Conversion of free indirect speech into direct speech or into narration by an external narrator (voice-over) cannot be avoided, it does, however, pose many problems because of the potential ambiguousness (polyphony) of free indirect speech. In such cases, Chabrol often finds effective solutions which bring the film closer to Flaubert’s style. Nevertheless, it remains clear t...
An analysis of the masking of speech by competing speech using self-report data.
Agus, Trevor R; Akeroyd, Michael A; Noble, William; Bhullar, Navjot
2009-01-01
Many of the items in the "Speech, Spatial, and Qualities of Hearing" scale questionnaire [S. Gatehouse and W. Noble, Int. J. Audiol. 43, 85-99 (2004)] are concerned with speech understanding in a variety of backgrounds, both speech and nonspeech. To study if this self-report data reflected informational masking, previously collected data on 414 people were analyzed. The lowest scores (greatest difficulties) were found for the two items in which there were two speech targets, with successively higher scores for competing speech (six items), energetic masking (one item), and no masking (three items). The results suggest significant masking by competing speech in everyday listening situations.
Sensorimotor oscillations prior to speech onset reflect altered motor networks in adults who stutter
Directory of Open Access Journals (Sweden)
Anna-Maria Mersov
2016-09-01
Full Text Available Adults who stutter (AWS have demonstrated atypical coordination of motor and sensory regions during speech production. Yet little is known of the speech-motor network in AWS in the brief time window preceding audible speech onset. The purpose of the current study was to characterize neural oscillations in the speech-motor network during preparation for and execution of overt speech production in AWS using magnetoencephalography (MEG. Twelve AWS and twelve age-matched controls were presented with 220 words, each word embedded in a carrier phrase. Controls were presented with the same word list as their matched AWS participant. Neural oscillatory activity was localized using minimum-variance beamforming during two time periods of interest: speech preparation (prior to speech onset and speech execution (following speech onset. Compared to controls, AWS showed stronger beta (15-25Hz suppression in the speech preparation stage, followed by stronger beta synchronization in the bilateral mouth motor cortex. AWS also recruited the right mouth motor cortex significantly earlier in the speech preparation stage compared to controls. Exaggerated motor preparation is discussed in the context of reduced coordination in the speech-motor network of AWS. It is further proposed that exaggerated beta synchronization may reflect a more strongly inhibited motor system that requires a stronger beta suppression to disengage prior to speech initiation. These novel findings highlight critical differences in the speech-motor network of AWS that occur prior to speech onset and emphasize the need to investigate further the speech-motor assembly in the stuttering population.
The pathways for intelligible speech: multivariate and univariate perspectives.
Evans, S; Kyong, J S; Rosen, S; Golestani, N; Warren, J E; McGettigan, C; Mourão-Miranda, J; Wise, R J S; Scott, S K
2014-09-01
An anterior pathway, concerned with extracting meaning from sound, has been identified in nonhuman primates. An analogous pathway has been suggested in humans, but controversy exists concerning the degree of lateralization and the precise location where responses to intelligible speech emerge. We have demonstrated that the left anterior superior temporal sulcus (STS) responds preferentially to intelligible speech (Scott SK, Blank CC, Rosen S, Wise RJS. 2000. Identification of a pathway for intelligible speech in the left temporal lobe. Brain. 123:2400-2406.). A functional magnetic resonance imaging study in Cerebral Cortex used equivalent stimuli and univariate and multivariate analyses to argue for the greater importance of bilateral posterior when compared with the left anterior STS in responding to intelligible speech (Okada K, Rong F, Venezia J, Matchin W, Hsieh IH, Saberi K, Serences JT,Hickok G. 2010. Hierarchical organization of human auditory cortex: evidence from acoustic invariance in the response to intelligible speech. 20: 2486-2495.). Here, we also replicate our original study, demonstrating that the left anterior STS exhibits the strongest univariate response and, in decoding using the bilateral temporal cortex, contains the most informative voxels showing an increased response to intelligible speech. In contrast, in classifications using local "searchlights" and a whole brain analysis, we find greater classification accuracy in posterior rather than anterior temporal regions. Thus, we show that the precise nature of the multivariate analysis used will emphasize different response profiles associated with complex sound to speech processing. © The Author 2013. Published by Oxford University Press.
SPEECH DISORDERS IN PRIMARY SCHOOL STUDENTS OF ISFAHAN (1998-9
Directory of Open Access Journals (Sweden)
B SHAFIEI
2002-06-01
Full Text Available Introduction. The aim of this study was to describe frequency of speech disorders in primary school students.
Methods. In a cross-sectional study 300 first and second grade primary school students were examined for speech disorders.
Results. From 300 subjects, 280 were normal (without speech disorders, 15 had articulation disorders, 2 had voice disorders, 3 had resonance disorders and no one had fluency disorders.
Discussion. The findings of this study are supported by fomer study in other countries, except frequency of fluency disorders which may due to low sample size of present study.
Speech and language development in 2-year-old children with cerebral palsy.
Hustad, Katherine C; Allison, Kristen; McFadd, Emily; Riehle, Katherine
2014-06-01
We examined early speech and language development in children who had cerebral palsy. Questions addressed whether children could be classified into early profile groups on the basis of speech and language skills and whether there were differences on selected speech and language measures among groups. Speech and language assessments were completed on 27 children with CP who were between the ages of 24 and 30 months (mean age 27.1 months; SD 1.8). We examined several measures of expressive and receptive language, along with speech intelligibility. Two-step cluster analysis was used to identify homogeneous groups of children based on their performance on the seven dependent variables characterizing speech and language performance. Three groups of children identified were those not yet talking (44% of the sample); those whose talking abilities appeared to be emerging (41% of the sample); and those who were established talkers (15% of the sample). Group differences were evident on all variables except receptive language skills. 85% of 2-year-old children with CP in this study had clinical speech and/or language delays relative to age expectations. Findings suggest that children with CP should receive speech and language assessment and treatment at or before 2 years of age.
Neural pathways for visual speech perception
Directory of Open Access Journals (Sweden)
Lynne E Bernstein
2014-12-01
Full Text Available This paper examines the questions, what levels of speech can be perceived visually, and how is visual speech represented by the brain? Review of the literature leads to the conclusions that every level of psycholinguistic speech structure (i.e., phonetic features, phonemes, syllables, words, and prosody can be perceived visually, although individuals differ in their abilities to do so; and that there are visual modality-specific representations of speech qua speech in higher-level vision brain areas. That is, the visual system represents the modal patterns of visual speech. The suggestion that the auditory speech pathway receives and represents visual speech is examined in light of neuroimaging evidence on the auditory speech pathways. We outline the generally agreed-upon organization of the visual ventral and dorsal pathways and examine several types of visual processing that might be related to speech through those pathways, specifically, face and body, orthography, and sign language processing. In this context, we examine the visual speech processing literature, which reveals widespread diverse patterns activity in posterior temporal cortices in response to visual speech stimuli. We outline a model of the visual and auditory speech pathways and make several suggestions: (1 The visual perception of speech relies on visual pathway representations of speech qua speech. (2 A proposed site of these representations, the temporal visual speech area (TVSA has been demonstrated in posterior temporal cortex, ventral and posterior to multisensory posterior superior temporal sulcus (pSTS. (3 Given that visual speech has dynamic and configural features, its representations in feedforward visual pathways are expected to integrate these features, possibly in TVSA.
Gonzalez, Jorge E.; Acosta, Sandra; Davis, Heather; Pollard-Durodola, Sharolyn; Saenz, Laura; Soares, Denise; Resendez, Nora; Zhu, Leina
2017-01-01
Research Findings: This study investigated the association between Mexican American maternal education and socioeconomic status (SES) and child vocabulary as mediated by parental reading beliefs, home literacy environment (HLE), and parent-child shared reading frequency. As part of a larger study, maternal reports of education level, SES, HLE, and…
Preschoolers' Emergent Literacy Skills: The Mediating Role of Maternal Reading Beliefs
Cottone, Elizabeth Ann
2012-01-01
Research Findings: The purpose of this paper is to explore the association between maternal reading beliefs and children's emergent literacy outcomes in light of maternal education. Furthermore, I consider whether maternal reading beliefs may mediate the association between maternal education level and children's print knowledge and phonological…
Part-of-speech effects on text-to-speech synthesis
CSIR Research Space (South Africa)
Schlunz, GI
2010-11-01
Full Text Available One of the goals of text-to-speech (TTS) systems is to produce natural-sounding synthesised speech. Towards this end various natural language processing (NLP) tasks are performed to model the prosodic aspects of the TTS voice. One of the fundamental...
2010-05-12
...] Telecommunications Relay Services and Speech-to-Speech Services for Individuals With Hearing and Speech Disabilities... proposed compensation rates for Interstate TRS, Speech-to-Speech Services (STS), Captioned Telephone... costs reported in the data submitted to NECA by VRS providers. In this regard, document DA 10-761 also...
Speech Rate Entrainment in Children and Adults With and Without Autism Spectrum Disorder.
Wynn, Camille J; Borrie, Stephanie A; Sellers, Tyra P
2018-05-03
Conversational entrainment, a phenomenon whereby people modify their behaviors to match their communication partner, has been evidenced as critical to successful conversation. It is plausible that deficits in entrainment contribute to the conversational breakdowns and social difficulties exhibited by people with autism spectrum disorder (ASD). This study examined speech rate entrainment in children and adult populations with and without ASD. Sixty participants including typically developing children, children with ASD, typically developed adults, and adults with ASD participated in a quasi-conversational paradigm with a pseudoconfederate. The confederate's speech rate was digitally manipulated to create slow and fast speech rate conditions. Typically developed adults entrained their speech rate in the quasi-conversational paradigm, using a faster rate during the fast speech rate conditions and a slower rate during the slow speech rate conditions. This entrainment pattern was not evident in adults with ASD or in children populations. Findings suggest that speech rate entrainment is a developmentally acquired skill and offers preliminary evidence of speech rate entrainment deficits in adults with ASD. Impairments in this area may contribute to the conversational breakdowns and social difficulties experienced by this population. Future work is needed to advance this area of inquiry.
Gallardo, L.F.; Möller, S.; Beerends, J.
2017-01-01
The performance of automatic speech recognition based on coded-decoded speech heavily depends on the quality of the transmitted signals, determined by channel impairments. This paper examines relationships between speech recognition performance and measurements of speech quality and intelligibility
[Non-speech oral motor treatment efficacy for children with developmental speech sound disorders].
Ygual-Fernandez, A; Cervera-Merida, J F
2016-01-01
In the treatment of speech disorders by means of speech therapy two antagonistic methodological approaches are applied: non-verbal ones, based on oral motor exercises (OME), and verbal ones, which are based on speech processing tasks with syllables, phonemes and words. In Spain, OME programmes are called 'programas de praxias', and are widely used and valued by speech therapists. To review the studies conducted on the effectiveness of OME-based treatments applied to children with speech disorders and the theoretical arguments that could justify, or not, their usefulness. Over the last few decades evidence has been gathered about the lack of efficacy of this approach to treat developmental speech disorders and pronunciation problems in populations without any neurological alteration of motor functioning. The American Speech-Language-Hearing Association has advised against its use taking into account the principles of evidence-based practice. The knowledge gathered to date on motor control shows that the pattern of mobility and its corresponding organisation in the brain are different in speech and other non-verbal functions linked to nutrition and breathing. Neither the studies on their effectiveness nor the arguments based on motor control studies recommend the use of OME-based programmes for the treatment of pronunciation problems in children with developmental language disorders.
Speech parts as Poisson processes.
Badalamenti, A F
2001-09-01
This paper presents evidence that six of the seven parts of speech occur in written text as Poisson processes, simple or recurring. The six major parts are nouns, verbs, adjectives, adverbs, prepositions, and conjunctions, with the interjection occurring too infrequently to support a model. The data consist of more than the first 5000 words of works by four major authors coded to label the parts of speech, as well as periods (sentence terminators). Sentence length is measured via the period and found to be normally distributed with no stochastic model identified for its occurrence. The models for all six speech parts but the noun significantly distinguish some pairs of authors and likewise for the joint use of all words types. Any one author is significantly distinguished from any other by at least one word type and sentence length very significantly distinguishes each from all others. The variety of word type use, measured by Shannon entropy, builds to about 90% of its maximum possible value. The rate constants for nouns are close to the fractions of maximum entropy achieved. This finding together with the stochastic models and the relations among them suggest that the noun may be a primitive organizer of written text.
Acoustic properties of naturally produced clear speech at normal speaking rates
Krause, Jean C.; Braida, Louis D.
2004-01-01
Sentences spoken ``clearly'' are significantly more intelligible than those spoken ``conversationally'' for hearing-impaired listeners in a variety of backgrounds [Picheny et al., J. Speech Hear. Res. 28, 96-103 (1985); Uchanski et al., ibid. 39, 494-509 (1996); Payton et al., J. Acoust. Soc. Am. 95, 1581-1592 (1994)]. While producing clear speech, however, talkers often reduce their speaking rate significantly [Picheny et al., J. Speech Hear. Res. 29, 434-446 (1986); Uchanski et al., ibid. 39, 494-509 (1996)]. Yet speaking slowly is not solely responsible for the intelligibility benefit of clear speech (over conversational speech), since a recent study [Krause and Braida, J. Acoust. Soc. Am. 112, 2165-2172 (2002)] showed that talkers can produce clear speech at normal rates with training. This finding suggests that clear speech has inherent acoustic properties, independent of rate, that contribute to improved intelligibility. Identifying these acoustic properties could lead to improved signal processing schemes for hearing aids. To gain insight into these acoustical properties, conversational and clear speech produced at normal speaking rates were analyzed at three levels of detail (global, phonological, and phonetic). Although results suggest that talkers may have employed different strategies to achieve clear speech at normal rates, two global-level properties were identified that appear likely to be linked to the improvements in intelligibility provided by clear/normal speech: increased energy in the 1000-3000-Hz range of long-term spectra and increased modulation depth of low frequency modulations of the intensity envelope. Other phonological and phonetic differences associated with clear/normal speech include changes in (1) frequency of stop burst releases, (2) VOT of word-initial voiceless stop consonants, and (3) short-term vowel spectra.
Speech and Hearing Problems among Older People.
Carstenson, Blue
1978-01-01
Findings from speech and hearing tests of older people in South Dakota community senior programs indicate the need for better testing and therapy procedures. Lipreading may be more effective than hearing aids, and factors other than hearing may be involved. Some problems and needs are noted. (MF)
2010-09-03
...] Telecommunications Relay Services and Speech-to-Speech Services for Individuals With Hearing and Speech Disabilities...; speech-to-speech (STS); pay-per-call (900) calls; types of calls; and equal access to interexchange... of a report, due April 16, 2011, addressing whether it is necessary for the waivers to remain in...
Harrison, Linda J; McLeod, Sharynne
2010-04-01
To determine risk and protective factors for speech and language impairment in early childhood. Data are presented for a nationally representative sample of 4,983 children participating in the Longitudinal Study of Australian Children (described in McLeod & Harrison, 2009). Thirty-one child, parent, family, and community factors previously reported as being predictors of speech and language impairment were tested as predictors of (a) parent-rated expressive speech/language concern and (b) receptive language concern, (c) use of speech-language pathology services, and (d) low receptive vocabulary. Bivariate logistic regression analyses confirmed 29 of the identified factors. However, when tested concurrently with other predictors in multivariate analyses, only 19 remained significant: 9 for 2-4 outcomes and 10 for 1 outcome. Consistent risk factors were being male, having ongoing hearing problems, and having a more reactive temperament. Protective factors were having a more persistent and sociable temperament and higher levels of maternal well-being. Results differed by outcome for having an older sibling, parents speaking a language other than English, and parental support for children's learning at home. Identification of children requiring speech and language assessment requires consideration of the context of family life as well as biological and psychosocial factors intrinsic to the child.
An investigation of maternal food intake and maternal food talk as predictors of child food intake.
DeJesus, Jasmine M; Gelman, Susan A; Viechnicki, Gail B; Appugliese, Danielle P; Miller, Alison L; Rosenblum, Katherine L; Lumeng, Julie C
2018-08-01
Though parental modeling is thought to play a critical role in promoting children's healthy eating, little research has examined maternal food intake and maternal food talk as independent predictors of children's food intake. The present study examines maternal food talk during a structured eating protocol, in which mothers and their children had the opportunity to eat a series of familiar and unfamiliar vegetables and desserts. Several aspects of maternal talk during the protocol were coded, including overall food talk, directives, pronoun use, and questions. This study analyzed the predictors of maternal food talk and whether maternal food talk and maternal food intake predicted children's food intake during the protocol. Higher maternal body mass index (BMI) predicted lower amounts of food talk, pronoun use, and questions. Higher child BMI z-scores predicted more first person pronouns and more wh-questions within maternal food talk. Mothers of older children used fewer directives, fewer second person pronouns, and fewer yes/no questions. However, maternal food talk (overall and specific types of food talk) did not predict children's food intake. Instead, the most robust predictor of children's food intake during this protocol was the amount of food that mothers ate while sitting with their children. These findings emphasize the importance of modeling healthy eating through action and have implications for designing interventions to provide parents with more effective tools to promote their children's healthy eating. Copyright © 2018 Elsevier Ltd. All rights reserved.
Maternal Education Gradients in Infant Health in Four South American Countries.
Wehby, George L; López-Camelo, Jorge S
2017-11-01
Objective We investigate gradients (i.e. differences) in infant health outcomes by maternal education in Argentina, Brazil, Chile, and Venezuela and explore channels related to father's education, household labor outcomes, and maternal health, fertility, and use of prenatal services and technology. Methods We employ secondary interview and birth record data similarly collected across a network of birth hospitals from the early 1980s through 2011 within the Latin American Collaborative Study of Congenital Anomalies (ECLAMC). Focusing on children without birth defects, we estimate gradients in several infant health outcomes including birth weight, gestational age, and hospital discharge status by maternal education using ordinary least squares regression models adjusting for several demographic factors. To explore channels, we add as covariates father's education, parental occupational activity, maternal health and fertility history, and use of prenatal services and technology and evaluate changes in the coefficient of maternal education. We use the same models for each country sample. Results We find important differences in gradients across countries. We find evidence for educational gradients in preterm birth in three countries but weaker evidence for gradients in fetal growth. The extent to which observed household and maternal factors explain these gradients based on changes in the regression coefficient of maternal education when controlling for these factors as covariates also varies between countries. In contrast, we generally find evidence across all countries that higher maternal education is associated with increased use of prenatal care services and technology. Conclusions Our findings suggest that differences in infant health by maternal education and their underlying mechanisms vary and are not necessarily generalizable across countries. However, the positive association between maternal education and use of prenatal services and technology is more
A Joint Approach for Single-Channel Speaker Identification and Speech Separation
DEFF Research Database (Denmark)
Mowlaee, Pejman; Saeidi, Rahim; Christensen, Mads Græsbøll
2012-01-01
) accuracy, here, we report the objective and subjective results as well. The results show that the proposed system performs as well as the best of the state-of-the-art in terms of perceived quality while its performance in terms of speaker identification and automatic speech recognition results......In this paper, we present a novel system for joint speaker identification and speech separation. For speaker identification a single-channel speaker identification algorithm is proposed which provides an estimate of signal-to-signal ratio (SSR) as a by-product. For speech separation, we propose...... a sinusoidal model-based algorithm. The speech separation algorithm consists of a double-talk/single-talk detector followed by a minimum mean square error estimator of sinusoidal parameters for finding optimal codevectors from pre-trained speaker codebooks. In evaluating the proposed system, we start from...
Environmental Contamination of Normal Speech.
Harley, Trevor A.
1990-01-01
Environmentally contaminated speech errors (irrelevant words or phrases derived from the speaker's environment and erroneously incorporated into speech) are hypothesized to occur at a high level of speech processing, but with a relatively late insertion point. The data indicate that speech production processes are not independent of other…
Perriman, Noelyn; Davis, Deborah
2016-06-01
The objective of this systematic integrative review is to identify, summarise and communicate the findings of research relating to tools that measure maternal satisfaction with continuity of maternity care models. In so doing the most appropriate, reliable and valid tool that can be used to measure maternal satisfaction with continuity of maternity care will be determined. A systematic integrative review of published and unpublished literature was undertaken using selected databases. Research papers were included if they measured maternal satisfaction in a continuity model of maternity care, were published in English after 1999 and if they included (or made available) the instrument used to measure satisfaction. Six hundred and thirty two unique papers were identified and after applying the selection criteria, four papers were included in the review. Three of these originated in Australia and one in Canada. The primary focus of all papers was not on the development of a tool to measure maternal satisfaction but on the comparison of outcomes in different models of care. The instruments developed varied in terms of the degree to which they were tested for validity and reliability. Women's satisfaction with maternity services is an important measure of quality. Most satisfaction surveys in maternity appear to reflect fragmented models of care though continuity of care models are increasing in line with the evidence demonstrating their effectiveness. It is important that robust tools are developed for this context and that there is some consistency in the way this is measured and reported for the purposes of benchmarking and quality improvement. Copyright © 2016 Australian College of Midwives. Published by Elsevier Ltd. All rights reserved.
When speaker identity is unavoidable: Neural processing of speaker identity cues in natural speech.
Tuninetti, Alba; Chládková, Kateřina; Peter, Varghese; Schiller, Niels O; Escudero, Paola
2017-11-01
Speech sound acoustic properties vary largely across speakers and accents. When perceiving speech, adult listeners normally disregard non-linguistic variation caused by speaker or accent differences, in order to comprehend the linguistic message, e.g. to correctly identify a speech sound or a word. Here we tested whether the process of normalizing speaker and accent differences, facilitating the recognition of linguistic information, is found at the level of neural processing, and whether it is modulated by the listeners' native language. In a multi-deviant oddball paradigm, native and nonnative speakers of Dutch were exposed to naturally-produced Dutch vowels varying in speaker, sex, accent, and phoneme identity. Unexpectedly, the analysis of mismatch negativity (MMN) amplitudes elicited by each type of change shows a large degree of early perceptual sensitivity to non-linguistic cues. This finding on perception of naturally-produced stimuli contrasts with previous studies examining the perception of synthetic stimuli wherein adult listeners automatically disregard acoustic cues to speaker identity. The present finding bears relevance to speech normalization theories, suggesting that at an unattended level of processing, listeners are indeed sensitive to changes in fundamental frequency in natural speech tokens. Copyright © 2017 Elsevier Inc. All rights reserved.
Lu, Lingxi; Bao, Xiaohan; Chen, Jing; Qu, Tianshu; Wu, Xihong; Li, Liang
2018-05-01
Under a noisy "cocktail-party" listening condition with multiple people talking, listeners can use various perceptual/cognitive unmasking cues to improve recognition of the target speech against informational speech-on-speech masking. One potential unmasking cue is the emotion expressed in a speech voice, by means of certain acoustical features. However, it was unclear whether emotionally conditioning a target-speech voice that has none of the typical acoustical features of emotions (i.e., an emotionally neutral voice) can be used by listeners for enhancing target-speech recognition under speech-on-speech masking conditions. In this study we examined the recognition of target speech against a two-talker speech masker both before and after the emotionally neutral target voice was paired with a loud female screaming sound that has a marked negative emotional valence. The results showed that recognition of the target speech (especially the first keyword in a target sentence) was significantly improved by emotionally conditioning the target speaker's voice. Moreover, the emotional unmasking effect was independent of the unmasking effect of the perceived spatial separation between the target speech and the masker. Also, (skin conductance) electrodermal responses became stronger after emotional learning when the target speech and masker were perceptually co-located, suggesting an increase of listening efforts when the target speech was informationally masked. These results indicate that emotionally conditioning the target speaker's voice does not change the acoustical parameters of the target-speech stimuli, but the emotionally conditioned vocal features can be used as cues for unmasking target speech.
Maternal Employment and Early Adolescent Substance Use.
Hillman, Stephen B.; Sawilowsky, Shlomo S.
1991-01-01
Examined effects of maternal employment on use of alcohol, cigarettes, marijuana, and other drugs by ninth graders (n=48). Comparison of maternal employment patterns (full-time versus part-time versus not employed outside the home) indicated no significant differences in substance use behavior among adolescents. Findings support literature on…
Multilevel Analysis in Analyzing Speech Data
Guddattu, Vasudeva; Krishna, Y.
2011-01-01
The speech produced by human vocal tract is a complex acoustic signal, with diverse applications in phonetics, speech synthesis, automatic speech recognition, speaker identification, communication aids, speech pathology, speech perception, machine translation, hearing research, rehabilitation and assessment of communication disorders and many…
Directory of Open Access Journals (Sweden)
Violeta TORTEVSKA
1997-06-01
Full Text Available The modern way of living in which the communication becomes a basic and upbringing factor and regulator of the relations isolates children with hard individual, family, educative and social problems.The speech and language disorders are the most remarkable symptoms pointing out the complex of defects in the communicative activities, reduced cognitive functions and cerebral dysfunction's.The modern conception in the rehabilitation field leads to a full engagement of the children’s closest environment and especially parents.The study will include the work of the speech therapist with children with a diagnosis tardy speech development (alalia and developing dysphasia in the hearing, speech and voice rehabilitation institute-Skopje, and its role introducing the parents for their right access and the systematic conduction of the rehabilitation proceedings-especially stimulating the motors and speech development.The speech therapist’s task is to find out a way and to apply means by which the children with central damages could build their speech and lingual system and to help the parents through instructive and advisory work into the comprehension of the phases and stages of that system.The conclusion is that the proceedings of the early treatment with the children with central damages are naturally caused by the difference of their early supplementation. The suggestions that are referring to what should be substituted, how much it should be substituted and how it should be done leads to the frames of the early therapeutical access.
Groenewold, Rimke; Bastiaanse, Roelien; Nickels, Lyndsey; Huiskes, Mike
2014-01-01
Background: Previous studies have shown that in semi-spontaneous speech, individuals with Broca's and anomic aphasia produce relatively many direct speech constructions. It has been claimed that in "healthy" communication direct speech constructions contribute to the liveliness, and indirectly to the comprehensibility, of speech.…
Speech Enhancement by MAP Spectral Amplitude Estimation Using a Super-Gaussian Speech Model
Directory of Open Access Journals (Sweden)
Lotter Thomas
2005-01-01
Full Text Available This contribution presents two spectral amplitude estimators for acoustical background noise suppression based on maximum a posteriori estimation and super-Gaussian statistical modelling of the speech DFT amplitudes. The probability density function of the speech spectral amplitude is modelled with a simple parametric function, which allows a high approximation accuracy for Laplace- or Gamma-distributed real and imaginary parts of the speech DFT coefficients. Also, the statistical model can be adapted to optimally fit the distribution of the speech spectral amplitudes for a specific noise reduction system. Based on the super-Gaussian statistical model, computationally efficient maximum a posteriori speech estimators are derived, which outperform the commonly applied Ephraim-Malah algorithm.
Directory of Open Access Journals (Sweden)
Vincent Aubanel
2016-08-01
Full Text Available A growing body of evidence shows that brain oscillations track speech. This mechanism is thought to maximise processing efficiency by allocating resources to important speech information, effectively parsing speech into units of appropriate granularity for further decoding. However, some aspects of this mechanism remain unclear. First, while periodicity is an intrinsic property of this physiological mechanism, speech is only quasi-periodic, so it is not clear whether periodicity would present an advantage in processing. Second, it is still a matter of debate which aspect of speech triggers or maintains cortical entrainment, from bottom-up cues such as fluctuations of the amplitude envelope of speech to higher level linguistic cues such as syntactic structure. We present data from a behavioural experiment assessing the effect of isochronous retiming of speech on speech perception in noise. Two types of anchor points were defined for retiming speech, namely syllable onsets and amplitude envelope peaks. For each anchor point type, retiming was implemented at two hierarchical levels, a slow time scale around 2.5 Hz and a fast time scale around 4 Hz. Results show that while any temporal distortion resulted in reduced speech intelligibility, isochronous speech anchored to P-centers (approximated by stressed syllable vowel onsets was significantly more intelligible than a matched anisochronous retiming, suggesting a facilitative role of periodicity defined on linguistically motivated units in processing speech in noise.
Broderick, Michael P; Anderson, Andrew J; Di Liberto, Giovanni M; Crosse, Michael J; Lalor, Edmund C
2018-03-05
People routinely hear and understand speech at rates of 120-200 words per minute [1, 2]. Thus, speech comprehension must involve rapid, online neural mechanisms that process words' meanings in an approximately time-locked fashion. However, electrophysiological evidence for such time-locked processing has been lacking for continuous speech. Although valuable insights into semantic processing have been provided by the "N400 component" of the event-related potential [3-6], this literature has been dominated by paradigms using incongruous words within specially constructed sentences, with less emphasis on natural, narrative speech comprehension. Building on the discovery that cortical activity "tracks" the dynamics of running speech [7-9] and psycholinguistic work demonstrating [10-12] and modeling [13-15] how context impacts on word processing, we describe a new approach for deriving an electrophysiological correlate of natural speech comprehension. We used a computational model [16] to quantify the meaning carried by words based on how semantically dissimilar they were to their preceding context and then regressed this measure against electroencephalographic (EEG) data recorded from subjects as they listened to narrative speech. This produced a prominent negativity at a time lag of 200-600 ms on centro-parietal EEG channels, characteristics common to the N400. Applying this approach to EEG datasets involving time-reversed speech, cocktail party attention, and audiovisual speech-in-noise demonstrated that this response was very sensitive to whether or not subjects understood the speech they heard. These findings demonstrate that, when successfully comprehending natural speech, the human brain responds to the contextual semantic content of each word in a relatively time-locked fashion. Copyright © 2018 Elsevier Ltd. All rights reserved.
DEFF Research Database (Denmark)
Poulsen, Torben
2000-01-01
An introduction is given to the the anatomy and the function of the ear, basic psychoacoustic matters (hearing threshold, loudness, masking), the speech signal and speech intelligibility. The lecture note is written for the course: Fundamentals of Acoustics and Noise Control (51001)......An introduction is given to the the anatomy and the function of the ear, basic psychoacoustic matters (hearing threshold, loudness, masking), the speech signal and speech intelligibility. The lecture note is written for the course: Fundamentals of Acoustics and Noise Control (51001)...
Automated analysis of free speech predicts psychosis onset in high-risk youths
Bedi, Gillinder; Carrillo, Facundo; Cecchi, Guillermo A; Slezak, Diego Fernández; Sigman, Mariano; Mota, Natália B; Ribeiro, Sidarta; Javitt, Daniel C; Copelli, Mauro; Corcoran, Cheryl M
2015-01-01
Background/Objectives: Psychiatry lacks the objective clinical tests routinely used in other specializations. Novel computerized methods to characterize complex behaviors such as speech could be used to identify and predict psychiatric illness in individuals. AIMS: In this proof-of-principle study, our aim was to test automated speech analyses combined with Machine Learning to predict later psychosis onset in youths at clinical high-risk (CHR) for psychosis. Methods: Thirty-four CHR youths (11 females) had baseline interviews and were assessed quarterly for up to 2.5 years; five transitioned to psychosis. Using automated analysis, transcripts of interviews were evaluated for semantic and syntactic features predicting later psychosis onset. Speech features were fed into a convex hull classification algorithm with leave-one-subject-out cross-validation to assess their predictive value for psychosis outcome. The canonical correlation between the speech features and prodromal symptom ratings was computed. Results: Derived speech features included a Latent Semantic Analysis measure of semantic coherence and two syntactic markers of speech complexity: maximum phrase length and use of determiners (e.g., which). These speech features predicted later psychosis development with 100% accuracy, outperforming classification from clinical interviews. Speech features were significantly correlated with prodromal symptoms. Conclusions: Findings support the utility of automated speech analysis to measure subtle, clinically relevant mental state changes in emergent psychosis. Recent developments in computer science, including natural language processing, could provide the foundation for future development of objective clinical tests for psychiatry. PMID:27336038
Effect of gap detection threshold on consistency of speech in children with speech sound disorder.
Sayyahi, Fateme; Soleymani, Zahra; Akbari, Mohammad; Bijankhan, Mahmood; Dolatshahi, Behrooz
2017-02-01
The present study examined the relationship between gap detection threshold and speech error consistency in children with speech sound disorder. The participants were children five to six years of age who were categorized into three groups of typical speech, consistent speech disorder (CSD) and inconsistent speech disorder (ISD).The phonetic gap detection threshold test was used for this study, which is a valid test comprised six syllables with inter-stimulus intervals between 20-300ms. The participants were asked to listen to the recorded stimuli three times and indicate whether they heard one or two sounds. There was no significant difference between the typical and CSD groups (p=0.55), but there were significant differences in performance between the ISD and CSD groups and the ISD and typical groups (p=0.00). The ISD group discriminated between speech sounds at a higher threshold. Children with inconsistent speech errors could not distinguish speech sounds during time-limited phonetic discrimination. It is suggested that inconsistency in speech is a representation of inconsistency in auditory perception, which causes by high gap detection threshold. Copyright © 2016 Elsevier Ltd. All rights reserved.
Maternal Depression and Developmental Disability: Research Critique
Bailey, Donald B., Jr.; Golden, Robert N.; Roberts, Jane; Ford, Amy
2007-01-01
Maternal depression in families having a child with a disability has been the subject of considerable research over the past 25 years. This review was designed to describe the literature on maternal depression, critique its research methodology, identify consensus findings across studies, and make recommendations for future research. A particular…
McCreery, Ryan W.; Walker, Elizabeth A.; Spratford, Meredith; Oleson, Jacob; Bentler, Ruth; Holte, Lenore; Roush, Patricia
2015-01-01
Objectives Progress has been made in recent years in the provision of amplification and early intervention for children who are hard of hearing. However, children who use hearing aids (HA) may have inconsistent access to their auditory environment due to limitations in speech audibility through their HAs or limited HA use. The effects of variability in children’s auditory experience on parent-report auditory skills questionnaires and on speech recognition in quiet and in noise were examined for a large group of children who were followed as part of the Outcomes of Children with Hearing Loss study. Design Parent ratings on auditory development questionnaires and children’s speech recognition were assessed for 306 children who are hard of hearing. Children ranged in age from 12 months to 9 years of age. Three questionnaires involving parent ratings of auditory skill development and behavior were used, including the LittlEARS Auditory Questionnaire, Parents Evaluation of Oral/Aural Performance in Children Rating Scale, and an adaptation of the Speech, Spatial and Qualities of Hearing scale. Speech recognition in quiet was assessed using the Open and Closed set task, Early Speech Perception Test, Lexical Neighborhood Test, and Phonetically-balanced Kindergarten word lists. Speech recognition in noise was assessed using the Computer-Assisted Speech Perception Assessment. Children who are hard of hearing were compared to peers with normal hearing matched for age, maternal educational level and nonverbal intelligence. The effects of aided audibility, HA use and language ability on parent responses to auditory development questionnaires and on children’s speech recognition were also examined. Results Children who are hard of hearing had poorer performance than peers with normal hearing on parent ratings of auditory skills and had poorer speech recognition. Significant individual variability among children who are hard of hearing was observed. Children with greater
Conversation electrified: ERP correlates of speech act recognition in underspecified utterances.
Directory of Open Access Journals (Sweden)
Rosa S Gisladottir
Full Text Available The ability to recognize speech acts (verbal actions in conversation is critical for everyday interaction. However, utterances are often underspecified for the speech act they perform, requiring listeners to rely on the context to recognize the action. The goal of this study was to investigate the time-course of auditory speech act recognition in action-underspecified utterances and explore how sequential context (the prior action impacts this process. We hypothesized that speech acts are recognized early in the utterance to allow for quick transitions between turns in conversation. Event-related potentials (ERPs were recorded while participants listened to spoken dialogues and performed an action categorization task. The dialogues contained target utterances that each of which could deliver three distinct speech acts depending on the prior turn. The targets were identical across conditions, but differed in the type of speech act performed and how it fit into the larger action sequence. The ERP results show an early effect of action type, reflected by frontal positivities as early as 200 ms after target utterance onset. This indicates that speech act recognition begins early in the turn when the utterance has only been partially processed. Providing further support for early speech act recognition, actions in highly constraining contexts did not elicit an ERP effect to the utterance-final word. We take this to show that listeners can recognize the action before the final word through predictions at the speech act level. However, additional processing based on the complete utterance is required in more complex actions, as reflected by a posterior negativity at the final word when the speech act is in a less constraining context and a new action sequence is initiated. These findings demonstrate that sentence comprehension in conversational contexts crucially involves recognition of verbal action which begins as soon as it can.
Ogunfunmi, Tokunbo
2010-01-01
It is becoming increasingly apparent that all forms of communication-including voice-will be transmitted through packet-switched networks based on the Internet Protocol (IP). Therefore, the design of modern devices that rely on speech interfaces, such as cell phones and PDAs, requires a complete and up-to-date understanding of the basics of speech coding. Outlines key signal processing algorithms used to mitigate impairments to speech quality in VoIP networksOffering a detailed yet easily accessible introduction to the field, Principles of Speech Coding provides an in-depth examination of the
Computerized System to Aid Deaf Children in Speech Learning
National Research Council Canada - National Science Library
Riella, Rodrigo
2001-01-01
.... The aim of this analyzer is not to find the distinction between spoken words, main objective of a speech recognizer but to calculate a level of correctness in the toggle of a specific word, Voice...
Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
Directory of Open Access Journals (Sweden)
Warrick Roseboom
2011-04-01
Full Text Available Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room.
Maternal employment and early childhood overweight: findings from the UK Millennium Cohort Study.
Hawkins, S S; Cole, T J; Law, C
2008-01-01
In most developed countries, maternal employment has increased rapidly. Changing patterns of family life have been suggested to be contributing to the rising prevalence of childhood obesity. Our primary objective was to examine the relationship between maternal and partner employment and overweight in children aged 3 years. Our secondary objective was to investigate factors related to early childhood overweight only among mothers in employment. Cohort study. A total of 13 113 singleton children aged 3 years in the Millennium Cohort Study, born between 2000 and 2002 in the United Kingdom, who had complete height/weight data and parental employment histories. Parents were interviewed when the child was aged 9 months and 3 years, and the child's height and weight were measured at 3 years. Overweight (including obesity) was defined by the International Obesity Task Force cut-offs. A total of 23% (3085) of children were overweight at 3 years. Any maternal employment after the child's birth was associated with early childhood overweight (odds ratio (OR) [95% confidence interval (CI)]; 1.14 [1.00, 1.29]), after adjustment for potential confounding and mediating factors. Children were more likely to be overweight for every 10 h a mother worked per week (OR [95% CI]; 1.10 [1.04, 1.17]), after adjustment. An interaction with household income revealed that this relationship was only significant for children from households with an annual income of pound33 000 ($57 750) or higher. There was no evidence for an association between early childhood overweight and whether or for how many hours the partner worked, or with mothers' or partners' duration of employment. These relationships were also evident among mothers in employment. Independent risk factors for early childhood overweight were consistent with the published literature. Long hours of maternal employment, rather than lack of money may impede young children's access to healthy foods and physical activity. Policies
Whole-exome sequencing supports genetic heterogeneity in childhood apraxia of speech
Worthey, Elizabeth A; Raca, Gordana; Laffin, Jennifer J; Wilk, Brandon M; Harris, Jeremy M; Jakielski, Kathy J; Dimmock, David P; Strand, Edythe A; Shriberg, Lawrence D
2013-01-01
Background Childhood apraxia of speech (CAS) is a rare, severe, persistent pediatric motor speech disorder with associated deficits in sensorimotor, cognitive, language, learning and affective processes. Among other neurogenetic origins, CAS is the disorder segregating with a mutation in FOXP2 in a widely studied, multigenerational London family. We report the first whole-exome sequencing (WES) findings from a cohort of 10 unrelated participants, ages 3 to 19 years, with well-characterized CA...
Adank, Patti
2012-01-01
The role of speech production mechanisms in difficult speech comprehension is the subject of on-going debate in speech science. Two Activation Likelihood Estimation (ALE) analyses were conducted on neuroimaging studies investigating difficult speech comprehension or speech production. Meta-analysis 1 included 10 studies contrasting comprehension…
Error Consistency in Acquired Apraxia of Speech with Aphasia: Effects of the Analysis Unit
Haley, Katarina L.; Cunningham, Kevin T.; Eaton, Catherine Torrington; Jacks, Adam
2018-01-01
Purpose: Diagnostic recommendations for acquired apraxia of speech (AOS) have been contradictory concerning whether speech sound errors are consistent or variable. Studies have reported divergent findings that, on face value, could argue either for or against error consistency as a diagnostic criterion. The purpose of this study was to explain…
Metaheuristic applications to speech enhancement
Kunche, Prajna
2016-01-01
This book serves as a basic reference for those interested in the application of metaheuristics to speech enhancement. The major goal of the book is to explain the basic concepts of optimization methods and their use in heuristic optimization in speech enhancement to scientists, practicing engineers, and academic researchers in speech processing. The authors discuss why it has been a challenging problem for researchers to develop new enhancement algorithms that aid in the quality and intelligibility of degraded speech. They present powerful optimization methods to speech enhancement that can help to solve the noise reduction problems. Readers will be able to understand the fundamentals of speech processing as well as the optimization techniques, how the speech enhancement algorithms are implemented by utilizing optimization methods, and will be given the tools to develop new algorithms. The authors also provide a comprehensive literature survey regarding the topic.
The Use of Electropalatography in the Treatment of Acquired Apraxia of Speech.
Mauszycki, Shannon C; Wright, Sandra; Dingus, Nicole; Wambaugh, Julie L
2016-12-01
This investigation was designed to examine the effects of an articulatory-kinematic treatment in conjunction with visual biofeedback (VBFB) via electropalatography (EPG) on the accuracy of articulation for acquired apraxia of speech (AOS). A multiple-baseline design across participants and behaviors was used with 4 individuals with chronic AOS and aphasia. Accuracy of target speech sounds in treated and untreated phrases in probe sessions served as the dependent variable. Participants received an articulatory-kinematic treatment in combination with VBFB, which was sequentially applied to 3 stimulus sets composed of 2-word phrases with a target speech sound for each set. Positive changes in articulatory accuracy were observed for participants for the majority of treated speech sounds. Also, there was generalization to untreated phrases for most trained speech sounds. Two participants had better long-term maintenance of treated speech sounds in both trained and untrained stimuli. Findings indicate EPG may be a potential treatment tool for AOS. It appears that individuals with AOS can benefit from VBFB via EPG in improving articulatory accuracy. However, further research is needed to determine if VBFB is more advantageous than behavioral treatments that have been proven effective in improving speech production for speakers with AOS.
Davidow, Jason H.
2014-01-01
Background: Metronome-paced speech results in the elimination, or substantial reduction, of stuttering moments. The cause of fluency during this fluency-inducing condition is unknown. Several investigations have reported changes in speech pattern characteristics from a control condition to a metronome-paced speech condition, but failure to control…
TongueToSpeech (TTS): Wearable wireless assistive device for augmented speech.
Marjanovic, Nicholas; Piccinini, Giacomo; Kerr, Kevin; Esmailbeigi, Hananeh
2017-07-01
Speech is an important aspect of human communication; individuals with speech impairment are unable to communicate vocally in real time. Our team has developed the TongueToSpeech (TTS) device with the goal of augmenting speech communication for the vocally impaired. The proposed device is a wearable wireless assistive device that incorporates a capacitive touch keyboard interface embedded inside a discrete retainer. This device connects to a computer, tablet or a smartphone via Bluetooth connection. The developed TTS application converts text typed by the tongue into audible speech. Our studies have concluded that an 8-contact point configuration between the tongue and the TTS device would yield the best user precision and speed performance. On average using the TTS device inside the oral cavity takes 2.5 times longer than the pointer finger using a T9 (Text on 9 keys) keyboard configuration to type the same phrase. In conclusion, we have developed a discrete noninvasive wearable device that allows the vocally impaired individuals to communicate in real time.
Non-right handed primary progressive apraxia of speech.
Botha, Hugo; Duffy, Joseph R; Whitwell, Jennifer L; Strand, Edythe A; Machulda, Mary M; Spychalla, Anthony J; Tosakulwong, Nirubol; Senjem, Matthew L; Knopman, David S; Petersen, Ronald C; Jack, Clifford R; Lowe, Val J; Josephs, Keith A
2018-07-15
In recent years a large and growing body of research has greatly advanced our understanding of primary progressive apraxia of speech. Handedness has emerged as one potential marker of selective vulnerability in degenerative diseases. This study evaluated the clinical and imaging findings in non-right handed compared to right handed participants in a prospective cohort diagnosed with primary progressive apraxia of speech. A total of 30 participants were included. Compared to the expected rate in the population, there was a higher prevalence of non-right handedness among those with primary progressive apraxia of speech (6/30, 20%). Small group numbers meant that these results did not reach statistical significance, although the effect sizes were moderate-to-large. There were no clinical differences between right handed and non-right handed participants. Bilateral hypometabolism was seen in primary progressive apraxia of speech compared to controls, with non-right handed participants showing more right hemispheric involvement. This is the first report of a higher rate of non-right handedness in participants with isolated apraxia of speech, which may point to an increased vulnerability for developing this disorder among non-right handed participants. This challenges prior hypotheses about a relative protective effect of non-right handedness for tau-related neurodegeneration. We discuss potential avenues for future research to investigate the relationship between handedness and motor disorders more generally. Copyright © 2018 Elsevier B.V. All rights reserved.
Electrophysiological evidence for speech-specific audiovisual integration.
Baart, Martijn; Stekelenburg, Jeroen J; Vroomen, Jean
2014-01-01
Lip-read speech is integrated with heard speech at various neural levels. Here, we investigated the extent to which lip-read induced modulations of the auditory N1 and P2 (measured with EEG) are indicative of speech-specific audiovisual integration, and we explored to what extent the ERPs were modulated by phonetic audiovisual congruency. In order to disentangle speech-specific (phonetic) integration from non-speech integration, we used Sine-Wave Speech (SWS) that was perceived as speech by half of the participants (they were in speech-mode), while the other half was in non-speech mode. Results showed that the N1 obtained with audiovisual stimuli peaked earlier than the N1 evoked by auditory-only stimuli. This lip-read induced speeding up of the N1 occurred for listeners in speech and non-speech mode. In contrast, if listeners were in speech-mode, lip-read speech also modulated the auditory P2, but not if listeners were in non-speech mode, thus revealing speech-specific audiovisual binding. Comparing ERPs for phonetically congruent audiovisual stimuli with ERPs for incongruent stimuli revealed an effect of phonetic stimulus congruency that started at ~200 ms after (in)congruence became apparent. Critically, akin to the P2 suppression, congruency effects were only observed if listeners were in speech mode, and not if they were in non-speech mode. Using identical stimuli, we thus confirm that audiovisual binding involves (partially) different neural mechanisms for sound processing in speech and non-speech mode. © 2013 Published by Elsevier Ltd.
Van Ackeren, Markus Johannes; Barbero, Francesca M; Mattioni, Stefania; Bottini, Roberto
2018-01-01
The occipital cortex of early blind individuals (EB) activates during speech processing, challenging the notion of a hard-wired neurobiology of language. But, at what stage of speech processing do occipital regions participate in EB? Here we demonstrate that parieto-occipital regions in EB enhance their synchronization to acoustic fluctuations in human speech in the theta-range (corresponding to syllabic rate), irrespective of speech intelligibility. Crucially, enhanced synchronization to the intelligibility of speech was selectively observed in primary visual cortex in EB, suggesting that this region is at the interface between speech perception and comprehension. Moreover, EB showed overall enhanced functional connectivity between temporal and occipital cortices that are sensitive to speech intelligibility and altered directionality when compared to the sighted group. These findings suggest that the occipital cortex of the blind adopts an architecture that allows the tracking of speech material, and therefore does not fully abstract from the reorganized sensory inputs it receives. PMID:29338838
Pattern Recognition Methods and Features Selection for Speech Emotion Recognition System.
Partila, Pavol; Voznak, Miroslav; Tovarek, Jaromir
2015-01-01
The impact of the classification method and features selection for the speech emotion recognition accuracy is discussed in this paper. Selecting the correct parameters in combination with the classifier is an important part of reducing the complexity of system computing. This step is necessary especially for systems that will be deployed in real-time applications. The reason for the development and improvement of speech emotion recognition systems is wide usability in nowadays automatic voice controlled systems. Berlin database of emotional recordings was used in this experiment. Classification accuracy of artificial neural networks, k-nearest neighbours, and Gaussian mixture model is measured considering the selection of prosodic, spectral, and voice quality features. The purpose was to find an optimal combination of methods and group of features for stress detection in human speech. The research contribution lies in the design of the speech emotion recognition system due to its accuracy and efficiency.
Phifer, Gregg, Ed.
The 17 articles in this collection deal with theoretical and practical freedom of speech issues. The topics include: freedom of speech in Marquette Park, Illinois; Nazis in Skokie, Illinois; freedom of expression in the Confederate States of America; Robert M. LaFollette's arguments for free speech and the rights of Congress; the United States…
Medeiros, Kara F; Cress, Cynthia J
2016-06-01
Maternal directive and responsive behaviors were compared for 25 mothers and children with complex communication needs using two types of toys (familiar and unfamiliar toys). Each type of toy play was conducted with and without a single message speech-generating communication device (SGD) programmed to say "more." Rate percentages of coded intervals for maternal directive and responsive behaviors were analyzed using repeated measures ANOVAs. Results indicated that mothers used significantly more responsive behaviors when playing with their own familiar toys than with unfamiliar toys, but no differences in directiveness between types of play. Mothers showed no differences in responsivity or directiveness when the SGD was added to play interactions, indicating that the SGD did not introduce task features that detracted from the mothers' existing levels of responsivity with their children. Clinical implications are discussed.
Lin, Wan-Chien; Chang, Shin-Yow; Chen, Yi-Ting; Lee, Hsin-Chien; Chen, Yi-Hua
2017-09-01
Recently, studies have begun emphasizing paternal involvement during the perinatal period and its impact on maternal health. However, most studies have assessed maternal perception and focused on adolescents or minority groups in Western countries. Therefore, the current study investigated the association between paternal involvement and maternal postnatal depression and anxiety, along with the effects of maternal job status in the Asian society of Taiwan. This study recruited pregnant women in the first trimester of pregnancy as well as their partners on prenatal visits from July 2011 to September 2013 at four selected hospitals in metropolitan areas of Taipei, Taiwan. In total, 593 parental pairs completed the first interview and responded to the follow-up questionnaires until 6 months postpartum. Self-reported data were collected, and multiple logistic regression models were used for analyses. Lower paternal childcare and nursing frequency was independently associated with an increased risk of maternal postpartum depression (adjusted odds ratio (OR) =4.33, 95% confidence interval (CI)=1.34-13.98), particularly among unemployed mothers. Furthermore, among unemployed mothers, the risk of postnatal anxiety was 3.14 times higher in couples with fathers spending less time with the child, compared with couples with fathers spending more time (95% CI=1.10-8.98). However, no significant findings were obtained for employed mothers. The high prevalence of maternal postnatal emotional disturbances warrants continual consideration. Higher paternal involvement in childcare arrangements should be emphasized to aid in ameliorating these maternal emotional disturbances, particularly among unemployed mothers. Copyright © 2017 Elsevier B.V. All rights reserved.
Fine-grained pitch processing of music and speech in congenital amusia.
Tillmann, Barbara; Rusconi, Elena; Traube, Caroline; Butterworth, Brian; Umiltà, Carlo; Peretz, Isabelle
2011-12-01
Congenital amusia is a lifelong disorder of music processing that has been ascribed to impaired pitch perception and memory. The present study tested a large group of amusics (n=17) and provided evidence that their pitch deficit affects pitch processing in speech to a lesser extent: Fine-grained pitch discrimination was better in spoken syllables than in acoustically matched tones. Unlike amusics, control participants performed fine-grained pitch discrimination better for musical material than for verbal material. These findings suggest that pitch extraction can be influenced by the nature of the material (music vs speech), and that amusics' pitch deficit is not restricted to musical material, but extends to segmented speech events. © 2011 Acoustical Society of America
Drijvers, L.; Özyürek, A.
2017-01-01
Purpose: This study investigated whether and to what extent iconic co-speech gestures contribute to information from visible speech to enhance degraded speech comprehension at different levels of noise-vocoding. Previous studies of the contributions of these 2 visual articulators to speech
Legal decision-making by people with aphasia: critical incidents for speech pathologists.
Ferguson, Alison; Duffield, Gemma; Worrall, Linda
2010-01-01
assessment practices involved some standardized testing, but this was stressed by all participants to be of lesser importance than informal observations of function. Speech pathologists emphasized the importance of multiple observations, and multimodal means of communication. The findings indicate that speech pathologists are currently playing an active role when questions arise regarding capacity for legal and related decision-making by people with aphasia. At the same time, the findings support the need for further research to develop guidelines for practice and to build educational experiences for students and novice clinicians to assist them when they engage with the complex case management issues in this area. 2010 Royal College of Speech & Language Therapists.
Neuroanatomical correlates of childhood apraxia of speech: A connectomic approach.
Fiori, Simona; Guzzetta, Andrea; Mitra, Jhimli; Pannek, Kerstin; Pasquariello, Rosa; Cipriani, Paola; Tosetti, Michela; Cioni, Giovanni; Rose, Stephen E; Chilosi, Anna
2016-01-01
Childhood apraxia of speech (CAS) is a paediatric speech sound disorder in which precision and consistency of speech movements are impaired. Most children with idiopathic CAS have normal structural brain MRI. We hypothesize that children with CAS have altered structural connectivity in speech/language networks compared to controls and that these altered connections are related to functional speech/language measures. Whole brain probabilistic tractography, using constrained spherical deconvolution, was performed for connectome generation in 17 children with CAS and 10 age-matched controls. Fractional anisotropy (FA) was used as a measure of connectivity and the connections with altered FA between CAS and controls were identified. Further, the relationship between altered FA and speech/language scores was determined. Three intra-hemispheric/interhemispheric subnetworks showed reduction of FA in CAS compared to controls, including left inferior (opercular part) and superior (dorsolateral, medial and orbital part) frontal gyrus, left superior and middle temporal gyrus and left post-central gyrus (subnetwork 1); right supplementary motor area, left middle and inferior (orbital part) frontal gyrus, left precuneus and cuneus, right superior occipital gyrus and right cerebellum (subnetwork 2); right angular gyrus, right superior temporal gyrus and right inferior occipital gyrus (subnetwork 3). Reduced FA of some connections correlated with diadochokinesis, oromotor skills, expressive grammar and poor lexical production in CAS. These findings provide evidence of structural connectivity anomalies in children with CAS across specific brain regions involved in speech/language function. We propose altered connectivity as a possible epiphenomenon of complex pathogenic mechanisms in CAS which need further investigation.
The contribution of dynamic visual cues to audiovisual speech perception.
Jaekl, Philip; Pesquita, Ana; Alsius, Agnes; Munhall, Kevin; Soto-Faraco, Salvador
2015-08-01
Seeing a speaker's facial gestures can significantly improve speech comprehension, especially in noisy environments. However, the nature of the visual information from the speaker's facial movements that is relevant for this enhancement is still unclear. Like auditory speech signals, visual speech signals unfold over time and contain both dynamic configural information and luminance-defined local motion cues; two information sources that are thought to engage anatomically and functionally separate visual systems. Whereas, some past studies have highlighted the importance of local, luminance-defined motion cues in audiovisual speech perception, the contribution of dynamic configural information signalling changes in form over time has not yet been assessed. We therefore attempted to single out the contribution of dynamic configural information to audiovisual speech processing. To this aim, we measured word identification performance in noise using unimodal auditory stimuli, and with audiovisual stimuli. In the audiovisual condition, speaking faces were presented as point light displays achieved via motion capture of the original talker. Point light displays could be isoluminant, to minimise the contribution of effective luminance-defined local motion information, or with added luminance contrast, allowing the combined effect of dynamic configural cues and local motion cues. Audiovisual enhancement was found in both the isoluminant and contrast-based luminance conditions compared to an auditory-only condition, demonstrating, for the first time the specific contribution of dynamic configural cues to audiovisual speech improvement. These findings imply that globally processed changes in a speaker's facial shape contribute significantly towards the perception of articulatory gestures and the analysis of audiovisual speech. Copyright © 2015 Elsevier Ltd. All rights reserved.
[Precautionary maternity leave in Tirol].
Ludescher, K; Baumgartner, E; Roner, A; Brezinka, C
1998-01-01
precautionary maternity leave - little effort is made to find a low-risk workplace for a pregnant employee.
Jeon, Jin Yong; Hong, Joo Young; Jang, Hyung Suk; Kim, Jae Hyeon
2015-12-01
It is necessary to consider not only annoyance of interior noises but also speech privacy to achieve acoustic comfort in a passenger car of a high-speed train because speech from other passengers can be annoying. This study aimed to explore an optimal acoustic environment to satisfy speech privacy and reduce annoyance in a passenger car. Two experiments were conducted using speech sources and compartment noise of a high speed train with varying speech-to-noise ratios (SNRA) and background noise levels (BNL). Speech intelligibility was tested in experiment I, and in experiment II, perceived speech privacy, annoyance, and acoustic comfort of combined sounds with speech and background noise were assessed. The results show that speech privacy and annoyance were significantly influenced by the SNRA. In particular, the acoustic comfort was evaluated as acceptable when the SNRA was less than -6 dB for both speech privacy and noise annoyance. In addition, annoyance increased significantly as the BNL exceeded 63 dBA, whereas the effect of the background-noise level on the speech privacy was not significant. These findings suggest that an optimal level of interior noise in a passenger car might exist between 59 and 63 dBA, taking normal speech levels into account.
The role of training structure in perceptual learning of accented speech.
Tzeng, Christina Y; Alexander, Jessica E D; Sidaras, Sabrina K; Nygaard, Lynne C
2016-11-01
Foreign-accented speech contains multiple sources of variation that listeners learn to accommodate. Extending previous findings showing that exposure to high-variation training facilitates perceptual learning of accented speech, the current study examines to what extent the structure of training materials affects learning. During training, native adult speakers of American English transcribed sentences spoken in English by native Spanish-speaking adults. In Experiment 1, training stimuli were blocked by speaker, sentence, or randomized with respect to speaker and sentence (Variable training). At test, listeners transcribed novel English sentences produced by unfamiliar Spanish-accented speakers. Listeners' transcription accuracy was highest in the Variable condition, suggesting that varying both speaker identity and sentence across training trials enabled listeners to generalize their learning to novel speakers and linguistic content. Experiment 2 assessed the extent to which ordering of training tokens by a single factor, speaker intelligibility, would facilitate speaker-independent accent learning, finding that listeners' test performance did not reliably differ from that in the no-training control condition. Overall, these results suggest that the structure of training exposure, specifically trial-to-trial variation on both speaker's voice and linguistic content, facilitates learning of the systematic properties of accented speech. The current findings suggest a crucial role of training structure in optimizing perceptual learning. Beyond characterizing the types of variation listeners encode in their representations of spoken utterances, theories of spoken language processing should incorporate the role of training structure in learning lawful variation in speech. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Zion Golumbic, Elana M.; Poeppel, David; Schroeder, Charles E.
2012-01-01
The human capacity for processing speech is remarkable, especially given that information in speech unfolds over multiple time scales concurrently. Similarly notable is our ability to filter out of extraneous sounds and focus our attention on one conversation, epitomized by the ‘Cocktail Party’ effect. Yet, the neural mechanisms underlying on-line speech decoding and attentional stream selection are not well understood. We review findings from behavioral and neurophysiological investigations that underscore the importance of the temporal structure of speech for achieving these perceptual feats. We discuss the hypothesis that entrainment of ambient neuronal oscillations to speech’s temporal structure, across multiple time-scales, serves to facilitate its decoding and underlies the selection of an attended speech stream over other competing input. In this regard, speech decoding and attentional stream selection are examples of ‘active sensing’, emphasizing an interaction between proactive and predictive top-down modulation of neuronal dynamics and bottom-up sensory input. PMID:22285024
Stekelenburg, J.J.; Keetels, M.N.; Vroomen, J.H.M.
2018-01-01
Numerous studies have demonstrated that the vision of lip movements can alter the perception of auditory speech syllables (McGurk effect). While there is ample evidence for integration of text and auditory speech, there are only a few studies on the orthographic equivalent of the McGurk effect.
Several articles addressing topics in speech research are presented. The topics include: exploring the functional significance of physiological tremor: A biospectroscopic approach; differences between experienced and inexperienced listeners to deaf speech; a language-oriented view of reading and its disabilities; Phonetic factors in letter detection; categorical perception; Short-term recall by deaf signers of American sign language; a common basis for auditory sensory storage in perception and immediate memory; phonological awareness and verbal short-term memory; initiation versus execution time during manual and oral counting by stutterers; trading relations in the perception of speech by five-year-old children; the role of the strap muscles in pitch lowering; phonetic validation of distinctive features; consonants and syllable boundaires; and vowel information in postvocalic frictions.
Represented Speech in Qualitative Health Research
DEFF Research Database (Denmark)
Musaeus, Peter
2017-01-01
Represented speech refers to speech where we reference somebody. Represented speech is an important phenomenon in everyday conversation, health care communication, and qualitative research. This case will draw first from a case study on physicians’ workplace learning and second from a case study...... on nurses’ apprenticeship learning. The aim of the case is to guide the qualitative researcher to use own and others’ voices in the interview and to be sensitive to represented speech in everyday conversation. Moreover, reported speech matters to health professionals who aim to represent the voice...... of their patients. Qualitative researchers and students might learn to encourage interviewees to elaborate different voices or perspectives. Qualitative researchers working with natural speech might pay attention to how people talk and use represented speech. Finally, represented speech might be relevant...
Word-Form Familiarity Bootstraps Infant Speech Segmentation
Altvater-Mackensen, Nicole; Mani, Nivedita
2013-01-01
At about 7 months of age, infants listen longer to sentences containing familiar words--but not deviant pronunciations of familiar words (Jusczyk & Aslin, 1995). This finding suggests that infants are able to segment familiar words from fluent speech and that they store words in sufficient phonological detail to recognize deviations from a…
Spectral integration in speech and non-speech sounds
Jacewicz, Ewa
2005-04-01
Spectral integration (or formant averaging) was proposed in vowel perception research to account for the observation that a reduction of the intensity of one of two closely spaced formants (as in /u/) produced a predictable shift in vowel quality [Delattre et al., Word 8, 195-210 (1952)]. A related observation was reported in psychoacoustics, indicating that when the components of a two-tone periodic complex differ in amplitude and frequency, its perceived pitch is shifted toward that of the more intense tone [Helmholtz, App. XIV (1875/1948)]. Subsequent research in both fields focused on the frequency interval that separates these two spectral components, in an attempt to determine the size of the bandwidth for spectral integration to occur. This talk will review the accumulated evidence for and against spectral integration within the hypothesized limit of 3.5 Bark for static and dynamic signals in speech perception and psychoacoustics. Based on similarities in the processing of speech and non-speech sounds, it is suggested that spectral integration may reflect a general property of the auditory system. A larger frequency bandwidth, possibly close to 3.5 Bark, may be utilized in integrating acoustic information, including speech, complex signals, or sound quality of a violin.
Measurement of speech parameters in casual speech of dementia patients
Ossewaarde, Roelant; Jonkers, Roel; Jalvingh, Fedor; Bastiaanse, Yvonne
Measurement of speech parameters in casual speech of dementia patients Roelant Adriaan Ossewaarde1,2, Roel Jonkers1, Fedor Jalvingh1,3, Roelien Bastiaanse1 1CLCG, University of Groningen (NL); 2HU University of Applied Sciences Utrecht (NL); 33St. Marienhospital - Vechta, Geriatric Clinic Vechta
Directory of Open Access Journals (Sweden)
Terfa T. Alakali
2017-01-01
Full Text Available This paper examined the phenomenon of hate speech and foul language on social media platforms in Nigeria, and assessed their moral and legal consequences in the society and to journalism practice. It used both quantitative and qualitative methodology to investigate the phenomenon. In the first place, the paper employed the survey research methodology to sample 384 respondents using questionnaire and focus group discussion as instruments for data collection. Findings from the research indicate that promoting hate speech and foul language on social media have moral and legal consequences in the society and to journalism practice. Findings also show that although, the respondents understand that hate speech and foul language attract legal consequences, they do not know what obligations are created by law against perpetrators of hate speech and foul language in Nigeria. The paper therefore, adopted the qualitative, doctrinal and analytical methodology to discuss the legal consequences and obligations created against perpetrators of hate speech and foul language in Nigeria. The paper concluded based on the findings that hate speech and foul language is prevalent on social media platforms in Nigeria and that there are adequate legal provisions to curb the phenomenon in Nigeria. It recommends among others things that the Nigerian government and NGOs should sponsor monitoring projects like the UMATI in Kenya to better understand the use of hate speech and that monitoring agencies set up under the legal regime should adopt mechanisms to identify and remove hate speech content on social media platforms in Nigeria.
Shriberg, Lawrence D.; Strand, Edythe A.; Fourakis, Marios; Jakielski, Kathy J.; Hall, Sheryl D.; Karlsson, Heather B.; Mabie, Heather L.; McSweeny, Jane L.; Tilkens, Christie M.; Wilson, David L.
2017-01-01
Purpose: The purpose of this 2nd article in this supplement is to report validity support findings for the Pause Marker (PM), a proposed single-sign diagnostic marker of childhood apraxia of speech (CAS). Method: PM scores and additional perceptual and acoustic measures were obtained from 296 participants in cohorts with idiopathic and…
Maternal Household Decision-Making Autonomy and Adolescent Education in Honduras.
Hendrick, C Emily; Marteleto, Leticia
2017-06-01
Maternal decision-making autonomy has been linked to positive outcomes for children's health and well-being early in life in low- and middle-income countries throughout the world. However, there is a dearth of research examining if and how maternal autonomy continues to influence children's outcomes into adolescence and whether it impacts other domains of children's lives beyond health, such as their education. The goal of this study was to determine whether high maternal decision-making was associated with school enrollment for secondary school-aged youth in Honduras. Further, we aimed to assess whether the relationships between maternal autonomy and school enrollment varied by adolescents' environmental contexts and individual characteristics such as gender. Our analytical sample included 6,579 adolescents ages 12-16 living with their mothers from the Honduran Demographic and Health Survey (DHS) 2011-12. We used stepwise logistic regression models to investigate the association between maternal household decision-making autonomy and adolescents' school enrollment. Our findings suggest that adolescents, especially girls, benefit from their mothers' high decision-making autonomy. Findings suggest that maternal decision-making autonomy promotes adolescents' school enrollment above and beyond other maternal, household, and regional influences.
Music and speech distractors disrupt sensorimotor synchronization: effects of musical training.
Białuńska, Anita; Dalla Bella, Simone
2017-12-01
Humans display a natural tendency to move to the beat of music, more than to the rhythm of any other auditory stimulus. We typically move with music, but rarely with speech. This proclivity is apparent early during development and can be further developed over the years via joint dancing, singing, or instrument playing. Synchronization of movement to the beat can thus improve with age, but also with musical experience. In a previous study, we found that music perturbed synchronization with a metronome more than speech fragments; music superiority disappeared when distractors shared isochrony and the same meter (Dalla Bella et al., PLoS One 8(8):e71945, 2013). Here, we examined if the interfering effect of music and speech distractors in a synchronization task is influenced by musical training. Musicians and non-musicians synchronized by producing finger force pulses to the sounds of a metronome while music and speech distractors were presented at one of various phase relationships with respect to the target. Distractors were familiar musical excerpts and fragments of children poetry comparable in terms of beat/stress isochrony. Music perturbed synchronization with the metronome more than speech did in both groups. However, the difference in synchronization error between music and speech distractors was smaller for musicians than for non-musicians, especially when the peak force of movement is reached. These findings point to a link between musical training and timing of sensorimotor synchronization when reacting to music and speech distractors.
Development of The Viking Speech Scale to classify the speech of children with cerebral palsy.
Pennington, Lindsay; Virella, Daniel; Mjøen, Tone; da Graça Andrada, Maria; Murray, Janice; Colver, Allan; Himmelmann, Kate; Rackauskaite, Gija; Greitane, Andra; Prasauskiene, Audrone; Andersen, Guro; de la Cruz, Javier
2013-10-01
Surveillance registers monitor the prevalence of cerebral palsy and the severity of resulting impairments across time and place. The motor disorders of cerebral palsy can affect children's speech production and limit their intelligibility. We describe the development of a scale to classify children's speech performance for use in cerebral palsy surveillance registers, and its reliability across raters and across time. Speech and language therapists, other healthcare professionals and parents classified the speech of 139 children with cerebral palsy (85 boys, 54 girls; mean age 6.03 years, SD 1.09) from observation and previous knowledge of the children. Another group of health professionals rated children's speech from information in their medical notes. With the exception of parents, raters reclassified children's speech at least four weeks after their initial classification. Raters were asked to rate how easy the scale was to use and how well the scale described the child's speech production using Likert scales. Inter-rater reliability was moderate to substantial (k>.58 for all comparisons). Test-retest reliability was substantial to almost perfect for all groups (k>.68). Over 74% of raters found the scale easy or very easy to use; 66% of parents and over 70% of health care professionals judged the scale to describe children's speech well or very well. We conclude that the Viking Speech Scale is a reliable tool to describe the speech performance of children with cerebral palsy, which can be applied through direct observation of children or through case note review. Copyright © 2013 Elsevier Ltd. All rights reserved.
Johari, Karim; Behroozmand, Roozbeh
2017-08-01
Skilled movement is mediated by motor commands executed with extremely fine temporal precision. The question of how the brain incorporates temporal information to perform motor actions has remained unanswered. This study investigated the effect of stimulus temporal predictability on response timing of speech and hand movement. Subjects performed a randomized vowel vocalization or button press task in two counterbalanced blocks in response to temporally-predictable and unpredictable visual cues. Results indicated that speech and hand reaction time was decreased for predictable compared with unpredictable stimuli. This finding suggests that a temporal predictive code is established to capture temporal dynamics of sensory cues in order to produce faster movements in responses to predictable stimuli. In addition, results revealed a main effect of modality, indicating faster hand movement compared with speech. We suggest that this effect is accounted for by the inherent complexity of speech production compared with hand movement. Lastly, we found that movement inhibition was faster than initiation for both hand and speech, suggesting that movement initiation requires a longer processing time to coordinate activities across multiple regions in the brain. These findings provide new insights into the mechanisms of temporal information processing during initiation and inhibition of speech and hand movement. Copyright © 2017 Elsevier B.V. All rights reserved.
The impact of language co-activation on L1 and L2 speech fluency.
Bergmann, Christopher; Sprenger, Simone A; Schmid, Monika S
2015-10-01
Fluent speech depends on the availability of well-established linguistic knowledge and routines for speech planning and articulation. A lack of speech fluency in late second-language (L2) learners may point to a deficiency of these representations, due to incomplete acquisition. Experiments on bilingual language processing have shown, however, that there are strong reasons to believe that multilingual speakers experience co-activation of the languages they speak. We have studied to what degree language co-activation affects fluency in the speech of bilinguals, comparing a monolingual German control group with two bilingual groups: 1) first-language (L1) attriters, who have fully acquired German before emigrating to an L2 English environment, and 2) immersed L2 learners of German (L1: English). We have analysed the temporal fluency and the incidence of disfluency markers (pauses, repetitions and self-corrections) in spontaneous film retellings. Our findings show that learners to speak more slowly than controls and attriters. Also, on each count, the speech of at least one of the bilingual groups contains more disfluency markers than the retellings of the control group. Generally speaking, both bilingual groups-learners and attriters-are equally (dis)fluent and significantly more disfluent than the monolingual speakers. Given that the L1 attriters are unaffected by incomplete acquisition, we interpret these findings as evidence for language competition during speech production. Copyright © 2015. Published by Elsevier B.V.
Drijvers, Linda; Ozyurek, Asli
2017-01-01
Purpose: This study investigated whether and to what extent iconic co-speech gestures contribute to information from visible speech to enhance degraded speech comprehension at different levels of noise-vocoding. Previous studies of the contributions of these 2 visual articulators to speech comprehension have only been performed separately. Method:…
Production planning and coronal stop deletion in spontaneous speech
Directory of Open Access Journals (Sweden)
James Tanner
2017-06-01
Full Text Available Many phonological processes can be affected by segmental context spanning word boundaries, which often lead to variable outcomes. This paper tests the idea that some of this variability can be explained by reference to production planning. We examine coronal stop deletion (CSD, a variable process conditioned by preceding and upcoming phonological context, in a corpus of spontaneous British English speech, as a means of investigating a number of variables associated with planning: Prosodic boundary strength, word frequency, conditional probability of the following word, and speech rate. From the perspective of production planning, (1 prosodic boundaries should affect deletion rate independently of following context; (2 given the locality of production planning, the effect of the following context should decrease at stronger prosodic boundaries; and (3 other factors affecting planning scope should modulate the effect of upcoming phonological material above and beyond the modulating effect of prosodic boundaries. We build a statistical model of CSD realization, using pause length as a quantitative proxy for boundary strength, and find support for these predictions. These findings are compatible with the hypothesis that the locality of production planning constrains variability in speech production, and have practical implications for work on CSD and other variable processes.
The development of visual speech perception in Mandarin Chinese-speaking children.
Chen, Liang; Lei, Jianghua
2017-01-01
The present study aimed to investigate the development of visual speech perception in Chinese-speaking children. Children aged 7, 13 and 16 were asked to visually identify both consonant and vowel sounds in Chinese as quickly and accurately as possible. Results revealed (1) an increase in accuracy of visual speech perception between ages 7 and 13 after which the accuracy rate either stagnates or drops; and (2) a U-shaped development pattern in speed of perception with peak performance in 13-year olds. Results also showed that across all age groups, the overall levels of accuracy rose, whereas the response times fell for simplex finals, complex finals and initials. These findings suggest that (1) visual speech perception in Chinese is a developmental process that is acquired over time and is still fine-tuned well into late adolescence; (2) factors other than cross-linguistic differences in phonological complexity and degrees of reliance on visual information are involved in development of visual speech perception.
Pattern Recognition Methods and Features Selection for Speech Emotion Recognition System
Directory of Open Access Journals (Sweden)
Pavol Partila
2015-01-01
Full Text Available The impact of the classification method and features selection for the speech emotion recognition accuracy is discussed in this paper. Selecting the correct parameters in combination with the classifier is an important part of reducing the complexity of system computing. This step is necessary especially for systems that will be deployed in real-time applications. The reason for the development and improvement of speech emotion recognition systems is wide usability in nowadays automatic voice controlled systems. Berlin database of emotional recordings was used in this experiment. Classification accuracy of artificial neural networks, k-nearest neighbours, and Gaussian mixture model is measured considering the selection of prosodic, spectral, and voice quality features. The purpose was to find an optimal combination of methods and group of features for stress detection in human speech. The research contribution lies in the design of the speech emotion recognition system due to its accuracy and efficiency.
The Effect of Furlow Palatoplasty Timing on Speech Outcomes in Submucous Cleft Palate.
Swanson, Jordan W; Mitchell, Brianne T; Cohen, Marilyn; Solot, Cynthia; Jackson, Oksana; Low, David; Bartlett, Scott P; Taylor, Jesse A
2017-08-01
Because some patients with submucous cleft palate (SMCP) are asymptomatic, surgical treatment is conventionally delayed until hypernasal resonance is identified during speech production. We aim to identify whether speech outcomes after repair of a SMCP is influenced by age of repair. We retrospectively studied nonsyndromic children with SMCP. Speech results, before and after any surgical treatment or physical management of the palate were compared using the Pittsburgh Weighted Speech Scoring system. Furlow palatoplasty was performed on 40 nonsyndromic patients with SMCP, and 26 patients were not surgically treated. Total composite speech scores improved significantly among children repaired between 3 and 4 years of age (P = 0.02), but not older than 4 years (P = 0.63). Twelve (86%) of 14 patients repaired who are older than 4 years had borderline or incompetent speech (composite Pittsburgh Weighted Speech Scoring ≥3) compared with 2 (29%) of 7 repaired between 3 and 4 years of age (P = 0.0068), despite worse prerepair scores in the latter group. Resonance improved in children repaired who are older than 4 years, but articulation errors persisted to a greater degree than those treated before 4 years of age (P = 0.01.) CONCLUSIONS: Submucous cleft palate repair before 4 years of age appears associated with lower ultimate rates of borderline or incompetent speech. Speech of patients repaired at or after 4 years of age seems to be characterized by persistent misarticulation. These findings highlight the importance of timely diagnosis and management.
Terband, H; Maassen, B; Guenther, F H; Brumberg, J
2014-01-01
Differentiating the symptom complex due to phonological-level disorders, speech delay and pediatric motor speech disorders is a controversial issue in the field of pediatric speech and language pathology. The present study investigated the developmental interaction between neurological deficits in auditory and motor processes using computational modeling with the DIVA model. In a series of computer simulations, we investigated the effect of a motor processing deficit alone (MPD), and the effect of a motor processing deficit in combination with an auditory processing deficit (MPD+APD) on the trajectory and endpoint of speech motor development in the DIVA model. Simulation results showed that a motor programming deficit predominantly leads to deterioration on the phonological level (phonemic mappings) when auditory self-monitoring is intact, and on the systemic level (systemic mapping) if auditory self-monitoring is impaired. These findings suggest a close relation between quality of auditory self-monitoring and the involvement of phonological vs. motor processes in children with pediatric motor speech disorders. It is suggested that MPD+APD might be involved in typically apraxic speech output disorders and MPD in pediatric motor speech disorders that also have a phonological component. Possibilities to verify these hypotheses using empirical data collected from human subjects are discussed. The reader will be able to: (1) identify the difficulties in studying disordered speech motor development; (2) describe the differences in speech motor characteristics between SSD and subtype CAS; (3) describe the different types of learning that occur in the sensory-motor system during babbling and early speech acquisition; (4) identify the neural control subsystems involved in speech production; (5) describe the potential role of auditory self-monitoring in developmental speech disorders. Copyright © 2014 Elsevier Inc. All rights reserved.
Current trends in small vocabulary speech recognition for equipment control
Doukas, Nikolaos; Bardis, Nikolaos G.
2017-09-01
Speech recognition systems allow human - machine communication to acquire an intuitive nature that approaches the simplicity of inter - human communication. Small vocabulary speech recognition is a subset of the overall speech recognition problem, where only a small number of words need to be recognized. Speaker independent small vocabulary recognition can find significant applications in field equipment used by military personnel. Such equipment may typically be controlled by a small number of commands that need to be given quickly and accurately, under conditions where delicate manual operations are difficult to achieve. This type of application could hence significantly benefit by the use of robust voice operated control components, as they would facilitate the interaction with their users and render it much more reliable in times of crisis. This paper presents current challenges involved in attaining efficient and robust small vocabulary speech recognition. These challenges concern feature selection, classification techniques, speaker diversity and noise effects. A state machine approach is presented that facilitates the voice guidance of different equipment in a variety of situations.
Speech enhancement using emotion dependent codebooks
Naidu, D.H.R.; Srinivasan, S.
2012-01-01
Several speech enhancement approaches utilize trained models of clean speech data, such as codebooks, Gaussian mixtures, and hidden Markov models. These models are typically trained on neutral clean speech data, without any emotion. However, in practical scenarios, emotional speech is a common
Early Maternal Time Investment and Early Child Outcomes
Del Bono, Emilia; Francesconi, Marco; Kelly, Yvonne; Sacker, Amanda
2014-01-01
Using large longitudinal survey data from the UK Millennium Cohort Study, this paper estimates the relationship between maternal time inputs and early child development. We find that maternal time is a quantitatively important determinant of skill formation and that its effect declines with child age. There is evidence of long-term effects of early maternal time inputs on later outcomes, especially in the case of cognitive skill development. In the case of non-cognitive development, the evide...
The Length of Maternity Leave and Family Health
DEFF Research Database (Denmark)
Beuchert-Pedersen, Louise Voldby; Humlum, Maria Knoth; Vejlin, Rune Majlund
We study the relationship between the length of maternity leave and the physical and psychological health of the family. Using a reform of the parental leave scheme in Denmark that increased the number of weeks of leave with full benefit compensation, we estimate the effect of the lenght...... of maternity leave on a range of health indicators including the number of hospital admissions for both mother and child and the probability of the mother receiving antidepressants. The reform led to an increase in average post-birth maternity leave matters for child or maternal health outcomes and thus we...... complement the existing evidence on maternity leave expansions that tends to find limited effects on children's later deveopmental, educational, and labor market outcomes. Our results suggest that any beneficial effects of increasing the lenght of maternity leave are greater for low-resource families....
The phonological memory profile of preschool children who make atypical speech sound errors.
Waring, Rebecca; Eadie, Patricia; Rickard Liow, Susan; Dodd, Barbara
2018-01-01
Previous research indicates that children with speech sound disorders (SSD) have underlying phonological memory deficits. The SSD population, however, is diverse. While children who make consistent atypical speech errors (phonological disorder/PhDis) are known to have executive function deficits in rule abstraction and cognitive flexibility, little is known about their memory profile. Sixteen monolingual preschool children with atypical speech errors (PhDis) were matched individually to age-and-gender peers with typically developing speech (TDS). The two groups were compared on forward recall of familiar words (pointing response), reverse recall of familiar words (pointing response), and reverse recall of digits (spoken response) and a receptive vocabulary task. There were no differences between children with TDS and children with PhDis on forward recall or vocabulary tasks. However, children with TDS significantly outperformed children with PhDis on the two reverse recall tasks. Findings suggest that atypical speech errors are associated with impaired phonological working memory, implicating executive function impairment in specific subtypes of SSD.
Integrating speech in time depends on temporal expectancies and attention.
Scharinger, Mathias; Steinberg, Johanna; Tavano, Alessandro
2017-08-01
Sensory information that unfolds in time, such as in speech perception, relies on efficient chunking mechanisms in order to yield optimally-sized units for further processing. Whether or not two successive acoustic events receive a one-unit or a two-unit interpretation seems to depend on the fit between their temporal extent and a stipulated temporal window of integration. However, there is ongoing debate on how flexible this temporal window of integration should be, especially for the processing of speech sounds. Furthermore, there is no direct evidence of whether attention may modulate the temporal constraints on the integration window. For this reason, we here examine how different word durations, which lead to different temporal separations of sound onsets, interact with attention. In an Electroencephalography (EEG) study, participants actively and passively listened to words where word-final consonants were occasionally omitted. Words had either a natural duration or were artificially prolonged in order to increase the separation of speech sound onsets. Omission responses to incomplete speech input, originating in left temporal cortex, decreased when the critical speech sound was separated from previous sounds by more than 250 msec, i.e., when the separation was larger than the stipulated temporal window of integration (125-150 msec). Attention, on the other hand, only increased omission responses for stimuli with natural durations. We complemented the event-related potential (ERP) analyses by a frequency-domain analysis on the stimulus presentation rate. Notably, the power of stimulation frequency showed the same duration and attention effects than the omission responses. We interpret these findings on the background of existing research on temporal integration windows and further suggest that our findings may be accounted for within the framework of predictive coding. Copyright © 2017 Elsevier Ltd. All rights reserved.
STRUCTURAL AND STRATEGIC ASPECTS OF PROFESSIONALLY ORIENTED SPEECH OF A PSYCHOLOGIST MEDIATOR
Directory of Open Access Journals (Sweden)
Iryna Levchyk
2016-12-01
Full Text Available The article presents characteristic speech patterns of psychologist-mediator on the basis of five staged model of his professional speech behavior that involves the following five speech activities: introductory talks with the conflict parties; clarifying of the parties’ positions; finding the optimal solution to the problem; persuasion in the legality of a compromise; execution of the agreement between the parties. Each of these stages of the mediation process in terms of mental and speech activities of a specialist have been analyzed and subsequently the structure of mediator’s communication has been derived. The concept of a "strategy of verbal behavior" considering professional activity of a psychologist-mediator has been described in terms of its correlation with the type of negotiation behaviors of disputants. The basic types of opponents’ behavior in negotiations ‒ namely avoidance, concession, denial, aggression have been specified. The compliance of strategy of speech of mediator’s behavior to his chosen style of mediation has been discovered. The tactics and logic of mediator’s speech behavior according to the stages of mediation conversation have been determined. It has been found out that the mediator’s tactics implies application of specific professional speech skills to conduct a dialogue in accordance with the chosen strategy as well as emotional and verbal reaction of conflict sides in the process of communication.
Brouwer, Susanne; Van Engen, Kristin J.; Calandruccio, Lauren; Bradlow, Ann R.
2012-01-01
This study examined whether speech-on-speech masking is sensitive to variation in the degree of similarity between the target and the masker speech. Three experiments investigated whether speech-in-speech recognition varies across different background speech languages (English vs Dutch) for both English and Dutch targets, as well as across variation in the semantic content of the background speech (meaningful vs semantically anomalous sentences), and across variation in listener status vis-à-vis the target and masker languages (native, non-native, or unfamiliar). The results showed that the more similar the target speech is to the masker speech (e.g., same vs different language, same vs different levels of semantic content), the greater the interference on speech recognition accuracy. Moreover, the listener’s knowledge of the target and the background language modulate the size of the release from masking. These factors had an especially strong effect on masking effectiveness in highly unfavorable listening conditions. Overall this research provided evidence that that the degree of target-masker similarity plays a significant role in speech-in-speech recognition. The results also give insight into how listeners assign their resources differently depending on whether they are listening to their first or second language. PMID:22352516
Speech-specificity of two audiovisual integration effects
DEFF Research Database (Denmark)
Eskelund, Kasper; Tuomainen, Jyrki; Andersen, Tobias
2010-01-01
Seeing the talker’s articulatory mouth movements can influence the auditory speech percept both in speech identification and detection tasks. Here we show that these audiovisual integration effects also occur for sine wave speech (SWS), which is an impoverished speech signal that naïve observers...... often fail to perceive as speech. While audiovisual integration in the identification task only occurred when observers were informed of the speech-like nature of SWS, integration occurred in the detection task both for informed and naïve observers. This shows that both speech-specific and general...... mechanisms underlie audiovisual integration of speech....
Recognizing speech in a novel accent: the motor theory of speech perception reframed.
Moulin-Frier, Clément; Arbib, Michael A
2013-08-01
The motor theory of speech perception holds that we perceive the speech of another in terms of a motor representation of that speech. However, when we have learned to recognize a foreign accent, it seems plausible that recognition of a word rarely involves reconstruction of the speech gestures of the speaker rather than the listener. To better assess the motor theory and this observation, we proceed in three stages. Part 1 places the motor theory of speech perception in a larger framework based on our earlier models of the adaptive formation of mirror neurons for grasping, and for viewing extensions of that mirror system as part of a larger system for neuro-linguistic processing, augmented by the present consideration of recognizing speech in a novel accent. Part 2 then offers a novel computational model of how a listener comes to understand the speech of someone speaking the listener's native language with a foreign accent. The core tenet of the model is that the listener uses hypotheses about the word the speaker is currently uttering to update probabilities linking the sound produced by the speaker to phonemes in the native language repertoire of the listener. This, on average, improves the recognition of later words. This model is neutral regarding the nature of the representations it uses (motor vs. auditory). It serve as a reference point for the discussion in Part 3, which proposes a dual-stream neuro-linguistic architecture to revisits claims for and against the motor theory of speech perception and the relevance of mirror neurons, and extracts some implications for the reframing of the motor theory.
Advocate: A Distributed Architecture for Speech-to-Speech Translation
2009-01-01
tecture, are either wrapped natural-language processing ( NLP ) components or objects developed from scratch using the architecture’s API. GATE is...framework, we put together a demonstration Arabic -to- English speech translation system using both internally developed ( Arabic speech recognition and MT...conditions of our Arabic S2S demonstration system described earlier. Once again, the data size was varied and eighty identical requests were
Improving the speech intelligibility in classrooms
Lam, Choi Ling Coriolanus
One of the major acoustical concerns in classrooms is the establishment of effective verbal communication between teachers and students. Non-optimal acoustical conditions, resulting in reduced verbal communication, can cause two main problems. First, they can lead to reduce learning efficiency. Second, they can also cause fatigue, stress, vocal strain and health problems, such as headaches and sore throats, among teachers who are forced to compensate for poor acoustical conditions by raising their voices. Besides, inadequate acoustical conditions can induce the usage of public address system. Improper usage of such amplifiers or loudspeakers can lead to impairment of students' hearing systems. The social costs of poor classroom acoustics will be large to impair the learning of children. This invisible problem has far reaching implications for learning, but is easily solved. Many researches have been carried out that they have accurately and concisely summarized the research findings on classrooms acoustics. Though, there is still a number of challenging questions remaining unanswered. Most objective indices for speech intelligibility are essentially based on studies of western languages. Even several studies of tonal languages as Mandarin have been conducted, there is much less on Cantonese. In this research, measurements have been done in unoccupied rooms to investigate the acoustical parameters and characteristics of the classrooms. The speech intelligibility tests, which based on English, Mandarin and Cantonese, and the survey were carried out on students aged from 5 years old to 22 years old. It aims to investigate the differences in intelligibility between English, Mandarin and Cantonese of the classrooms in Hong Kong. The significance on speech transmission index (STI) related to Phonetically Balanced (PB) word scores will further be developed. Together with developed empirical relationship between the speech intelligibility in classrooms with the variations
Roman, Adrienne S; Pisoni, David B; Kronenberger, William G; Faulkner, Kathleen F
Noise-vocoded speech is a valuable research tool for testing experimental hypotheses about the effects of spectral degradation on speech recognition in adults with normal hearing (NH). However, very little research has utilized noise-vocoded speech with children with NH. Earlier studies with children with NH focused primarily on the amount of spectral information needed for speech recognition without assessing the contribution of neurocognitive processes to speech perception and spoken word recognition. In this study, we first replicated the seminal findings reported by ) who investigated effects of lexical density and word frequency on noise-vocoded speech perception in a small group of children with NH. We then extended the research to investigate relations between noise-vocoded speech recognition abilities and five neurocognitive measures: auditory attention (AA) and response set, talker discrimination, and verbal and nonverbal short-term working memory. Thirty-one children with NH between 5 and 13 years of age were assessed on their ability to perceive lexically controlled words in isolation and in sentences that were noise-vocoded to four spectral channels. Children were also administered vocabulary assessments (Peabody Picture Vocabulary test-4th Edition and Expressive Vocabulary test-2nd Edition) and measures of AA (NEPSY AA and response set and a talker discrimination task) and short-term memory (visual digit and symbol spans). Consistent with the findings reported in the original ) study, we found that children perceived noise-vocoded lexically easy words better than lexically hard words. Words in sentences were also recognized better than the same words presented in isolation. No significant correlations were observed between noise-vocoded speech recognition scores and the Peabody Picture Vocabulary test-4th Edition using language quotients to control for age effects. However, children who scored higher on the Expressive Vocabulary test-2nd Edition
Speech graphs provide a quantitative measure of thought disorder in psychosis.
Directory of Open Access Journals (Sweden)
Natalia B Mota
Full Text Available BACKGROUND: Psychosis has various causes, including mania and schizophrenia. Since the differential diagnosis of psychosis is exclusively based on subjective assessments of oral interviews with patients, an objective quantification of the speech disturbances that characterize mania and schizophrenia is in order. In principle, such quantification could be achieved by the analysis of speech graphs. A graph represents a network with nodes connected by edges; in speech graphs, nodes correspond to words and edges correspond to semantic and grammatical relationships. METHODOLOGY/PRINCIPAL FINDINGS: To quantify speech differences related to psychosis, interviews with schizophrenics, manics and normal subjects were recorded and represented as graphs. Manics scored significantly higher than schizophrenics in ten graph measures. Psychopathological symptoms such as logorrhea, poor speech, and flight of thoughts were grasped by the analysis even when verbosity differences were discounted. Binary classifiers based on speech graph measures sorted schizophrenics from manics with up to 93.8% of sensitivity and 93.7% of specificity. In contrast, sorting based on the scores of two standard psychiatric scales (BPRS and PANSS reached only 62.5% of sensitivity and specificity. CONCLUSIONS/SIGNIFICANCE: The results demonstrate that alterations of the thought process manifested in the speech of psychotic patients can be objectively measured using graph-theoretical tools, developed to capture specific features of the normal and dysfunctional flow of thought, such as divergence and recurrence. The quantitative analysis of speech graphs is not redundant with standard psychometric scales but rather complementary, as it yields a very accurate sorting of schizophrenics and manics. Overall, the results point to automated psychiatric diagnosis based not on what is said, but on how it is said.
Using the Speech Transmission Index for predicting non-native speech intelligibility
Wijngaarden, S.J. van; Bronkhorst, A.W.; Houtgast, T.; Steeneken, H.J.M.
2004-01-01
While the Speech Transmission Index ~STI! is widely applied for prediction of speech intelligibility in room acoustics and telecommunication engineering, it is unclear how to interpret STI values when non-native talkers or listeners are involved. Based on subjectively measured psychometric functions
Maas, Edwin; Mailend, Marja-Liisa
2012-01-01
Purpose: The purpose of this article is to present an argument for the use of online reaction time (RT) methods to the study of apraxia of speech (AOS) and to review the existing small literature in this area and the contributions it has made to our fundamental understanding of speech planning (deficits) in AOS. Method: Following a brief…
Role of working memory and lexical knowledge in perceptual restoration of interrupted speech.
Nagaraj, Naveen K; Magimairaj, Beula M
2017-12-01
The role of working memory (WM) capacity and lexical knowledge in perceptual restoration (PR) of missing speech was investigated using the interrupted speech perception paradigm. Speech identification ability, which indexed PR, was measured using low-context sentences periodically interrupted at 1.5 Hz. PR was measured for silent gated, low-frequency speech noise filled, and low-frequency fine-structure and envelope filled interrupted conditions. WM capacity was measured using verbal and visuospatial span tasks. Lexical knowledge was assessed using both receptive vocabulary and meaning from context tests. Results showed that PR was better for speech noise filled condition than other conditions tested. Both receptive vocabulary and verbal WM capacity explained unique variance in PR for the speech noise filled condition, but were unrelated to performance in the silent gated condition. It was only receptive vocabulary that uniquely predicted PR for fine-structure and envelope filled conditions. These findings suggest that the contribution of lexical knowledge and verbal WM during PR depends crucially on the information content that replaced the silent intervals. When perceptual continuity was partially restored by filler speech noise, both lexical knowledge and verbal WM capacity facilitated PR. Importantly, for fine-structure and envelope filled interrupted conditions, lexical knowledge was crucial for PR.
Terband, H.; Maassen, B.; Guenther, F.H.; Brumberg, J.
2014-01-01
Background/Purpose Differentiating the symptom complex due to phonological-level disorders, speech delay and pediatric motor speech disorders is a controversial issue in the field of pediatric speech and language pathology. The present study investigated the developmental interaction between neurological deficits in auditory and motor processes using computational modeling with the DIVA model. Method In a series of computer simulations, we investigated the effect of a motor processing deficit alone (MPD), and the effect of a motor processing deficit in combination with an auditory processing deficit (MPD+APD) on the trajectory and endpoint of speech motor development in the DIVA model. Results Simulation results showed that a motor programming deficit predominantly leads to deterioration on the phonological level (phonemic mappings) when auditory self-monitoring is intact, and on the systemic level (systemic mapping) if auditory self-monitoring is impaired. Conclusions These findings suggest a close relation between quality of auditory self-monitoring and the involvement of phonological vs. motor processes in children with pediatric motor speech disorders. It is suggested that MPD+APD might be involved in typically apraxic speech output disorders and MPD in pediatric motor speech disorders that also have a phonological component. Possibilities to verify these hypotheses using empirical data collected from human subjects are discussed. PMID:24491630
Individual differences in language and working memory affect children's speech recognition in noise.
McCreery, Ryan W; Spratford, Meredith; Kirby, Benjamin; Brennan, Marc
2017-05-01
We examined how cognitive and linguistic skills affect speech recognition in noise for children with normal hearing. Children with better working memory and language abilities were expected to have better speech recognition in noise than peers with poorer skills in these domains. As part of a prospective, cross-sectional study, children with normal hearing completed speech recognition in noise for three types of stimuli: (1) monosyllabic words, (2) syntactically correct but semantically anomalous sentences and (3) semantically and syntactically anomalous word sequences. Measures of vocabulary, syntax and working memory were used to predict individual differences in speech recognition in noise. Ninety-six children with normal hearing, who were between 5 and 12 years of age. Higher working memory was associated with better speech recognition in noise for all three stimulus types. Higher vocabulary abilities were associated with better recognition in noise for sentences and word sequences, but not for words. Working memory and language both influence children's speech recognition in noise, but the relationships vary across types of stimuli. These findings suggest that clinical assessment of speech recognition is likely to reflect underlying cognitive and linguistic abilities, in addition to a child's auditory skills, consistent with the Ease of Language Understanding model.
DEFF Research Database (Denmark)
Jørgensen, Søren; Dau, Torsten
2011-01-01
conditions by comparing predictions to measured data from [Kjems et al. (2009). J. Acoust. Soc. Am. 126 (3), 1415-1426] where speech is mixed with four different interferers, including speech-shaped noise, bottle noise, car noise, and cafe noise. The model accounts well for the differences in intelligibility......The speech-based envelope power spectrum model (sEPSM) [Jørgensen and Dau (2011). J. Acoust. Soc. Am., 130 (3), 1475–1487] estimates the envelope signal-to-noise ratio (SNRenv) of distorted speech and accurately describes the speech recognition thresholds (SRT) for normal-hearing listeners...... observed for the different interferers. None of the standardized models successfully describe these data....
Chen, Zhaocong; Wong, Francis C K; Jones, Jeffery A; Li, Weifeng; Liu, Peng; Chen, Xi; Liu, Hanjun
2015-08-17
Speech perception and production are intimately linked. There is evidence that speech motor learning results in changes to auditory processing of speech. Whether speech motor control benefits from perceptual learning in speech, however, remains unclear. This event-related potential study investigated whether speech-sound learning can modulate the processing of feedback errors during vocal pitch regulation. Mandarin speakers were trained to perceive five Thai lexical tones while learning to associate pictures with spoken words over 5 days. Before and after training, participants produced sustained vowel sounds while they heard their vocal pitch feedback unexpectedly perturbed. As compared to the pre-training session, the magnitude of vocal compensation significantly decreased for the control group, but remained consistent for the trained group at the post-training session. However, the trained group had smaller and faster N1 responses to pitch perturbations and exhibited enhanced P2 responses that correlated significantly with their learning performance. These findings indicate that the cortical processing of vocal pitch regulation can be shaped by learning new speech-sound associations, suggesting that perceptual learning in speech can produce transfer effects to facilitating the neural mechanisms underlying the online monitoring of auditory feedback regarding vocal production.
Energy Technology Data Exchange (ETDEWEB)
Hogden, J.
1996-11-05
The goal of the proposed research is to test a statistical model of speech recognition that incorporates the knowledge that speech is produced by relatively slow motions of the tongue, lips, and other speech articulators. This model is called Maximum Likelihood Continuity Mapping (Malcom). Many speech researchers believe that by using constraints imposed by articulator motions, we can improve or replace the current hidden Markov model based speech recognition algorithms. Unfortunately, previous efforts to incorporate information about articulation into speech recognition algorithms have suffered because (1) slight inaccuracies in our knowledge or the formulation of our knowledge about articulation may decrease recognition performance, (2) small changes in the assumptions underlying models of speech production can lead to large changes in the speech derived from the models, and (3) collecting measurements of human articulator positions in sufficient quantity for training a speech recognition algorithm is still impractical. The most interesting (and in fact, unique) quality of Malcom is that, even though Malcom makes use of a mapping between acoustics and articulation, Malcom can be trained to recognize speech using only acoustic data. By learning the mapping between acoustics and articulation using only acoustic data, Malcom avoids the difficulties involved in collecting articulator position measurements and does not require an articulatory synthesizer model to estimate the mapping between vocal tract shapes and speech acoustics. Preliminary experiments that demonstrate that Malcom can learn the mapping between acoustics and articulation are discussed. Potential applications of Malcom aside from speech recognition are also discussed. Finally, specific deliverables resulting from the proposed research are described.
Cleft Audit Protocol for Speech (CAPS-A): A Comprehensive Training Package for Speech Analysis
Sell, D.; John, A.; Harding-Bell, A.; Sweeney, T.; Hegarty, F.; Freeman, J.
2009-01-01
Background: The previous literature has largely focused on speech analysis systems and ignored process issues, such as the nature of adequate speech samples, data acquisition, recording and playback. Although there has been recognition of the need for training on tools used in speech analysis associated with cleft palate, little attention has been…
Groenewold, Rimke; Bastiaanse, Roelien; Nickels, Lyndsey; Huiskes, Mike
2014-01-01
Background: Previous studies have shown that in semi-spontaneous speech, individuals with Broca's and anomic aphasia produce relatively many direct speech constructions. It has been claimed that in 'healthy' communication direct speech constructions contribute to the liveliness, and indirectly to
Maternal employment and the health of low-income young children.
Gennetian, Lisa A; Hill, Heather D; London, Andrew S; Lopoo, Leonard M
2010-05-01
This study examines whether maternal employment affects the health status of low-income, elementary-school-aged children using instrumental variables estimation and experimental data from a welfare-to-work program implemented in the early 1990s. Maternal report of child health status is predicted as a function of exogenous variation in maternal employment associated with random assignment to the experimental group. IV estimates show a modest adverse effect of maternal employment on children's health. Making use of data from another welfare-to-work program we propose that any adverse effect on child health may be tempered by increased family income and access to public health insurance coverage, findings with direct relevance to a number of current policy discussions. In a secondary analysis using fixed effects techniques on longitudinal survey data collected in 1998 and 2001, we find a comparable adverse effect of maternal employment on child health that supports the external validity of our primary result.
Stevenson, Ryan A; Nelms, Caitlin E; Baum, Sarah H; Zurkovsky, Lilia; Barense, Morgan D; Newhouse, Paul A; Wallace, Mark T
2015-01-01
Over the next 2 decades, a dramatic shift in the demographics of society will take place, with a rapid growth in the population of older adults. One of the most common complaints with healthy aging is a decreased ability to successfully perceive speech, particularly in noisy environments. In such noisy environments, the presence of visual speech cues (i.e., lip movements) provide striking benefits for speech perception and comprehension, but previous research suggests that older adults gain less from such audiovisual integration than their younger peers. To determine at what processing level these behavioral differences arise in healthy-aging populations, we administered a speech-in-noise task to younger and older adults. We compared the perceptual benefits of having speech information available in both the auditory and visual modalities and examined both phoneme and whole-word recognition across varying levels of signal-to-noise ratio. For whole-word recognition, older adults relative to younger adults showed greater multisensory gains at intermediate SNRs but reduced benefit at low SNRs. By contrast, at the phoneme level both younger and older adults showed approximately equivalent increases in multisensory gain as signal-to-noise ratio decreased. Collectively, the results provide important insights into both the similarities and differences in how older and younger adults integrate auditory and visual speech cues in noisy environments and help explain some of the conflicting findings in previous studies of multisensory speech perception in healthy aging. These novel findings suggest that audiovisual processing is intact at more elementary levels of speech perception in healthy-aging populations and that deficits begin to emerge only at the more complex word-recognition level of speech signals. Copyright © 2015 Elsevier Inc. All rights reserved.
Objective support for subjective reports of successful inner speech in two people with aphasia.
Hayward, William; Snider, Sarah F; Luta, George; Friedman, Rhonda B; Turkeltaub, Peter E
2016-01-01
People with aphasia frequently report being able to say a word correctly in their heads, even if they are unable to say that word aloud. It is difficult to know what is meant by these reports of "successful inner speech". We probe the experience of successful inner speech in two people with aphasia. We show that these reports are associated with correct overt speech and phonologically related nonword errors, that they relate to word characteristics associated with ease of lexical access but not ease of production, and that they predict whether or not individual words are relearned during anomia treatment. These findings suggest that reports of successful inner speech are meaningful and may be useful to study self-monitoring in aphasia, to better understand anomia, and to predict treatment outcomes. Ultimately, the study of inner speech in people with aphasia could provide critical insights that inform our understanding of normal language.
Precision of working memory for speech sounds.
Joseph, Sabine; Iverson, Paul; Manohar, Sanjay; Fox, Zoe; Scott, Sophie K; Husain, Masud
2015-01-01
Memory for speech sounds is a key component of models of verbal working memory (WM). But how good is verbal WM? Most investigations assess this using binary report measures to derive a fixed number of items that can be stored. However, recent findings in visual WM have challenged such "quantized" views by employing measures of recall precision with an analogue response scale. WM for speech sounds might rely on both continuous and categorical storage mechanisms. Using a novel speech matching paradigm, we measured WM recall precision for phonemes. Vowel qualities were sampled from a formant space continuum. A probe vowel had to be adjusted to match the vowel quality of a target on a continuous, analogue response scale. Crucially, this provided an index of the variability of a memory representation around its true value and thus allowed us to estimate how memories were distorted from the original sounds. Memory load affected the quality of speech sound recall in two ways. First, there was a gradual decline in recall precision with increasing number of items, consistent with the view that WM representations of speech sounds become noisier with an increase in the number of items held in memory, just as for vision. Based on multidimensional scaling (MDS), the level of noise appeared to be reflected in distortions of the formant space. Second, as memory load increased, there was evidence of greater clustering of participants' responses around particular vowels. A mixture model captured both continuous and categorical responses, demonstrating a shift from continuous to categorical memory with increasing WM load. This suggests that direct acoustic storage can be used for single items, but when more items must be stored, categorical representations must be used.
Gao, Yayue; Wang, Qian; Ding, Yu; Wang, Changming; Li, Haifeng; Wu, Xihong; Qu, Tianshu; Li, Liang
2017-01-01
Human listeners are able to selectively attend to target speech in a noisy environment with multiple-people talking. Using recordings of scalp electroencephalogram (EEG), this study investigated how selective attention facilitates the cortical representation of target speech under a simulated "cocktail-party" listening condition with speech-on-speech masking. The result shows that the cortical representation of target-speech signals under the multiple-people talking condition was specifically improved by selective attention relative to the non-selective-attention listening condition, and the beta-band activity was most strongly modulated by selective attention. Moreover, measured with the Granger Causality value, selective attention to the single target speech in the mixed-speech complex enhanced the following four causal connectivities for the beta-band oscillation: the ones (1) from site FT7 to the right motor area, (2) from the left frontal area to the right motor area, (3) from the central frontal area to the right motor area, and (4) from the central frontal area to the right frontal area. However, the selective-attention-induced change in beta-band causal connectivity from the central frontal area to the right motor area, but not other beta-band causal connectivities, was significantly correlated with the selective-attention-induced change in the cortical beta-band representation of target speech. These findings suggest that under the "cocktail-party" listening condition, the beta-band oscillation in EEGs to target speech is specifically facilitated by selective attention to the target speech that is embedded in the mixed-speech complex. The selective attention-induced unmasking of target speech may be associated with the improved beta-band functional connectivity from the central frontal area to the right motor area, suggesting a top-down attentional modulation of the speech-motor process.
Kayasith, Prakasith; Theeramunkong, Thanaruk
It is a tedious and subjective task to measure severity of a dysarthria by manually evaluating his/her speech using available standard assessment methods based on human perception. This paper presents an automated approach to assess speech quality of a dysarthric speaker with cerebral palsy. With the consideration of two complementary factors, speech consistency and speech distinction, a speech quality indicator called speech clarity index (Ψ) is proposed as a measure of the speaker's ability to produce consistent speech signal for a certain word and distinguished speech signal for different words. As an application, it can be used to assess speech quality and forecast speech recognition rate of speech made by an individual dysarthric speaker before actual exhaustive implementation of an automatic speech recognition system for the speaker. The effectiveness of Ψ as a speech recognition rate predictor is evaluated by rank-order inconsistency, correlation coefficient, and root-mean-square of difference. The evaluations had been done by comparing its predicted recognition rates with ones predicted by the standard methods called the articulatory and intelligibility tests based on the two recognition systems (HMM and ANN). The results show that Ψ is a promising indicator for predicting recognition rate of dysarthric speech. All experiments had been done on speech corpus composed of speech data from eight normal speakers and eight dysarthric speakers.
Automated Speech Rate Measurement in Dysarthria
Martens, Heidi; Dekens, Tomas; Van Nuffelen, Gwen; Latacz, Lukas; Verhelst, Werner; De Bodt, Marc
2015-01-01
Purpose: In this study, a new algorithm for automated determination of speech rate (SR) in dysarthric speech is evaluated. We investigated how reliably the algorithm calculates the SR of dysarthric speech samples when compared with calculation performed by speech-language pathologists. Method: The new algorithm was trained and tested using Dutch…
Oommen, Elizabeth R; McCarthy, John W
2015-03-01
In childhood apraxia of speech (CAS), children exhibit varying levels of speech intelligibility depending on the nature of errors in articulation and prosody. Augmentative and alternative communication (AAC) strategies are beneficial, and commonly adopted with children with CAS. This study focused on the decision-making process and strategies adopted by speech-language pathologists (SLPs) when simultaneously implementing interventions that focused on natural speech and AAC. Eight SLPs, with significant clinical experience in CAS and AAC interventions, participated in an online focus group. Thematic analysis revealed eight themes: key decision-making factors; treatment history and rationale; benefits; challenges; therapy strategies and activities; collaboration with team members; recommendations; and other comments. Results are discussed along with clinical implications and directions for future research.
Speech Recognition on Mobile Devices
DEFF Research Database (Denmark)
Tan, Zheng-Hua; Lindberg, Børge
2010-01-01
in the mobile context covering motivations, challenges, fundamental techniques and applications. Three ASR architectures are introduced: embedded speech recognition, distributed speech recognition and network speech recognition. Their pros and cons and implementation issues are discussed. Applications within......The enthusiasm of deploying automatic speech recognition (ASR) on mobile devices is driven both by remarkable advances in ASR technology and by the demand for efficient user interfaces on such devices as mobile phones and personal digital assistants (PDAs). This chapter presents an overview of ASR...
Song and speech: examining the link between singing talent and speech imitation ability.
Christiner, Markus; Reiterer, Susanne M
2013-01-01
In previous research on speech imitation, musicality, and an ability to sing were isolated as the strongest indicators of good pronunciation skills in foreign languages. We, therefore, wanted to take a closer look at the nature of the ability to sing, which shares a common ground with the ability to imitate speech. This study focuses on whether good singing performance predicts good speech imitation. Forty-one singers of different levels of proficiency were selected for the study and their ability to sing, to imitate speech, their musical talent and working memory were tested. Results indicated that singing performance is a better indicator of the ability to imitate speech than the playing of a musical instrument. A multiple regression revealed that 64% of the speech imitation score variance could be explained by working memory together with educational background and singing performance. A second multiple regression showed that 66% of the speech imitation variance of completely unintelligible and unfamiliar language stimuli (Hindi) could be explained by working memory together with a singer's sense of rhythm and quality of voice. This supports the idea that both vocal behaviors have a common grounding in terms of vocal and motor flexibility, ontogenetic and phylogenetic development, neural orchestration and auditory memory with singing fitting better into the category of "speech" on the productive level and "music" on the acoustic level. As a result, good singers benefit from vocal and motor flexibility, productively and cognitively, in three ways. (1) Motor flexibility and the ability to sing improve language and musical function. (2) Good singers retain a certain plasticity and are open to new and unusual sound combinations during adulthood both perceptually and productively. (3) The ability to sing improves the memory span of the auditory working memory.
Song and speech: examining the link between singing talent and speech imitation ability
Directory of Open Access Journals (Sweden)
Markus eChristiner
2013-11-01
Full Text Available In previous research on speech imitation, musicality and an ability to sing were isolated as the strongest indicators of good pronunciation skills in foreign languages. We, therefore, wanted to take a closer look at the nature of the ability to sing, which shares a common ground with the ability to imitate speech. This study focuses on whether good singing performance predicts good speech imitation. Fourty-one singers of different levels of proficiency were selected for the study and their ability to sing, to imitate speech, their musical talent and working memory were tested. Results indicated that singing performance is a better indicator of the ability to imitate speech than the playing of a musical instrument. A multiple regression revealed that 64 % of the speech imitation score variance could be explained by working memory together with educational background and singing performance. A second multiple regression showed that 66 % of the speech imitation variance of completely unintelligible and unfamiliar language stimuli (Hindi could be explained by working memory together with a singer’s sense of rhythm and quality of voice. This supports the idea that both vocal behaviors have a common grounding in terms of vocal and motor flexibility, ontogenetic and phylogenetic development, neural orchestration and sound memory with singing fitting better into the category of "speech" on the productive level and "music" on the acoustic level. As a result, good singers benefit from vocal and motor flexibility, productively and cognitively, in three ways. 1. Motor flexibility and the ability to sing improve language and musical function. 2. Good singers retain a certain plasticity and are open to new and unusual sound combinations during adulthood both perceptually and productively. 3. The ability to sing improves the memory span of the auditory short term memory.
Maternal Employment and Children's Academic Achievement: Parenting Styles as Mediating Variable.
Beyer, Sylvia
1995-01-01
Provides a review and integration of findings on the effects of parenting styles and maternal employment on children's academic achievement. Presents a model in which it is argued that maternal employment status has little, if any, direct effect on children's academic achievement. Suggests maternal employment affects parenting styles, which in…
Sandor, Aniko; Moses, Haifa
2016-01-01
Speech alarms have been used extensively in aviation and included in International Building Codes (IBC) and National Fire Protection Association's (NFPA) Life Safety Code. However, they have not been implemented on space vehicles. Previous studies conducted at NASA JSC showed that speech alarms lead to faster identification and higher accuracy. This research evaluated updated speech and tone alerts in a laboratory environment and in the Human Exploration Research Analog (HERA) in a realistic setup.
Zhao, Wanying; Riggs, Kevin; Schindler, Igor; Holle, Henning
2018-02-21
Language and action naturally occur together in the form of cospeech gestures, and there is now convincing evidence that listeners display a strong tendency to integrate semantic information from both domains during comprehension. A contentious question, however, has been which brain areas are causally involved in this integration process. In previous neuroimaging studies, left inferior frontal gyrus (IFG) and posterior middle temporal gyrus (pMTG) have emerged as candidate areas; however, it is currently not clear whether these areas are causally or merely epiphenomenally involved in gesture-speech integration. In the present series of experiments, we directly tested for a potential critical role of IFG and pMTG by observing the effect of disrupting activity in these areas using transcranial magnetic stimulation in a mixed gender sample of healthy human volunteers. The outcome measure was performance on a Stroop-like gesture task (Kelly et al., 2010a), which provides a behavioral index of gesture-speech integration. Our results provide clear evidence that disrupting activity in IFG and pMTG selectively impairs gesture-speech integration, suggesting that both areas are causally involved in the process. These findings are consistent with the idea that these areas play a joint role in gesture-speech integration, with IFG regulating strategic semantic access via top-down signals acting upon temporal storage areas. SIGNIFICANCE STATEMENT Previous neuroimaging studies suggest an involvement of inferior frontal gyrus and posterior middle temporal gyrus in gesture-speech integration, but findings have been mixed and due to methodological constraints did not allow inferences of causality. By adopting a virtual lesion approach involving transcranial magnetic stimulation, the present study provides clear evidence that both areas are causally involved in combining semantic information arising from gesture and speech. These findings support the view that, rather than being
Freedom of Speech Newsletter, September, 1975.
Allen, Winfred G., Jr., Ed.
The Freedom of Speech Newsletter is the communication medium for the Freedom of Speech Interest Group of the Western Speech Communication Association. The newsletter contains such features as a statement of concern by the National Ad Hoc Committee Against Censorship; Reticence and Free Speech, an article by James F. Vickrey discussing the subtle…
Maternal response to child affect: Role of maternal depression and relationship quality.
Morgan, Judith K; Ambrosia, Marigrace; Forbes, Erika E; Cyranowski, Jill M; Amole, Marlissa C; Silk, Jennifer S; Elliott, Rosalind D; Swartz, Holly A
2015-11-15
Maternal depression is associated with negative outcomes for offspring, including increased incidence of child psychopathology. Quality of mother-child relationships can be compromised among affectively ill dyads, such as those characterized by maternal depression and child psychopathology, and negatively impact outcomes bidirectionally. Little is known about the neural mechanisms that may modulate depressed mothers' responses to their psychiatrically ill children during middle childhood and adolescence, partially because of a need for ecologically valid personally relevant fMRI tasks that might most effectively elicit these neural mechanisms. The current project evaluated maternal response to child positive and negative affective video clips in 19 depressed mothers with psychiatrically ill offspring using a novel fMRI task. The task elicited activation in the ventral striatum when mothers viewed positive clips and insula when mothers viewed negative clips of their own (versus unfamiliar) children. Both types of clips elicited activation in regions associated with affect regulation and self-related and social processing. Greater lifetime number of depressive episodes, comorbid anxiety, and poor mother-child relationship quality all emerged as predictors of maternal response to child affect. Findings may be specific to dyads with psychiatrically ill children. Altered neural response to child affect may be an important characteristic of chronic maternal depression and may impact mother-child relationships negatively. Existing interventions for depression may be improved by helping mothers respond to their children's affect more adaptively. Copyright © 2015 Elsevier B.V. All rights reserved.
An investigation of co-speech gesture production during action description in Parkinson's disease.
Cleary, Rebecca A; Poliakoff, Ellen; Galpin, Adam; Dick, Jeremy P R; Holler, Judith
2011-12-01
Parkinson's disease (PD) can impact enormously on speech communication. One aspect of non-verbal behaviour closely tied to speech is co-speech gesture production. In healthy people, co-speech gestures can add significant meaning and emphasis to speech. There is, however, little research into how this important channel of communication is affected in PD. The present study provides a systematic analysis of co-speech gestures which spontaneously accompany the description of actions in a group of PD patients (N = 23, Hoehn and Yahr Stage III or less) and age-matched healthy controls (N = 22). The analysis considers different co-speech gesture types, using established classification schemes from the field of gesture research. The analysis focuses on the rate of these gestures as well as on their qualitative nature. In doing so, the analysis attempts to overcome several methodological shortcomings of research in this area. Contrary to expectation, gesture rate was not significantly affected in our patient group, with relatively mild PD. This indicates that co-speech gestures could compensate for speech problems. However, while gesture rate seems unaffected, the qualitative precision of gestures representing actions was significantly reduced. This study demonstrates the feasibility of carrying out fine-grained, detailed analyses of gestures in PD and offers insights into an as yet neglected facet of communication in patients with PD. Based on the present findings, an important next step is the closer investigation of the qualitative changes in gesture (including different communicative situations) and an analysis of the heterogeneity in co-speech gesture production in PD. Copyright © 2011 Elsevier Ltd. All rights reserved.
Automatic speech recognition used for evaluation of text-to-speech systems
Czech Academy of Sciences Publication Activity Database
Vích, Robert; Nouza, J.; Vondra, Martin
-, č. 5042 (2008), s. 136-148 ISSN 0302-9743 R&D Projects: GA AV ČR 1ET301710509; GA AV ČR 1QS108040569 Institutional research plan: CEZ:AV0Z20670512 Keywords : speech recognition * speech processing Subject RIV: JA - Electronics ; Optoelectronics, Electrical Engineering
Former Auctioneer Finds Voice After Aphasia
... Aphasia Follow us Former Auctioneer Finds Voice After Aphasia Speech impairment changed his life One unremarkable September ... 10 Tips for Communicating with Someone who has Aphasia Talk to them in a quiet, calm, relaxed ...
SynFace—Speech-Driven Facial Animation for Virtual Speech-Reading Support
Directory of Open Access Journals (Sweden)
Giampiero Salvi
2009-01-01
Full Text Available This paper describes SynFace, a supportive technology that aims at enhancing audio-based spoken communication in adverse acoustic conditions by providing the missing visual information in the form of an animated talking head. Firstly, we describe the system architecture, consisting of a 3D animated face model controlled from the speech input by a specifically optimised phonetic recogniser. Secondly, we report on speech intelligibility experiments with focus on multilinguality and robustness to audio quality. The system, already available for Swedish, English, and Flemish, was optimised for German and for Swedish wide-band speech quality available in TV, radio, and Internet communication. Lastly, the paper covers experiments with nonverbal motions driven from the speech signal. It is shown that turn-taking gestures can be used to affect the flow of human-human dialogues. We have focused specifically on two categories of cues that may be extracted from the acoustic signal: prominence/emphasis and interactional cues (turn-taking/back-channelling.
Maternal employment and birth outcomes
DEFF Research Database (Denmark)
Wüst, Miriam
selection of mothers between pregnancies drives the results, I focus on mothers whose change in employment status is likely not to be driven by underlying health (unemployed mothers and students). Given generous welfare bene ts and strict workplace regulations in Denmark, my findings support a residual......I use Danish survey and administrative data to examine the impact of maternal employment during pregnancy on birth outcomes. As healthier mothers are more likely to work and health shocks to mothers may impact employment and birth outcomes, I combine two strategies: First, I control extensively...... for time-varying factors that may correlate with employment and birth outcomes, such as pre-pregnancy family income and maternal occupation, pregnancy-related health shocks, maternal sick listing, and health behaviors (smoking and alcohol consumption). Second, to account for remaining time...
An analysis of the masking of speech by competing speech using self-report data (L)
Agus, Trevor R.; Akeroyd, Michael A.; Noble, William; Bhullar, Navjot
2009-01-01
Many of the items in the “Speech, Spatial, and Qualities of Hearing” scale questionnaire [S. Gatehouse and W. Noble, Int. J. Audiol.43, 85–99 (2004)] are concerned with speech understanding in a variety of backgrounds, both speech and nonspeech. To study if this self-report data reflected informational masking, previously collected data on 414 people were analyzed. The lowest scores (greatest difficulties) were found for the two items in which there were two speech targets, with successively ...
Shearer, William M.
Written for students in the fields of speech correction and audiology, the text deals with the following: structures involved in respiration; the skeleton and the processes of inhalation and exhalation; phonation and pitch, the larynx, and esophageal speech; muscles involved in articulation; muscles involved in resonance; and the anatomy of the…
Bortolussi, Michael R.; Vidulich, Michael A.
1991-01-01
The potential benefit of speech as a control modality has been investigated with mixed results. Earlier studies suggests that speech controls can reduce the potential of manual control overloads and improve time-sharing performance. However, these benefits were not without costs. Pilots reported higher workload levels associated with the use of speech controls. To further investigate these previous findings, an experiment was conducted in a simulation of an advanced single-pilot, scout/attack helicopter at NASA-Ames' ICAB (interchangeable cab) facility. Objective performance data suggested that speech control modality was effective in reducing interference of discrete, time-shared responses during continuous flight control activity. Subjective ratings, however, indicated that the speech control modality increased workload. Post-flight debriefing indicated that these results were mainly due to the increased effort to speak precisely to a less than perfect voice recognition system.
Maternal high-fat diet during pregnancy and lactation affects hepatic ...
Indian Academy of Sciences (India)
2017-04-18
Apr 18, 2017 ... pregnancy, several studies suggest that maternal obesity might lead to insulin ... High-fat food consumption by maternal dams during re- stricted periods ..... Our findings might be of significance in uncovering the association of.
Maternal Mortality At The State Specialist Hospital Bauchi, Northern ...
African Journals Online (AJOL)
Objective: To analyse and document our experiences with maternal mortality with the view of finding the trends over the last seven years, common causes and attributing socio-demographic factors. Design: A prospective analysis of maternal mortality. Setting: State Specialists Hospital Bauchi, Bauchi Northeastern Nigeria.
Speech Entrainment Compensates for Broca's Area Damage
Fridriksson, Julius; Basilakos, Alexandra; Hickok, Gregory; Bonilha, Leonardo; Rorden, Chris
2015-01-01
Speech entrainment (SE), the online mimicking of an audiovisual speech model, has been shown to increase speech fluency in patients with Broca's aphasia. However, not all individuals with aphasia benefit from SE. The purpose of this study was to identify patterns of cortical damage that predict a positive response SE's fluency-inducing effects. Forty-four chronic patients with left hemisphere stroke (15 female) were included in this study. Participants completed two tasks: 1) spontaneous speech production, and 2) audiovisual SE. Number of different words per minute was calculated as a speech output measure for each task, with the difference between SE and spontaneous speech conditions yielding a measure of fluency improvement. Voxel-wise lesion-symptom mapping (VLSM) was used to relate the number of different words per minute for spontaneous speech, SE, and SE-related improvement to patterns of brain damage in order to predict lesion locations associated with the fluency-inducing response to speech entrainment. Individuals with Broca's aphasia demonstrated a significant increase in different words per minute during speech entrainment versus spontaneous speech. A similar pattern of improvement was not seen in patients with other types of aphasia. VLSM analysis revealed damage to the inferior frontal gyrus predicted this response. Results suggest that SE exerts its fluency-inducing effects by providing a surrogate target for speech production via internal monitoring processes. Clinically, these results add further support for the use of speech entrainment to improve speech production and may help select patients for speech entrainment treatment. PMID:25989443
Castral, T C; Warnock, F; Dos Santos, C B; Daré, M F; Moreira, A C; Antonini, S R R; Scochi, C G S
2015-03-01
Maternal kangaroo care (MKC) is a naturalistic intervention that alleviates neonatal pain, and mothers are assumed to play a stress regulatory role in MKC. Yet, no MKC infant pain study has examined relationship between maternal and infant stress reactivity concurrently, or whether post-partum depression and/or anxiety (PPDA) alters maternal and neonatal stress response and the regulatory effects of MKC. To examine the concordance of salivary cortisol reactivity between 42 mothers and their stable preterm infants during routine infant heel lance (HL) while in MKC and to compare salivary cortisol between groups of mothers with and without PPDA and their infants. Maternal and infant salivary cortisol samples were collected pre-HL and 20 min post-HL with two additional maternal samples at night and in the morning. Mothers and infants were allocated to with PPDA versus without PPDA study groups on the basis of maternal post-natal mental health assessment scores. Higher mothers' cortisol pre-HL was weakly associated with higher infants' salivary cortisol in response to the HL procedure. Maternal depression and/or anxiety were not associated with infants' cortisol. During HL, both groups of mothers and infants showed no change in salivary cortisol. Concordance between mother and infant salivary cortisol supports the maternal stress regulatory role in MKC. MKC may have stress regulatory benefits for mothers and their preterm infants during HL independent of PPDA. Future MKC studies that target mothers with altered mood will help to build on these findings. © 2014 European Pain Federation - EFIC®
Prosody Analysis of Mariah Carey’s 1999 Billboard Award Speech and Its Contribution to EFL
Directory of Open Access Journals (Sweden)
Heri Setiawan
2017-07-01
Full Text Available Prosodic elements are the essential elements in human beings’ speech that can be used to know what the speakers of an utterance exactly mean when they are speaking. This research was a work of analysis on prosody of speech to know the emotion and the meaning of utterances. Descriptive qualitative type of research model was employed in this research. The object of the research was Mariah Carey’s speech when she was receiving a special award named the Billboard artist of the decade in 1999. The researcher was interested in not only knowing about Mariah Carey’s emotion and meaning when she was delivering her speech but also finding other uses of prosodic elements which were not found in the supporting previous literatures used in this research. Moreover, the possible contribution which these research findings could contribute to the EFL teaching and learning was also addressed. It is evident that prosody definitely has significant contribution in EFL teaching and learning especially in improving speaking and reading comprehension skills
Investigations on search methods for speech recognition using weighted finite state transducers
Rybach, David
2014-01-01
The search problem in the statistical approach to speech recognition is to find the most likely word sequence for an observed speech signal using a combination of knowledge sources, i.e. the language model, the pronunciation model, and the acoustic models of phones. The resulting search space is enormous. Therefore, an efficient search strategy is required to compute the result with a feasible amount of time and memory. The structured statistical models as well as their combination, the searc...
Association of maternal age with child health: A Japanese longitudinal study.
Directory of Open Access Journals (Sweden)
Tsuguhiko Kato
Full Text Available Average maternal age at birth has been rising steadily in Western and some Asian countries. Older maternal age has been associated with adverse pregnancy and birth outcomes; however, studies on the relationship between maternal age and young children's health remain scarce. Therefore, we sought to investigate the association of maternal age with child health outcomes in the Japanese population. We analyzed data from two birth cohorts of the nationwide Japanese Longitudinal Survey of Babies in 21st Century (n2001 = 47,715 and n2010 = 38,554. We estimated risks of unintentional injuries and hospital admissions at 18 and 66 months according to maternal age, controlling for the following potential confounders: parental education; maternal parity, smoking status, and employment status; household income; paternal age, and sex of the child. We also included the following as potential mediators: preterm births and birthweight. We observed a decreasing trend in the risks of children's unintentional injuries and hospital admissions at 18 months according to maternal age in both cohorts. In the 2001 cohort, compared to mothers 40.0 years, respectively, controlling for confounders. Our findings were in line with previous findings from population-based studies conducted in the United Kingdom and Canada suggesting that older maternal age may be beneficial for early child health.
Basilakos, Alexandra; Rorden, Chris; Bonilha, Leonardo; Moser, Dana; Fridriksson, Julius
2015-06-01
Acquired apraxia of speech (AOS) is a motor speech disorder caused by brain damage. AOS often co-occurs with aphasia, a language disorder in which patients may also demonstrate speech production errors. The overlap of speech production deficits in both disorders has raised questions on whether AOS emerges from a unique pattern of brain damage or as a subelement of the aphasic syndrome. The purpose of this study was to determine whether speech production errors in AOS and aphasia are associated with distinctive patterns of brain injury. Forty-three patients with history of a single left-hemisphere stroke underwent comprehensive speech and language testing. The AOS Rating Scale was used to rate speech errors specific to AOS versus speech errors that can also be associated with both AOS and aphasia. Localized brain damage was identified using structural magnetic resonance imaging, and voxel-based lesion-impairment mapping was used to evaluate the relationship between speech errors specific to AOS, those that can occur in AOS or aphasia, and brain damage. The pattern of brain damage associated with AOS was most strongly associated with damage to cortical motor regions, with additional involvement of somatosensory areas. Speech production deficits that could be attributed to AOS or aphasia were associated with damage to the temporal lobe and the inferior precentral frontal regions. AOS likely occurs in conjunction with aphasia because of the proximity of the brain areas supporting speech and language, but the neurobiological substrate for each disorder differs. © 2015 American Heart Association, Inc.
Bowers, Andrew; Saltuklaroglu, Tim; Harkrider, Ashley; Cuellar, Megan
2013-01-01
Constructivist theories propose that articulatory hypotheses about incoming phonetic targets may function to enhance perception by limiting the possibilities for sensory analysis. To provide evidence for this proposal, it is necessary to map ongoing, high-temporal resolution changes in sensorimotor activity (i.e., the sensorimotor μ rhythm) to accurate speech and non-speech discrimination performance (i.e., correct trials.). Sixteen participants (15 female and 1 male) were asked to passively listen to or actively identify speech and tone-sweeps in a two-force choice discrimination task while the electroencephalograph (EEG) was recorded from 32 channels. The stimuli were presented at signal-to-noise ratios (SNRs) in which discrimination accuracy was high (i.e., 80-100%) and low SNRs producing discrimination performance at chance. EEG data were decomposed using independent component analysis and clustered across participants using principle component methods in EEGLAB. ICA revealed left and right sensorimotor µ components for 14/16 and 13/16 participants respectively that were identified on the basis of scalp topography, spectral peaks, and localization to the precentral and postcentral gyri. Time-frequency analysis of left and right lateralized µ component clusters revealed significant (pFDRspeech discrimination trials relative to chance trials following stimulus offset. Findings are consistent with constructivist, internal model theories proposing that early forward motor models generate predictions about likely phonemic units that are then synthesized with incoming sensory cues during active as opposed to passive processing. Future directions and possible translational value for clinical populations in which sensorimotor integration may play a functional role are discussed.
A NOVEL APPROACH TO STUTTERED SPEECH CORRECTION
Directory of Open Access Journals (Sweden)
Alim Sabur Ajibola
2016-06-01
Full Text Available Stuttered speech is a dysfluency rich speech, more prevalent in males than females. It has been associated with insufficient air pressure or poor articulation, even though the root causes are more complex. The primary features include prolonged speech and repetitive speech, while some of its secondary features include, anxiety, fear, and shame. This study used LPC analysis and synthesis algorithms to reconstruct the stuttered speech. The results were evaluated using cepstral distance, Itakura-Saito distance, mean square error, and likelihood ratio. These measures implied perfect speech reconstruction quality. ASR was used for further testing, and the results showed that all the reconstructed speech samples were perfectly recognized while only three samples of the original speech were perfectly recognized.
Prisoner Fasting as Symbolic Speech: The Ultimate Speech-Action Test.
Sneed, Don; Stonecipher, Harry W.
The ultimate test of the speech-action dichotomy, as it relates to symbolic speech to be considered by the courts, may be the fasting of prison inmates who use hunger strikes to protest the conditions of their confinement or to make political statements. While hunger strikes have been utilized by prisoners for years as a means of protest, it was…
Implicit Talker Training Improves Comprehension of Auditory Speech in Noise
Directory of Open Access Journals (Sweden)
Jens Kreitewolf
2017-09-01
Full Text Available Previous studies have shown that listeners are better able to understand speech when they are familiar with the talker’s voice. In most of these studies, talker familiarity was ensured by explicit voice training; that is, listeners learned to identify the familiar talkers. In the real world, however, the characteristics of familiar talkers are learned incidentally, through communication. The present study investigated whether speech comprehension benefits from implicit voice training; that is, through exposure to talkers’ voices without listeners explicitly trying to identify them. During four training sessions, listeners heard short sentences containing a single verb (e.g., “he writes”, spoken by one talker. The sentences were mixed with noise, and listeners identified the verb within each sentence while their speech-reception thresholds (SRT were measured. In a final test session, listeners performed the same task, but this time they heard different sentences spoken by the familiar talker and three unfamiliar talkers. Familiar and unfamiliar talkers were counterbalanced across listeners. Half of the listeners performed a test session in which the four talkers were presented in separate blocks (blocked paradigm. For the other half, talkers varied randomly from trial to trial (interleaved paradigm. The results showed that listeners had lower SRT when the speech was produced by the familiar talker than the unfamiliar talkers. The type of talker presentation (blocked vs. interleaved had no effect on this familiarity benefit. These findings suggest that listeners implicitly learn talker-specific information during a speech-comprehension task, and exploit this information to improve the comprehension of novel speech material from familiar talkers.
Aziz, Azza Adel; Shohdi, Sahar; Osman, Dalia Mostafa; Habib, Emad Iskander
2010-06-01
Childhood apraxia of speech is a neurological childhood speech-sound disorder in which the precision and consistency of movements underlying speech are impaired in the absence of neuromuscular deficits. Children with childhood apraxia of speech and those with multiple phonological disorder share some common phonological errors that can be misleading in diagnosis. This study posed a question about a possible significant difference in language, speech and non-speech oral performances between children with childhood apraxia of speech, multiple phonological disorder and normal children that can be used for a differential diagnostic purpose. 30 pre-school children between the ages of 4 and 6 years served as participants. Each of these children represented one of 3 possible subject-groups: Group 1: multiple phonological disorder; Group 2: suspected cases of childhood apraxia of speech; Group 3: control group with no communication disorder. Assessment procedures included: parent interviews; testing of non-speech oral motor skills and testing of speech skills. Data showed that children with suspected childhood apraxia of speech showed significantly lower language score only in their expressive abilities. Non-speech tasks did not identify significant differences between childhood apraxia of speech and multiple phonological disorder groups except for those which required two sequential motor performances. In speech tasks, both consonant and vowel accuracy were significantly lower and inconsistent in childhood apraxia of speech group than in the multiple phonological disorder group. Syllable number, shape and sequence accuracy differed significantly in the childhood apraxia of speech group than the other two groups. In addition, children with childhood apraxia of speech showed greater difficulty in processing prosodic features indicating a clear need to address these variables for differential diagnosis and treatment of children with childhood apraxia of speech. Copyright (c
Individual differneces in degraded speech perception
Carbonell, Kathy M.
One of the lasting concerns in audiology is the unexplained individual differences in speech perception performance even for individuals with similar audiograms. One proposal is that there are cognitive/perceptual individual differences underlying this vulnerability and that these differences are present in normal hearing (NH) individuals but do not reveal themselves in studies that use clear speech produced in quiet (because of a ceiling effect). However, previous studies have failed to uncover cognitive/perceptual variables that explain much of the variance in NH performance on more challenging degraded speech tasks. This lack of strong correlations may be due to either examining the wrong measures (e.g., working memory capacity) or to there being no reliable differences in degraded speech performance in NH listeners (i.e., variability in performance is due to measurement noise). The proposed project has 3 aims; the first, is to establish whether there are reliable individual differences in degraded speech performance for NH listeners that are sustained both across degradation types (speech in noise, compressed speech, noise-vocoded speech) and across multiple testing sessions. The second aim is to establish whether there are reliable differences in NH listeners' ability to adapt their phonetic categories based on short-term statistics both across tasks and across sessions; and finally, to determine whether performance on degraded speech perception tasks are correlated with performance on phonetic adaptability tasks, thus establishing a possible explanatory variable for individual differences in speech perception for NH and hearing impaired listeners.
Modeling the Development of Audiovisual Cue Integration in Speech Perception.
Getz, Laura M; Nordeen, Elke R; Vrabic, Sarah C; Toscano, Joseph C
2017-03-21
Adult speech perception is generally enhanced when information is provided from multiple modalities. In contrast, infants do not appear to benefit from combining auditory and visual speech information early in development. This is true despite the fact that both modalities are important to speech comprehension even at early stages of language acquisition. How then do listeners learn how to process auditory and visual information as part of a unified signal? In the auditory domain, statistical learning processes provide an excellent mechanism for acquiring phonological categories. Is this also true for the more complex problem of acquiring audiovisual correspondences, which require the learner to integrate information from multiple modalities? In this paper, we present simulations using Gaussian mixture models (GMMs) that learn cue weights and combine cues on the basis of their distributional statistics. First, we simulate the developmental process of acquiring phonological categories from auditory and visual cues, asking whether simple statistical learning approaches are sufficient for learning multi-modal representations. Second, we use this time course information to explain audiovisual speech perception in adult perceivers, including cases where auditory and visual input are mismatched. Overall, we find that domain-general statistical learning techniques allow us to model the developmental trajectory of audiovisual cue integration in speech, and in turn, allow us to better understand the mechanisms that give rise to unified percepts based on multiple cues.
Meijers, A.W.M.; Tsohatzidis, S.L.
2007-01-01
From its early development in the 1960s, speech act theory always had an individualistic orientation. It focused exclusively on speech acts performed by individual agents. Paradigmatic examples are ‘I promise that p’, ‘I order that p’, and ‘I declare that p’. There is a single speaker and a single
Speech perception in autism spectrum disorder: An activation likelihood estimation meta-analysis.
Tryfon, Ana; Foster, Nicholas E V; Sharda, Megha; Hyde, Krista L
2018-02-15
Autism spectrum disorder (ASD) is often characterized by atypical language profiles and auditory and speech processing. These can contribute to aberrant language and social communication skills in ASD. The study of the neural basis of speech perception in ASD can serve as a potential neurobiological marker of ASD early on, but mixed results across studies renders it difficult to find a reliable neural characterization of speech processing in ASD. To this aim, the present study examined the functional neural basis of speech perception in ASD versus typical development (TD) using an activation likelihood estimation (ALE) meta-analysis of 18 qualifying studies. The present study included separate analyses for TD and ASD, which allowed us to examine patterns of within-group brain activation as well as both common and distinct patterns of brain activation across the ASD and TD groups. Overall, ASD and TD showed mostly common brain activation of speech processing in bilateral superior temporal gyrus (STG) and left inferior frontal gyrus (IFG). However, the results revealed trends for some distinct activation in the TD group showing additional activation in higher-order brain areas including left superior frontal gyrus (SFG), left medial frontal gyrus (MFG), and right IFG. These results provide a more reliable neural characterization of speech processing in ASD relative to previous single neuroimaging studies and motivate future work to investigate how these brain signatures relate to behavioral measures of speech processing in ASD. Copyright © 2017 Elsevier B.V. All rights reserved.
Mother-child conversation in different social classes and communicative settings.
Hoff-Ginsberg, E
1991-08-01
30 working-class and 33 upper-middle-class mothers were videotaped in dyadic interaction with their 18-29-month-old children in 4 settings--mealtime, dressing, book reading, and toy play. Samples of the mothers' adult-directed speech also were collected. There were significant social class differences in the mothers' child-directed speech and some parallel social class differences in the mothers' adult-directed speech. These findings suggested that some social class differences in child-directed speech may be instances of more general class differences in language use. There also were main effects of communicative setting on mothers' child-directed speech and interaction effects in which setting moderated the size of the class differences in maternal speech. These findings suggested that the amount of time mothers spend interacting with their children in different contexts may be at least as important an influence on children's linguistic experience as are average characteristics of their mothers' speech.
Telomere length is longer in women with late maternal age
DEFF Research Database (Denmark)
Fagan, Erin; Sun, Fangui; Bae, Harold
2017-01-01
OBJECTIVE:: Maternal age at birth of last child has been associated with maternal longevity. The aim of this study was to determine whether older women with a history of late maternal age at last childbirth had a longer leukocyte telomere length than those with maternal age at last childbirth of 29...... died, but were at least 70 years old, were studied. Logistic regression models using generalized estimating equations were used to determine the association between tertiles of telomere length and maternal age at last childbirth, adjusting for covariates. RESULTS:: Age at birth of the last child...... in the first tertile. CONCLUSIONS:: These findings show an association between longer leukocyte telomere length and a later maternal age at birth of last child, suggesting that extended maternal age at last childbirth may be a marker for longevity....
Perceptual restoration of degraded speech is preserved with advancing age.
Saija, Jefta D; Akyürek, Elkan G; Andringa, Tjeerd C; Başkent, Deniz
2014-02-01
Cognitive skills, such as processing speed, memory functioning, and the ability to divide attention, are known to diminish with aging. The present study shows that, despite these changes, older adults can successfully compensate for degradations in speech perception. Critically, the older participants of this study were not pre-selected for high performance on cognitive tasks, but only screened for normal hearing. We measured the compensation for speech degradation using phonemic restoration, where intelligibility of degraded speech is enhanced using top-down repair mechanisms. Linguistic knowledge, Gestalt principles of perception, and expectations based on situational and linguistic context are used to effectively fill in the inaudible masked speech portions. A positive compensation effect was previously observed only with young normal hearing people, but not with older hearing-impaired populations, leaving the question whether the lack of compensation was due to aging or due to age-related hearing problems. Older participants in the present study showed poorer intelligibility of degraded speech than the younger group, as expected from previous reports of aging effects. However, in conditions that induce top-down restoration, a robust compensation was observed. Speech perception by the older group was enhanced, and the enhancement effect was similar to that observed with the younger group. This effect was even stronger with slowed-down speech, which gives more time for cognitive processing. Based on previous research, the likely explanations for these observations are that older adults can overcome age-related cognitive deterioration by relying on linguistic skills and vocabulary that they have accumulated over their lifetime. Alternatively, or simultaneously, they may use different cerebral activation patterns or exert more mental effort. This positive finding on top-down restoration skills by the older individuals suggests that new cognitive training methods
Chan, S M
2011-07-01
Although the influences of parenting on children's development of social competence have been well established, research on the underlying mechanisms of this link is relatively limited. The present study examined children's coping strategies as a mediator of the effects of maternal authoritativeness and maternal inductive responses on their social competence. The mothers of 183 Hong Kong Chinese children aged 6 to 8 years (89 girls and 94 boys) reported on their adoption of authoritative parenting and their responses to their children's expressions of emotion, and rated their children's adoption of constructive coping strategies. The children's teachers reported on the children's prosocial behaviour, and rated their level of peer acceptance at school. A model of maternal authoritativeness and supportive maternal responses affecting children's social competence is presented. The study results show that the effects of authoritative parenting on children's adoption of constructive coping strategies were mediated by supportive maternal responses to children's expression of emotion, and that the effects of maternal authoritativeness and maternal responses on children's social competence were mediated by children's coping strategies. These results suggest that school personnel should organize training programmes on emotion-coping strategies for both parents and children. The findings imply that positive parenting facilitates children's acquisition of constructive emotion-coping strategies. Programmes on emotion-coping strategies should be introduced for both parents and school children. © 2011 Blackwell Publishing Ltd.
Hurkmans, Joost; Jonkers, Roel; de Bruijn, Madeleen; Boonstra, Anne M.; Hartman, Paul P.; Arendzen, Hans; Reinders - Messelink, Heelen
2015-01-01
Background: Several studies using musical elements in the treatment of neurological language and speech disorders have reported improvement of speech production. One such programme, Speech-Music Therapy for Aphasia (SMTA), integrates speech therapy and music therapy (MT) to treat the individual with
Current trends in multilingual speech processing
Indian Academy of Sciences (India)
2016-08-26
; speech-to-speech translation; language identification. ... interest owing to two strong driving forces. Firstly, technical advances in speech recognition and synthesis are posing new challenges and opportunities to researchers.
Do long-term tongue piercings affect speech quality?
Heinen, Esther; Birkholz, Peter; Willmes, Klaus; Neuschaefer-Rube, Christiane
2017-10-01
To explore possible effects of tongue piercing on perceived speech quality. Using a quasi-experimental design, we analyzed the effect of tongue piercing on speech in a perception experiment. Samples of spontaneous speech and read speech were recorded from 20 long-term pierced and 20 non-pierced individuals (10 males, 10 females each). The individuals having a tongue piercing were recorded with attached and removed piercing. The audio samples were blindly rated by 26 female and 20 male laypersons and by 5 female speech-language pathologists with regard to perceived speech quality along 5 dimensions: speech clarity, speech rate, prosody, rhythm and fluency. We found no statistically significant differences for any of the speech quality dimensions between the pierced and non-pierced individuals, neither for the read nor for the spontaneous speech. In addition, neither length nor position of piercing had a significant effect on speech quality. The removal of tongue piercings had no effects on speech performance either. Rating differences between laypersons and speech-language pathologists were not dependent on the presence of a tongue piercing. People are able to perfectly adapt their articulation to long-term tongue piercings such that their speech quality is not perceptually affected.
Basilakos, Alexandra; Rorden, Chris; Bonilha, Leonardo; Moser, Dana; Fridriksson, Julius
2015-01-01
Background and Purpose Acquired apraxia of speech (AOS) is a motor speech disorder caused by brain damage. AOS often co-occurs with aphasia, a language disorder in which patients may also demonstrate speech production errors. The overlap of speech production deficits in both disorders has raised questions regarding if AOS emerges from a unique pattern of brain damage or as a sub-element of the aphasic syndrome. The purpose of this study was to determine whether speech production errors in AOS and aphasia are associated with distinctive patterns of brain injury. Methods Forty-three patients with history of a single left-hemisphere stroke underwent comprehensive speech and language testing. The Apraxia of Speech Rating Scale was used to rate speech errors specific to AOS versus speech errors that can also be associated with AOS and/or aphasia. Localized brain damage was identified using structural MRI, and voxel-based lesion-impairment mapping was used to evaluate the relationship between speech errors specific to AOS, those that can occur in AOS and/or aphasia, and brain damage. Results The pattern of brain damage associated with AOS was most strongly associated with damage to cortical motor regions, with additional involvement of somatosensory areas. Speech production deficits that could be attributed to AOS and/or aphasia were associated with damage to the temporal lobe and the inferior pre-central frontal regions. Conclusion AOS likely occurs in conjunction with aphasia due to the proximity of the brain areas supporting speech and language, but the neurobiological substrate for each disorder differs. PMID:25908457
Progressive apraxia of speech as a window into the study of speech planning processes.
Laganaro, Marina; Croisier, Michèle; Bagou, Odile; Assal, Frédéric
2012-09-01
We present a 3-year follow-up study of a patient with progressive apraxia of speech (PAoS), aimed at investigating whether the theoretical organization of phonetic encoding is reflected in the progressive disruption of speech. As decreased speech rate was the most striking pattern of disruption during the first 2 years, durational analyses were carried out longitudinally on syllables excised from spontaneous, repetition and reading speech samples. The crucial result of the present study is the demonstration of an effect of syllable frequency on duration: the progressive disruption of articulation rate did not affect all syllables in the same way, but followed a gradient that was function of the frequency of use of syllable-sized motor programs. The combination of data from this case of PAoS with previous psycholinguistic and neurolinguistic data, points to a frequency organization of syllable-sized speech-motor plans. In this study we also illustrate how studying PAoS can be exploited in theoretical and clinical investigations of phonetic encoding as it represents a unique opportunity to investigate speech while it progressively disrupts. Copyright © 2011 Elsevier Srl. All rights reserved.
Maternal-Zygotic Epistasis and the Evolution of Genetic Diseases
Directory of Open Access Journals (Sweden)
Nicholas K. Priest
2010-01-01
Full Text Available Many birth defects and genetic diseases are expressed in individuals that do not carry the disease causing alleles. Genetic diseases observed in offspring can be caused by gene expression in mothers and by interactions between gene expression in mothers and offspring. It is not clear whether the underlying pattern of gene expression (maternal versus offspring affects the incidence of genetic disease. Here we develop a 2-locus population genetic model with epistatic interactions between a maternal gene and a zygotic gene to address this question. We show that maternal effect genes that affect disease susceptibility in offspring persist longer and at higher frequencies in a population than offspring genes with the same effects. We find that specific forms of maternal-zygotic epistasis can maintain disease causing alleles at high frequencies over a range of plausible values. Our findings suggest that the strength and form of epistasis and the underlying pattern of gene expression may greatly influence the prevalence of human genetic diseases.
DEFF Research Database (Denmark)
Madsen, Sara Miay Kim; Whiteford, Kelly L.; Oxenham, Andrew J.
2017-01-01
Recent studies disagree on whether musicians have an advantage over non-musicians in understanding speech in noise. However, it has been suggested that musicians may be able to use diferences in fundamental frequency (F0) to better understand target speech in the presence of interfering talkers....... Here we studied a relatively large (N=60) cohort of young adults, equally divided between nonmusicians and highly trained musicians, to test whether the musicians were better able to understand speech either in noise or in a two-talker competing speech masker. The target speech and competing speech...... were presented with either their natural F0 contours or on a monotone F0, and the F0 diference between the target and masker was systematically varied. As expected, speech intelligibility improved with increasing F0 diference between the target and the two-talker masker for both natural and monotone...
A Persian Cued Speech Website Fromthe Deaf Professionals’ Views
Directory of Open Access Journals (Sweden)
Guita Movallali
2013-10-01
Full Text Available Objectives: Increasingly people are using the internet to find information about medical and educational issues and one of the simplest ways to obtain information is internet. Persian Cued Speech is a very new system to Iranian families with deaf child and the professionals and a few educators have enough knowledge about it, so the purpose of this study was to introduce Persian Cued Speech website to deaf educators and rehabilitation professionals and assess their views about the website and their accessibility to important information through its use. Methods:The sample group was randomly selected fromdeaf educators and rehabilitation professionals working in different educational settings for deaf children in Tehran, our capital .They completed a questionnaire which was adopted from different website assessing questionnaires. Researchers also completed an interview with the sample group. Results: Our findings show that from the deaf educators and rehabilitation professionals point of view,the Persian Cued Speech website was a helpful and informing website. They also wanted more pictures and videos,bigger font sizes and more practical parts.So we decide to change some parts of the website to be more usable for them. Discussion: Using informational websites can be a very helpful tool in the internet area. Helpful websites are those which are more accessible, readable and appropriately designed and are user-friendly.
Novel Techniques for Dialectal Arabic Speech Recognition
Elmahdy, Mohamed; Minker, Wolfgang
2012-01-01
Novel Techniques for Dialectal Arabic Speech describes approaches to improve automatic speech recognition for dialectal Arabic. Since speech resources for dialectal Arabic speech recognition are very sparse, the authors describe how existing Modern Standard Arabic (MSA) speech data can be applied to dialectal Arabic speech recognition, while assuming that MSA is always a second language for all Arabic speakers. In this book, Egyptian Colloquial Arabic (ECA) has been chosen as a typical Arabic dialect. ECA is the first ranked Arabic dialect in terms of number of speakers, and a high quality ECA speech corpus with accurate phonetic transcription has been collected. MSA acoustic models were trained using news broadcast speech. In order to cross-lingually use MSA in dialectal Arabic speech recognition, the authors have normalized the phoneme sets for MSA and ECA. After this normalization, they have applied state-of-the-art acoustic model adaptation techniques like Maximum Likelihood Linear Regression (MLLR) and M...
Depression, Sensation Seeking, and Maternal Smoking as Predictors of Adolescent Cigarette Smoking
Directory of Open Access Journals (Sweden)
Judy van de Venne
2006-01-01
Full Text Available The purpose of this study was to examine maternal and adolescent depression, maternal and teen sensation seeking, and maternal smoking, and their associations with adolescent smoking. Data were collected from a sample of 47 male and 66 female adolescents (ages 11—18 years and their mothers from three different health clinics. The findings indicated that maternal sensation seeking was linked indirectly with adolescent smoking through teen sensation seeking, both of which were significantly associated with teen smoking (β = 0.29, p < 0.001 and β = 0.32, p < 0.001, respectively. Teen depression was associated positively with teen smoking (β = 0.24, p < 0.01 when controlling for sensation seeking behaviors. Maternal smoking was also directly linked to adolescent smoking (β = 0.20, p < 0.05. These findings underscore a potentially important role of sensation seeking in the origins of adolescent smoking, and clarify pathways of influence with regard to maternal attitudes and behaviors in subsequent teenage nicotine use.
Speech and Communication Disorders
... to being completely unable to speak or understand speech. Causes include Hearing disorders and deafness Voice problems, ... or those caused by cleft lip or palate Speech problems like stuttering Developmental disabilities Learning disorders Autism ...
Yao, Bo; Belin, Pascal; Scheepers, Christoph
2012-04-15
In human communication, direct speech (e.g., Mary said, "I'm hungry") is perceived as more vivid than indirect speech (e.g., Mary said that she was hungry). This vividness distinction has previously been found to underlie silent reading of quotations: Using functional magnetic resonance imaging (fMRI), we found that direct speech elicited higher brain activity in the temporal voice areas (TVA) of the auditory cortex than indirect speech, consistent with an "inner voice" experience in reading direct speech. Here we show that listening to monotonously spoken direct versus indirect speech quotations also engenders differential TVA activity. This suggests that individuals engage in top-down simulations or imagery of enriched supra-segmental acoustic representations while listening to monotonous direct speech. The findings shed new light on the acoustic nature of the "inner voice" in understanding direct speech. Copyright © 2012 Elsevier Inc. All rights reserved.
The mechanism of speech processing in congenital amusia: evidence from Mandarin speakers.
Directory of Open Access Journals (Sweden)
Fang Liu
Full Text Available Congenital amusia is a neuro-developmental disorder of pitch perception that causes severe problems with music processing but only subtle difficulties in speech processing. This study investigated speech processing in a group of Mandarin speakers with congenital amusia. Thirteen Mandarin amusics and thirteen matched controls participated in a set of tone and intonation perception tasks and two pitch threshold tasks. Compared with controls, amusics showed impaired performance on word discrimination in natural speech and their gliding tone analogs. They also performed worse than controls on discriminating gliding tone sequences derived from statements and questions, and showed elevated thresholds for pitch change detection and pitch direction discrimination. However, they performed as well as controls on word identification, and on statement-question identification and discrimination in natural speech. Overall, tasks that involved multiple acoustic cues to communicative meaning were not impacted by amusia. Only when the tasks relied mainly on pitch sensitivity did amusics show impaired performance compared to controls. These findings help explain why amusia only affects speech processing in subtle ways. Further studies on a larger sample of Mandarin amusics and on amusics of other language backgrounds are needed to consolidate these results.
The mechanism of speech processing in congenital amusia: evidence from Mandarin speakers.
Liu, Fang; Jiang, Cunmei; Thompson, William Forde; Xu, Yi; Yang, Yufang; Stewart, Lauren
2012-01-01
Congenital amusia is a neuro-developmental disorder of pitch perception that causes severe problems with music processing but only subtle difficulties in speech processing. This study investigated speech processing in a group of Mandarin speakers with congenital amusia. Thirteen Mandarin amusics and thirteen matched controls participated in a set of tone and intonation perception tasks and two pitch threshold tasks. Compared with controls, amusics showed impaired performance on word discrimination in natural speech and their gliding tone analogs. They also performed worse than controls on discriminating gliding tone sequences derived from statements and questions, and showed elevated thresholds for pitch change detection and pitch direction discrimination. However, they performed as well as controls on word identification, and on statement-question identification and discrimination in natural speech. Overall, tasks that involved multiple acoustic cues to communicative meaning were not impacted by amusia. Only when the tasks relied mainly on pitch sensitivity did amusics show impaired performance compared to controls. These findings help explain why amusia only affects speech processing in subtle ways. Further studies on a larger sample of Mandarin amusics and on amusics of other language backgrounds are needed to consolidate these results.
Processing melodic contour and speech intonation in congenital amusics with Mandarin Chinese.
Jiang, Cunmei; Hamm, Jeff P; Lim, Vanessa K; Kirk, Ian J; Yang, Yufang
2010-07-01
Congenital amusia is a disorder in the perception and production of musical pitch. It has been suggested that early exposure to a tonal language may compensate for the pitch disorder (Peretz, 2008). If so, it is reasonable to expect that there would be different characterizations of pitch perception in music and speech in congenital amusics who speak a tonal language, such as Mandarin. In this study, a group of 11 adults with amusia whose first language was Mandarin were tested with melodic contour and speech intonation discrimination and identification tasks. The participants with amusia were impaired in discriminating and identifying melodic contour. These abnormalities were also detected in identifying both speech and non-linguistic analogue derived patterns for the Mandarin intonation tasks. In addition, there was an overall trend for the participants with amusia to show deficits with respect to controls in the intonation discrimination tasks for both speech and non-linguistic analogues. These findings suggest that the amusics' melodic pitch deficits may extend to the perception of speech, and could potentially result in some language deficits in those who speak a tonal language. Copyright (c) 2010 Elsevier Ltd. All rights reserved.
Speech of people with autism: Echolalia and echolalic speech
Błeszyński, Jacek Jarosław
2013-01-01
Speech of people with autism is recognised as one of the basic diagnostic, therapeutic and theoretical problems. One of the most common symptoms of autism in children is echolalia, described here as being of different types and severity. This paper presents the results of studies into different levels of echolalia, both in normally developing children and in children diagnosed with autism, discusses the differences between simple echolalia and echolalic speech - which can be considered to b...
Neuroanatomical correlates of childhood apraxia of speech: A connectomic approach
Directory of Open Access Journals (Sweden)
Simona Fiori
2016-01-01
These findings provide evidence of structural connectivity anomalies in children with CAS across specific brain regions involved in speech/language function. We propose altered connectivity as a possible epiphenomenon of complex pathogenic mechanisms in CAS which need further investigation.
A Diagnostic Marker to Discriminate Childhood Apraxia of Speech from Speech Delay: Introduction
Shriberg, Lawrence D.; Strand, Edythe A.; Fourakis, Marios; Jakielski, Kathy J.; Hall, Sheryl D.; Karlsson, Heather B.; Mabie, Heather L.; McSweeny, Jane L.; Tilkens, Christie M.; Wilson, David L.
2017-01-01
Purpose: The goal of this article is to introduce the pause marker (PM), a single-sign diagnostic marker proposed to discriminate early or persistent childhood apraxia of speech (CAS) from speech delay.
Accuracy of Cochlear Implant Recipients on Speech Reception in Background Music
Gfeller, Kate; Turner, Christopher; Oleson, Jacob; Kliethermes, Stephanie; Driscoll, Virginia
2012-01-01
Objectives This study (a) examined speech recognition abilities of cochlear implant (CI) recipients in the spectrally complex listening condition of three contrasting types of background music, and (b) compared performance based upon listener groups: CI recipients using conventional long-electrode (LE) devices, Hybrid CI recipients (acoustic plus electric stimulation), and normal-hearing (NH) adults. Methods We tested 154 LE CI recipients using varied devices and strategies, 21 Hybrid CI recipients, and 49 NH adults on closed-set recognition of spondees presented in three contrasting forms of background music (piano solo, large symphony orchestra, vocal solo with small combo accompaniment) in an adaptive test. Outcomes Signal-to-noise thresholds for speech in music (SRTM) were examined in relation to measures of speech recognition in background noise and multi-talker babble, pitch perception, and music experience. Results SRTM thresholds varied as a function of category of background music, group membership (LE, Hybrid, NH), and age. Thresholds for speech in background music were significantly correlated with measures of pitch perception and speech in background noise thresholds; auditory status was an important predictor. Conclusions Evidence suggests that speech reception thresholds in background music change as a function of listener age (with more advanced age being detrimental), structural characteristics of different types of music, and hearing status (residual hearing). These findings have implications for everyday listening conditions such as communicating in social or commercial situations in which there is background music. PMID:23342550
Maternal Methyl Supplemented Diets and Effects on Offspring Health
Directory of Open Access Journals (Sweden)
Rachel J. O'Neill
2014-08-01
Full Text Available Women seeking to become pregnant and pregnant women are currently advised to consume high amounts of folic acid and other methyl donors to prevent neural tube defects in their offspring. These diets can alter methylation patterns of several biomolecules, including nucleic acids and histone proteins. Limited animal model data suggests that developmental exposure to these maternal methyl supplemented (MS diets leads to beneficial epimutations. However, other rodent and humans studies have yielded opposing findings with such diets leading to promiscuous epimutations that are likely associated with negative health outcomes. Conflict exists to whether these maternal diets are preventative or exacerbate the risk for ASD in children. This review will discuss the findings to date on the potential beneficial and aversive effects of maternal MS diets. We will also consider how other factors might influence the effects of MS diets. Current data suggest that there is cause for concern as maternal MS diets may lead to epimutations that underpin various diseases, including neurobehavioral disorders. Further studies are needed to explore the comprehensive effects maternal MS diets have on the offspring epigenome and subsequent overall health.
Typical versus delayed speech onset influences verbal reporting of autistic interests.
Chiodo, Liliane; Majerus, Steve; Mottron, Laurent
2017-01-01
The distinction between autism and Asperger syndrome has been abandoned in the DSM-5. However, this clinical categorization largely overlaps with the presence or absence of a speech onset delay which is associated with clinical, cognitive, and neural differences. It is unknown whether these different speech development pathways and associated cognitive differences are involved in the heterogeneity of the restricted interests that characterize autistic adults. This study tested the hypothesis that speech onset delay, or conversely, early mastery of speech, orients the nature and verbal reporting of adult autistic interests. The occurrence of a priori defined descriptors for perceptual and thematic dimensions were determined, as well as the perceived function and benefits, in the response of autistic people to a semi-structured interview on their intense interests. The number of words, grammatical categories, and proportion of perceptual / thematic descriptors were computed and compared between groups by variance analyses. The participants comprised 40 autistic adults grouped according to the presence ( N = 20) or absence ( N = 20) of speech onset delay, as well as 20 non-autistic adults, also with intense interests, matched for non-verbal intelligence using Raven's Progressive Matrices. The overall nature, function, and benefit of intense interests were similar across autistic subgroups, and between autistic and non-autistic groups. However, autistic participants with a history of speech onset delay used more perceptual than thematic descriptors when talking about their interests, whereas the opposite was true for autistic individuals without speech onset delay. This finding remained significant after controlling for linguistic differences observed between the two groups. Verbal reporting, but not the nature or positive function, of intense interests differed between adult autistic individuals depending on their speech acquisition history: oral reporting of
Individual differences in language and working memory affect children’s speech recognition in noise
McCreery, Ryan W.; Spratford, Meredith; Kirby, Benjamin; Brennan, Marc
2017-01-01
Objective We examined how cognitive and linguistic skills affect speech recognition in noise for children with normal hearing. Children with better working memory and language abilities were expected to have better speech recognition in noise than peers with poorer skills in these domains. Design As part of a prospective, cross-sectional study, children with normal hearing completed speech recognition in noise for three types of stimuli: (1) monosyllabic words, (2) syntactically correct but semantically anomalous sentences and (3) semantically and syntactically anomalous word sequences. Measures of vocabulary, syntax and working memory were used to predict individual differences in speech recognition in noise. Study sample Ninety-six children with normal hearing, who were between 5 and 12 years of age. Results Higher working memory was associated with better speech recognition in noise for all three stimulus types. Higher vocabulary abilities were associated with better recognition in noise for sentences and word sequences, but not for words. Conclusions Working memory and language both influence children’s speech recognition in noise, but the relationships vary across types of stimuli. These findings suggest that clinical assessment of speech recognition is likely to reflect underlying cognitive and linguistic abilities, in addition to a child’s auditory skills, consistent with the Ease of Language Understanding model. PMID:27981855
Facilitated auditory detection for speech sounds
Directory of Open Access Journals (Sweden)
Carine eSignoret
2011-07-01
Full Text Available If it is well known that knowledge facilitates higher cognitive functions, such as visual and auditory word recognition, little is known about the influence of knowledge on detection, particularly in the auditory modality. Our study tested the influence of phonological and lexical knowledge on auditory detection. Words, pseudo words and complex non phonological sounds, energetically matched as closely as possible, were presented at a range of presentation levels from sub threshold to clearly audible. The participants performed a detection task (Experiments 1 and 2 that was followed by a two alternative forced choice recognition task in Experiment 2. The results of this second task in Experiment 2 suggest a correct recognition of words in the absence of detection with a subjective threshold approach. In the detection task of both experiments, phonological stimuli (words and pseudo words were better detected than non phonological stimuli (complex sounds, presented close to the auditory threshold. This finding suggests an advantage of speech for signal detection. An additional advantage of words over pseudo words was observed in Experiment 2, suggesting that lexical knowledge could also improve auditory detection when listeners had to recognize the stimulus in a subsequent task. Two simulations of detection performance performed on the sound signals confirmed that the advantage of speech over non speech processing could not be attributed to energetic differences in the stimuli.
DEFF Research Database (Denmark)
Olesen, Morten
In order to obtain articulatory analysis of speech production the model is improved. the standard model, as used in LPC analysis, to a large extent only models the acoustic properties of speech signal as opposed to articulatory modelling of the speech production. In spite of this the LPC model...... is by far the most widely used model in speech technology....
Shin, Yu-Jeong; Ko, Seung-O
2015-12-01
Velopharyngeal dysfunction in cleft palate patients following the primary palate repair may result in nasal air emission, hypernasality, articulation disorder and poor intelligibility of speech. Among conservative treatment methods, speech aid prosthesis combined with speech therapy is widely used method. However because of its long time of treatment more than a year and low predictability, some clinicians prefer a surgical intervention. Thus, the purpose of this report was to increase an attention on the effectiveness of speech aid prosthesis by introducing a case that was successfully treated. In this clinical report, speech bulb reduction program with intensive speech therapy was applied for a patient with velopharyngeal dysfunction and it was rapidly treated by 5months which was unusually short period for speech aid therapy. Furthermore, advantages of pre-operative speech aid therapy were discussed.
Gao, Yayue; Wang, Qian; Ding, Yu; Wang, Changming; Li, Haifeng; Wu, Xihong; Qu, Tianshu; Li, Liang
2017-01-01
Human listeners are able to selectively attend to target speech in a noisy environment with multiple-people talking. Using recordings of scalp electroencephalogram (EEG), this study investigated how selective attention facilitates the cortical representation of target speech under a simulated “cocktail-party” listening condition with speech-on-speech masking. The result shows that the cortical representation of target-speech signals under the multiple-people talking condition was specifically improved by selective attention relative to the non-selective-attention listening condition, and the beta-band activity was most strongly modulated by selective attention. Moreover, measured with the Granger Causality value, selective attention to the single target speech in the mixed-speech complex enhanced the following four causal connectivities for the beta-band oscillation: the ones (1) from site FT7 to the right motor area, (2) from the left frontal area to the right motor area, (3) from the central frontal area to the right motor area, and (4) from the central frontal area to the right frontal area. However, the selective-attention-induced change in beta-band causal connectivity from the central frontal area to the right motor area, but not other beta-band causal connectivities, was significantly correlated with the selective-attention-induced change in the cortical beta-band representation of target speech. These findings suggest that under the “cocktail-party” listening condition, the beta-band oscillation in EEGs to target speech is specifically facilitated by selective attention to the target speech that is embedded in the mixed-speech complex. The selective attention-induced unmasking of target speech may be associated with the improved beta-band functional connectivity from the central frontal area to the right motor area, suggesting a top-down attentional modulation of the speech-motor process. PMID:28239344
Speech Intelligibility Evaluation for Mobile Phones
DEFF Research Database (Denmark)
Jørgensen, Søren; Cubick, Jens; Dau, Torsten
2015-01-01
In the development process of modern telecommunication systems, such as mobile phones, it is common practice to use computer models to objectively evaluate the transmission quality of the system, instead of time-consuming perceptual listening tests. Such models have typically focused on the quality...... of the transmitted speech, while little or no attention has been provided to speech intelligibility. The present study investigated to what extent three state-of-the art speech intelligibility models could predict the intelligibility of noisy speech transmitted through mobile phones. Sentences from the Danish...... Dantale II speech material were mixed with three different kinds of background noise, transmitted through three different mobile phones, and recorded at the receiver via a local network simulator. The speech intelligibility of the transmitted sentences was assessed by six normal-hearing listeners...
Radiological evaluation of esophageal speech on total laryngectomee
International Nuclear Information System (INIS)
Chung, Tae Sub; Suh, Jung Ho; Kim, Dong Ik; Kim, Gwi Eon; Hong, Won Phy; Lee, Won Sang
1988-01-01
Total laryngectomee requires some form of alaryngeal speech for communication. Generally, esophageal speech is regarded as the most available and comfortable technique for alaryngeal speech. But esophageal speech is difficult to train, so many patients are unable to attain esophageal speech for communication. To understand mechanism of esophageal of esophageal speech on total laryngectomee, evaluation of anatomical change of the pharyngoesophageal segment is very important. We used video fluoroscopy for evaluation of pharyngesophageal segment during esophageal speech. Eighteen total laryngectomees were evaluated with video fluoroscopy from Dec. 1986 to May 1987 at Y.U.M.C. Our results were as follows: 1. Peseudoglottis is the most important factor for esophageal speech, which is visualized in 7 cases among 8 cases of excellent esophageal speech group. 2. Two cases of longer A-P diameter at the pseudoglottis have the best quality of esophageal speech than others. 3. Two cases of mucosal vibration at the pharyngoesophageal segment can make excellent esophageal speech. 4. The cases of failed esophageal speech are poor aerophagia in 6 cases, abscence of pseudoglottis in 4 cases and poor air ejection in 3 cases. 5. Aerophagia synchronizes with diaphragmatic motion in 8 cases of excellent esophageal speech.
Laffin, Jennifer J S; Raca, Gordana; Jackson, Craig A; Strand, Edythe A; Jakielski, Kathy J; Shriberg, Lawrence D
2012-11-01
The goal of this study was to identify new candidate genes and genomic copy-number variations associated with a rare, severe, and persistent speech disorder termed childhood apraxia of speech. Childhood apraxia of speech is the speech disorder segregating with a mutation in FOXP2 in a multigenerational London pedigree widely studied for its role in the development of speech-language in humans. A total of 24 participants who were suspected to have childhood apraxia of speech were assessed using a comprehensive protocol that samples speech in challenging contexts. All participants met clinical-research criteria for childhood apraxia of speech. Array comparative genomic hybridization analyses were completed using a customized 385K Nimblegen array (Roche Nimblegen, Madison, WI) with increased coverage of genes and regions previously associated with childhood apraxia of speech. A total of 16 copy-number variations with potential consequences for speech-language development were detected in 12 or half of the 24 participants. The copy-number variations occurred on 10 chromosomes, 3 of which had two to four candidate regions. Several participants were identified with copy-number variations in two to three regions. In addition, one participant had a heterozygous FOXP2 mutation and a copy-number variation on chromosome 2, and one participant had a 16p11.2 microdeletion and copy-number variations on chromosomes 13 and 14. Findings support the likelihood of heterogeneous genomic pathways associated with childhood apraxia of speech.
Directory of Open Access Journals (Sweden)
Andreas Maier
2010-01-01
Full Text Available In patients suffering from head and neck cancer, speech intelligibility is often restricted. For assessment and outcome measurements, automatic speech recognition systems have previously been shown to be appropriate for objective and quick evaluation of intelligibility. In this study we investigate the applicability of the method to speech disorders caused by head and neck cancer. Intelligibility was quantified by speech recognition on recordings of a standard text read by 41 German laryngectomized patients with cancer of the larynx or hypopharynx and 49 German patients who had suffered from oral cancer. The speech recognition provides the percentage of correctly recognized words of a sequence, that is, the word recognition rate. Automatic evaluation was compared to perceptual ratings by a panel of experts and to an age-matched control group. Both patient groups showed significantly lower word recognition rates than the control group. Automatic speech recognition yielded word recognition rates which complied with experts' evaluation of intelligibility on a significant level. Automatic speech recognition serves as a good means with low effort to objectify and quantify the most important aspect of pathologic speech—the intelligibility. The system was successfully applied to voice and speech disorders.
Egami, Sonoko
2007-06-01
This study examined the impact of mothers' adherence to "maternal love" on maternal emotional expression toward their children. It was postulated that adherence to "maternal love" (defined as the tendency to accept and obey blindly the traditional maternal role and sociocultural belief in "desirable mothers") would have both positive and negative effects on maternal emotional expression, depending on the mothers' occupational status and satisfaction in workplace. The results showed an interaction between mothers' adherence to "maternal love" and the mothers' satisfaction in the workplace, which affected their expression of emotion. When satisfaction in the workplace was rated in the middle, it was positively associated with positive emotional expression. When satisfaction in the workplace was rated as high, it was both positively and negatively associated with positive emotional expression for full-time workers. Moreover, when satisfaction in the workplace was rated as in the middle, it was negatively associated with negative emotional expression, and when satisfaction in the workplace was rated as low or high, it was positively associated with negative emotional expression for all workers. These findings confirmed that mothers' adherence to "maternal love" is "the double-edged sword".
On speech recognition during anaesthesia
DEFF Research Database (Denmark)
Alapetite, Alexandre
2007-01-01
This PhD thesis in human-computer interfaces (informatics) studies the case of the anaesthesia record used during medical operations and the possibility to supplement it with speech recognition facilities. Problems and limitations have been identified with the traditional paper-based anaesthesia...... and inaccuracies in the anaesthesia record. Supplementing the electronic anaesthesia record interface with speech input facilities is proposed as one possible solution to a part of the problem. The testing of the various hypotheses has involved the development of a prototype of an electronic anaesthesia record...... interface with speech input facilities in Danish. The evaluation of the new interface was carried out in a full-scale anaesthesia simulator. This has been complemented by laboratory experiments on several aspects of speech recognition for this type of use, e.g. the effects of noise on speech recognition...
Directory of Open Access Journals (Sweden)
Renata Aparecida Leite
2014-03-01
Full Text Available OBJECTIVES: This study investigated whether neurophysiologic responses (auditory evoked potentials differ between typically developed children and children with phonological disorders and whether these responses are modified in children with phonological disorders after speech therapy. METHODS: The participants included 24 typically developing children (Control Group, mean age: eight years and ten months and 23 children clinically diagnosed with phonological disorders (Study Group, mean age: eight years and eleven months. Additionally, 12 study group children were enrolled in speech therapy (Study Group 1, and 11 were not enrolled in speech therapy (Study Group 2. The subjects were submitted to the following procedures: conventional audiological, auditory brainstem response, auditory middle-latency response, and P300 assessments. All participants presented with normal hearing thresholds. The study group 1 subjects were reassessed after 12 speech therapy sessions, and the study group 2 subjects were reassessed 3 months after the initial assessment. Electrophysiological results were compared between the groups. RESULTS: Latency differences were observed between the groups (the control and study groups regarding the auditory brainstem response and the P300 tests. Additionally, the P300 responses improved in the study group 1 children after speech therapy. CONCLUSION: The findings suggest that children with phonological disorders have impaired auditory brainstem and cortical region pathways that may benefit from speech therapy.
2013-10-23
...] Telecommunications Relay Services and Speech-to-Speech Services for Individuals With Hearing and Speech Disabilities... for telecommunications relay services (TRS) by eliminating standards for Internet-based relay services... comments, identified by CG Docket No. 03-123, by any of the following methods: Electronic Filers: Comments...
Sandor, A.; Moses, H. R.
2016-01-01
Currently on the International Space Station (ISS) and other space vehicles Caution & Warning (C&W) alerts are represented with various auditory tones that correspond to the type of event. This system relies on the crew's ability to remember what each tone represents in a high stress, high workload environment when responding to the alert. Furthermore, crew receive a year or more in advance of the mission that makes remembering the semantic meaning of the alerts more difficult. The current system works for missions conducted close to Earth where ground operators can assist as needed. On long duration missions, however, they will need to work off-nominal events autonomously. There is evidence that speech alarms may be easier and faster to recognize, especially during an off-nominal event. The Information Presentation Directed Research Project (FY07-FY09) funded by the Human Research Program included several studies investigating C&W alerts. The studies evaluated tone alerts currently in use with NASA flight deck displays along with candidate speech alerts. A follow-on study used four types of speech alerts to investigate how quickly various types of auditory alerts with and without a speech component - either at the beginning or at the end of the tone - can be identified. Even though crew were familiar with the tone alert from training or direct mission experience, alerts starting with a speech component were identified faster than alerts starting with a tone. The current study replicated the results from the previous study in a more rigorous experimental design to determine if the candidate speech alarms are ready for transition to operations or if more research is needed. Four types of alarms (caution, warning, fire, and depressurization) were presented to participants in both tone and speech formats in laboratory settings and later in the Human Exploration Research Analog (HERA). In the laboratory study, the alerts were presented by software and participants were
Stacey, Tomasina; Thompson, John M D; Mitchell, Edwin A; Ekeroma, Alec; Zuccollo, Jane; McCowan, Lesley M E
2011-12-01
Maternal perception of decreased fetal movements has been associated with adverse pregnancy outcomes, including stillbirth. Little is known about other aspects of perceived fetal activity. The objective of this study was to explore the relationship between maternal perception of fetal activity and late stillbirth (≥28 wk gestation) risk. Participants were women with a singleton, late stillbirth without congenital abnormality, born between July 2006 and June 2009 in Auckland, New Zealand. Two control women with ongoing pregnancies were randomly selected at the same gestation at which the stillbirth occurred. Detailed demographic and fetal movement data were collected by way of interview in the first few weeks after the stillbirth, or at the equivalent gestation for control women. A total of 155/215 (72%) women who experienced a stillbirth and 310/429 (72%) control group women consented to participate in the study. Maternal perception of increased strength and frequency of fetal movements, fetal hiccups, and frequent vigorous fetal activity were all associated with a reduced risk of late stillbirth. In contrast, perception of decreased strength of fetal movement was associated with a more than twofold increased risk of late stillbirth (aOR: 2.37; 95% CI: 1.29-4.35). A single episode of vigorous fetal activity was associated with an almost sevenfold increase in late stillbirth risk (aOR: 6.81; 95% CI: 3.01-15.41) compared with no unusually vigorous activity. Our study suggests that maternal perception of increasing fetal activity throughout the last 3 months of pregnancy is a sign of fetal well-being, whereas perception of reduced fetal movements is associated with increased risk of late stillbirth. © 2011, Copyright the Authors. Journal compilation © 2011, Wiley Periodicals, Inc.
Common cues to emotion in the dynamic facial expressions of speech and song.
Livingstone, Steven R; Thompson, William F; Wanderley, Marcelo M; Palmer, Caroline
2015-01-01
Speech and song are universal forms of vocalization that may share aspects of emotional expression. Research has focused on parallels in acoustic features, overlooking facial cues to emotion. In three experiments, we compared moving facial expressions in speech and song. In Experiment 1, vocalists spoke and sang statements each with five emotions. Vocalists exhibited emotion-dependent movements of the eyebrows and lip corners that transcended speech-song differences. Vocalists' jaw movements were coupled to their acoustic intensity, exhibiting differences across emotion and speech-song. Vocalists' emotional movements extended beyond vocal sound to include large sustained expressions, suggesting a communicative function. In Experiment 2, viewers judged silent videos of vocalists' facial expressions prior to, during, and following vocalization. Emotional intentions were identified accurately for movements during and after vocalization, suggesting that these movements support the acoustic message. Experiment 3 compared emotional identification in voice-only, face-only, and face-and-voice recordings. Emotion judgements for voice-only singing were poorly identified, yet were accurate for all other conditions, confirming that facial expressions conveyed emotion more accurately than the voice in song, yet were equivalent in speech. Collectively, these findings highlight broad commonalities in the facial cues to emotion in speech and song, yet highlight differences in perception and acoustic-motor production.
Visualizing structures of speech expressiveness
DEFF Research Database (Denmark)
Herbelin, Bruno; Jensen, Karl Kristoffer; Graugaard, Lars
2008-01-01
Speech is both beautiful and informative. In this work, a conceptual study of the speech, through investigation of the tower of Babel, the archetypal phonemes, and a study of the reasons of uses of language is undertaken in order to create an artistic work investigating the nature of speech. The ....... The artwork is presented at the Re:New festival in May 2008....
A Clinician Survey of Speech and Non-Speech Characteristics of Neurogenic Stuttering
Theys, Catherine; van Wieringen, Astrid; De Nil, Luc F.
2008-01-01
This study presents survey data on 58 Dutch-speaking patients with neurogenic stuttering following various neurological injuries. Stroke was the most prevalent cause of stuttering in our patients, followed by traumatic brain injury, neurodegenerative diseases, and other causes. Speech and non-speech characteristics were analyzed separately for…
Directory of Open Access Journals (Sweden)
William Andrew Dunlop
2016-08-01
Full Text Available Autism Spectrum Disorder (ASD, characterised by impaired communication skills and repetitive behaviours, can also result in differences in sensory perception. Individuals with ASD often perform normally in simple auditory tasks but poorly compared to typically developed (TD individuals on complex auditory tasks like discriminating speech from complex background noise. A common trait of individuals with ASD is hypersensitivity to auditory stimulation. No studies to our knowledge consider whether hypersensitivity to sounds is related to differences in speech-in-noise discrimination. We provide novel evidence that individuals with high-functioning ASD show poor performance compared to TD individuals in a speech-in-noise discrimination task with an attentionally demanding background noise, but not in a purely energetic noise. Further, we demonstrate in our small sample that speech-hypersensitivity does not appear to predict performance in the speech-in-noise task. The findings support the argument that an attentional deficit, rather than a perceptual deficit, affects the ability of individuals with ASD to discriminate speech from background noise. Finally, we piloted a novel questionnaire that measures difficulty hearing in noisy environments, and sensitivity to non-verbal and verbal sounds. Psychometric analysis using 128 TD participants provided novel evidence for a difference in sensitivity to non-verbal and verbal sounds, and these findings were reinforced by participants with ASD who also completed the questionnaire. The study was limited by a small and high-functioning sample of participants with ASD. Future work could test larger sample sizes and include lower-functioning ASD participants.
Automatic Speech Signal Analysis for Clinical Diagnosis and Assessment of Speech Disorders
Baghai-Ravary, Ladan
2013-01-01
Automatic Speech Signal Analysis for Clinical Diagnosis and Assessment of Speech Disorders provides a survey of methods designed to aid clinicians in the diagnosis and monitoring of speech disorders such as dysarthria and dyspraxia, with an emphasis on the signal processing techniques, statistical validity of the results presented in the literature, and the appropriateness of methods that do not require specialized equipment, rigorously controlled recording procedures or highly skilled personnel to interpret results. Such techniques offer the promise of a simple and cost-effective, yet objective, assessment of a range of medical conditions, which would be of great value to clinicians. The ideal scenario would begin with the collection of examples of the clients’ speech, either over the phone or using portable recording devices operated by non-specialist nursing staff. The recordings could then be analyzed initially to aid diagnosis of conditions, and subsequently to monitor the clients’ progress and res...
Temporal modulations in speech and music.
Ding, Nai; Patel, Aniruddh D; Chen, Lin; Butler, Henry; Luo, Cheng; Poeppel, David
2017-10-01
Speech and music have structured rhythms. Here we discuss a major acoustic correlate of spoken and musical rhythms, the slow (0.25-32Hz) temporal modulations in sound intensity and compare the modulation properties of speech and music. We analyze these modulations using over 25h of speech and over 39h of recordings of Western music. We show that the speech modulation spectrum is highly consistent across 9 languages (including languages with typologically different rhythmic characteristics). A different, but similarly consistent modulation spectrum is observed for music, including classical music played by single instruments of different types, symphonic, jazz, and rock. The temporal modulations of speech and music show broad but well-separated peaks around 5 and 2Hz, respectively. These acoustically dominant time scales may be intrinsic features of speech and music, a possibility which should be investigated using more culturally diverse samples in each domain. Distinct modulation timescales for speech and music could facilitate their perceptual analysis and its neural processing. Copyright © 2017 Elsevier Ltd. All rights reserved.
The Effect of Maternal Employment on Children's Academic Performance
Rachel Dunifon; Anne Toft Hansen; Sean Nicholson; Lisbeth Palmhøj Nielsen
2013-01-01
Using a Danish data set that follows 135,000 Danish children from birth through 9th grade, we examine the effect of maternal employment during a child's first three and first 15 years on that child's grade point average in 9th grade. We address the endogeneity of employment by including a rich set of household control variables, instrumenting for employment with the gender- and education-specific local unemployment rate, and by including maternal fixed effects. We find that maternal employmen...
Fetal MRI: incidental findings in the mother
Energy Technology Data Exchange (ETDEWEB)
Abdullah, Selwan B. [University of Maryland Medical Center, Diagnostic Radiology and Nuclear Medicine, Baltimore, MD (United States); University of Minnesota, Medical School, Minneapolis, MN (United States); Dietz, Kelly R.; Holm, Tara L. [University of Minnesota, Department of Radiology, Minneapolis, MN (United States)
2016-11-15
Fetal magnetic resonance imaging (MRI) is a routinely used tool in prenatal diagnosis; however, there is a lack of studies evaluating incidental findings observed in the mother. This study describes and quantifies incidental findings observed in the mother during fetal MRI. We reviewed all fetal MRI studies at the University of Minnesota Medical Center from February 2008 to September 2014. Two pediatric radiologists retrospectively conducted a consensus evaluation. The maternal findings were categorized into neurologic, gynecologic, urinary, gastrointestinal and musculoskeletal. Hydronephrosis consistent with the stage of pregnancy was recorded but was not included as an abnormal finding. Abnormal findings were classified into three groups, depending on their clinical significance: level I (low), level II (medium) and level III (high). We evaluated 332 pregnant patients with a mean age of 29.3 years and a mean gestational age of 29 weeks. Of these, 55.4% had at least 1 incidental finding, for a total of 262 incidental maternal findings. Of the 262 abnormalities, 113 (43.1%) were neurologic, 69 were gynecologic (26.3%), 36 (13.7%) urinary, 24 (9.2%) gastrointestinal and 20 (7.6%) musculoskeletal. Of the 262 incidental findings, 237 (90.5%) were level I, 24 (9.2%) were level II and 1 (0.4%) was level III. Our results suggest that although the vast majority of incidental maternal findings are benign, more significant findings are still encountered and should be expected. (orig.)
Fetal MRI: incidental findings in the mother
International Nuclear Information System (INIS)
Abdullah, Selwan B.; Dietz, Kelly R.; Holm, Tara L.
2016-01-01
Fetal magnetic resonance imaging (MRI) is a routinely used tool in prenatal diagnosis; however, there is a lack of studies evaluating incidental findings observed in the mother. This study describes and quantifies incidental findings observed in the mother during fetal MRI. We reviewed all fetal MRI studies at the University of Minnesota Medical Center from February 2008 to September 2014. Two pediatric radiologists retrospectively conducted a consensus evaluation. The maternal findings were categorized into neurologic, gynecologic, urinary, gastrointestinal and musculoskeletal. Hydronephrosis consistent with the stage of pregnancy was recorded but was not included as an abnormal finding. Abnormal findings were classified into three groups, depending on their clinical significance: level I (low), level II (medium) and level III (high). We evaluated 332 pregnant patients with a mean age of 29.3 years and a mean gestational age of 29 weeks. Of these, 55.4% had at least 1 incidental finding, for a total of 262 incidental maternal findings. Of the 262 abnormalities, 113 (43.1%) were neurologic, 69 were gynecologic (26.3%), 36 (13.7%) urinary, 24 (9.2%) gastrointestinal and 20 (7.6%) musculoskeletal. Of the 262 incidental findings, 237 (90.5%) were level I, 24 (9.2%) were level II and 1 (0.4%) was level III. Our results suggest that although the vast majority of incidental maternal findings are benign, more significant findings are still encountered and should be expected. (orig.)
Acoustic richness modulates the neural networks supporting intelligible speech processing.
Lee, Yune-Sang; Min, Nam Eun; Wingfield, Arthur; Grossman, Murray; Peelle, Jonathan E
2016-03-01
The information contained in a sensory signal plays a critical role in determining what neural processes are engaged. Here we used interleaved silent steady-state (ISSS) functional magnetic resonance imaging (fMRI) to explore how human listeners cope with different degrees of acoustic richness during auditory sentence comprehension. Twenty-six healthy young adults underwent scanning while hearing sentences that varied in acoustic richness (high vs. low spectral detail) and syntactic complexity (subject-relative vs. object-relative center-embedded clause structures). We manipulated acoustic richness by presenting the stimuli as unprocessed full-spectrum speech, or noise-vocoded with 24 channels. Importantly, although the vocoded sentences were spectrally impoverished, all sentences were highly intelligible. These manipulations allowed us to test how intelligible speech processing was affected by orthogonal linguistic and acoustic demands. Acoustically rich speech showed stronger activation than acoustically less-detailed speech in a bilateral temporoparietal network with more pronounced activity in the right hemisphere. By contrast, listening to sentences with greater syntactic complexity resulted in increased activation of a left-lateralized network including left posterior lateral temporal cortex, left inferior frontal gyrus, and left dorsolateral prefrontal cortex. Significant interactions between acoustic richness and syntactic complexity occurred in left supramarginal gyrus, right superior temporal gyrus, and right inferior frontal gyrus, indicating that the regions recruited for syntactic challenge differed as a function of acoustic properties of the speech. Our findings suggest that the neural systems involved in speech perception are finely tuned to the type of information available, and that reducing the richness of the acoustic signal dramatically alters the brain's response to spoken language, even when intelligibility is high. Copyright © 2015 Elsevier
Sensorimotor speech disorders in Parkinson's disease: Programming and execution deficits
Directory of Open Access Journals (Sweden)
Karin Zazo Ortiz
Full Text Available ABSTRACT Introduction: Dysfunction in the basal ganglia circuits is a determining factor in the physiopathology of the classic signs of Parkinson's disease (PD and hypokinetic dysarthria is commonly related to PD. Regarding speech disorders associated with PD, the latest four-level framework of speech complicates the traditional view of dysarthria as a motor execution disorder. Based on findings that dysfunctions in basal ganglia can cause speech disorders, and on the premise that the speech deficits seen in PD are not related to an execution motor disorder alone but also to a disorder at the motor programming level, the main objective of this study was to investigate the presence of sensorimotor disorders of programming (besides the execution disorders previously described in PD patients. Methods: A cross-sectional study was conducted in a sample of 60 adults matched for gender, age and education: 30 adult patients diagnosed with idiopathic PD (PDG and 30 healthy adults (CG. All types of articulation errors were reanalyzed to investigate the nature of these errors. Interjections, hesitations and repetitions of words or sentences (during discourse were considered typical disfluencies; blocking, episodes of palilalia (words or syllables were analyzed as atypical disfluencies. We analysed features including successive self-initiated trial, phoneme distortions, self-correction, repetition of sounds and syllables, prolonged movement transitions, additions or omissions of sounds and syllables, in order to identify programming and/or execution failures. Orofacial agility was also investigated. Results: The PDG had worse performance on all sensorimotor speech tasks. All PD patients had hypokinetic dysarthria. Conclusion: The clinical characteristics found suggest both execution and programming sensorimotor speech disorders in PD patients.
Language and motor speech skills in children with cerebral palsy
Pirila, Sija; van der Meere, Jaap; Pentikainen, Taina; Ruusu-Niemi, Pirjo; Korpela, Raija; Kilpinen, Jenni; Nieminen, Pirkko; Ruusu-Niemin, P; Kilpinen, R
2007-01-01
The aim of the study was to investigate associations between the severity of motor limitations, cognitive difficulties, language and motor speech problems in children with cerebral palsy. Also, the predictive power of neonatal cranial ultrasound findings on later outcome was investigated. For this
Song and speech: examining the link between singing talent and speech imitation ability
Christiner, Markus; Reiterer, Susanne M.
2013-01-01
In previous research on speech imitation, musicality, and an ability to sing were isolated as the strongest indicators of good pronunciation skills in foreign languages. We, therefore, wanted to take a closer look at the nature of the ability to sing, which shares a common ground with the ability to imitate speech. This study focuses on whether good singing performance predicts good speech imitation. Forty-one singers of different levels of proficiency were selected for the study and their ability to sing, to imitate speech, their musical talent and working memory were tested. Results indicated that singing performance is a better indicator of the ability to imitate speech than the playing of a musical instrument. A multiple regression revealed that 64% of the speech imitation score variance could be explained by working memory together with educational background and singing performance. A second multiple regression showed that 66% of the speech imitation variance of completely unintelligible and unfamiliar language stimuli (Hindi) could be explained by working memory together with a singer's sense of rhythm and quality of voice. This supports the idea that both vocal behaviors have a common grounding in terms of vocal and motor flexibility, ontogenetic and phylogenetic development, neural orchestration and auditory memory with singing fitting better into the category of “speech” on the productive level and “music” on the acoustic level. As a result, good singers benefit from vocal and motor flexibility, productively and cognitively, in three ways. (1) Motor flexibility and the ability to sing improve language and musical function. (2) Good singers retain a certain plasticity and are open to new and unusual sound combinations during adulthood both perceptually and productively. (3) The ability to sing improves the memory span of the auditory working memory. PMID:24319438
Coppens-Hofman, Marjolein C; Terband, Hayo R; Maassen, Ben A M; van Schrojenstein Lantman-De Valk, Henny M J; van Zaalen-op't Hof, Yvonne; Snik, Ad F M
2013-01-01
In individuals with an intellectual disability, speech dysfluencies are more common than in the general population. In clinical practice, these fluency disorders are generally diagnosed and treated as stuttering rather than cluttering. To characterise the type of dysfluencies in adults with intellectual disabilities and reported speech difficulties with an emphasis on manifestations of stuttering and cluttering, which distinction is to help optimise treatment aimed at improving fluency and intelligibility. The dysfluencies in the spontaneous speech of 28 adults (18-40 years; 16 men) with mild and moderate intellectual disabilities (IQs 40-70), who were characterised as poorly intelligible by their caregivers, were analysed using the speech norms for typically developing adults and children. The speakers were subsequently assigned to different diagnostic categories by relating their resulting dysfluency profiles to mean articulatory rate and articulatory rate variability. Twenty-two (75%) of the participants showed clinically significant dysfluencies, of which 21% were classified as cluttering, 29% as cluttering-stuttering and 25% as clear cluttering at normal articulatory rate. The characteristic pattern of stuttering did not occur. The dysfluencies in the speech of adults with intellectual disabilities and poor intelligibility show patterns that are specific for this population. Together, the results suggest that in this specific group of dysfluent speakers interventions should be aimed at cluttering rather than stuttering. The reader will be able to (1) describe patterns of dysfluencies in the speech of adults with intellectual disabilities that are specific for this group of people, (2) explain that a high rate of dysfluencies in speech is potentially a major determiner of poor intelligibility in adults with ID and (3) describe suggestions for intervention focusing on cluttering rather than stuttering in dysfluent speakers with ID. Copyright © 2013 Elsevier Inc
Directory of Open Access Journals (Sweden)
Hoang Van Minh
2016-02-01
Full Text Available Background: Knowledge of the aggregate effects of multiple socioeconomic vulnerabilities is important for shedding light on the determinants of growing health inequalities and inequities in maternal healthcare. Objective: This paper describes patterns of inequity in maternal healthcare utilization and analyzes associations between inequity and multiple socioeconomic vulnerabilities among women in Vietnam. Design: This is a repeated cross-sectional study using data from the Vietnam Multiple Indicator Cluster Surveys 2000, 2006, and 2011. Two maternal healthcare indicators were selected: (1 skilled antenatal care and (2 skilled delivery care. Four types of socioeconomic vulnerabilities – low education, ethnic minority, poverty, and rural location – were assessed both as separate explanatory variables and as composite indicators (combinations of three and four vulnerabilities. Pairwise comparisons and adjusted odds ratios were used to assess socioeconomic inequities in maternal healthcare. Results: In all three surveys, there were increases across the survey years in both the proportions of women who received antenatal care by skilled staff (68.6% in 2000, 90.8% in 2006, and 93.7% in 2011 and the proportions of women who gave birth with assistance from skilled staff (69.9% in 2000, 87.7% in 2006, and 92.9% in 2011. The receipt of antenatal care by skilled staff and birth assistance from skilled health personnel were less common among vulnerable women, especially those with multiple vulnerabilities. Conclusions: Even though Vietnam has improved its coverage of maternal healthcare on average, policies should target maternal healthcare utilization among women with multiple socioeconomic vulnerabilities. Both multisectoral social policies and health policies are needed to tackle multiple vulnerabilities more effectively by identifying those who are poor, less educated, live in rural areas, and belong to ethnic minority groups.
Shriberg, Lawrence D.; Strand, Edythe A.; Fourakis, Marios; Jakielski, Kathy J.; Hall, Sheryl D.; Karlsson, Heather B.; Mabie, Heather L.; McSweeny, Jane L.; Tilkens, Christie M.; Wilson, David L.
2017-01-01
Purpose: Previous articles in this supplement described rationale for and development of the pause marker (PM), a diagnostic marker of childhood apraxia of speech (CAS), and studies supporting its validity and reliability. The present article assesses the theoretical coherence of the PM with speech processing deficits in CAS. Method: PM and other…
Moharir, Madhavi; Barnett, Noel; Taras, Jillian; Cole, Martha; Ford-Jones, E Lee; Levin, Leo
2014-01-01
Failure to recognize and intervene early in speech and language delays can lead to multifaceted and potentially severe consequences for early child development and later literacy skills. While routine evaluations of speech and language during well-child visits are recommended, there is no standardized (office) approach to facilitate this. Furthermore, extensive wait times for speech and language pathology consultation represent valuable lost time for the child and family. Using speech and language expertise, and paediatric collaboration, key content for an office-based tool was developed. early and accurate identification of speech and language delays as well as children at risk for literacy challenges; appropriate referral to speech and language services when required; and teaching and, thus, empowering parents to create rich and responsive language environments at home. Using this tool, in combination with the Canadian Paediatric Society's Read, Speak, Sing and Grow Literacy Initiative, physicians will be better positioned to offer practical strategies to caregivers to enhance children's speech and language capabilities. The tool represents a strategy to evaluate speech and language delays. It depicts age-specific linguistic/phonetic milestones and suggests interventions. The tool represents a practical interim treatment while the family is waiting for formal speech and language therapy consultation.
Abortion and compelled physician speech.
Orentlicher, David
2015-01-01
Informed consent mandates for abortion providers may infringe the First Amendment's freedom of speech. On the other hand, they may reinforce the physician's duty to obtain informed consent. Courts can promote both doctrines by ensuring that compelled physician speech pertains to medical facts about abortion rather than abortion ideology and that compelled speech is truthful and not misleading. © 2015 American Society of Law, Medicine & Ethics, Inc.
Directory of Open Access Journals (Sweden)
McNeel Gordon Jantzen
2014-03-01
Full Text Available Musicians have a more accurate temporal and tonal representation of auditory stimuli than their non-musician counterparts (Kraus & Chandrasekaran, 2010; Parbery-Clark, Skoe, & Kraus, 2009; Zendel & Alain, 2008; Musacchia, Sams, Skoe, & Kraus, 2007. Musicians who are adept at the production and perception of music are also more sensitive to key acoustic features of speech such as voice onset timing and pitch. Together, these data suggest that musical training may enhance the processing of acoustic information for speech sounds. In the current study, we sought to provide neural evidence that musicians process speech and music in a similar way. We hypothesized that for musicians, right hemisphere areas traditionally associated with music are also engaged for the processing of speech sounds. In contrast we predicted that in non-musicians processing of speech sounds would be localized to traditional left hemisphere language areas. Speech stimuli differing in voice onset time was presented using a dichotic listening paradigm. Subjects either indicated aural location for a specified speech sound or identified a specific speech sound from a directed aural location. Musical training effects and organization of acoustic features were reflected by activity in source generators of the P50. This included greater activation of right middle temporal gyrus (MTG and superior temporal gyrus (STG in musicians. The findings demonstrate recruitment of right hemisphere in musicians for discriminating speech sounds and a putative broadening of their language network. Musicians appear to have an increased sensitivity to acoustic features and enhanced selective attention to temporal features of speech that is facilitated by musical training and supported, in part, by right hemisphere homologues of established speech processing regions of the brain.
Ghaedrahmat, Mahdi; Alavi Nia, Parviz; Biria, Reza
2016-01-01
This pragmatic study investigated the speech act of thanking as used by non-native speakers of English. The study was an attempt to find whether the pragmatic awareness of Iranian EFL learners could be improved through explicit instruction of the structure of the speech act of "Thanking". In fact, this study aimed to find out if there…
Benesty, Jacob; Chen, Jingdong
2006-01-01
We live in a noisy world! In all applications (telecommunications, hands-free communications, recording, human-machine interfaces, etc.) that require at least one microphone, the signal of interest is usually contaminated by noise and reverberation. As a result, the microphone signal has to be ""cleaned"" with digital signal processing tools before it is played out, transmitted, or stored.This book is about speech enhancement. Different well-known and state-of-the-art methods for noise reduction, with one or multiple microphones, are discussed. By speech enhancement, we mean not only noise red
Association of maternal diabetes and child asthma.
Azad, Meghan B; Becker, Allan B; Kozyrskyj, Anita L
2013-06-01
Perinatal programming is an emerging theory for the fetal origins of chronic disease. Maternal asthma and environmental tobacco smoke (ETS) are two of the best-known triggers for the perinatal programming of asthma, while the potential role of maternal diabetes has not been widely studied. To determine if maternal diabetes is associated with child asthma, and if so, whether it modifies the effects of ETS exposure and maternal asthma. We studied 3,574 Canadian children, aged 7-8 years, enrolled in a population-based birth cohort. Standardized questionnaires were completed by the children's parents, and data were analyzed by multivariable logistic regression. Asthma was reported in 442 children (12.4%). Compared to those without asthma, asthmatic children were more likely to have mothers (P = 0.003), but not fathers (P = 0.89), with diabetes. Among children without maternal history of diabetes, the likelihood of child asthma was 1.4-fold higher in those exposed to ETS (adjusted odds ratio, 1.40; 95% confidence interval, 1.13-1.73), and 3.6-fold higher in those with maternal asthma (3.59; 2.71-4.76). Among children born to diabetic mothers, these risks were amplified to 5.7-fold (5.68; 1.18-27.37) and 11.3-fold (11.30; 2.26-56.38), respectively. In the absence of maternal asthma or ETS, maternal diabetes was not associated with child asthma (0.65, 0.16-2.56). Our findings suggest that maternal diabetes may contribute to the perinatal programming of child asthma by amplifying the detrimental effects of ETS exposure and maternal asthma. Copyright © 2012 Wiley Periodicals, Inc.
Effect of speech rate variation on acoustic phone stability in Afrikaans speech recognition
CSIR Research Space (South Africa)
Badenhorst, JAC
2007-11-01
Full Text Available The authors analyse the effect of speech rate variation on Afrikaans phone stability from an acoustic perspective. Specifically they introduce two techniques for the acoustic analysis of speech rate variation, apply these techniques to an Afrikaans...
Ellis, Rachel J; Rönnberg, Jerker
2015-01-01
Proactive interference (PI) is the capacity to resist interference to the acquisition of new memories from information stored in the long-term memory. Previous research has shown that PI correlates significantly with the speech-in-noise recognition scores of younger adults with normal hearing. In this study, we report the results of an experiment designed to investigate the extent to which tests of visual PI relate to the speech-in-noise recognition scores of older adults with hearing loss, in aided and unaided conditions. The results suggest that measures of PI correlate significantly with speech-in-noise recognition only in the unaided condition. Furthermore the relation between PI and speech-in-noise recognition differs to that observed in younger listeners without hearing loss. The findings suggest that the relation between PI tests and the speech-in-noise recognition scores of older adults with hearing loss relates to capability of the test to index cognitive flexibility.
Directory of Open Access Journals (Sweden)
Rachel Jane Ellis
2015-08-01
Full Text Available Proactive interference (PI is the capacity to resist interference to the acquisition of new memories from information stored in the long-term memory. Previous research has shown that PI correlates significantly with the speech-in-noise recognition scores of younger adults with normal hearing. In this study, we report the results of an experiment designed to investigate the extent to which tests of visual PI relate to the speech-in-noise recognition scores of older adults with hearing loss, in aided and unaided conditions. The results suggest that measures of PI correlate significantly with speech-in-noise recognition only in the unaided condition. Furthermore the relation between PI and speech-in-noise recognition differs to that observed in younger listeners without hearing loss. The findings suggest that the relation between PI tests and the speech-in-noise recognition scores of older adults with hearing loss relates to capability of the test to index cognitive flexibility.
Hitch, Graham J.; And Others
1991-01-01
Reports on experiments to determine effects of overt speech on children's use of inner speech in short-term memory. Word length and phonemic similarity had greater effects on older children and when pictures were labeled at presentation. Suggests that speaking or listening to speech activates an internal articulatory loop. (Author/GH)
A CLINICAL STUDY OF MATERNAL DEATHS DUE TO PPH
Directory of Open Access Journals (Sweden)
Basavana Gowda
2015-03-01
Full Text Available OBJECTIVES: A study of maternal death conducted to evaluate various factors responsible for maternal deaths. To identify complications in pregnancy, a childbirth which result in maternal death, and to identify opportunities for preventive intervention and understand the events leading to death; so that improving maternal health and reducing maternal mortality rate significantly. To analyze the causes and epidemiological amounts maternal mortality e.g. age parity, socioeconomic status and literacy. In order to reduce maternal mortality and to implement safe motherhood program and complications of pregnancy and to find out safe motherhood program. METHODS: The data collected was a retrograde by a proforma containing particulars of the diseased, detailed history and relatives were interviewed for additional information. The data collected was analysed. RESULTS: Maternal mortality rate in our own institution is 200/ 100,000 live births. Among 30 maternal deaths, 56% deaths (17 were among low socio - economic status, groups 60% deaths among unbooked 53.5% deaths more along illiterates evidenced by direct and indirect deaths about 25% of deaths were preventable. CONCLUSION: Maternal mortality is a global problem, facing every country in the world. Target specific interventions are needed for specific population. Fifth millennium development goal (MDG is to reduce maternal mortality by 75% by the year 2015, worthwhile investment for every case provider, results that investing on mothers
Phonetic recalibration of speech by text
Keetels, M.N.; Schakel, L.; de Bonte, M.; Vroomen, J.
2016-01-01
Listeners adjust their phonetic categories to cope with variations in the speech signal (phonetic recalibration). Previous studies have shown that lipread speech (and word knowledge) can adjust the perception of ambiguous speech and can induce phonetic adjustments (Bertelson, Vroomen, & de Gelder in
Epoch-based analysis of speech signals
Indian Academy of Sciences (India)
on speech production characteristics, but also helps in accurate analysis of speech. .... include time delay estimation, speech enhancement from single and multi- ...... log. (. E[k]. ∑K−1 l=0. E[l]. ) ,. (7) where K is the number of samples in the ...
Andreas Maier; Tino Haderlein; Florian Stelzle; Elmar Nöth; Emeka Nkenke; Frank Rosanowski; Anne Schützenberger; Maria Schuster
2010-01-01
In patients suffering from head and neck cancer, speech intelligibility is often restricted. For assessment and outcome measurements, automatic speech recognition systems have previously been shown to be appropriate for objective and quick evaluation of intelligibility. In this study we investigate the applicability of the method to speech disorders caused by head and neck cancer. Intelligibility was quantified by speech recognition on recordings of a standard text read by 41 German laryngect...
Gifford, René H; Noble, Jack H; Camarata, Stephen M; Sunderhaus, Linsey W; Dwyer, Robert T; Dawant, Benoit M; Dietrich, Mary S; Labadie, Robert F
2018-01-01
Adult cochlear implant (CI) recipients demonstrate a reliable relationship between spectral modulation detection and speech understanding. Prior studies documenting this relationship have focused on postlingually deafened adult CI recipients-leaving an open question regarding the relationship between spectral resolution and speech understanding for adults and children with prelingual onset of deafness. Here, we report CI performance on the measures of speech recognition and spectral modulation detection for 578 CI recipients including 477 postlingual adults, 65 prelingual adults, and 36 prelingual pediatric CI users. The results demonstrated a significant correlation between spectral modulation detection and various measures of speech understanding for 542 adult CI recipients. For 36 pediatric CI recipients, however, there was no significant correlation between spectral modulation detection and speech understanding in quiet or in noise nor was spectral modulation detection significantly correlated with listener age or age at implantation. These findings suggest that pediatric CI recipients might not depend upon spectral resolution for speech understanding in the same manner as adult CI recipients. It is possible that pediatric CI users are making use of different cues, such as those contained within the temporal envelope, to achieve high levels of speech understanding. Further investigation is warranted to investigate the relationship between spectral and temporal resolution and speech recognition to describe the underlying mechanisms driving peripheral auditory processing in pediatric CI users.
Directory of Open Access Journals (Sweden)
Joshua FRYE
2017-07-01
Full Text Available The Nobel Peace Prize has long been considered the premier peace prize in the world. According to Geir Lundestad, Secretary of the Nobel Committee, of the 300 some peace prizes awarded worldwide, “none is in any way as well known and as highly respected as the Nobel Peace Prize” (Lundestad, 2001. Nobel peace speech is a unique and significant international site of public discourse committed to articulating the universal grammar of peace. Spanning over 100 years of sociopolitical history on the world stage, Nobel Peace Laureates richly represent an important cross-section of domestic and international issues increasingly germane to many publics. Communication scholars’ interest in this rhetorical genre has increased in the past decade. Yet, the norm has been to analyze a single speech artifact from a prestigious or controversial winner rather than examine the collection of speeches for generic commonalities of import. In this essay, we analyze the discourse of Nobel peace speech inductively and argue that the organizing principle of the Nobel peace speech genre is the repetitive form of normative liberal principles and values that function as rhetorical topoi. These topoi include freedom and justice and appeal to the inviolable, inborn right of human beings to exercise certain political and civil liberties and the expectation of equality of protection from totalitarian and tyrannical abuses. The significance of this essay to contemporary communication theory is to expand our theoretical understanding of rhetoric’s role in the maintenance and development of an international and cross-cultural vocabulary for the grammar of peace.
Directory of Open Access Journals (Sweden)
Mike Wald
2006-12-01
Full Text Available The potential use of Automatic Speech Recognition to assist receptive communication is explored. The opportunities and challenges that this technology presents students and staff to provide captioning of speech online or in classrooms for deaf or hard of hearing students and assist blind, visually impaired or dyslexic learners to read and search learning material more readily by augmenting synthetic speech with natural recorded real speech is also discussed and evaluated. The automatic provision of online lecture notes, synchronised with speech, enables staff and students to focus on learning and teaching issues, while also benefiting learners unable to attend the lecture or who find it difficult or impossible to take notes at the same time as listening, watching and thinking.
Economic consequences of maternal illness in rural Bangladesh.
Powell-Jackson, Timothy; Hoque, Mohammad Enamul
2012-07-01
We use panel data on household consumption combined with information taken from the medical records of women who gave birth in health facilities to explore the economic consequences of maternal ill health, in the context of a rural population in Bangladesh. The findings suggest that there is a large reduction in household resources associated with maternal illness, driven almost entirely by spending on health care. In spite of this loss of resources, we find that households are able to fully insure consumption against maternal ill health, although confidence intervals are unable to rule out a small effect. Households in our study area are shown to have good access to informal credit (whether it be from local money lenders or family relatives), and this appears critical in helping to smooth consumption in response to these health shocks, at least in the short term. Copyright © 2011 John Wiley & Sons, Ltd.
Motor Speech Sequence Learning in Adults Who Stutter
Directory of Open Access Journals (Sweden)
Mahsa Aghazamani
2018-04-01
Conclusion The results of this study showed that PWS show improvement in accuracy, reaction time and sequence duration variables from day 1 to day 3. Also, PWS show more substantial number of errors compared to PNS, but this difference was not significant between the two groups. Similar results were obtained for the reaction time. Results of this study demonstrated that PWS show slower sequence duration compared to PNS. Some studies suggested that this could be because people who stutter use a control strategy to reduce the number of errors, although many studies suggested that this may indicate motor learning. According to speech motor skills hypothesis, it can be concluded that people who stutter have limitations in motor speech learning abilities. The findings of the present study could have clinical implication for the treatment of stuttering.
Salam, Rehana A; Lassi, Zohra S; Das, Jai K; Bhutta, Zulfiqar A
2014-09-04
District level healthcare serves as a nexus between community and district level facilities. Inputs at the district level can be broadly divided into governance and accountability mechanisms; leadership and supervision; financial platforms; and information systems. This paper aims to evaluate the effectivness of district level inputs for imporving maternal and newborn health. We considered all available systematic reviews published before May 2013 on the pre-defined district level interventions and included 47 systematic reviews. Evidence suggests that supervision positively influenced provider's practice, knowledge and client/provider satisfaction. Involving local opinion leaders to promote evidence-based practice improved compliance to the desired practice. Audit and feedback mechanisms and tele-medicine were found to be associated with improved immunization rates and mammogram uptake. User-directed financial schemes including maternal vouchers, user fee exemption and community based health insurance showed significant impact on maternal health service utilization with voucher schemes showing the most significant positive impact across all range of outcomes including antenatal care, skilled birth attendant, institutional delivery, complicated delivery and postnatal care. We found insufficient evidence to support or refute the use of electronic health record systems and telemedicine technology to improve maternal and newborn health specific outcomes. There is dearth of evidence on the effectiveness of district level inputs to improve maternal newborn health outcomes. Future studies should evaluate the impact of supervision and monitoring; electronic health record and tele-communication interventions in low-middle-income countries.
Attentional Demand of Speech in Children and Adolescents with Developmental Stuttering
Directory of Open Access Journals (Sweden)
Hajar Bahrami
2012-09-01
Full Text Available Background & Objective: Stuttering is a prevalent disorder in children and adolescents. Because attention is the only fuel resource for cognitive functions and the language have high cognitive functions, then it is possible that speech difficulties are related to attention deficit. The purpose of this study was to investigate the attentional demand of speech in children and adolescents with developmental stuttering. Materials & Methods: It is a dependent measurement study though which 30 school students (8-13 yr. were selected by convenience sampling and speech therapist´s detection from Shahriyar. The instruments were used in this research consist of: a text for reading, a device for recording of speech, and stuttering severity instrument-3 (SSI-3. The research was implemented in two conditions: single task (only reading and dual task (reading along finger tapping task. The data were analyzed using T- test. Results: Findings show that stuttering severity increased in dual task condition (divided attention. Conclusion: This result suggests that a decreased attentional capacity in children with developmental stuttering cause an increase in the number of stuttering words. With a better understanding of attentional functions of stuttering people as an important cognitive variables, we can take a step toward recognizing cognitive vulnerability of disorder. Therefore, intervention programs for children with developmental stuttering should pay attention to cognitive deficits and prior to speech interventions, the cognitive deficits should be eliminated with neuropsychological implements. With the improvement of neurological base of speech which is the first point of that in the brain, the considerable improvement may be seen in the stuttering severity.
Adriana Iordache
2015-01-01
The article analyzes the specificities of Romanian hate speech over a period of twelve years through a qualitative analysis of 384 Decisions of the National Council for Combating Discrimination. The study employs a coding methodology which allows one to separate decisions according to the group that was the victim of hate speech. The article finds that stereotypes employed are similar to those encountered in the international literature. The main target of hate speech is the Roma, who are ...
Narayan, Angela; Cicchetti, Dante; Rogosch, Fred A.; Toth, Sheree L.
2014-01-01
Research has documented that maternal expressed emotion-criticism (EE-Crit) from the Five-Minute Speech Sample (FMSS) predicts family conflict and children’s externalizing behavior in clinical and community samples. However, studies have not examined EE-Crit in maltreating or separated/divorced families, or whether these family risks exacerbate the links between EE-Crit and family conflict and externalizing behavior. The current study examined the associations between maternal EE-Crit, maltreatment, and separation/divorce, and whether maltreatment and separation/divorce moderated associations between EE-Crit and children’s externalizing problems, and EE-Crit and family conflict. Participants included 123 children (M = 8.01 years, SD = 1.58; 64.2% males) from maltreating (n = 83) or low-income, comparison (n = 40) families, and 123 mothers (n = 48 separated/divorced). Mothers completed the FMSS for EE-Crit and the Family Environment Scale for family conflict. Maltreatment was coded with the Maltreatment Classification System using information from official Child Protection Services (CPS) reports from the Department of Human Services (DHS). Trained summer camp counselors rated children’s externalizing behavior. Maltreatment was directly associated with higher externalizing problems, and separation/divorce, but not maltreatment, moderated the association between EE-Crit and externalizing behavior. Analyses pertaining to family conflict were not significant. Findings indicate that maltreatment is a direct risk factor for children’s externalizing behavior and separation/divorce is a vulnerability factor for externalizing behavior in family contexts with high maternal EE-Crit. Intervention, prevention, and policy efforts to promote resilience in high-risk families may be effective in targeting maltreating and critical parents, especially those with co-occurring separation/divorce. PMID:25037461
Tchoungui Oyono, Lilly; Pascoe, Michelle; Singh, Shajila
2018-05-17
The purpose of this study was to determine the prevalence of speech and language disorders in French-speaking preschool-age children in Yaoundé, the capital city of Cameroon. A total of 460 participants aged 3-5 years were recruited from the 7 communes of Yaoundé using a 2-stage cluster sampling method. Speech and language assessment was undertaken using a standardized speech and language test, the Evaluation du Langage Oral (Khomsi, 2001), which was purposefully renormed on the sample. A predetermined cutoff of 2 SDs below the normative mean was applied to identify articulation, expressive language, and receptive language disorders. Fluency and voice disorders were identified using clinical judgment by a speech-language pathologist. Overall prevalence was calculated as follows: speech disorders, 14.7%; language disorders, 4.3%; and speech and language disorders, 17.1%. In terms of disorders, prevalence findings were as follows: articulation disorders, 3.6%; expressive language disorders, 1.3%; receptive language disorders, 3%; fluency disorders, 8.4%; and voice disorders, 3.6%. Prevalence figures are higher than those reported for other countries and emphasize the urgent need to develop speech and language services for the Cameroonian population.
McLeod, Sharynne; Verdon, Sarah; Bowen, Caroline
2013-01-01
A major challenge for the speech-language pathology profession in many cultures is to address the mismatch between the "linguistic homogeneity of the speech-language pathology profession and the linguistic diversity of its clientele" (Caesar & Kohler, 2007, p. 198). This paper outlines the development of the Multilingual Children with Speech Sound Disorders: Position Paper created to guide speech-language pathologists' (SLPs') facilitation of multilingual children's speech. An international expert panel was assembled comprising 57 researchers (SLPs, linguists, phoneticians, and speech scientists) with knowledge about multilingual children's speech, or children with speech sound disorders. Combined, they had worked in 33 countries and used 26 languages in professional practice. Fourteen panel members met for a one-day workshop to identify key points for inclusion in the position paper. Subsequently, 42 additional panel members participated online to contribute to drafts of the position paper. A thematic analysis was undertaken of the major areas of discussion using two data sources: (a) face-to-face workshop transcript (133 pages) and (b) online discussion artifacts (104 pages). Finally, a moderator with international expertise in working with children with speech sound disorders facilitated the incorporation of the panel's recommendations. The following themes were identified: definitions, scope, framework, evidence, challenges, practices, and consideration of a multilingual audience. The resulting position paper contains guidelines for providing services to multilingual children with speech sound disorders (http://www.csu.edu.au/research/multilingual-speech/position-paper). The paper is structured using the International Classification of Functioning, Disability and Health: Children and Youth Version (World Health Organization, 2007) and incorporates recommendations for (a) children and families, (b) SLPs' assessment and intervention, (c) SLPs' professional
Shahin, Antoine J; Shen, Stanley; Kerlin, Jess R
2017-01-01
We examined the relationship between tolerance for audiovisual onset asynchrony (AVOA) and the spectrotemporal fidelity of the spoken words and the speaker's mouth movements. In two experiments that only varied in the temporal order of sensory modality, visual speech leading (exp1) or lagging (exp2) acoustic speech, participants watched intact and blurred videos of a speaker uttering trisyllabic words and nonwords that were noise vocoded with 4-, 8-, 16-, and 32-channels. They judged whether the speaker's mouth movements and the speech sounds were in-sync or out-of-sync . Individuals perceived synchrony (tolerated AVOA) on more trials when the acoustic speech was more speech-like (8 channels and higher vs. 4 channels), and when visual speech was intact than blurred (exp1 only). These findings suggest that enhanced spectrotemporal fidelity of the audiovisual (AV) signal prompts the brain to widen the window of integration promoting the fusion of temporally distant AV percepts.
Kane, Peter E., Ed.
This issue of "Free Speech" contains the following articles: "Daniel Schoor Relieved of Reporting Duties" by Laurence Stern, "The Sellout at CBS" by Michael Harrington, "Defending Dan Schorr" by Tome Wicker, "Speech to the Washington Press Club, February 25, 1976" by Daniel Schorr, "Funds…
Noise and pitch interact during the cortical segregation of concurrent speech.
Bidelman, Gavin M; Yellamsetty, Anusha
2017-08-01
Behavioral studies reveal listeners exploit intrinsic differences in voice fundamental frequency (F0) to segregate concurrent speech sounds-the so-called "F0-benefit." More favorable signal-to-noise ratio (SNR) in the environment, an extrinsic acoustic factor, similarly benefits the parsing of simultaneous speech. Here, we examined the neurobiological substrates of these two cues in the perceptual segregation of concurrent speech mixtures. We recorded event-related brain potentials (ERPs) while listeners performed a speeded double-vowel identification task. Listeners heard two concurrent vowels whose F0 differed by zero or four semitones presented in either clean (no noise) or noise-degraded (+5 dB SNR) conditions. Behaviorally, listeners were more accurate in correctly identifying both vowels for larger F0 separations but F0-benefit was more pronounced at more favorable SNRs (i.e., pitch × SNR interaction). Analysis of the ERPs revealed that only the P2 wave (∼200 ms) showed a similar F0 x SNR interaction as behavior and was correlated with listeners' perceptual F0-benefit. Neural classifiers applied to the ERPs further suggested that speech sounds are segregated neurally within 200 ms based on SNR whereas segregation based on pitch occurs later in time (400-700 ms). The earlier timing of extrinsic SNR compared to intrinsic F0-based segregation implies that the cortical extraction of speech from noise is more efficient than differentiating speech based on pitch cues alone, which may recruit additional cortical processes. Findings indicate that noise and pitch differences interact relatively early in cerebral cortex and that the brain arrives at the identities of concurrent speech mixtures as early as ∼200 ms. Copyright © 2017 Elsevier B.V. All rights reserved.
APPRECIATING SPEECH THROUGH GAMING
Directory of Open Access Journals (Sweden)
Mario T Carreon
2014-06-01
Full Text Available This paper discusses the Speech and Phoneme Recognition as an Educational Aid for the Deaf and Hearing Impaired (SPREAD application and the ongoing research on its deployment as a tool for motivating deaf and hearing impaired students to learn and appreciate speech. This application uses the Sphinx-4 voice recognition system to analyze the vocalization of the student and provide prompt feedback on their pronunciation. The packaging of the application as an interactive game aims to provide additional motivation for the deaf and hearing impaired student through visual motivation for them to learn and appreciate speech.