Sample records for spoken language performance

  1. The relation of the number of languages spoken to performance in different cognitive abilities in old age. (United States)

    Ihle, Andreas; Oris, Michel; Fagot, Delphine; Kliegel, Matthias


    Findings on the association of speaking different languages with cognitive functioning in old age are inconsistent and inconclusive so far. Therefore, the present study set out to investigate the relation of the number of languages spoken to cognitive performance and its interplay with several other markers of cognitive reserve in a large sample of older adults. Two thousand eight hundred and twelve older adults served as sample for the present study. Psychometric tests on verbal abilities, basic processing speed, and cognitive flexibility were administered. In addition, individuals were interviewed on their different languages spoken on a regular basis, educational attainment, occupation, and engaging in different activities throughout adulthood. Higher number of languages regularly spoken was significantly associated with better performance in verbal abilities and processing speed, but unrelated to cognitive flexibility. Regression analyses showed that the number of languages spoken predicted cognitive performance over and above leisure activities/physical demand of job/gainful activity as respective additional predictor, but not over and above educational attainment/cognitive level of job as respective additional predictor. There was no significant moderation of the association of the number of languages spoken with cognitive performance in any model. Present data suggest that speaking different languages on a regular basis may additionally contribute to the build-up of cognitive reserve in old age. Yet, this may not be universal, but linked to verbal abilities and basic cognitive processing speed. Moreover, it may be dependent on other types of cognitive stimulation that individuals also engaged in during their life course.

  2. Professionals' Guidance about Spoken Language Multilingualism and Spoken Language Choice for Children with Hearing Loss (United States)

    Crowe, Kathryn; McLeod, Sharynne


    The purpose of this research was to investigate factors that influence professionals' guidance of parents of children with hearing loss regarding spoken language multilingualism and spoken language choice. Sixteen professionals who provide services to children and young people with hearing loss completed an online survey, rating the importance of…

  3. Deep bottleneck features for spoken language identification.

    Directory of Open Access Journals (Sweden)

    Bing Jiang

    Full Text Available A key problem in spoken language identification (LID is to design effective representations which are specific to language information. For example, in recent years, representations based on both phonotactic and acoustic features have proven their effectiveness for LID. Although advances in machine learning have led to significant improvements, LID performance is still lacking, especially for short duration speech utterances. With the hypothesis that language information is weak and represented only latently in speech, and is largely dependent on the statistical properties of the speech content, existing representations may be insufficient. Furthermore they may be susceptible to the variations caused by different speakers, specific content of the speech segments, and background noise. To address this, we propose using Deep Bottleneck Features (DBF for spoken LID, motivated by the success of Deep Neural Networks (DNN in speech recognition. We show that DBFs can form a low-dimensional compact representation of the original inputs with a powerful descriptive and discriminative capability. To evaluate the effectiveness of this, we design two acoustic models, termed DBF-TV and parallel DBF-TV (PDBF-TV, using a DBF based i-vector representation for each speech utterance. Results on NIST language recognition evaluation 2009 (LRE09 show significant improvements over state-of-the-art systems. By fusing the output of phonotactic and acoustic approaches, we achieve an EER of 1.08%, 1.89% and 7.01% for 30 s, 10 s and 3 s test utterances respectively. Furthermore, various DBF configurations have been extensively evaluated, and an optimal system proposed.

  4. Spoken Grammar and Its Role in the English Language Classroom (United States)

    Hilliard, Amanda


    This article addresses key issues and considerations for teachers wanting to incorporate spoken grammar activities into their own teaching and also focuses on six common features of spoken grammar, with practical activities and suggestions for teaching them in the language classroom. The hope is that this discussion of spoken grammar and its place…

  5. Speech-Language Pathologists: Vital Listening and Spoken Language Professionals (United States)

    Houston, K. Todd; Perigoe, Christina B.


    Determining the most effective methods and techniques to facilitate the spoken language development of individuals with hearing loss has been a focus of practitioners for centuries. Due to modern advances in hearing technology, earlier identification of hearing loss, and immediate enrollment in early intervention, children with hearing loss are…

  6. How Does the Linguistic Distance Between Spoken and Standard Language in Arabic Affect Recall and Recognition Performances During Verbal Memory Examination. (United States)

    Taha, Haitham


    The current research examined how Arabic diglossia affects verbal learning memory. Thirty native Arab college students were tested using auditory verbal memory test that was adapted according to the Rey Auditory Verbal Learning Test and developed in three versions: Pure spoken language version (SL), pure standard language version (SA), and phonologically similar version (PS). The result showed that for immediate free-recall, the performances were better for the SL and the PS conditions compared to the SA one. However, for the parts of delayed recall and recognition, the results did not reveal any significant consistent effect of diglossia. Accordingly, it was suggested that diglossia has a significant effect on the storage and short term memory functions but not on long term memory functions. The results were discussed in light of different approaches in the field of bilingual memory.

  7. How Does the Linguistic Distance between Spoken and Standard Language in Arabic Affect Recall and Recognition Performances during Verbal Memory Examination (United States)

    Taha, Haitham


    The current research examined how Arabic diglossia affects verbal learning memory. Thirty native Arab college students were tested using auditory verbal memory test that was adapted according to the Rey Auditory Verbal Learning Test and developed in three versions: Pure spoken language version (SL), pure standard language version (SA), and…

  8. Spoken Indian language identification: a review of features and ...

    Indian Academy of Sciences (India)



    Apr 12, 2018 ... languages and can be used for the purposes of spoken language identification. Keywords. SLID .... branch of linguistics to study the sound structure of human language. ... countries, work in the area of Indian language identification has not ...... English and speech database has been collected over tele-.

  9. ELSIE: The Quick Reaction Spoken Language Translation (QRSLT)

    National Research Council Canada - National Science Library

    Montgomery, Christine


    The objective of this effort was to develop a prototype, hand-held or body-mounted spoken language translator to assist military and law enforcement personnel in interacting with non-English-speaking people...

  10. Using Spoken Language to Facilitate Military Transportation Planning

    National Research Council Canada - National Science Library

    Bates, Madeleine; Ellard, Dan; Peterson, Pat; Shaked, Varda


    .... In an effort to demonstrate the relevance of SIS technology to real-world military applications, BBN has undertaken the task of providing a spoken language interface to DART, a system for military...

  11. The employment of a spoken language computer applied to an air traffic control task. (United States)

    Laveson, J. I.; Silver, C. A.


    Assessment of the merits of a limited spoken language (56 words) computer in a simulated air traffic control (ATC) task. An airport zone approximately 60 miles in diameter with a traffic flow simulation ranging from single-engine to commercial jet aircraft provided the workload for the controllers. This research determined that, under the circumstances of the experiments carried out, the use of a spoken-language computer would not improve the controller performance.

  12. Attentional Capture of Objects Referred to by Spoken Language (United States)

    Salverda, Anne Pier; Altmann, Gerry T. M.


    Participants saw a small number of objects in a visual display and performed a visual detection or visual-discrimination task in the context of task-irrelevant spoken distractors. In each experiment, a visual cue was presented 400 ms after the onset of a spoken word. In experiments 1 and 2, the cue was an isoluminant color change and participants…

  13. Improving Spoken Language Outcomes for Children With Hearing Loss: Data-driven Instruction. (United States)

    Douglas, Michael


    To assess the effects of data-driven instruction (DDI) on spoken language outcomes of children with cochlear implants and hearing aids. Retrospective, matched-pairs comparison of post-treatment speech/language data of children who did and did not receive DDI. Private, spoken-language preschool for children with hearing loss. Eleven matched pairs of children with cochlear implants who attended the same spoken language preschool. Groups were matched for age of hearing device fitting, time in the program, degree of predevice fitting hearing loss, sex, and age at testing. Daily informal language samples were collected and analyzed over a 2-year period, per preschool protocol. Annual informal and formal spoken language assessments in articulation, vocabulary, and omnibus language were administered at the end of three time intervals: baseline, end of year one, and end of year two. The primary outcome measures were total raw score performance of spontaneous utterance sentence types and syntax element use as measured by the Teacher Assessment of Spoken Language (TASL). In addition, standardized assessments (the Clinical Evaluation of Language Fundamentals--Preschool Version 2 (CELF-P2), the Expressive One-Word Picture Vocabulary Test (EOWPVT), the Receptive One-Word Picture Vocabulary Test (ROWPVT), and the Goldman-Fristoe Test of Articulation 2 (GFTA2)) were also administered and compared with the control group. The DDI group demonstrated significantly higher raw scores on the TASL each year of the study. The DDI group also achieved statistically significant higher scores for total language on the CELF-P and expressive vocabulary on the EOWPVT, but not for articulation nor receptive vocabulary. Post-hoc assessment revealed that 78% of the students in the DDI group achieved scores in the average range compared with 59% in the control group. The preliminary results of this study support further investigation regarding DDI to investigate whether this method can consistently


    Directory of Open Access Journals (Sweden)

    Jelena Kuvač Kraljević


    Full Text Available Interest in spoken-language corpora has increased over the past two decades leading to the development of new corpora and the discovery of new facets of spoken language. These types of corpora represent the most comprehensive data source about the language of ordinary speakers. Such corpora are based on spontaneous, unscripted speech defined by a variety of styles, registers and dialects. The aim of this paper is to present the Croatian Adult Spoken Language Corpus (HrAL, its structure and its possible applications in different linguistic subfields. HrAL was built by sampling spontaneous conversations among 617 speakers from all Croatian counties, and it comprises more than 250,000 tokens and more than 100,000 types. Data were collected during three time slots: from 2010 to 2012, from 2014 to 2015 and during 2016. HrAL is today available within TalkBank, a large database of spoken-language corpora covering different languages (, in the Conversational Analyses corpora within the subsection titled Conversational Banks. Data were transcribed, coded and segmented using the transcription format Codes for Human Analysis of Transcripts (CHAT and the Computerised Language Analysis (CLAN suite of programmes within the TalkBank toolkit. Speech streams were segmented into communication units (C-units based on syntactic criteria. Most transcripts were linked to their source audios. The TalkBank is public free, i.e. all data stored in it can be shared by the wider community in accordance with the basic rules of the TalkBank. HrAL provides information about spoken grammar and lexicon, discourse skills, error production and productivity in general. It may be useful for sociolinguistic research and studies of synchronic language changes in Croatian.

  15. Auditory and verbal memory predictors of spoken language skills in children with cochlear implants. (United States)

    de Hoog, Brigitte E; Langereis, Margreet C; van Weerdenburg, Marjolijn; Keuning, Jos; Knoors, Harry; Verhoeven, Ludo


    Large variability in individual spoken language outcomes remains a persistent finding in the group of children with cochlear implants (CIs), particularly in their grammatical development. In the present study, we examined the extent of delay in lexical and morphosyntactic spoken language levels of children with CIs as compared to those of a normative sample of age-matched children with normal hearing. Furthermore, the predictive value of auditory and verbal memory factors in the spoken language performance of implanted children was analyzed. Thirty-nine profoundly deaf children with CIs were assessed using a test battery including measures of lexical, grammatical, auditory and verbal memory tests. Furthermore, child-related demographic characteristics were taken into account. The majority of the children with CIs did not reach age-equivalent lexical and morphosyntactic language skills. Multiple linear regression analyses revealed that lexical spoken language performance in children with CIs was best predicted by age at testing, phoneme perception, and auditory word closure. The morphosyntactic language outcomes of the CI group were best predicted by lexicon, auditory word closure, and auditory memory for words. Qualitatively good speech perception skills appear to be crucial for lexical and grammatical development in children with CIs. Furthermore, strongly developed vocabulary skills and verbal memory abilities predict morphosyntactic language skills. Copyright © 2016 Elsevier Ltd. All rights reserved.

  16. Spoken language outcomes after hemispherectomy: factoring in etiology. (United States)

    Curtiss, S; de Bode, S; Mathern, G W


    We analyzed postsurgery linguistic outcomes of 43 hemispherectomy patients operated on at UCLA. We rated spoken language (Spoken Language Rank, SLR) on a scale from 0 (no language) to 6 (mature grammar) and examined the effects of side of resection/damage, age at surgery/seizure onset, seizure control postsurgery, and etiology on language development. Etiology was defined as developmental (cortical dysplasia and prenatal stroke) and acquired pathology (Rasmussen's encephalitis and postnatal stroke). We found that clinical variables were predictive of language outcomes only when they were considered within distinct etiology groups. Specifically, children with developmental etiologies had lower SLRs than those with acquired pathologies (p =.0006); age factors correlated positively with higher SLRs only for children with acquired etiologies (p =.0006); right-sided resections led to higher SLRs only for the acquired group (p =.0008); and postsurgery seizure control correlated positively with SLR only for those with developmental etiologies (p =.0047). We argue that the variables considered are not independent predictors of spoken language outcome posthemispherectomy but should be viewed instead as characteristics of etiology. Copyright 2001 Elsevier Science.

  17. A real-time spoken-language system for interactive problem-solving, combining linguistic and statistical technology for improved spoken language understanding (United States)

    Moore, Robert C.; Cohen, Michael H.


    Under this effort, SRI has developed spoken-language technology for interactive problem solving, featuring real-time performance for up to several thousand word vocabularies, high semantic accuracy, habitability within the domain, and robustness to many sources of variability. Although the technology is suitable for many applications, efforts to date have focused on developing an Air Travel Information System (ATIS) prototype application. SRI's ATIS system has been evaluated in four ARPA benchmark evaluations, and has consistently been at or near the top in performance. These achievements are the result of SRI's technical progress in speech recognition, natural-language processing, and speech and natural-language integration.

  18. Prosodic Parallelism – comparing spoken and written language

    Directory of Open Access Journals (Sweden)

    Richard Wiese


    Full Text Available The Prosodic Parallelism hypothesis claims adjacent prosodic categories to prefer identical branching of internal adjacent constituents. According to Wiese and Speyer (2015, this preference implies feet contained in the same phonological phrase to display either binary or unary branching, but not different types of branching. The seemingly free schwa-zero alternations at the end of some words in German make it possible to test this hypothesis. The hypothesis was successfully tested by conducting a corpus study which used large-scale bodies of written German. As some open questions remain, and as it is unclear whether Prosodic Parallelism is valid for the spoken modality as well, the present study extends this inquiry to spoken German. As in the previous study, the results of a corpus analysis recruiting a variety of linguistic constructions are presented. The Prosodic Parallelism hypothesis can be demonstrated to be valid for spoken German as well as for written German. The paper thus contributes to the question whether prosodic preferences are similar between the spoken and written modes of a language. Some consequences of the results for the production of language are discussed.

  19. Inferring Speaker Affect in Spoken Natural Language Communication


    Pon-Barry, Heather Roberta


    The field of spoken language processing is concerned with creating computer programs that can understand human speech and produce human-like speech. Regarding the problem of understanding human speech, there is currently growing interest in moving beyond speech recognition (the task of transcribing the words in an audio stream) and towards machine listening—interpreting the full spectrum of information in an audio stream. One part of machine listening, the problem that this thesis focuses on, ...

  20. Grammatical Deviations in the Spoken and Written Language of Hebrew-Speaking Children With Hearing Impairments. (United States)

    Tur-Kaspa, Hana; Dromi, Esther


    The present study reports a detailed analysis of written and spoken language samples of Hebrew-speaking children aged 11-13 years who are deaf. It focuses on the description of various grammatical deviations in the two modalities. Participants were 13 students with hearing impairments (HI) attending special classrooms integrated into two elementary schools in Tel Aviv, Israel, and 9 students with normal hearing (NH) in regular classes in these same schools. Spoken and written language samples were collected from all participants using the same five preplanned elicitation probes. Students with HI were found to display significantly more grammatical deviations than their NH peers in both their spoken and written language samples. Most importantly, between-modality differences were noted. The participants with HI exhibited significantly more grammatical deviations in their written language samples than in their spoken samples. However, the distribution of grammatical deviations across categories was similar in the two modalities. The most common grammatical deviations in order of their frequency were failure to supply obligatory morphological markers, failure to mark grammatical agreement, and the omission of a major syntactic constituent in a sentence. Word order violations were rarely recorded in the Hebrew samples. Performance differences in the two modalities encourage clinicians and teachers to facilitate target linguistic forms in diverse communication contexts. Furthermore, the identification of linguistic targets for intervention must be based on the unique grammatical structure of the target language.

  1. Spoken Language Understanding Systems for Extracting Semantic Information from Speech

    CERN Document Server

    Tur, Gokhan


    Spoken language understanding (SLU) is an emerging field in between speech and language processing, investigating human/ machine and human/ human communication by leveraging technologies from signal processing, pattern recognition, machine learning and artificial intelligence. SLU systems are designed to extract the meaning from speech utterances and its applications are vast, from voice search in mobile devices to meeting summarization, attracting interest from both commercial and academic sectors. Both human/machine and human/human communications can benefit from the application of SLU, usin

  2. Spoken Spanish Language Development at the High School Level: A Mixed-Methods Study (United States)

    Moeller, Aleidine J.; Theiler, Janine


    Communicative approaches to teaching language have emphasized the centrality of oral proficiency in the language acquisition process, but research investigating oral proficiency has been surprisingly limited, yielding an incomplete understanding of spoken language development. This study investigated the development of spoken language at the high…

  3. Factors Influencing Verbal Intelligence and Spoken Language in Children with Phenylketonuria. (United States)

    Soleymani, Zahra; Keramati, Nasrin; Rohani, Farzaneh; Jalaei, Shohre


    To determine verbal intelligence and spoken language of children with phenylketonuria and to study the effect of age at diagnosis and phenylalanine plasma level on these abilities. Cross-sectional. Children with phenylketonuria were recruited from pediatric hospitals in 2012. Normal control subjects were recruited from kindergartens in Tehran. 30 phenylketonuria and 42 control subjects aged 4-6.5 years. Skills were compared between 3 phenylketonuria groups categorized by age at diagnosis/treatment, and between the phenylketonuria and control groups. Scores on Wechsler Preschool and Primary Scale of Intelligence for verbal and total intelligence, and Test of Language Development-Primary, third edition for spoken language, listening, speaking, semantics, syntax, and organization. The performance of control subjects was significantly better than that of early-treated subjects for all composite quotients from Test of Language Development and verbal intelligence (Pphenylketonuria subjects.

  4. Asian/Pacific Islander Languages Spoken by English Learners (ELs). Fast Facts (United States)

    Office of English Language Acquisition, US Department of Education, 2015


    The Office of English Language Acquisition (OELA) has synthesized key data on English learners (ELs) into two-page PDF sheets, by topic, with graphics, plus key contacts. The topics for this report on Asian/Pacific Islander languages spoken by English Learners (ELs) include: (1) Top 10 Most Common Asian/Pacific Islander Languages Spoken Among ELs:…

  5. Native Language Spoken as a Risk Marker for Tooth Decay. (United States)

    Carson, J; Walker, L A; Sanders, B J; Jones, J E; Weddell, J A; Tomlin, A M


    The purpose of this study was to assess dmft, the number of decayed, missing (due to caries), and/ or filled primary teeth, of English-speaking and non-English speaking patients of a hospital based pediatric dental clinic under the age of 72 months to determine if native language is a risk marker for tooth decay. Records from an outpatient dental clinic which met the inclusion criteria were reviewed. Patient demographics and dmft score were recorded, and the patients were separated into three groups by the native language spoken by their parents: English, Spanish and all other languages. A total of 419 charts were assessed: 253 English-speaking, 126 Spanish-speaking, and 40 other native languages. After accounting for patient characteristics, dmft was significantly higher for the other language group than for the English-speaking (p0.05). Those patients under 72 months of age whose parents' native language is not English or Spanish, have the highest risk for increased dmft when compared to English and Spanish speaking patients. Providers should consider taking additional time to educate patients and their parents, in their native language, on the importance of routine dental care and oral hygiene.

  6. Pointing and Reference in Sign Language and Spoken Language: Anchoring vs. Identifying (United States)

    Barberà, Gemma; Zwets, Martine


    In both signed and spoken languages, pointing serves to direct an addressee's attention to a particular entity. This entity may be either present or absent in the physical context of the conversation. In this article we focus on pointing directed to nonspeaker/nonaddressee referents in Sign Language of the Netherlands (Nederlandse Gebarentaal,…

  7. The Listening and Spoken Language Data Repository: Design and Project Overview (United States)

    Bradham, Tamala S.; Fonnesbeck, Christopher; Toll, Alice; Hecht, Barbara F.


    Purpose: The purpose of the Listening and Spoken Language Data Repository (LSL-DR) was to address a critical need for a systemwide outcome data-monitoring program for the development of listening and spoken language skills in highly specialized educational programs for children with hearing loss highlighted in Goal 3b of the 2007 Joint Committee…

  8. Development of Mandarin spoken language after pediatric cochlear implantation. (United States)

    Li, Bei; Soli, Sigfrid D; Zheng, Yun; Li, Gang; Meng, Zhaoli


    The purpose of this study was to evaluate early spoken language development in young Mandarin-speaking children during the first 24 months after cochlear implantation, as measured by receptive and expressive vocabulary growth rates. Growth rates were compared with those of normally hearing children and with growth rates for English-speaking children with cochlear implants. Receptive and expressive vocabularies were measured with the simplified short form (SSF) version of the Mandarin Communicative Development Inventory (MCDI) in a sample of 112 pediatric implant recipients at baseline, 3, 6, 12, and 24 months after implantation. Implant ages ranged from 1 to 5 years. Scores were expressed in terms of normal equivalent ages, allowing normalized vocabulary growth rates to be determined. Scores for English-speaking children were re-expressed in these terms, allowing direct comparisons of Mandarin and English early spoken language development. Vocabulary growth rates during the first 12 months after implantation were similar to those for normally hearing children less than 16 months of age. Comparisons with growth rates for normally hearing children 16-30 months of age showed that the youngest implant age group (1-2 years) had an average growth rate of 0.68 that of normally hearing children; while the middle implant age group (2-3 years) had an average growth rate of 0.65; and the oldest implant age group (>3 years) had an average growth rate of 0.56, significantly less than the other two rates. Growth rates for English-speaking children with cochlear implants were 0.68 in the youngest group, 0.54 in the middle group, and 0.57 in the oldest group. Growth rates in the middle implant age groups for the two languages differed significantly. The SSF version of the MCDI is suitable for assessment of Mandarin language development during the first 24 months after cochlear implantation. Effects of implant age and duration of implantation can be compared directly across

  9. Language and Culture in the Multiethnic Community: Spoken Language Assessment. (United States)

    Matluck, Joseph H.; Mace-Matluck, Betty J.

    This paper discusses the sociolinguistic problems inherent in multilingual testing, and the accompanying dangers of cultural bias in either the visuals or the language used in a given test. The first section discusses English-speaking Americans' perception of foreign speakers in terms of: (1) physical features; (2) speech, specifically vocabulary,…

  10. Monitoring the Performance of Human and Automated Scores for Spoken Responses (United States)

    Wang, Zhen; Zechner, Klaus; Sun, Yu


    As automated scoring systems for spoken responses are increasingly used in language assessments, testing organizations need to analyze their performance, as compared to human raters, across several dimensions, for example, on individual items or based on subgroups of test takers. In addition, there is a need in testing organizations to establish…

  11. Foreign Language Tutoring in Oral Conversations Using Spoken Dialog Systems (United States)

    Lee, Sungjin; Noh, Hyungjong; Lee, Jonghoon; Lee, Kyusong; Lee, Gary Geunbae

    Although there have been enormous investments into English education all around the world, not many differences have been made to change the English instruction style. Considering the shortcomings for the current teaching-learning methodology, we have been investigating advanced computer-assisted language learning (CALL) systems. This paper aims at summarizing a set of POSTECH approaches including theories, technologies, systems, and field studies and providing relevant pointers. On top of the state-of-the-art technologies of spoken dialog system, a variety of adaptations have been applied to overcome some problems caused by numerous errors and variations naturally produced by non-native speakers. Furthermore, a number of methods have been developed for generating educational feedback that help learners develop to be proficient. Integrating these efforts resulted in intelligent educational robots — Mero and Engkey — and virtual 3D language learning games, Pomy. To verify the effects of our approaches on students' communicative abilities, we have conducted a field study at an elementary school in Korea. The results showed that our CALL approaches can be enjoyable and fruitful activities for students. Although the results of this study bring us a step closer to understanding computer-based education, more studies are needed to consolidate the findings.

  12. Auditory and verbal memory predictors of spoken language skills in children with cochlear implants

    NARCIS (Netherlands)

    Hoog, B.E. de; Langereis, M.C.; Weerdenburg, M. van; Keuning, J.; Knoors, H.; Verhoeven, L.


    BACKGROUND: Large variability in individual spoken language outcomes remains a persistent finding in the group of children with cochlear implants (CIs), particularly in their grammatical development. AIMS: In the present study, we examined the extent of delay in lexical and morphosyntactic spoken

  13. Auditory and verbal memory predictors of spoken language skills in children with cochlear implants

    NARCIS (Netherlands)

    Hoog, B.E. de; Langereis, M.C.; Weerdenburg, M.W.C. van; Keuning, J.; Knoors, H.E.T.; Verhoeven, L.T.W.


    Background: Large variability in individual spoken language outcomes remains a persistent finding in the group of children with cochlear implants (CIs), particularly in their grammatical development. Aims: In the present study, we examined the extent of delay in lexical and morphosyntactic spoken

  14. Sign Language and Spoken Language for Children With Hearing Loss: A Systematic Review. (United States)

    Fitzpatrick, Elizabeth M; Hamel, Candyce; Stevens, Adrienne; Pratt, Misty; Moher, David; Doucet, Suzanne P; Neuss, Deirdre; Bernstein, Anita; Na, Eunjung


    Permanent hearing loss affects 1 to 3 per 1000 children and interferes with typical communication development. Early detection through newborn hearing screening and hearing technology provide most children with the option of spoken language acquisition. However, no consensus exists on optimal interventions for spoken language development. To conduct a systematic review of the effectiveness of early sign and oral language intervention compared with oral language intervention only for children with permanent hearing loss. An a priori protocol was developed. Electronic databases (eg, Medline, Embase, CINAHL) from 1995 to June 2013 and gray literature sources were searched. Studies in English and French were included. Two reviewers screened potentially relevant articles. Outcomes of interest were measures of auditory, vocabulary, language, and speech production skills. All data collection and risk of bias assessments were completed and then verified by a second person. Grades of Recommendation, Assessment, Development, and Evaluation (GRADE) was used to judge the strength of evidence. Eleven cohort studies met inclusion criteria, of which 8 included only children with severe to profound hearing loss with cochlear implants. Language development was the most frequently reported outcome. Other reported outcomes included speech and speech perception. Several measures and metrics were reported across studies, and descriptions of interventions were sometimes unclear. Very limited, and hence insufficient, high-quality evidence exists to determine whether sign language in combination with oral language is more effective than oral language therapy alone. More research is needed to supplement the evidence base. Copyright © 2016 by the American Academy of Pediatrics.

  15. Neural organization of linguistic short-term memory is sensory modality-dependent: evidence from signed and spoken language. (United States)

    Pa, Judy; Wilson, Stephen M; Pickell, Herbert; Bellugi, Ursula; Hickok, Gregory


    Despite decades of research, there is still disagreement regarding the nature of the information that is maintained in linguistic short-term memory (STM). Some authors argue for abstract phonological codes, whereas others argue for more general sensory traces. We assess these possibilities by investigating linguistic STM in two distinct sensory-motor modalities, spoken and signed language. Hearing bilingual participants (native in English and American Sign Language) performed equivalent STM tasks in both languages during functional magnetic resonance imaging. Distinct, sensory-specific activations were seen during the maintenance phase of the task for spoken versus signed language. These regions have been previously shown to respond to nonlinguistic sensory stimulation, suggesting that linguistic STM tasks recruit sensory-specific networks. However, maintenance-phase activations common to the two languages were also observed, implying some form of common process. We conclude that linguistic STM involves sensory-dependent neural networks, but suggest that sensory-independent neural networks may also exist.

  16. Iconicity as a general property of language: evidence from spoken and signed languages

    Directory of Open Access Journals (Sweden)

    Pamela Perniss


    Full Text Available Current views about language are dominated by the idea of arbitrary connections between linguistic form and meaning. However, if we look beyond the more familiar Indo-European languages and also include both spoken and signed language modalities, we find that motivated, iconic form-meaning mappings are, in fact, pervasive in language. In this paper, we review the different types of iconic mappings that characterize languages in both modalities, including the predominantly visually iconic mappings in signed languages. Having shown that iconic mapping are present across languages, we then proceed to review evidence showing that language users (signers and speakers exploit iconicity in language processing and language acquisition. While not discounting the presence and importance of arbitrariness in language, we put forward the idea that iconicity need also be recognized as a general property of language, which may serve the function of reducing the gap between linguistic form and conceptual representation to allow the language system to hook up to motor and perceptual experience.

  17. Predictors of spoken language development following pediatric cochlear implantation. (United States)

    Boons, Tinne; Brokx, Jan P L; Dhooge, Ingeborg; Frijns, Johan H M; Peeraer, Louis; Vermeulen, Anneke; Wouters, Jan; van Wieringen, Astrid


    Although deaf children with cochlear implants (CIs) are able to develop good language skills, the large variability in outcomes remains a significant concern. The first aim of this study was to evaluate language skills in children with CIs to establish benchmarks. The second aim was to make an estimation of the optimal age at implantation to provide maximal opportunities for the child to achieve good language skills afterward. The third aim was to gain more insight into the causes of variability to set recommendations for optimizing the rehabilitation process of prelingually deaf children with CIs. Receptive and expressive language development of 288 children who received CIs by age five was analyzed in a retrospective multicenter study. Outcome measures were language quotients (LQs) on the Reynell Developmental Language Scales and Schlichting Expressive Language Test at 1, 2, and 3 years after implantation. Independent predictive variables were nine child-related, environmental, and auditory factors. A series of multiple regression analyses determined the amount of variance in expressive and receptive language outcomes attributable to each predictor when controlling for the other variables. Simple linear regressions with age at first fitting and independent samples t tests demonstrated that children implanted before the age of two performed significantly better on all tests than children who were implanted at an older age. The mean LQ was 0.78 with an SD of 0.18. A child with an LQ lower than 0.60 (= 0.78-0.18) within 3 years after implantation was labeled as a weak performer compared with other deaf children implanted before the age of two. Contralateral stimulation with a second CI or a hearing aid and the absence of additional disabilities were related to better language outcomes. The effect of environmental factors, comprising multilingualism, parental involvement, and communication mode increased over time. Three years after implantation, the total multiple

  18. Spoken Word Recognition in Adolescents with Autism Spectrum Disorders and Specific Language Impairment (United States)

    Loucas, Tom; Riches, Nick; Baird, Gillian; Pickles, Andrew; Simonoff, Emily; Chandler, Susie; Charman, Tony


    Spoken word recognition, during gating, appears intact in specific language impairment (SLI). This study used gating to investigate the process in adolescents with autism spectrum disorders plus language impairment (ALI). Adolescents with ALI, SLI, and typical language development (TLD), matched on nonverbal IQ listened to gated words that varied…

  19. Semantic Fluency in Deaf Children Who Use Spoken and Signed Language in Comparison with Hearing Peers (United States)

    Marshall, C. R.; Jones, A.; Fastelli, A.; Atkinson, J.; Botting, N.; Morgan, G.


    Background: Deafness has an adverse impact on children's ability to acquire spoken languages. Signed languages offer a more accessible input for deaf children, but because the vast majority are born to hearing parents who do not sign, their early exposure to sign language is limited. Deaf children as a whole are therefore at high risk of language…

  20. Resting-state low-frequency fluctuations reflect individual differences in spoken language learning (United States)

    Deng, Zhizhou; Chandrasekaran, Bharath; Wang, Suiping; Wong, Patrick C.M.


    A major challenge in language learning studies is to identify objective, pre-training predictors of success. Variation in the low-frequency fluctuations (LFFs) of spontaneous brain activity measured by resting-state functional magnetic resonance imaging (RS-fMRI) has been found to reflect individual differences in cognitive measures. In the present study, we aimed to investigate the extent to which initial spontaneous brain activity is related to individual differences in spoken language learning. We acquired RS-fMRI data and subsequently trained participants on a sound-to-word learning paradigm in which they learned to use foreign pitch patterns (from Mandarin Chinese) to signal word meaning. We performed amplitude of spontaneous low-frequency fluctuation (ALFF) analysis, graph theory-based analysis, and independent component analysis (ICA) to identify functional components of the LFFs in the resting-state. First, we examined the ALFF as a regional measure and showed that regional ALFFs in the left superior temporal gyrus were positively correlated with learning performance, whereas ALFFs in the default mode network (DMN) regions were negatively correlated with learning performance. Furthermore, the graph theory-based analysis indicated that the degree and local efficiency of the left superior temporal gyrus were positively correlated with learning performance. Finally, the default mode network and several task-positive resting-state networks (RSNs) were identified via the ICA. The “competition” (i.e., negative correlation) between the DMN and the dorsal attention network was negatively correlated with learning performance. Our results demonstrate that a) spontaneous brain activity can predict future language learning outcome without prior hypotheses (e.g., selection of regions of interest – ROIs) and b) both regional dynamics and network-level interactions in the resting brain can account for individual differences in future spoken language learning success

  1. Resting-state low-frequency fluctuations reflect individual differences in spoken language learning. (United States)

    Deng, Zhizhou; Chandrasekaran, Bharath; Wang, Suiping; Wong, Patrick C M


    A major challenge in language learning studies is to identify objective, pre-training predictors of success. Variation in the low-frequency fluctuations (LFFs) of spontaneous brain activity measured by resting-state functional magnetic resonance imaging (RS-fMRI) has been found to reflect individual differences in cognitive measures. In the present study, we aimed to investigate the extent to which initial spontaneous brain activity is related to individual differences in spoken language learning. We acquired RS-fMRI data and subsequently trained participants on a sound-to-word learning paradigm in which they learned to use foreign pitch patterns (from Mandarin Chinese) to signal word meaning. We performed amplitude of spontaneous low-frequency fluctuation (ALFF) analysis, graph theory-based analysis, and independent component analysis (ICA) to identify functional components of the LFFs in the resting-state. First, we examined the ALFF as a regional measure and showed that regional ALFFs in the left superior temporal gyrus were positively correlated with learning performance, whereas ALFFs in the default mode network (DMN) regions were negatively correlated with learning performance. Furthermore, the graph theory-based analysis indicated that the degree and local efficiency of the left superior temporal gyrus were positively correlated with learning performance. Finally, the default mode network and several task-positive resting-state networks (RSNs) were identified via the ICA. The "competition" (i.e., negative correlation) between the DMN and the dorsal attention network was negatively correlated with learning performance. Our results demonstrate that a) spontaneous brain activity can predict future language learning outcome without prior hypotheses (e.g., selection of regions of interest--ROIs) and b) both regional dynamics and network-level interactions in the resting brain can account for individual differences in future spoken language learning success

  2. The effect of written text on comprehension of spoken English as a foreign language. (United States)

    Diao, Yali; Chandler, Paul; Sweller, John


    Based on cognitive load theory, this study investigated the effect of simultaneous written presentations on comprehension of spoken English as a foreign language. Learners' language comprehension was compared while they used 3 instructional formats: listening with auditory materials only, listening with a full, written script, and listening with simultaneous subtitled text. Listening with the presence of a script and subtitles led to better understanding of the scripted and subtitled passage but poorer performance on a subsequent auditory passage than listening with the auditory materials only. These findings indicated that where the intention was learning to listen, the use of a full script or subtitles had detrimental effects on the construction and automation of listening comprehension schemas.

  3. Tonal Language Background and Detecting Pitch Contour in Spoken and Musical Items (United States)

    Stevens, Catherine J.; Keller, Peter E.; Tyler, Michael D.


    An experiment investigated the effect of tonal language background on discrimination of pitch contour in short spoken and musical items. It was hypothesized that extensive exposure to a tonal language attunes perception of pitch contour. Accuracy and reaction times of adult participants from tonal (Thai) and non-tonal (Australian English) language…

  4. Delayed Anticipatory Spoken Language Processing in Adults with Dyslexia—Evidence from Eye-tracking. (United States)

    Huettig, Falk; Brouwer, Susanne


    It is now well established that anticipation of upcoming input is a key characteristic of spoken language comprehension. It has also frequently been observed that literacy influences spoken language processing. Here, we investigated whether anticipatory spoken language processing is related to individuals' word reading abilities. Dutch adults with dyslexia and a control group participated in two eye-tracking experiments. Experiment 1 was conducted to assess whether adults with dyslexia show the typical language-mediated eye gaze patterns. Eye movements of both adults with and without dyslexia closely replicated earlier research: spoken language is used to direct attention to relevant objects in the environment in a closely time-locked manner. In Experiment 2, participants received instructions (e.g., 'Kijk naar de(COM) afgebeelde piano(COM)', look at the displayed piano) while viewing four objects. Articles (Dutch 'het' or 'de') were gender marked such that the article agreed in gender only with the target, and thus, participants could use gender information from the article to predict the target object. The adults with dyslexia anticipated the target objects but much later than the controls. Moreover, participants' word reading scores correlated positively with their anticipatory eye movements. We conclude by discussing the mechanisms by which reading abilities may influence predictive language processing. Copyright © 2015 John Wiley & Sons, Ltd.


    Directory of Open Access Journals (Sweden)

    Aris Novi


    Full Text Available Spoken text differs from written one in its features of context dependency, turn-taking organization, and dynamic structure. EFL learners; however, sometime find it difficult to produce typical characteristics of spoken language, particularly in casual talk. When they are asked to conduct a conversation, some of them tend to be script-based which is considered unnatural. Using the theory of Thornburry (2005, this paper aims to analyze characteristics of spoken language in casual conversation which cover spontaneity, interactivity, interpersonality, and coherence. This study used discourse analysis to reveal four features in turns and moves of three casual conversations. The findings indicate that not all sub-features used in the conversation. In this case, the spontaneity features were used 132 times; the interactivity features were used 1081 times; the interpersonality features were used 257 times; while the coherence features (negotiation features were used 526 times. Besides, the results also present that some participants seem to dominantly produce some sub-features naturally and vice versa. Therefore, this finding is expected to be beneficial to provide a model of how spoken interaction should be carried out. More importantly, it could raise English teachers or lecturers‘ awareness in teaching features of spoken language, so that, the students could develop their communicative competence as the native speakers of English do.

  6. Intervention Effects on Spoken-Language Outcomes for Children with Autism: A Systematic Review and Meta-Analysis (United States)

    Hampton, L. H.; Kaiser, A. P.


    Background: Although spoken-language deficits are not core to an autism spectrum disorder (ASD) diagnosis, many children with ASD do present with delays in this area. Previous meta-analyses have assessed the effects of intervention on reducing autism symptomatology, but have not determined if intervention improves spoken language. This analysis…

  7. Word reading skill predicts anticipation of upcoming spoken language input: a study of children developing proficiency in reading. (United States)

    Mani, Nivedita; Huettig, Falk


    Despite the efficiency with which language users typically process spoken language, a growing body of research finds substantial individual differences in both the speed and accuracy of spoken language processing potentially attributable to participants' literacy skills. Against this background, the current study took a look at the role of word reading skill in listeners' anticipation of upcoming spoken language input in children at the cusp of learning to read; if reading skills affect predictive language processing, then children at this stage of literacy acquisition should be most susceptible to the effects of reading skills on spoken language processing. We tested 8-year-olds on their prediction of upcoming spoken language input in an eye-tracking task. Although children, like in previous studies to date, were successfully able to anticipate upcoming spoken language input, there was a strong positive correlation between children's word reading skills (but not their pseudo-word reading and meta-phonological awareness or their spoken word recognition skills) and their prediction skills. We suggest that these findings are most compatible with the notion that the process of learning orthographic representations during reading acquisition sharpens pre-existing lexical representations, which in turn also supports anticipation of upcoming spoken words. Copyright © 2014 Elsevier Inc. All rights reserved.

  8. What Comes First, What Comes Next: Information Packaging in Written and Spoken Language

    Directory of Open Access Journals (Sweden)

    Vladislav Smolka


    Full Text Available The paper explores similarities and differences in the strategies of structuring information at sentence level in spoken and written language, respectively. In particular, it is concerned with the position of the rheme in the sentence in the two different modalities of language, and with the application and correlation of the end-focus and the end-weight principles. The assumption is that while there is a general tendency in both written and spoken language to place the focus in or close to the final position, owing to the limitations imposed by short-term memory capacity (and possibly by other factors, for the sake of easy processibility, it may occasionally be more felicitous in spoken language to place the rhematic element in the initial position or at least close to the beginning of the sentence. The paper aims to identify differences in the function of selected grammatical structures in written and spoken language, respectively, and to point out circumstances under which initial focus is a convenient alternative to the usual end-focus principle.

  9. Cognitive Predictors of Spoken Word Recognition in Children With and Without Developmental Language Disorders. (United States)

    Evans, Julia L; Gillam, Ronald B; Montgomery, James W


    This study examined the influence of cognitive factors on spoken word recognition in children with developmental language disorder (DLD) and typically developing (TD) children. Participants included 234 children (aged 7;0-11;11 years;months), 117 with DLD and 117 TD children, propensity matched for age, gender, socioeconomic status, and maternal education. Children completed a series of standardized assessment measures, a forward gating task, a rapid automatic naming task, and a series of tasks designed to examine cognitive factors hypothesized to influence spoken word recognition including phonological working memory, updating, attention shifting, and interference inhibition. Spoken word recognition for both initial and final accept gate points did not differ for children with DLD and TD controls after controlling target word knowledge in both groups. The 2 groups also did not differ on measures of updating, attention switching, and interference inhibition. Despite the lack of difference on these measures, for children with DLD, attention shifting and interference inhibition were significant predictors of spoken word recognition, whereas updating and receptive vocabulary were significant predictors of speed of spoken word recognition for the children in the TD group. Contrary to expectations, after controlling for target word knowledge, spoken word recognition did not differ for children with DLD and TD controls; however, the cognitive processing factors that influenced children's ability to recognize the target word in a stream of speech differed qualitatively for children with and without DLDs.


    Directory of Open Access Journals (Sweden)

    Eri Rusnawati


    Full Text Available Tujuan dari penelitian ini adalah untuk menggambarkan penerapan metode Communicative Language Teaching/CLT untuk pembelajaran spoken recount. Penelitian ini menelaah data yang kualitatif. Penelitian ini mengambarkan fenomena yang terjadi di dalam kelas. Data studi ini adalah perilaku dan respon para siswa dalam pembelajaran spoken recount dengan menggunakan metode CLT. Subjek penelitian ini adalah para siswa kelas X SMA Negeri 1 Kuaro yang terdiri dari 34 siswa. Observasi dan wawancara dilakukan dalam rangka untuk mengumpulkan data dalam mengajarkan spoken recount melalui tiga aktivitas (presentasi, bermain-peran, serta melakukan prosedur. Dalam penelitian ini ditemukan beberapa hal antara lain bahwa CLT meningkatkan kemampuan berbicara siswa dalam pembelajaran recount. Berdasarkan pada grafik peningkatan, disimpulkan bahwa tata bahasa, kosakata, pengucapan, kefasihan, serta performa siswa mengalami peningkatan. Ini berarti bahwa performa spoken recount dari para siswa meningkat. Andaikata presentasi ditempatkan di bagian akhir dari langkah-langkah aktivitas, peforma spoken recount para siswa bahkan akan lebih baik lagi. Kesimpulannya adalah bahwa implementasi metode CLT beserta tiga praktiknya berkontribusi pada peningkatan kemampuan berbicara para siswa dalam pembelajaran recount dan bahkan metode CLT mengarahkan mereka untuk memiliki keberanian dalam mengonstruksi komunikasi yang bermakna dengan percaya diri. Kata kunci: Communicative Language Teaching (CLT, recount, berbicara, respon siswa

  11. On-Line Syntax: Thoughts on the Temporality of Spoken Language (United States)

    Auer, Peter


    One fundamental difference between spoken and written language has to do with the "linearity" of speaking in time, in that the temporal structure of speaking is inherently the outcome of an interactive process between speaker and listener. But despite the status of "linearity" as one of Saussure's fundamental principles, in practice little more…

  12. Acquisition of graphic communication by a young girl without comprehension of spoken language. (United States)

    von Tetzchner, S; Øvreeide, K D; Jørgensen, K K; Ormhaug, B M; Oxholm, B; Warme, R

    To describe a graphic-mode communication intervention involving a girl with intellectual impairment and autism who did not develop comprehension of spoken language. The aim was to teach graphic-mode vocabulary that reflected her interests, preferences, and the activities and routines of her daily life, by providing sufficient cues to the meanings of the graphic representations so that she would not need to comprehend spoken instructions. An individual case study design was selected, including the use of written records, participant observation, and registration of the girl's graphic vocabulary and use of graphic signs and other communicative expressions. While the girl's comprehension (and hence use) of spoken language remained lacking over a 3-year period, she acquired an active use of over 80 photographs and pictograms. The girl was able to cope better with the cognitive and attentional requirements of graphic communication than those of spoken language and manual signs, which had been focused in earlier interventions. Her achievements demonstrate that it is possible for communication-impaired children to learn to use an augmentative and alternative communication system without speech comprehension, provided the intervention utilizes functional strategies and non-language cues to the meaning of the graphic representations that are taught.

  13. Expected Test Scores for Preschoolers with a Cochlear Implant Who Use Spoken Language (United States)

    Nicholas, Johanna G.; Geers, Ann E.


    Purpose: The major purpose of this study was to provide information about expected spoken language skills of preschool-age children who are deaf and who use a cochlear implant. A goal was to provide "benchmarks" against which those skills could be compared, for a given age at implantation. We also examined whether parent-completed…

  14. Personality Structure in the Trait Lexicon of Hindi, a Major Language Spoken in India

    NARCIS (Netherlands)

    Singh, Jitendra K.; Misra, Girishwar; De Raad, Boele


    The psycho-lexical approach is extended to Hindi, a major language spoken in India. From both the dictionary and from Hindi novels, a huge set of personality descriptors was put together, ultimately reduced to a manageable set of 295 trait terms. Both self and peer ratings were collected on those

  15. Developing and Testing EVALOE: A Tool for Assessing Spoken Language Teaching and Learning in the Classroom (United States)

    Gràcia, Marta; Vega, Fàtima; Galván-Bovaira, Maria José


    Broadly speaking, the teaching of spoken language in Spanish schools has not been approached in a systematic way. Changes in school practices are needed in order to allow all children to become competent speakers and to understand and construct oral texts that are appropriate in different contexts and for different audiences both inside and…

  16. A Spoken-Language Intervention for School-Aged Boys with Fragile X Syndrome (United States)

    McDuffie, Andrea; Machalicek, Wendy; Bullard, Lauren; Nelson, Sarah; Mello, Melissa; Tempero-Feigles, Robyn; Castignetti, Nancy; Abbeduto, Leonard


    Using a single case design, a parent-mediated spoken-language intervention was delivered to three mothers and their school-aged sons with fragile X syndrome, the leading inherited cause of intellectual disability. The intervention was embedded in the context of shared storytelling using wordless picture books and targeted three empirically derived…

  17. Phonotactic spoken language identification with limited training data

    CSIR Research Space (South Africa)

    Peche, M


    Full Text Available The authors investigate the addition of a new language, for which limited resources are available, to a phonotactic language identification system. Two classes of approaches are studied: in the first class, only existing phonetic recognizers...

  18. Cochlear implants and spoken language processing abilities: Review and assessment of the literature


    Peterson, Nathaniel R.; Pisoni, David B.; Miyamoto, Richard T.


    Cochlear implants (CIs) process sounds electronically and then transmit electric stimulation to the cochlea of individuals with sensorineural deafness, restoring some sensation of auditory perception. Many congenitally deaf CI recipients achieve a high degree of accuracy in speech perception and develop near-normal language skills. Post-lingually deafened implant recipients often regain the ability to understand and use spoken language with or without the aid of visual input (i.e. lip reading...

  19. The effects of sign language on spoken language acquisition in children with hearing loss: a systematic review protocol. (United States)

    Fitzpatrick, Elizabeth M; Stevens, Adrienne; Garritty, Chantelle; Moher, David


    Permanent childhood hearing loss affects 1 to 3 per 1000 children and frequently disrupts typical spoken language acquisition. Early identification of hearing loss through universal newborn hearing screening and the use of new hearing technologies including cochlear implants make spoken language an option for most children. However, there is no consensus on what constitutes optimal interventions for children when spoken language is the desired outcome. Intervention and educational approaches ranging from oral language only to oral language combined with various forms of sign language have evolved. Parents are therefore faced with important decisions in the first months of their child's life. This article presents the protocol for a systematic review of the effects of using sign language in combination with oral language intervention on spoken language acquisition. Studies addressing early intervention will be selected in which therapy involving oral language intervention and any form of sign language or sign support is used. Comparison groups will include children in early oral language intervention programs without sign support. The primary outcomes of interest to be examined include all measures of auditory, vocabulary, language, speech production, and speech intelligibility skills. We will include randomized controlled trials, controlled clinical trials, and other quasi-experimental designs that include comparator groups as well as prospective and retrospective cohort studies. Case-control, cross-sectional, case series, and case studies will be excluded. Several electronic databases will be searched (for example, MEDLINE, EMBASE, CINAHL, PsycINFO) as well as grey literature and key websites. We anticipate that a narrative synthesis of the evidence will be required. We will carry out meta-analysis for outcomes if clinical similarity, quantity and quality permit quantitative pooling of data. We will conduct subgroup analyses if possible according to severity

  20. Retinoic acid signaling: a new piece in the spoken language puzzle

    Directory of Open Access Journals (Sweden)

    Jon-Ruben eVan Rhijn


    Full Text Available Speech requires precise motor control and rapid sequencing of highly complex vocal musculature. Despite its complexity, most people produce spoken language effortlessly. This is due to activity in distributed neuronal circuitry including cortico-striato-thalamic loops that control speech-motor output. Understanding the neuro-genetic mechanisms that encode these pathways will shed light on how humans can effortlessly and innately use spoken language and could elucidate what goes wrong in speech-language disorders.FOXP2 was the first single gene identified to cause speech and language disorder. Individuals with FOXP2 mutations display a severe speech deficit that also includes receptive and expressive language impairments. The underlying neuro-molecular mechanisms controlled by FOXP2, which will give insight into our capacity for speech-motor control, are only beginning to be unraveled. Recently FOXP2 was found to regulate genes involved in retinoic acid signaling and to modify the cellular response to retinoic acid, a key regulator of brain development. Herein we explore the evidence that FOXP2 and retinoic acid signaling function in the same pathways. We present evidence at molecular, cellular and behavioral levels that suggest an interplay between FOXP2 and retinoic acid that may be important for fine motor control and speech-motor output. We propose that retinoic acid signaling is an exciting new angle from which to investigate how neurogenetic mechanisms can contribute to the (spoken language ready brain.

  1. ORIGINAL ARTICLES How do doctors learn the spoken language of ...

    African Journals Online (AJOL)


    Jul 1, 2009 ... and cultural metaphors of illness as part of language learning. The theory of .... role.21 Even in a military setting, where soldiers learnt Korean or Spanish as part of ... own language – a cross-cultural survey. Brit J Gen Pract ...

  2. Predictors of Spoken Language Development Following Pediatric Cochlear Implantation

    NARCIS (Netherlands)

    Johan Frijns; prof. Dr. Louis Peeraer; van Wieringen; Ingeborg Dhooge; Vermeulen; Jan Brokx; Tinne Boons; Wouters


    Objectives: Although deaf children with cochlear implants (CIs) are able to develop good language skills, the large variability in outcomes remains a significant concern. The first aim of this study was to evaluate language skills in children with CIs to establish benchmarks. The second aim was to

  3. Brain basis of phonological awareness for spoken language in children and its disruption in dyslexia. (United States)

    Kovelman, Ioulia; Norton, Elizabeth S; Christodoulou, Joanna A; Gaab, Nadine; Lieberman, Daniel A; Triantafyllou, Christina; Wolf, Maryanne; Whitfield-Gabrieli, Susan; Gabrieli, John D E


    Phonological awareness, knowledge that speech is composed of syllables and phonemes, is critical for learning to read. Phonological awareness precedes and predicts successful transition from language to literacy, and weakness in phonological awareness is a leading cause of dyslexia, but the brain basis of phonological awareness for spoken language in children is unknown. We used functional magnetic resonance imaging to identify the neural correlates of phonological awareness using an auditory word-rhyming task in children who were typical readers or who had dyslexia (ages 7-13) and a younger group of kindergarteners (ages 5-6). Typically developing children, but not children with dyslexia, recruited left dorsolateral prefrontal cortex (DLPFC) when making explicit phonological judgments. Kindergarteners, who were matched to the older children with dyslexia on standardized tests of phonological awareness, also recruited left DLPFC. Left DLPFC may play a critical role in the development of phonological awareness for spoken language critical for reading and in the etiology of dyslexia.

  4. How sensory-motor systems impact the neural organization for language: direct contrasts between spoken and signed language (United States)

    Emmorey, Karen; McCullough, Stephen; Mehta, Sonya; Grabowski, Thomas J.


    To investigate the impact of sensory-motor systems on the neural organization for language, we conducted an H215O-PET study of sign and spoken word production (picture-naming) and an fMRI study of sign and audio-visual spoken language comprehension (detection of a semantically anomalous sentence) with hearing bilinguals who are native users of American Sign Language (ASL) and English. Directly contrasting speech and sign production revealed greater activation in bilateral parietal cortex for signing, while speaking resulted in greater activation in bilateral superior temporal cortex (STC) and right frontal cortex, likely reflecting auditory feedback control. Surprisingly, the language production contrast revealed a relative increase in activation in bilateral occipital cortex for speaking. We speculate that greater activation in visual cortex for speaking may actually reflect cortical attenuation when signing, which functions to distinguish self-produced from externally generated visual input. Directly contrasting speech and sign comprehension revealed greater activation in bilateral STC for speech and greater activation in bilateral occipital-temporal cortex for sign. Sign comprehension, like sign production, engaged bilateral parietal cortex to a greater extent than spoken language. We hypothesize that posterior parietal activation in part reflects processing related to spatial classifier constructions in ASL and that anterior parietal activation may reflect covert imitation that functions as a predictive model during sign comprehension. The conjunction analysis for comprehension revealed that both speech and sign bilaterally engaged the inferior frontal gyrus (with more extensive activation on the left) and the superior temporal sulcus, suggesting an invariant bilateral perisylvian language system. We conclude that surface level differences between sign and spoken languages should not be dismissed and are critical for understanding the neurobiology of language

  5. Spoken language identification based on the enhanced self-adjusting extreme learning machine approach (United States)

    Tiun, Sabrina; AL-Dhief, Fahad Taha; Sammour, Mahmoud A. M.


    Spoken Language Identification (LID) is the process of determining and classifying natural language from a given content and dataset. Typically, data must be processed to extract useful features to perform LID. The extracting features for LID, based on literature, is a mature process where the standard features for LID have already been developed using Mel-Frequency Cepstral Coefficients (MFCC), Shifted Delta Cepstral (SDC), the Gaussian Mixture Model (GMM) and ending with the i-vector based framework. However, the process of learning based on extract features remains to be improved (i.e. optimised) to capture all embedded knowledge on the extracted features. The Extreme Learning Machine (ELM) is an effective learning model used to perform classification and regression analysis and is extremely useful to train a single hidden layer neural network. Nevertheless, the learning process of this model is not entirely effective (i.e. optimised) due to the random selection of weights within the input hidden layer. In this study, the ELM is selected as a learning model for LID based on standard feature extraction. One of the optimisation approaches of ELM, the Self-Adjusting Extreme Learning Machine (SA-ELM) is selected as the benchmark and improved by altering the selection phase of the optimisation process. The selection process is performed incorporating both the Split-Ratio and K-Tournament methods, the improved SA-ELM is named Enhanced Self-Adjusting Extreme Learning Machine (ESA-ELM). The results are generated based on LID with the datasets created from eight different languages. The results of the study showed excellent superiority relating to the performance of the Enhanced Self-Adjusting Extreme Learning Machine LID (ESA-ELM LID) compared with the SA-ELM LID, with ESA-ELM LID achieving an accuracy of 96.25%, as compared to the accuracy of SA-ELM LID of only 95.00%. PMID:29672546

  6. Spoken language identification based on the enhanced self-adjusting extreme learning machine approach. (United States)

    Albadr, Musatafa Abbas Abbood; Tiun, Sabrina; Al-Dhief, Fahad Taha; Sammour, Mahmoud A M


    Spoken Language Identification (LID) is the process of determining and classifying natural language from a given content and dataset. Typically, data must be processed to extract useful features to perform LID. The extracting features for LID, based on literature, is a mature process where the standard features for LID have already been developed using Mel-Frequency Cepstral Coefficients (MFCC), Shifted Delta Cepstral (SDC), the Gaussian Mixture Model (GMM) and ending with the i-vector based framework. However, the process of learning based on extract features remains to be improved (i.e. optimised) to capture all embedded knowledge on the extracted features. The Extreme Learning Machine (ELM) is an effective learning model used to perform classification and regression analysis and is extremely useful to train a single hidden layer neural network. Nevertheless, the learning process of this model is not entirely effective (i.e. optimised) due to the random selection of weights within the input hidden layer. In this study, the ELM is selected as a learning model for LID based on standard feature extraction. One of the optimisation approaches of ELM, the Self-Adjusting Extreme Learning Machine (SA-ELM) is selected as the benchmark and improved by altering the selection phase of the optimisation process. The selection process is performed incorporating both the Split-Ratio and K-Tournament methods, the improved SA-ELM is named Enhanced Self-Adjusting Extreme Learning Machine (ESA-ELM). The results are generated based on LID with the datasets created from eight different languages. The results of the study showed excellent superiority relating to the performance of the Enhanced Self-Adjusting Extreme Learning Machine LID (ESA-ELM LID) compared with the SA-ELM LID, with ESA-ELM LID achieving an accuracy of 96.25%, as compared to the accuracy of SA-ELM LID of only 95.00%.

  7. The interface between spoken and written language: developmental disorders. (United States)

    Hulme, Charles; Snowling, Margaret J


    We review current knowledge about reading development and the origins of difficulties in learning to read. We distinguish between the processes involved in learning to decode print, and the processes involved in reading for meaning (reading comprehension). At a cognitive level, difficulties in learning to read appear to be predominantly caused by deficits in underlying oral language skills. The development of decoding skills appears to depend critically upon phonological language skills, and variations in phoneme awareness, letter-sound knowledge and rapid automatized naming each appear to be causally related to problems in learning to read. Reading comprehension difficulties in contrast appear to be critically dependent on a range of oral language comprehension skills (including vocabulary knowledge and grammatical, morphological and pragmatic skills).

  8. Loops of Spoken Language i Danish Broadcasting Corporation News

    DEFF Research Database (Denmark)

    le Fevre Jakobsen, Bjarne


    The tempo of Danish television news broadcasts has changed markedly over the past 40 years, while the language has essentially always been conservative, and remains so today. The development in the tempo of the broadcasts has gone through a number of phases from a newsreader in a rigid structure...


    African Journals Online (AJOL)

    In the face of globalisation, the scale of communication is increasing from being merely .... capital goods and services across national frontiers involving too, political contexts of ... auditory and audiovisual entertainment, the use of English dominates. The language .... manners, entertainment, sports, the legal system, etc.

  10. Three-dimensional grammar in the brain: Dissociating the neural correlates of natural sign language and manually coded spoken language. (United States)

    Jednoróg, Katarzyna; Bola, Łukasz; Mostowski, Piotr; Szwed, Marcin; Boguszewski, Paweł M; Marchewka, Artur; Rutkowski, Paweł


    In several countries natural sign languages were considered inadequate for education. Instead, new sign-supported systems were created, based on the belief that spoken/written language is grammatically superior. One such system called SJM (system językowo-migowy) preserves the grammatical and lexical structure of spoken Polish and since 1960s has been extensively employed in schools and on TV. Nevertheless, the Deaf community avoids using SJM for everyday communication, its preferred language being PJM (polski język migowy), a natural sign language, structurally and grammatically independent of spoken Polish and featuring classifier constructions (CCs). Here, for the first time, we compare, with fMRI method, the neural bases of natural vs. devised communication systems. Deaf signers were presented with three types of signed sentences (SJM and PJM with/without CCs). Consistent with previous findings, PJM with CCs compared to either SJM or PJM without CCs recruited the parietal lobes. The reverse comparison revealed activation in the anterior temporal lobes, suggesting increased semantic combinatory processes in lexical sign comprehension. Finally, PJM compared with SJM engaged left posterior superior temporal gyrus and anterior temporal lobe, areas crucial for sentence-level speech comprehension. We suggest that activity in these two areas reflects greater processing efficiency for naturally evolved sign language. Copyright © 2015 Elsevier Ltd. All rights reserved.

  11. A step beyond local observations with a dialog aware bidirectional GRU network for Spoken Language Understanding


    Vukotic , Vedran; Raymond , Christian; Gravier , Guillaume


    International audience; Architectures of Recurrent Neural Networks (RNN) recently become a very popular choice for Spoken Language Understanding (SLU) problems; however, they represent a big family of different architectures that can furthermore be combined to form more complex neural networks. In this work, we compare different recurrent networks, such as simple Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM) networks, Gated Memory Units (GRU) and their bidirectional versions,...

  12. Verbal short-term memory development and spoken language outcomes in deaf children with cochlear implants. (United States)

    Harris, Michael S; Kronenberger, William G; Gao, Sujuan; Hoen, Helena M; Miyamoto, Richard T; Pisoni, David B


    Cochlear implants (CIs) help many deaf children achieve near-normal speech and language (S/L) milestones. Nevertheless, high levels of unexplained variability in S/L outcomes are limiting factors in improving the effectiveness of CIs in deaf children. The objective of this study was to longitudinally assess the role of verbal short-term memory (STM) and working memory (WM) capacity as a progress-limiting source of variability in S/L outcomes after CI in children. Longitudinal study of 66 children with CIs for prelingual severe-to-profound hearing loss. Outcome measures included performance on digit span forward (DSF), digit span backward (DSB), and four conventional S/L measures that examined spoken-word recognition (Phonetically Balanced Kindergarten word test), receptive vocabulary (Peabody Picture Vocabulary Test ), sentence-recognition skills (Hearing in Noise Test), and receptive and expressive language functioning (Clinical Evaluation of Language Fundamentals Fourth Edition Core Language Score; CELF). Growth curves for DSF and DSB in the CI sample over time were comparable in slope, but consistently lagged in magnitude relative to norms for normal-hearing peers of the same age. For DSF and DSB, 50.5% and 44.0%, respectively, of the CI sample scored more than 1 SD below the normative mean for raw scores across all ages. The first (baseline) DSF score significantly predicted all endpoint scores for the four S/L measures, and DSF slope (growth) over time predicted CELF scores. DSF baseline and slope accounted for an additional 13 to 31% of variance in S/L scores after controlling for conventional predictor variables such as: chronological age at time of testing, age at time of implantation, communication mode (auditory-oral communication versus total communication), and maternal education. Only DSB baseline scores predicted endpoint language scores on Peabody Picture Vocabulary Test and CELF. DSB slopes were not significantly related to any endpoint S/L measures

  13. Spoken Language Activation Alters Subsequent Sign Language Activation in L2 Learners of American Sign Language (United States)

    Williams, Joshua T.; Newman, Sharlene D.


    A large body of literature has characterized unimodal monolingual and bilingual lexicons and how neighborhood density affects lexical access; however there have been relatively fewer studies that generalize these findings to bimodal (M2) second language (L2) learners of sign languages. The goal of the current study was to investigate parallel…

  14. The Attitudes and Motivation of Children towards Learning Rarely Spoken Foreign Languages: A Case Study from Saudi Arabia (United States)

    Al-Nofaie, Haifa


    This article discusses the attitudes and motivations of two Saudi children learning Japanese as a foreign language (hence JFL), a language which is rarely spoken in the country. Studies regarding children's motivation for learning foreign languages that are not widely spread in their contexts in informal settings are scarce. The aim of the study…

  15. The missing foundation in teacher education: Knowledge of the structure of spoken and written language. (United States)

    Moats, L C


    Reading research supports the necessity for directly teaching concepts about linguistic structure to beginning readers and to students with reading and spelling difficulties. In this study, experienced teachers of reading, language arts, and special education were tested to determine if they have the requisite awareness of language elements (e.g., phonemes, morphemes) and of how these elements are represented in writing (e.g., knowledge of sound-symbol correspondences). The results were surprisingly poor, indicating that even motivated and experienced teachers typically understand too little about spoken and written language structure to be able to provide sufficient instruction in these areas. The utility of language structure knowledge for instructional planning, for assessment of student progress, and for remediation of literacy problems is discussed.The teachers participating in the study subsequently took a course focusing on phonemic awareness training, spoken-written language relationships, and careful analysis of spelling and reading behavior in children. At the end of the course, the teachers judged this information to be essential for teaching and advised that it become a prerequisite for certification. Recommendations for requirements and content of teacher education programs are presented.

  16. Phonologic-graphemic transcodifier for Portuguese Language spoken in Brazil (PLB) (United States)

    Fragadasilva, Francisco Jose; Saotome, Osamu; Deoliveira, Carlos Alberto

    An automatic speech-to-text transformer system, suited to unlimited vocabulary, is presented. The basic acoustic unit considered are the allophones of the phonemes corresponding to the Portuguese language spoken in Brazil (PLB). The input to the system is a phonetic sequence, from a former step of isolated word recognition of slowly spoken speech. In a first stage, the system eliminates phonetic elements that don't belong to PLB. Using knowledge sources such as phonetics, phonology, orthography, and PLB specific lexicon, the output is a sequence of written words, ordered by probabilistic criterion that constitutes the set of graphemic possibilities to that input sequence. Pronunciation differences of some regions of Brazil are considered, but only those that cause differences in phonological transcription, because those of phonetic level are absorbed, during the transformation to phonological level. In the final stage, all possible written words are analyzed for orthography and grammar point of view, to eliminate the incorrect ones.

  17. The Beneficial Role of L1 Spoken Language Skills on Initial L2 Sign Language Learning: Cognitive and Linguistic Predictors of M2L2 Acquisition (United States)

    Williams, Joshua T.; Darcy, Isabelle; Newman, Sharlene D.


    Understanding how language modality (i.e., signed vs. spoken) affects second language outcomes in hearing adults is important both theoretically and pedagogically, as it can determine the specificity of second language (L2) theory and inform how best to teach a language that uses a new modality. The present study investigated which…

  18. Porting a spoken language identification systen to a new environment.

    CSIR Research Space (South Africa)

    Peche, M


    Full Text Available A speech processing system is often required to perform in a different environment than the one for which it was initially developed. In such a case, data from the new environment may be more limited in quantity and of poorer quality than...

  19. Spoken language identification system adaptation in under-resourced environments

    CSIR Research Space (South Africa)

    Kleynhans, N


    Full Text Available Speech Recognition (ASR) systems in the developing world is severely inhibited. Given that few task-specific corpora exist and speech technology systems perform poorly when deployed in a new environment, we investigate the use of acoustic model adaptation...

  20. Spoken language development in oral preschool children with permanent childhood deafness. (United States)

    Sarant, Julia Z; Holt, Colleen M; Dowell, Richard C; Rickards, Field W; Blamey, Peter J


    This article documented spoken language outcomes for preschool children with hearing loss and examined the relationships between language abilities and characteristics of children such as degree of hearing loss, cognitive abilities, age at entry to early intervention, and parent involvement in children's intervention programs. Participants were evaluated using a combination of the Child Development Inventory, the Peabody Picture Vocabulary Test, and the Preschool Clinical Evaluation of Language Fundamentals depending on their age at the time of assessment. Maternal education, cognitive ability, and family involvement were also measured. Over half of the children who participated in this study had poor language outcomes overall. No significant differences were found in language outcomes on any of the measures for children who were diagnosed early and those diagnosed later. Multiple regression analyses showed that family participation, degree of hearing loss, and cognitive ability significantly predicted language outcomes and together accounted for almost 60% of the variance in scores. This article highlights the importance of family participation in intervention programs to enable children to achieve optimal language outcomes. Further work may clarify the effects of early diagnosis on language outcomes for preschool children.

  1. Spoken language achieves robustness and evolvability by exploiting degeneracy and neutrality. (United States)

    Winter, Bodo


    As with biological systems, spoken languages are strikingly robust against perturbations. This paper shows that languages achieve robustness in a way that is highly similar to many biological systems. For example, speech sounds are encoded via multiple acoustically diverse, temporally distributed and functionally redundant cues, characteristics that bear similarities to what biologists call "degeneracy". Speech is furthermore adequately characterized by neutrality, with many different tongue configurations leading to similar acoustic outputs, and different acoustic variants understood as the same by recipients. This highlights the presence of a large neutral network of acoustic neighbors for every speech sound. Such neutrality ensures that a steady backdrop of variation can be maintained without impeding communication, assuring that there is "fodder" for subsequent evolution. Thus, studying linguistic robustness is not only important for understanding how linguistic systems maintain their functioning upon the background of noise, but also for understanding the preconditions for language evolution. © 2014 WILEY Periodicals, Inc.

  2. Positive Emotional Language in the Final Words Spoken Directly Before Execution. (United States)

    Hirschmüller, Sarah; Egloff, Boris


    How do individuals emotionally cope with the imminent real-world salience of mortality? DeWall and Baumeister as well as Kashdan and colleagues previously provided support that an increased use of positive emotion words serves as a way to protect and defend against mortality salience of one's own contemplated death. Although these studies provide important insights into the psychological dynamics of mortality salience, it remains an open question how individuals cope with the immense threat of mortality prior to their imminent actual death. In the present research, we therefore analyzed positivity in the final words spoken immediately before execution by 407 death row inmates in Texas. By using computerized quantitative text analysis as an objective measure of emotional language use, our results showed that the final words contained a significantly higher proportion of positive than negative emotion words. This emotional positivity was significantly higher than (a) positive emotion word usage base rates in spoken and written materials and (b) positive emotional language use with regard to contemplated death and attempted or actual suicide. Additional analyses showed that emotional positivity in final statements was associated with a greater frequency of language use that was indicative of self-references, social orientation, and present-oriented time focus as well as with fewer instances of cognitive-processing, past-oriented, and death-related word use. Taken together, our findings offer new insights into how individuals cope with the imminent real-world salience of mortality.

  3. Activating gender stereotypes during online spoken language processing: evidence from Visual World Eye Tracking. (United States)

    Pyykkönen, Pirita; Hyönä, Jukka; van Gompel, Roger P G


    This study used the visual world eye-tracking method to investigate activation of general world knowledge related to gender-stereotypical role names in online spoken language comprehension in Finnish. The results showed that listeners activated gender stereotypes elaboratively in story contexts where this information was not needed to build coherence. Furthermore, listeners made additional inferences based on gender stereotypes to revise an already established coherence relation. Both results are consistent with mental models theory (e.g., Garnham, 2001). They are harder to explain by the minimalist account (McKoon & Ratcliff, 1992) which suggests that people limit inferences to those needed to establish coherence in discourse.

  4. Students who are deaf and hard of hearing and use sign language: considerations and strategies for developing spoken language and literacy skills. (United States)

    Nussbaum, Debra; Waddy-Smith, Bettie; Doyle, Jane


    There is a core body of knowledge, experience, and skills integral to facilitating auditory, speech, and spoken language development when working with the general population of students who are deaf and hard of hearing. There are additional issues, strategies, and challenges inherent in speech habilitation/rehabilitation practices essential to the population of deaf and hard of hearing students who also use sign language. This article will highlight philosophical and practical considerations related to practices used to facilitate spoken language development and associated literacy skills for children and adolescents who sign. It will discuss considerations for planning and implementing practices that acknowledge and utilize a student's abilities in sign language, and address how to link these skills to developing and using spoken language. Included will be considerations for children from early childhood through high school with a broad range of auditory access, language, and communication characteristics. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  5. Let's all speak together! Exploring the masking effects of various languages on spoken word identification in multi-linguistic babble. (United States)

    Gautreau, Aurore; Hoen, Michel; Meunier, Fanny


    This study aimed to characterize the linguistic interference that occurs during speech-in-speech comprehension by combining offline and online measures, which included an intelligibility task (at a -5 dB Signal-to-Noise Ratio) and 2 lexical decision tasks (at a -5 dB and 0 dB SNR) that were performed with French spoken target words. In these 3 experiments we always compared the masking effects of speech backgrounds (i.e., 4-talker babble) that were produced in the same language as the target language (i.e., French) or in unknown foreign languages (i.e., Irish and Italian) to the masking effects of corresponding non-speech backgrounds (i.e., speech-derived fluctuating noise). The fluctuating noise contained similar spectro-temporal information as babble but lacked linguistic information. At -5 dB SNR, both tasks revealed significantly divergent results between the unknown languages (i.e., Irish and Italian) with Italian and French hindering French target word identification to a similar extent, whereas Irish led to significantly better performances on these tasks. By comparing the performances obtained with speech and fluctuating noise backgrounds, we were able to evaluate the effect of each language. The intelligibility task showed a significant difference between babble and fluctuating noise for French, Irish and Italian, suggesting acoustic and linguistic effects for each language. However, the lexical decision task, which reduces the effect of post-lexical interference, appeared to be more accurate, as it only revealed a linguistic effect for French. Thus, although French and Italian had equivalent masking effects on French word identification, the nature of their interference was different. This finding suggests that the differences observed between the masking effects of Italian and Irish can be explained at an acoustic level but not at a linguistic level.

  6. Effects of early auditory experience on the spoken language of deaf children at 3 years of age. (United States)

    Nicholas, Johanna Grant; Geers, Ann E


    By age 3, typically developing children have achieved extensive vocabulary and syntax skills that facilitate both cognitive and social development. Substantial delays in spoken language acquisition have been documented for children with severe to profound deafness, even those with auditory oral training and early hearing aid use. This study documents the spoken language skills achieved by orally educated 3-yr-olds whose profound hearing loss was identified and hearing aids fitted between 1 and 30 mo of age and who received a cochlear implant between 12 and 38 mo of age. The purpose of the analysis was to examine the effects of age, duration, and type of early auditory experience on spoken language competence at age 3.5 yr. The spoken language skills of 76 children who had used a cochlear implant for at least 7 mo were evaluated via standardized 30-minute language sample analysis, a parent-completed vocabulary checklist, and a teacher language-rating scale. The children were recruited from and enrolled in oral education programs or therapy practices across the United States. Inclusion criteria included presumed deaf since birth, English the primary language of the home, no other known conditions that interfere with speech/language development, enrolled in programs using oral education methods, and no known problems with the cochlear implant lasting more than 30 days. Strong correlations were obtained among all language measures. Therefore, principal components analysis was used to derive a single Language Factor score for each child. A number of possible predictors of language outcome were examined, including age at identification and intervention with a hearing aid, duration of use of a hearing aid, pre-implant pure-tone average (PTA) threshold with a hearing aid, PTA threshold with a cochlear implant, and duration of use of a cochlear implant/age at implantation (the last two variables were practically identical because all children were tested between 40 and 44

  7. Cochlear implants and spoken language processing abilities: review and assessment of the literature. (United States)

    Peterson, Nathaniel R; Pisoni, David B; Miyamoto, Richard T


    Cochlear implants (CIs) process sounds electronically and then transmit electric stimulation to the cochlea of individuals with sensorineural deafness, restoring some sensation of auditory perception. Many congenitally deaf CI recipients achieve a high degree of accuracy in speech perception and develop near-normal language skills. Post-lingually deafened implant recipients often regain the ability to understand and use spoken language with or without the aid of visual input (i.e. lip reading). However, there is wide variation in individual outcomes following cochlear implantation, and some CI recipients never develop useable speech and oral language skills. The causes of this enormous variation in outcomes are only partly understood at the present time. The variables most strongly associated with language outcomes are age at implantation and mode of communication in rehabilitation. Thus, some of the more important factors determining success of cochlear implantation are broadly related to neural plasticity that appears to be transiently present in deaf individuals. In this article we review the expected outcomes of cochlear implantation, potential predictors of those outcomes, the basic science regarding critical and sensitive periods, and several new research directions in the field of cochlear implantation.

  8. Semantic Richness and Word Learning in Children with Hearing Loss Who Are Developing Spoken Language: A Single Case Design Study (United States)

    Lund, Emily; Douglas, W. Michael; Schuele, C. Melanie


    Children with hearing loss who are developing spoken language tend to lag behind children with normal hearing in vocabulary knowledge. Thus, researchers must validate instructional practices that lead to improved vocabulary outcomes for children with hearing loss. The purpose of this study was to investigate how semantic richness of instruction…

  9. Rethinking spoken fluency


    McCarthy, Michael


    This article re-examines the notion of spoken fluency. Fluent and fluency are terms commonly used in everyday, lay language, and fluency, or lack of it, has social consequences. The article reviews the main approaches to understanding and measuring spoken fluency and suggest that spoken fluency is best understood as an interactive achievement, and offers the metaphor of ‘confluence’ to replace the term fluency. Many measures of spoken fluency are internal and monologue-based, whereas evidence...

  10. Phonological Interference in the Spoken English Performance of the ...

    African Journals Online (AJOL)

    ... An International Journal of Language, Literature and Gender Studies ... as a result of the systemic differences that exist between both language systems (Izon ... and pay particular attention to the differences in the phonological systems of the ...

  11. Foreign body aspiration and language spoken at home: 10-year review. (United States)

    Choroomi, S; Curotta, J


    To review foreign body aspiration cases encountered over a 10-year period in a tertiary paediatric hospital, and to assess correlation between foreign body type and language spoken at home. Retrospective chart review of all children undergoing direct laryngobronchoscopy for foreign body aspiration over a 10-year period. Age, sex, foreign body type, complications, hospital stay and home language were analysed. At direct laryngobronchoscopy, 132 children had foreign body aspiration (male:female ratio 1.31:1; mean age 32 months (2.67 years)). Mean hospital stay was 2.0 days. Foreign bodies most commonly comprised food matter (53/132; 40.1 per cent), followed by non-food matter (44/132; 33.33 per cent), a negative endoscopy (11/132; 8.33 per cent) and unknown composition (24/132; 18.2 per cent). Most parents spoke English (92/132, 69.7 per cent; vs non-English-speaking 40/132, 30.3 per cent), but non-English-speaking patients had disproportionately more food foreign bodies, and significantly more nut aspirations (p = 0.0065). Results constitute level 2b evidence. Patients from non-English speaking backgrounds had a significantly higher incidence of food (particularly nut) aspiration. Awareness-raising and public education is needed in relevant communities to prevent certain foods, particularly nuts, being given to children too young to chew and swallow them adequately.

  12. Primary phonological planning units in spoken word production are language-specific: Evidence from an ERP study. (United States)

    Wang, Jie; Wong, Andus Wing-Kuen; Wang, Suiping; Chen, Hsuan-Chih


    It is widely acknowledged in Germanic languages that segments are the primary planning units at the phonological encoding stage of spoken word production. Mixed results, however, have been found in Chinese, and it is still unclear what roles syllables and segments play in planning Chinese spoken word production. In the current study, participants were asked to first prepare and later produce disyllabic Mandarin words upon picture prompts and a response cue while electroencephalogram (EEG) signals were recorded. Each two consecutive pictures implicitly formed a pair of prime and target, whose names shared the same word-initial atonal syllable or the same word-initial segments, or were unrelated in the control conditions. Only syllable repetition induced significant effects on event-related brain potentials (ERPs) after target onset: a widely distributed positivity in the 200- to 400-ms interval and an anterior positivity in the 400- to 600-ms interval. We interpret these to reflect syllable-size representations at the phonological encoding and phonetic encoding stages. Our results provide the first electrophysiological evidence for the distinct role of syllables in producing Mandarin spoken words, supporting a language specificity hypothesis about the primary phonological units in spoken word production.

  13. Emergent Literacy Skills in Preschool Children With Hearing Loss Who Use Spoken Language: Initial Findings From the Early Language and Literacy Acquisition (ELLA) Study. (United States)

    Werfel, Krystal L


    The purpose of this study was to compare change in emergent literacy skills of preschool children with and without hearing loss over a 6-month period. Participants included 19 children with hearing loss and 14 children with normal hearing. Children with hearing loss used amplification and spoken language. Participants completed measures of oral language, phonological processing, and print knowledge twice at a 6-month interval. A series of repeated-measures analyses of variance were used to compare change across groups. Main effects of time were observed for all variables except phonological recoding. Main effects of group were observed for vocabulary, morphosyntax, phonological memory, and concepts of print. Interaction effects were observed for phonological awareness and concepts of print. Children with hearing loss performed more poorly than children with normal hearing on measures of oral language, phonological memory, and conceptual print knowledge. Two interaction effects were present. For phonological awareness and concepts of print, children with hearing loss demonstrated less positive change than children with normal hearing. Although children with hearing loss generally demonstrated a positive growth in emergent literacy skills, their initial performance was lower than that of children with normal hearing, and rates of change were not sufficient to catch up to the peers over time.

  14. Emergent Literacy Skills in Preschool Children with Hearing Loss Who Use Spoken Language: Initial Findings from the Early Language and Literacy Acquisition (ELLA) Study (United States)

    Werfel, Krystal L.


    Purpose: The purpose of this study was to compare change in emergent literacy skills of preschool children with and without hearing loss over a 6-month period. Method: Participants included 19 children with hearing loss and 14 children with normal hearing. Children with hearing loss used amplification and spoken language. Participants completed…

  15. Identification of four class emotion from Indonesian spoken language using acoustic and lexical features (United States)

    Kasyidi, Fatan; Puji Lestari, Dessi


    One of the important aspects in human to human communication is to understand emotion of each party. Recently, interactions between human and computer continues to develop, especially affective interaction where emotion recognition is one of its important components. This paper presents our extended works on emotion recognition of Indonesian spoken language to identify four main class of emotions: Happy, Sad, Angry, and Contentment using combination of acoustic/prosodic features and lexical features. We construct emotion speech corpus from Indonesia television talk show where the situations are as close as possible to the natural situation. After constructing the emotion speech corpus, the acoustic/prosodic and lexical features are extracted to train the emotion model. We employ some machine learning algorithms such as Support Vector Machine (SVM), Naive Bayes, and Random Forest to get the best model. The experiment result of testing data shows that the best model has an F-measure score of 0.447 by using only the acoustic/prosodic feature and F-measure score of 0.488 by using both acoustic/prosodic and lexical features to recognize four class emotion using the SVM RBF Kernel.

  16. Impact of cognitive function and dysarthria on spoken language and perceived speech severity in multiple sclerosis (United States)

    Feenaughty, Lynda

    Purpose: The current study sought to investigate the separate effects of dysarthria and cognitive status on global speech timing, speech hesitation, and linguistic complexity characteristics and how these speech behaviors impose on listener impressions for three connected speech tasks presumed to differ in cognitive-linguistic demand for four carefully defined speaker groups; 1) MS with cognitive deficits (MSCI), 2) MS with clinically diagnosed dysarthria and intact cognition (MSDYS), 3) MS without dysarthria or cognitive deficits (MS), and 4) healthy talkers (CON). The relationship between neuropsychological test scores and speech-language production and perceptual variables for speakers with cognitive deficits was also explored. Methods: 48 speakers, including 36 individuals reporting a neurological diagnosis of MS and 12 healthy talkers participated. The three MS groups and control group each contained 12 speakers (8 women and 4 men). Cognitive function was quantified using standard clinical tests of memory, information processing speed, and executive function. A standard z-score of ≤ -1.50 indicated deficits in a given cognitive domain. Three certified speech-language pathologists determined the clinical diagnosis of dysarthria for speakers with MS. Experimental speech tasks of interest included audio-recordings of an oral reading of the Grandfather passage and two spontaneous speech samples in the form of Familiar and Unfamiliar descriptive discourse. Various measures of spoken language were of interest. Suprasegmental acoustic measures included speech and articulatory rate. Linguistic speech hesitation measures included pause frequency (i.e., silent and filled pauses), mean silent pause duration, grammatical appropriateness of pauses, and interjection frequency. For the two discourse samples, three standard measures of language complexity were obtained including subordination index, inter-sentence cohesion adequacy, and lexical diversity. Ten listeners


    Directory of Open Access Journals (Sweden)

    Leni Amalia Suek


    Full Text Available The maintenance of community languages of migrant students is heavily determined by language use and language attitudes. The superiority of a dominant language over a community language contributes to attitudes of migrant students toward their native languages. When they perceive their native languages as unimportant language, they will reduce the frequency of using that language even though at home domain. Solutions provided for a problem of maintaining community languages should be related to language use and attitudes of community languages, which are developed mostly in two important domains, school and family. Hence, the valorization of community language should be promoted not only in family but also school domains. Several programs such as community language school and community language program can be used for migrant students to practice and use their native languages. Since educational resources such as class session, teachers and government support are limited; family plays significant roles to stimulate positive attitudes toward community language and also to develop the use of native languages.

  18. Birds, primates, and spoken language origins: behavioral phenotypes and neurobiological substrates. (United States)

    Petkov, Christopher I; Jarvis, Erich D


    Vocal learners such as humans and songbirds can learn to produce elaborate patterns of structurally organized vocalizations, whereas many other vertebrates such as non-human primates and most other bird groups either cannot or do so to a very limited degree. To explain the similarities among humans and vocal-learning birds and the differences with other species, various theories have been proposed. One set of theories are motor theories, which underscore the role of the motor system as an evolutionary substrate for vocal production learning. For instance, the motor theory of speech and song perception proposes enhanced auditory perceptual learning of speech in humans and song in birds, which suggests a considerable level of neurobiological specialization. Another, a motor theory of vocal learning origin, proposes that the brain pathways that control the learning and production of song and speech were derived from adjacent motor brain pathways. Another set of theories are cognitive theories, which address the interface between cognition and the auditory-vocal domains to support language learning in humans. Here we critically review the behavioral and neurobiological evidence for parallels and differences between the so-called vocal learners and vocal non-learners in the context of motor and cognitive theories. In doing so, we note that behaviorally vocal-production learning abilities are more distributed than categorical, as are the auditory-learning abilities of animals. We propose testable hypotheses on the extent of the specializations and cross-species correspondences suggested by motor and cognitive theories. We believe that determining how spoken language evolved is likely to become clearer with concerted efforts in testing comparative data from many non-human animal species.


    Directory of Open Access Journals (Sweden)

    HOREA Ioana-Claudia


    Full Text Available Spoken English may sometimes cause us to face a peculiar problem in respect of the reception and the decoding of auditive signals, which might lead to mishearings. Risen from erroneous perception, from a lack in understanding the communication and an involuntary mental replacement of a certain element or structure by a more familiar one, these mistakes are most frequently encountered in the case of listening to songs, where the melodic line can facilitate the development of confusion by its somewhat altered intonation, which produces the so called mondegreens. Still, instances can be met in all domains of verbal communication, as proven in several examples noticed during classes of English as a foreign language (EFL taught to non-philological subjects. Production and perceptions of language depend on a series of elements that influence the encoding and the decoding of the message. These filters belong to both psychological and semantic categories which can either interfere with the accuracy of emission and reception. Poor understanding of a notion or concept combined with a more familiar relation with a similarly sounding one will result in unconsciously picking the structure which is better known. This means ‘hearing’ something else than it had been said, something closer to the receiver’s preoccupations and baggage of knowledge than the original structure or word. Some mishearings become particularly relevant as they concern teaching English for Specific Purposes (ESP. Such are those encountered during classes of Business English or in English for Law. Though not very likely to occur too often, given an intuitively felt inaccuracy - as the terms are known by the users to need to be more specialised -, such examples are still not ignorable. Thus, we consider they deserve a higher degree of attention, as they might become quite relevant in the global context of an increasing work force migration and a spread of multinational companies.

  20. Development of a spoken language identification system for South African languages

    CSIR Research Space (South Africa)

    Peché, M


    Full Text Available , and complicates the design of the system as a whole. Current benchmark results are established by the National Institute of Standards and Technology (NIST) Language Recognition Evaluation (LRE) [12]. Initially started in 1996, the next evaluation was in 2003..., Gunnar Evermann, Mark Gales, Thomas Hain, Dan Kershaw, Gareth Moore, Julian Odell, Dave Ollason, Dan Povey, Valtcho Valtchev, and Phil Woodland: “The HTK book. Revised for HTK version 3.3”, Online:, 2005. [11] M.A. Zissman...

  1. The role of planum temporale in processing accent variation in spoken language comprehension.

    NARCIS (Netherlands)

    Adank, P.M.; Noordzij, M.L.; Hagoort, P.


    A repetitionsuppression functional magnetic resonance imaging paradigm was used to explore the neuroanatomical substrates of processing two types of acoustic variationspeaker and accentduring spoken sentence comprehension. Recordings were made for two speakers and two accents: Standard Dutch and a

  2. Does it really matter whether students' contributions are spoken versus typed in an intelligent tutoring system with natural language? (United States)

    D'Mello, Sidney K; Dowell, Nia; Graesser, Arthur


    There is the question of whether learning differs when students speak versus type their responses when interacting with intelligent tutoring systems with natural language dialogues. Theoretical bases exist for three contrasting hypotheses. The speech facilitation hypothesis predicts that spoken input will increase learning, whereas the text facilitation hypothesis predicts typed input will be superior. The modality equivalence hypothesis claims that learning gains will be equivalent. Previous experiments that tested these hypotheses were confounded by automated speech recognition systems with substantial error rates that were detected by learners. We addressed this concern in two experiments via a Wizard of Oz procedure, where a human intercepted the learner's speech and transcribed the utterances before submitting them to the tutor. The overall pattern of the results supported the following conclusions: (1) learning gains associated with spoken and typed input were on par and quantitatively higher than a no-intervention control, (2) participants' evaluations of the session were not influenced by modality, and (3) there were no modality effects associated with differences in prior knowledge and typing proficiency. Although the results generally support the modality equivalence hypothesis, highly motivated learners reported lower cognitive load and demonstrated increased learning when typing compared with speaking. We discuss the implications of our findings for intelligent tutoring systems that can support typed and spoken input.

  3. Audiovisual spoken word recognition as a clinical criterion for sensory aids efficiency in Persian-language children with hearing loss. (United States)

    Oryadi-Zanjani, Mohammad Majid; Vahab, Maryam; Bazrafkan, Mozhdeh; Haghjoo, Asghar


    The aim of this study was to examine the role of audiovisual speech recognition as a clinical criterion of cochlear implant or hearing aid efficiency in Persian-language children with severe-to-profound hearing loss. This research was administered as a cross-sectional study. The sample size was 60 Persian 5-7 year old children. The assessment tool was one of subtests of Persian version of the Test of Language Development-Primary 3. The study included two experiments: auditory-only and audiovisual presentation conditions. The test was a closed-set including 30 words which were orally presented by a speech-language pathologist. The scores of audiovisual word perception were significantly higher than auditory-only condition in the children with normal hearing (Paudiovisual presentation conditions (P>0.05). The audiovisual spoken word recognition can be applied as a clinical criterion to assess the children with severe to profound hearing loss in order to find whether cochlear implant or hearing aid has been efficient for them or not; i.e. if a child with hearing impairment who using CI or HA can obtain higher scores in audiovisual spoken word recognition than auditory-only condition, his/her auditory skills have appropriately developed due to effective CI or HA as one of the main factors of auditory habilitation. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  4. The role of planum temporale in processing accent variation in spoken language comprehension

    NARCIS (Netherlands)

    Adank, P.M.; Noordzij, M.L.; Hagoort, P.


    A repetition–suppression functional magnetic resonance imaging paradigm was used to explore the neuroanatomical substrates of processing two types of acoustic variation—speaker and accent—during spoken sentence comprehension. Recordings were made for two speakers and two accents: Standard Dutch and

  5. The Relationship between Intrinsic Couplings of the Visual Word Form Area with Spoken Language Network and Reading Ability in Children and Adults

    Directory of Open Access Journals (Sweden)

    Yu Li


    Full Text Available Reading plays a key role in education and communication in modern society. Learning to read establishes the connections between the visual word form area (VWFA and language areas responsible for speech processing. Using resting-state functional connectivity (RSFC and Granger Causality Analysis (GCA methods, the current developmental study aimed to identify the difference in the relationship between the connections of VWFA-language areas and reading performance in both adults and children. The results showed that: (1 the spontaneous connectivity between VWFA and the spoken language areas, i.e., the left inferior frontal gyrus/supramarginal gyrus (LIFG/LSMG, was stronger in adults compared with children; (2 the spontaneous functional patterns of connectivity between VWFA and language network were negatively correlated with reading ability in adults but not in children; (3 the causal influence from LIFG to VWFA was negatively correlated with reading ability only in adults but not in children; (4 the RSFCs between left posterior middle frontal gyrus (LpMFG and VWFA/LIFG were positively correlated with reading ability in both adults and children; and (5 the causal influence from LIFG to LSMG was positively correlated with reading ability in both groups. These findings provide insights into the relationship between VWFA and the language network for reading, and the role of the unique features of Chinese in the neural circuits of reading.

  6. Social inclusion for children with hearing loss in listening and spoken Language early intervention: an exploratory study. (United States)

    Constantinescu-Sharpe, Gabriella; Phillips, Rebecca L; Davis, Aleisha; Dornan, Dimity; Hogan, Anthony


    Social inclusion is a common focus of listening and spoken language (LSL) early intervention for children with hearing loss. This exploratory study compared the social inclusion of young children with hearing loss educated using a listening and spoken language approach with population data. A framework for understanding the scope of social inclusion is presented in the Background. This framework guided the use of a shortened, modified version of the Longitudinal Study of Australian Children (LSAC) to measure two of the five facets of social inclusion ('education' and 'interacting with society and fulfilling social goals'). The survey was completed by parents of children with hearing loss aged 4-5 years who were educated using a LSL approach (n = 78; 37% who responded). These responses were compared to those obtained for typical hearing children in the LSAC dataset (n = 3265). Analyses revealed that most children with hearing loss had comparable outcomes to those with typical hearing on the 'education' and 'interacting with society and fulfilling social roles' facets of social inclusion. These exploratory findings are positive and warrant further investigation across all five facets of the framework to identify which factors influence social inclusion.

  7. Is spoken Danish less intelligible than Swedish?

    NARCIS (Netherlands)

    Gooskens, Charlotte; van Heuven, Vincent J.; van Bezooijen, Renee; Pacilly, Jos J. A.


    The most straightforward way to explain why Danes understand spoken Swedish relatively better than Swedes understand spoken Danish would be that spoken Danish is intrinsically a more difficult language to understand than spoken Swedish. We discuss circumstantial evidence suggesting that Danish is

  8. Does Discourse Congruence Influence Spoken Language Comprehension before Lexical Association? Evidence from Event-Related Potentials (United States)

    Boudewyn, Megan A.; Gordon, Peter C.; Long, Debra; Polse, Lara; Swaab, Tamara Y.


    The goal of this study was to examine how lexical association and discourse congruence affect the time course of processing incoming words in spoken discourse. In an ERP norming study, we presented prime-target pairs in the absence of a sentence context to obtain a baseline measure of lexical priming. We observed a typical N400 effect when participants heard critical associated and unassociated target words in word pairs. In a subsequent experiment, we presented the same word pairs in spoken discourse contexts. Target words were always consistent with the local sentence context, but were congruent or not with the global discourse (e.g., “Luckily Ben had picked up some salt and pepper/basil”, preceded by a context in which Ben was preparing marinara sauce (congruent) or dealing with an icy walkway (incongruent). ERP effects of global discourse congruence preceded those of local lexical association, suggesting an early influence of the global discourse representation on lexical processing, even in locally congruent contexts. Furthermore, effects of lexical association occurred earlier in the congruent than incongruent condition. These results differ from those that have been obtained in studies of reading, suggesting that the effects may be unique to spoken word recognition. PMID:23002319

  9. Auditory-cognitive training improves language performance in prelingually deafened cochlear implant recipients. (United States)

    Ingvalson, Erin M; Young, Nancy M; Wong, Patrick C M


    Phonological and working memory skills have been shown to be important for the development of spoken language. Children who use a cochlear implant (CI) show performance deficits relative to normal hearing (NH) children on all constructs: phonological skills, working memory, and spoken language. Given that phonological skills and working memory have been shown to be important for spoken language development in NH children, we hypothesized that training these foundational skills would result in improved spoken language performance in CI-using children. Nineteen prelingually deafened CI-using children aged 4- to 7-years-old participated. All children had been using their implants for at least one year and were matched on pre-implant hearing thresholds, hearing thresholds at study enrollment, and non-verbal IQ. Children were assessed on expressive vocabulary, listening language, spoken language, and composite language. Ten children received four weeks of training on phonological skills including rhyme, sound blending, and sound discrimination and auditory working memory. The remaining nine children continued with their normal classroom activities for four weeks. Language assessments were repeated following the training/control period. Children who received combined phonological-working memory training showed significant gains on expressive and composite language scores. Children who did not receive training showed no significant improvements at post-test. On average, trained children had gain scores of 6.35 points on expressive language and gain scores of 6.15 points whereas the untrained children had test-retest gain scores of 2.89 points for expressive language and 2.56 for composite language. Our results suggest that training to improve the phonological and working memory skills in CI-using children may lead to improved language performance. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.


    Directory of Open Access Journals (Sweden)

    Ritva Takkinen


    Full Text Available In this paper the use and quality of the evaluative language produced by a bilingual child in a story-telling situation is analysed. The subject, an 11-year-old Finnish boy, Jimmy, is bilingual in Finnish sign language (FinSL and spoken Finnish.He was born deaf but got a cochlear implant at the age of five.The data consist of a spoken and a signed version of “The Frog Story”. The analysis shows that evaluative devices and expressions differ in the spoken and signed stories told by the child. In his Finnish story he uses mostly lexical devices – comments on a character and the character’s actions as well as quoted speech occasionally combined with prosodic features. In his FinSL story he uses both lexical and paralinguistic devices in a balanced way.

  11. Spoken Dialogue Systems

    CERN Document Server

    Jokinen, Kristiina


    Considerable progress has been made in recent years in the development of dialogue systems that support robust and efficient human-machine interaction using spoken language. Spoken dialogue technology allows various interactive applications to be built and used for practical purposes, and research focuses on issues that aim to increase the system's communicative competence by including aspects of error correction, cooperation, multimodality, and adaptation in context. This book gives a comprehensive view of state-of-the-art techniques that are used to build spoken dialogue systems. It provides

  12. A Multilingual Approach to Analysing Standardized Test Results: Immigrant Primary School Children and the Role of Languages Spoken in a Bi-/Multilingual Community (United States)

    De Angelis, Gessica


    The present study adopts a multilingual approach to analysing the standardized test results of primary school immigrant children living in the bi-/multilingual context of South Tyrol, Italy. The standardized test results are from the Invalsi test administered across Italy in 2009/2010. In South Tyrol, several languages are spoken on a daily basis…

  13. How and When Accentuation Influences Temporally Selective Attention and Subsequent Semantic Processing during On-Line Spoken Language Comprehension: An ERP Study (United States)

    Li, Xiao-qing; Ren, Gui-qin


    An event-related brain potentials (ERP) experiment was carried out to investigate how and when accentuation influences temporally selective attention and subsequent semantic processing during on-line spoken language comprehension, and how the effect of accentuation on attention allocation and semantic processing changed with the degree of…

  14. Self-Ratings of Spoken Language Dominance: A Multilingual Naming Test (MINT) and Preliminary Norms for Young and Aging Spanish-English Bilinguals (United States)

    Gollan, Tamar H.; Weissberger, Gali H.; Runnqvist, Elin; Montoya, Rosa I.; Cera, Cynthia M.


    This study investigated correspondence between different measures of bilingual language proficiency contrasting self-report, proficiency interview, and picture naming skills. Fifty-two young (Experiment 1) and 20 aging (Experiment 2) Spanish-English bilinguals provided self-ratings of proficiency level, were interviewed for spoken proficiency, and…

  15. Long-term memory traces for familiar spoken words in tonal languages as revealed by the Mismatch Negativity

    Directory of Open Access Journals (Sweden)

    Naiphinich Kotchabhakdi


    Full Text Available Mismatch negativity (MMN, a primary response to an acoustic change and an index of sensory memory, was used to investigate the processing of the discrimination between familiar and unfamiliar Consonant-Vowel (CV speech contrasts. The MMN was elicited by rare familiar words presented among repetitive unfamiliar words. Phonetic and phonological contrasts were identical in all conditions. MMN elicited by the familiar word deviant was larger than that elicited by the unfamiliar word deviant. The presence of syllable contrast did significantly alter the word-elicited MMN in amplitude and scalp voltage field distribution. Thus, our results indicate the existence of word-related MMN enhancement largely independent of the word status of the standard stimulus. This enhancement may reflect the presence of a longterm memory trace for familiar spoken words in tonal languages.

  16. "Now We Have Spoken." (United States)

    Zimmer, Patricia Moore


    Describes the author's experiences directing a play translated and acted in Korean. Notes that she had to get familiar with the sound of the language spoken fluently, to see how an actor's thought is discerned when the verbal language is not understood. Concludes that so much of understanding and communication unfolds in ways other than with…

  17. Quarterly Data for Spoken Language Preferences of Social Security Retirement and Survivor Claimants (2014-2015) (United States)

    Social Security Administration — This data set provides quarterly volumes for language preferences at the national level of individuals filing claims for Retirement and Survivor benefits for fiscal...

  18. Quarterly Data for Spoken Language Preferences of Social Security Retirement and Survivor Claimants (2016-onwards) (United States)

    Social Security Administration — This data set provides quarterly volumes for language preferences at the national level of individuals filing claims for Retirement and Survivor benefits from fiscal...

  19. Development of lexical-semantic language system: N400 priming effect for spoken words in 18- and 24-month old children. (United States)

    Rämä, Pia; Sirri, Louah; Serres, Josette


    Our aim was to investigate whether developing language system, as measured by a priming task for spoken words, is organized by semantic categories. Event-related potentials (ERPs) were recorded during a priming task for spoken words in 18- and 24-month-old monolingual French learning children. Spoken word pairs were either semantically related (e.g., train-bike) or unrelated (e.g., chicken-bike). The results showed that the N400-like priming effect occurred in 24-month-olds over the right parietal-occipital recording sites. In 18-month-olds the effect was observed similarly to 24-month-olds only in those children with higher word production ability. The results suggest that words are categorically organized in the mental lexicon of children at the age of 2 years and even earlier in children with a high vocabulary. Copyright © 2013 Elsevier Inc. All rights reserved.

  20. Phonological processing of rhyme in spoken language and location in sign language by deaf and hearing participants: a neurophysiological study. (United States)

    Colin, C; Zuinen, T; Bayard, C; Leybaert, J


    Sign languages (SL), like oral languages (OL), organize elementary, meaningless units into meaningful semantic units. Our aim was to compare, at behavioral and neurophysiological levels, the processing of the location parameter in French Belgian SL to that of the rhyme in oral French. Ten hearing and 10 profoundly deaf adults performed a rhyme judgment task in OL and a similarity judgment on location in SL. Stimuli were pairs of pictures. As regards OL, deaf subjects' performances, although above chance level, were significantly lower than that of hearing subjects, suggesting that a metaphonological analysis is possible for deaf people but rests on phonological representations that are less precise than in hearing people. As regards SL, deaf subjects scores indicated that a metaphonological judgment may be performed on location. The contingent negative variation (CNV) evoked by the first picture of a pair was similar in hearing subjects in OL and in deaf subjects in OL and SL. However, an N400 evoked by the second picture of the non-rhyming pairs was evidenced only in hearing subjects in OL. The absence of N400 in deaf subjects may be interpreted as the failure to associate two words according to their rhyme in OL or to their location in SL. Although deaf participants can perform metaphonological judgments in OL, they differ from hearing participants both behaviorally and in ERP. Judgment of location in SL is possible for deaf signers, but, contrary to rhyme judgment in hearing participants, does not elicit any N400. Copyright © 2013 Elsevier Masson SAS. All rights reserved.

  1. Task-Oriented Spoken Dialog System for Second-Language Learning (United States)

    Kwon, Oh-Woog; Kim, Young-Kil; Lee, Yunkeun


    This paper introduces a Dialog-Based Computer Assisted second-Language Learning (DB-CALL) system using task-oriented dialogue processing technology. The system promotes dialogue with a second-language learner for a specific task, such as purchasing tour tickets, ordering food, passing through immigration, etc. The dialog system plays a role of a…

  2. Web-based mini-games for language learning that support spoken interaction

    CSIR Research Space (South Africa)

    Strik, H


    Full Text Available The European ‘Lifelong Learning Programme’ (LLP) project ‘Games Online for Basic Language learning’ (GOBL) aimed to provide youths and adults wishing to improve their basic language skills access to materials for the development of communicative...

  3. Propositional Density in Spoken and Written Language of Czech-Speaking Patients with Mild Cognitive Impairment (United States)

    Smolík, Filip; Stepankova, Hana; Vyhnálek, Martin; Nikolai, Tomáš; Horáková, Karolína; Matejka, Štepán


    Purpose Propositional density (PD) is a measure of content richness in language production that declines in normal aging and more profoundly in dementia. The present study aimed to develop a PD scoring system for Czech and use it to compare PD in language productions of older people with amnestic mild cognitive impairment (aMCI) and control…

  4. Language Outcomes in Deaf or Hard of Hearing Teenagers Who Are Spoken Language Users: Effects of Universal Newborn Hearing Screening and Early Confirmation. (United States)

    Pimperton, Hannah; Kreppner, Jana; Mahon, Merle; Stevenson, Jim; Terlektsi, Emmanouela; Worsfold, Sarah; Yuen, Ho Ming; Kennedy, Colin R

    This study aimed to examine whether (a) exposure to universal newborn hearing screening (UNHS) and b) early confirmation of hearing loss were associated with benefits to expressive and receptive language outcomes in the teenage years for a cohort of spoken language users. It also aimed to determine whether either of these two variables was associated with benefits to relative language gain from middle childhood to adolescence within this cohort. The participants were drawn from a prospective cohort study of a population sample of children with bilateral permanent childhood hearing loss, who varied in their exposure to UNHS and who had previously had their language skills assessed at 6-10 years. Sixty deaf or hard of hearing teenagers who were spoken language users and a comparison group of 38 teenagers with normal hearing completed standardized measures of their receptive and expressive language ability at 13-19 years. Teenagers exposed to UNHS did not show significantly better expressive (adjusted mean difference, 0.40; 95% confidence interval [CI], -0.26 to 1.05; d = 0.32) or receptive (adjusted mean difference, 0.68; 95% CI, -0.56 to 1.93; d = 0.28) language skills than those who were not. Those who had their hearing loss confirmed by 9 months of age did not show significantly better expressive (adjusted mean difference, 0.43; 95% CI, -0.20 to 1.05; d = 0.35) or receptive (adjusted mean difference, 0.95; 95% CI, -0.22 to 2.11; d = 0.42) language skills than those who had it confirmed later. In all cases, effect sizes were of small size and in favor of those exposed to UNHS or confirmed by 9 months. Subgroup analysis indicated larger beneficial effects of early confirmation for those deaf or hard of hearing teenagers without cochlear implants (N = 48; 80% of the sample), and these benefits were significant in the case of receptive language outcomes (adjusted mean difference, 1.55; 95% CI, 0.38 to 2.71; d = 0.78). Exposure to UNHS did not account for significant

  5. Bilateral Versus Unilateral Cochlear Implants in Children: A Study of Spoken Language Outcomes (United States)

    Harris, David; Bennet, Lisa; Bant, Sharyn


    Objectives: Although it has been established that bilateral cochlear implants (CIs) offer additional speech perception and localization benefits to many children with severe to profound hearing loss, whether these improved perceptual abilities facilitate significantly better language development has not yet been clearly established. The aims of this study were to compare language abilities of children having unilateral and bilateral CIs to quantify the rate of any improvement in language attributable to bilateral CIs and to document other predictors of language development in children with CIs. Design: The receptive vocabulary and language development of 91 children was assessed when they were aged either 5 or 8 years old by using the Peabody Picture Vocabulary Test (fourth edition), and either the Preschool Language Scales (fourth edition) or the Clinical Evaluation of Language Fundamentals (fourth edition), respectively. Cognitive ability, parent involvement in children’s intervention or education programs, and family reading habits were also evaluated. Language outcomes were examined by using linear regression analyses. The influence of elements of parenting style, child characteristics, and family background as predictors of outcomes were examined. Results: Children using bilateral CIs achieved significantly better vocabulary outcomes and significantly higher scores on the Core and Expressive Language subscales of the Clinical Evaluation of Language Fundamentals (fourth edition) than did comparable children with unilateral CIs. Scores on the Preschool Language Scales (fourth edition) did not differ significantly between children with unilateral and bilateral CIs. Bilateral CI use was found to predict significantly faster rates of vocabulary and language development than unilateral CI use; the magnitude of this effect was moderated by child age at activation of the bilateral CI. In terms of parenting style, high levels of parental involvement, low amounts of

  6. Bilateral versus unilateral cochlear implants in children: a study of spoken language outcomes. (United States)

    Sarant, Julia; Harris, David; Bennet, Lisa; Bant, Sharyn


    Although it has been established that bilateral cochlear implants (CIs) offer additional speech perception and localization benefits to many children with severe to profound hearing loss, whether these improved perceptual abilities facilitate significantly better language development has not yet been clearly established. The aims of this study were to compare language abilities of children having unilateral and bilateral CIs to quantify the rate of any improvement in language attributable to bilateral CIs and to document other predictors of language development in children with CIs. The receptive vocabulary and language development of 91 children was assessed when they were aged either 5 or 8 years old by using the Peabody Picture Vocabulary Test (fourth edition), and either the Preschool Language Scales (fourth edition) or the Clinical Evaluation of Language Fundamentals (fourth edition), respectively. Cognitive ability, parent involvement in children's intervention or education programs, and family reading habits were also evaluated. Language outcomes were examined by using linear regression analyses. The influence of elements of parenting style, child characteristics, and family background as predictors of outcomes were examined. Children using bilateral CIs achieved significantly better vocabulary outcomes and significantly higher scores on the Core and Expressive Language subscales of the Clinical Evaluation of Language Fundamentals (fourth edition) than did comparable children with unilateral CIs. Scores on the Preschool Language Scales (fourth edition) did not differ significantly between children with unilateral and bilateral CIs. Bilateral CI use was found to predict significantly faster rates of vocabulary and language development than unilateral CI use; the magnitude of this effect was moderated by child age at activation of the bilateral CI. In terms of parenting style, high levels of parental involvement, low amounts of screen time, and more time spent

  7. Yearly Data for Spoken Language Preferences of Supplemental Security Income Blind and Disabled Applicants (2016 Onwards) (United States)

    Social Security Administration — This data set provides annual volumes for language preferences at the national level of individuals filing claims for SSI Blind and Disabled benefits from federal...

  8. Yearly Data for Spoken Language Preferences of Supplemental Security Income (Blind & Disabled) (2011-2015) (United States)

    Social Security Administration — This data set provides annual volumes for language preferences at the national level of individuals filing claims for ESRD Medicare benefits for federal fiscal years...

  9. Yearly Data for Spoken Language Preferences of Supplemental Security Income Aged Applicants (2011-Onward) (United States)

    Social Security Administration — This data set provides annual volumes for language preferences at the national level of individuals filing claims for SSI Aged benefits from federal fiscal year 2011...

  10. Yearly Data for Spoken Language Preferences of Social Security Disability Insurance Claimants (2011-2015) (United States)

    Social Security Administration — This data set provides annual volumes for language preferences at the national level of individuals filing claims for ESRD Medicare benefits for federal fiscal years...

  11. Yearly Data for Spoken Language Preferences of End Stage Renal Disease Medicare Claimants (2011-2015) (United States)

    Social Security Administration — This data set provides annual volumes for language preferences at the national level of individuals filing claims for ESRD Medicare benefits for federal fiscal year...

  12. Quarterly Data for Spoken Language Preferences of Supplemental Security Income Aged Applicants (2014-2015) (United States)

    Social Security Administration — This data set provides quarterly volumes for language preferences at the national level of individuals filing claims for SSI Aged benefits for fiscal years 2014 -...

  13. Quarterly Data for Spoken Language Preferences of End Stage Renal Disease Medicare Claimants (2014-2015) (United States)

    Social Security Administration — This data set provides quarterly volumes for language preferences at the national level of individuals filing claims for ESRD Medicare benefits for fiscal years 2014...

  14. Yearly Data for Spoken Language Preferences of End Stage Renal Disease Medicare Claimants (2016 Onwards) (United States)

    Social Security Administration — This data set provides annual volumes for language preferences at the national level of individuals filing claims for ESRD Medicare benefits for federal fiscal year...

  15. How Spoken Language Comprehension is Achieved by Older Listeners in Difficult Listening Situations. (United States)

    Schneider, Bruce A; Avivi-Reich, Meital; Daneman, Meredyth


    Comprehending spoken discourse in noisy situations is likely to be more challenging to older adults than to younger adults due to potential declines in the auditory, cognitive, or linguistic processes supporting speech comprehension. These challenges might force older listeners to reorganize the ways in which they perceive and process speech, thereby altering the balance between the contributions of bottom-up versus top-down processes to speech comprehension. The authors review studies that investigated the effect of age on listeners' ability to follow and comprehend lectures (monologues), and two-talker conversations (dialogues), and the extent to which individual differences in lexical knowledge and reading comprehension skill relate to individual differences in speech comprehension. Comprehension was evaluated after each lecture or conversation by asking listeners to answer multiple-choice questions regarding its content. Once individual differences in speech recognition for words presented in babble were compensated for, age differences in speech comprehension were minimized if not eliminated. However, younger listeners benefited more from spatial separation than did older listeners. Vocabulary knowledge predicted the comprehension scores of both younger and older listeners when listening was difficult, but not when it was easy. However, the contribution of reading comprehension to listening comprehension appeared to be independent of listening difficulty in younger adults but not in older adults. The evidence suggests (1) that most of the difficulties experienced by older adults are due to age-related auditory declines, and (2) that these declines, along with listening difficulty, modulate the degree to which selective linguistic and cognitive abilities are engaged to support listening comprehension in difficult listening situations. When older listeners experience speech recognition difficulties, their attentional resources are more likely to be deployed to

  16. Case report: acquisition of three spoken languages by a child with a cochlear implant. (United States)

    Francis, Alexander L; Ho, Diana Wai Lam


    There have been only two reports of multilingual cochlear implant users to date, and both of these were postlingually deafened adults. Here we report the case of a 6-year-old early-deafened child who is acquiring Cantonese, English and Mandarin in Hong Kong. He and two age-matched peers with similar educational backgrounds were tested using common, standardized tests of vocabulary and expressive and receptive language skills (Peabody Picture Vocabulary Test (Revised) and Reynell Developmental Language Scales version II). Results show that this child is acquiring Cantonese, English and Mandarin to a degree comparable to two classmates with normal hearing and similar educational and social backgrounds.

  17. Assessing Spoken Language Competence in Children with Selective Mutism: Using Parents as Test Presenters (United States)

    Klein, Evelyn R.; Armstrong, Sharon Lee; Shipon-Blum, Elisa


    Children with selective mutism (SM) display a failure to speak in select situations despite speaking when comfortable. The purpose of this study was to obtain valid assessments of receptive and expressive language in 33 children (ages 5 to 12) with SM. Because some children with SM will speak to parents but not a professional, another purpose was…

  18. Cross-Sensory Correspondences and Symbolism in Spoken and Written Language (United States)

    Walker, Peter


    Lexical sound symbolism in language appears to exploit the feature associations embedded in cross-sensory correspondences. For example, words incorporating relatively high acoustic frequencies (i.e., front/close rather than back/open vowels) are deemed more appropriate as names for concepts associated with brightness, lightness in weight,…

  19. Grammatical number processing and anticipatory eye movements are not tightly coordinated in English spoken language comprehension

    Directory of Open Access Journals (Sweden)

    Brian eRiordan


    Full Text Available Recent studies of eye movements in world-situated language comprehension have demonstrated that rapid processing of morphosyntactic information – e.g., grammatical gender and number marking – can produce anticipatory eye movements to referents in the visual scene. We investigated how type of morphosyntactic information and the goals of language users in comprehension affected eye movements, focusing on the processing of grammatical number morphology in English-speaking adults. Participants’ eye movements were recorded as they listened to simple English declarative (There are the lions. and interrogative (Where are the lions? sentences. In Experiment 1, no differences were observed in speed to fixate target referents when grammatical number information was informative relative to when it was not. The same result was obtained in a speeded task (Experiment 2 and in a task using mixed sentence types (Experiment 3. We conclude that grammatical number processing in English and eye movements to potential referents are not tightly coordinated. These results suggest limits on the role of predictive eye movements in concurrent linguistic and scene processing. We discuss how these results can inform and constrain predictive approaches to language processing.

  20. Spoken Word Recognition Errors in Speech Audiometry: A Measure of Hearing Performance?

    Directory of Open Access Journals (Sweden)

    Martine Coene


    Full Text Available This report provides a detailed analysis of incorrect responses from an open-set spoken word-repetition task which is part of a Dutch speech audiometric test battery. Single-consonant confusions were analyzed from 230 normal hearing participants in terms of the probability of choice of a particular response on the basis of acoustic-phonetic, lexical, and frequency variables. The results indicate that consonant confusions are better predicted by lexical knowledge than by acoustic properties of the stimulus word. A detailed analysis of the transmission of phonetic features indicates that “voicing” is best preserved whereas “manner of articulation” yields most perception errors. As consonant confusion matrices are often used to determine the degree and type of a patient’s hearing impairment, to predict a patient’s gain in hearing performance with hearing devices and to optimize the device settings in view of maximum output, the observed findings are highly relevant for the audiological practice. Based on our findings, speech audiometric outcomes provide a combined auditory-linguistic profile of the patient. The use of confusion matrices might therefore not be the method best suited to measure hearing performance. Ideally, they should be complemented by other listening task types that are known to have less linguistic bias, such as phonemic discrimination.

  1. Human inferior colliculus activity relates to individual differences in spoken language learning. (United States)

    Chandrasekaran, Bharath; Kraus, Nina; Wong, Patrick C M


    A challenge to learning words of a foreign language is encoding nonnative phonemes, a process typically attributed to cortical circuitry. Using multimodal imaging methods [functional magnetic resonance imaging-adaptation (fMRI-A) and auditory brain stem responses (ABR)], we examined the extent to which pretraining pitch encoding in the inferior colliculus (IC), a primary midbrain structure, related to individual variability in learning to successfully use nonnative pitch patterns to distinguish words in American English-speaking adults. fMRI-A indexed the efficiency of pitch representation localized to the IC, whereas ABR quantified midbrain pitch-related activity with millisecond precision. In line with neural "sharpening" models, we found that efficient IC pitch pattern representation (indexed by fMRI) related to superior neural representation of pitch patterns (indexed by ABR), and consequently more successful word learning following sound-to-meaning training. Our results establish a critical role for the IC in speech-sound representation, consistent with the established role for the IC in the representation of communication signals in other animal models.

  2. SNOT-22: psychometric properties and cross-cultural adaptation into the Portuguese language spoken in Brazil. (United States)

    Caminha, Guilherme Pilla; Melo Junior, José Tavares de; Hopkins, Claire; Pizzichini, Emilio; Pizzichini, Marcia Margaret Menezes


    Rhinosinusitis is a highly prevalent disease and a major cause of high medical costs. It has been proven to have an impact on the quality of life through generic health-related quality of life assessments. However, generic instruments may not be able to factor in the effects of interventions and treatments. SNOT-22 is a major disease-specific instrument to assess quality of life for patients with rhinosinusitis. Nevertheless, there is still no validated SNOT-22 version in our country. Cross-cultural adaptation of the SNOT-22 into Brazilian Portuguese and assessment of its psychometric properties. The Brazilian version of the SNOT-22 was developed according to international guidelines and was broken down into nine stages: 1) Preparation 2) Translation 3) Reconciliation 4) Back-translation 5) Comparison 6) Evaluation by the author of the SNOT-22 7) Revision by committee of experts 8) Cognitive debriefing 9) Final version. Second phase: prospective study consisting of a verification of the psychometric properties, by analyzing internal consistency and test-retest reliability. Cultural adaptation showed adequate understanding, acceptability and psychometric properties. We followed the recommended steps for the cultural adaptation of the SNOT-22 into Portuguese language, producing a tool for the assessment of patients with sinonasal disorders of clinical importance and for scientific studies.

  3. Use of Automated Scoring in Spoken Language Assessments for Test Takers with Speech Impairments. Research Report. ETS RR-17-42 (United States)

    Loukina, Anastassia; Buzick, Heather


    This study is an evaluation of the performance of automated speech scoring for speakers with documented or suspected speech impairments. Given that the use of automated scoring of open-ended spoken responses is relatively nascent and there is little research to date that includes test takers with disabilities, this small exploratory study focuses…

  4. Intentional and Reactive Inhibition during Spoken-Word Stroop Task Performance in People with Aphasia (United States)

    Pompon, Rebecca Hunting; McNeil, Malcolm R.; Spencer, Kristie A.; Kendall, Diane L.


    Purpose: The integrity of selective attention in people with aphasia (PWA) is currently unknown. Selective attention is essential for everyday communication, and inhibition is an important part of selective attention. This study explored components of inhibition--both intentional and reactive inhibition--during spoken-word production in PWA and in…

  5. Early use of orthographic information in spoken word recognition: Event-related potential evidence from the Korean language. (United States)

    Kwon, Youan; Choi, Sungmook; Lee, Yoonhyoung


    This study examines whether orthographic information is used during prelexical processes in spoken word recognition by investigating ERPs during spoken word processing for Korean words. Differential effects due to orthographic syllable neighborhood size and sound-to-spelling consistency on P200 and N320 were evaluated by recording ERPs from 42 participants during a lexical decision task. The results indicate that P200 was smaller for words whose orthographic syllable neighbors are large in number rather than those that are small. In addition, a word with a large orthographic syllable neighborhood elicited a smaller N320 effect than a word with a small orthographic syllable neighborhood only when the word had inconsistent sound-to-spelling mapping. The results provide support for the assumption that orthographic information is used early during the prelexical spoken word recognition process. © 2015 Society for Psychophysiological Research.

  6. A randomized trial comparison of the effects of verbal and pictorial naturalistic communication strategies on spoken language for young children with autism. (United States)

    Schreibman, Laura; Stahmer, Aubyn C


    Presently there is no consensus on the specific behavioral treatment of choice for targeting language in young nonverbal children with autism. This randomized clinical trial compared the effectiveness of a verbally-based intervention, Pivotal Response Training (PRT) to a pictorially-based behavioral intervention, the Picture Exchange Communication System (PECS) on the acquisition of spoken language by young (2-4 years), nonverbal or minimally verbal (≤9 words) children with autism. Thirty-nine children were randomly assigned to either the PRT or PECS condition. Participants received on average 247 h of intervention across 23 weeks. Dependent measures included overall communication, expressive vocabulary, pictorial communication and parent satisfaction. Children in both intervention groups demonstrated increases in spoken language skills, with no significant difference between the two conditions. Seventy-eight percent of all children exited the program with more than 10 functional words. Parents were very satisfied with both programs but indicated PECS was more difficult to implement.

  7. Automated Scoring of L2 Spoken English with Random Forests (United States)

    Kobayashi, Yuichiro; Abe, Mariko


    The purpose of the present study is to assess second language (L2) spoken English using automated scoring techniques. Automated scoring aims to classify a large set of learners' oral performance data into a small number of discrete oral proficiency levels. In automated scoring, objectively measurable features such as the frequencies of lexical and…

  8. Evaluating Attributions of Delay and Confusion in Young Bilinguals: Special Insights from Infants Acquiring a Signed and a Spoken Language. (United States)

    Petitto, Laura Ann; Holowka, Siobhan


    Examines whether early simultaneous bilingual language exposure causes children to be language delayed or confused. Cites research suggesting normal and parallel linguistic development occurs in each language in young children and young children's dual language developments are similar to monolingual language acquisition. Research on simultaneous…

  9. How appropriate are the English language test requirements for non-UK-trained nurses? A qualitative study of spoken communication in UK hospitals. (United States)

    Sedgwick, Carole; Garner, Mark


    Non-native speakers of English who hold nursing qualifications from outside the UK are required to provide evidence of English language competence by achieving a minimum overall score of Band 7 on the International English Language Testing System (IELTS) academic test. To describe the English language required to deal with the daily demands of nursing in the UK. To compare these abilities with the stipulated levels on the language test. A tracking study was conducted with 4 nurses, and focus groups with 11 further nurses. The transcripts of the interviews and focus groups were analysed thematically for recurrent themes. These findings were then compared with the requirements of the IELTS spoken test. The study was conducted outside the participants' working shifts in busy London hospitals. The participants in the tracking study were selected opportunistically;all were trained in non-English speaking countries. Snowball sampling was used for the focus groups, of whom 4 were non-native and 7 native speakers of English. In the tracking study, each of the 4 nurses was interviewed on four occasions, outside the workplace, and as close to the end of a shift as possible. They were asked to recount their spoken interactions during the course of their shift. The participants in the focus groups were asked to describe their typical interactions with patients, family members, doctors, and nursing colleagues. They were prompted to recall specific instances of frequently-occurring communication problems. All interactions were audio-recorded, with the participants' permission,and transcribed. Nurses are at the centre of communication for patient care. They have to use appropriate registers to communicate with a range of health professionals, patients and their families. They must elicit information, calm and reassure, instruct, check procedures, ask for and give opinions,agree and disagree. Politeness strategies are needed to avoid threats to face. They participate in medical

  10. Language configurations of degree-related denotations in the spoken production of a group of Colombian EFL university students: A corpus-based study

    Directory of Open Access Journals (Sweden)

    Wilder Yesid Escobar


    Full Text Available Recognizing that developing the competences needed to appropriately use linguistic resources according to contextual characteristics (pragmatics is as important as the cultural-imbedded linguistic knowledge itself (semantics and that both are equally essential to form competent speakers of English in foreign language contexts, we feel this research relies on corpus linguistics to analyze both the scope and the limitations of the sociolinguistic knowledge and the communicative skills of English students at the university level. To such end, a linguistic corpus was assembled, compared to an existing corpus of native speakers, and analyzed in terms of the frequency, overuse, underuse, misuse, ambiguity, success, and failure of the linguistic parameters used in speech acts. The findings herein describe the linguistic configurations employed to modify levels and degrees of descriptions (salient sematic theme exhibited in the EFL learners´ corpus appealing to the sociolinguistic principles governing meaning making and language use which are constructed under the social conditions of the environments where the language is naturally spoken for sociocultural exchange.

  11. Dichotic listening performance predicts language comprehension. (United States)

    Asbjørnsen, Arve E; Helland, Turid


    Dichotic listening performance is considered a reliable and valid procedure for the assessment of language lateralisation in the brain. However, the documentation of a relationship between language functions and dichotic listening performance is sparse, although it is accepted that dichotic listening measures language perception. In particular, language comprehension should show close correspondence to perception of language stimuli. In the present study, we tested samples of reading-impaired and normally achieving children between 10 and 13 years of age with tests of reading skills, language comprehension, and dichotic listening to consonant-vowel (CV) syllables. A high correlation between the language scores and the dichotic listening performance was expected. However, since the left ear score is believed to be an error when assessing language laterality, covariation was expected for the right ear scores only. In addition, directing attention to one ear input was believed to reduce the influence of random factors, and thus show a more concise estimate of left hemisphere language capacity. Thus, a stronger correlation between language comprehension skills and the dichotic listening performance when attending to the right ear was expected. The analyses yielded a positive correlation between the right ear score in DL and language comprehension, an effect that was stronger when attending to the right ear. The present results confirm the assumption that dichotic listening with CV syllables measures an aspect of language perception and language skills that is related to general language comprehension.

  12. Quarterly Data for Spoken Language Preferences of Supplemental Security Income Blind and Disabled Applicants (2014-2015) (United States)

    Social Security Administration — This data set provides quarterly volumes for language preferences at the national level of individuals filing claims for SSI Blind and Disabled benefits for fiscal...

  13. Social Security Administration - Quarterly Data for Spoken Language Preferences of Supplemental Security Income Blind and Disabled Applicants (2016-onwards) (United States)

    Social Security Administration — This data set provides quarterly volumes for language preferences at the national level of individuals filing claims for SSI Blind and Disabled benefits from fiscal...

  14. A Mother Tongue Spoken Mainly by Fathers. (United States)

    Corsetti, Renato


    Reviews what is known about Esperanto as a home language and first language. Recorded cases of Esperanto-speaking families are known since 1919, and in nearly all of the approximately 350 families documented, the language is spoken to the children by the father. The data suggests that this "artificial bilingualism" can be as successful…

  15. Spoken word recognition in young tone language learners: Age-dependent effects of segmental and suprasegmental variation. (United States)

    Ma, Weiyi; Zhou, Peng; Singh, Leher; Gao, Liqun


    The majority of the world's languages rely on both segmental (vowels, consonants) and suprasegmental (lexical tones) information to contrast the meanings of individual words. However, research on early language development has mostly focused on the acquisition of vowel-consonant languages. Developmental research comparing sensitivity to segmental and suprasegmental features in young tone learners is extremely rare. This study examined 2- and 3-year-old monolingual tone learners' sensitivity to vowels and tones. Experiment 1a tested the influence of vowel and tone variation on novel word learning. Vowel and tone variation hindered word recognition efficiency in both age groups. However, tone variation hindered word recognition accuracy only in 2-year-olds, while 3-year-olds were insensitive to tone variation. Experiment 1b demonstrated that 3-year-olds could use tones to learn new words when additional support was provided, and additionally, that Tone 3 words were exceptionally difficult to learn. Experiment 2 confirmed a similar pattern of results when children were presented with familiar words. This study is the first to show that despite the importance of tones in tone languages, vowels maintain primacy over tones in young children's word recognition and that tone sensitivity in word learning and recognition changes between 2 and 3years of age. The findings suggest that early lexical processes are more tightly constrained by variation in vowels than by tones. Copyright © 2016 Elsevier B.V. All rights reserved.

  16. A Pilot Study of Telepractice for Teaching Listening and Spoken Language to Mandarin-Speaking Children with Congenital Hearing Loss (United States)

    Chen, Pei-Hua; Liu, Ting-Wei


    Telepractice provides an alternative form of auditory-verbal therapy (eAVT) intervention through videoconferencing; this can be of immense benefit for children with hearing loss, especially those living in rural or remote areas. The effectiveness of eAVT for the language development of Mandarin-speaking preschoolers with hearing loss was…

  17. Spoken language and everyday functioning in 5-year-old children using hearing aids or cochlear implants. (United States)

    Cupples, Linda; Ching, Teresa Yc; Button, Laura; Seeto, Mark; Zhang, Vicky; Whitfield, Jessica; Gunnourie, Miriam; Martin, Louise; Marnane, Vivienne


    This study investigated the factors influencing 5-year language, speech and everyday functioning of children with congenital hearing loss. Standardised tests including PLS-4, PPVT-4 and DEAP were directly administered to children. Parent reports on language (CDI) and everyday functioning (PEACH) were collected. Regression analyses were conducted to examine the influence of a range of demographic variables on outcomes. Participants were 339 children enrolled in the Longitudinal Outcomes of Children with Hearing Impairment (LOCHI) study. Children's average receptive and expressive language scores were approximately 1 SD below the mean of typically developing children, and scores on speech production and everyday functioning were more than 1 SD below. Regression models accounted for 70-23% of variance in scores across different tests. Earlier CI switch-on and higher non-verbal ability were associated with better outcomes in most domains. Earlier HA fitting and use of oral communication were associated with better outcomes on directly administered language assessments. Severity of hearing loss and maternal education influenced outcomes of children with HAs. The presence of additional disabilities affected outcomes of children with CIs. The findings provide strong evidence for the benefits of early HA fitting and early CI for improving children's outcomes.

  18. Approaches for Language Identification in Mismatched Environments (United States)


    domain adaptation, unsupervised learning , deep neural networks, bottleneck features 1. Introduction and task Spoken language identification (LID) is...the process of identifying the language in a spoken speech utterance. In recent years, great improvements in LID system performance have been the case in practice. Lastly, we conduct an out-of-set experiment where VoA data from 9 other languages (Amharic, Creole, Croatian, English

  19. Designing Second Language Performance Assessments. Technical Report. (United States)

    Norris, John M.; Brown, James Dean; Hudson, Thom; Yoshioka, Jim

    This technical report focuses on the decision-making potential provided by second language performance assessments. First, performance assessment is situated within the broader discussion of alternatives in language assessment and in educational assessment in general. Next, issues in performance assessment design, implementation, reliability, and…

  20. Czech spoken in Bohemia and Moravia

    NARCIS (Netherlands)

    Šimáčková, Š.; Podlipský, V.J.; Chládková, K.


    As a western Slavic language of the Indo-European family, Czech is closest to Slovak and Polish. It is spoken as a native language by nearly 10 million people in the Czech Republic (Czech Statistical Office n.d.). About two million people living abroad, mostly in the USA, Canada, Austria, Germany,

  1. Sentence Repetition in Deaf Children with Specific Language Impairment in British Sign Language (United States)

    Marshall, Chloë; Mason, Kathryn; Rowley, Katherine; Herman, Rosalind; Atkinson, Joanna; Woll, Bencie; Morgan, Gary


    Children with specific language impairment (SLI) perform poorly on sentence repetition tasks across different spoken languages, but until now, this methodology has not been investigated in children who have SLI in a signed language. Users of a natural sign language encode different sentence meanings through their choice of signs and by altering…


    Directory of Open Access Journals (Sweden)

    Süleyman Kaan YALÇIN (M.A.H.


    Full Text Available Language is an object realized in two ways; spokenlanguage and written language. Each language can havethe characteristics of a spoken language, however, everylanguage can not have the characteristics of a writtenlanguage since there are some requirements for alanguage to be deemed as a written language. Theserequirements are selection, coding, standardization andbecoming widespread. It is necessary for a language tomeet these requirements in either natural or artificial wayso to be deemed as a written language (standardlanguage.Turkish language, which developed as a singlewritten language till 13th century, was divided intolanguages as West Turkish and North-East Turkish bymeeting the requirements of a written language in anatural way. Following this separation and through anatural process, it showed some differences in itself;however, the policy of converting the spoken language ofeach Turkish clan into their written language -the policypursued by Russia in a planned way- turned Turkish,which came to 20th century as a few written languagesinto20 different written languages. Implementation ofdiscriminatory language policies suggested by missionerssuch as Slinky and Ostramov to Russian Government,imposing of Cyrillic alphabet full of different andunnecessary signs on each Turkish clan by force andothering activities of Soviet boarding schools opened hadconsiderable effects on the said process.This study aims at explaining that the conversionof spoken languages of Turkish societies in Russia intotheir written languages did not result from a naturalprocess; the historical development of Turkish languagewhich is shaped as 20 separate written languages onlybecause of the pressure exerted by political will; and how the Russian subjected language concept -which is thememory of a nation- to an artificial process.

  3. Impact of family language and testing language on reading performance in a bilingual educational context. (United States)

    Elosua Oliden, Paula; Mujika Lizaso, Josu


    When different languages co-exist in one area, or when one person speaks more than one language, the impact of language on psychological and educational assessment processes can be considerable. The aim of this work was to study the impact of testing language in a community with two official languages: Spanish and Basque. By taking the PISA 2009 Reading Comprehension Test as a basis for analysis, four linguistic groups were defined according to the language spoken at home and the test language. Psychometric equivalence between test forms and differences in results among the four language groups were analyzed. The comparison of competence means took into account the effects of the index of socioeconomic and cultural status (ISEC) and gender. One reading unit with differential item functioning was detected. The reading competence means were considerably higher in the monolingual Spanish-Spanish group. No differences were found between the language groups based on family language when the test was conducted in Basque. The study illustrates the importance of taking into account psychometric, linguistic and sociolinguistic factors in linguistically diverse assessment contexts.

  4. Orthographic Facilitation in Chinese Spoken Word Recognition: An ERP Study (United States)

    Zou, Lijuan; Desroches, Amy S.; Liu, Youyi; Xia, Zhichao; Shu, Hua


    Orthographic influences in spoken word recognition have been previously examined in alphabetic languages. However, it is unknown whether orthographic information affects spoken word recognition in Chinese, which has a clean dissociation between orthography (O) and phonology (P). The present study investigated orthographic effects using event…

  5. Novel Spoken Word Learning in Adults with Developmental Dyslexia (United States)

    Conner, Peggy S.


    A high percentage of individuals with dyslexia struggle to learn unfamiliar spoken words, creating a significant obstacle to foreign language learning after early childhood. The origin of spoken-word learning difficulties in this population, generally thought to be related to the underlying literacy deficit, is not well defined (e.g., Di Betta…

  6. Spoken Grammar for Chinese Learners

    Institute of Scientific and Technical Information of China (English)



    Currently, the concept of spoken grammar has been mentioned among Chinese teachers. However, teach-ers in China still have a vague idea of spoken grammar. Therefore this dissertation examines what spoken grammar is and argues that native speakers’ model of spoken grammar needs to be highlighted in the classroom teaching.

  7. Language Motivation, Metacognitive Strategies and Language Performance: A Cause and Effect Correlation


    Ag. Bambang Setiyadi; - Mahpul; Muhammad Sukirlan; Bujang Rahman


    Studies on motivation in language learning have been well documented. The role of motivation in determining the use of learning strategies has been identified and the correlation between motivation and language performance has been determined. However, how language motivation in EFL context is classified and how language motivation is inter-correlated with the use of metacognitive and language performance has still not become widespread in the literature on language learning. The current stud...

  8. Language Barrier And The Performance of Secondary School Students in EnglishLanguage in Katsina Metropolis


    Nwabudike Christopher Eziafa; Ojoko E. A.; George Anaso Nwaorah


    This research work centres on Language Barrier and the Performance of Secondary School Students in English Language in Katsina Metropolis. The study identifies the causes of failure in English Language in secondary schools, the factors responsible for the inability of students to learn English language as a second language and the effect of mother tongue interference on the performance of students in English language in the study area. Data for this study  were collected through the use of st...

  9. Bilinguals Show Weaker Lexical Access during Spoken Sentence Comprehension (United States)

    Shook, Anthony; Goldrick, Matthew; Engstler, Caroline; Marian, Viorica


    When bilinguals process written language, they show delays in accessing lexical items relative to monolinguals. The present study investigated whether this effect extended to spoken language comprehension, examining the processing of sentences with either low or high semantic constraint in both first and second languages. English-German…

  10. Orthographic effects in spoken word recognition: Evidence from Chinese. (United States)

    Qu, Qingqing; Damian, Markus F


    Extensive evidence from alphabetic languages demonstrates a role of orthography in the processing of spoken words. Because alphabetic systems explicitly code speech sounds, such effects are perhaps not surprising. However, it is less clear whether orthographic codes are involuntarily accessed from spoken words in languages with non-alphabetic systems, in which the sound-spelling correspondence is largely arbitrary. We investigated the role of orthography via a semantic relatedness judgment task: native Mandarin speakers judged whether or not spoken word pairs were related in meaning. Word pairs were either semantically related, orthographically related, or unrelated. Results showed that relatedness judgments were made faster for word pairs that were semantically related than for unrelated word pairs. Critically, orthographic overlap on semantically unrelated word pairs induced a significant increase in response latencies. These findings indicate that orthographic information is involuntarily accessed in spoken-word recognition, even in a non-alphabetic language such as Chinese.

  11. Pulmonary Function Affects Language Performance in Aging

    Directory of Open Access Journals (Sweden)

    Lewina O Lee


    Full Text Available Background Good pulmonary function (PF is associated with preservation of cognitive performance, primarily of executive functions, in aging (Albert et al., 1995; Chyou et al., 1996; Emery, Finkel, & Pedersen, 2012; Yohannes & Gindo, 2013. The contribution of PF to older adults’ language abilities, however, has never been explored, to our knowledge. We addressed this gap by examining the effects of PF on older adults’ language functions, as measured by naming and sentence processing accuracy. We predicted similar effects as found for executive functions, given the positive associations between executive functions and sentence processing in aging (e.g., Goral et al., 2011. Methods Data were collected from 190 healthy adults aged 55 to 84 years (M = 71.1, SD = 8.1, with no history of neurological or psychiatric disorders. Procedure PF was measured prior to language testing. Measures included forced expiratory volume in 1 second (FEV1 and forced vital capacity (FVC. Language functions were assessed through performance on computer-administered lexical retrieval and sentence processing tasks. Sentence processing was measured using two auditory comprehension tasks: one, of embedded sentences (ES, the other, of sentences with multiple negatives (MN. Lexical retrieval was measured using the Boston Naming Test (BNT and Action Naming Test (ANT. Performance was scored for percent accuracy. Additionally, lexical retrieval was evaluated with a phonemic fluency task (FAS, which also taps executive function abilities. Statistical Analyses Multiple regression was used to examine the association between pulmonary and language functions, adjusting for age, education, gender, history of respiratory illness, current level of physical activities, and current and past smoking. Results Better PF was associated with better sentence processing and lexical retrieval on naming tasks, but not with phonemic fluency, after adjusting for covariates. Higher FVC was

  12. Spoken Sentence Production in College Students with Dyslexia: Working Memory and Vocabulary Effects (United States)

    Wiseheart, Rebecca; Altmann, Lori J. P.


    Background: Individuals with dyslexia demonstrate syntactic difficulties on tasks of language comprehension, yet little is known about spoken language production in this population. Aims: To investigate whether spoken sentence production in college students with dyslexia is less proficient than in typical readers, and to determine whether group…

  13. Code-switched English pronunciation modeling for Swahili spoken term detection

    CSIR Research Space (South Africa)

    Kleynhans, N


    Full Text Available Computer Science 81 ( 2016 ) 128 – 135 5th Workshop on Spoken Language Technology for Under-resourced Languages, SLTU 2016, 9-12 May 2016, Yogyakarta, Indonesia Code-switched English Pronunciation Modeling for Swahili Spoken Term Detection Neil...

  14. Phonological Analysis of University Students’ Spoken Discourse

    Directory of Open Access Journals (Sweden)

    Clara Herlina


    Full Text Available The study of discourse is the study of using language in actual use. In this article, the writer is trying to investigate the phonological features, either segmental or supra-segmental, in the spoken discourse of Indonesian university students. The data were taken from the recordings of 15 conversations by 30 students of Bina Nusantara University who are taking English Entrant subject (TOEFL –IBT. Finally, the writer is in opinion that the students are still influenced by their first language in their spoken discourse. This results in English with Indonesian accent. Even though it does not cause misunderstanding at the moment, this may become problematic if they have to communicate in the real world.  

  15. When novel sentences spoken or heard for the first time in the history of the universe are not enough: toward a dual-process model of language. (United States)

    Van Lancker Sidtis, Diana


    Although interest in the language sciences was previously focused on newly created sentences, more recently much attention has turned to the importance of formulaic expressions in normal and disordered communication. Also referred to as formulaic expressions and made up of speech formulas, idioms, expletives, serial and memorized speech, slang, sayings, clichés, and conventional expressions, non-propositional language forms a large proportion of every speaker's competence, and may be differentially disturbed in neurological disorders. This review aims to examine non-propositional speech with respect to linguistic descriptions, psycholinguistic experiments, sociolinguistic studies, child language development, clinical language disorders, and neurological studies. Evidence from numerous sources reveals differentiated and specialized roles for novel and formulaic verbal functions, and suggests that generation of novel sentences and management of prefabricated expressions represent two legitimate and separable processes in language behaviour. A preliminary model of language behaviour that encompasses unitary and compositional properties and their integration in everyday language use is proposed. Integration and synchronizing of two disparate processes in language behaviour, formulaic and novel, characterizes normal communicative function and contributes to creativity in language. This dichotomy is supported by studies arising from other disciplines in neurology and psychology. Further studies are necessary to determine in what ways the various categories of formulaic expressions are related, and how these categories are processed by the brain. Better understanding of how non-propositional categories of speech are stored and processed in the brain can lead to better informed treatment strategies in language disorders.

  16. Spoken sentence production in college students with dyslexia: working memory and vocabulary effects. (United States)

    Wiseheart, Rebecca; Altmann, Lori J P


    Individuals with dyslexia demonstrate syntactic difficulties on tasks of language comprehension, yet little is known about spoken language production in this population. To investigate whether spoken sentence production in college students with dyslexia is less proficient than in typical readers, and to determine whether group differences can be attributable to cognitive differences between groups. Fifty-one college students with and without dyslexia were asked to produce sentences from stimuli comprising a verb and two nouns. Verb types varied in argument structure and morphological form and nouns varied in animacy. Outcome measures were precision (measured by fluency, grammaticality and completeness) and efficiency (measured by response times). Vocabulary and working memory tests were also administered and used as predictors of sentence production performance. Relative to non-dyslexic peers, students with dyslexia responded significantly slower and produced sentences that were significantly less precise in terms of fluency, grammaticality and completeness. The primary predictors of precision and efficiency were working memory, which differed between groups, and vocabulary, which did not. College students with dyslexia were significantly less facile and flexible on this spoken sentence-production task than typical readers, which is consistent with previous studies of school-age children with dyslexia. Group differences in performance were traced primarily to limited working memory, and were somewhat mitigated by strong vocabulary. © 2017 Royal College of Speech and Language Therapists.

  17. Cross-Cultural Differences in Beliefs and Practices that Affect the Language Spoken to Children: Mothers with Indian and Western Heritage (United States)

    Simmons, Noreen; Johnston, Judith


    Background: Speech-language pathologists often advise families about interaction patterns that will facilitate language learning. This advice is typically based on research with North American families of European heritage and may not be culturally suited for non-Western families. Aims: The goal of the project was to identify differences in the…

  18. Narrative skills in deaf children who use spoken English: Dissociations between macro and microstructural devices. (United States)

    Jones, -A C; Toscano, E; Botting, N; Marshall, C-R; Atkinson, J R; Denmark, T; Herman, -R; Morgan, G


    Previous research has highlighted that deaf children acquiring spoken English have difficulties in narrative development relative to their hearing peers both in terms of macro-structure and with micro-structural devices. The majority of previous research focused on narrative tasks designed for hearing children that depend on good receptive language skills. The current study compared narratives of 6 to 11-year-old deaf children who use spoken English (N=59) with matched for age and non-verbal intelligence hearing peers. To examine the role of general language abilities, single word vocabulary was also assessed. Narratives were elicited by the retelling of a story presented non-verbally in video format. Results showed that deaf and hearing children had equivalent macro-structure skills, but the deaf group showed poorer performance on micro-structural components. Furthermore, the deaf group gave less detailed responses to inferencing probe questions indicating poorer understanding of the story's underlying message. For deaf children, micro-level devices most strongly correlated with the vocabulary measure. These findings suggest that deaf children, despite spoken language delays, are able to convey the main elements of content and structure in narrative but have greater difficulty in using grammatical devices more dependent on finer linguistic and pragmatic skills. Crown Copyright © 2016. Published by Elsevier Ltd. All rights reserved.

  19. Sign language: an international handbook

    NARCIS (Netherlands)

    Pfau, R.; Steinbach, M.; Woll, B.


    Sign language linguists show here that all the questions relevant to the linguistic investigation of spoken languages can be asked about sign languages. Conversely, questions that sign language linguists consider - even if spoken language researchers have not asked them yet - should also be asked of

  20. Cross-cultural differences in beliefs and practices that affect the language spoken to children: mothers with Indian and Western heritage. (United States)

    Simmons, Noreen; Johnston, Judith


    Speech-language pathologists often advise families about interaction patterns that will facilitate language learning. This advice is typically based on research with North American families of European heritage and may not be culturally suited for non-Western families. The goal of the project was to identify differences in the beliefs and practices of Indian and Euro-Canadian mothers that would affect patterns of talk to children. A total of 47 Indian mothers and 51 Euro-Canadian mothers of preschool age children completed a written survey concerning child-rearing practices and beliefs, especially those about talk to children. Discriminant analyses indicated clear cross-cultural differences and produced functions that could predict group membership with a 96% accuracy rate. Items contributing most to these functions concerned the importance of family, perceptions of language learning, children's use of language in family and society, and interactions surrounding text. Speech-language pathologists who wish to adapt their services for families of Indian heritage should remember the centrality of the family, the likelihood that there will be less emphasis on early independence and achievement, and the preference for direct instruction.

  1. Examining Transcription, Autonomy and Reflective Practice in Language Development (United States)

    Cooke, Simon D.


    This pilot study explores language development among a class of L2 students who were required to transcribe and reflect upon spoken performances. The class was given tasks for self and peer-evaluation and afforded the opportunity to assume more responsibility for assessing language development of both themselves and their peers. Several studies…

  2. Syllable Frequency and Spoken Word Recognition: An Inhibitory Effect. (United States)

    González-Alvarez, Julio; Palomar-García, María-Angeles


    Research has shown that syllables play a relevant role in lexical access in Spanish, a shallow language with a transparent syllabic structure. Syllable frequency has been shown to have an inhibitory effect on visual word recognition in Spanish. However, no study has examined the syllable frequency effect on spoken word recognition. The present study tested the effect of the frequency of the first syllable on recognition of spoken Spanish words. A sample of 45 young adults (33 women, 12 men; M = 20.4, SD = 2.8; college students) performed an auditory lexical decision on 128 Spanish disyllabic words and 128 disyllabic nonwords. Words were selected so that lexical and first syllable frequency were manipulated in a within-subject 2 × 2 design, and six additional independent variables were controlled: token positional frequency of the second syllable, number of phonemes, position of lexical stress, number of phonological neighbors, number of phonological neighbors that have higher frequencies than the word, and acoustical durations measured in milliseconds. Decision latencies and error rates were submitted to linear mixed models analysis. Results showed a typical facilitatory effect of the lexical frequency and, importantly, an inhibitory effect of the first syllable frequency on reaction times and error rates. © The Author(s) 2016.

  3. Word frequencies in written and spoken English based on the British National Corpus

    CERN Document Server

    Leech, Geoffrey; Wilson, Andrew (All Of Lancaster University)


    Word Frequencies in Written and Spoken English is a landmark volume in the development of vocabulary frequency studies. Whereas previous books have in general given frequency information about the written language only, this book provides information on both speech and writing. It not only gives information about the language as a whole, but also about the differences between spoken and written English, and between different spoken and written varieties of the language. The frequencies are derived from a wide ranging and up-to-date corpus of English: the British Na

  4. The language issue and academic performance at a South African ...

    African Journals Online (AJOL)

    Academic performance at universities in South Africa is a cause of concern. It is widely acknowledged that there are a variety of factors that contribute to poor academic performance, but language is regarded as one of the most important issues in this discussion. In this article, the relationship between language and ...

  5. Spoken word recognition without a TRACE (United States)

    Hannagan, Thomas; Magnuson, James S.; Grainger, Jonathan


    How do we map the rapid input of spoken language onto phonological and lexical representations over time? Attempts at psychologically-tractable computational models of spoken word recognition tend either to ignore time or to transform the temporal input into a spatial representation. TRACE, a connectionist model with broad and deep coverage of speech perception and spoken word recognition phenomena, takes the latter approach, using exclusively time-specific units at every level of representation. TRACE reduplicates featural, phonemic, and lexical inputs at every time step in a large memory trace, with rich interconnections (excitatory forward and backward connections between levels and inhibitory links within levels). As the length of the memory trace is increased, or as the phoneme and lexical inventory of the model is increased to a realistic size, this reduplication of time- (temporal position) specific units leads to a dramatic proliferation of units and connections, begging the question of whether a more efficient approach is possible. Our starting point is the observation that models of visual object recognition—including visual word recognition—have grappled with the problem of spatial invariance, and arrived at solutions other than a fully-reduplicative strategy like that of TRACE. This inspires a new model of spoken word recognition that combines time-specific phoneme representations similar to those in TRACE with higher-level representations based on string kernels: temporally independent (time invariant) diphone and lexical units. This reduces the number of necessary units and connections by several orders of magnitude relative to TRACE. Critically, we compare the new model to TRACE on a set of key phenomena, demonstrating that the new model inherits much of the behavior of TRACE and that the drastic computational savings do not come at the cost of explanatory power. PMID:24058349

  6. Processing spoken lectures in resource-scarce environments

    CSIR Research Space (South Africa)

    Van Heerden, CJ


    Full Text Available and then adapting or training new models using the segmented spoken lectures. The eventual systems perform quite well, aligning more than 90% of a selected set of target words successfully....

  7. Directionality effects in simultaneous language interpreting: the case of sign language interpreters in The Netherlands. (United States)

    Van Dijk, Rick; Boers, Eveline; Christoffels, Ingrid; Hermans, Daan


    The quality of interpretations produced by sign language interpreters was investigated. Twenty-five experienced interpreters were instructed to interpret narratives from (a) spoken Dutch to Sign Language of The Netherlands (SLN), (b) spoken Dutch to Sign Supported Dutch (SSD), and (c) SLN to spoken Dutch. The quality of the interpreted narratives was assessed by 5 certified sign language interpreters who did not participate in the study. Two measures were used to assess interpreting quality: the propositional accuracy of the interpreters' interpretations and a subjective quality measure. The results showed that the interpreted narratives in the SLN-to-Dutch interpreting direction were of lower quality (on both measures) than the interpreted narratives in the Dutch-to-SLN and Dutch-to-SSD directions. Furthermore, interpreters who had begun acquiring SLN when they entered the interpreter training program performed as well in all 3 interpreting directions as interpreters who had acquired SLN from birth.

  8. Evaluating the spoken English proficiency of graduates of foreign medical schools. (United States)

    Boulet, J R; van Zanten, M; McKinley, D W; Gary, N E


    The purpose of this study was to gather additional evidence for the validity and reliability of spoken English proficiency ratings provided by trained standardized patients (SPs) in high-stakes clinical skills examination. Over 2500 candidates who took the Educational Commission for Foreign Medical Graduates' (ECFMG) Clinical Skills Assessment (CSA) were studied. The CSA consists of 10 or 11 timed clinical encounters. Standardized patients evaluate spoken English proficiency and interpersonal skills in every encounter. Generalizability theory was used to estimate the consistency of spoken English ratings. Validity coefficients were calculated by correlating summary English ratings with CSA scores and other external criterion measures. Mean spoken English ratings were also compared by various candidate background variables. The reliability of the spoken English ratings, based on 10 independent evaluations, was high. The magnitudes of the associated variance components indicated that the evaluation of a candidate's spoken English proficiency is unlikely to be affected by the choice of cases or SPs used in a given assessment. Proficiency in spoken English was related to native language (English versus other) and scores from the Test of English as a Foreign Language (TOEFL). The pattern of the relationships, both within assessment components and with external criterion measures, suggests that valid measures of spoken English proficiency are obtained. This result, combined with the high reproducibility of the ratings over encounters and SPs, supports the use of trained SPs to measure spoken English skills in a simulated medical environment.

  9. Performance of Senior Tourism Students in Using Foreign Language

    Directory of Open Access Journals (Sweden)

    Dr. Dexter R. Buted


    Full Text Available The study generally intended to reckon the previous and present condition of senior tourism students with regards on their foreign language class. Specifically, it described the profile of the professors teaching foreign language; determined the senior tourism student’s performances on their foreign language class; assessed the teaching strategies used by the professors; tested the significant relationship between the performances of the students to the teaching strategies used; and lastly, proposed an action plan to help tourism students in the study of foreign language. The researchers used the descriptive method of research, with one hundred seventy-eight (178 respondents composed of all senior tourism students who are enrolled in foreign language class. The result of the study revealed that the professors who are teaching foreign language are 61 years old and above, masters degree holder, 10 years and above, with a unit of 21 and can speak Spanish. Also, the students are able to speak and comprehend Mandarin, French and Spanish. The teaching techniques used by the professors in teaching the language was giving and evaluating student’s performance more often. Moreover, the performances of the students in foreign language are affected by the teaching strategies used by the professors. And a proposed plan was formulated to improve foreign language subject of the study

  10. Spoken grammar awareness raising: Does it affect the listening ability of Iranian EFL learners?

    Directory of Open Access Journals (Sweden)

    Mojgan Rashtchi


    Full Text Available Advances in spoken corpora analysis have brought about new insights into language pedagogy and have led to an awareness of the characteristics of spoken language. Current findings have shown that grammar of spoken language is different from written language. However, most listening and speaking materials are concocted based on written grammar and lack core spoken language features. The aim of the present study was to explore the question whether awareness of spoken grammar features could affect learners’ comprehension of real-life conversations. To this end, 45 university students in two intact classes participated in a listening course employing corpus-based materials. The instruction of the spoken grammar features to the experimental group was done overtly through awareness raising tasks, whereas the control group, though exposed to the same materials, was not provided with such tasks for learning the features. The results of the independent samples t tests revealed that the learners in the experimental group comprehended everyday conversations much better than those in the control group. Additionally, the highly positive views of spoken grammar held by the learners, which was elicited by means of a retrospective questionnaire, were generally comparable to those reported in the literature.

  11. A Comparison between Written and Spoken Narratives in Aphasia (United States)

    Behrns, Ingrid; Wengelin, Asa; Broberg, Malin; Hartelius, Lena


    The aim of the present study was to explore how a personal narrative told by a group of eight persons with aphasia differed between written and spoken language, and to compare this with findings from 10 participants in a reference group. The stories were analysed through holistic assessments made by 60 participants without experience of aphasia…

  12. Lexicon Optimization for Dutch Speech Recognition in Spoken Document Retrieval

    NARCIS (Netherlands)

    Ordelman, Roeland J.F.; van Hessen, Adrianus J.; de Jong, Franciska M.G.

    In this paper, ongoing work concerning the language modelling and lexicon optimization of a Dutch speech recognition system for Spoken Document Retrieval is described: the collection and normalization of a training data set and the optimization of our recognition lexicon. Effects on lexical coverage

  13. Lexicon optimization for Dutch speech recognition in spoken document retrieval

    NARCIS (Netherlands)

    Ordelman, Roeland J.F.; van Hessen, Adrianus J.; de Jong, Franciska M.G.; Dalsgaard, P.; Lindberg, B.; Benner, H.


    In this paper, ongoing work concerning the language modelling and lexicon optimization of a Dutch speech recognition system for Spoken Document Retrieval is described: the collection and normalization of a training data set and the optimization of our recognition lexicon. Effects on lexical coverage

  14. Automated Metadata Extraction for Semantic Access to Spoken Word Archives

    NARCIS (Netherlands)

    de Jong, Franciska M.G.; Heeren, W.F.L.; van Hessen, Adrianus J.; Ordelman, Roeland J.F.; Nijholt, Antinus; Ruiz Miyares, L.; Alvarez Silva, M.R.


    Archival practice is shifting from the analogue to the digital world. A specific subset of heritage collections that impose interesting challenges for the field of language and speech technology are spoken word archives. Given the enormous backlog at audiovisual archives of unannotated materials and

  15. Error detection in spoken human-machine interaction

    NARCIS (Netherlands)

    Krahmer, E.J.; Swerts, M.G.J.; Theune, M.; Weegels, M.F.


    Given the state of the art of current language and speech technology, errors are unavoidable in present-day spoken dialogue systems. Therefore, one of the main concerns in dialogue design is how to decide whether or not the system has understood the user correctly. In human-human communication,

  16. Error detection in spoken human-machine interaction

    NARCIS (Netherlands)

    Krahmer, E.; Swerts, M.; Theune, Mariet; Weegels, M.

    Given the state of the art of current language and speech technology, errors are unavoidable in present-day spoken dialogue systems. Therefore, one of the main concerns in dialogue design is how to decide whether or not the system has understood the user correctly. In human-human communication,

  17. L1 and L2 Spoken Word Processing: Evidence from Divided Attention Paradigm. (United States)

    Shafiee Nahrkhalaji, Saeedeh; Lotfi, Ahmad Reza; Koosha, Mansour


    The present study aims to reveal some facts concerning first language (L 1 ) and second language (L 2 ) spoken-word processing in unbalanced proficient bilinguals using behavioral measures. The intention here is to examine the effects of auditory repetition word priming and semantic priming in first and second languages of these bilinguals. The other goal is to explore the effects of attention manipulation on implicit retrieval of perceptual and conceptual properties of spoken L 1 and L 2 words. In so doing, the participants performed auditory word priming and semantic priming as memory tests in their L 1 and L 2 . In a half of the trials of each experiment, they carried out the memory test while simultaneously performing a secondary task in visual modality. The results revealed that effects of auditory word priming and semantic priming were present when participants processed L 1 and L 2 words in full attention condition. Attention manipulation could reduce priming magnitude in both experiments in L 2 . Moreover, L 2 word retrieval increases the reaction times and reduces accuracy on the simultaneous secondary task to protect its own accuracy and speed.

  18. The influence of second language teaching on undergraduate mathematics performance (United States)

    Gerber, Ans; Engelbrecht, Johann; Harding, Ansie; Rogan, John


    Understanding abstract concepts and ideas in mathematics, if instruction takes place in the first language of the student, is difficult. Yet worldwide students often have to master mathematics via a second or third language. The majority of students in South Africa — a country with eleven official languages — has to face this difficulty. In a quantitative study of first year calculus students, we investigated two groups of students. For one group tuition took place in their home language; for the second group, tuition was in English, a second or even a third language. Performance data on their secondary mathematics and first year tertiary calculus were analysed. The study showed that there was no significant difference between the adjusted means of the entire group of first language learners and the entire group of second language learners. Neither was there any statistically significant difference between the performances of the two groups of second language learners (based on the adjusted means). Yet, there did seem to be a significant difference between the achievement of Afrikaans students attending Afrikaans lectures and Afrikaans students attending English lectures.

  19. Multi-Language Programming Environments for High Performance Java Computing


    Vladimir Getov; Paul Gray; Sava Mintchev; Vaidy Sunderam


    Recent developments in processor capabilities, software tools, programming languages and programming paradigms have brought about new approaches to high performance computing. A steadfast component of this dynamic evolution has been the scientific community’s reliance on established scientific packages. As a consequence, programmers of high‐performance applications are reluctant to embrace evolving languages such as Java. This paper describes the Java‐to‐C Interface (JCI) tool which provides ...

  20. User-Centred Design for Chinese-Oriented Spoken English Learning System (United States)

    Yu, Ping; Pan, Yingxin; Li, Chen; Zhang, Zengxiu; Shi, Qin; Chu, Wenpei; Liu, Mingzhuo; Zhu, Zhiting


    Oral production is an important part in English learning. Lack of a language environment with efficient instruction and feedback is a big issue for non-native speakers' English spoken skill improvement. A computer-assisted language learning system can provide many potential benefits to language learners. It allows adequate instructions and instant…


    Directory of Open Access Journals (Sweden)

    Gabriel de Ávila Othero


    Full Text Available In this paper, we present the results of our work with automatic morphological annotation of excerpts from a corpus of spoken language – belonging to the VARSUL project – using the free morphosyntatic tagger Aelius. We present 20 texts containing 154,530 words, annotated automatically and corrected manually. This paper presents the tagger Aelius and our work of manual review of the texts, as well as our suggestions for improvements of the tool, concerning aspects of oral texts. We verify the performance of morphosyntactic tagging a spoken language corpus, an unprecedented challenge for the tagger. Based on the errors of the tagger, we try to infer certain patterns of annotation to overcome limitations presented by the program, and we propose suggestions for implementations in order to allow Aelius to tag spoken language corpora in a more effective way, specially treating cases such as interjections, apheresis, onomatopeia and conversational markers.

  2. Vývoj sociální kognice českých neslyšících dětí — uživatelů českého znakového jazyka a uživatelů mluvené češtiny: adaptace testové baterie : Development of Social Cognition in Czech Deaf Children — Czech Sign Language Users and Czech Spoken Language Users: Adaptation of a Test Battery

    Directory of Open Access Journals (Sweden)

    Andrea Hudáková


    Full Text Available The present paper describes the process of an adaptation of a set of tasks for testing theory-of-mind competencies, Theory of Mind Task Battery, for the use with the population of Czech Deaf children — both users of Czech Sign Language as well as those using spoken Czech.

  3. Improving Language Learning Strategies and Performance of Pre-Service Language Teachers through a CALLA-TBLT Model (United States)

    Guapacha Chamorro, Maria Eugenia; Benavidez Paz, Luis Humberto


    This paper reports an action-research study on language learning strategies in tertiary education at a Colombian university. The study aimed at improving the English language performance and language learning strategies use of 33 first-year pre-service language teachers by combining elements from two models: the cognitive academic language…

  4. Developing a corpus of spoken language variability (United States)

    Carmichael, Lesley; Wright, Richard; Wassink, Alicia Beckford


    We are developing a novel, searchable corpus as a research tool for investigating phonetic and phonological phenomena across various speech styles. Five speech styles have been well studied independently in previous work: reduced (casual), careful (hyperarticulated), citation (reading), Lombard effect (speech in noise), and ``motherese'' (child-directed speech). Few studies to date have collected a wide range of styles from a single set of speakers, and fewer yet have provided publicly available corpora. The pilot corpus includes recordings of (1) a set of speakers participating in a variety of tasks designed to elicit the five speech styles, and (2) casual peer conversations and wordlists to illustrate regional vowels. The data include high-quality recordings and time-aligned transcriptions linked to text files that can be queried. Initial measures drawn from the database provide comparison across speech styles along the following acoustic dimensions: MLU (changes in unit duration); relative intra-speaker intensity changes (mean and dynamic range); and intra-speaker pitch values (minimum, maximum, mean, range). The corpus design will allow for a variety of analyses requiring control of demographic and style factors, including hyperarticulation variety, disfluencies, intonation, discourse analysis, and detailed spectral measures.

  5. Circadian Rhythms and Second Language Performance (United States)

    de Bot, Kees; Fang, Fang


    Human behavior is not constant over the hours of the day, and there are considerable individual differences. Some people raise early and go to bed early and have their peek performance early in the day ("larks") while others tend to go to bed late and get up late and have their best performance later in the day ("owls"). In…

  6. Towards Adaptive Spoken Dialog Systems

    CERN Document Server

    Schmitt, Alexander


    In Monitoring Adaptive Spoken Dialog Systems, authors Alexander Schmitt and Wolfgang Minker investigate statistical approaches that allow for recognition of negative dialog patterns in Spoken Dialog Systems (SDS). The presented stochastic methods allow a flexible, portable and  accurate use.  Beginning with the foundations of machine learning and pattern recognition, this monograph examines how frequently users show negative emotions in spoken dialog systems and develop novel approaches to speech-based emotion recognition using hybrid approach to model emotions. The authors make use of statistical methods based on acoustic, linguistic and contextual features to examine the relationship between the interaction flow and the occurrence of emotions using non-acted  recordings several thousand real users from commercial and non-commercial SDS. Additionally, the authors present novel statistical methods that spot problems within a dialog based on interaction patterns. The approaches enable future SDS to offer m...

  7. Neural Correlates of High Performance in Foreign Language Vocabulary Learning (United States)

    Macedonia, Manuela; Muller, Karsten; Friederici, Angela D.


    Learning vocabulary in a foreign language is a laborious task which people perform with varying levels of success. Here, we investigated the neural underpinning of high performance on this task. In a within-subjects paradigm, participants learned 92 vocabulary items under two multimodal conditions: one condition paired novel words with iconic…

  8. Potential Ambiguity Translation Performances within Legal Language Institutional Nomenclature

    Directory of Open Access Journals (Sweden)

    Oţăt Diana


    Full Text Available Motivated by a paradoxical corollary of ambiguities in legal documents and especially in contract texts, the current paper underpins a dichotomy approach to unintended ambiguities aiming to establish a referential framework for the occurrence rate of translation ambiguities within the legal language nomenclature. The research focus is on a twofold situation since ambiguities may. on the one hand, arise dining the translation process, generated by the translator’s lack of competence, i.e. inadequate use of English regarding the special nature of legal language, or. on the other hand, they may be simply transferred from the source language into the target language without even noticing the potential ambiguous situation, i.e. culture-bound ambiguities. Hence, the paper proposes a contrastive analysis in order to localize the occurrence of lexical, structural, and socio-cultural ambiguities triggered by the use of the term performance and its Romanian equivalents in a number of sales contracts.


    Directory of Open Access Journals (Sweden)

    I Nengah Sudipa


    Full Text Available This article investigates the spoken ability for German students using Bahasa Indonesia (BI. They have studied it for six weeks in IBSN Program at Udayana University, Bali-Indonesia. The data was collected at the time the students sat for the mid-term oral test and was further analyzed with reference to the standard usage of BI. The result suggests that most students managed to express several concepts related to (1 LOCATION; (2 TIME; (3 TRANSPORT; (4 PURPOSE; (5 TRANSACTION; (6 IMPRESSION; (7 REASON; (8 FOOD AND BEVERAGE, and (9 NUMBER AND PERSON. The only problem few students might encounter is due to the influence from their own language system called interference, especially in word order.

  10. Recognizing Young Readers' Spoken Questions (United States)

    Chen, Wei; Mostow, Jack; Aist, Gregory


    Free-form spoken input would be the easiest and most natural way for young children to communicate to an intelligent tutoring system. However, achieving such a capability poses a challenge both to instruction design and to automatic speech recognition. To address the difficulties of accepting such input, we adopt the framework of predictable…

  11. Improving Language Learning Strategies and Performance of Pre-Service Language Teachers Through a CALLA-TBLT Model

    Directory of Open Access Journals (Sweden)

    Maria Eugenia Guapacha Chamorro


    Full Text Available This paper reports an action-research study on language learning strategies in tertiary education at a Colombian university. The study aimed at improving the English language performance and language learning strategies use of 33 first-year pre-service language teachers by combining elements from two models: the cognitive academic language learning approach and task-based language teaching. Data were gathered through surveys, a focus group, students’ and teachers’ journals, language tests, and documentary analysis. Results evidenced that the students improved in speaking, writing, grammar, vocabulary and in their language learning strategies repertoire. As a conclusion, explicit strategy instruction in the proposed model resulted in a proper combination to improve learners’ language learning strategies and performance.

  12. Ragnar Rommetveit's Approach to Everyday Spoken Dialogue from Within. (United States)

    Kowal, Sabine; O'Connell, Daniel C


    The following article presents basic concepts and methods of Ragnar Rommetveit's (born 1924) hermeneutic-dialogical approach to everyday spoken dialogue with a focus on both shared consciousness and linguistically mediated meaning. He developed this approach originally in his engagement of mainstream linguistic and psycholinguistic research of the 1960s and 1970s. He criticized this research tradition for its individualistic orientation and its adherence to experimental methodology which did not allow the engagement of interactively established meaning and understanding in everyday spoken dialogue. As a social psychologist influenced by phenomenological philosophy, Rommetveit opted for an alternative conceptualization of such dialogue as a contextualized, partially private world, temporarily co-established by interlocutors on the basis of shared consciousness. He argued that everyday spoken dialogue should be investigated from within, i.e., from the perspectives of the interlocutors and from a psychology of the second person. Hence, he developed his approach with an emphasis on intersubjectivity, perspectivity and perspectival relativity, meaning potential of utterances, and epistemic responsibility of interlocutors. In his methods, he limited himself for the most part to casuistic analyses, i.e., logical analyses of fictitious examples to argue for the plausibility of his approach. After many years of experimental research on language, he pursued his phenomenologically oriented research on dialogue in English-language publications from the late 1980s up to 2003. During that period, he engaged psycholinguistic research on spoken dialogue carried out by Anglo-American colleagues only occasionally. Although his work remained unfinished and open to development, it provides both a challenging alternative and supplement to current Anglo-American research on spoken dialogue and some overlap therewith.

  13. Criteria for the segmentation of spoken input into individual utterances


    Mast, Marion; Maier, Elisabeth; Schmitz, Birte


    This report describes how spoken language turns are segmented into utterances in the framework of the verbmobil project. The problem of segmenting turns is directly related to the task of annotating a discourse with dialogue act information: an utterance can be characterized as a stretch of dialogue that is attributed one dialogue act. Unfortunately, this rule in many cases is insufficient and many doubtful cases remain. We tried to at least reduce the number of unclear cases by providing a n...

  14. Autonomous Language Learning on Twitter: Performing Affiliation with Target Language Users through #hashtags (United States)

    Solmaz, Osman


    The purpose of this study is to examine the potential of social networking sites for autonomous language learners, specifically the role of hashtag literacies in learners' affiliation performances with native speakers. Informed by ecological approach and guided by Zappavigna's (2012) concepts of "searchable talk" and "ambient…

  15. Phonological Short-Term Memory, Working Memory and Foreign Language Performance in Intensive Language Learning (United States)

    Kormos, Judit; Safar, Anna


    In our research we addressed the question what the relationship is between phonological short-term and working memory capacity and performance in an end-of-year reading, writing, listening, speaking and use of English test. The participants of our study were 121 secondary school students aged 15-16 in the first intensive language training year of…


    NARCIS (Netherlands)

    Marti, U-V.; Bunke, H.


    In this paper we present a number of language models and their behavior in the recognition of unconstrained handwritten English sentences. We use the perplexity to compare the different models and their prediction power, and relate it to the performance of a recognition system under different

  17. Four Language Skills Performance, Academic Achievement, and Learning Strategy Use in Preservice Teacher Training Programs (United States)

    Shawer, Saad Fathy


    This article examines the differences in language learning strategies (LLS) use between preservice teachers of English as a foreign language (EFL) and Arabic as a second language (ASL). It also examines the relationship between LLS use and language performance (academic achievement and four language skills) among ASL students. The study made use…

  18. A Prerequisite to L1 Homophone Effects in L2 Spoken-Word Recognition (United States)

    Nakai, Satsuki; Lindsay, Shane; Ota, Mitsuhiko


    When both members of a phonemic contrast in L2 (second language) are perceptually mapped to a single phoneme in one's L1 (first language), L2 words containing a member of that contrast can spuriously activate L2 words in spoken-word recognition. For example, upon hearing cattle, Dutch speakers of English are reported to experience activation…

  19. Time course of Chinese monosyllabic spoken word recognition: evidence from ERP analyses. (United States)

    Zhao, Jingjing; Guo, Jingjing; Zhou, Fengying; Shu, Hua


    Evidence from event-related potential (ERP) analyses of English spoken words suggests that the time course of English word recognition in monosyllables is cumulative. Different types of phonological competitors (i.e., rhymes and cohorts) modulate the temporal grain of ERP components differentially (Desroches, Newman, & Joanisse, 2009). The time course of Chinese monosyllabic spoken word recognition could be different from that of English due to the differences in syllable structure between the two languages (e.g., lexical tones). The present study investigated the time course of Chinese monosyllabic spoken word recognition using ERPs to record brain responses online while subjects listened to spoken words. During the experiment, participants were asked to compare a target picture with a subsequent picture by judging whether or not these two pictures belonged to the same semantic category. The spoken word was presented between the two pictures, and participants were not required to respond during its presentation. We manipulated phonological competition by presenting spoken words that either matched or mismatched the target picture in one of the following four ways: onset mismatch, rime mismatch, tone mismatch, or syllable mismatch. In contrast to the English findings, our findings showed that the three partial mismatches (onset, rime, and tone mismatches) equally modulated the amplitudes and time courses of the N400 (a negative component that peaks about 400ms after the spoken word), whereas, the syllable mismatched words elicited an earlier and stronger N400 than the three partial mismatched words. The results shed light on the important role of syllable-level awareness in Chinese spoken word recognition and also imply that the recognition of Chinese monosyllabic words might rely more on global similarity of the whole syllable structure or syllable-based holistic processing rather than phonemic segment-based processing. We interpret the differences in spoken word

  20. Serbian heritage language schools in the Netherlands through the eyes of the parents

    NARCIS (Netherlands)

    Palmen, Andrej

    It is difficult to find the exact number of other languages spoken besides Dutch in the Netherlands. A study showed that a total of 96 other languages are spoken by students attending Dutch primary and secondary schools. The variety of languages spoken shows the growth of linguistic diversity in the

  1. Singing can facilitate foreign language learning. (United States)

    Ludke, Karen M; Ferreira, Fernanda; Overy, Katie


    This study presents the first experimental evidence that singing can facilitate short-term paired-associate phrase learning in an unfamiliar language (Hungarian). Sixty adult participants were randomly assigned to one of three "listen-and-repeat" learning conditions: speaking, rhythmic speaking, or singing. Participants in the singing condition showed superior overall performance on a collection of Hungarian language tests after a 15-min learning period, as compared with participants in the speaking and rhythmic speaking conditions. This superior performance was statistically significant (p sing" learning method can facilitate verbatim memory for spoken foreign language phrases.

  2. The influence of talker and foreign-accent variability on spoken word identification. (United States)

    Bent, Tessa; Holt, Rachael Frush


    In spoken word identification and memory tasks, stimulus variability from numerous sources impairs performance. In the current study, the influence of foreign-accent variability on spoken word identification was evaluated in two experiments. Experiment 1 used a between-subjects design to test word identification in noise in single-talker and two multiple-talker conditions: multiple talkers with the same accent and multiple talkers with different accents. Identification performance was highest in the single-talker condition, but there was no difference between the single-accent and multiple-accent conditions. Experiment 2 further explored word recognition for multiple talkers in single-accent versus multiple-accent conditions using a mixed design. A detriment to word recognition was observed in the multiple-accent condition compared to the single-accent condition, but the effect differed across the language backgrounds tested. These results demonstrate that the processing of foreign-accent variation may influence word recognition in ways similar to other sources of variability (e.g., speaking rate or style) in that the inclusion of multiple foreign accents can result in a small but significant performance decrement beyond the multiple-talker effect.

  3. Multi-Language Programming Environments for High Performance Java Computing

    Directory of Open Access Journals (Sweden)

    Vladimir Getov


    Full Text Available Recent developments in processor capabilities, software tools, programming languages and programming paradigms have brought about new approaches to high performance computing. A steadfast component of this dynamic evolution has been the scientific community’s reliance on established scientific packages. As a consequence, programmers of high‐performance applications are reluctant to embrace evolving languages such as Java. This paper describes the Java‐to‐C Interface (JCI tool which provides application programmers wishing to use Java with immediate accessibility to existing scientific packages. The JCI tool also facilitates rapid development and reuse of existing code. These benefits are provided at minimal cost to the programmer. While beneficial to the programmer, the additional advantages of mixed‐language programming in terms of application performance and portability are addressed in detail within the context of this paper. In addition, we discuss how the JCI tool is complementing other ongoing projects such as IBM’s High‐Performance Compiler for Java (HPCJ and IceT’s metacomputing environment.

  4. How Do Typically Developing Deaf Children and Deaf Children with Autism Spectrum Disorder Use the Face When Comprehending Emotional Facial Expressions in British Sign Language? (United States)

    Denmark, Tanya; Atkinson, Joanna; Campbell, Ruth; Swettenham, John


    Facial expressions in sign language carry a variety of communicative features. While emotion can modulate a spoken utterance through changes in intonation, duration and intensity, in sign language specific facial expressions presented concurrently with a manual sign perform this function. When deaf adult signers cannot see facial features, their…

  5. Mixed-Language High-Performance Computing for Plasma Simulations

    Directory of Open Access Journals (Sweden)

    Quanming Lu


    Full Text Available Java is receiving increasing attention as the most popular platform for distributed computing. However, programmers are still reluctant to embrace Java as a tool for writing scientific and engineering applications due to its still noticeable performance drawbacks compared with other programming languages such as Fortran or C. In this paper, we present a hybrid Java/Fortran implementation of a parallel particle-in-cell (PIC algorithm for plasma simulations. In our approach, the time-consuming components of this application are designed and implemented as Fortran subroutines, while less calculation-intensive components usually involved in building the user interface are written in Java. The two types of software modules have been glued together using the Java native interface (JNI. Our mixed-language PIC code was tested and its performance compared with pure Java and Fortran versions of the same algorithm on a Sun E6500 SMP system and a Linux cluster of Pentium~III machines.

  6. Language interoperability for high-performance parallel scientific components

    International Nuclear Information System (INIS)

    Elliot, N; Kohn, S; Smolinski, B


    With the increasing complexity and interdisciplinary nature of scientific applications, code reuse is becoming increasingly important in scientific computing. One method for facilitating code reuse is the use of components technologies, which have been used widely in industry. However, components have only recently worked their way into scientific computing. Language interoperability is an important underlying technology for these component architectures. In this paper, we present an approach to language interoperability for a high-performance parallel, component architecture being developed by the Common Component Architecture (CCA) group. Our approach is based on Interface Definition Language (IDL) techniques. We have developed a Scientific Interface Definition Language (SIDL), as well as bindings to C and Fortran. We have also developed a SIDL compiler and run-time library support for reference counting, reflection, object management, and exception handling (Babel). Results from using Babel to call a standard numerical solver library (written in C) from C and Fortran show that the cost of using Babel is minimal, where as the savings in development time and the benefits of object-oriented development support for C and Fortran far outweigh the costs

  7. Introducing Spoken Dialogue Systems into Intelligent Environments

    CERN Document Server

    Heinroth, Tobias


    Introducing Spoken Dialogue Systems into Intelligent Environments outlines the formalisms of a novel knowledge-driven framework for spoken dialogue management and presents the implementation of a model-based Adaptive Spoken Dialogue Manager(ASDM) called OwlSpeak. The authors have identified three stakeholders that potentially influence the behavior of the ASDM: the user, the SDS, and a complex Intelligent Environment (IE) consisting of various devices, services, and task descriptions. The theoretical foundation of a working ontology-based spoken dialogue description framework, the prototype implementation of the ASDM, and the evaluation activities that are presented as part of this book contribute to the ongoing spoken dialogue research by establishing the fertile ground of model-based adaptive spoken dialogue management. This monograph is ideal for advanced undergraduate students, PhD students, and postdocs as well as academic and industrial researchers and developers in speech and multimodal interactive ...

  8. L[subscript 1] and L[subscript 2] Spoken Word Processing: Evidence from Divided Attention Paradigm (United States)

    Shafiee Nahrkhalaji, Saeedeh; Lotfi, Ahmad Reza; Koosha, Mansour


    The present study aims to reveal some facts concerning first language (L[subscript 1]) and second language (L[subscript 2]) spoken-word processing in unbalanced proficient bilinguals using behavioral measures. The intention here is to examine the effects of auditory repetition word priming and semantic priming in first and second languages of…

  9. Alpha and theta brain oscillations index dissociable processes in spoken word recognition. (United States)

    Strauß, Antje; Kotz, Sonja A; Scharinger, Mathias; Obleser, Jonas


    Slow neural oscillations (~1-15 Hz) are thought to orchestrate the neural processes of spoken language comprehension. However, functional subdivisions within this broad range of frequencies are disputed, with most studies hypothesizing only about single frequency bands. The present study utilizes an established paradigm of spoken word recognition (lexical decision) to test the hypothesis that within the slow neural oscillatory frequency range, distinct functional signatures and cortical networks can be identified at least for theta- (~3-7 Hz) and alpha-frequencies (~8-12 Hz). Listeners performed an auditory lexical decision task on a set of items that formed a word-pseudoword continuum: ranging from (1) real words over (2) ambiguous pseudowords (deviating from real words only in one vowel; comparable to natural mispronunciations in speech) to (3) pseudowords (clearly deviating from real words by randomized syllables). By means of time-frequency analysis and spatial filtering, we observed a dissociation into distinct but simultaneous patterns of alpha power suppression and theta power enhancement. Alpha exhibited a parametric suppression as items increasingly matched real words, in line with lowered functional inhibition in a left-dominant lexical processing network for more word-like input. Simultaneously, theta power in a bilateral fronto-temporal network was selectively enhanced for ambiguous pseudowords only. Thus, enhanced alpha power can neurally 'gate' lexical integration, while enhanced theta power might index functionally more specific ambiguity-resolution processes. To this end, a joint analysis of both frequency bands provides neural evidence for parallel processes in achieving spoken word recognition. Copyright © 2014 Elsevier Inc. All rights reserved.

  10. Bilingualism alters brain functional connectivity between "control" regions and "language" regions: Evidence from bimodal bilinguals. (United States)

    Li, Le; Abutalebi, Jubin; Zou, Lijuan; Yan, Xin; Liu, Lanfang; Feng, Xiaoxia; Wang, Ruiming; Guo, Taomei; Ding, Guosheng


    Previous neuroimaging studies have revealed that bilingualism induces both structural and functional neuroplasticity in the dorsal anterior cingulate cortex (dACC) and the left caudate nucleus (LCN), both of which are associated with cognitive control. Since these "control" regions should work together with other language regions during language processing, we hypothesized that bilingualism may also alter the functional interaction between the dACC/LCN and language regions. Here we tested this hypothesis by exploring the functional connectivity (FC) in bimodal bilinguals and monolinguals using functional MRI when they either performed a picture naming task with spoken language or were in resting state. We found that for bimodal bilinguals who use spoken and sign languages, the FC of the dACC with regions involved in spoken language (e.g. the left superior temporal gyrus) was stronger in performing the task, but weaker in the resting state as compared to monolinguals. For the LCN, its intrinsic FC with sign language regions including the left inferior temporo-occipital part and right inferior and superior parietal lobules was increased in the bilinguals. These results demonstrate that bilingual experience may alter the brain functional interaction between "control" regions and "language" regions. For different control regions, the FC alters in different ways. The findings also deepen our understanding of the functional roles of the dACC and LCN in language processing. Copyright © 2015 Elsevier Ltd. All rights reserved.

  11. Biomechanically Preferred Consonant-Vowel Combinations Fail to Appear in Adult Spoken Corpora (United States)

    Whalen, D. H.; Giulivi, Sara; Nam, Hosung; Levitt, Andrea G.; Hallé, Pierre; Goldstein, Louis M.


    Certain consonant/vowel (CV) combinations are more frequent than would be expected from the individual C and V frequencies alone, both in babbling and, to a lesser extent, in adult language, based on dictionary counts: Labial consonants co-occur with central vowels more often than chance would dictate; coronals co-occur with front vowels, and velars with back vowels (Davis & MacNeilage, 1994). Plausible biomechanical explanations have been proposed, but it is also possible that infants are mirroring the frequency of the CVs that they hear. As noted, previous assessments of adult language were based on dictionaries; these “type” counts are incommensurate with the babbling measures, which are necessarily “token” counts. We analyzed the tokens in two spoken corpora for English, two for French and one for Mandarin. We found that the adult spoken CV preferences correlated with the type counts for Mandarin and French, not for English. Correlations between the adult spoken corpora and the babbling results had all three possible outcomes: significantly positive (French), uncorrelated (Mandarin), and significantly negative (English). There were no correlations of the dictionary data with the babbling results when we consider all nine combinations of consonants and vowels. The results indicate that spoken frequencies of CV combinations can differ from dictionary (type) counts and that the CV preferences apparent in babbling are biomechanically driven and can ignore the frequencies of CVs in the ambient spoken language. PMID:23420980

  12. Modeling Longitudinal Changes in Older Adults’ Memory for Spoken Discourse: Findings from the ACTIVE Cohort (United States)

    Payne, Brennan R.; Gross, Alden L.; Parisi, Jeanine M.; Sisco, Shannon M.; Stine-Morrow, Elizabeth A. L.; Marsiske, Michael; Rebok, George W.


    Episodic memory shows substantial declines with advancing age, but research on longitudinal trajectories of spoken discourse memory (SDM) in older adulthood is limited. Using parallel process latent growth curve models, we examined 10 years of longitudinal data from the no-contact control group (N = 698) of the Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE) randomized controlled trial in order to test (a) the degree to which SDM declines with advancing age, (b) predictors of these age-related declines, and (c) the within-person relationship between longitudinal changes in SDM and longitudinal changes in fluid reasoning and verbal ability over 10 years, independent of age. Individuals who were younger, White, had more years of formal education, were male, and had better global cognitive function and episodic memory performance at baseline demonstrated greater levels of SDM on average. However, only age at baseline uniquely predicted longitudinal changes in SDM, such that declines accelerated with greater age. Independent of age, within-person decline in reasoning ability over the 10-year study period was substantially correlated with decline in SDM (r = .87). An analogous association with SDM did not hold for verbal ability. The findings suggest that longitudinal declines in fluid cognition are associated with reduced spoken language comprehension. Unlike findings from memory for written prose, preserved verbal ability may not protect against developmental declines in memory for speech. PMID:24304364


    Directory of Open Access Journals (Sweden)

    Arijana Krišković


    Full Text Available Th e characterization of metonymy as a conceptual tool for guiding inferencing in language has opened a new fi eld of study in cognitive linguistics and pragmatics. To appreciate the value of metonymy for pragmatic inferencing, metonymy should not be viewed as performing only its prototypical referential function. Metonymic mappings are operative in speech acts at the level of reference, predication, proposition and illocution. Th e aim of this paper is to study the role of metonymy in pragmatic inferencing in spoken discourse in televison interviews. Case analyses of authentic utterances classifi ed as illocutionary metonymies following the pragmatic typology of metonymic functions are presented. Th e inferencing processes are facilitated by metonymic connections existing between domains or subdomains in the same functional domain. It has been widely accepted by cognitive linguists that universal human knowledge and embodiment are essential for the interpretation of metonymy. Th is analysis points to the role of cultural background knowledge in understanding target meanings. All these aspects of metonymic connections are exploited in complex inferential processes in spoken discourse. In most cases, metaphoric mappings are also a part of utterance interpretation.

  14. The Arabic Natural Language Processing: Introduction and Challenges

    Directory of Open Access Journals (Sweden)

    Boukhatem Nadera


    Full Text Available Arabic is a Semitic language spoken by more than 330 million people as a native language, in an area extending from the Arabian/Persian Gulf in the East to the Atlantic Ocean in the West. Moreover, it is the language in which 1.4 billion Muslims around the world perform their daily prayers. Over the last few years, Arabic natural language processing (ANLP has gained increasing importance, and several state of the art systems have been developed for a wide range of applications.

  15. Minority Languages and Performative Genres: The Case of Irish Language Stand-Up Comedy (United States)

    Moriarty, Mairead


    This article will examine the potential for language change from the bottom-up given the new domains in which minority languages are present as a result of the process of language mobility. Drawing on a theoretical notion of sociolinguistic scales, this article aims to discuss how the position of the Irish language has been reconfigured. From this…

  16. Development and Relationships Between Phonological Awareness, Morphological Awareness and Word Reading in Spoken and Standard Arabic

    Directory of Open Access Journals (Sweden)

    Rachel Schiff


    Full Text Available This study addressed the development of and the relationship between foundational metalinguistic skills and word reading skills in Arabic. It compared Arabic-speaking children’s phonological awareness (PA, morphological awareness, and voweled and unvoweled word reading skills in spoken and standard language varieties separately in children across five grade levels from childhood to adolescence. Second, it investigated whether skills developed in the spoken variety of Arabic predict reading in the standard variety. Results indicate that although individual differences between students in PA are eliminated toward the end of elementary school in both spoken and standard language varieties, gaps in morphological awareness and in reading skills persisted through junior and high school years. The results also show that the gap in reading accuracy and fluency between Spoken Arabic (SpA and Standard Arabic (StA was evident in both voweled and unvoweled words. Finally, regression analyses showed that morphological awareness in SpA contributed to reading fluency in StA, i.e., children’s early morphological awareness in SpA explained variance in children’s gains in reading fluency in StA. These findings have important theoretical and practical contributions for Arabic reading theory in general and they extend the previous work regarding the cross-linguistic relevance of foundational metalinguistic skills in the first acquired language to reading in a second language, as in societal bilingualism contexts, or a second language variety, as in diglossic contexts.

  17. Development and Relationships Between Phonological Awareness, Morphological Awareness and Word Reading in Spoken and Standard Arabic (United States)

    Schiff, Rachel; Saiegh-Haddad, Elinor


    This study addressed the development of and the relationship between foundational metalinguistic skills and word reading skills in Arabic. It compared Arabic-speaking children’s phonological awareness (PA), morphological awareness, and voweled and unvoweled word reading skills in spoken and standard language varieties separately in children across five grade levels from childhood to adolescence. Second, it investigated whether skills developed in the spoken variety of Arabic predict reading in the standard variety. Results indicate that although individual differences between students in PA are eliminated toward the end of elementary school in both spoken and standard language varieties, gaps in morphological awareness and in reading skills persisted through junior and high school years. The results also show that the gap in reading accuracy and fluency between Spoken Arabic (SpA) and Standard Arabic (StA) was evident in both voweled and unvoweled words. Finally, regression analyses showed that morphological awareness in SpA contributed to reading fluency in StA, i.e., children’s early morphological awareness in SpA explained variance in children’s gains in reading fluency in StA. These findings have important theoretical and practical contributions for Arabic reading theory in general and they extend the previous work regarding the cross-linguistic relevance of foundational metalinguistic skills in the first acquired language to reading in a second language, as in societal bilingualism contexts, or a second language variety, as in diglossic contexts. PMID:29686633

  18. A high level language for a high performance computer (United States)

    Perrott, R. H.


    The proposed computational aerodynamic facility will join the ranks of the supercomputers due to its architecture and increased execution speed. At present, the languages used to program these supercomputers have been modifications of programming languages which were designed many years ago for sequential machines. A new programming language should be developed based on the techniques which have proved valuable for sequential programming languages and incorporating the algorithmic techniques required for these supercomputers. The design objectives for such a language are outlined.

  19. Digital Language Death (United States)

    Kornai, András


    Of the approximately 7,000 languages spoken today, some 2,500 are generally considered endangered. Here we argue that this consensus figure vastly underestimates the danger of digital language death, in that less than 5% of all languages can still ascend to the digital realm. We present evidence of a massive die-off caused by the digital divide. PMID:24167559

  20. Digital language death.

    Directory of Open Access Journals (Sweden)

    András Kornai

    Full Text Available Of the approximately 7,000 languages spoken today, some 2,500 are generally considered endangered. Here we argue that this consensus figure vastly underestimates the danger of digital language death, in that less than 5% of all languages can still ascend to the digital realm. We present evidence of a massive die-off caused by the digital divide.

  1. Rapid modulation of spoken word recognition by visual primes. (United States)

    Okano, Kana; Grainger, Jonathan; Holcomb, Phillip J


    In a masked cross-modal priming experiment with ERP recordings, spoken Japanese words were primed with words written in one of the two syllabary scripts of Japanese. An early priming effect, peaking at around 200ms after onset of the spoken word target, was seen in left lateral electrode sites for Katakana primes, and later effects were seen for both Hiragana and Katakana primes on the N400 ERP component. The early effect is thought to reflect the efficiency with which words in Katakana script make contact with sublexical phonological representations involved in spoken language comprehension, due to the particular way this script is used by Japanese readers. This demonstrates fast-acting influences of visual primes on the processing of auditory target words, and suggests that briefly presented visual primes can influence sublexical processing of auditory target words. The later N400 priming effects, on the other hand, most likely reflect cross-modal influences on activity at the level of whole-word phonology and semantics.

  2. Changes in N400 Topography Following Intensive Speech Language Therapy for Individuals with Aphasia (United States)

    Wilson, K. Ryan; O'Rourke, Heather; Wozniak, Linda A.; Kostopoulos, Ellina; Marchand, Yannick; Newman, Aaron J.


    Our goal was to characterize the effects of intensive aphasia therapy on the N400, an electrophysiological index of lexical-semantic processing. Immediately before and after 4 weeks of intensive speech-language therapy, people with aphasia performed a task in which they had to determine whether spoken words were a "match" or a "mismatch" to…

  3. Language Development in Children with Language Disorders: An Introduction to Skinner's Verbal Behavior and the Techniques for Initial Language Acquisition (United States)

    Casey, Laura Baylot; Bicard, David F.


    Language development in typically developing children has a very predictable pattern beginning with crying, cooing, babbling, and gestures along with the recognition of spoken words, comprehension of spoken words, and then one word utterances. This predictable pattern breaks down for children with language disorders. This article will discuss…

  4. Improving Written Language Performance of Adolescents with Asperger Syndrome (United States)

    Delano, Monica E


    The effects of a multicomponent intervention involving self-regulated strategy development delivered via video self-modeling on the written language performance of 3 students with Asperger syndrome were examined. During intervention sessions, each student watched a video of himself performing strategies for increasing the number of words written and the number of functional essay elements. He then wrote a persuasive essay. The number of words written and number of functional essay elements included in each essay were measured. Each student demonstrated gains in the number of words written and number of functional essay elements. Maintenance of treatment effects at follow-up varied across targets and participants. Implications for future research are suggested. PMID:17624076

  5. Multiclausal Utterances Aren't Just for Big Kids: A Framework for Analysis of Complex Syntax Production in Spoken Language of Preschool- and Early School-Age Children (United States)

    Arndt, Karen Barako; Schuele, C. Melanie


    Complex syntax production emerges shortly after the emergence of two-word combinations in oral language and continues to develop through the school-age years. This article defines a framework for the analysis of complex syntax in the spontaneous language of preschool- and early school-age children. The purpose of this article is to provide…

  6. Screening for Language Delay: Growth Trajectories of Language Ability in Low- and High-Performing Children (United States)

    Klem, Marianne; Hagtvet, Bente; Hulme, Charles; Gustafsson, Jan-Eric


    Purpose: This study investigated the stability and growth of preschool language skills and explores latent class analysis as an approach for identifying children at risk of language impairment. Method: The authors present data from a large-scale 2-year longitudinal study, in which 600 children were assessed with a language-screening tool…

  7. Word level language identification in online multilingual communication

    NARCIS (Netherlands)

    Nguyen, Dong-Phuong; Dogruoz, A. Seza


    Multilingual speakers switch between languages in online and spoken communication. Analyses of large scale multilingual data require automatic language identification at the word level. For our experiments with multilingual online discussions, we first tag the language of individual words using

  8. A performance analysis for evaluation of programming languages ...

    African Journals Online (AJOL)

    In Nigeria, several programming Languages exist from general purpose to special purpose programming languages that are used in one application domain. People always find difficulties about which programming language should be learnt and adopt to develop particular software. In this paper, three (3) most commonly ...

  9. Webster's word power better English grammar improve your written and spoken English

    CERN Document Server

    Kirkpatrick, Betty


    With questions and answer sections throughout, this book helps you to improve your written and spoken English through understanding the structure of the English language. This is a thorough and useful book with all parts of speech and grammar explained. Used by ELT self-study students.

  10. Between Syntax and Pragmatics: The Causal Conjunction Protože in Spoken and Written Czech

    Czech Academy of Sciences Publication Activity Database

    Čermáková, Anna; Komrsková, Zuzana; Kopřivová, Marie; Poukarová, Petra

    -, 25.04.2017 (2017), s. 393-414 ISSN 2509-9507 R&D Projects: GA ČR GA15-01116S Institutional support: RVO:68378092 Keywords : Causality * Discourse marker * Spoken language * Czech Subject RIV: AI - Linguistics OBOR OECD: Linguistics

  11. Language modeling for automatic speech recognition of inflective languages an applications-oriented approach using lexical data

    CERN Document Server

    Donaj, Gregor


    This book covers language modeling and automatic speech recognition for inflective languages (e.g. Slavic languages), which represent roughly half of the languages spoken in Europe. These languages do not perform as well as English in speech recognition systems and it is therefore harder to develop an application with sufficient quality for the end user. The authors describe the most important language features for the development of a speech recognition system. This is then presented through the analysis of errors in the system and the development of language models and their inclusion in speech recognition systems, which specifically address the errors that are relevant for targeted applications. The error analysis is done with regard to morphological characteristics of the word in the recognized sentences. The book is oriented towards speech recognition with large vocabularies and continuous and even spontaneous speech. Today such applications work with a rather small number of languages compared to the nu...

  12. Visual Sonority Modulates Infants' Attraction to Sign Language (United States)

    Stone, Adam; Petitto, Laura-Ann; Bosworth, Rain


    The infant brain may be predisposed to identify perceptually salient cues that are common to both signed and spoken languages. Recent theory based on spoken languages has advanced sonority as one of these potential language acquisition cues. Using a preferential looking paradigm with an infrared eye tracker, we explored visual attention of hearing…

  13. Bimodal Bilingual Language Development of Hearing Children of Deaf Parents (United States)

    Hofmann, Kristin; Chilla, Solveig


    Adopting a bimodal bilingual language acquisition model, this qualitative case study is the first in Germany to investigate the spoken and sign language development of hearing children of deaf adults (codas). The spoken language competence of six codas within the age range of 3;10 to 6;4 is assessed by a series of standardised tests (SETK 3-5,…

  14. English Foreign Language Learners Kinesics on Teaching Performance

    Directory of Open Access Journals (Sweden)

    Hajjah Zulianti


    Full Text Available This research entitle English foreign language learners kinesics on teaching performance aims to mention and to describe the forms and the function of kinesics used by EFL learners on teaching performance, and to describe the importance of kinesics in teaching activity. This research is descriptive qualitative research. The data of the research are taken from EFL learners’ teaching performance on sixth semester at STKIP PGRI Bandar Lampung. The researcher observes the learners’ kinesics in teaching activity by using observing method and noting technique. In analyzing the data, the researcher uses description method. The result shows that there are twenty kinds of kinesics acted by the trainee, those are sitting in relaxing, arms crossed in front of the chest, standing in relaxing, walking around the class, checking the time, stroking the chin or beard, smile, happily surprised, wrinkle forehead, nodding head, shaking head, thumbs up, pointing finger, counting hand, waving hand, looking up, eye following, squinting, look in eye and breaking or making eye contact. Keywords: Kinesics, EFL Learners, Teaching Performance

  15. Intelligent Performance Analysis with a Natural Language Interface (United States)

    Juuso, Esko K.


    Performance improvement is taken as the primary goal in the asset management. Advanced data analysis is needed to efficiently integrate condition monitoring data into the operation and maintenance. Intelligent stress and condition indices have been developed for control and condition monitoring by combining generalized norms with efficient nonlinear scaling. These nonlinear scaling methodologies can also be used to handle performance measures used for management since management oriented indicators can be presented in the same scale as intelligent condition and stress indices. Performance indicators are responses of the process, machine or system to the stress contributions analyzed from process and condition monitoring data. Scaled values are directly used in intelligent temporal analysis to calculate fluctuations and trends. All these methodologies can be used in prognostics and fatigue prediction. The meanings of the variables are beneficial in extracting expert knowledge and representing information in natural language. The idea of dividing the problems into the variable specific meanings and the directions of interactions provides various improvements for performance monitoring and decision making. The integrated temporal analysis and uncertainty processing facilitates the efficient use of domain expertise. Measurements can be monitored with generalized statistical process control (GSPC) based on the same scaling functions.

  16. 125 The Fading Phase of Igbo Language and Culture: Path to its ...

    African Journals Online (AJOL)


    favour of foreign language (and culture). They also ... native language, and children are unable to learn a language not spoken ... shielding them off their mother tongue”. ..... the effect endangered language has on the existence of the owners.

  17. Stability in Chinese and Malay heritage languages as a source of divergence

    NARCIS (Netherlands)

    Aalberse, S.; Moro, F.; Braunmüller, K.; Höder, S.; Kühl, K.


    This article discusses Malay and Chinese heritage languages as spoken in the Netherlands. Heritage speakers are dominant in another language and use their heritage language less. Moreover, they have qualitatively and quantitatively different input from monolinguals. Heritage languages are often

  18. Stability in Chinese and Malay heritage languages as a source of divergence

    NARCIS (Netherlands)

    Aalberse, S.; Moro, F.R.; Braunmüller, K.; Höder, S.; Kühl, K.


    This article discusses Malay and Chinese heritage languages as spoken in the Netherlands. Heritage speakers are dominant in another language and use their heritage language less. Moreover, they have qualitatively and quantitatively different input from monolinguals. Heritage languages are often

  19. Spoken Narrative Assessment: A Supplementary Measure of Children's Creativity (United States)

    Wong, Miranda Kit-Yi; So, Wing Chee


    This study developed a spoken narrative (i.e., storytelling) assessment as a supplementary measure of children's creativity. Both spoken and gestural contents of children's spoken narratives were coded to assess their verbal and nonverbal creativity. The psychometric properties of the coding system for the spoken narrative assessment were…

  20. Language

    DEFF Research Database (Denmark)

    Sanden, Guro Refsum


    Purpose: – The purpose of this paper is to analyse the consequences of globalisation in the area of corporate communication, and investigate how language may be managed as a strategic resource. Design/methodology/approach: – A review of previous studies on the effects of globalisation on corporate...... communication and the implications of language management initiatives in international business. Findings: – Efficient language management can turn language into a strategic resource. Language needs analyses, i.e. linguistic auditing/language check-ups, can be used to determine the language situation...... of a company. Language policies and/or strategies can be used to regulate a company’s internal modes of communication. Language management tools can be deployed to address existing and expected language needs. Continuous feedback from the front line ensures strategic learning and reduces the risk of suboptimal...

  1. The effect of Content and Language Integrated Learning (CLIL) on English performance and self-confidence

    NARCIS (Netherlands)

    Jansma, Marrit; Minnaert, Alexander; Klinkenberg, Edwin


    In this study, it was investigated whether third language teaching through Content and Language Integrated Learning (CLIL) was more effective than teaching a third language as an isolated subject. By means of a cross-sectional study design, English vocabulary, speaking performance and

  2. Assessing the Relation between Language Comprehension and Performance in General Chemistry (United States)

    Pyburn, Daniel T.; Pazicni, Samuel; Benassi, Victor A.; Tappin, Elizabeth E.


    Few studies have focused specifically on the role that language plays in learning chemistry. We report here an investigation into the ability of language comprehension measures to predict performance in university introductory chemistry courses. This work is informed by theories of language comprehension, which posit that high-skilled…

  3. Language and False-Belief Task Performance in Children with Autism Spectrum Disorder (United States)

    Farrar, M. Jeffrey; Seung, Hye Kyeung; Lee, Hyeonjin


    Purpose: Language is related to false-belief (FB) understanding in both typically developing children and children with autism spectrum disorder (ASD). The current study examined the role of complementation and general language in FB understanding. Of interest was whether language plays similar or different roles in the groups' FB performance.…

  4. Foreign Language Listening Anxiety and Listening Performance: Conceptualizations and Causal Relationships (United States)

    Zhang, Xian


    This study used structural equation modeling to explore the possible causal relations between foreign language (English) listening anxiety and English listening performance. Three hundred participants learning English as a foreign language (FL) completed the foreign language listening anxiety scale (FLLAS) and IELTS test twice with an interval of…

  5. Basic speech recognition for spoken dialogues

    CSIR Research Space (South Africa)

    Van Heerden, C


    Full Text Available Spoken dialogue systems (SDSs) have great potential for information access in the developing world. However, the realisation of that potential requires the solution of several challenging problems, including the development of sufficiently accurate...

  6. The gender congruency effect during bilingual spoken-word recognition (United States)

    Morales, Luis; Paolieri, Daniela; Dussias, Paola E.; Valdés kroff, Jorge R.; Gerfen, Chip; Bajo, María Teresa


    We investigate the ‘gender-congruency’ effect during a spoken-word recognition task using the visual world paradigm. Eye movements of Italian–Spanish bilinguals and Spanish monolinguals were monitored while they viewed a pair of objects on a computer screen. Participants listened to instructions in Spanish (encuentra la bufanda / ‘find the scarf’) and clicked on the object named in the instruction. Grammatical gender of the objects’ name was manipulated so that pairs of objects had the same (congruent) or different (incongruent) gender in Italian, but gender in Spanish was always congruent. Results showed that bilinguals, but not monolinguals, looked at target objects less when they were incongruent in gender, suggesting a between-language gender competition effect. In addition, bilinguals looked at target objects more when the definite article in the spoken instructions provided a valid cue to anticipate its selection (different-gender condition). The temporal dynamics of gender processing and cross-language activation in bilinguals are discussed. PMID:28018132

  7. Guest Comment: Universal Language Requirement. (United States)

    Sherwood, Bruce Arne


    Explains that reading English among Scientists is almost universal, however, there are enormous problems with spoken English. Advocates the use of Esperanto as a viable alternative, and as a language requirement for graduate work. (GA)

  8. Language, Thought and Memory in Linguistic Performance, A Thought View. (United States)

    Lado, Robert

    The contrasting views of Saussure and Bloomfield ("mentalist" versus "mechanist"), the hypotheses of Whorf showing the influence of language on certain habits of thought, and Chomsky's notion of generative transformational grammar in the context of language use are reviewed. The author notes the limits of these systems and suggests that in dealing…

  9. Effects of Auditory and Visual Priming on the Identification of Spoken Words. (United States)

    Shigeno, Sumi


    This study examined the effects of preceding contextual stimuli, either auditory or visual, on the identification of spoken target words. Fifty-one participants (29% males, 71% females; mean age = 24.5 years, SD = 8.5) were divided into three groups: no context, auditory context, and visual context. All target stimuli were spoken words masked with white noise. The relationships between the context and target stimuli were as follows: identical word, similar word, and unrelated word. Participants presented with context experienced a sequence of six context stimuli in the form of either spoken words or photographs. Auditory and visual context conditions produced similar results, but the auditory context aided word identification more than the visual context in the similar word relationship. We discuss these results in the light of top-down processing, motor theory, and the phonological system of language.

  10. Unsupervised Language Acquisition (United States)

    de Marcken, Carl


    This thesis presents a computational theory of unsupervised language acquisition, precisely defining procedures for learning language from ordinary spoken or written utterances, with no explicit help from a teacher. The theory is based heavily on concepts borrowed from machine learning and statistical estimation. In particular, learning takes place by fitting a stochastic, generative model of language to the evidence. Much of the thesis is devoted to explaining conditions that must hold for this general learning strategy to arrive at linguistically desirable grammars. The thesis introduces a variety of technical innovations, among them a common representation for evidence and grammars, and a learning strategy that separates the ``content'' of linguistic parameters from their representation. Algorithms based on it suffer from few of the search problems that have plagued other computational approaches to language acquisition. The theory has been tested on problems of learning vocabularies and grammars from unsegmented text and continuous speech, and mappings between sound and representations of meaning. It performs extremely well on various objective criteria, acquiring knowledge that causes it to assign almost exactly the same structure to utterances as humans do. This work has application to data compression, language modeling, speech recognition, machine translation, information retrieval, and other tasks that rely on either structural or stochastic descriptions of language.

  11. The reliability of language performance measurement in language sample analysis of children aged 5-6 years

    Directory of Open Access Journals (Sweden)

    Zahra Soleymani


    Full Text Available Background and Aim: The language sample analysis (LSA is more common in other languages than Persian to study language development and assess language pathology. We studied some psychometric properties of language sample analysis in this research such as content validity of written story and its pictures, test-retest reliability, and inter-rater reliability.Methods: We wrote a story based on Persian culture from Schneider’s study. The validity of written story and drawn pictures was approved by experts. To study test-retest reliability, 30 children looked at the pictures and told their own story twice with 7-10 days interval. Children generated the story themselves and tester did not give any cue about the story. Their audio-taped story was transcribed and analyzed. Sentence and word structures were detected in the analysis.Results: Mean of experts' agreement with the validity of written story was 92.28 percent. Experts scored the quality of pictures high and excellent. There was correlation between variables in sentence and word structure (p<0.05 in test-retest, except complex sentences (p=0.137. The agreement rate was 97.1 percent in inter-rater reliability assessment of transcription. The results of inter-rater reliability of language analysis showed that correlation coefficients were significant.Conclusion: The results confirmed that the tool was valid for eliciting language sample. The consistency of language performance in repeated measurement varied from mild to high in language sample analysis approach.

  12. Use of spoken and written Japanese did not protect Japanese-American men from cognitive decline in late life. (United States)

    Crane, Paul K; Gruhl, Jonathan C; Erosheva, Elena A; Gibbons, Laura E; McCurry, Susan M; Rhoads, Kristoffer; Nguyen, Viet; Arani, Keerthi; Masaki, Kamal; White, Lon


    Spoken bilingualism may be associated with cognitive reserve. Mastering a complicated written language may be associated with additional reserve. We sought to determine if midlife use of spoken and written Japanese was associated with lower rates of late life cognitive decline. Participants were second-generation Japanese-American men from the Hawaiian island of Oahu, born 1900-1919, free of dementia in 1991, and categorized based on midlife self-reported use of spoken and written Japanese (total n included in primary analysis = 2,520). Cognitive functioning was measured with the Cognitive Abilities Screening Instrument scored using item response theory. We used mixed effects models, controlling for age, income, education, smoking status, apolipoprotein E e4 alleles, and number of study visits. Rates of cognitive decline were not related to use of spoken or written Japanese. This finding was consistent across numerous sensitivity analyses. We did not find evidence to support the hypothesis that multilingualism is associated with cognitive reserve.

  13. Interferência da língua falada na escrita de crianças: processos de apagamento da oclusiva dental /d/ e da vibrante final /r/ Interference of the spoken language on children's writing: cancellation processes of the dental occlusive /d/ and final vibrant /r/

    Directory of Open Access Journals (Sweden)

    Socorro Cláudia Tavares de Sousa


    Full Text Available O presente trabalho tem como objetivo investigar a influência da língua falada na escrita de crianças em relação aos fenômenos do cancelamento da dental /d/ e da vibrante final /r/. Elaboramos e aplicamos um instrumento de pesquisa em alunos do Ensino Fundamental em escolas de Fortaleza. Para a análise dos dados obtidos, utilizamos o software SPSS. Os resultados nos revelaram que o sexo masculino e as palavras polissílabas são fatores que influenciam, de forma parcial, a realização da variável dependente /no/ e que os verbos e o nível de escolaridade são elementos condicionadores para o cancelamento da vibrante final /r/.The present study aims to investigate the influence of the spoken language in children's writing in relation to the phenomena of cancellation of dental /d/ and final vibrant /r/. We elaborated and applied a research instrument to children from primary school in Fortaleza. We used the software SPSS to analyze the data. The results showed that the male sex and the words which have three or more syllable are factors that influence, in part, the realization of the dependent variable /no/ and that verbs and level of education are conditioners elements for the cancellation of the final vibrant /r/.

  14. a performance analysis for evaluation of programming languages ...

    African Journals Online (AJOL)

    Mohammed et al.

    PROGRAMMING LANGUAGES BASED ON MOBILE COMPUTING. FOR NIGERIA ... Finally, is suitable for data Transfer using upload scheme. Keywords: ... INTRODUCTION .... java, Julia, python, matlab, mathematica and Ruby by.

  15. The Effects of the Classroom Performance System on Student Participation, Attendance, and Achievement (United States)

    Termos, Mohamad Hani


    The Classroom Performance System (CPS) is an instructional technology that increases student performance and promotes active learning. This study assessed the effect of the CPS on student participation, attendance, and achievement in multicultural college-level anatomy and physiology classes, where students' first spoken language is not English.…

  16. Implications of Hegel's Theories of Language on Second Language Teaching (United States)

    Wu, Manfred


    This article explores the implications of Hegel's theories of language on second language (L2) teaching. Three among the various concepts in Hegel's theories of language are selected. They are the crucial role of intersubjectivity; the primacy of the spoken over the written form; and the importance of the training of form or grammar. Applying…

  17. Inuit Sign Language: a contribution to sign language typology

    NARCIS (Netherlands)

    Schuit, J.; Baker, A.; Pfau, R.


    Sign language typology is a fairly new research field and typological classifications have yet to be established. For spoken languages, these classifications are generally based on typological parameters; it would thus be desirable to establish these for sign languages. In this paper, different

  18. A domain specific language for performance portable molecular dynamics algorithms (United States)

    Saunders, William Robert; Grant, James; Müller, Eike Hermann


    Developers of Molecular Dynamics (MD) codes face significant challenges when adapting existing simulation packages to new hardware. In a continuously diversifying hardware landscape it becomes increasingly difficult for scientists to be experts both in their own domain (physics/chemistry/biology) and specialists in the low level parallelisation and optimisation of their codes. To address this challenge, we describe a "Separation of Concerns" approach for the development of parallel and optimised MD codes: the science specialist writes code at a high abstraction level in a domain specific language (DSL), which is then translated into efficient computer code by a scientific programmer. In a related context, an abstraction for the solution of partial differential equations with grid based methods has recently been implemented in the (Py)OP2 library. Inspired by this approach, we develop a Python code generation system for molecular dynamics simulations on different parallel architectures, including massively parallel distributed memory systems and GPUs. We demonstrate the efficiency of the auto-generated code by studying its performance and scalability on different hardware and compare it to other state-of-the-art simulation packages. With growing data volumes the extraction of physically meaningful information from the simulation becomes increasingly challenging and requires equally efficient implementations. A particular advantage of our approach is the easy expression of such analysis algorithms. We consider two popular methods for deducing the crystalline structure of a material from the local environment of each atom, show how they can be expressed in our abstraction and implement them in the code generation framework.

  19. Spanish as a Second Language when L1 Is Quechua: Endangered Languages and the SLA Researcher (United States)

    Kalt, Susan E.


    Spanish is one of the most widely spoken languages in the world. Quechua is the largest indigenous language family to constitute the first language (L1) of second language (L2) Spanish speakers. Despite sheer number of speakers and typologically interesting contrasts, Quechua-Spanish second language acquisition is a nearly untapped research area,…

  20. Language performance in siblings of nonverbal children with autism. (United States)

    Levy, Yonata; Bar-Yuda, Chanit


    The study focuses on language and cognitive abilities of siblings of the linguistically most affected children with autism (i.e. siblings of nonverbal children - SIBS-ANV). Twenty-eight SIBS-ANV (17 boys), ages 4-9 years, took part in the study. All children attended regular schools, and none had received a diagnosis of autism. Controls were 27 typically developing children (SIBS-TD; 16 boys) matched to the SIBS-ANV on age, family background, socioeconomic status and type of school they attended. Significant IQ differences, as well as language differences as measured on the Clinical Evaluation of Language Fundamentals (CELF), emerged between SIBS-ANV and SIBS-TD. However, differences in the language scores mostly disappeared when PIQ and FSIQ were controlled for. Furthermore, grammatical analysis of spontaneous speech samples produced in the course of testing did not reveal any significant differences between the groups. These results add to recent work suggesting that language deficits may not be part of the Broad Autism Phenotype (BAP). It further suggests that the cognitive deficit characteristic of nonverbal people with autism may be familial.

  1. Visual-motor integration performance in children with severe specific language impairment. (United States)

    Nicola, K; Watter, P


    This study investigated (1) the visual-motor integration (VMI) performance of children with severe specific language impairment (SLI), and any effect of age, gender, socio-economic status and concomitant speech impairment; and (2) the relationship between language and VMI performance. It is hypothesized that children with severe SLI would present with VMI problems irrespective of gender and socio-economic status; however, VMI deficits will be more pronounced in younger children and those with concomitant speech impairment. Furthermore, it is hypothesized that there will be a relationship between VMI and language performance, particularly in receptive scores. Children enrolled between 2000 and 2008 in a school dedicated to children with severe speech-language impairments were included, if they met the criteria for severe SLI with or without concomitant speech impairment which was verified by a government organization. Results from all initial standardized language and VMI assessments found during a retrospective review of chart files were included. The final study group included 100 children (males = 76), from 4 to 14 years of age with mean language scores at least 2SD below the mean. For VMI performance, 52% of the children scored below -1SD, with 25% of the total group scoring more than 1.5SD below the mean. Age, gender and the addition of a speech impairment did not impact on VMI performance; however, children living in disadvantaged suburbs scored significantly better than children residing in advantaged suburbs. Receptive language scores of the Clinical Evaluation of Language Fundamentals was the only score associated with and able to predict VMI performance. A small subgroup of children with severe SLI will also have poor VMI skills. The best predictor of poor VMI is receptive language scores on the Clinical Evaluation of Language Fundamentals. Children with poor receptive language performance may benefit from VMI assessment and multidisciplinary

  2. Language Classroom Risk-Taking Behavior in a Performed Culture-Based Program


    Stephen D. Luft


    While several studies have investigated the role of risk-taking in language learning, the findings of these studies may not be generalizable to language learning where the performed culture approach (PCA) is used. This study describes the relationship between language learning and risk-taking in PCA, and the relationship between risk-taking and personal study habits, teaching style, daily grading, and classroom dynamics. Data were collected by means of a questionnaire. Th...


    Directory of Open Access Journals (Sweden)

    Puji Astuti


    Full Text Available The contribution of cooperative learning (CL in promoting second and foreign language learning has been widely acknowledged. Little scholarly attention, however, has been given to revealing how this teaching method works and promotes learners’ improved communicative competence. This qualitative case study explores the important role that individual accountability in CL plays in giving English as a Foreign Language (EFL learners in Indonesia the opportunity to use the target language of English. While individual accountability is a principle of and one of the activities in CL, it is currently under studied, thus little is known about how it enhances EFL learning. This study aims to address this gap by conducting a constructivist grounded theory analysis on participant observation, in-depth interview, and document analysis data drawn from two secondary school EFL teachers, 77 students in the observed classrooms, and four focal students. The analysis shows that through individual accountability in CL, the EFL learners had opportunities to use the target language, which may have contributed to the attainment of communicative competence—the goal of the EFL instruction. More specifically, compared to the use of conventional group work in the observed classrooms, through the activities of individual accountability in CL, i.e., performances and peer interaction, the EFL learners had more opportunities to use spoken English. The present study recommends that teachers, especially those new to CL, follow the preset procedure of selected CL instructional strategies or structures in order to recognize the activities within individual accountability in CL and understand how these activities benefit students.

  4. Structural borrowing: The case of Kenyan Sign Language (KSL) and ...

    African Journals Online (AJOL)

    Kenyan Sign Language (KSL) is a visual gestural language used by members of the deaf community in Kenya. Kiswahili on the other hand is a Bantu language that is used as the national language of Kenya. The two are world's apart, one being a spoken language and the other a signed language and thus their “… basic ...

  5. Language Classroom Risk-Taking Behavior in a Performed Culture-Based Program

    Directory of Open Access Journals (Sweden)

    Stephen D. Luft


    Full Text Available While several studies have investigated the role of risk-taking in language learning, the findings of these studies may not be generalizable to language learning where the performed culture approach (PCA is used. This study describes the relationship between language learning and risk-taking in PCA, and the relationship between risk-taking and personal study habits, teaching style, daily grading, and classroom dynamics. Data were collected by means of a questionnaire. This study finds that risk-taking behavior has a moderate positive relationship with student performance in PCA. While questionnaire items related to teaching style and classroom dynamics are not found to significantly correlate with students’ risk-taking behavior, some items related to daily grading and personal study habits are found to have a moderate positive relationship with risk-taking behavior. Based on these findings, it is recommended that further research investigate the relationship between assessment and risktaking in language learning. As second language acquisition researchers have investigated the role of affective variables in language learning, risk-taking has frequently been identified as a variable linked with success (Beebe, 1983; Ely, 1986; Naiman, Frolich, Stern, & Todesco, 1978; Rubin, 1975; Samimy & Pardin, 1994; Samimy & Tabuse, 1992. However, it is difficult to apply these findings to language classrooms that use the performed culture approach (PCA, an approach to the teaching of East Asian languages, for two reasons: (a PCA’s focus on the learning of a foreign culture could mean that greater risk is involved in 106 Luft language learning than in a typical language classroom; (b PCA creates a language learning experience for which the risks involved are different than those in language classrooms where other approaches are used.

  6. English Second Language, General, Special Education, and Speech/Language Personal Teacher Efficacy, English Language Arts Scientifically-Validated Intervention Practice, and Working Memory Development of English Language Learners in High and Low Performing Elementary Schools (United States)

    Brown, Barbara J.


    The researcher investigated teacher factors contributing to English language arts (ELA) achievement of English language learners (ELLs) over 2 consecutive years, in high and low performing elementary schools with a Hispanic/Latino student population greater than or equal to 30 percent. These factors included personal teacher efficacy, teacher…

  7. Listening Comprehension Performance Viewed in the Light of Emotional Intelligence and Foreign Language Listening Anxiety (United States)

    Valizadeh, Mohammad Reza; Alavinia, Parviz


    The researchers in the current study were after probing the potential relationship between emotional intelligence, foreign language listening anxiety (FLLA), and listening comprehension performance of Iranian EFL learners. To this end, 233 participants, studying English language and literature at three different Universities in Urmia, were…

  8. Semiotic Structure and Meaning Making: The Performance of English Language Learners on Mathematics Tests (United States)

    Solano-Flores, Guillermo; Barnett-Clarke, Carne; Kachchaf, Rachel R.


    We examined the performance of English language learners (ELLs) and non-ELLs on Grade 4 and Grade 5 mathematics content knowledge (CK) and academic language (AL) tests. CK and AL items had different semiotic loads (numbers of different types of semiotic features) and different semiotic structures (relative frequencies of different semiotic…

  9. Performance of African American Preschool Children from Low-Income Families on Expressive Language Measures (United States)

    Qi, Cathy H.; Kaiser, Ann P.; Marley, Scott C.; Milan, Stephanie


    The purposes of the study were to determine (a) the ability of two spontaneous language measures, mean length of utterance in morphemes (MLU-m) and number of different words (NDW), to identify African American preschool children at low and high levels of language ability; (b) whether child chronological age was related to the performance of either…

  10. Language Skills and Labor Market Performance of Immigrants in the Netherlands

    NARCIS (Netherlands)

    Yao, Y.; van Ours, J.C.


    Many immigrants in the Netherlands have poor Dutch language skills. They face problems in speaking and reading Dutch. Our paper investigates how these prob- lems affect their labor market performance in terms of employment, hours of work and wages. We find that for female immigrants language

  11. Investigating Sentence Processing and Language Segmentation in Explaining Children's Performance on a Sentence-Span Task (United States)

    Mainela-Arnold, Elina; Misra, Maya; Miller, Carol; Poll, Gerard H.; Park, Ji Sook


    Background: Children with poor language abilities tend to perform poorly on verbal working memory tasks. This result has been interpreted as evidence that limitations in working memory capacity may interfere with the development of a mature linguistic system. However, it is possible that language abilities, such as the efficiency of sentence…

  12. The Relationships among Group Size, Participation, and Performance of Programming Language Learning Supported with Online Forums (United States)

    Shaw, Ruey-Shiang


    This study examined the relationships among group size, participation, and learning performance factors when learning a programming language in a computer-supported collaborative learning (CSCL) context. An online forum was used as the CSCL environment for learning the Microsoft ASP.NET programming language. The collaborative-learning experiment…

  13. Mobile Information Access with Spoken Query Answering

    DEFF Research Database (Denmark)

    Brøndsted, Tom; Larsen, Henrik Legind; Larsen, Lars Bo


    window focused over the part which most likely contains an answer to the query. The two systems are integrated into a full spoken query answering system. The prototype can answer queries and questions within the chosen football (soccer) test domain, but the system has the flexibility for being ported...

  14. SPOKEN AYACUCHO QUECHUA, UNITS 11-20. (United States)



  15. SPOKEN CUZCO QUECHUA, UNITS 7-12. (United States)



  16. SPOKEN COCHABAMBA QUECHUA, UNITS 13-24. (United States)



  17. SPOKEN AYACUCHO QUECHUA. UNITS 1-10. (United States)



  18. A Grammar of Spoken Brazilian Portuguese. (United States)

    Thomas, Earl W.

    This is a first-year text of Portuguese grammar based on the Portuguese of moderately educated Brazilians from the area around Rio de Janeiro. Spoken idiomatic usage is emphasized. An important innovation is found in the presentation of verb tenses; they are presented in the order in which the native speaker learns them. The text is intended to…

  19. Towards Affordable Disclosure of Spoken Word Archives

    NARCIS (Netherlands)

    Ordelman, Roeland J.F.; Heeren, W.F.L.; Huijbregts, M.A.H.; Hiemstra, Djoerd; de Jong, Franciska M.G.; Larson, M; Fernie, K; Oomen, J; Cigarran, J.


    This paper presents and discusses ongoing work aiming at affordable disclosure of real-world spoken word archives in general, and in particular of a collection of recorded interviews with Dutch survivors of World War II concentration camp Buchenwald. Given such collections, the least we want to be

  20. Towards Affordable Disclosure of Spoken Heritage Archives

    NARCIS (Netherlands)

    Larson, M; Ordelman, Roeland J.F.; Heeren, W.F.L.; Fernie, K; de Jong, Franciska M.G.; Huijbregts, M.A.H.; Oomen, J; Hiemstra, Djoerd


    This paper presents and discusses ongoing work aiming at affordable disclosure of real-world spoken heritage archives in general, and in particular of a collection of recorded interviews with Dutch survivors of World War II concentration camp Buchenwald. Given such collections, we at least want to

  1. Mapping Students' Spoken Conceptions of Equality (United States)

    Anakin, Megan


    This study expands contemporary theorising about students' conceptions of equality. A nationally representative sample of New Zealand students' were asked to provide a spoken numerical response and an explanation as they solved an arithmetic additive missing number problem. Students' responses were conceptualised as acts of communication and…

  2. Parental mode of communication is essential for speech and language outcomes in cochlear implanted children

    DEFF Research Database (Denmark)

    Percy-Smith, Lone; Cayé-Thomasen, Per; Breinegaard, Nina


    The present study demonstrates a very strong effect of the parental communication mode on the auditory capabilities and speech/language outcome for cochlear implanted children. The children exposed to spoken language had higher odds of scoring high in all tests applied and the findings suggest...... a very clear benefit of spoken language communication with a cochlear implanted child....

  3. Business Spoken English Learning Strategies for Chinese Enterprise Staff

    Institute of Scientific and Technical Information of China (English)

    Han Li


    This study addresses the issue of promoting effective Business Spoken English of Enterprise Staff in China.It aims to assess the assessment of spoken English learning methods and identify the difficulties of learning English oral expression concerned business area.It also provides strategies for enhancing Enterprise Staff’s level of Business Spoken English.

  4. Language and Literacy: The Case of India. (United States)

    Sridhar, Kamal K.

    Language and literacy issues in India are reviewed in terms of background, steps taken to combat illiteracy, and some problems associated with literacy. The following facts are noted: India has 106 languages spoken by more than 685 million people, there are several minor script systems, a major language has different dialects, a language may use…

  5. Universal brain signature of proficient reading: Evidence from four contrasting languages. (United States)

    Rueckl, Jay G; Paz-Alonso, Pedro M; Molfese, Peter J; Kuo, Wen-Jui; Bick, Atira; Frost, Stephen J; Hancock, Roeland; Wu, Denise H; Mencl, William Einar; Duñabeitia, Jon Andoni; Lee, Jun-Ren; Oliver, Myriam; Zevin, Jason D; Hoeft, Fumiko; Carreiras, Manuel; Tzeng, Ovid J L; Pugh, Kenneth R; Frost, Ram


    We propose and test a theoretical perspective in which a universal hallmark of successful literacy acquisition is the convergence of the speech and orthographic processing systems onto a common network of neural structures, regardless of how spoken words are represented orthographically in a writing system. During functional MRI, skilled adult readers of four distinct and highly contrasting languages, Spanish, English, Hebrew, and Chinese, performed an identical semantic categorization task to spoken and written words. Results from three complementary analytic approaches demonstrate limited language variation, with speech-print convergence emerging as a common brain signature of reading proficiency across the wide spectrum of selected languages, whether their writing system is alphabetic or logographic, whether it is opaque or transparent, and regardless of the phonological and morphological structure it represents.

  6. Syllable frequency and word frequency effects in spoken and written word production in a non-alphabetic script

    Directory of Open Access Journals (Sweden)

    Qingfang eZhang


    Full Text Available The effects of word frequency and syllable frequency are well-established phenomena in domain such as spoken production in alphabetic languages. Chinese, as a non-alphabetic language, presents unique lexical and phonological properties in speech production. For example, the proximate unit of phonological encoding is syllable in Chinese but segments in Dutch, French or English. The present study investigated the effects of word frequency and syllable frequency, and their interaction in Chinese written and spoken production. Significant facilitatory word frequency and syllable frequency effects were observed in spoken as well as in written production. The syllable frequency effect in writing indicated that phonological properties (i.e., syllabic frequency constrain orthographic output via a lexical route, at least, in Chinese written production. However, the syllable frequency effect over repetitions was divergent in both modalities: it was significant in the former two repetitions in spoken whereas it was significant in the second repetition only in written. Due to the fragility of the syllable frequency effect in writing, we suggest that the phonological influence in handwritten production is not mandatory and universal, and it is modulated by experimental manipulations. This provides evidence for the orthographic autonomy hypothesis, rather than the phonological mediation hypothesis. The absence of an interaction between word frequency and syllable frequency showed that the syllable frequency effect is independent of the word frequency effect in spoken and written output modalities. The implications of these results on written production models are discussed.

  7. Cortical networks for vision and language in dyslexic and normal children of variable socio-economic status. (United States)

    Monzalvo, Karla; Fluss, Joel; Billard, Catherine; Dehaene, Stanislas; Dehaene-Lambertz, Ghislaine


    In dyslexia, anomalous activations have been described in both left temporo-parietal language cortices and in left ventral visual occipito-temporal cortex. However, the reproducibility, task-dependency, and presence of these brain anomalies in childhood rather than adulthood remain debated. We probed the large-scale organization of ventral visual and spoken language areas in dyslexic children using minimal target-detection tasks that were performed equally well by all groups. In 23 normal and 23 dyslexic 10-year-old children from two different socio-economic status (SES) backgrounds, we compared fMRI activity to visually presented houses, faces, and written strings, and to spoken sentences in the native or in a foreign language. Our results confirm a disorganization of both ventral visual and spoken language areas in dyslexic children. Visually, dyslexic children showed a normal lateral-to-medial mosaic of preferences, as well as normal responses to houses and checkerboards, but a reduced activation to words in the visual word form area (VWFA) and to faces in the right fusiform face area (FFA). Auditorily, dyslexic children exhibited reduced responses to speech in posterior temporal cortex, left insula and supplementary motor area, as well as reduced responses to maternal language in subparts of the planum temporale, left basal language area and VWFA. By correlating these two findings, we identify spoken-language predictors of VWFA activation to written words, which differ for dyslexic and normal readers. Similarities in fMRI deficits in both SES groups emphasize the existence of a core set of brain activation anomalies in dyslexia, regardless of culture, language and SES, without however resolving whether these anomalies are a cause or a consequence of impaired reading. Copyright © 2012 Elsevier Inc. All rights reserved.

  8. Advanced applications of natural language processing for performing information extraction

    CERN Document Server

    Rodrigues, Mário


    This book explains how can be created information extraction (IE) applications that are able to tap the vast amount of relevant information available in natural language sources: Internet pages, official documents such as laws and regulations, books and newspapers, and social web. Readers are introduced to the problem of IE and its current challenges and limitations, supported with examples. The book discusses the need to fill the gap between documents, data, and people, and provides a broad overview of the technology supporting IE. The authors present a generic architecture for developing systems that are able to learn how to extract relevant information from natural language documents, and illustrate how to implement working systems using state-of-the-art and freely available software tools. The book also discusses concrete applications illustrating IE uses.   ·         Provides an overview of state-of-the-art technology in information extraction (IE), discussing achievements and limitations for t...

  9. Working Memory Load Affects Processing Time in Spoken Word Recognition: Evidence from Eye-Movements (United States)

    Hadar, Britt; Skrzypek, Joshua E.; Wingfield, Arthur; Ben-David, Boaz M.


    In daily life, speech perception is usually accompanied by other tasks that tap into working memory capacity. However, the role of working memory on speech processing is not clear. The goal of this study was to examine how working memory load affects the timeline for spoken word recognition in ideal listening conditions. We used the “visual world” eye-tracking paradigm. The task consisted of spoken instructions referring to one of four objects depicted on a computer monitor (e.g., “point at the candle”). Half of the trials presented a phonological competitor to the target word that either overlapped in the initial syllable (onset) or at the last syllable (offset). Eye movements captured listeners' ability to differentiate the target noun from its depicted phonological competitor (e.g., candy or sandal). We manipulated working memory load by using a digit pre-load task, where participants had to retain either one (low-load) or four (high-load) spoken digits for the duration of a spoken word recognition trial. The data show that the high-load condition delayed real-time target discrimination. Specifically, a four-digit load was sufficient to delay the point of discrimination between the spoken target word and its phonological competitor. Our results emphasize the important role working memory plays in speech perception, even when performed by young adults in ideal listening conditions. PMID:27242424

  10. Componential Skills in Second Language Development of Bilingual Children with Specific Language Impairment (United States)

    Verhoeven, Ludo; Steenge, Judit; van Leeuwe, Jan; van Balkom, Hans


    In this study, we investigated which componential skills can be distinguished in the second language (L2) development of 140 bilingual children with specific language impairment in the Netherlands, aged 6-11 years, divided into 3 age groups. L2 development was assessed by means of spoken language tasks representing different language skills…

  11. Family Language Policy and School Language Choice: Pathways to Bilingualism and Multilingualism in a Canadian Context (United States)

    Slavkov, Nikolay


    This article reports on a survey with 170 school-age children growing up with two or more languages in the Canadian province of Ontario where English is the majority language, French is a minority language, and numerous other minority languages may be spoken by immigrant or Indigenous residents. Within this context the study focuses on minority…

  12. Fourth International Workshop on Spoken Dialog Systems

    CERN Document Server

    Rosset, Sophie; Garnier-Rizet, Martine; Devillers, Laurence; Natural Interaction with Robots, Knowbots and Smartphones : Putting Spoken Dialog Systems into Practice


    These proceedings presents the state-of-the-art in spoken dialog systems with applications in robotics, knowledge access and communication. It addresses specifically: 1. Dialog for interacting with smartphones; 2. Dialog for Open Domain knowledge access; 3. Dialog for robot interaction; 4. Mediated dialog (including crosslingual dialog involving Speech Translation); and, 5. Dialog quality evaluation. These articles were presented at the IWSDS 2012 workshop.

  13. Micro Language Planning and Cultural Renaissance in Botswana (United States)

    Alimi, Modupe M.


    Many African countries exhibit complex patterns of language use because of linguistic pluralism. The situation is often compounded by the presence of at least one foreign language that is either the official or second language. The language situation in Botswana depicts this complex pattern. Out of the 26 languages spoken in the country, including…

  14. Technology Assisted Language Learning is a silver bullet for enhancing Language competence and performance: A Case Study

    Directory of Open Access Journals (Sweden)

    Jameel Ahmad


    Full Text Available Technology Assisted Language Learning (TALL is an infallible means to develop profound knowledge and wide range of language skills. It instills in EFL learners an illimitable passion for task-based and skills oriented learning rather than rote memorization. New technological gadgets have commoditized a broad-based learning and teaching avenues and brought the whole learning process to life. A vast variety of authentic online- learning resources, motivational visual prompts, exciting videos, web-based interactivity and customizable language software, email, discussion forums, Skype, Twitter, apps, Internet mobiles, Facebook and YouTube have become obtrusive tools to enhance competence and performance in EFL teaching and learning realms. Technology can also provide various types of scaffolding for students learning to read. Nevertheless, instructors can also enhance their pedagogical effectiveness. However, the main focus of interest in this study is to ascertain to what extent the modern technological devices augment learners’ competence and performance specifically in vocabulary learning, grammatical accuracy and listening/ speaking skills. The remarkable scores of empirical surveys conducted in the present study reveal that TALL does assist learners to improve listening / speaking skills, pronunciation, extensive vocabulary and grammatical accuracy. The findings also manifest that the hybridity, instantaneity and super-diversity of digital learning lay far-reaching impact on learners' motivation for learning and incredibly maneuver learners to immerse in the whole learning process.

  15. Deaf children’s non-verbal working memory is impacted by their language experience

    Directory of Open Access Journals (Sweden)

    Chloe eMarshall


    Full Text Available Recent studies suggest that deaf children perform more poorly on working memory tasks compared to hearing children, but do not say whether this poorer performance arises directly from deafness itself or from deaf children’s reduced language exposure. The issue remains unresolved because findings come from (1 tasks that are verbal as opposed to non-verbal, and (2 involve deaf children who use spoken communication and therefore may have experienced impoverished input and delayed language acquisition. This is in contrast to deaf children who have been exposed to a sign language since birth from Deaf parents (and who therefore have native language-learning opportunities. A more direct test of how the type and quality of language exposure impacts working memory is to use measures of non-verbal working memory (NVWM and to compare hearing children with two groups of deaf signing children: those who have had native exposure to a sign language, and those who have experienced delayed acquisition compared to their native-signing peers. In this study we investigated the relationship between NVWM and language in three groups aged 6-11 years: hearing children (n=27, deaf native users of British Sign Language (BSL; n=7, and deaf children non native signers (n=19. We administered a battery of non-verbal reasoning, NVWM, and language tasks. We examined whether the groups differed on NVWM scores, and if language tasks predicted scores on NVWM tasks. For the two NVWM tasks, the non-native signers performed less accurately than the native signer and hearing groups (who did not differ from one another. Multiple regression analysis revealed that the vocabulary measure predicted scores on NVWM tasks. Our results suggest that whatever the language modality – spoken or signed – rich language experience from birth, and the good language skills that result from this early age of aacquisition, play a critical role in the development of NVWM and in performance on NVWM

  16. Schools and Languages in India. (United States)

    Harrison, Brian


    A brief review of Indian education focuses on special problems caused by overcrowded schools, insufficient funding, and the status of education itself in the Indian social structure. Language instruction in India, a complex issue due largely to the numerous official languages currently spoken, is commented on with special reference to the problem…

  17. Selected Topics from LVCSR Research for Asian Languages at Tokyo Tech (United States)

    Furui, Sadaoki

    This paper presents our recent work in regard to building Large Vocabulary Continuous Speech Recognition (LVCSR) systems for the Thai, Indonesian, and Chinese languages. For Thai, since there is no word boundary in the written form, we have proposed a new method for automatically creating word-like units from a text corpus, and applied topic and speaking style adaptation to the language model to recognize spoken-style utterances. For Indonesian, we have applied proper noun-specific adaptation to acoustic modeling, and rule-based English-to-Indonesian phoneme mapping to solve the problem of large variation in proper noun and English word pronunciation in a spoken-query information retrieval system. In spoken Chinese, long organization names are frequently abbreviated, and abbreviated utterances cannot be recognized if the abbreviations are not included in the dictionary. We have proposed a new method for automatically generating Chinese abbreviations, and by expanding the vocabulary using the generated abbreviations, we have significantly improved the performance of spoken query-based search.

  18. Estimating Spoken Dialog System Quality with User Models

    CERN Document Server

    Engelbrecht, Klaus-Peter


    Spoken dialog systems have the potential to offer highly intuitive user interfaces, as they allow systems to be controlled using natural language. However, the complexity inherent in natural language dialogs means that careful testing of the system must be carried out from the very beginning of the design process.   This book examines how user models can be used to support such early evaluations in two ways:  by running simulations of dialogs, and by estimating the quality judgments of users. First, a design environment supporting the creation of dialog flows, the simulation of dialogs, and the analysis of the simulated data is proposed.  How the quality of user simulations may be quantified with respect to their suitability for both formative and summative evaluation is then discussed. The remainder of the book is dedicated to the problem of predicting quality judgments of users based on interaction data. New modeling approaches are presented, which process the dialogs as sequences, and which allow knowl...

  19. Benchmarking and performance enhancement framework for multi-staging object-oriented languages

    Directory of Open Access Journals (Sweden)

    Ahmed H. Yousef


    Full Text Available This paper focuses on verifying the readiness, feasibility, generality and usefulness of multi-staging programming in software applications. We present a benchmark designed to evaluate the performance gain of different multi-staging programming (MSP languages implementations of object oriented languages. The benchmarks in this suite cover different tests that range from classic simple examples (like matrix algebra to advanced examples (like encryption and image processing. The benchmark is applied to compare the performance gain of two different MSP implementations (Mint and Metaphor that are built on object oriented languages (Java and C# respectively. The results concerning the application of this benchmark on these languages are presented and analysed. The measurement technique used in benchmarking leads to the development of a language independent performance enhancement framework that allows the programmer to select which code segments need staging. The framework also enables the programmer to verify the effectiveness of staging on the application performance. The framework is applied to a real case study. The case study results showed the effectiveness of the framework to achieve significant performance enhancement.

  20. Pronoun forms and courtesy in spoken language in Tunja, Colombia

    Directory of Open Access Journals (Sweden)

    Gloria Avendaño de Barón


    Full Text Available This article presents the results of a research project whose aims were the following: to determine the frequency of the use of pronoun forms in polite treatment sumercé, usted and tú, according to differences in gender, age and level of education, among speakers in Tunja; to describe the sociodiscursive variations and to explain the relationship between usage and courtesy. The methodology of the Project for the Sociolinguistic Study of Spanish in Spain and in Latin America (PRESEEA was used, and a sample of 54 speakers was taken. The results indicate that the most frequently used pronoun in Tunja to express friendliness and affection is sumercé, followed by usted and tú; women and men of different generations and levels of education alternate the use of these three forms in the context of narrative, descriptive, argumentative and explanatory speech.

  1. Assessing spoken-language educational interpreting: Measuring up ...

    African Journals Online (AJOL)

    Kate H

    10. 10. Content. 25. Conveying interaction between lecturer and students. 5. 5. Managing equipment. Breath control and volume. 5. Intonation and voice quality. Use of coping techniques. Pronunciation and general clarity. Error correction. Fluency of interpreting product (hesitations, silences etc.) Interpreting competency. 15.

  2. Endowing Spoken Language Dialogue System with Emotional Intelligence

    DEFF Research Database (Denmark)

    André, Elisabeth; Rehm, Matthias; Minker, Wolfgang


    While most dialogue systems restrict themselves to the adjustment of the propositional contents, our work concentrates on the generation of stylistic variations in order to improve the user’s perception of the interaction. To accomplish this goal, our approach integrates a social theory of polite...... of politeness with a cognitive theory of emotions. We propose a hierarchical selection process for politeness behaviors in order to enable the refinement of decisions in case additional context information becomes available....

  3. Usable, Real-Time, Interactive Spoken Language Systems (United States)


    Similarly, we included derivations (mostly plurals and possessives) of many open-class words in the domnain. We also added about 400 concatenated word...UueraiCe’l~ usinig a system of’ ’realization 1111C, %%. hiCh map) thle gr-aimmlatcal relation anl argumlent bears to the head onto thle semantic relatio ...syntactic categories as well. Representations of this form contain significantly more internal structure than specialized sublanguage models. This can be

  4. Cognitive Load Theory: An Empirical Study of Anxiety and Task Performance in Language Learning (United States)

    Chen, I-Jung; Chang, Chi-Cheng


    Introduction: This study explores the relationship among three variables--cognitive load, foreign language anxiety, and task performance. Cognitive load refers to the load imposed on working memory while performing a particular task. The authors hypothesized that anxiety consumes the resources of working memory, leaving less capacity for cognitive…

  5. A Performance-Based Teacher Education Curriculum in the Language Arts (United States)

    Rudman, Masha


    Under a feasibility grant awarded by the U.S. Department of Health, Education and Welfare for a Model Elementary Teacher Education Program (METEP), the University of Massachusetts' School of Education set up a language arts education program based on performance criteria, in that it is the performance of the student that is crucial, not the method…

  6. A Domain Specific Language for Performance Evaluation of Medical Imaging Systems

    NARCIS (Netherlands)

    van den Berg, Freek; Remke, Anne Katharina Ingrid; Haverkort, Boudewijn R.H.M.; Turau, Volker; Kwiatkowska, Marta; Mangharam, Rahul; Weyer, Christoph


    We propose iDSL, a domain specific language and toolbox for performance evaluation of Medical Imaging Systems. iDSL provides transformations to MoDeST models, which are in turn converted into UPPAAL and discrete-event MODES models. This enables automated performance evaluation by means of model

  7. Intensity of Multilingual Language Use Predicts Cognitive Performance in Some Multilingual Older Adults (United States)

    Keijzer, Merel; de Bot, Kees


    Cognitive advantages for bilinguals have inconsistently been observed in different populations, with different operationalisations of bilingualism, cognitive performance, and the process by which language control transfers to cognitive control. This calls for studies investigating which aspects of multilingualism drive a cognitive advantage, in which populations and under which conditions. This study reports on two cognitive tasks coupled with an extensive background questionnaire on health, wellbeing, personality, language knowledge and language use, administered to 387 older adults in the northern Netherlands, a small but highly multilingual area. Using linear mixed effects regression modeling, we find that when different languages are used frequently in different contexts, enhanced attentional control is observed. Subsequently, a PLS regression model targeting also other influential factors yielded a two-component solution whereby only more sensitive measures of language proficiency and language usage in different social contexts were predictive of cognitive performance above and beyond the contribution of age, gender, income and education. We discuss these findings in light of previous studies that try to uncover more about the nature of bilingualism and the cognitive processes that may drive an advantage. With an unusually large sample size our study advocates for a move away from dichotomous, knowledge-based operationalisations of multilingualism and offers new insights for future studies at the individual level. PMID:29783764


    Directory of Open Access Journals (Sweden)

    Elvira Küün


    Full Text Available The goal of this project in Estonia was to determine what languages are spoken by students from the 2nd to the 5th year of basic school at their homes in Tallinn, the capital of Estonia. At the same time, this problem was also studied in other segregated regions of Estonia: Kohtla-Järve and Maardu. According to the database of the population census from the year 2000 (Estonian Statistics Executive Office's census 2000, there are representatives of 142 ethnic groups living in Estonia, speaking a total of 109 native languages. At the same time, the database doesn’t state which languages are spoken at homes. The material presented in this article belongs to the research topic “Home Language of Basic School Students in Tallinn” from years 2007–2008, specifically financed and ordered by the Estonian Ministry of Education and Research (grant No. ETF 7065 in the framework of an international study called “Multilingual Project”. It was determined what language is dominating in everyday use, what are the factors for choosing the language for communication, what are the preferred languages and language skills. This study reflects the actual trends of the language situation in these cities.

  9. Lateralized theta wave connectivity and language performance in 2- to 5-year-old children. (United States)

    Kikuchi, Mitsuru; Shitamichi, Kiyomi; Yoshimura, Yuko; Ueno, Sanae; Remijn, Gerard B; Hirosawa, Tetsu; Munesue, Toshio; Tsubokawa, Tsunehisa; Haruta, Yasuhiro; Oi, Manabu; Higashida, Haruhiro; Minabe, Yoshio


    Recent neuroimaging studies support the view that a left-lateralized brain network is crucial for language development in children. However, no previous studies have demonstrated a clear link between lateralized brain functional network and language performance in preschool children. Magnetoencephalography (MEG) is a noninvasive brain imaging technique and is a practical neuroimaging method for use in young children. MEG produces a reference-free signal, and is therefore an ideal tool to compute coherence between two distant cortical rhythms. In the present study, using a custom child-sized MEG system, we investigated brain networks while 78 right-handed preschool human children (32-64 months; 96% were 3-4 years old) listened to stories with moving images. The results indicated that left dominance of parietotemporal coherence in theta band activity (6-8 Hz) was specifically correlated with higher performance of language-related tasks, whereas this laterality was not correlated with nonverbal cognitive performance, chronological age, or head circumference. Power analyses did not reveal any specific frequencies that contributed to higher language performance. Our results suggest that it is not the left dominance in theta oscillation per se, but the left-dominant phase-locked connectivity via theta oscillation that contributes to the development of language ability in young children.

  10. World Language Teacher Candidate Performance on Edtpa: An Exploratory Study (United States)

    Hildebrandt, Susan A.; Swanson, Pete


    Federal and state legislation continues to promote teacher accountability in the United States. The new edTPA, a subject-specific teacher performance assessment, is purported to measure beginning teacher readiness and is being pilot tested and implemented for licensure and certification decisions across the country. In this exploratory…

  11. Recording voiceover the spoken word in media

    CERN Document Server

    Blakemore, Tom


    The only book on the market to specifically address its audience, Recording Voiceover is the comprehensive guide for engineers looking to understand the aspects of capturing the spoken word.Discussing all phases of the recording session, Recording Voiceover addresses everything from microphone recommendations for voice recording to pre-production considerations, including setting up the studio, working with and directing the voice talent, and strategies for reducing or eliminating distracting noise elements found in human speech.Recording Voiceover features in-depth, specific recommendations f

  12. On the Usability of Spoken Dialogue Systems

    DEFF Research Database (Denmark)

    Larsen, Lars Bo

     This work is centred on the methods and problems associated with defining and measuring the usability of Spoken Dialogue Systems (SDS). The starting point is the fact that speech based interfaces has several times during the last 20 years fallen short of the high expectations and predictions held...... by industry, researchers and analysts. Several studies in the literature of SDS indicate that this can be ascribed to a lack of attention from the speech technology community towards the usability of such systems. The experimental results presented in this work are based on a field trial with the OVID home...

  13. The Functional Organisation of the Fronto-Temporal Language System: Evidence from Syntactic and Semantic Ambiguity (United States)

    Rodd, Jennifer M.; Longe, Olivia A.; Randall, Billi; Tyler, Lorraine K.


    Spoken language comprehension is known to involve a large left-dominant network of fronto-temporal brain regions, but there is still little consensus about how the syntactic and semantic aspects of language are processed within this network. In an fMRI study, volunteers heard spoken sentences that contained either syntactic or semantic ambiguities…

  14. Notes from the Field: Lolak--Another Moribund Language of Indonesia, with Supporting Audio (United States)

    Lobel, Jason William; Paputungan, Ade Tatak


    This paper consists of a short multimedia introduction to Lolak, a near-extinct Greater Central Philippine language traditionally spoken in three small communities on the island of Sulawesi in Indonesia. In addition to being one of the most underdocumented languages in the area, it is also spoken by one of the smallest native speaker populations…

  15. Bimodal bilingualism as multisensory training?: Evidence for improved audiovisual speech perception after sign language exposure. (United States)

    Williams, Joshua T; Darcy, Isabelle; Newman, Sharlene D


    The aim of the present study was to characterize effects of learning a sign language on the processing of a spoken language. Specifically, audiovisual phoneme comprehension was assessed before and after 13 weeks of sign language exposure. L2 ASL learners performed this task in the fMRI scanner. Results indicated that L2 American Sign Language (ASL) learners' behavioral classification of the speech sounds improved with time compared to hearing nonsigners. Results indicated increased activation in the supramarginal gyrus (SMG) after sign language exposure, which suggests concomitant increased phonological processing of speech. A multiple regression analysis indicated that learner's rating on co-sign speech use and lipreading ability was correlated with SMG activation. This pattern of results indicates that the increased use of mouthing and possibly lipreading during sign language acquisition may concurrently improve audiovisual speech processing in budding hearing bimodal bilinguals. Copyright © 2015 Elsevier B.V. All rights reserved.

  16. Teaching English as a "Second Language" in Kenya and the United States: Convergences and Divergences (United States)

    Roy-Campbell, Zaline M.


    English is spoken in five countries as the native language and in numerous other countries as an official language and the language of instruction. In countries where English is the native language, it is taught to speakers of other languages as an additional language to enable them to participate in all domains of life of that country. In many…

  17. Bridging the Gap: The Development of Appropriate Educational Strategies for Minority Language Communities in the Philippines (United States)

    Dekker, Diane; Young, Catherine


    There are more than 6000 languages spoken by the 6 billion people in the world today--however, those languages are not evenly divided among the world's population--over 90% of people globally speak only about 300 majority languages--the remaining 5700 languages being termed "minority languages". These languages represent the…

  18. Phonological Sketch of the Sida Language of Luang Namtha, Laos

    Directory of Open Access Journals (Sweden)

    Nathan Badenoch


    Full Text Available This paper describes the phonology of the Sida language, a Tibeto-Burman language spoken by approximately 3,900 people in Laos and Vietnam. The data presented here are the variety spoken in Luang Namtha province of northwestern Laos, and focuses on a synchronic description of the fundamentals of the Sida phonological systems. Several issues of diachronic interest are also discussed in the context of the diversity of the Southern Loloish group of languages, many of which are spoken in Laos and have not yet been described in detail.

  19. Semantic and phonological schema influence spoken word learning and overnight consolidation. (United States)

    Havas, Viktória; Taylor, Jsh; Vaquero, Lucía; de Diego-Balaguer, Ruth; Rodríguez-Fornells, Antoni; Davis, Matthew H


    We studied the initial acquisition and overnight consolidation of new spoken words that resemble words in the native language (L1) or in an unfamiliar, non-native language (L2). Spanish-speaking participants learned the spoken forms of novel words in their native language (Spanish) or in a different language (Hungarian), which were paired with pictures of familiar or unfamiliar objects, or no picture. We thereby assessed, in a factorial way, the impact of existing knowledge (schema) on word learning by manipulating both semantic (familiar vs unfamiliar objects) and phonological (L1- vs L2-like novel words) familiarity. Participants were trained and tested with a 12-hr intervening period that included overnight sleep or daytime awake. Our results showed (1) benefits of sleep to recognition memory that were greater for words with L2-like phonology and (2) that learned associations with familiar but not unfamiliar pictures enhanced recognition memory for novel words. Implications for complementary systems accounts of word learning are discussed.

  20. [Hemispheric asymmetry modulation for language processing in aging: meta-analysis of studies using the dichotic listening test]. (United States)

    Vanhoucke, Elodie; Cousin, Emilie; Baciu, Monica


    Growing evidence suggests that age impacts on interhemispheric representation of language. Dichotic listening test allows assessing language lateralization for spoken language and it generally reveals right-ear/left-hemisphere (LH) predominance for language in young adult subjects. According to reported results, elderly would display increasing LH predominance in some studies or stable LH language lateralization for language in others ones. The aim of this study was to depict the main pattern of results in respect with the effect of normal aging on the hemisphere specialization for language by using dichotic listening test. A meta-analysis based on 11 studies has been performed. The inter-hemisphere asymmetry does not seem to increase according to age. A supplementary qualitative analysis suggests that right-ear advantage seems to increase between 40 and 49 y old and becomes stable or decreases after 55 y old, suggesting right-ear/LH decline.

  1. A grammar of Abui : A Papuan language of Alor

    NARCIS (Netherlands)

    Kratochvil, František


    This work contains the first comprehensive description of Abui, a language of the Trans New Guinea family spoken approximately by 16,000 speakers in the central part of the Alor Island in Eastern Indonesia. The description focuses on the northern dialect of Abui as spoken in the village

  2. Determination of language lateralization using functional MRI during the performance of shiritori tasks in neurosurgery patients

    International Nuclear Information System (INIS)

    Takayama, Hideichi; Kobayashi, Masahito; Sugishita, Morihiro; Onozuka, Satoshi; Kawase, Takeshi


    Assessment of language lateralization is crucial in patients considered for neurological surgery. The authors used functional MRI (fMRI) in conjunction with shiritori, a kind of word-generation task as paradigms, to determine language lateralization in the patients. We used a 1.5 Tesla magnetic resonance imaging devise with an echo-planar imaging sequence. Thirty-two patients undergoing neurological surgery would alternately rest and silently perform shiritori during fMRI acquisition. Language lateralization was determined in 29 out of 32 patients. Twenty-two patients were considered as left-hemisphere dominant and seven were right-hemisphere dominant. Brain activation was seen in the prefrontal area, premotor area, superior temporal gyrus and parietal lobe of the dominant hemisphere, which is consistent with the results in normal adults. Language lateralization was particularly useful in a case of meningioma in the left lateral ventricle and in a case of AVM in the left temporoparietal region. fMRI with shiritori tasks revealed right-hemisphere dominance in both cases, which was also confirmed by intracarotid amobarbital (Wada) testing. Both lesions were treated successfully without causing any further deficit to the patients' language function. These results suggest that fMRI with shiritori tasks can be used to assess language lateralization non-invasively, compared with the current techniques, such as intracarotid amobarbital testing and cortical electrostimulation mapping. Thus, fMRI with shiritori tasks has significant clinical potential as a presurgical evaluation tool. (author)

  3. Determination of language lateralization using functional MRI during the performance of shiritori tasks in neurosurgery patients

    Energy Technology Data Exchange (ETDEWEB)

    Takayama, Hideichi; Kobayashi, Masahito [Mihara Memorial Hospital, Isesaki, Gunma (Japan); Sugishita, Morihiro; Onozuka, Satoshi; Kawase, Takeshi


    Assessment of language lateralization is crucial in patients considered for neurological surgery. The authors used functional MRI (fMRI) in conjunction with shiritori, a kind of word-generation task as paradigms, to determine language lateralization in the patients. We used a 1.5 Tesla magnetic resonance imaging devise with an echo-planar imaging sequence. Thirty-two patients undergoing neurological surgery would alternately rest and silently perform shiritori during fMRI acquisition. Language lateralization was determined in 29 out of 32 patients. Twenty-two patients were considered as left-hemisphere dominant and seven were right-hemisphere dominant. Brain activation was seen in the prefrontal area, premotor area, superior temporal gyrus and parietal lobe of the dominant hemisphere, which is consistent with the results in normal adults. Language lateralization was particularly useful in a case of meningioma in the left lateral ventricle and in a case of AVM in the left temporoparietal region. fMRI with shiritori tasks revealed right-hemisphere dominance in both cases, which was also confirmed by intracarotid amobarbital (Wada) testing. Both lesions were treated successfully without causing any further deficit to the patients' language function. These results suggest that fMRI with shiritori tasks can be used to assess language lateralization non-invasively, compared with the current techniques, such as intracarotid amobarbital testing and cortical electrostimulation mapping. Thus, fMRI with shiritori tasks has significant clinical potential as a presurgical evaluation tool. (author)

  4. Language Mediated Concept Activation in Bilingual Memory Facilitates Cognitive Flexibility

    Directory of Open Access Journals (Sweden)

    Anatoliy V. Kharkhurin


    Full Text Available This is the first attempt of empirical investigation of language mediated concept activation (LMCA in bilingual memory as a cognitive mechanism facilitating divergent thinking. Russian–English bilingual and Russian monolingual college students were tested on a battery of tests including among others Abbreviated Torrance Tests for Adults assessing divergent thinking traits and translingual priming (TLP test assessing the LMCA. The latter was designed as a lexical decision priming test, in which a prime and a target were not related in Russian (language of testing, but were related through their translation equivalents in English (spoken only by bilinguals. Bilinguals outperformed their monolingual counterparts on divergent thinking trait of cognitive flexibility, and bilinguals’ performance on this trait could be explained by their TLP effect. Age of second language acquisition and proficiency in this language were found to relate to the TLP effect, and therefore were proposed to influence the directionality and strength of connections in bilingual memory.

  5. Beyond Languages, beyond Modalities: Transforming the Study of Semiotic Repertoires (United States)

    Kusters, Annelies; Spotti, Massimiliano; Swanwick, Ruth; Tapio, Elina


    This paper presents a critical examination of key concepts in the study of (signed and spoken) language and multimodality. It shows how shifts in conceptual understandings of language use, moving from bilingualism to multilingualism and (trans)languaging, have resulted in the revitalisation of the concept of language repertoires. We discuss key…

  6. Acquiring the Language of Learning: The Performance of Hawaiian Preschool Children on the Preschool Language Assessment Instrument (PLAI). (United States)

    Martini, Mary

    The Preschool Language Assessment Instrument (PLAI) was designed as a diagnostic tool for 3- to 6-year-old children to assess children's abilities to use language to solve thinking problems typically posed by teachers. The PLAI was developed after observing middle-class teachers in preschool classrooms encourage children to use language in…

  7. Analysis of English language learner performance on the biology Massachusetts comprehensive assessment system: The impact of english proficiency, first language characteristics, and late-entry ELL status (United States)

    Mitchell, Mary A.

    This study analyzed English language learner (ELL) performance on the June 2012 Biology MCAS, namely on item attributes of domain, cognitive skill, and linguistic complexity. It examined the impact of English proficiency, Latinate first language, first language orthography, and late-entry ELL status. The results indicated that English proficiency was a strong predictor of performance and that ELLs at higher levels of English proficiency overwhelmingly passed. The results further indicated that English proficiency introduced a construct-irrelevant variance on the Biology MCAS and raised validity issues for using this assessment at lower levels of English proficiency. This study also found that ELLs with a Latinate first language consistently had statistically significant lower performance. Late-entry ELL status did not predict Biology MCAS performance.

  8. Teaching and Learning Sign Language as a “Foreign” Language ...

    African Journals Online (AJOL)

    In recent years, there has been a growing debate in the United States, Europe, and Australia about the nature of the Deaf community as a cultural community,1 and the recognition of signed languages as “real” or “legitimate” languages comparable in all meaningful ways to spoken languages. An important element of this ...

  9. The Impact of Biculturalism on Language and Literacy Development: Teaching Chinese English Language Learners (United States)

    Palmer, Barbara C.; Chen, Chia-I; Chang, Sara; Leclere, Judith T.


    According to the 2000 United States Census, Americans age five and older who speak a language other than English at home grew 47 percent over the preceding decade. This group accounts for slightly less than one in five Americans (17.9%). Among the minority languages spoken in the United States, Asian-language speakers, including Chinese and other…

  10. Spoken Grammar: Where Are We and Where Are We Going? (United States)

    Carter, Ronald; McCarthy, Michael


    This article synthesises progress made in the description of spoken (especially conversational) grammar over the 20 years since the authors published a paper in this journal arguing for a re-thinking of grammatical description and pedagogy based on spoken corpus evidence. We begin with a glance back at the 16th century and the teaching of Latin…

  11. Attention to spoken word planning: Chronometric and neuroimaging evidence

    NARCIS (Netherlands)

    Roelofs, A.P.A.


    This article reviews chronometric and neuroimaging evidence on attention to spoken word planning, using the WEAVER++ model as theoretical framework. First, chronometric studies on the time to initiate vocal responding and gaze shifting suggest that spoken word planning may require some attention,

  12. Spoken commands control robot that handles radioactive materials

    International Nuclear Information System (INIS)

    Phelan, P.F.; Keddy, C.; Beugelsdojk, T.J.


    Several robotic systems have been developed by Los Alamos National Laboratory to handle radioactive material. Because of safety considerations, the robotic system must be under direct human supervision and interactive control continuously. In this paper, we describe the implementation of a voice-recognition system that permits this control, yet allows the robot to perform complex preprogrammed manipulations without the operator's intervention. To provide better interactive control, we connected to the robot's control computer, a speech synthesis unit, which provides audible feedback to the operator. Thus upon completion of a task or if an emergency arises, an appropriate spoken message can be reported by the control computer. The training programming and operation of this commercially available system are discussed, as are the practical problems encountered during operations

  13. Reading Test Performance of English-Language Learners Using an English Dictionary. (United States)

    Albus, Debra; Thurlow, Martha; Liu, Kristin; Bielinski, John


    The authors examined the effects of a simplified English dictionary accommodation on the reading-test performance of Hmong English-language learners (ELLs). Participants included a control group of 69 non-ELL students and an experimental group of 133 Hmong ELLs from 3 urban middle schools in Minnesota. In a randomized counterbalanced design, all…

  14. Impact of Formulas, Language and Instruction on Student Performance on Cost-Volume-Profit Problems (United States)

    Johnson, Benny G.; Sargent, Carol Springer


    This study investigated how three factors impacted performance on cost-volume-profit homework problems: language, formula use, and instruction. Students enrolled in Introduction to Financial Accounting (the first principles of accounting course) and Managerial Accounting (the second principles of accounting course) from eight different US colleges…

  15. Perceived Attachment Security to Father, Academic Self-Concept and School Performance in Language Mastery (United States)

    Bacro, Fabien


    This study examined the relations between 8-12-year-olds' perceived attachment security to father, academic self-concept and school performance in language mastery. One hundred and twenty two French students' perceptions of attachment to mother and to father were explored with the Security Scale and their academic self-concept was assessed with…

  16. Learner Performance in Mandarin Immersion and High School World Language Programs: A Comparison (United States)

    Xu, Xiaoqiu; Padilla, Amado M.; Silva, Duarte M.


    This study compared the Mandarin performance of elementary immersion program students and high school world language program students in the same school district. A cross-sectional design was employed to gather information on Mandarin proficiency of fourth and fifth graders and Level 4 and Level 5 (AP Chinese) high school students who took the…

  17. Phonological and Non-Phonological Language Skills as Predictors of Early Reading Performance (United States)

    Batson-Magnuson, LuAnn


    Accurate prediction of early childhood reading performance could help identify at-risk students, aid in the development of evidence-based intervention strategies, and further our theoretical understanding of reading development. This study assessed the validity of the Developmental Indicator for the Assessment of Learning (DIAL) language-based…

  18. Performed Culture: An Approach to East Asian Language Pedagogy. Pathways to Advanced Skills Series, Volume 11 (United States)

    Christensen, Matthew; Warnick, Paul


    This book is a general introduction to the performed culture approach, which trains students how to express themselves in a way that native speakers of the target culture feel appropriate in given situations. Target readership includes Chinese, Japanese, and Korean language teachers and graduate students. Chapters of this book include: (1)…

  19. Science as a second language: Analysis of Emergent Bilinguals performance and the impact of English language proficiency and first language characteristics on the Colorado measures of academic success for science (United States)

    Bruno, Joanna K.

    In an age when communication is highly important and states across the nation, including Colorado, have adopted Common Core State Standards, the need for academic language is even more important than ever. The language of science has been compared to a second language in that it uses specific discourse patterns, semantic rules, and a very specific vocabulary. There is a need for educators to better understand how language impacts academic achievement, specifically concerning Emergent Bilinguals (EBs). Research has identified the need to study the role language plays in content assessments and the impact they have on EBs performance (Abedi, 2008b; Abedi, Hofestter & Lord, 2004; Abedi & Lord, 2001). Since language is the means through which content knowledge is assessed, it is important to analyze this aspect of learning. A review of literature identified the need to create more reliable and valid content assessments for EBs (Abedi, 2008b) and to further study the impact of English proficiency on EBs performance on standardized assessments (Solorzano, 2008; Wolf, & Leon, 2009). This study contributes to the literature by analyzing EBs performance on a state-level science content assessment, taking into consideration English language proficiency, receptive versus productive elements of language, and students' home language. This study further contributes by discussing the relationship between language proficiency, and the different strands of science (physical, life, and earth) on the state science assessment. Finally, this study demonstrates that home language, English language proficiency, and receptive and productive elements of language are predictive of EBs' achievement on the CMAS for science, overall and by strand. It is the blending of the social (listening and speaking) with the academic (reading and writing) that is also important and possibly more important.

  20. Teaching natural language to computers


    Corneli, Joseph; Corneli, Miriam


    "Natural Language," whether spoken and attended to by humans, or processed and generated by computers, requires networked structures that reflect creative processes in semantic, syntactic, phonetic, linguistic, social, emotional, and cultural modules. Being able to produce novel and useful behavior following repeated practice gets to the root of both artificial intelligence and human language. This paper investigates the modalities involved in language-like applications that computers -- and ...

  1. Integration of multimodal MRI data via PCA to explain language performance

    Directory of Open Access Journals (Sweden)

    N.E. Kucukboyaci


    Conclusions: Quantitative MRI measures from T1 and diffusion-weighted scans are unlikely to represent perfectly orthogonal vectors of disease in individuals with epilepsy. On the contrary, they exhibit highly intercorrelated PCs in their factor structures, which is consistent with an underlying pathological process that affects both the cortical and the subcortical structures simultaneously. In addition to hippocampal volume, the PCs of diffusion weighted measures (FA and MD increase the sensitivity and specificity for determining naming impairment in patients with TLE. These findings underline the importance of combining multimodal imaging measures to better predict language performance in TLE that could extend to other patients with prominent language impairments.

  2. Assessing spoken word recognition in children who are deaf or hard of hearing: a translational approach. (United States)

    Kirk, Karen Iler; Prusick, Lindsay; French, Brian; Gotch, Chad; Eisenberg, Laurie S; Young, Nancy


    Under natural conditions, listeners use both auditory and visual speech cues to extract meaning from speech signals containing many sources of variability. However, traditional clinical tests of spoken word recognition routinely employ isolated words or sentences produced by a single talker in an auditory-only presentation format. The more central cognitive processes used during multimodal integration, perceptual normalization, and lexical discrimination that may contribute to individual variation in spoken word recognition performance are not assessed in conventional tests of this kind. In this article, we review our past and current research activities aimed at developing a series of new assessment tools designed to evaluate spoken word recognition in children who are deaf or hard of hearing. These measures are theoretically motivated by a current model of spoken word recognition and also incorporate "real-world" stimulus variability in the form of multiple talkers and presentation formats. The goal of this research is to enhance our ability to estimate real-world listening skills and to predict benefit from sensory aid use in children with varying degrees of hearing loss. American Academy of Audiology.

  3. I Feel You: The Design and Evaluation of a Domotic Affect-Sensitive Spoken Conversational Agent

    Directory of Open Access Journals (Sweden)

    Juan Manuel Montero


    Full Text Available We describe the work on infusion of emotion into a limited-task autonomous spoken conversational agent situated in the domestic environment, using a need-inspired task-independent emotion model (NEMO. In order to demonstrate the generation of affect through the use of the model, we describe the work of integrating it with a natural-language mixed-initiative HiFi-control spoken conversational agent (SCA. NEMO and the host system communicate externally, removing the need for the Dialog Manager to be modified, as is done in most existing dialog systems, in order to be adaptive. The first part of the paper concerns the integration between NEMO and the host agent. The second part summarizes the work on automatic affect prediction, namely, frustration and contentment, from dialog features, a non-conventional source, in the attempt of moving towards a more user-centric approach. The final part reports the evaluation results obtained from a user study, in which both versions of the agent (non-adaptive and emotionally-adaptive were compared. The results provide substantial evidences with respect to the benefits of adding emotion in a spoken conversational agent, especially in mitigating users’ frustrations and, ultimately, improving their satisfaction.

  4. The socially weighted encoding of spoken words: a dual-route approach to speech perception. (United States)

    Sumner, Meghan; Kim, Seung Kyung; King, Ed; McGowan, Kevin B


    Spoken words are highly variable. A single word may never be uttered the same way twice. As listeners, we regularly encounter speakers of different ages, genders, and accents, increasing the amount of variation we face. How listeners understand spoken words as quickly and adeptly as they do despite this variation remains an issue central to linguistic theory. We propose that learned acoustic patterns are mapped simultaneously to linguistic representations and to social representations. In doing so, we illuminate a paradox that results in the literature from, we argue, the focus on representations and the peripheral treatment of word-level phonetic variation. We consider phonetic variation more fully and highlight a growing body of work that is problematic for current theory: words with different pronunciation variants are recognized equally well in immediate processing tasks, while an atypical, infrequent, but socially idealized form is remembered better in the long-term. We suggest that the perception of spoken words is socially weighted, resulting in sparse, but high-resolution clusters of socially idealized episodes that are robust in immediate processing and are more strongly encoded, predicting memory inequality. Our proposal includes a dual-route approach to speech perception in which listeners map acoustic patterns in speech to linguistic and social representations in tandem. This approach makes novel predictions about the extraction of information from the speech signal, and provides a framework with which we can ask new questions. We propose that language comprehension, broadly, results from the integration of both linguistic and social information.

  5. Language performance and auditory evoked fields in 2- to 5-year-old children. (United States)

    Yoshimura, Yuko; Kikuchi, Mitsuru; Shitamichi, Kiyomi; Ueno, Sanae; Remijn, Gerard B; Haruta, Yasuhiro; Oi, Manabu; Munesue, Toshio; Tsubokawa, Tsunehisa; Higashida, Haruhiro; Minabe, Yoshio


    Language development progresses at a dramatic rate in preschool children. As rapid temporal processing of speech signals is important in daily colloquial environments, we performed magnetoencephalography (MEG) to investigate the linkage between speech-evoked responses during rapid-rate stimulus presentation (interstimulus interval language performance in 2- to 5-year-old children (n = 59). Our results indicated that syllables with this short stimulus interval evoked detectable P50m, but not N100m, in most participants, indicating a marked influence of longer neuronal refractory period for stimulation. The results of equivalent dipole estimation showed that the intensity of the P50m component in the left hemisphere was positively correlated with language performance (conceptual inference ability). The observed positive correlations were suggested to reflect the maturation of synaptic organisation or axonal maturation and myelination underlying the acquisition of linguistic abilities. The present study is among the first to use MEG to study brain maturation pertaining to language abilities in preschool children. © 2012 The Authors. European Journal of Neuroscience © 2012 Federation of European Neuroscience Societies and Blackwell Publishing Ltd.

  6. Does the speaker's voice quality influence children's performance on a language comprehension test? (United States)

    Lyberg-Åhlander, Viveka; Haake, Magnus; Brännström, Jonas; Schötz, Susanne; Sahlén, Birgitta


    A small number of studies have explored children's perception of speakers' voice quality and its possible influence on language comprehension. The aim of this explorative study was to investigate the relationship between the examiner's voice quality, the child's performance on a digital version of a language comprehension test, the Test for Reception of Grammar (TROG-2), and two measures of cognitive functioning. The participants were (n = 86) mainstreamed 8-year old children with typical language development. Two groups of children (n = 41/45) were presented with the TROG-2 through recordings of one female speaker: one group was presented with a typical voice and the other with a simulated dysphonic voice. Significant associations were found between executive functioning and language comprehension. The results also showed that children listening to the dysphonic voice achieved significantly lower scores for more difficult sentences ("the man but not the horse jumps") and used more self-corrections on simpler sentences ("the girl is sitting"). Findings suggest that a dysphonic speaker's voice may force the child to allocate capacity to the processing of the voice signal at the expense of comprehension. The findings have implications for clinical and research settings where standardized language tests are used.


    Directory of Open Access Journals (Sweden)

    Elva Yohana


    Full Text Available The primary condition for successful in second or foreign language learning is providing an adequate environment. It is as a medium of increasing the students’ language exposure in order to be able to success in acquiring second or foreign language profciency. This study was designed to propose the adequate English language input that can decrease the students’ anxiety in reading comprehension performance. Of the four skills, somehow reading can be regarded as especially important because reading is assumed to be the central means for learning new information. Some students, however, still encounter many problems in reading. It is because of their anxiety when they are reading. Providing and creating an interesting-contextual reading material and gratifed teachers can make out this problem which occurs mostly in Indonesian’s classrooms. It revealed that the younger learners of English there do not received adequate amount of the target language input in their learning of English. Hence, it suggested the adoption of extensive reading programs as the most effective means in the creation of an input-rich environment in EFL learning contexts. Besides they also give suggestion to book writers and publisher to provide myriad books that appropriate and readable for their students.

  8. Speaking two languages with different number naming systems: What implications for magnitude judgments in bilinguals at different stages of language acquisition? (United States)

    Van Rinsveld, Amandine; Schiltz, Christine; Landerl, Karin; Brunner, Martin; Ugen, Sonja


    Differences between languages in terms of number naming systems may lead to performance differences in number processing. The current study focused on differences concerning the order of decades and units in two-digit number words (i.e., unit-decade order in German but decade-unit order in French) and how they affect number magnitude judgments. Participants performed basic numerical tasks, namely two-digit number magnitude judgments, and we used the compatibility effect (Nuerk et al. in Cognition 82(1):B25-B33, 2001) as a hallmark of language influence on numbers. In the first part we aimed to understand the influence of language on compatibility effects in adults coming from German or French monolingual and German-French bilingual groups (Experiment 1). The second part examined how this language influence develops at different stages of language acquisition in individuals with increasing bilingual proficiency (Experiment 2). Language systematically influenced magnitude judgments such that: (a) The spoken language(s) modulated magnitude judgments presented as Arabic digits, and (b) bilinguals' progressive language mastery impacted magnitude judgments presented as number words. Taken together, the current results suggest that the order of decades and units in verbal numbers may qualitatively influence magnitude judgments in bilinguals and monolinguals, providing new insights into how number processing can be influenced by language(s).

  9. On the Performance of the Python Programming Language for Serial and Parallel Scientific Computations

    Directory of Open Access Journals (Sweden)

    Xing Cai


    Full Text Available This article addresses the performance of scientific applications that use the Python programming language. First, we investigate several techniques for improving the computational efficiency of serial Python codes. Then, we discuss the basic programming techniques in Python for parallelizing serial scientific applications. It is shown that an efficient implementation of the array-related operations is essential for achieving good parallel performance, as for the serial case. Once the array-related operations are efficiently implemented, probably using a mixed-language implementation, good serial and parallel performance become achievable. This is confirmed by a set of numerical experiments. Python is also shown to be well suited for writing high-level parallel programs.

  10. The Peculiarities of the Adverbs Functioning of the Dialect Spoken in the v. Shevchenkove, Kiliya district, Odessa Region

    Directory of Open Access Journals (Sweden)

    Maryna Delyusto


    Full Text Available The article gives new evidence about the adverb as a part of the grammatical system of the Ukrainian steppe dialect spread in the area between the Danube and the Dniester rivers. The author proves that the grammatical system of the dialect spoken in the v. Shevchenkove, Kiliya district, Odessa region is determined by the historical development of the Ukrainian language rather than the influence of neighboring dialects.


    Directory of Open Access Journals (Sweden)

    Marites Piguing HILAO


    Full Text Available Mobile phone technology that has a huge impact on students’ lives in the digital age may offer a new type of learning. The use of effective tool to support learning can be affected by the factor of gender. The current research compared how male and female students perceived mobile phones as a language learning tool, used mobile phones to learn English and developed their learning performance. A five-point rating scale questionnaire was used to collect data from 122 students, comprising 65 females and 57 males. They were enrolled in a fundamental English course where mobile phone usage was integrated in certain language learning tasks with an aim to facilitate learning. The findings demonstrated that male and female students did not differ in their usage, attitudes toward mobile phone uses for language learning as well as their learning performance at a significance level. In addition, the constraints of using mobile phone for learning that students identified in an open-ended question included the small screen and keyboard the most, followed by intrusiveness of SMS background knowledge, and limited memory of mobile phone. The implication for classroom practice was proposed in how mobile phone can be fully incorporated into the instructional process in order to enhance learner engagement. The results of this study are important for teachers when implementing the mobile phone technology in language teaching. They can be used as a guideline of how mobile phone can be fully incorporated into the instructional process in order to enhance learner engagement.

  12. language choice, code-switching and code- mixing in biase

    African Journals Online (AJOL)


    Finance and Economic Planning, Cross River and Akwa ... See Table 1. Table 1: Indigenous Languages Spoken in Biase ... used in education, in business, in religion, in the media ... far back as the seventeenth (17th) century (King. 1844).

  13. The language of football

    DEFF Research Database (Denmark)

    Rossing, Niels Nygaard; Skrubbeltrang, Lotte Stausgaard


    levels (Schein, 2004) in which each player and his actions can be considered an artefact - a concrete symbol in motion embedded in espoused values and basic assumptions. Therefore, the actions of each dialect are strongly connected to the underlying understanding of football. By document and video......The language of football: A cultural analysis of selected World Cup nations. This essay describes how actions on the football field relate to the nations’ different cultural understanding of football and how these actions become spoken dialects within a language of football. Saussure reasoned...... language to have two components: a language system and language users (Danesi, 2003). Consequently, football can be characterized as a language containing a system with specific rules of the game and users with actual choices and actions within the game. All football players can be considered language...

  14. Lecturing in one’s first language or in English as a lingua franca

    DEFF Research Database (Denmark)

    Preisler, Bent


    The demand for internationalization puts pressure on Danish universities to use English as the language of instruction instead of or in addition to the local language(s). The purpose of this study – though proceeding from the belief that true internationalization seeks to exploit all linguistic...... and multilingual classroom. This case study concerns Danish university teachers' spoken discourse and interaction with students in a Danish-language versus English-language classroom. The data are video recordings of classroom interaction at the University of Roskilde, Denmark. The focus is on the relationship...... between linguistic-pragmatic performance and academic authenticity for university teachers teaching courses in both English and Danish, based on recent sociolinguistic concepts such as “persona,” “stylization,” and “authenticity.” The analysis suggests that it is crucial for teachers' ability...

  15. The Road to Language Learning Is Not Entirely Iconic: Iconicity, Neighborhood Density, and Frequency Facilitate Acquisition of Sign Language. (United States)

    Caselli, Naomi K; Pyers, Jennie E


    Iconic mappings between words and their meanings are far more prevalent than once estimated and seem to support children's acquisition of new words, spoken or signed. We asked whether iconicity's prevalence in sign language overshadows two other factors known to support the acquisition of spoken vocabulary: neighborhood density (the number of lexical items phonologically similar to the target) and lexical frequency. Using mixed-effects logistic regressions, we reanalyzed 58 parental reports of native-signing deaf children's productive acquisition of 332 signs in American Sign Language (ASL; Anderson & Reilly, 2002) and found that iconicity, neighborhood density, and lexical frequency independently facilitated vocabulary acquisition. Despite differences in iconicity and phonological structure between signed and spoken language, signing children, like children learning a spoken language, track statistical information about lexical items and their phonological properties and leverage this information to expand their vocabulary.

  16. Conceptual representation of verbs in bilinguals: semantic field effects and a second-language performance paradox. (United States)

    Segalowitz, Norman; de Almeida, Roberto G


    It is well known that bilinguals perform better in their first language (L1) than in their second lanaguage (L2) in a wide range of linguistic tasks. In recent studies, however, the authors have found that bilingual participants can demonstrate faster response times to L1 stimuli than to L2 stimuli in one classification task and the reverse in a different classification task. In the current study, they investigated the reasons for this "L2-better-than-L1" effect. English-French bilinguals performed one word relatedness and two categorization tasks with verbs of motion (e.g., run) and psychological verbs (e.g., admire) in both languages. In the word relatedness task, participants judged how closely related pairs of verbs from both categories were. In a speeded semantic categorization task, participants classified the verbs according to their semantic category (psychological or motion). In an arbitrary classification task, participants had to learn how verbs had been assigned to two arbitrary categories. Participants performed better in L1 in the semantic classification task but paradoxically better in L2 in the arbitrary classification task. To account for these effects, the authors used the ratings from the word relatedness task to plot three-dimensional "semantic fields" for the verbs. Cross-language field differences were found to be significantly related to the paradoxical performance and to fluency levels. The results have implications for understanding of how bilinguals represent verbs in the mental lexicon. Copyright 2002 Elsevier Science (USA).

  17. "We communicated that way for a reason": language practices and language ideologies among hearing adults whose parents are deaf. (United States)

    Pizer, Ginger; Walters, Keith; Meier, Richard P


    Families with deaf parents and hearing children are often bilingual and bimodal, with both a spoken language and a signed one in regular use among family members. When interviewed, 13 American hearing adults with deaf parents reported widely varying language practices, sign language abilities, and social affiliations with Deaf and Hearing communities. Despite this variation, the interviewees' moral judgments of their own and others' communicative behavior suggest that these adults share a language ideology concerning the obligation of all family members to expend effort to overcome potential communication barriers. To our knowledge, such a language ideology is not similarly pervasive among spoken-language bilingual families, raising the question of whether there is something unique about family bimodal bilingualism that imposes different rights and responsibilities on family members than spoken-language family bilingualism does. This ideology unites an otherwise diverse group of interviewees, where each one preemptively denied being a "typical CODA [children of deaf adult]."

  18. Comparison of Word Intelligibility in Spoken and Sung Phrases

    Directory of Open Access Journals (Sweden)

    Lauren B. Collister


    Full Text Available Twenty listeners were exposed to spoken and sung passages in English produced by three trained vocalists. Passages included representative words extracted from a large database of vocal lyrics, including both popular and classical repertoires. Target words were set within spoken or sung carrier phrases. Sung carrier phrases were selected from classical vocal melodies. Roughly a quarter of all words sung by an unaccompanied soloist were misheard. Sung passages showed a seven-fold decrease in intelligibility compared with their spoken counterparts. The perceptual mistakes occurring with vowels replicate previous studies showing the centralization of vowels. Significant confusions are also evident for consonants, especially voiced stops and nasals.

  19. Germanic heritage languages in North America: Acquisition, attrition and change


    Johannessen, Janne Bondi; Salmons, Joseph C.; Westergaard, Marit; Anderssen, Merete; Arnbjörnsdóttir, Birna; Allen, Brent; Pierce, Marc; Boas, Hans C.; Roesch, Karen; Brown, Joshua R.; Putnam, Michael; Åfarli, Tor A.; Newman, Zelda Kahan; Annear, Lucas; Speth, Kristin


    This book presents new empirical findings about Germanic heritage varieties spoken in North America: Dutch, German, Pennsylvania Dutch, Icelandic, Norwegian, Swedish, West Frisian and Yiddish, and varieties of English spoken both by heritage speakers and in communities after language shift. The volume focuses on three critical issues underlying the notion of ‘heritage language’: acquisition, attrition and change. The book offers theoretically-informed discussions of heritage language processe...

  20. The road to language learning is iconic: evidence from British Sign Language. (United States)

    Thompson, Robin L; Vinson, David P; Woll, Bencie; Vigliocco, Gabriella


    An arbitrary link between linguistic form and meaning is generally considered a universal feature of language. However, iconic (i.e., nonarbitrary) mappings between properties of meaning and features of linguistic form are also widely present across languages, especially signed languages. Although recent research has shown a role for sign iconicity in language processing, research on the role of iconicity in sign-language development has been mixed. In this article, we present clear evidence that iconicity plays a role in sign-language acquisition for both the comprehension and production of signs. Signed languages were taken as a starting point because they tend to encode a higher degree of iconic form-meaning mappings in their lexicons than spoken languages do, but our findings are more broadly applicable: Specifically, we hypothesize that iconicity is fundamental to all languages (signed and spoken) and that it serves to bridge the gap between linguistic form and human experience.

  1. An fMRI study of concreteness effects during spoken word recognition in aging. Preservation or attenuation?

    Directory of Open Access Journals (Sweden)

    Tracy eRoxbury


    Full Text Available It is unclear whether healthy aging influences concreteness effects (ie. the processing advantage seen for concrete over abstract words and its associated neural mechanisms. We conducted an fMRI study on young and older healthy adults performing auditory lexical decisions on concrete versus abstract words. We found that spoken comprehension of concrete and abstract words appears relatively preserved for healthy older individuals, including the concreteness effect. This preserved performance was supported by altered activity in left hemisphere regions including the inferior and middle frontal gyri, angular gyrus, and fusiform gyrus. This pattern is consistent with age-related compensatory mechanisms supporting spoken word processing.

  2. Moving conceptualizations of language and literacy in SLA

    DEFF Research Database (Denmark)

    Laursen, Helle Pia

    in various technological environments, we see an increase in scholarship that highlights the mixing and chaining of spoken, written and visual modalities and how written and visual often precede or overrule spoken language. There seems to be a mismatch between current day language practices......, in language education and in language practices. As a consequence of this and in the light of the increasing mobility and linguistic diversity in Europe, in this colloquium, we address the need for a (re)conceptualization of the relation between language and literacy. Drawing on data from different settings...

  3. Use of Spoken and Written Japanese Did Not Protect Japanese-American Men From Cognitive Decline in Late Life (United States)

    Gruhl, Jonathan C.; Erosheva, Elena A.; Gibbons, Laura E.; McCurry, Susan M.; Rhoads, Kristoffer; Nguyen, Viet; Arani, Keerthi; Masaki, Kamal; White, Lon


    Objectives. Spoken bilingualism may be associated with cognitive reserve. Mastering a complicated written language may be associated with additional reserve. We sought to determine if midlife use of spoken and written Japanese was associated with lower rates of late life cognitive decline. Methods. Participants were second-generation Japanese-American men from the Hawaiian island of Oahu, born 1900–1919, free of dementia in 1991, and categorized based on midlife self-reported use of spoken and written Japanese (total n included in primary analysis = 2,520). Cognitive functioning was measured with the Cognitive Abilities Screening Instrument scored using item response theory. We used mixed effects models, controlling for age, income, education, smoking status, apolipoprotein E e4 alleles, and number of study visits. Results. Rates of cognitive decline were not related to use of spoken or written Japanese. This finding was consistent across numerous sensitivity analyses. Discussion. We did not find evidence to support the hypothesis that multilingualism is associated with cognitive reserve. PMID:20639282

  4. An investigation of Chinese university EFL learner’s foreign language reading anxiety, reading strategy use and reading comprehension performance


    Zhongshe Lu; Meihua Liu


    The present study explored the interrelations between foreign language (FL) reading anxiety, FL reading strategy use and their interactive effect on FL reading comprehension performance at the tertiary level in China. Analyses of the survey data collected from 1702 university students yielded the following results: (a) Both Foreign Language Reading Anxiety Scale (FLRAS) and Foreign Language Reading Strategy Use Scale (FLRSUS) had important subcomponents, (b) more than half of the stu...

  5. Preschoolers' Performance on the Brazilian Adaptation of the Preschool Language Assessment Instrument - Second Edition. (United States)

    Lindau, Tâmara Andrade; Rossi, Natalia Freitas; Giacheti, Celia Maria


    The objective was to test whether the Brazilian version of the Preschool Language Assessment Instrument - Second Edition (PLAI-2) has the potential to assess and identify differences in typical language development of Portuguese-speaking preschoolers. The study included 354 children of both genders with typical language development who were between the ages of 3 years and 5 years 11 months. The version of the PLAI-2 previously translated into Brazilian Portuguese was used to assess the communication skills of these preschool-age children. Statistically significant differences were found between the age groups, and the raw score tended to increase as a function of age. With nonstandardized assessments, the performances of the younger groups revealed behavioral profiles (e.g., nonresponsive, impulsive behavior) that directly influenced the evaluation. The findings of this study show that the PLAI-2 is effective in identifying differences in language development among Brazilian children of preschool age. Future research should include studies validating and standardizing these findings. © 2016 S. Karger AG, Basel.

  6. Moyamoya disease: impact on the performance of oral and written language. (United States)

    Lamônica, Dionísia Aparecida Cusin; Ribeiro, Camila da Costa; Ferraz, Plínio Marcos Duarte Pinto; Tabaquim, Maria de Lourdes Merighi

    Moyamoya disease is an unusual form of occlusive, cerebrovascular disorder that affects the arteries of the central nervous system, causing acquired language alterations and learning difficulties. The study aim was to describe the oral/written language and cognitive skills in a seven-year-and-seven-month-old girl diagnosed with Moyamoya disease. The assessment consisted of interviews with her parents and application of the following instruments: Observation of Communicative Behavior, Peabody Picture Vocabulary Test, Academic Performance Test, Profile of Phonological Awareness, Raven's Progressive Matrices Test, Special Scale, Wechsler Intelligence Scale for Children, Bender Visual Motor Gestalt Test, and Wisconsin Card Sorting Test. Two episodes of stroke in the left and right temporal-parietal and left frontal areas occurred until the age of six years and five months. Revascularization surgery and medication treatment were conducted. The audiologic and ophthalmologic assessments indicated normality. At the time of the study, the girl was attending the second grade of elementary school. She presented changes in oral and written language (syllabic-alphabetic), non-naming of all graphemes, low arithmetic and writing means, reading skill below first grade level and psycholinguistic delay, and pre-school level phonological processing skills. The psychological evaluation indicated satisfactory intellectual level; however, it also showed cognitive performance impairment in verbal and execution tasks and limitations on graphic-perceptual-motor skills and sequential logic organization. The stroke episodes influenced the performance of learning processes, affecting the analysis, integration, and interpretation of relevant visual and auditory information.

  7. Sign Language Interpreting in Theatre: Using the Human Body to Create Pictures of the Human Soul

    Directory of Open Access Journals (Sweden)

    Michael Richardson


    Full Text Available This paper explores theatrical interpreting for Deaf spectators, a specialism that both blurs the separation between translation and interpreting, and replaces these potentials with a paradigm in which the translator's body is central to the production of the target text. Meaningful written translations of dramatic texts into sign language are not currently possible. For Deaf people to access Shakespeare or Moliere in their own language usually means attending a sign language interpreted performance, a typically disappointing experience that fails to provide accessibility or to fulfil the potential of a dynamically equivalent theatrical translation. I argue that when such interpreting events fail, significant contributory factors are the challenges involved in producing such a target text and the insufficient embodiment of that text. The second of these factors suggests that the existing conference and community models of interpreting are insufficient in describing theatrical interpreting. I propose that a model drawn from Theatre Studies, namely psychophysical acting, might be more effective for conceptualising theatrical interpreting. I also draw on theories from neurological research into the Mirror Neuron System to suggest that a highly visual and physical approach to performance (be that by actors or interpreters is more effective in building a strong actor-spectator interaction than a performance in which meaning is conveyed by spoken words. Arguably this difference in language impact between signed and spoken is irrelevant to hearing audiences attending spoken language plays, but I suggest that for all theatre translators the implications are significant: it is not enough to create a literary translation as the target text; it is also essential to produce a text that suggests physicality. The aim should be the creation of a text which demands full expression through the body, the best picture of the human soul and the fundamental medium

  8. Speech, gesture and the origins of language

    NARCIS (Netherlands)

    Levelt, W.J.M.


    During the second half of the 19th century, the psychology of language was invented as a discipline for the sole purpose of explaining the evolution of spoken language. These efforts culminated in Wilhelm Wundt’s monumental Die Sprache of 1900, which outlined the psychological mechanisms involved in

  9. Iconic Factors and Language Word Order (United States)

    Moeser, Shannon Dawn


    College students were presented with an artificial language in which spoken nonsense words were correlated with visual references. Inferences regarding vocabulary acquisition were drawn, and it was suggested that the processing of the language was mediated through a semantic memory system. (CK)



    Boluwaji Oshodi


    There are conflicting claims among scholars on whether the structural outputs of the types of English spoken in countries where English is used as a second language gives such speech forms the status of varieties of English. This study examined those morphological features considered to be marked features of the variety spoken in Nigeria according to Kirkpatrick (2011) and the variety spoken in Malaysia by considering the claims of the Missing Surface Inflection Hypothesis (MSIH) a Second Lan...

  11. Navigating Hybridized Language Learning Spaces through Translanguaging Pedagogy: Dual Language Preschool Teachers' Languaging Practices in Support of Emergent Bilingual Children's Performance of Academic Discourse (United States)

    Gort, Mileidis; Sembiante, Sabrina Francesca


    In recent years, there has been a growing interest among policymakers, practitioners, and researchers in early bilingual development and the unique role of the educational setting's language policy in this development. In this article, we describe how one dual language preschool teacher, in partnership with two co-teachers, navigated the tensions…

  12. Autosegmental Representation of Epenthesis in the Spoken French ...

    African Journals Online (AJOL)

    Nneka Umera-Okeke

    ... spoken French of IUFLs. Key words: IUFLs, Epenthensis, Ijebu dialect, Autosegmental phonology .... Ambiguities may result: salmi "strait" vs. salami. (An exception is that in .... tiers of segments. In the picture given us by classical generative.

  13. Aphasia, an acquired language disorder

    African Journals Online (AJOL)


    Oct 11, 2009 ... In this article we will review the types of aphasia, an approach to its diagnosis, aphasia subtypes, rehabilitation and prognosis. ... language processing in both the written and spoken forms.6 ... The angular gyrus (Brodman area 39) is located at the .... of his or her quality of life, emotional state, sense of well-.

  14. Cohesion as interaction in ELF spoken discourse

    Directory of Open Access Journals (Sweden)

    T. Christiansen


    Full Text Available Hitherto, most research into cohesion has concentrated on texts (usually written only in standard Native Speaker English – e.g. Halliday and Hasan (1976. By contrast, following on the work in anaphora of such scholars as Reinhart (1983 and Cornish (1999, Christiansen (2011 describes cohesion as an interac­tive process focusing on the link between text cohesion and discourse coherence. Such a consideration of cohesion from the perspective of discourse (i.e. the process of which text is the product -- Widdowson 1984, p. 100 is especially relevant within a lingua franca context as the issue of different variations of ELF and inter-cultural concerns (Guido 2008 add extra dimensions to the complex multi-code interaction. In this case study, six extracts of transcripts (approximately 1000 words each, taken from the VOICE corpus (2011 of conference question and answer sessions (spoken interaction set in multicultural university con­texts are analysed in depth by means of a qualitative method.

  15. Talker and background noise specificity in spoken word recognition memory


    Cooper, Angela; Bradlow, Ann R.


    Prior research has demonstrated that listeners are sensitive to changes in the indexical (talker-specific) characteristics of speech input, suggesting that these signal-intrinsic features are integrally encoded in memory for spoken words. Given that listeners frequently must contend with concurrent environmental noise, to what extent do they also encode signal-extrinsic details? Native English listeners’ explicit memory for spoken English monosyllabic and disyllabic words was assessed as a fu...

  16. Spectrotemporal processing drives fast access to memory traces for spoken words. (United States)

    Tavano, A; Grimm, S; Costa-Faidella, J; Slabu, L; Schröger, E; Escera, C


    The Mismatch Negativity (MMN) component of the event-related potentials is generated when a detectable spectrotemporal feature of the incoming sound does not match the sensory model set up by preceding repeated stimuli. MMN is enhanced at frontocentral scalp sites for deviant words when compared to acoustically similar deviant pseudowords, suggesting that automatic access to long-term memory traces for spoken words contributes to MMN generation. Does spectrotemporal feature matching also drive automatic lexical access? To test this, we recorded human auditory event-related potentials (ERPs) to disyllabic spoken words and pseudowords within a passive oddball paradigm. We first aimed at replicating the word-related MMN enhancement effect for Spanish, thereby adding to the available cross-linguistic evidence (e.g., Finnish, English). We then probed its resilience to spectrotemporal perturbation by inserting short (20 ms) and long (120 ms) silent gaps between first and second syllables of deviant and standard stimuli. A significantly enhanced, frontocentrally distributed MMN to deviant words was found for stimuli with no gap. The long gap yielded no deviant word MMN, showing that prior expectations of word form limits in a given language influence deviance detection processes. Crucially, the insertion of a short gap suppressed deviant word MMN enhancement at frontocentral sites. We propose that spectrotemporal point-wise matching constitutes a core mechanism for fast serial computations in audition and language, bridging sensory and long-term memory systems. Copyright © 2012 Elsevier Inc. All rights reserved.

  17. Probabilistic Approaches to Examining Linguistic Features of Test Items and Their Effect on the Performance of English Language Learners (United States)

    Solano-Flores, Guillermo


    This article addresses validity and fairness in the testing of English language learners (ELLs)--students in the United States who are developing English as a second language. It discusses limitations of current approaches to examining the linguistic features of items and their effect on the performance of ELL students. The article submits that…

  18. Receptive and expressive language performance in children with and without Cleft Lip and Palate. (United States)

    Lamônica, Dionísia Aparecida Cusin; Silva-Mori, Mariana Jales Felix da; Ribeiro, Camila da Costa; Maximino, Luciana Paula


    To compare the performance in the abilities of receptive and expressive language of children with cleft lip and palate with that of children without cleft lip and palate with typical 12 to 36-month chronological development. The sample consisted of 60 children aged 12 and 36 months: 30 with cleft lip and palate diagnosis and 30 without cleft lip and palate diagnosis with typical development. The groups were paired according to gender, age (in months), and socioeconomic level. The procedures consisted of analysis of medical records, anamnesis with family members, and valuation of the Early Language Milestone Scale (ELMS). The chart analysis showed 63.34% of the children with unilateral cleft lip and palate, 16.66% with bilateral incisive transforamen cleft, and 20% with post-foramen cleft. Children with cleft lip and palate underwent surgeries (lip repair and/or palatoplasty) at the recommended ages and participated in early intervention programs; 40% presented recurrent otitis history, and 50% attended schools. Statistical analysis included the use of the Mann Whitney test with significance level of p cleft lip and palate showed statistically significant low performance in receptive and expressive language compared with children without cleft lip and palate.

  19. Effects of Music and Tonal Language Experience on Relative Pitch Performance. (United States)

    Ngo, Mary Kim; Vu, Kim-Phuong L; Strybel, Thomas Z


    We examined the interaction between music and tone language experience as related to relative pitch processing by having participants judge the direction and magnitude of pitch changes in a relative pitch task. Participants' performance on this relative pitch task was assessed using the Cochran-Weiss-Shanteau (CWS) index of expertise, based on a ratio of discrimination over consistency in participants' relative pitch judgments. Testing took place in 2 separate sessions on different days to assess the effects of practice on participants' performance. Participants also completed the Montreal Battery of Evaluation of Amusia (MBEA), an existing measure comprising subtests aimed at evaluating relative pitch processing abilities. Musicians outperformed nonmusicians on both the relative pitch task, as measured by the CWS index, and the MBEA, but tonal language speakers outperformed non-tonal language speakers only on the MBEA. A closer look at the discrimination and consistency component scores of the CWS index revealed that musicians were better at discriminating different pitches and more consistent in their assessments of the direction and magnitude of relative pitch change.

  20. Making a Difference: Language Teaching for Intercultural and International Dialogue (United States)

    Byram, Michael; Wagner, Manuela


    Language teaching has long been associated with teaching in a country or countries where a target language is spoken, but this approach is inadequate. In the contemporary world, language teaching has a responsibility to prepare learners for interaction with people of other cultural backgrounds, teaching them skills and attitudes as well as…

  1. Australian Aboriginal Deaf People and Aboriginal Sign Language (United States)

    Power, Des


    Many Australian Aboriginal people use a sign language ("hand talk") that mirrors their local spoken language and is used both in culturally appropriate settings when speech is taboo or counterindicated and for community communication. The characteristics of these languages are described, and early European settlers' reports of deaf…

  2. Regional Sign Language Varieties in Contact: Investigating Patterns of Accommodation (United States)

    Stamp, Rose; Schembri, Adam; Evans, Bronwen G.; Cormier, Kearsy


    Short-term linguistic accommodation has been observed in a number of spoken language studies. The first of its kind in sign language research, this study aims to investigate the effects of regional varieties in contact and lexical accommodation in British Sign Language (BSL). Twenty-five participants were recruited from Belfast, Glasgow,…

  3. How Facebook Can Revitalise Local Languages: Lessons from Bali (United States)

    Stern, Alissa Joy


    For a language to survive, it must be spoken and passed down to the next generation. But how can we engage teenagers--so crucial for language transmission--to use and value their local tongue when they are bombarded by pressures from outside and from within their society to only speak national and international languages? This paper analyses the…

  4. El Espanol como Idioma Universal (Spanish as a Universal Language) (United States)

    Mijares, Jose


    A proposal to transform Spanish into a universal language because it possesses the prerequisites: it is a living language, spoken in several countries; it is a natural language; and it uses the ordinary alphabet. Details on simplification and standardization are given. (Text is in Spanish.) (AMH)

  5. Language Planning for Venezuela: The Role of English. (United States)

    Kelsey, Irving; Serrano, Jose

    A rationale for teaching foreign languages in Venezuelan schools is discussed. An included sociolinguistic profile of Venezuela indicates that Spanish is the sole language of internal communication needs. Other languages spoken in Venezuela serve primarily a group function among the immigrant and indigenous communities. However, the teaching of…

  6. Dilemmatic Aspects of Language Policies in a Trilingual Preschool Group (United States)

    Puskás, Tünde; Björk-Willén, Polly


    This article explores dilemmatic aspects of language policies in a preschool group in which three languages (Swedish, Romani and Arabic) are spoken on an everyday basis. The article highlights the interplay between policy decisions on the societal level, the teachers' interpretations of these policies, as well as language practices on the micro…

  7. Dynamic Effects of Performance-Avoidance Goal Orientation on Student Achievement in Language and Mathematics. (United States)

    Stamovlasis, Dimitrios; Gonida, Sofia-Eleftheria N


    The present study used achievement goal theory (AGT) as a theoretical framework and examined the role of mastery and performance goals, both performance-approach and performance-avoidance, on school achieve-ment within the nonlinear dynamical systems (NDS) perspective. A series of cusp catastrophe models were applied on students' achievement in a number of school subjects, such as mathematics and language for elementary school and algebra, geometry, ancient and modern Greek language for high school, using achievement goal orientations as control variables. The participants (N=224) were students attending fifth and eighth grade (aged 11 and 14, respectively) in public schools located in northern Greece. Cusp analysis based on the probability density function was carried out by two procedures, the maximum likelihood and the least squares. The results showed that performance-approach goals had no linear effect on achievement, while the cusp models implementing mastery goals as the asymmetry factor and performance-avoidance as the bifurcation, proved superior to their linear alternatives. The results of the study based on NDS support the multiple goal perspective within AGT. Theoretical issues, educational implications and future directions are discussed.

  8. Delay or deficit? Spelling processes in children with specific language impairment. (United States)

    Larkin, Rebecca F; Williams, Gareth J; Blaggan, Samarita


    Few studies have explored the phonological, morphological and orthographic spellings skills of children with specific language impairment (SLI) simultaneously. Fifteen children with SLI (mean age=113.07 months, SD=8.61) completed language and spelling tasks alongside chronological-age controls and spelling-age controls. While the children with SLI showed a deficit in phonological spelling, they performed comparably to spelling-age controls on morphological spelling skills, and there were no differences between the three groups in producing orthographically legal spellings. The results also highlighted the potential importance of adequate non-word repetition skills in relation to effective spelling skills, and demonstrated that not all children with spoken language impairments show marked spelling difficulties. Findings are discussed in relation to theory, educational assessment and practice. As a result of this activity, readers will describe components of spoken language that predict children's morphological and phonological spelling performance. As a result of this activity, readers will describe how the spelling skills of children with SLI compare to age-matched and spelling age-matched control children. Readers will be able to interpret the variability in spelling performance seen in children with SLI. Copyright © 2013 Elsevier Inc. All rights reserved.

  9. Textese and use of texting by children with typical language development and Specific Language Impairment

    NARCIS (Netherlands)

    Blom, E.; van Dijk, C.; Vasić, N.; van Witteloostuijn, M.; Avrutin, S.

    The purpose of this study was to investigate texting and textese, which is the special register used for sending brief text messages, across children with typical development (TD) and children with Specific Language Impairment (SLI). Using elicitation techniques, texting and spoken language messages

  10. Textese and use of texting by children with typical language development and Specific Language Impairment

    NARCIS (Netherlands)

    Blom, W.B.T.; van Dijk, Chantal; Vasic, Nada; van Witteloostuijn, Merel; Avrutin, S.


    The purpose of this study was to investigate texting and textese, which is the special register used for sending brief text messages, across children with typical development (TD) and children with Specific Language Impairment (SLI). Using elicitation techniques, texting and spoken language messages

  11. First Steps to Endangered Language Documentation: The Kalasha Language, a Case Study (United States)

    Mela-Athanasopoulou, Elizabeth


    The present paper based on extensive fieldwork D conducted on Kalasha, an endangered language spoken in the three small valleys in Chitral District of Northwestern Pakistan, exposes a spontaneous dialogue-based elicitation of linguistic material used for the description and documentation of the language. After a brief display of the basic typology…

  12. How much exposure to English is necessary for a bilingual toddler to perform like a monolingual peer in language tests? (United States)

    Cattani, Allegra; Abbot-Smith, Kirsten; Farag, Rafalla; Krott, Andrea; Arreckx, Frédérique; Dennis, Ian; Floccia, Caroline


    Bilingual children are under-referred due to an ostensible expectation that they lag behind their monolingual peers in their English acquisition. The recommendations of the Royal College of Speech and Language Therapists (RCSLT) state that bilingual children should be assessed in both the languages known by the children. However, despite these recommendations, a majority of speech and language professionals report that they assess bilingual children only in English as bilingual children come from a wide array of language backgrounds and standardized language measures are not available for the majority of these. Moreover, even when such measures do exist, they are not tailored for bilingual children. It was asked whether a cut-off exists in the proportion of exposure to English at which one should expect a bilingual toddler to perform as well as a monolingual on a test standardized for monolingual English-speaking children. Thirty-five bilingual 2;6-year-olds exposed to British English plus an additional language and 36 British monolingual toddlers were assessed on the auditory component of the Preschool Language Scale, British Picture Vocabulary Scale and an object-naming measure. All parents completed the Oxford Communicative Development Inventory (Oxford CDI) and an exposure questionnaire that assessed the proportion of English in the language input. Where the CDI existed in the bilingual's additional language, these data were also collected. Hierarchical regression analyses found the proportion of exposure to English to be the main predictor of the performance of bilingual toddlers. Bilingual toddlers who received 60% exposure to English or more performed like their monolingual peers on all measures. K-means cluster analyses and Levene variance tests confirmed the estimated English exposure cut-off at 60% for all language measures. Finally, for one additional language for which we had multiple participants, additional language CDI production scores were

  13. Expressive Language Development in 45 Cochlear Implanted Children Following 2 Years of Implantation

    Directory of Open Access Journals (Sweden)

    Seyed Basir Hashemi


    Full Text Available Objectives: Profound hearing loss encounters children with delay in speech and language. As it is known language acquisition in young deaf children is a lengthy process, but cochlear implanted children have better spoken language skills than if they had not received the device. According to the importance of cochlear implant in deaf child's language development, this study evaluates the effect of different variables on child's language performance. Methods: 45 cochlear implanted children were tested, all of whom had used the device for at least 2 years. In order to evaluate the children, the NEWSHA test which is fitted for Persian speaking children was performed and language development of the children was compared through stepwise discriminant analysis. Results: After evaluation of the effect of different variables like child's age of implantation, participating in rehabilitation classes, parent's cooperation and their level of education, we came to a conclusion that the child's age of implantation and rehabilitation program significantly develop the child's language performance. Discussion: The value of cochlear implant in improvement of deaf children in speech, language perception, production and comprehension is confirmed by different studies which have been done on cochlear implanted children. Also, the present study indicates that language development in cochlear implanted children is highly related to their age of implantation and rehabilitation program.

  14. Grammar of Kove: An Austronesian Language of the West New Britain Province, Papua New Guinea (United States)

    Sato, Hiroko


    This dissertation is a descriptive grammar of Kove, an Austronesian language spoken in the West New Britain Province of Papua New Guinea. Kove is primarily spoken in 18 villages, including some on the small islands north of New Britain. There are about 9,000 people living in the area, but many are not fluent speakers of Kove. The dissertation…

  15. On the Conventionalization of Mouth Actions in Australian Sign Language. (United States)

    Johnston, Trevor; van Roekel, Jane; Schembri, Adam


    This study investigates the conventionalization of mouth actions in Australian Sign Language. Signed languages were once thought of as simply manual languages because the hands produce the signs which individually and in groups are the symbolic units most easily equated with the words, phrases and clauses of spoken languages. However, it has long been acknowledged that non-manual activity, such as movements of the body, head and the face play a very important role. In this context, mouth actions that occur while communicating in signed languages have posed a number of questions for linguists: are the silent mouthings of spoken language words simply borrowings from the respective majority community spoken language(s)? Are those mouth actions that are not silent mouthings of spoken words conventionalized linguistic units proper to each signed language, culturally linked semi-conventional gestural units shared by signers with members of the majority speaking community, or even gestures and expressions common to all humans? We use a corpus-based approach to gather evidence of the extent of the use of mouth actions in naturalistic Australian Sign Language-making comparisons with other signed languages where data is available--and the form/meaning pairings that these mouth actions instantiate.

  16. Toward a tactile language for human-robot interaction: two studies of tacton learning and performance. (United States)

    Barber, Daniel J; Reinerman-Jones, Lauren E; Matthews, Gerald


    Two experiments were performed to investigate the feasibility for robot-to-human communication of a tactile language using a lexicon of standardized tactons (tactile icons) within a sentence. Improvements in autonomous systems technology and a growing demand within military operations are spurring interest in communication via vibrotactile displays. Tactile communication may become an important element of human-robot interaction (HRI), but it requires the development of messaging capabilities approaching the communication power of the speech and visual signals used in the military. In Experiment 1 (N = 38), we trained participants to identify sets of directional, dynamic, and static tactons and tested performance and workload following training. In Experiment 2 (N = 76), we introduced an extended training procedure and tested participants' ability to correctly identify two-tacton phrases. We also investigated the impact of multitasking on performance and workload. Individual difference factors were assessed. Experiment 1 showed that participants found dynamic and static tactons difficult to learn, but the enhanced training procedure in Experiment 2 produced competency in performance for all tacton categories. Participants in the latter study also performed well on two-tacton phrases and when multitasking. However, some deficits in performance and elevation of workload were observed. Spatial ability predicted some aspects of performance in both studies. Participants may be trained to identify both single tactons and tacton phrases, demonstrating the feasibility of developing a tactile language for HRI. Tactile communication may be incorporated into multi-modal communication systems for HRI. It also has potential for human-human communication in challenging environments. © 2014, Human Factors and Ergonomics Society.

  17. Visualization of Distributed Data Structures for High Performance Fortran-Like Languages

    Directory of Open Access Journals (Sweden)

    Rainer Koppler


    Full Text Available This article motivates the usage of graphics and visualization for efficient utilization of High Performance Fortran's (HPF's data distribution facilities. It proposes a graphical toolkit consisting of exploratory and estimation tools which allow the programmer to navigate through complex distributions and to obtain graphical ratings with respect to load distribution and communication. The toolkit has been implemented in a mapping design and visualization tool which is coupled with a compilation system for the HPF predecessor Vienna Fortran. Since this language covers a superset of HPF's facilities, the tool may also be used for visualization of HPF data structures.

  18. Advances in natural language processing. (United States)

    Hirschberg, Julia; Manning, Christopher D


    Natural language processing employs computational techniques for the purpose of learning, understanding, and producing human language content. Early computational approaches to language research focused on automating the analysis of the linguistic structure of language and developing basic technologies such as machine translation, speech recognition, and speech synthesis. Today's researchers refine and make use of such tools in real-world applications, creating spoken dialogue systems and speech-to-speech translation engines, mining social media for information about health or finance, and identifying sentiment and emotion toward products and services. We describe successes and challenges in this rapidly advancing area. Copyright © 2015, American Association for the Advancement of Science.

  19. V2S: Voice to Sign Language Translation System for Malaysian Deaf People (United States)

    Mean Foong, Oi; Low, Tang Jung; La, Wai Wan

    The process of learning and understand the sign language may be cumbersome to some, and therefore, this paper proposes a solution to this problem by providing a voice (English Language) to sign language translation system using Speech and Image processing technique. Speech processing which includes Speech Recognition is the study of recognizing the words being spoken, regardless of whom the speaker is. This project uses template-based recognition as the main approach in which the V2S system first needs to be trained with speech pattern based on some generic spectral parameter set. These spectral parameter set will then be stored as template in a database. The system will perform the recognition process through matching the parameter set of the input speech with the stored templates to finally display the sign language in video format. Empirical results show that the system has 80.3% recognition rate.

  20. Performance of Low-Income Dual Language Learners Attending English-Only Schools on the Clinical Evaluation of Language Fundamentals-Fourth Edition, Spanish. (United States)

    Barragan, Beatriz; Castilla-Earls, Anny; Martinez-Nieto, Lourdes; Restrepo, M Adelaida; Gray, Shelley


    The aim of this study was to examine the performance of a group of Spanish-speaking, dual language learners (DLLs) who were attending English-only schools and came from low-income and low-parental education backgrounds on the Clinical Evaluation of Language Fundamentals-Fourth Edition, Spanish (CELF-4S; Semel, Wiig, & Secord, 2006). Spanish-speaking DLLs (N = 656), ages 5;0 (years;months) to 7;11, were tested for language impairment (LI) using the core language score of the CELF-4S and the English Structured Photographic Expressive Language Test (Dawson, Stout, & Eyer, 2003). A subsample (n = 299) was additionally tested using a Spanish language sample analysis and a newly developed Spanish morphosyntactic measure, for identification of children with LI and to conduct a receiver operating characteristics curve analysis. Over 50% of the sample scored more than 1 SD below the mean on the core language score. In our subsample, the sensitivity of the CELF-4S was 94%, and specificity was 65%, using a cutoff score of 85 as suggested in the manual. Using an empirically derived cutoff score of 78, the sensitivity was 86%, and the specificity was 80%. Results suggest that the CELF-4S overidentifies low-income Spanish-English DLLs attending English-only schools as presenting with LI. For this sample, 1 in every 3 Latino children from low socioeconomic status was incorrectly identified with LI. Clinicians should be cautious when using the CELF-4S to evaluate low-income Spanish-English DLLs and ensure that they have converging evidence before making diagnostic decisions.

  1. Openness to language and value diversity fosters multicultural team creativity and performance

    DEFF Research Database (Denmark)

    Lauring, Jakob; Paunova, Minna; Butler, Christina


    Multicultural teams are increasingly employed by organizations as a way to achieve international coordination, particularly when creativity and innovation are desired. Unfortunately, studies often fail to demonstrate the purported benefits associated with these teams, reporting difficulties with ...... open to value diversity. We conclude that multicultural teams are in want of a climate that is welcoming to both linguistic and cultural differences and suggest implications for theory and practice....... with communication and social integration, inhibiting creativity and performance. A survey-based study of multicultural academic teams (n = 1085) demonstrates that teams that are open to language diversity are more creative and perform better. We observe that performance is enhanced even further when teams are also...

  2. Working memory affects older adults' use of context in spoken-word recognition. (United States)

    Janse, Esther; Jesse, Alexandra


    Many older listeners report difficulties in understanding speech in noisy situations. Working memory and other cognitive skills may modulate older listeners' ability to use context information to alleviate the effects of noise on spoken-word recognition. In the present study, we investigated whether verbal working memory predicts older adults' ability to immediately use context information in the recognition of words embedded in sentences, presented in different listening conditions. In a phoneme-monitoring task, older adults were asked to detect as fast and as accurately as possible target phonemes in sentences spoken by a target speaker. Target speech was presented without noise, with fluctuating speech-shaped noise, or with competing speech from a single distractor speaker. The gradient measure of contextual probability (derived from a separate offline rating study) affected the speed of recognition. Contextual facilitation was modulated by older listeners' verbal working memory (measured with a backward digit span task) and age across listening conditions. Working memory and age, as well as hearing loss, were also the most consistent predictors of overall listening performance. Older listeners' immediate benefit from context in spoken-word recognition thus relates to their ability to keep and update a semantic representation of the sentence content in working memory.

  3. The socially-weighted encoding of spoken words: A dual-route approach to speech perception

    Directory of Open Access Journals (Sweden)

    Meghan eSumner


    Full Text Available Spoken words are highly variable. A single word may never be uttered the same way twice. As listeners, we regularly encounter speakers of different ages, genders, and accents, increasing the amount of variation we face. How listeners understand spoken words as quickly and adeptly as they do despite this variation remains an issue central to linguistic theory. We propose that learned acoustic patterns are mapped simultaneously to linguistic representations and to social representations. In doing so, we illuminate a paradox that results in the literature from, we argue, the focus on representations and the peripheral treatment of word-level phonetic variation. We consider phonetic variation more fully and highlight a growing body of work that is problematic for current theory: Words with different pronunciation variants are recognized equally well in immediate processing tasks, while an atypical, infrequent, but socially-idealized form is remembered better in the long-term. We suggest that the perception of spoken words is socially-weighted, resulting in sparse, but high-resolution clusters of socially-idealized episodes that are robust in immediate processing and are more strongly encoded, predicting memory inequality. Our proposal includes a dual-route approach to speech perception in which listeners map acoustic patterns in speech to linguistic and social representations in tandem. This approach makes novel predictions about the extraction of information from the speech signal, and provides a framework with which we can ask new questions. We propose that language comprehension, broadly, results from the integration of both linguistic and social information.

  4. Learning trajectories for speech motor performance in children with specific language impairment. (United States)

    Richtsmeier, Peter T; Goffman, Lisa


    Children with specific language impairment (SLI) often perform below expected levels, including on tests of motor skill and in learning tasks, particularly procedural learning. In this experiment we examined the possibility that children with SLI might also have a motor learning deficit. Twelve children with SLI and thirteen children with typical development (TD) produced complex nonwords in an imitation task. Productions were collected across three blocks, with the first and second blocks on the same day and the third block one week later. Children's lip movements while producing the nonwords were recorded using an Optotrak camera system. Movements were then analyzed for production duration and stability. Movement analyses indicated that both groups of children produced shorter productions in later blocks (corroborated by an acoustic analysis), and the rate of change was comparable for the TD and SLI groups. A nonsignificant trend for more stable productions was also observed in both groups. SLI is regularly accompanied by a motor deficit, and this study does not dispute that. However, children with SLI learned to make more efficient productions at a rate similar to their peers with TD, revealing some modification of the motor deficit associated with SLI. The reader will learn about deficits commonly associated with specific language impairment (SLI) that often occur alongside the hallmark language deficit. The authors present an experiment showing that children with SLI improved speech motor performance at a similar rate compared to typically developing children. The implication is that speech motor learning is not impaired in children with SLI. Copyright © 2015 Elsevier Inc. All rights reserved.

  5. Language Shift or Increased Bilingualism in South Africa: Evidence from Census Data (United States)

    Posel, Dorrit; Zeller, Jochen


    In the post-apartheid era, South Africa has adopted a language policy that gives official status to 11 languages (English, Afrikaans, and nine Bantu languages). However, English has remained the dominant language of business, public office, and education, and some research suggests that English is increasingly being spoken in domestic settings.…

  6. Key Data on Teaching Languages at School in Europe. 2017 Edition. Eurydice Report (United States)

    Baïdak, Nathalie; Balcon, Marie-Pascale; Motiejunaite, Akvile


    Linguistic diversity is part of Europe's DNA. It embraces not only the official languages of Member States, but also the regional and/or minority languages spoken for centuries on European territory, as well as the languages brought by the various waves of migrants. The coexistence of this variety of languages constitutes an asset, but it is also…

  7. Mapudungun According to Its Speakers: Mapuche Intellectuals and the Influence of Standard Language Ideology (United States)

    Lagos, Cristián; Espinoza, Marco; Rojas, Darío


    In this paper, we analyse the cultural models (or folk theory of language) that the Mapuche intellectual elite have about Mapudungun, the native language of the Mapuche people still spoken today in Chile as the major minority language. Our theoretical frame is folk linguistics and studies of language ideology, but we have also taken an applied…

  8. Working with the Bilingual Child Who Has a Language Delay. Meeting Learning Challenges (United States)

    Greenspan, Stanley I.


    It is very important to determine if a bilingual child's language delay is simply in English or also in the child's native language. Understandably, many children have higher levels of language development in the language spoken at home. To discover if this is the case, observe the child talking with his parents. Sometimes, even without…

  9. Extrinsic Cognitive Load Impairs Spoken Word Recognition in High- and Low-Predictability Sentences. (United States)

    Hunter, Cynthia R; Pisoni, David B

    Listening effort (LE) induced by speech degradation reduces performance on concurrent cognitive tasks. However, a converse effect of extrinsic cognitive load on recognition of spoken words in sentences has not been shown. The aims of the present study were to (a) examine the impact of extrinsic cognitive load on spoken word recognition in a sentence recognition task and (b) determine whether cognitive load and/or LE needed to understand spectrally degraded speech would differentially affect word recognition in high- and low-predictability sentences. Downstream effects of speech degradation and sentence predictability on the cognitive load task were also examined. One hundred twenty young adults identified sentence-final spoken words in high- and low-predictability Speech Perception in Noise sentences. Cognitive load consisted of a preload of short (low-load) or long (high-load) sequences of digits, presented visually before each spoken sentence and reported either before or after identification of the sentence-final word. LE was varied by spectrally degrading sentences with four-, six-, or eight-channel noise vocoding. Level of spectral degradation and order of report (digits first or words first) were between-participants variables. Effects of cognitive load, sentence predictability, and speech degradation on accuracy of sentence-final word identification as well as recall of preload digit sequences were examined. In addition to anticipated main effects of sentence predictability and spectral degradation on word recognition, we found an effect of cognitive load, such that words were identified more accurately under low load than high load. However, load differentially affected word identification in high- and low-predictability sentences depending on the level of sentence degradation. Under severe spectral degradation (four-channel vocoding), the effect of cognitive load on word identification was present for high-predictability sentences but not for low

  10. Comparisons of IQ in Children With and Without Cochlear Implants: Longitudinal Findings and Associations With Language. (United States)

    Cejas, Ivette; Mitchell, Christine M; Hoffman, Michael; Quittner, Alexandra L


    To make longitudinal comparisons of intelligence quotient (IQ) in children with cochlear implants (CIs) and typical hearing peers from early in development to the school-age period. Children with additional comorbidities and CIs were also evaluated. To estimate the impact of socioeconomic status and oral language on school-age cognitive performance. This longitudinal study evaluated nonverbal IQ in a multicenter, national sample of 147 children with CIs and 75 typically hearing peers. IQ was evaluated at baseline, prior to cochlear implantation, using the Bayley Scales of Infant and Toddler Development and the Leiter International Performance Scale. School-age IQ was assessed using the Wechsler Intelligence Scales for Children. For the current study, only the Perceptual Reasoning and Processing Speed indices were administered. Oral language was evaluated using the Comprehensive Assessment of Spoken Language. Children in the CI group scored within the normal range of intelligence at both time points. However, children with additional comorbidities scored significantly worse on the Processing Speed, but not the Perceptual Reasoning Index. Maternal education and language were significantly related to school-age IQ in both groups. Importantly, language was the strongest predictor of intellectual functioning in both children with CIs and normal hearing. These results suggest that children using cochlear implants perform similarly to hearing peers on measures of intelligence, but those with severe comorbidities are at-risk for cognitive deficits. Despite the strong link between socioeconomic status and intelligence, this association was no longer significant once spoken language performance was accounted for. These results reveal the important contributions that early intervention programs, which emphasize language and parent training, contribute to cognitive functioning in school-age children with CIs. For families from economically disadvantaged backgrounds, who are at

  11. Language and verbal reasoning skills in adolescents with 10 or more years of cochlear implant experience. (United States)

    Geers, Ann E; Sedey, Allison L


    The purpose of this study is to identify factors predictive of successful English language outcomes in adolescents who received a cochlear implant (CI) between 2 and 5 yrs of age. All 112 participants had been part of a previous study examining English language outcomes at the age of 8 and 9 yrs with CIs. The participants were given a battery of language and verbal reasoning tests in their preferred communication mode along with measures of working memory (digit span) and verbal rehearsal speed (sentence repetition duration). The degree to which students' language performance was enhanced when sign was added to spoken language was estimated at both test sessions. Multiple linear regression analyses were used to document factors contributing to overall language outcomes. A substantial proportion of the adolescents obtained test scores within or above 1SD compared with hearing age-mates in the tests' normative samples: 71% on a verbal intelligence test, 68% on a measure of language content, 71% on receptive vocabulary, and 74% on expressive vocabulary. Improvement in verbal intelligence scores over an 8-yr interval exceeded expectation based on age-mates in the test's normative sample. Better English language outcomes were associated with shorter duration of deafness before cochlear implantation, higher nonverbal intelligence, higher family socioeconomic status, longer digit spans, and faster verbal rehearsal speed as measured by sentence repetition rate. Students whose current receptive vocabulary scores were not enhanced by the addition of signs also exhibited higher English language scores than those without sign enhancement; however, sign enhancement demonstrated in the elementary school years was not predictive of later high-school language skills. Results of this study support the provision of CIs to children at the youngest age possible. In addition, it highlights the substantial role that cognition plays in later language outcomes. Although the students' use

  12. Written cohesion in children with and without language learning disabilities. (United States)

    Koutsoftas, Anthony D; Petersen, Victoria


    Cohesion refers to the linguistic elements of discourse that contribute to its continuity and is an important element to consider as part of written language intervention, especially in children with language learning disabilities (LLD). There is substantial evidence that children with LLD perform more poorly than typically developing (TD) peers on measures of cohesion in spoken language and on written transcription measures; however, there is far less research comparing groups on cohesion as a measure of written language across genres. The current study addresses this gap through the following two aims. First, to describe and compare cohesion in narrative and expository writing samples of children with and without language learning disabilities. Second, to relate measures of cohesion to written transcription and translation measures, oral language, and writing quality. Fifty intermediate-grade children produced one narrative and one expository writing sample from which measures of written cohesion were obtained. These included the frequency, adequacy and complexity of referential and conjunctive ties. Expository samples resulted in more complex cohesive ties and children with TD used more complex ties than peers with LLD. Different relationships among cohesion measures and writing were observed for narrative verse expository samples. Findings from this study demonstrate cohesion as a discourse-level measure of written transcription and how the use of cohesion can vary by genre and group (LLD, TD). Clinical implications for assessment, intervention, and future research are provided. © 2016 Royal College of Speech and Language Therapists.

  13. Differential Performance by English Language Learners on an Inquiry-Based Science Assessment (United States)

    Turkan, Sultan; Liu, Ou Lydia


    The performance of English language learners (ELLs) has been a concern given the rapidly changing demographics in US K-12 education. This study aimed to examine whether students' English language status has an impact on their inquiry science performance. Differential item functioning (DIF) analysis was conducted with regard to ELL status on an inquiry-based science assessment, using a multifaceted Rasch DIF model. A total of 1,396 seventh- and eighth-grade students took the science test, including 313 ELL students. The results showed that, overall, non-ELLs significantly outperformed ELLs. Of the four items that showed DIF, three favored non-ELLs while one favored ELLs. The item that favored ELLs provided a graphic representation of a science concept within a family context. There is some evidence that constructed-response items may help ELLs articulate scientific reasoning using their own words. Assessment developers and teachers should pay attention to the possible interaction between linguistic challenges and science content when designing assessment for and providing instruction to ELLs.

  14. Language and the origin of numerical concepts. (United States)

    Gelman, Rochel; Gallistel, C R


    Reports of research with the Pirahã and Mundurukú Amazonian Indians of Brazil lend themselves to discussions of the role of language in the origin of numerical concepts. The research findings indicate that, whether or not humans have an extensive counting list, they share with nonverbal animals a language-independent representation of number, with limited, scale-invariant precision. What causal role, then, does knowledge of the language of counting serve? We consider the strong Whorfian proposal, that of linguistic determinism; the weak Whorfian hypothesis, that language influences how we think; and that the "language of thought" maps to spoken language or symbol systems.

  15. Development of brain networks involved in spoken word processing of Mandarin Chinese. (United States)

    Cao, Fan; Khalid, Kainat; Lee, Rebecca; Brennan, Christine; Yang, Yanhui; Li, Kuncheng; Bolger, Donald J; Booth, James R


    Developmental differences in phonological and orthographic processing of Chinese spoken words were examined in 9-year-olds, 11-year-olds and adults using functional magnetic resonance imaging (fMRI). Rhyming and spelling judgments were made to two-character words presented sequentially in the auditory modality. Developmental comparisons between adults and both groups of children combined showed that age-related changes in activation in visuo-orthographic regions depended on a task. There were developmental increases in the left inferior temporal gyrus and the right inferior occipital gyrus in the spelling task, suggesting more extensive visuo-orthographic processing in a task that required access to these representations. Conversely, there were developmental decreases in activation in the left fusiform gyrus and left middle occipital gyrus in the rhyming task, suggesting that the development of reading is marked by reduced involvement of orthography in a spoken language task that does not require access to these orthographic representations. Developmental decreases may arise from the existence of extensive homophony (auditory words that have multiple spellings) in Chinese. In addition, we found that 11-year-olds and adults showed similar activation in the left superior temporal gyrus across tasks, with both groups showing greater activation than 9-year-olds. This pattern suggests early development of perceptual representations of phonology. In contrast, 11-year-olds and 9-year-olds showed similar activation in the left inferior frontal gyrus across tasks, with both groups showing weaker activation than adults. This pattern suggests late development of controlled retrieval and selection of lexical representations. Altogether, this study suggests differential effects of character acquisition on development of components of the language network in Chinese as compared to previous reports on alphabetic languages. Published by Elsevier Inc.

  16. Early Sign Language Exposure and Cochlear Implantation Benefits. (United States)

    Geers, Ann E; Mitchell, Christine M; Warner-Czyz, Andrea; Wang, Nae-Yuh; Eisenberg, Laurie S


    Most children with hearing loss who receive cochlear implants (CI) learn spoken language, and parents must choose early on whether to use sign language to accompany speech at home. We address whether parents' use of sign language before and after CI positively influences auditory-only speech recognition, speech intelligibility, spoken language, and reading outcomes. Three groups of children with CIs from a nationwide database who differed in the duration of early sign language exposure provided in their homes were compared in their progress through elementary grades. The groups did not differ in demographic, auditory, or linguistic characteristics before implantation. Children without early sign language exposure achieved better speech recognition skills over the first 3 years postimplant and exhibited a statistically significant advantage in spoken language and reading near the end of elementary grades over children exposed to sign language. Over 70% of children without sign language exposure achieved age-appropriate spoken language compared with only 39% of those exposed for 3 or more years. Early speech perception predicted speech intelligibility in middle elementary grades. Children without sign language exposure produced speech that was more intelligible (mean = 70%) than those exposed to sign language (mean = 51%). This study provides the most compelling support yet available in CI literature for the benefits of spoken language input for promoting verbal development in children implanted by 3 years of age. Contrary to earlier published assertions, there was no advantage to parents' use of sign language either before or after CI. Copyright © 2017 by the American Academy of Pediatrics.

  17. Multiple Language Use Influences Oculomotor Task Performance: Neurophysiological Evidence of a Shared Substrate between Language and Motor Control.

    Directory of Open Access Journals (Sweden)

    Karin Heidlmayr

    Full Text Available In the present electroencephalographical study, we asked to which extent executive control processes are shared by both the language and motor domain. The rationale was to examine whether executive control processes whose efficiency is reinforced by the frequent use of a second language can lead to a benefit in the control of eye movements, i.e. a non-linguistic activity. For this purpose, we administrated to 19 highly proficient late French-German bilingual participants and to a control group of 20 French monolingual participants an antisaccade task, i.e. a specific motor task involving control. In this task, an automatic saccade has to be suppressed while a voluntary eye movement in the opposite direction has to be carried out. Here, our main hypothesis is that an advantage in the antisaccade task should be observed in the bilinguals if some properties of the control processes are shared between linguistic and motor domains. ERP data revealed clear differences between bilinguals and monolinguals. Critically, we showed an increased N2 effect size in bilinguals, thought to reflect better efficiency to monitor conflict, combined with reduced effect sizes on markers reflecting inhibitory control, i.e. cue-locked positivity, the target-locked P3 and the saccade-locked presaccadic positivity (PSP. Moreover, effective connectivity analyses (dynamic causal modelling; DCM on the neuronal source level indicated that bilinguals rely more strongly on ACC-driven control while monolinguals rely on PFC-driven control. Taken together, our combined ERP and effective connectivity findings may reflect a dynamic interplay between strengthened conflict monitoring, associated with subsequently more efficient inhibition in bilinguals. Finally, L2 proficiency and immersion experience constitute relevant factors of the language background that predict efficiency of inhibition. To conclude, the present study provided ERP and effective connectivity evidence for domain


    Directory of Open Access Journals (Sweden)

    Huma Imran Khan


    Full Text Available The prime purpose of this study was to explore a correlation between bilingual instruction in an ESL class and the class performance of the ESL learners at the secondary level. Quantitative research method was used to evaluate the test performance of 60 ESL learners divided into two groups: One was the controlled group (which was given instructions in L2 only and the other was the treatment group (which was given instructions in both L1 and L2 in Public School settings. Apart from the students, 15 language teachers’ feedback upon their perceptions of L1 usage in L2 classrooms was taken by using the Likert scale feedback forms. The results confirmed that the instructions given bilingually are directly associated to improved learner outcomes and teachers’ responses for the usage of L1 in classrooms showed a strong positive response.

  19. The impact of visual layout factors on performance in Web pages: a cross-language study. (United States)

    Parush, Avi; Shwarts, Yonit; Shtub, Avy; Chandra, M Jeya


    Visual layout has a strong impact on performance and is a critical factor in the design of graphical user interfaces (GUIs) and Web pages. Many design guidelines employed in Web page design were inherited from human performance literature and GUI design studies and practices. However, few studies have investigated the more specific patterns of performance with Web pages that may reflect some differences between Web page and GUI design. We investigated interactions among four visual layout factors in Web page design (quantity of links, alignment, grouping indications, and density) in two experiments: one with pages in Hebrew, entailing right-to-left reading, and the other with English pages, entailing left-to-right reading. Some performance patterns (measured by search times and eye movements) were similar between languages. Performance was particularly poor in pages with many links and variable densities, but it improved with the presence of uniform density. Alignment was not shown to be a performance-enhancing factor. The findings are discussed in terms of the similarities and differences in the impact of layout factors between GUIs and Web pages. Actual or potential applications of this research include specific guidelines for Web page design.

  20. Brain-to-text: Decoding spoken phrases from phone representations in the brain

    Directory of Open Access Journals (Sweden)

    Christian eHerff


    Full Text Available It has long been speculated whether communication between humans and machines based on natural speech related cortical activity is possible. Over the past decade, studies have suggested that it is feasible to recognize isolated aspects of speech from neural signals, such as auditory features, phones or one of a few isolated words. However, until now it remained an unsolved challenge to decode continuously spoken speech from the neural substrate associated with speech and language processing. Here, we show for the first time that continuously spoken speech can be decoded into the expressed words from intracranial electrocorticographic (ECoG recordings. Specifically, we implemented a system, which we call Brain-To-Text that models single phones, employs techniques from automatic speech recognition (ASR, and thereby transforms brain activity while speaking into the corresponding textual representation. Our results demonstrate that our system achieved word error rates as low as 25% and phone error rates below 50%. Additionally, our approach contributes to the current understanding of the neural basis of continuous speech production by identifying those cortical regions that hold substantial information about individual phones. In conclusion, the Brain-To-Text system described in this paper represents an important step towards human-machine communication based on imagined speech.

  1. The Bakayat SpokenText Tradition The Articulation of Religious Value and Social Discourse of Sasak Community in Lombok

    Directory of Open Access Journals (Sweden)

    I Made Suyasa


    narrative structure, which was seen from the narrative, rhythmic, and storytelling style. The bakayat tradition had two functions, namelythe manifest function and latent function, and there was a shifting function. The manifest function of the Sasak bakayat tradition was that it was used as the Islamic preaching media, while the latent function was that it was used as the educational, social, economic, and customs media. The shifting function of the tradition was that it was used as an entertainment that described the function of social identity, local knowledge, and competition activity. Based on the cultural and social context, the tradition implied that the bakayat tradition had historical, political, aesthetic, and social significances. The translation during the performance of the tradition was accompanied with the meaning interpretation or review conveyed in Sasak language. Each performance had new meaning interpretation of the text; the dominant meaning of the text was about the singleness of God, negotiation, ambiguity, and rejection. The articulation of the Sasak bakayat text was a form of new meaning. The articulation of the religious value and social discourse of the bakayat text was about the story of Prophet Joseph made up of the spoken-text form using the typical Sasak song.

  2. The determinants of spoken and written picture naming latencies. (United States)

    Bonin, Patrick; Chalard, Marylène; Méot, Alain; Fayol, Michel


    The influence of nine variables on the latencies to write down or to speak aloud the names of pictures taken from Snodgrass and Vanderwart (1980) was investigated in French adults. The major determinants of both written and spoken picture naming latencies were image variability, image agreement and age of acquisition. To a lesser extent, name agreement was also found to have an impact in both production modes. The implications of the findings for theoretical views of both spoken and written picture naming are discussed.

  3. Rhythm in language acquisition. (United States)

    Langus, Alan; Mehler, Jacques; Nespor, Marina


    Spoken language is governed by rhythm. Linguistic rhythm is hierarchical and the rhythmic hierarchy partially mimics the prosodic as well as the morpho-syntactic hierarchy of spoken language. It can thus provide learners with cues about the structure of the language they are acquiring. We identify three universal levels of linguistic rhythm - the segmental level, the level of the metrical feet and the phonological phrase level - and discuss why primary lexical stress is not rhythmic. We survey experimental evidence on rhythm perception in young infants and native speakers of various languages to determine the properties of linguistic rhythm that are present at birth, those that mature during the first year of life and those that are shaped by the linguistic environment of language learners. We conclude with a discussion of the major gaps in current knowledge on linguistic rhythm and highlight areas of interest for future research that are most likely to yield significant insights into the nature, the perception, and the usefulness of linguistic rhythm. Copyright © 2016 Elsevier Ltd. All rights reserved.

  4. From spoken narratives to domain knowledge: mining linguistic data for medical image understanding. (United States)

    Guo, Xuan; Yu, Qi; Alm, Cecilia Ovesdotter; Calvelli, Cara; Pelz, Jeff B; Shi, Pengcheng; Haake, Anne R


    Extracting useful visual clues from medical images allowing accurate diagnoses requires physicians' domain knowledge acquired through years of systematic study and clinical training. This is especially true in the dermatology domain, a medical specialty that requires physicians to have image inspection experience. Automating or at least aiding such efforts requires understanding physicians' reasoning processes and their use of domain knowledge. Mining physicians' references to medical concepts in narratives during image-based diagnosis of a disease is an interesting research topic that can help reveal experts' reasoning processes. It can also be a useful resource to assist with design of information technologies for image use and for image case-based medical education systems. We collected data for analyzing physicians' diagnostic reasoning processes by conducting an experiment that recorded their spoken descriptions during inspection of dermatology images. In this paper we focus on the benefit of physicians' spoken descriptions and provide a general workflow for mining medical domain knowledge based on linguistic data from these narratives. The challenge of a medical image case can influence the accuracy of the diagnosis as well as how physicians pursue the diagnostic process. Accordingly, we define two lexical metrics for physicians' narratives--lexical consensus score and top N relatedness score--and evaluate their usefulness by assessing the diagnostic challenge levels of corresponding medical images. We also report on clustering medical images based on anchor concepts obtained from physicians' medical term usage. These analyses are based on physicians' spoken narratives that have been preprocessed by incorporating the Unified Medical Language System for detecting medical concepts. The image rankings based on lexical consensus score and on top 1 relatedness score are well correlated with those based on challenge levels (Spearman correlation>0.5 and Kendall

  5. A Comparative Study of Chinese and Japanese Language Performance in Advertising

    Directory of Open Access Journals (Sweden)

    Chen Ting


    Full Text Available Advertising as a propaganda tool, especially for the propaganda of commodities, is known by the public. However, advertising is not only a tool, but also a carrier of culture. It can reflect the culture of a particular society. We collected and contrasted a lot of the Sino-Japanese advertisement language in the thesis, analyzed the ubiquitous language differences in these advertisement languages, and also explored the differences between two cultures hidden beneath these language differences.

  6. Medial olivocochlear function in children with poor speech-in-noise performance and language disorder. (United States)

    Rocha-Muniz, Caroline Nunes; Mamede Carvallo, Renata Mota; Schochat, Eliane


    Contralateral masking of transient-evoked otoacoustic emissions is a phenomenon that suggests an inhibitory effect of the olivocochlear efferent auditory pathway. Many studies have been inconclusive in demonstrating a clear connection between this system and a behavioral speech-in-noise listening skill. The purpose of this study was to investigate the activation of a medial olivocochlear (MOC) efferent in children with poor speech-in-noise (PSIN) performance and children with language impairment and PSIN (SLI + PSIN). Transient evoked otoacoustic emissions (TEOAEs) with and without contralateral white noise were tested in 52 children (between 6 and 12 years). These children were arranged in three groups: typical development (TD) (n = 25), PSIN (n = 14) and SLI + PSI (n = 13). PSIN and SLI + PSI groups presented reduced otoacoustic emission suppression in comparison with the TD group. Our finding suggests differences in MOC function among children with typical development and children with poor SIN and language problems. Copyright © 2017 Elsevier B.V. All rights reserved.

  7. Analyzing Student Performance and Attitudes toward Textual versus Iconic Programming Languages (United States)

    Lin, Janet Mei-Chuen; Yang, Mei-Ching


    In this study half of 52 sixth graders learned to program in MSWLogo and the other half in Drape. An analysis of students' test scores revealed that Drape (an iconic language) seemed to have a steeper learning curve than MSWLogo (a textual language). However, as students gradually became more familiar with either language, the difference in…

  8. [Prosody, speech input and language acquisition]. (United States)

    Jungheim, M; Miller, S; Kühn, D; Ptok, M


    In order to acquire language, children require speech input. The prosody of the speech input plays an important role. In most cultures adults modify their code when communicating with children. Compared to normal speech this code differs especially with regard to prosody. For this review a selective literature search in PubMed and Scopus was performed. Prosodic characteristics are a key feature of spoken language. By analysing prosodic features, children gain knowledge about underlying grammatical structures. Child-directed speech (CDS) is modified in a way that meaningful sequences are highlighted acoustically so that important information can be extracted from the continuous speech flow more easily. CDS is said to enhance the representation of linguistic signs. Taking into consideration what has previously been described in the literature regarding the perception of suprasegmentals, CDS seems to be able to support language acquisition due to the correspondence of prosodic and syntactic units. However, no findings have been reported, stating that the linguistically reduced CDS could hinder first language acquisition.

  9. Technological Devices Improving System of Translating Languages: What About their Usefulness on the Applicability in Medicine and Health Sciences?

    Directory of Open Access Journals (Sweden)

    Adilia Maria Pires Sciarra


    Full Text Available ABSTRACT INTRODUCTION: In a world in which global communication is becoming ever more important and in which English is increasingly positioned as the pre-eminent international language, that is, English as a Lingua Franca refers to the use of English as a medium of communication between peoples of different languages. It is important to highlight the positive advances in communication in health, provided by technology. OBJECTIVE: To present an overview on some technological devices of translating languages provided by the Web as well as to point out some advantages and disadvantages specially using Google Translate in Medicine and Health Sciences. METHODS: A bibliographical survey was performed to provide an overview on the usefulness of online translators for applicability using written and spoken languages. RESULTS: As we have to consider this question to be further surely answered, this study could present some advantages and disadvantages in using translating online devices. CONCLUSION: Considering Medicine and Health Sciences as expressive into the human scientific knowledge to be spread worldwidely; technological devices available on communication should be used to overcome some language barriers either written or spoken, but with some caution depending on the context of their applicability.

  10. Technological Devices Improving System of Translating Languages: What About their Usefulness on the Applicability in Medicine and Health Sciences? (United States)

    Sciarra, Adilia Maria Pires; Batigália, Fernando; Oliveira, Marcos Aurélio Barboza de


    In a world in which global communication is becoming ever more important and in which English is increasingly positioned as the pre-eminent international language, that is, English as a Lingua Franca refers to the use of English as a medium of communication between peoples of different languages. It is important to highlight the positive advances in communication in health, provided by technology. To present an overview on some technological devices of translating languages provided by the Web as well as to point out some advantages and disadvantages specially using Google Translate in Medicine and Health Sciences. A bibliographical survey was performed to provide an overview on the usefulness of online translators for applicability using written and spoken languages. As we have to consider this question to be further surely answered, this study could present some advantages and disadvantages in using translating online devices. Considering Medicine and Health Sciences as expressive into the human scientific knowledge to be spread worldwidely; technological devices available on communication should be used to overcome some language barriers either written or spoken, but with some caution depending on the context of their applicability.

  11. A Transcription Scheme for Languages Employing the Arabic Script Motivated by Speech Processing Application

    National Research Council Canada - National Science Library

    Ganjavi, Shadi; Georgiou, Panayiotis G; Narayanan, Shrikanth


    ... (The DARPA Babylon Program; Narayanan, 2003). In this paper, we discuss transcription systems needed for automated spoken language processing applications in Persian that uses the Arabic script for writing...

  12. Online learning in speech and language therapy: Student performance and attitudes. (United States)

    Upton, Dominic


    Behavioural studies form an essential component of the Speech and Language Therapy (SLT) undergraduate degree. This study aimed to produce online teaching material in behavioural studies suitable for undergraduate SLT students, explore students' views on the online material, record their performance when taught through this innovative method and compare their performance to a group taught through the traditional lecture based method. Finally, it aimed to explore the relationship between engagement with the module and performance. SLT students completed an online health psychology/sociology module and their performance was compared to students who completed a traditional lecture based course. Student evaluations of the online course were also recorded as was their engagement with the online module. Results suggested that there was no significant difference between students taught through an online medium compared to those taught through "traditional lectures". An evaluation survey suggested that students appeared to enjoy the material although there was some reluctance to develop an independent learning style. Online learning has a great deal to offer SLT education. However, material has to be developed that can both engage and motivate learners, thereby enhancing student independent learning.

  13. Connecting the Dots for English Language Learners: How Odds-Beating Elementary School Educators Monitor and Use Student Performance Data (United States)

    Wilcox, Kristen Campbell; Gregory, Karen; Yu, Lisa


    This article reports on findings from a multiple case study investigating the nature of educators' approaches toward monitoring English language learners' (ELLs) performance and using data to improve instruction and apply appropriate interventions. Six New York elementary schools where ELLs' performance was better than predicted (i.e.…

  14. Long-term temporal tracking of speech rate affects spoken-word recognition. (United States)

    Baese-Berk, Melissa M; Heffner, Christopher C; Dilley, Laura C; Pitt, Mark A; Morrill, Tuuli H; McAuley, J Devin


    Humans unconsciously track a wide array of distributional characteristics in their sensory environment. Recent research in spoken-language processing has demonstrated that the speech rate surrounding a target region within an utterance influences which words, and how many words, listeners hear later in that utterance. On the basis of hypotheses that listeners track timing information in speech over long timescales, we investigated the possibility that the perception of words is sensitive to speech rate over such a timescale (e.g., an extended conversation). Results demonstrated that listeners tracked variation in the overall pace of speech over an extended duration (analogous to that of a conversation that listeners might have outside the lab) and that this global speech rate influenced which words listeners reported hearing. The effects of speech rate became stronger over time. Our findings are consistent with the hypothesis that neural entrainment by speech occurs on multiple timescales, some lasting more than an hour. © The Author(s) 2014.

  15. Bilingual recognition memory: stronger performance but weaker levels-of-processing effects in the less fluent language. (United States)

    Francis, Wendy S; Gutiérrez, Marisela


    The effects of bilingual proficiency on recognition memory were examined in an experiment with Spanish-English bilinguals. Participants learned lists of words in English and Spanish under shallow- and deep-encoding conditions. Overall, hit rates were higher, discrimination greater, and response times shorter in the nondominant language, consistent with effects previously observed for lower frequency words. Levels-of-processing effects in hit rates, discrimination, and response time were stronger in the dominant language. Specifically, with shallow encoding, the advantage for the nondominant language was larger than with deep encoding. The results support the idea that memory performance in the nondominant language is impacted by both the greater demand for cognitive resources and the lower familiarity of the words.

  16. Allophones, not phonemes in spoken-word recognition

    NARCIS (Netherlands)

    Mitterer, H.A.; Reinisch, E.; McQueen, J.M.


    What are the phonological representations that listeners use to map information about the segmental content of speech onto the mental lexicon during spoken-word recognition? Recent evidence from perceptual-learning paradigms seems to support (context-dependent) allophones as the basic

  17. The Impact of Orthographic Consistency on German Spoken Word Identification (United States)

    Beyermann, Sandra; Penke, Martina


    An auditory lexical decision experiment was conducted to find out whether sound-to-spelling consistency has an impact on German spoken word processing, and whether such an impact is different at different stages of reading development. Four groups of readers (school children in the second, third and fifth grades, and university students)…

  18. Individual Differences in Online Spoken Word Recognition: Implications for SLI (United States)

    McMurray, Bob; Samelson, Vicki M.; Lee, Sung Hee; Tomblin, J. Bruce


    Thirty years of research has uncovered the broad principles that characterize spoken word processing across listeners. However, there have been few systematic investigations of individual differences. Such an investigation could help refine models of word recognition by indicating which processing parameters are likely to vary, and could also have…

  19. Spoken Word Recognition of Chinese Words in Continuous Speech (United States)

    Yip, Michael C. W.


    The present study examined the role of positional probability of syllables played in recognition of spoken word in continuous Cantonese speech. Because some sounds occur more frequently at the beginning position or ending position of Cantonese syllables than the others, so these kinds of probabilistic information of syllables may cue the locations…

  20. Using the Corpus of Spoken Afrikaans to generate an Afrikaans ...

    African Journals Online (AJOL)

    This paper presents two chatbot systems, ALICE and. Elizabeth, illustrating the dialogue knowledge representation and pattern matching techniques of each. We discuss the problems which arise when using the. Corpus of Spoken Afrikaans (Korpus Gesproke Afrikaans) to retrain the ALICE chatbot system with human ...

  1. Oral and Literate Strategies in Spoken and Written Narratives. (United States)

    Tannen, Deborah


    Discusses comparative analysis of spoken and written versions of a narrative to demonstrate that features which have been identified as characterizing oral discourse are also found in written discourse and that the written short story combines syntactic complexity expected in writing with features which create involvement expected in speaking.…

  2. Pedagogy for Liberation: Spoken Word Poetry in Urban Schools (United States)

    Fiore, Mia


    The Black Arts Movement of the 1960s and 1970s, hip hop of the 1980s and early 1990s, and spoken word poetry have each attempted to initiate the dialogical process outlined by Paulo Freire as necessary in overturning oppression. Each art form has done this by critically engaging with the world and questioning dominant systems of power. However,…

  3. "Context and Spoken Word Recognition in a Novel Lexicon": Correction (United States)

    Revill, Kathleen Pirog; Tanenhaus, Michael K.; Aslin, Richard N.


    Reports an error in "Context and spoken word recognition in a novel lexicon" by Kathleen Pirog Revill, Michael K. Tanenhaus and Richard N. Aslin ("Journal of Experimental Psychology: Learning, Memory, and Cognition," 2008[Sep], Vol 34[5], 1207-1223). Figure 9 was inadvertently duplicated as Figure 10. Figure 9 in the original article was correct.…

  4. Producing complex spoken numerals for time and space

    NARCIS (Netherlands)

    Meeuwissen, M.H.W.


    This thesis addressed the spoken production of complex numerals for time and space. The production of complex numerical expressions like those involved in telling time (e.g., 'quarter to four') or producing house numbers (e.g., 'two hundred forty-five') has been almost completely ignored. Yet, adult

  5. A memory-based shallow parser for spoken Dutch

    NARCIS (Netherlands)

    Canisius, S.V.M.; van den Bosch, A.; Decadt, B.; Hoste, V.; De Pauw, G.


    We describe the development of a Dutch memory-based shallow parser. The availability of large treebanks for Dutch, such as the one provided by the Spoken Dutch Corpus, allows memory-based learners to be trained on examples of shallow parsing taken from the treebank, and act as a shallow parser after

  6. In a Manner of Speaking: Assessing Frequent Spoken Figurative Idioms to Assist ESL/EFL Teachers (United States)

    Grant, Lynn E.


    This article outlines criteria to define a figurative idiom, and then compares the frequent figurative idioms identified in two sources of spoken American English (academic and contemporary) to their frequency in spoken British English. This is done by searching the spoken part of the British National Corpus (BNC), to see whether they are frequent…

  7. Instructional Benefits of Spoken Words: A Review of Cognitive Load Factors (United States)

    Kalyuga, Slava


    Spoken words have always been an important component of traditional instruction. With the development of modern educational technology tools, spoken text more often replaces or supplements written or on-screen textual representations. However, there could be a cognitive load cost involved in this trend, as spoken words can have both benefits and…

  8. The relationship between spoken English proficiency and participation in higher education, employment and income from two Australian censuses. (United States)

    Blake, Helen L; Mcleod, Sharynne; Verdon, Sarah; Fuller, Gail


    Proficiency in the language of the country of residence has implications for an individual's level of education, employability, income and social integration. This paper explores the relationship between the spoken English proficiency of residents of Australia on census day and their educational level, employment and income to provide insight into multilingual speakers' ability to participate in Australia as an English-dominant society. Data presented are derived from two Australian censuses i.e. 2006 and 2011 of over 19 million people. The proportion of Australians who reported speaking a language other than English at home was 21.5% in the 2006 census and 23.2% in the 2011 census. Multilingual speakers who also spoke English very well were more likely to have post-graduate qualifications, full-time employment and high income than monolingual English-speaking Australians. However, multilingual speakers who reported speaking English not well were much less likely to have post-graduate qualifications or full-time employment than monolingual English-speaking Australians. These findings provide insight into the socioeconomic and educational profiles of multilingual speakers, which will inform the understanding of people such as speech-language pathologists who provide them with support. The results indicate spoken English proficiency may impact participation in Australian society. These findings challenge the "monolingual mindset" by demonstrating that outcomes for multilingual speakers in education, employment and income are higher than for monolingual speakers.

  9. The Plausibility of Tonal Evolution in the Malay Dialect Spoken in Thailand: Evidence from an Acoustic Study

    Directory of Open Access Journals (Sweden)

    Phanintra Teeranon


    Full Text Available The F0 values of vowels following voiceless consonants are higher than those of vowels following voiced consonants; high vowels have a higher F0 than low vowels. It has also been found that when high vowels follow voiced consonants, the F0 values decrease. In contrast, low vowels following voiceless consonants show increasing F0 values. In other words, the voicing of initial consonants has been found to counterbalance the intrinsic F0 values of high and low vowels (House and Fairbanks 1953, Lehiste and Peterson 1961, Lehiste 1970, Laver 1994, Teeranon 2006. To test whether these three findings are applicable to a disyllabic language, the F0 values of high and low vowels following voiceless and voiced consonants were studied in a Malay dialect of the Austronesian language family spoken in Pathumthani Province, Thailand. The data was collected from three male informants, aged 30-35. The Praat program was used for acoustic analysis. The findings revealed the influence of the voicing of initial consonants on the F0 of vowels to be greater than that of the influence of vowel height. Evidence from this acoustic study shows the plausibility for the Malay dialect spoken in Pathumthani to become a tonal language by the influence of initial consonants rather by the influence of the high-low vowel dimension.

  10. "We call it Springbok-German!": language contact in the German communities in South Africa.


    Franke, Katharina


    Varieties of German are spoken all over the world, some of which have been maintained for prolonged periods of time. As a result, these transplanted varieties often show traces of the ongoing language contact as specific to their particular context. This thesis explores one such transplanted German language variety – Springbok- German – as spoken by a small subset of German Lutherans in South Africa. Specifically, this study takes as its focus eight rural German communities acr...

  11. Neural dynamics of morphological processing in spoken word comprehension: Laterality and automaticity

    Directory of Open Access Journals (Sweden)

    Caroline M. Whiting


    Full Text Available Rapid and automatic processing of grammatical complexity is argued to take place during speech comprehension, engaging a left-lateralised fronto-temporal language network. Here we address how neural activity in these regions is modulated by the grammatical properties of spoken words. We used combined magneto- and electroencephalography (MEG, EEG to delineate the spatiotemporal patterns of activity that support the recognition of morphologically complex words in English with inflectional (-s and derivational (-er affixes (e.g. bakes, baker. The mismatch negativity (MMN, an index of linguistic memory traces elicited in a passive listening paradigm, was used to examine the neural dynamics elicited by morphologically complex words. Results revealed an initial peak 130-180 ms after the deviation point with a major source in left superior temporal cortex. The localisation of this early activation showed a sensitivity to two grammatical properties of the stimuli: 1 the presence of morphological complexity, with affixed words showing increased left-laterality compared to non-affixed words; and 2 the grammatical category, with affixed verbs showing greater left-lateralisation in inferior frontal gyrus compared to affixed nouns (bakes vs. beaks. This automatic brain response was additionally sensitive to semantic coherence (the meaning of the stem vs. the meaning of the whole form in fronto-temporal regions. These results demonstrate that the spatiotemporal pattern of neural activity in spoken word processing is modulated by the presence of morphological structure, predominantly engaging the left-hemisphere’s fronto-temporal language network, and does not require focused attention on the linguistic input.

  12. Morphosyntactic constructs in the development of spoken and written Hebrew text production. (United States)

    Ravid, Dorit; Zilberbuch, Shoshana


    This study examined the distribution of two Hebrew nominal structures-N-N compounds and denominal adjectives-in spoken and written texts of two genres produced by 90 native-speaking participants in three age groups: eleven/twelve-year-olds (6th graders), sixteen/seventeen-year-olds (11th graders), and adults. The two constructions are later linguistic acquisitions, part of the profound lexical and syntactic changes that occur in language development during the school years. They are investigated in the context of learning how modality (speech vs. writing) and genre (biographical vs. expository texts) affect the production of continuous discourse. Participants were asked to speak and write about two topics, one biographical, describing the life of a public figure or of a friend; and another, expository, discussing one of ten topics such as the cinema, cats, or higher academic studies. N-N compounding was found to be the main device of complex subcategorization in Hebrew discourse, unrelated to genre. Denominal adjectives are a secondary subcategorizing device emerging only during the late teen years, a linguistic resource untapped until very late, more restricted to specific text types than N-N compounding, and characteristic of expository writing. Written texts were found to be denser than spoken texts lexically and syntactically as measured by number of novel N-N compounds and denominal adjectives per clause, and in older age groups this difference was found to be more pronounced. The paper contributes to our understanding of how the syntax/lexicon interface changes with age, modality and genre in the context of later language acquisition.

  13. Young children's communication and literacy: a qualitative study of language in the inclusive preschool. (United States)

    Kliewer, C


    Interactive and literacy-based language use of young children within the context of an inclusive preschool classroom was explored. An interpretivist framework and qualitative research methods, including participant observation, were used to examine and analyze language in five preschool classes that were composed of children with and without disabilities. Children's language use included spoken, written, signed, and typed. Results showed complex communicative and literacy language use on the part of young children outside conventional adult perspectives. Also, children who used expressive methods other than speech were often left out of the contexts where spoken language was richest and most complex.

  14. [Assessment of pragmatics from verbal spoken data]. (United States)

    Gallardo-Paúls, B


    Pragmatic assessment is usually complex, long and sophisticated, especially for professionals who lack specific linguistic education and interact with impaired speakers. To design a quick method of assessment that will provide a quick general evaluation of the pragmatic effectiveness of neurologically affected speakers. This first filter will allow us to decide whether a detailed analysis of the altered categories should follow. Our starting point was the PerLA (perception, language and aphasia) profile of pragmatic assessment designed for the comprehensive analysis of conversational data in clinical linguistics; this was then converted into a quick questionnaire. A quick protocol of pragmatic assessment is proposed and the results found in a group of children with attention deficit hyperactivity disorder are discussed.

  15. Learning Performance Enhancement Using Computer-Assisted Language Learning by Collaborative Learning Groups

    Directory of Open Access Journals (Sweden)

    Ya-huei Wang


    Full Text Available This study attempted to test whether the use of computer-assisted language learning (CALL and innovative collaborative learning could be more effective than the use of traditional collaborative learning in improving students’ English proficiencies. A true experimental design was used in the study. Four randomly-assigned groups participated in the study: a traditional collaborative learning group (TCLG, 34 students, an innovative collaborative learning group (ICLG, 31 students, a CALL traditional collaborative learning group (CALLTCLG, 32 students, and a CALL innovative collaborative learning group (CALLICLG, 31 students. TOEIC (Test of English for International Communication listening, reading, speaking, and writing pre-test and post-test assessments were given to all students at an interval of sixteen weeks. Multivariate analysis of covariance (MANCOVA, multivariate analysis of variance (MANOVA, and analysis of variance (ANOVA were used to analyze the data. The results revealed that students who used CALL had significantly better learning performance than those who did not. Students in innovative collaborative learning had significantly better learning performances than those in traditional collaborative learning. Additionally, students using CALL innovative collaborative learning had better learning performances than those in CALL collaborative learning, those in innovative collaborative learning, and those in traditional collaborative learning.

  16. Effect of Integrating Computer-Assisted Language Learning in the TOEFL Performance of Nursing Students

    Directory of Open Access Journals (Sweden)

    Marlene R. Castillo


    Full Text Available The study aimed to determine the relationship between the use of a computer-assisted language learning (CALL software and the TOEFL performance of nursing students. This descriptive study used a pre-test and post-test and a survey questionnaire to gather data. Interviews and observations were also conducted to gain further insights in the use of the software. Results indicate that for the three areas of the TOEFL model practice test, there is a statistically significant increase in the gain scores of the students in the post-test after being exposed to the use of the software for two semesters. They performed well in Sentence Completion and Error Identification, both under Structure and Written Expression. On the other hand, they performed only fairly for Reading Comprehension and Vocabulary. For Structure and Written Expression, the areas of strength included correct choice of main verbs, subject pronouns, and comparative use of adjective while the areas of weaknesses include correlative conjunctions, adverb-related structures, and indefinite subject and verb agreement.


    Directory of Open Access Journals (Sweden)

    Ana Carolina Hecht


    Full Text Available Process of language shift is explained by many researchers since linguistic and anthropological perspectives. This area focuses on the correlations between social processes and changes in systems of use of a language. This article aims to address these issues. In particular, we analyze the links between educational-linguistic policy and the maintenance of the languages spoken in Argentina. In doing so, we explore this field taking into account the linguistic and educational policies implemented about indigenous languages in Argentina.

  18. A word by any other intonation: fMRI evidence for implicit memory traces for pitch contours of spoken words in adult brains.

    Directory of Open Access Journals (Sweden)

    Michael Inspector

    Full Text Available OBJECTIVES: Intonation may serve as a cue for facilitated recognition and processing of spoken words and it has been suggested that the pitch contour of spoken words is implicitly remembered. Thus, using the repetition suppression (RS effect of BOLD-fMRI signals, we tested whether the same spoken words are differentially processed in language and auditory brain areas depending on whether or not they retain an arbitrary intonation pattern. EXPERIMENTAL DESIGN: Words were presented repeatedly in three blocks for passive and active listening tasks. There were three prosodic conditions in each of which a different set of words was used and specific task-irrelevant intonation changes were applied: (i All words presented in a set flat monotonous pitch contour (ii Each word had an arbitrary pitch contour that was set throughout the three repetitions. (iii Each word had a different arbitrary pitch contour in each of its repetition. PRINCIPAL FINDINGS: The repeated presentations of words with a set pitch contour, resulted in robust behavioral priming effects as well as in significant RS of the BOLD signals in primary auditory cortex (BA 41, temporal areas (BA 21 22 bilaterally and in Broca's area. However, changing the intonation of the same words on each successive repetition resulted in reduced behavioral priming and the abolition of RS effects. CONCLUSIONS: Intonation patterns are retained in memory even when the intonation is task-irrelevant. Implicit memory traces for the pitch contour of spoken words were reflected in facilitated neuronal processing in auditory and language associated areas. Thus, the results lend support for the notion that prosody and specifically pitch contour is strongly associated with the memory representation of spoken words.

  19. Coaching Parents to Use Naturalistic Language and Communication Strategies (United States)

    Akamoglu, Yusuf; Dinnebeil, Laurie


    Naturalistic language and communication strategies (i.e., naturalistic teaching strategies) refer to practices that are used to promote the child's language and communication skills either through verbal (e.g., spoken words) or nonverbal (e.g., gestures, signs) interactions between an adult (e.g., parent, teacher) and a child. Use of naturalistic…

  20. A study of syllable codas in South African Sign Language

    African Journals Online (AJOL)

    Kate H

    A South African Sign Language Dictionary for Families with Young Deaf Children (SLED 2006) was used with permission ... Figure 1: Syllable structure of a CVC syllable in the word “bed”. In spoken languages .... often than not, there is a societal emphasis on 'fixing' a child's deafness and attempting to teach deaf children to ...

  1. Japanese Non Resident Language Refresher Course; 210 Hour Course. (United States)

    Defense Language Inst., Washington, DC.

    This military intelligence unit refresher course in Japanese is designed for 210 hours of audiolingual instruction. The materials, prepared by the Defense Language Institute, are intended for students with considerable intensive training in spoken and written Japanese who are preparing for a military language assignment. [Not available in hard…

  2. A grammar of Tadaksahak a northern Songhay language of Mali

    NARCIS (Netherlands)

    Christiansen-Bolli, Regula


    This dissertation is a descriptive grammar of the language Tadaksahak spoken by about 30,000 people living in the most eastern part of Mali. The four chapters of the book give 1. Information about the background of the group. 2. The phonological features of the language with the inventory of the

  3. Three languages from America in contact with Spanish

    NARCIS (Netherlands)

    Bakker, D.; Sakel, J.; Stolz, T.


    Long before Europeans reached the American shores for the first time, and forced their cultures upon the indigenous population, including their languages, a great many other languages were spoken on that continent. These dated back to the original discoverers of America, who probably came from the

  4. Social Class and Language Attitudes in Hong Kong (United States)

    Lai, Mee Ling


    This article examines the relation between social class and language attitudes through a triangulated study that analyses the attitudes of 836 secondary school students from different socioeconomic backgrounds toward the 3 official spoken languages used in postcolonial Hong Kong (HK; i.e., Cantonese, English, and Putonghua). The respondents were…

  5. Language studies in higher education in the Netherlands

    NARCIS (Netherlands)

    Zwarts, Frans; Silkens, B.


    Dutch is one of the two official languages of the Netherlands. It is the mother tongue of 15 million Dutchmen and 5,5 million Belgians. The second official language is Frisian, which is spoken by the 500,000 inhabitants of Friesland - a province of the Netherlands, 1,248 square miles in area, in the

  6. Universality versus language-specificity in listening to running speech

    NARCIS (Netherlands)

    Cutler, A.; Demuth, K.; McQueen, J.M.


    Recognizing spoken language involves automatic activation of multiple candidate words. The process of selection between candidates is made more efficient by inhibition of embedded words (like egg in beg) which leave a portion of the input stranded (here, b). Results from European languages suggest

  7. Intonational Division of a Speech Flow in the Kazakh Language (United States)

    Bazarbayeva, Zeynep M.; Zhalalova, Akshay M.; Ormakhanova, Yenlik N.; Ospangaziyeva, Nazgul B.; Karbozova, Bulbul D.


    The purpose of this research is to analyze the speech intonation of the French, Kazakh, English and Russian languages. The study considers intonation component functions (of melodics, duration, and intensity) in poetry and language spoken. It is defined that a set of prosodic means are used in order to convey the intonational specifics of sounding…

  8. A Shared Platform for Studying Second Language Acquisition (United States)

    MacWhinney, Brian


    The study of second language acquisition (SLA) can benefit from the same process of datasharing that has proven effective in areas such as first language acquisition and aphasiology. Researchers can work together to construct a shared platform that combines data from spoken and written corpora, online tutors, and Web-based experimentation. Many of…

  9. Theory of Mind and Language in Children with Cochlear Implants (United States)

    Remmel, Ethan; Peters, Kimberly


    Thirty children with cochlear implants (CI children), age range 3-12 years, and 30 children with normal hearing (NH children), age range 4-6 years, were tested on theory of mind and language measures. The CI children showed little to no delay on either theory of mind, relative to the NH children, or spoken language, relative to hearing norms. The…

  10. The role of foreign and indigenous languages in primary schools ...

    African Journals Online (AJOL)

    This article investigates the use of English and other African languages in Kenyan primary schools. English is a .... For a long time, the issue of the medium of instruction, in especially primary schools, has persisted in spite of .... mother tongue, they use this language for spoken classroom interaction in order to bring about.

  11. Mutual intelligibility between closely related language in Europe.

    NARCIS (Netherlands)

    Gooskens, Charlotte; van Heuven, Vincent; Golubovic, Jelena; Schüppert, Anja; Swarte, Femke; Voigt, Stefanie


    By means of a large-scale web-based investigation, we established the degree of mutual intelligibility of 16 closely related spoken languages within the Germanic, Slavic and Romance language families in Europe. We first present the results of a selection of 1833 listeners representing the mutual

  12. Mutual intelligibility between closely related languages in Europe

    NARCIS (Netherlands)

    Gooskens, C.; Heuven, van V.J.J.P.; Golubović, J.; Schüppert, A.; Swarte, F.; Voigt, S.


    By means of a large-scale web-based investigation, we established the degree of mutual intelligibility of 16 closely related spoken languages within the Germanic, Slavic and Romance language families in Europe. We first present the results of a selection of 1833 listeners representing the mutual

  13. Learning a Minoritized Language in a Majority Language Context: Student Agency and the Creation of Micro-Immersion Contexts (United States)

    DePalma, Renée


    This study investigates the self-reported experiences of students participating in a Galician language and culture course. Galician, a language historically spoken in northwestern Spain, has been losing ground with respect to Spanish, particularly in urban areas and among the younger generations. The research specifically focuses on informal…

  14. Translingualism and Second Language Acquisition: Language Ideologies of Gaelic Medium Education Teachers in a Linguistically Liminal Setting (United States)

    Knipe, John


    Scottish Gaelic, among the nearly 7,000 languages spoken in the world today, is endangered. In the 1980s the Gaelic Medium Education (GME) movement emerged with an emphasis on teaching students all subjects via this ancient tongue with the hope of revitalizing the language. Concomitantly, many linguists have called for problematizing traditional…

  15. Listening in first and second language

    NARCIS (Netherlands)

    Farrell, J.; Cutler, A.; Liontas, J.I.


    Listeners' recognition of spoken language involves complex decoding processes: The continuous speech stream must be segmented into its component words, and words must be recognized despite great variability in their pronunciation (due to talker differences, or to influence of phonetic context, or to

  16. Second Language Learners' Attitudes towards English Varieties (United States)

    Zhang, Weimin; Hu, Guiling


    This pilot project investigates second language (L2) learners' attitudes towards three varieties of English: American (AmE), British (BrE) and Australian (AuE). A 69-word passage spoken by a female speaker of each variety was used. Participants were 30 Chinese students pursuing Masters or Doctoral degrees in the United States, who listened to each…

  17. Bikol Dictionary. PALI Language Texts: Philippines. (United States)

    Mintz, Malcolm W.

    The Bikol language of the Philippines, spoken in the southernmost peninsula of Luzon Island and extending into the island provinces of Catanduanes and Masbate, is presented in this bilingual dictionary. An introduction explains the Bikol alphabet, orthographic representation (including policies adopted in writing Spanish and English loan words),…

  18. Word order in Russian Sign Language

    NARCIS (Netherlands)

    Kimmelman, V.


    The article discusses word order, the syntactic arrangement of words in a sentence, clause, or phrase as one of the most crucial aspects of grammar of any spoken language. It aims to investigate the order of the primary constituents which can either be subject, object, or verb of a simple

  19. Enhancing Language Material Availability Using Computers. (United States)

    Miyashita, Mizuki; Moll, Laura A.

    This paper describes the use of computer technology to produce an updated online Tohono O'odham dictionary. Spoken in southern Arizona and northern Mexico, Tohono O'odham (formerly Papago) and its close relative Akimel O'odham (Pima) had a total of about 25,000 speakers in 1988. Although the language is taught to school children through community…

  20. The English Language of the Nigeria Police (United States)

    Chinwe, Udo Victoria


    In the present day Nigeria, the quality of the English language spoken by Nigerians, is perceived to have been deteriorating and needs urgent attention. The proliferation of books and articles in the recent years can be seen as the native outcrop of its received attention and recognition as a matter of discourse. Evidently, every profession,…