WorldWideScience

Sample records for preparing speech texts

  1. Text To Speech System for Telugu Language

    OpenAIRE

    Siva kumar, M; E. Prakash Babu

    2014-01-01

    Telugu is one of the oldest languages in India. This paper describes the development of Telugu Text-to-Speech System (TTS).In Telugu TTS the input is Telugu text in Unicode. The voices are sampled from real recorded speech. The objective of a text to speech system is to convert an arbitrary text into its corresponding spoken waveform. Speech synthesis is a process of building machinery that can generate human-like speech from any text input to imitate human speakers. Text proc...

  2. Predicting Prosody from Text for Text-to-Speech Synthesis

    CERN Document Server

    Rao, K Sreenivasa

    2012-01-01

    Predicting Prosody from Text for Text-to-Speech Synthesis covers the specific aspects of prosody, mainly focusing on how to predict the prosodic information from linguistic text, and then how to exploit the predicted prosodic knowledge for various speech applications. Author K. Sreenivasa Rao discusses proposed methods along with state-of-the-art techniques for the acquisition and incorporation of prosodic knowledge for developing speech systems. Positional, contextual and phonological features are proposed for representing the linguistic and production constraints of the sound units present in the text. This book is intended for graduate students and researchers working in the area of speech processing.

  3. Text To Speech System for Telugu Language

    Directory of Open Access Journals (Sweden)

    M. Siva Kumar

    2014-03-01

    Full Text Available Telugu is one of the oldest languages in India. This paper describes the development of Telugu Text-to-Speech System (TTS.In Telugu TTS the input is Telugu text in Unicode. The voices are sampled from real recorded speech. The objective of a text to speech system is to convert an arbitrary text into its corresponding spoken waveform. Speech synthesis is a process of building machinery that can generate human-like speech from any text input to imitate human speakers. Text processing and speech generation are two main components of a text to speech system. To build a natural sounding speech synthesis system, it is essential that text processing component produce an appropriate sequence of phonemic units. Generation of sequence of phonetic units for a given standard word is referred to as letter to phoneme rule or text to phoneme rule. The complexity of these rules and their derivation depends upon the nature of the language. The quality of a speech synthesizer is judged by its closeness to the natural human voice and understandability. In this paper we described an approach to build a Telugu TTS system using concatenative synthesis method with syllable as a basic unit of concatenation.

  4. Multilingual Text Analysis for Text-to-Speech Synthesis

    CERN Document Server

    Sproat, R

    1996-01-01

    We present a model of text analysis for text-to-speech (TTS) synthesis based on (weighted) finite-state transducers, which serves as the text-analysis module of the multilingual Bell Labs TTS system. The transducers are constructed using a lexical toolkit that allows declarative descriptions of lexicons, morphological rules, numeral-expansion rules, and phonological rules, inter alia. To date, the model has been applied to eight languages: Spanish, Italian, Romanian, French, German, Russian, Mandarin and Japanese.

  5. Bimodal Emotion Recognition from Speech and Text

    Directory of Open Access Journals (Sweden)

    Weilin Ye

    2014-01-01

    Full Text Available This paper presents an approach to emotion recognition from speech signals and textual content. In the analysis of speech signals, thirty-seven acoustic features are extracted from the speech input. Two different classifiers Support Vector Machines (SVMs and BP neural network are adopted to classify the emotional states. In text analysis, we use the two-step classification method to recognize the emotional states. The final emotional state is determined based on the emotion outputs from the acoustic and textual analyses. In this paper we have two parallel classifiers for acoustic information and two serial classifiers for textual information, and a final decision is made by combing these classifiers in decision level fusion. Experimental results show that the emotion recognition accuracy of the integrated system is better than that of either of the two individual approaches.

  6. Evaluating Text-to-Speech Synthesizers

    Science.gov (United States)

    Cardoso, Walcir; Smith, George; Fuentes, Cesar Garcia

    2015-01-01

    Text-To-Speech (TTS) synthesizers have piqued the interest of researchers for their potential to enhance the L2 acquisition of writing (Kirstein, 2006), vocabulary and reading (Proctor, Dalton, & Grisham, 2007) and pronunciation (Cardoso, Collins, & White, 2012; Soler-Urzua, 2011). Despite their proven effectiveness, there is a need for…

  7. Texting while driving: is speech-based text entry less risky than handheld text entry?

    Science.gov (United States)

    He, J; Chaparro, A; Nguyen, B; Burge, R J; Crandall, J; Chaparro, B; Ni, R; Cao, S

    2014-11-01

    Research indicates that using a cell phone to talk or text while maneuvering a vehicle impairs driving performance. However, few published studies directly compare the distracting effects of texting using a hands-free (i.e., speech-based interface) versus handheld cell phone, which is an important issue for legislation, automotive interface design and driving safety training. This study compared the effect of speech-based versus handheld text entries on simulated driving performance by asking participants to perform a car following task while controlling the duration of a secondary text-entry task. Results showed that both speech-based and handheld text entries impaired driving performance relative to the drive-only condition by causing more variation in speed and lane position. Handheld text entry also increased the brake response time and increased variation in headway distance. Text entry using a speech-based cell phone was less detrimental to driving performance than handheld text entry. Nevertheless, the speech-based text entry task still significantly impaired driving compared to the drive-only condition. These results suggest that speech-based text entry disrupts driving, but reduces the level of performance interference compared to text entry with a handheld device. In addition, the difference in the distraction effect caused by speech-based and handheld text entry is not simply due to the difference in task duration.

  8. Speech Act Classification of German Advertising Texts

    Directory of Open Access Journals (Sweden)

    Артур Нарманович Мамедов

    2015-12-01

    Full Text Available This paper uses the theory of speech acts and the underlying concept of pragmalinguistics to determine the types of speech acts and their classification in the German advertising printed texts. We ascertain that the advertising of cars and accessories, household appliances and computer equipment, watches, fancy goods, food, pharmaceuticals, and financial, insurance, legal services and also airline advertising is dominated by a pragmatic principle, which is based on demonstrating information about the benefits of a product / service. This influences the frequent usage of certain speech acts. The dominant form of exposure is to inform the recipient-user about the characteristics of the advertised product. This information is fore-grounded by means of stylistic and syntactic constructions specific to the advertisement (participial constructions, appositional constructions which contribute to emphasize certain notional components within the framework of the advertising text. Stylistic and syntactic devices of reduction (parceling constructions convey the author's idea. Other means like repetitions, enumerations etc are used by the advertiser to strengthen his selling power. The advertiser focuses the attention of the consumer on the characteristics of the product seeking to convince him of the utility of the product and to influence his/ her buying behavior.

  9. Annotating Speech Corpus for Prosody Modeling in Indian Language Text to Speech Systems

    Directory of Open Access Journals (Sweden)

    Kiruthiga S

    2012-01-01

    Full Text Available A spoken language system, it may either be a speech synthesis or a speech recognition system, starts with building a speech corpora. We give a detailed survey of issues and a methodology that selects the appropriate speech unit in building a speech corpus for Indian language Text to Speech systems. The paper ultimately aims to improve the intelligibility of the synthesized speech in Text to Speech synthesis systems. To begin with, an appropriate text file should be selected for building the speech corpus. Then a corresponding speech file is generated and stored. This speech file is the phonetic representation of the selected text file. The speech file is processed in different levels viz., paragraphs, sentences, phrases, words, syllables and phones. These are called the speech units of the file. Researches have been done taking these units as the basic unit for processing. This paper analyses the researches done using phones, diphones, triphones, syllables and polysyllables as their basic unit for speech synthesis. The paper also provides a recommended set of combinations for polysyllables. Concatenative speech synthesis involves the concatenation of these basic units to synthesize an intelligent, natural sounding speech. The speech units are annotated with relevant prosodic information about each unit, manually or automatically, based on an algorithm. The database consisting of the units along with their annotated information is called as the annotated speech corpus. A Clustering technique is used in the annotated speech corpus that provides way to select the appropriate unit for concatenation, based on the lowest total join cost of the speech unit.

  10. Speect: a multilingual text-to-speech system

    CSIR Research Space (South Africa)

    Louw, JA

    2008-11-01

    Full Text Available This paper introduces a new multilingual text-to-speech system, which we call Speect (Speech synthesis with extensible architecture), aiming to address the shortcomings of using Festival as a research sytem and Flite as a deployment system in a...

  11. Aplikasi Dongeng Berbasis Text to Speech Untuk Platform Android

    Directory of Open Access Journals (Sweden)

    Uswatun Hasanah

    2017-01-01

    Full Text Available AbstrakAnak-anak suka mendengarkan dongeng karena memberikan kesempatan kepada mereka untuk berimajinasi dengan menggambarkan peristiwa pada dongeng ke dalam khayalan. Dikarenakan banyak anak-anak menyukai dongeng, banyak peluang pengembangan aplikasi dongeng seperti menggunakan text to speech. Maka dari itu dilakukan pengembangan aplikasi dongeng menggunakan konsep Text to Speech untuk mengkonversikan text dongeng menjadi suara sebagai media pengganti membaca dongeng. Tujuan penelitian ini adalah untuk membuat aplikasi dongeng text to speech untuk platform android. Metode yang digunakan dalam penelitian ini ialah menggunakan metode waterfall.Aplikasi ini dirancang dengan menggunakan pemograman java dan editor eclipse. Data yang digunakan berupa teks dongeng dan pemanfaatan google text to speech untuk menghasilkan keluaran suara.  Hasil akhir berupa aplikasi pada perangkat android, dan telah dilakukan pengujian. Hasil pengujian menunjukkan Aplikasi dongeng  text to speech sudah mampu membaca dongeng, walaupun dari segi intonasi suara belum alami sebagai mana manusia membaca dongeng.  Pengujian sistem dilakukan bahwa aplikasi ini hanya bisa digunakan pada perangkat android versi Lollipop (5.1.1 untuk menghasilkan keluaran suara bahasa Indonesia secara offline. Dari segi tampilan aplikasi, aplikasi ini belum Fleksibel sehingga hanya terlihat baik pada perangkat tertentu. Kata kunci : Aplikasi, Dongeng, Text to Speech, Android AbstractKids like hearing fable because it gives a chance for them to imagine by drawing an event into imagination. Since many kids like fable, there are many chances to create an application of fable that used text to speech. So from that, fable application text to speech is done to convert fable text into a voice as media of replacement from reading fable. The interest of kids is considered to emerge with this application in order to help parents easily in telling a fable to their kids. Parents are not necessary to

  12. Part-of-speech effects on text-to-speech synthesis

    CSIR Research Space (South Africa)

    Schlunz, GI

    2010-11-01

    Full Text Available One of the goals of text-to-speech (TTS) systems is to produce natural-sounding synthesised speech. Towards this end various natural language processing (NLP) tasks are performed to model the prosodic aspects of the TTS voice. One of the fundamental...

  13. Indian accent text-to-speech system for web browsing

    Indian Academy of Sciences (India)

    Aniruddha Sen; K Samudravijaya

    2002-02-01

    Incorporation of speech and Indian scripts can greatly enhance the accessibility of web information among common people. This paper describes a ‘web reader’ which ‘reads out’ the textual contents of a selected web page in Hindi or in English with Indian accent. The content of the page is downloaded and parsed into suitable textual form. It is then passed on to an indigenously developed text-to-speech system for Hindi/Indian English, to generate spoken output. The text-to-speech conversion is performed in three stages: text analysis, to establish pronunciation, phoneme to acoustic-phonetic parameter conversion and, lastly, parameter-to-speech conversion through a production model. Different types of voices are used to read special messages. The web reader detects the hypertext links in the web pages and gives the user the option to follow the link or continue perusing the current web page. The user can exercise the option either through a keyboard or via spoken commands. Future plans include refining the web parser, improvement of naturalness of synthetic speech and improving the robustness of the speech recognition system.

  14. Indonesian Text-To-Speech System Using Diphone Concatenative Synthesis

    Directory of Open Access Journals (Sweden)

    Sutarman

    2015-02-01

    Full Text Available In this paper, we describe the design and develop a database of Indonesian diphone synthesis using speech segment of recorded voice to be converted from text to speech and save it as audio file like WAV or MP3. In designing and develop a database of Indonesian diphone there are several steps to follow; First, developed Diphone database includes: create a list of sample of words consisting of diphones organized by prioritizing looking diphone located in the middle of a word if not at the beginning or end; recording the samples of words by segmentation. ;create diphones made with a tool Diphone Studio 1.3. Second, develop system using Microsoft Visual Delphi 6.0, includes: the conversion system from the input of numbers, acronyms, words, and sentences into representations diphone. There are two kinds of conversion (process alleged in analyzing the Indonesian text-to-speech system. One is to convert the text to be sounded to phonem and two, to convert the phonem to speech. Method used in this research is called Diphone Concatenative synthesis, in which recorded sound segments are collected. Every segment consists of a diphone (2 phonems. This synthesizer may produce voice with high level of naturalness. The Indonesian Text to Speech system can differentiate special phonemes like in ‘Beda’ and ‘Bedak’ but sample of other spesific words is necessary to put into the system. This Indonesia TTS system can handle texts with abbreviation, there is the facility to add such words.

  15. A text to speech interface for Universal Digital Library

    Institute of Scientific and Technical Information of China (English)

    PRAHALLAD Kishore; BLACK Alan

    2005-01-01

    The objective of Universal Digital Library (UDL) is to capture all books in digital format. A text to speech (TTS)interface for UDL portal would enable access to the digital content in voice mode, and also provide access to the digital content for illiterate and vision-impaired people. Our work focuses on design and implementation of text to speech interface for UDL portal primarily for Indian languages. This paper is aimed at identifying the issues involved in integrating text to speech system into UDL portal and describes the development process of Hindi, Telugu and Tamil voices under Festvox framework using unit selection techniques. We demonstrate the quality of the Tamil and Telugu voices and lay out the plan for integrating the TTS into the UDL portal.

  16. Rule-Based Storytelling Text-to-Speech (TTS Synthesis

    Directory of Open Access Journals (Sweden)

    Ramli Izzad

    2016-01-01

    Full Text Available In recent years, various real life applications such as talking books, gadgets and humanoid robots have drawn the attention to pursue research in the area of expressive speech synthesis. Speech synthesis is widely used in various applications. However, there is a growing need for an expressive speech synthesis especially for communication and robotic. In this paper, global and local rule are developed to convert neutral to storytelling style speech for the Malay language. In order to generate rules, modification of prosodic parameters such as pitch, intensity, duration, tempo and pauses are considered. Modification of prosodic parameters is examined by performing prosodic analysis on a story collected from an experienced female and male storyteller. The global and local rule is applied in sentence level and synthesized using HNM. Subjective tests are conducted to evaluate the synthesized storytelling speech quality of both rules based on naturalness, intelligibility, and similarity to the original storytelling speech. The results showed that global rule give a better result than local rule

  17. Part-of-speech tagging of Modern Hebrew text

    NARCIS (Netherlands)

    Bar-Haim, R.; Sima'an, K.; Winter, Y.

    2008-01-01

    Words in Semitic texts often consist of a concatenation of word segments, each corresponding to a part-of-speech (POS) category. Semitic words may be ambiguous with regard to their segmentation as well as to the POS tags assigned to each segment. When designing POS taggers for Semitic languages, a m

  18. Developing a Child Friendly Text-to-Speech System

    Directory of Open Access Journals (Sweden)

    Agnes Jacob

    2008-01-01

    Full Text Available This paper discusses the implementation details of a child friendly, good quality, English text-to-speech (TTS system that is phoneme-based, concatenative, easy to set up and use with little memory. Direct waveform concatenation and linear prediction coding (LPC are used. Most existing TTS systems are unit-selection based, which use standard speech databases available in neutral adult voices. Here reduced memory is achieved by the concatenation of phonemes and by replacing phonetic wave files with their LPC coefficients. Linguistic analysis was used to reduce the algorithmic complexity instead of signal processing techniques. Sufficient degree of customization and generalization catering to the needs of the child user had been included through the provision for vocabulary and voice selection to suit the requisites of the child. Prosody had also been incorporated. This inexpensive TTS system was implemented in MATLAB, with the synthesis presented by means of a graphical user interface (GUI, thus making it child friendly. This can be used not only as an interesting language learning aid for the normal child but it also serves as a speech aid to the vocally disabled child. The quality of the synthesized speech was evaluated using the mean opinion score (MOS.

  19. Speech-to-text: the next revelation for recording data.

    Science.gov (United States)

    Sutton, J

    1997-01-01

    There are numerous tools available for communicating and storing information. Administrators may want to consider speech-to-text-technology. Programs are available that are accurate, easy to use and easily integrated into existing hospital information systems. Computerized voice recognition systems that provide direct speech-to-text are available today and are often referred to as the "listening typewriter." You speak, the computer listens and then types out what it thinks it heard. Such a system has a vocabulary or collection of words that may be spoken by a user. The program analyzes the phonetic structure of words in the vocabulary. A database table compares sounds made by the voice to words in the vocabulary. Two types of recognition are available--realtime and deferred server-based. With realtime transcription, dictation appears on the workstation screen in front of the dictating radiologist. With the deferred system, the dictated session goes to a queue on a server and waits there to be processed, on operation often known as batch processing. Failure of the system can be prevented by clustering or having a redundant server that takes over for the primary server if it should fail. Options to consider for a speech transcription system include security for electronic signatures and a provision for quality assurance or editing professionals who will review documents for errors in recognition. Some software systems offer a program that examines the final document for new or misspelled words, while other systems require error correction to be completed within the last 10 dictated words. The distributed voice model contains a profile of users' pronunciations and allows the computer to recognize a voice that, associated with login ID, differentiates it from others. This model allows a physician to dictate from different locations. Although the cost of the hardware is impressive and requires a thorough cost benefit analysis, such a system can be an excellent solution

  20. Implementation of Text To Speech for Marathi Language Using Transcriptions Concept

    Directory of Open Access Journals (Sweden)

    Sangramsing N. Kayte

    2015-11-01

    Full Text Available This research paper presents the approach towards converting text to speech using new methodology. The text to speech conversion system enables user to enter text in Marathi and as output it gets sound. The paper presents the steps followed for converting text to speech for Marathi language and the algorithm used for it. The focus of this paper is based on the tokenisation process and the orthographic representation of the text that shows the mapping of letter to sound using the description of language’s phonetics. Here the main focus is on the text to IPA transcription concept. It is in fact, a system that translates text to IPA transcription which is the primary stage for text to speech conversion. The whole procedure for converting text to speech involves a great deal of time as it’s not an easy task and requires efforts.

  1. Speech-Language Pathology: Preparing Early Interventionists

    Science.gov (United States)

    Prelock, Patricia A.; Deppe, Janet

    2015-01-01

    The purpose of this article is to explain the role of speech-language pathology in early intervention. The expected credentials of professionals in the field are described, and the current numbers of practitioners serving young children are identified. Several resource documents available from the American Speech-­Language Hearing Association are…

  2. Auditory Support in Linguistically Diverse Classrooms: Factors Related to Bilingual Text-to-Speech Use

    Science.gov (United States)

    Van Laere, E.; Braak, J.

    2017-01-01

    Text-to-speech technology can act as an important support tool in computer-based learning environments (CBLEs) as it provides auditory input, next to on-screen text. Particularly for students who use a language at home other than the language of instruction (LOI) applied at school, text-to-speech can be useful. The CBLE E-Validiv offers content in…

  3. Auditory Support in Linguistically Diverse Classrooms: Factors Related to Bilingual Text-to-Speech Use

    Science.gov (United States)

    Van Laere, E.; Braak, J.

    2017-01-01

    Text-to-speech technology can act as an important support tool in computer-based learning environments (CBLEs) as it provides auditory input, next to on-screen text. Particularly for students who use a language at home other than the language of instruction (LOI) applied at school, text-to-speech can be useful. The CBLE E-Validiv offers content in…

  4. Automated Gesturing for Virtual Characters: Speech-driven and Text-driven Approaches

    Directory of Open Access Journals (Sweden)

    Goranka Zoric

    2006-04-01

    Full Text Available We present two methods for automatic facial gesturing of graphically embodied animated agents. In one case, conversational agent is driven by speech in automatic Lip Sync process. By analyzing speech input, lip movements are determined from the speech signal. Another method provides virtual speaker capable of reading plain English text and rendering it in a form of speech accompanied by the appropriate facial gestures. Proposed statistical model for generating virtual speaker’s facial gestures can be also applied as addition to lip synchronization process in order to obtain speech driven facial gesturing. In this case statistical model will be triggered with the input speech prosody instead of lexical analysis of the input text.

  5. A Spoken Access Approach for Chinese Text and Speech Information Retrieval.

    Science.gov (United States)

    Chien, Lee-Feng; Wang, Hsin-Min; Bai, Bo-Ren; Lin, Sun-Chein

    2000-01-01

    Presents an efficient spoken-access approach for both Chinese text and Mandarin speech information retrieval. Highlights include human-computer interaction via voice input, speech query recognition at the syllable level, automatic term suggestion, relevance feedback techniques, and experiments that show an improvement in the effectiveness of…

  6. An Arabic Text-To-Speech System Based on Artificial Neural Networks

    Directory of Open Access Journals (Sweden)

    Ghadeer Al-Said

    2009-01-01

    Full Text Available Problem statement: With the rapid advancement in information technology and communications, computer systems increasingly offer the users the opportunity to interact with information through speech. The interest in speech synthesis and in building voices is increasing. Worldwide, speech synthesizers have been developed for many popular languages English, Spanish and French and many researches and developments have been applied to those languages. Arabic on the other hand, has been given little attention compared to other languages of similar importance and the research in Arabic is still in its infancy. Based on these ideas, we introduced a system to transform Arabic text that was retrieved from a search engine into spoken words. Approach: We designed a text-to-speech system in which we used concatenative speech synthesis approach to synthesize Arabic text. The synthesizer was based on artificial neural networks, specifically the unsupervised learning paradigm. Different sizes of speech units had been used to produce spoken utterances, which are words, diphones and triphones. We also built a dictionary of 500 common words of Arabic. The smaller speech units (diphones and triphones used for synthesis were chosen to achieve unlimited vocabulary of speech, while the word units were used for synthesizing limited set of sentences. Results: The system showed very high accuracy in synthesizing the Arabic text and the output speech was highly intelligible. For the word and diphone unit experiments, we could reach an accuracy of 99% while for the triphone units we reached an accuracy of 86.5%. Conclusion: An Arabic text-to-speech synthesizer was built with the ability to produce unlimited number of words with high quality voice.

  7. Speech-To-Text Conversion STT System Using Hidden Markov Model HMM

    Directory of Open Access Journals (Sweden)

    Su Myat Mon

    2015-06-01

    Full Text Available Abstract Speech is an easiest way to communicate with each other. Speech processing is widely used in many applications like security devices household appliances cellular phones ATM machines and computers. The human computer interface has been developed to communicate or interact conveniently for one who is suffering from some kind of disabilities. Speech-to-Text Conversion STT systems have a lot of benefits for the deaf or dumb people and find their applications in our daily lives. In the same way the aim of the system is to convert the input speech signals into the text output for the deaf or dumb students in the educational fields. This paper presents an approach to extract features by using Mel Frequency Cepstral Coefficients MFCC from the speech signals of isolated spoken words. And Hidden Markov Model HMM method is applied to train and test the audio files to get the recognized spoken word. The speech database is created by using MATLAB.Then the original speech signals are preprocessed and these speech samples are extracted to the feature vectors which are used as the observation sequences of the Hidden Markov Model HMM recognizer. The feature vectors are analyzed in the HMM depending on the number of states.

  8. The principles of designing of algorithm for speech synthesis from texts written in Albanian language

    Directory of Open Access Journals (Sweden)

    Agni Dika

    2012-05-01

    Full Text Available The speech synthesis is artificial generation of human speech from written texts. For this purpose, adequate algorithms are designed, which then through relevant programs make it possible to synthesize texts to speech. The process of converting text into speech is also known as Text-To-Speech (TTS system [5]. In this paper are given basic principles to be used when designing a system to synthesize speech in Albanian language from written texts. Currently there are solutions that enable natural speech generation for various world languages. However, unfortunately these are not universal solutions to be used for other languages too, because the volume generated for other languages is incomprehensible and unnatural. For this reason, for every language one should seek solutions that address the specifics of it, always with the aim of generating voice to suit the nature of language. Generating systems that are currently used mainly rely on the use of the concatenation method [6], during which acoustic segments of text files are joined, which are previously digitized and stored as such in a database. For Albanian language, we consider that on the textual part of the database, as basic segments to be used are: the most frequent words, two-letters and letters [4]. However, in a particular part of the database are included various abbreviations, i.e. textual equivalents and their acoustics files, to be used also during the generation of appropriate speech. Whereas, with the aim of synthesizing the various numerical values written in the decimal system, in database were added values, respectively their corresponding sound files, whereby speech is generated for different numbers. The first part of the paper is a brief presentation of the Albanian language [1], respectively of the alphabet used in writing the language and its most frequent words.

  9. The effects of speech motor preparation on auditory perception

    Science.gov (United States)

    Myers, John

    Perception and action are coupled via bidirectional relationships between sensory and motor systems. Motor systems influence sensory areas by imparting a feedforward influence on sensory processing termed "motor efference copy" (MEC). MEC is suggested to occur in humans because speech preparation and production modulate neural measures of auditory cortical activity. However, it is not known if MEC can affect auditory perception. We tested the hypothesis that during speech preparation auditory thresholds will increase relative to a control condition, and that the increase would be most evident for frequencies that match the upcoming vocal response. Participants performed trials in a speech condition that contained a visual cue indicating a vocal response to prepare (one of two frequencies), followed by a go signal to speak. To determine threshold shifts, voice-matched or -mismatched pure tones were presented at one of three time points between the cue and target. The control condition was the same except the visual cues did not specify a response and subjects did not speak. For each participant, we measured f0 thresholds in isolation from the task in order to establish baselines. Results indicated that auditory thresholds were highest during speech preparation, relative to baselines and a non-speech control condition, especially at suprathreshold levels. Thresholds for tones that matched the frequency of planned responses gradually increased over time, but sharply declined for the mismatched tones shortly before targets. Findings support the hypothesis that MEC influences auditory perception by modulating thresholds during speech preparation, with some specificity relative to the planned response. The threshold increase in tasks vs. baseline may reflect attentional demands of the tasks.

  10. Text as a Supplement to Speech in Young and Older Adults a)

    Science.gov (United States)

    Krull, Vidya; Humes, Larry E.

    2015-01-01

    Objective The purpose of this experiment was to quantify the contribution of visual text to auditory speech recognition in background noise. Specifically, we tested the hypothesis that partially accurate visual text from an automatic speech recognizer could be used successfully to supplement speech understanding in difficult listening conditions in older adults, with normal or impaired hearing. Our working hypotheses were based on what is known regarding audiovisual speech perception in the elderly from speechreading literature. We hypothesized that: 1) combining auditory and visual text information will result in improved recognition accuracy compared to auditory or visual text information alone; 2) benefit from supplementing speech with visual text (auditory and visual enhancement) in young adults will be greater than that in older adults; and 3) individual differences in performance on perceptual measures would be associated with cognitive abilities. Design Fifteen young adults with normal hearing, fifteen older adults with normal hearing, and fifteen older adults with hearing loss participated in this study. All participants completed sentence recognition tasks in auditory-only, text-only, and combined auditory-text conditions. The auditory sentence stimuli were spectrally shaped to restore audibility for the older participants with impaired hearing. All participants also completed various cognitive measures, including measures of working memory, processing speed, verbal comprehension, perceptual and cognitive speed, processing efficiency, inhibition, and the ability to form wholes from parts. Group effects were examined for each of the perceptual and cognitive measures. Audiovisual benefit was calculated relative to performance on auditory-only and visual-text only conditions. Finally, the relationship between perceptual measures and other independent measures were examined using principal-component factor analyses, followed by regression analyses. Results

  11. The peculiarity of speech influense of the advertising texts

    Directory of Open Access Journals (Sweden)

    А.S. Тelеtov

    2015-12-01

    Full Text Available The aim of the article. The article shows that although over time advertising texts have an increasing influence on readers, listeners, viewers the responsibility of advertisers is not increased. The results of the analysis. Today advertising texts very often violate moral and ethical standards that directly conflict with the provisions of the Law of Ukraine «On Advertising». Linguistic manipulation of readers and viewers of advertising carried out both in official mass outdoor advertising, placed on special temporary and stationary designs, promotional materials, located in open areas, the outside of buildings, structures, elements of street equipment on the roadway of streets and roads, and directly on asphalt without any prior authorization. Moreover advertisers and advertising producers use so-called shocking advertising. Advertisers often «forget» that their advertising appeals have significant side effects. In the theory of semantic advertising exposure the concept of linguistic manipulatioin is given the first place. It involves appearance in the recipient unconscious incentive to modify his / her behavior which can be essential for advertiser. The essence of linguistic manipulation presents advertising information so that the consumers think that they come to some knowledge, conclusions, decisions on their own, so the attitude to such information is less critical. The basic techniques of linguistic manipulation in advertising are analyzed and classified (language game, the use of precedent texts (intertextuality, intriguing or provocative texts, incorrect and stupid remarks like, etc.. Conclusions and directions of further researches. The proposed synthesis and continuous regulation of advertising legislation can significantly reduce the negative impact of natural components of the existing advertising to the general public. It is clear that legal rules must encourage advertisers to create advertising product according to the

  12. Implementation of Phonetic Context Variable Length Unit Selection Module for Malay Text to Speech

    Directory of Open Access Journals (Sweden)

    Tian-Swee Tan

    2008-01-01

    Full Text Available Problem statement: The main problem with current Malay Text-To-Speech (MTTS synthesis system is the poor quality of the generated speech sound due to the inability of traditional TTS system to provide multiple choices of unit for generating more accurate synthesized speech. Approach: This study proposes a phonetic context variable length unit selection MTTS system that is capable of providing more natural and accurate unit selection for synthesized speech. It implemented a phonetic context algorithm for unit selection for MTTS. The unit selection method (without phonetic context may encounter the problem of selecting the speech unit from different sources and affect the quality of concatenation. This study proposes the design of speech corpus and unit selection method according to phonetic context so that it can select a string of continuous phoneme from same source instead of individual phoneme from different sources. This can further reduce the concatenation point and increase the quality of concatenation. The speech corpus was transcribed according to phonetic context to preserve the phonetic information. This method utilizes word base concatenation method. Firstly it will search through the speech corpus for the target word, if the target is found; it will be used for concatenation. If the word does not exist, then it will construct the words from phoneme sequence. Results: This system had been tested with 40 participants in Mean Opinion Score (MOS listening test with the average rates for naturalness, pronunciation and intelligibility are 3.9, 4.1 and 3.9. Conclusion/Recommendation: Through this study, a very first version of Corpus-based MTTS has been designed; it has improved the naturalness, pronunciation and intelligibility of synthetic speech. But it still has some lacking that need to be perfected such as the prosody module to support the phrasing analysis and intonation of input text to match with the waveform modifier.

  13. Web Voice Browser Based on an ISLPC Text-to-Speech Algorithm

    Institute of Scientific and Technical Information of China (English)

    LIAO Rikun; JI Yuefeng; LI Hui

    2006-01-01

    A kind of Web voice browser based on improved synchronous linear predictive coding (ISLPC) and Text-to-Speech (TTS) algorithm and Internet application was proposed. The paper analyzes the features of TTS system with ISLPC speech synthesis and discusses the design and implementation of ISLPC TTS-based Web voice browser. The browser integrates Web technology, Chinese information processing, artificial intelligence and the key technology of Chinese ISLPC speech synthesis. It's a visual and audible web browser that can improve information precision for network users. The evaluation results show that ISLPC-based TTS model has a better performance than other browsers in voice quality and capability of identifying Chinese characters.

  14. Use Pronunciation by Analogy for text to speech system in Persian language

    CERN Document Server

    Jowharpour, Ali; Yektaee, Mohammad hosein

    2011-01-01

    The interest in text to speech synthesis increased in the world .text to speech have been developed formany popular languages such as English, Spanish and French and many researches and developmentshave been applied to those languages. Persian on the other hand, has been given little attentioncompared to other languages of similar importance and the research in Persian is still in its infancy.Persian language possess many difficulty and exceptions that increase complexity of text to speechsystems. For example: short vowels is absent in written text or existence of homograph words. in thispaper we propose a new method for persian text to phonetic that base on pronunciations by analogy inwords, semantic relations and grammatical rules for finding proper phonetic. Keywords:PbA, text to speech, Persian language, FPbA

  15. Influence of GSM speech coding algorithms on the performance of text-independent speaker indentification

    OpenAIRE

    Grassi, Sara; Besacier, Laurent; DUFAUX, Alain; Ansorge, Michael; Pellandini, Fausto

    2006-01-01

    This paper investigates the influence, on theperformance of a text-independent speaker identification system, of the three speech coding algorithmsstandardized for use in the GSM wireless communication network. The speaker identification system isbased on Gaussian Mixture Models (GMM) classifiers. Only the influence of the speech codingalgorithms was taken into account. This was done bypassing the whole TIMIT database through eachcoding/decoding algorithm obtaining three transcodeddatabases. ...

  16. The role of speech prosody and text reading prosody in children's reading comprehension

    NARCIS (Netherlands)

    Veenendaal, N.J.; Groen, M.A.; Verhoeven, L.T.W.

    2014-01-01

    Text reading prosody has been associated with reading comprehension. However, text reading prosody is a reading-dependent measure that relies heavily on decoding skills. Investigation of the contribution of speech prosody - which is independent from reading skills - in addition to text reading proso

  17. The Role of Speech Prosody and Text Reading Prosody in Children's Reading Comprehension

    Science.gov (United States)

    Veenendaal, Nathalie J.; Groen, Margriet A.; Verhoeven, Ludo

    2014-01-01

    Background: Text reading prosody has been associated with reading comprehension. However, text reading prosody is a reading-dependent measure that relies heavily on decoding skills. Investigation of the contribution of speech prosody--which is independent from reading skills--in addition to text reading prosody, to reading comprehension could…

  18. The Role of Speech Prosody and Text Reading Prosody in Children's Reading Comprehension

    Science.gov (United States)

    Veenendaal, Nathalie J.; Groen, Margriet A.; Verhoeven, Ludo

    2014-01-01

    Background: Text reading prosody has been associated with reading comprehension. However, text reading prosody is a reading-dependent measure that relies heavily on decoding skills. Investigation of the contribution of speech prosody--which is independent from reading skills--in addition to text reading prosody, to reading comprehension could…

  19. A discourse model of affect for text-to-speech synthesis

    CSIR Research Space (South Africa)

    Schlunz, GI

    2013-12-01

    Full Text Available This paper introduces a model of affect to improve prosody in text-to-speech synthesis. It operates on the discourse level of text to predict the underlying linguistic factors that contribute towards emotional appraisal, rather than any particular...

  20. On advantage of seeing text and hearing speech

    Directory of Open Access Journals (Sweden)

    Živanović Jelena

    2011-01-01

    Full Text Available The aim of this study was to examine the effect of congruence between the sensory modality through which a concept can be experienced and the modality through which the word denoting that concept is perceived during word recognition. Words denoting concepts that can be experienced visually (e.g. “color” and words denoting concepts that can be experienced auditorily (e.g. “noise” were presented both visually and auditorily. We observed shorter processing latencies when there was a match between the modality through which a concept could be experienced and the modality through which a word denoting that concept was presented. In visual lexical decision task, “color” was recognized faster than “noise”, whereas in auditory lexical decision task, “noise” was recognized faster than “color”. The obtained pattern of results can not be accounted for by exclusive amodal theories, whereas it can be easily integrated in theories based on perceptual representations.

  1. Part-of-speech tagging and detection of social media texts

    OpenAIRE

    Neunerdt, Melanie

    2016-01-01

    This thesis contributes to sequence labeling tasks in the field of Natural Language Processing by introducing novel concepts, models and algorithms for Part-of-Speech (POS) tagging, social media text detection and Web page cleaning. First, the task of social media text classification in Web pages is addressed, where sequences of Web text segments are classified based on a high-dimensional feature vector. New features motivated by social media text characteristics are introduced and investigat...

  2. Orthographic Structuring of Human Speech and Texts Linguistic Application of Recurrence Quantification Analysis

    CERN Document Server

    Orsucci, F; Giuliani, A; Webber, C L; Zbilut, J P

    1997-01-01

    A methodology based upon recurrence quantification analysis is proposed for the study of orthographic structure of written texts. Five different orthographic data sets (20th century Italian poems, 20th century American poems, contemporary Swedish poems with their corresponding Italian translations, Italian speech samples, and American speech samples) were subjected to recurrence quantification analysis, a procedure which has been found to be diagnostically useful in the quantitative assessment of ordered series in fields such as physics, molecular dynamics, physiology, and general signal processing. Recurrence quantification was developed from recurrence plots as applied to the analysis of nonlinear, complex systems in the physical sciences, and is based on the computation of a distance matrix of the elements of an ordered series (in this case the letters consituting selected speech and poetic texts). From a strictly mathematical view, the results show the possibility of demonstrating invariance between diffe...

  3. Text Independent Speaker Recognition and Speaker Independent Speech Recognition Using Iterative Clustering Approach

    Directory of Open Access Journals (Sweden)

    A.Revathi

    2009-11-01

    Full Text Available This paper presents the effectiveness of perceptual features and iterative clustering approach forperforming both speech and speaker recognition. Procedure used for formation of training speech is differentfor developing training models for speaker independent speech and text independent speaker recognition. So,this work mainly emphasizes the utilization of clustering models developed for the training data to obtainbetter accuracy as 91%, 91% and 99.5% for mel frequency perceptual linear predictive cepstrum with respectto three categories such as speaker identification, isolated digit recognition and continuous speechrecognition. This feature also produces 9% as low equal error rate which is used as a performance measurefor speaker verification. The work is experimentally evaluated on the set of isolated digits and continuousspeeches from TI digits_1 and TI digits_2 database for speech recognition and on speeches of 50 speakersrandomly chosen from TIMIT database for speaker recognition. The noteworthy feature of speakerrecognition algorithm is to evaluate the testing procedure on identical messages of all the 50 speakers,theoretical validation of results using F-ratio and validation of results by statistical analysis using2 cdistribution.

  4. THE UNDERLYING PRINCIPLES OF SUSILO BAMBANG YUDHOYONO‘S THOUGHT PATTERNS IN HIS ENGLISH SPEECH TEXTS

    Directory of Open Access Journals (Sweden)

    Sulistya ningsih

    2014-10-01

    Full Text Available The underlying principles of thought patterns as shown in SBY's English Speeches Texts are made because there are different responses from the public, a part of public praise that SBY is a good president, and others claim and criticize him that  he is slow (Djalal, 2007: forward page. This title so far has not been investigated. This research was aimed at finding out:  the underlying principles of SBY’s thought patterns in his English Speech Texts related to Javanese philosophy. This research is qualitative. The data selected from SBY’s speech Texts were analyzed using semantic and pragmastylistic theory then were related to Javanese philosophy. The findings are the underlying principles of SBY’s thought patterns based on Javanese philosophy manifested in his English Speech Texts are: first is Memayu Hayuning Bawana, Ambrasta dur Hangkara means to reach safety, peace, happiness and well-being of the world and its contents, to keep the world maintained and harmony. Second, Rukun agawe santosa crah agawe bubrah  means to build the condition of harmony, and avoid conflict, because conflict can be harmful to both parties. Third, tepa selira means keep thinking not to offend others or lighten the burdens of others, tolerance. Fourth is ana rembug becik dirembug means thru negotiations can avoid conflict and achieve cooperation, safety, peace and prosperity. In sum, the world peace can be reached thru discussions without war, soft powers.

  5. BILINGUAL MULTIMODAL SYSTEM FOR TEXT-TO-AUDIOVISUAL SPEECH AND SIGN LANGUAGE SYNTHESIS

    Directory of Open Access Journals (Sweden)

    A. A. Karpov

    2014-09-01

    Full Text Available We present a conceptual model, architecture and software of a multimodal system for audio-visual speech and sign language synthesis by the input text. The main components of the developed multimodal synthesis system (signing avatar are: automatic text processor for input text analysis; simulation 3D model of human's head; computer text-to-speech synthesizer; a system for audio-visual speech synthesis; simulation 3D model of human’s hands and upper body; multimodal user interface integrating all the components for generation of audio, visual and signed speech. The proposed system performs automatic translation of input textual information into speech (audio information and gestures (video information, information fusion and its output in the form of multimedia information. A user can input any grammatically correct text in Russian or Czech languages to the system; it is analyzed by the text processor to detect sentences, words and characters. Then this textual information is converted into symbols of the sign language notation. We apply international «Hamburg Notation System» - HamNoSys, which describes the main differential features of each manual sign: hand shape, hand orientation, place and type of movement. On their basis the 3D signing avatar displays the elements of the sign language. The virtual 3D model of human’s head and upper body has been created using VRML virtual reality modeling language, and it is controlled by the software based on OpenGL graphical library. The developed multimodal synthesis system is a universal one since it is oriented for both regular users and disabled people (in particular, for the hard-of-hearing and visually impaired, and it serves for multimedia output (by audio and visual modalities of input textual information.

  6. The derivation of prosody for text-to-speech from prosodic sentence structure

    NARCIS (Netherlands)

    Quené, H.; Kager, R.W.J.

    1992-01-01

    Suprasegmental phenomena in synthetic speech should reflect the linguistic structure of the input text. An algorithm is described, which establishes the prosodic sentence structure (PSS). This can be achieved without exhaustive syntactic parsing, using a dictionary of 550 function words. Subsequentl

  7. An Evaluation of Text-to-Speech Synthesizers in the Foreign Language Classroom: Learners' Perceptions

    Science.gov (United States)

    Bione, Tiago; Grimshaw, Jennica; Cardoso, Walcir

    2016-01-01

    As stated in Cardoso, Smith, and Garcia Fuentes (2015), second language researchers and practitioners have explored the pedagogical capabilities of Text-To-Speech synthesizers (TTS) for their potential to enhance the acquisition of writing (e.g. Kirstein, 2006), vocabulary and reading (e.g. Proctor, Dalton, & Grisham, 2007), and pronunciation…

  8. The Effects of Word Prediction and Text-to-Speech on the Writing Process of Translating

    Science.gov (United States)

    Cunningham, Robert

    2013-01-01

    The purpose of this study was to determine the effects of the combination of word prediction and text-to-speech software on the writing process of translating. Participants for this study included 10 elementary and middle school students who had a diagnosis of disorder of written expression. A modified multiple case series was used to collect data…

  9. Review of Speech-to-Text Recognition Technology for Enhancing Learning

    Science.gov (United States)

    Shadiev, Rustam; Hwang, Wu-Yuin; Chen, Nian-Shing; Huang, Yueh-Min

    2014-01-01

    This paper reviewed literature from 1999 to 2014 inclusively on how Speech-to-Text Recognition (STR) technology has been applied to enhance learning. The first aim of this review is to understand how STR technology has been used to support learning over the past fifteen years, and the second is to analyze all research evidence to understand how…

  10. A Study of Text-to-Speech (TTS) in Children's English Learning

    Science.gov (United States)

    Huang, Yi-Ching; Liao, Lung-Chuan

    2015-01-01

    The purpose of this study was to explore the effects of the digital material incorporated into Text-to- Speech system for students' English spelling. The digital material was made on the basis of the Spelling Bee vocabulary list (approximately 300 words) issued by the selected school. 21 third graders from a private bilingual school in Taiwan were…

  11. Review of Speech-to-Text Recognition Technology for Enhancing Learning

    Science.gov (United States)

    Shadiev, Rustam; Hwang, Wu-Yuin; Chen, Nian-Shing; Huang, Yueh-Min

    2014-01-01

    This paper reviewed literature from 1999 to 2014 inclusively on how Speech-to-Text Recognition (STR) technology has been applied to enhance learning. The first aim of this review is to understand how STR technology has been used to support learning over the past fifteen years, and the second is to analyze all research evidence to understand how…

  12. Using Text-to-Speech Reading Support for an Adult with Mild Aphasia and Cognitive Impairment

    Science.gov (United States)

    Harvey, Judy; Hux, Karen; Snell, Jeffry

    2013-01-01

    This single case study served to examine text-to-speech (TTS) effects on reading rate and comprehension in an individual with mild aphasia and cognitive impairment. Findings showed faster reading, given TTS presented at a normal speaking rate, but no significant comprehension changes. TTS may support reading in people with aphasia when time…

  13. A Unified Framework for Multilingual Text-to-Speech Synthesis with SSML Specification as Interface

    Institute of Scientific and Technical Information of China (English)

    WU Zhiyong; CAO Guangqi; MENG M. Helen; CAI Lianhong

    2009-01-01

    This paper describes the design of a unified framework for a multilingual text-to-speech (TTS) synthesis engine -Crystal. The unified framework defines the common TTS modules for different languages and/or dialects. The interfaces between consecutive modules conform to the speech synthesis markup lan-guage (SSML) specification for standardization, interoperability, mulUlinguality, and extensibility. Detailed module divisions and implementation technologies for the unified framework are introduced, together with possible extensions for the algorithm research and evaluation of the TTS synthesis. Implementation of a mixed-language TTS system for Chinese Putonghua, Chinese Cantonese, and English demonstrates the feasibility of the proposed unified framework.

  14. DIFFICULTIES EMERGING IN THE PROCESS OF TEACHING RUSSIAN STUDENTS TO MAKE A SPEECH IN JAPANESE AT THE STAGE OF THE SPEECH TEXT PRELIMINARY DEVELOPMENT

    Directory of Open Access Journals (Sweden)

    N. L. Maksimenko

    2014-09-01

    Full Text Available The article gives a detailed description of difficulties that emerge in the process of teaching Russian students to make a speech in Japanese at the stage of the speech text preliminary development. The conducted research shows that in the process of acquiring speech making skills students have considerable difficulties and make a number of various mistakes at different stages (phases of oral speech production. We suppose that for developing an efficient model of teaching students to make a speech in Japanese it is necessary to introduce corrections into the speech production model which are directly connected with the peculiarities of making an oral presentation in Japanese at the initial stage of education.

  15. Rule-based Prosody Calculation for Marathi Text-to-Speech Synthesis

    Directory of Open Access Journals (Sweden)

    Sangramsing N. Kayte

    2015-11-01

    Full Text Available This research paper presents two empirical studies that examine the influence of different linguistic aspects on prosody in Marathi. First, we analyzed a Marathi corpus with respect to the effect of syntax and information status on prosody. Second, we conducted a listening test which investigated the prosodic realisation of constituents in the Marathi depending on their information status. The results were used to improve the prosody prediction in the Marathi text-to-speech synthesis system MARY.

  16. SOCIOLINGUISTIC FACTORS OF THE WRITTEN SPEECH NORMS APPROXIMATION IN LABOR MIGRANTS’ TEXTS

    Directory of Open Access Journals (Sweden)

    Utesheva Altynay Pazylovna

    2015-06-01

    Full Text Available The article focuses on the features of written Russian speech of labor migrants from different countries considering the norms of written speech. The empirical basis of the research is represented by the handwritten CVs of unemployed migrants from Vietnam and Uzbekistan, that were presented to the departments of the Federal Migration Service of the Russian Federation in the city of Volgograd. Written speech violations are classified according to the age groups which migrants belong to. The following sociolinguistic characteristics of the migrants are also taken into account: nationality, period of school education, higher education, document writing competence. Group 1 combined informants aged from 20 to 30, without higher education, who studied the Russian language at school in the new period of the collapse of the Soviet Union procedures or on their own. It is an educational institution with no experience compiling official documents and communication skills in Russian. Group 2 combined informants aged from 30 to 50, without higher education, who studied Russian at school by Soviet methods with experience of drawing up official documents and possessing basic communication skills to communicate in Russian. Group 3 combined informants aged 50 and older with secondary special education, who studied Russian at school by Soviet methods and actively developed communicative competence at the expense of everyday communication, reading books, listening to the radio and watching programs in Russian, with experience in drafting official documents. The features of migrants' written speech are manifested in specific language and speech mistakes, particularly in graphic, phonetic and genre rules violations. The general patterns of mistakes are registered. The mistakes are caused not only by language transfer and the Russian language competence, but also by sociolinguistic factors. The particular cross-language differences of migrants writing are

  17. The benefit obtained from visually displayed text from an automatic speech recognizer during listening to speech presented in noise

    NARCIS (Netherlands)

    Zekveld, A.A.; Kramer, S.E.; Kessens, J.M.; Vlaming, M.S.M.G.; Houtgast, T.

    2008-01-01

    OBJECTIVES: The aim of this study was to evaluate the benefit that listeners obtain from visually presented output from an automatic speech recognition (ASR) system during listening to speech in noise. DESIGN: Auditory-alone and audiovisual speech reception thresholds (SRTs) were measured. The SRT

  18. Text Detection and Recognition with Speech Output for Visually Challenged Person: A Review

    Directory of Open Access Journals (Sweden)

    Ms.Rupali D. Dharmale

    2015-03-01

    Full Text Available Reading text from scene, images and text boards is an exigent task for visually challenged persons. This task has been proposed to be carried out with the help of image processing. Since a long period of time, image processing has helped a lot in the field of object recognition and still an emerging area of research. The proposed system reads the text encountered in images and text boards with the aim to provide support to the visually challenged persons. Text detection and recognition in natural scene can give valuable information for many applications. In this work, an approach has been attempted to extract and recognize text from scene images and convert that recognized text into speech. This task can definitely be an empowering force in a visually challenged person's life and can be supportive in relieving them of their frustration of not being able to read whatever they want, thus enhancing the quality of their lives.

  19. Dialogue enabling speech-to-text user assistive agent system for hearing-impaired person.

    Science.gov (United States)

    Lee, Seongjae; Kang, Sunmee; Han, David K; Ko, Hanseok

    2016-06-01

    A novel approach for assisting bidirectional communication between people of normal hearing and hearing-impaired is presented. While the existing hearing-impaired assistive devices such as hearing aids and cochlear implants are vulnerable in extreme noise conditions or post-surgery side effects, the proposed concept is an alternative approach wherein spoken dialogue is achieved by means of employing a robust speech recognition technique which takes into consideration of noisy environmental factors without any attachment into human body. The proposed system is a portable device with an acoustic beamformer for directional noise reduction and capable of performing speech-to-text transcription function, which adopts a keyword spotting method. It is also equipped with an optimized user interface for hearing-impaired people, rendering intuitive and natural device usage with diverse domain contexts. The relevant experimental results confirm that the proposed interface design is feasible for realizing an effective and efficient intelligent agent for hearing-impaired.

  20. The Speect text-to-speech system entry for the Blizzard Challenge 2013

    CSIR Research Space (South Africa)

    Louw, JA

    2013-09-01

    Full Text Available , information structure and af- fect. In particular, [11] has applied the cognitive theory of [14], also known as the OCC model, to affect detec- tion from text. However, they have not been successful at modelling diverse emotions in synthesised speech [12]. Our... with respect to one’s attitudes (inter alia tastes). The goals, standards and attitudes of a person are the cognitive antecedents that determine whether his va- lenced reaction to the environment is positive or negative. A particular emotion is the consequent...

  1. A Novel FPGA Based Low Cost Solution for Tamil-text to Speech Synthesizer

    Directory of Open Access Journals (Sweden)

    T. Jayasankar

    2015-08-01

    Full Text Available This study presents a prior work of developing a single chip solution for Text-to-Speech synthesizer for Tamil (Tamil-TTS language. Though there are enormous works presented in the recent days to address TTS for their native languages, the motivation of this study is to develop a low-cost FPGA based solution for Tamil TTS synthesizer. This study uses the unique feature of Tamil language to eliminate the complexity involved in accessing a database of stored audio signals. It uses only the audio signals of consonants and vowels in the stored memory locations. The compound characters from the segmented input text are generated using a Direct Digital Synthesizer by operating at three different frequencies of phonetic interval units of Tamil. The proposed system is implemented in Cyclone IVE EP4CE115F29C7 FPGA device and the implementation results show that the proposed system outperforms the other similar methods in terms of memory utilization, text-to-speech time, area utilization and power dissipation. The accuracy of the system is examined with 25 native speakers and acceptable accuracy scale has been reached.

  2. Increase in Beta-Band Activity during Preparation for Overt Speech in Patients with Parkinson’s Disease

    Directory of Open Access Journals (Sweden)

    Peter Sörös

    2017-07-01

    Full Text Available Speech impairment is a frequent and often serious symptom of Parkinson’s disease (PD, characterized by a disorder of phonation, articulation and prosody. While research on the pathogenesis of the prominent limb motor symptoms has made considerable progress in recent years, the pathophysiology of PD speech impairment is still incompletely understood. To investigate the neural correlates of speech production in PD, EEG was recorded in 14 non-demented patients with idiopathic PD and preserved verbal fluency on regular dopaminergic medication (8 women; mean age ± SD: 69.5 ± 8.0 years. The control group consisted of 15 healthy age-matched individuals (7 women; age: 69.7 ± 7.0 years. All participants performed a visually-cued, overt speech production task; required utterances were papapa and pataka. During the preparatory phase of speech production, in a time window of 200–400 ms after presentation of the visual cue, β-power was significantly increased in PD patients compared to healthy controls. Previous research has shown that the physiological decrease of β-power preceding limb movement onset is delayed and smaller in PD patients off medication and normalizes under dopaminergic treatment. By contrast, our study demonstrates that β-power during preparation for speech production is higher in patients on dopaminergic therapy than controls. Thus, our results suggest that the mechanisms that regulate β-activity preceding limb movement and speech production differ in PD. The pathophysiological role of this increase in β-power during speech preparation needs to be determined.

  3. THE COMPOSITIONAL AND SPEECH ORGANIZATION OF REGULATION TEXT AS A REGULATORY DOCUMENT

    Directory of Open Access Journals (Sweden)

    Sharipova Roza Rifatovna

    2014-06-01

    Full Text Available The relevance of the study covered by this article is determined by the extension of the business communication scope, as well as the nessecity to upgrade the administrative activity of organizations which largely depends on the documentation quality. The documents are used in various communicative situations and reflect intercultural business relations, that is why the problem of studying the nature and functions of documents is urgent. Business communication involves interaction in different areas of activity, and a document is one of the main tools of regulating this process. The author studies a regulation, the document which ensures the systematization and adjustment of management process, reflects certain production processes and the order of their execution. Taking into account the complex of criteria (functioning level of document, specificity of business communication subjects, diversity of regulated processes, compositional and content, and speech organization of text, the author suggests to distinguish three types of regulations. The regulations of first type systemize the business activity at government level or corresponding administration. The regulations of second type are used to regulate external relations – with counter-agents, partners – during undetermined (long-term or determined (having starting and ending date validity period. The regulations of third type serve to regulate domestic relations within an organization and are mostly intended for staff. From the composition viewpoint, the regulations of all types represent the text consisting of several paginated sections; at this, the level of regulation functioning, the specificity of business communication subjects define the character of information – degree of its generality/detalization. The speech organization of studied documents is similar as it is characterized by use of lexis with process semantics and official clichés. The regulations differ in terminology

  4. Trainable prosodic model for standard Chinese Text-to-Speech system

    Institute of Scientific and Technical Information of China (English)

    TAO Jianhua; CAI Lianhong; ZHAO Shixia

    2001-01-01

    Putonghua prosody is characterized by its hierarchical structure when influenced by linguistic environments. Based on this, a neural network, with specially weighted factors and optimizing outputs, is described and applied to construct the Putonghua prosodic model in Text-to-Speech (TTS) system. Extensive tests show that the structure of the neural network characterizes the Putonghua prosody more exactly than traditional models. Learning rate is speeded up and computational precision is improved, which makes the whole prosodic model more efficient. Furthermore, the paper also stylizes the Putonghua syllable pitch contours with SPiS parameters (Syllable Pitch Stylized Parameters), and analyzes them in adjusting the syllable pitch. It shows that the SPiS parameters effectively characterize the Putonghua syllable pitch contours, and facilitate the establishment of the network model and the prosodic controlling.

  5. An Arabic Text-To-Speech System Based on Artificial Neural Networks

    OpenAIRE

    Ghadeer Al-Said; Moussa Abdallah

    2009-01-01

    Problem statement: With the rapid advancement in information technology and communications, computer systems increasingly offer the users the opportunity to interact with information through speech. The interest in speech synthesis and in building voices is increasing. Worldwide, speech synthesizers have been developed for many popular languages English, Spanish and French and many researches and developments have been applied to those languages. Arabic on the other hand, has been given littl...

  6. A token centric part-of-speech tagger for biomedical text.

    Science.gov (United States)

    Barrett, Neil; Weber-Jahnke, Jens

    2014-05-01

    Difficulties with part-of-speech (POS) tagging of biomedical text is accessing and annotating appropriate training corpora. These difficulties may result in POS taggers trained on corpora that differ from the tagger's target biomedical text (cross-domain tagging). In such cases where training and target corpora differ tagging accuracy decreases. This paper presents a POS tagger for cross-domain tagging called TcT. TcT estimates a tag's likelihood for a given token by combining token collocation probabilities and the token's tag probabilities calculated using a Naive Bayes classifier. We compared TcT to three POS taggers used in the biomedical domain (mxpost, Brill and TnT). We trained each tagger on a non-biomedical corpus and evaluated it on biomedical corpora. TcT was more accurate in cross-domain tagging than mxpost, Brill and TnT (respective averages 83.9, 81.0, 79.5 and 78.8). Our analysis of tagger performance suggests that lexical differences between corpora have more effect on tagging accuracy than originally considered by previous research work. Biomedical POS tagging algorithms may be modified to improve their cross-domain tagging accuracy without requiring extra training or large training data sets. Future work should reexamine POS tagging methods for biomedical text. This differs from the work to date that has focused on retraining existing POS taggers. Copyright © 2014 Elsevier B.V. All rights reserved.

  7. Lexicon, Genre and Local Discourse Organisation: French Speech Act Verbs and Journalistic Texts.

    Science.gov (United States)

    Monville-Burston, Monique; Waugh, Linda R.

    1998-01-01

    Analysis of French speech act verbs offers an account of how clarifying verbs (e.g., "preciser, souligner") are used in reported speech microstructures in news writing and how each verb has its own lexico-pragmatic specificity. The verbs contribute to textual cohesion and to clarification in terms of the relationship between the reported speech…

  8. Advances to the development of a basic Mexican sign-to-speech and text language translator

    Science.gov (United States)

    Garcia-Bautista, G.; Trujillo-Romero, F.; Diaz-Gonzalez, G.

    2016-09-01

    Sign Language (SL) is the basic alternative communication method between deaf people. However, most of the hearing people have trouble understanding the SL, making communication with deaf people almost impossible and taking them apart from daily activities. In this work we present an automatic basic real-time sign language translator capable of recognize a basic list of Mexican Sign Language (MSL) signs of 10 meaningful words, letters (A-Z) and numbers (1-10) and translate them into speech and text. The signs were collected from a group of 35 MSL signers executed in front of a Microsoft Kinect™ Sensor. The hand gesture recognition system use the RGB-D camera to build and storage data point clouds, color and skeleton tracking information. In this work we propose a method to obtain the representative hand trajectory pattern information. We use Euclidean Segmentation method to obtain the hand shape and Hierarchical Centroid as feature extraction method for images of numbers and letters. A pattern recognition method based on a Back Propagation Artificial Neural Network (ANN) is used to interpret the hand gestures. Finally, we use K-Fold Cross Validation method for training and testing stages. Our results achieve an accuracy of 95.71% on words, 98.57% on numbers and 79.71% on letters. In addition, an interactive user interface was designed to present the results in voice and text format.

  9. Construction of a screening instrument for Motor Speech Disorders: Standardization of phonetically balanced text "O Sol"

    Directory of Open Access Journals (Sweden)

    Ana P. Mendes

    2015-04-01

    Full Text Available Aging, neurodegenerative diseases or brain injuries are the main causes of motor speech disorders (MSD which impacted on the communication effectiveness and quality of life of humans with more than 60 years. MSD’s early identification is relevant to mitigate these effects. Objective: To ensure the psychometric criteria of validity, reliability and sensitivity of the European Portuguese (EP phonetically balanced text (PBT “O Sol” (The Sun for the norm speakers. Methodology: 55 subjects spoken of the three EP dialects with ages between [18-58] years participated on the validity and sensitivity criteria. Secondly, 10 subjects spoken of cen-tral-southern EP dialect with ages between [19-50] years participated on the reliability criteria. Subjects read aloud the PBT “O Sol”. Sound samples were captured with integrated Sony Linear PCM-D50 microphone recorder. International Phonetic Alphabet was used for transcription. Results: The PBT “The Sun” satisfied the seven construction pre-requisites of a PBT. It presented all EP phonemes and syllabic formats. For the three dialects, 6/38 phonemes presented significantly different absolute frequency averages (p <0.05. Inter-examiner agreement and intra-examiner were 82% and 91.3%, p <0.05, respectively. Conclusion: The PBT “O Sol” is valid, reliable and sensitive to dialectal variations of the EP.

  10. Investigating an Application of Speech-to-Text Recognition: A Study on Visual Attention and Learning Behaviour

    Science.gov (United States)

    Huang, Y-M.; Liu, C-J.; Shadiev, Rustam; Shen, M-H.; Hwang, W-Y.

    2015-01-01

    One major drawback of previous research on speech-to-text recognition (STR) is that most findings showing the effectiveness of STR for learning were based upon subjective evidence. Very few studies have used eye-tracking techniques to investigate visual attention of students on STR-generated text. Furthermore, not much attention was paid to…

  11. The school-based speech-language therapist: choosing multicultural texts.

    Science.gov (United States)

    Moodley, Saloshni; Chetty, Sandhya; Pahl, Jenny

    2005-01-01

    School-based speech-language therapists have a pivotal role in the transformation of education as directed by current education policy. The Revised National Curriculum Statement, for example, foregrounds a multicultural perspective in education, which impacts on the choice of Learning and Teaching Support Materials. Inappropriate support materials could create barriers to learning. Folktales were selected as an example of multicultural Learning and Teaching Support Materials. The responses of 10-year-old mainstream learners to five folktales reflecting a diversity of cultures were explored. Five girls and five boys in Grade 5 participated in the study, which was conducted in three phases. A questionnaire, a focus group interview, and audio-visual recordings were used to gather data. The qualitative method of constant comparison was used to analyse emerging themes. Five main themes were identified. Findings revealed that some participants responded most positively when folktales reflected their culture, gender, or physical characteristics. Participants' views on less familiar cultures were influenced by the mass media. The results highlighted the importance of the text as 'mirror' and as 'window'. The potential of folktales as multicultural Learning and Teaching Support Materials, the powerful influence of the educator on learners' responses, and the need for an anti-bias approach within education are discussed. Implications for future research and practice are highlighted.

  12. Timing of Gestures: Gestures Anticipating or Simultaneous With Speech as Indexes of Text Comprehension in Children and Adults.

    Science.gov (United States)

    Ianì, Francesco; Cutica, Ilaria; Bucciarelli, Monica

    2016-06-08

    The deep comprehension of a text is tantamount to the construction of an articulated mental model of that text. The number of correct recollections is an index of a learner's mental model of a text. We assume that another index of comprehension is the timing of the gestures produced during text recall; gestures are simultaneous with speech when the learner has built an articulated mental model of the text, whereas they anticipate the speech when the learner has built a less articulated mental model. The results of four experiments confirm the predictions deriving from our assumptions for both children and adults. Provided that the recollections are correct, the timing of gestures can differ and can be considered a further measure of the quality of the mental model, beyond the number of correct recollections.

  13. Re-Presenting Subversive Songs: Applying Strategies for Invention and Arrangement to Nontraditional Speech Texts

    Science.gov (United States)

    Charlesworth, Dacia

    2010-01-01

    Invention deals with the content of a speech, arrangement involves placing the content in an order that is most strategic, style focuses on selecting linguistic devices, such as metaphor, to make the message more appealing, memory assists the speaker in delivering the message correctly, and delivery ideally enables great reception of the message.…

  14. Re-Presenting Subversive Songs: Applying Strategies for Invention and Arrangement to Nontraditional Speech Texts

    Science.gov (United States)

    Charlesworth, Dacia

    2010-01-01

    Invention deals with the content of a speech, arrangement involves placing the content in an order that is most strategic, style focuses on selecting linguistic devices, such as metaphor, to make the message more appealing, memory assists the speaker in delivering the message correctly, and delivery ideally enables great reception of the message.…

  15. Reading Comprehension of an Inferential Text by Deaf Students with Cochlear Implants Using Cued Speech

    Science.gov (United States)

    Torres, Santiago; Rodriguez, Jose-Miguel; Garcia-Orza, Javier; Calleja, Marina

    2008-01-01

    The aim of this study was to explore the ability of children who are profoundly deaf to reach high levels of reading proficiency on an inferential reading task. In an experimental narrative reading task, four children with prelingual hearing loss who used cued speech (MOC group) were compared with 58 students with typical hearing: 30 peers at the…

  16. Supported eText: Effects of Text-to-Speech on Access and Achievement for High School Students with Disabilities

    Science.gov (United States)

    Izzo, Margo Vreeburg; Yurick, Amanda; McArrell, Bianca

    2009-01-01

    Students with disabilities often lack the skills required to access the general education curriculum and achieve success in school and postschool environments. Evidence suggests that using assistive technologies such as digital texts and translational supports enhances outcomes for these students (Anderson-Inman & Horney, 2007). The purpose of the…

  17. Increase in Beta-Band Activity during Preparation for Overt Speech in Patients with Parkinson's Disease.

    Science.gov (United States)

    Sörös, Peter; Doñamayor, Nuria; Wittke, Catharina; Al-Khaled, Mohamed; Brüggemann, Norbert; Münte, Thomas F

    2017-01-01

    Speech impairment is a frequent and often serious symptom of Parkinson's disease (PD), characterized by a disorder of phonation, articulation and prosody. While research on the pathogenesis of the prominent limb motor symptoms has made considerable progress in recent years, the pathophysiology of PD speech impairment is still incompletely understood. To investigate the neural correlates of speech production in PD, EEG was recorded in 14 non-demented patients with idiopathic PD and preserved verbal fluency on regular dopaminergic medication (8 women; mean age ± SD: 69.5 ± 8.0 years). The control group consisted of 15 healthy age-matched individuals (7 women; age: 69.7 ± 7.0 years). All participants performed a visually-cued, overt speech production task; required utterances were papapa and pataka. During the preparatory phase of speech production, in a time window of 200-400 ms after presentation of the visual cue, β-power was significantly increased in PD patients compared to healthy controls. Previous research has shown that the physiological decrease of β-power preceding limb movement onset is delayed and smaller in PD patients off medication and normalizes under dopaminergic treatment. By contrast, our study demonstrates that β-power during preparation for speech production is higher in patients on dopaminergic therapy than controls. Thus, our results suggest that the mechanisms that regulate β-activity preceding limb movement and speech production differ in PD. The pathophysiological role of this increase in β-power during speech preparation needs to be determined.

  18. Support vector machine and mel frequency Cepstral coefficient based algorithm for hand gestures and bidirectional speech to text device

    Science.gov (United States)

    Balbin, Jessie R.; Padilla, Dionis A.; Fausto, Janette C.; Vergara, Ernesto M.; Garcia, Ramon G.; Delos Angeles, Bethsedea Joy S.; Dizon, Neil John A.; Mardo, Mark Kevin N.

    2017-02-01

    This research is about translating series of hand gesture to form a word and produce its equivalent sound on how it is read and said in Filipino accent using Support Vector Machine and Mel Frequency Cepstral Coefficient analysis. The concept is to detect Filipino speech input and translate the spoken words to their text form in Filipino. This study is trying to help the Filipino deaf community to impart their thoughts through the use of hand gestures and be able to communicate to people who do not know how to read hand gestures. This also helps literate deaf to simply read the spoken words relayed to them using the Filipino speech to text system.

  19. Serving a diverse population: the role of speech-language pathology professional preparation programs.

    Science.gov (United States)

    Stewart, Sharon R; Gonzalez, Lori S

    2002-01-01

    A national survey of 228 program directors was conducted to determine how master's level professional preparation programs are meeting the challenge of preparing speech-language pathologists to provide quality services to an increasingly diverse population. A total of 91 respondents provided information regarding their programs' efforts to address diversity by increasing the diversity of speech-language pathology professionals, preparing students in research for diverse populations, and providing students with the didactic knowledge and clinical experience required to serve diverse populations. Results indicated that professional preparation programs continue to lag in their enrollment of minority students, but there are efforts actively to recruit and retain students from diverse groups. Much variation in preparation in research was found across programs. Graduate students are being presented with information concerning diversity issues, but clinical experiences vary greatly according to the geographic location of the preparation program and individual practicum placements. Implications of these findings for speech-language pathology preparation programs and other allied health programs are discussed.

  20. Parts-of-Speech Tagger Errors Do Not Necessarily Degrade Accuracy in Extracting Information from Biomedical Text

    CERN Document Server

    Ling, Maurice HT; Nicholas, Kevin R

    2008-01-01

    A recent study reported development of Muscorian, a generic text processing tool for extracting protein-protein interactions from text that achieved comparable performance to biomedical-specific text processing tools. This result was unexpected since potential errors from a series of text analysis processes is likely to adversely affect the outcome of the entire process. Most biomedical entity relationship extraction tools have used biomedical-specific parts-of-speech (POS) tagger as errors in POS tagging and are likely to affect subsequent semantic analysis of the text, such as shallow parsing. This study aims to evaluate the parts-of-speech (POS) tagging accuracy and attempts to explore whether a comparable performance is obtained when a generic POS tagger, MontyTagger, was used in place of MedPost, a tagger trained in biomedical text. Our results demonstrated that MontyTagger, Muscorian's POS tagger, has a POS tagging accuracy of 83.1% when tested on biomedical text. Replacing MontyTagger with MedPost did ...

  1. Nazareth College: Specialty Preparation for Speech-Language Pathologists to Work with Children Who Are Deaf and Hard of Hearing

    Science.gov (United States)

    Brown, Paula M.; Quenin, Cathy

    2010-01-01

    The specialty preparation program within the speech-language pathology master's degree program at Nazareth College in Rochester, New York, was designed to train speech-language pathologists to work with children who are deaf and hard of hearing, ages 0 to 21. The program is offered in collaboration with the Rochester Institute of Technology,…

  2. Does Use of Text-to-Speech and Related Read-Aloud Tools Improve Reading Comprehension for Students With Reading Disabilities? A Meta-Analysis.

    Science.gov (United States)

    Wood, Sarah G; Moxley, Jerad H; Tighe, Elizabeth L; Wagner, Richard K

    2017-01-01

    Text-to-speech and related read-aloud tools are being widely implemented in an attempt to assist students' reading comprehension skills. Read-aloud software, including text-to-speech, is used to translate written text into spoken text, enabling one to listen to written text while reading along. It is not clear how effective text-to-speech is at improving reading comprehension. This study addresses this gap in the research by conducting a meta-analysis on the effects of text-to-speech technology and related read-aloud tools on reading comprehension for students with reading difficulties. Random effects models yielded an average weighted effect size of ([Formula: see text] = .35, with a 95% confidence interval of .14 to .56, p text-to-speech technologies may assist students with reading comprehension. However, more studies are needed to further explore the moderating variables of text-to-speech and read-aloud tools' effectiveness for improving reading comprehension. Implications and recommendations for future research are discussed.

  3. When will a stuttering moment occur? The determining role of speech motor preparation.

    Science.gov (United States)

    Vanhoutte, Sarah; Cosyns, Marjan; van Mierlo, Pieter; Batens, Katja; Corthals, Paul; De Letter, Miet; Van Borsel, John; Santens, Patrick

    2016-06-01

    The present study aimed to evaluate whether increased activity related to speech motor preparation preceding fluently produced words reflects a successful compensation strategy in stuttering. For this purpose, a contingent negative variation (CNV) was evoked during a picture naming task and measured by use of electro-encephalography. A CNV is a slow, negative event-related potential known to reflect motor preparation generated by the basal ganglia-thalamo-cortical (BGTC) - loop. In a previous analysis, the CNV of 25 adults with developmental stuttering (AWS) was significantly increased, especially over the right hemisphere, compared to the CNV of 35 fluent speakers (FS) when both groups were speaking fluently (Vanhoutte et al., (2015) doi: 10.1016/j.neuropsychologia.2015.05.013). To elucidate whether this increase is a compensation strategy enabling fluent speech in AWS, the present analysis evaluated the CNV of 7 AWS who stuttered during this picture naming task. The CNV preceding AWS stuttered words was statistically compared to the CNV preceding AWS fluent words and FS fluent words. Though no difference emerged between the CNV of the AWS stuttered words and the FS fluent words, a significant reduction was observed when comparing the CNV preceding AWS stuttered words to the CNV preceding AWS fluent words. The latter seems to confirm the compensation hypothesis: the increased CNV prior to AWS fluent words is a successful compensation strategy, especially when it occurs over the right hemisphere. The words are produced fluently because of an enlarged activity during speech motor preparation. The left CNV preceding AWS stuttered words correlated negatively with stuttering frequency and severity suggestive for a link between the left BGTC - network and the stuttering pathology. Overall, speech motor preparatory activity generated by the BGTC - loop seems to have a determining role in stuttering. An important divergence between left and right hemisphere is

  4. Speaker-dependent Dictionary-based Speech Enhancement for Text-Dependent Speaker Verification

    DEFF Research Database (Denmark)

    Thomsen, Nicolai Bæk; Thomsen, Dennis Alexander Lehmann; Tan, Zheng-Hua

    2016-01-01

    The problem of text-dependent speaker verification under noisy conditions is becoming ever more relevant, due to increased usage for authentication in real-world applications. Classical methods for noise reduction such as spectral subtraction and Wiener filtering introduce distortion and do...... not perform well in this setting. In this work we compare the performance of different noise reduction methods under different noise conditions in terms of speaker verification when the text is known and the system is trained on clean data (mis-matched conditions). We furthermore propose a new approach based...

  5. Speaker-dependent Dictionary-based Speech Enhancement for Text-Dependent Speaker Verification

    DEFF Research Database (Denmark)

    Thomsen, Nicolai Bæk; Thomsen, Dennis Alexander Lehmann; Tan, Zheng-Hua

    2016-01-01

    The problem of text-dependent speaker verification under noisy conditions is becoming ever more relevant, due to increased usage for authentication in real-world applications. Classical methods for noise reduction such as spectral subtraction and Wiener filtering introduce distortion and do...

  6. Text-to-speech enhanced eBooks for emerging literacy development

    CSIR Research Space (South Africa)

    De Wet, Febe

    2015-09-01

    Full Text Available The purpose of this study was to measure the efficacy of an eBook to improve the vocabulary and word recognition skills in an Afrikaans speaking group of lower socio-economic status of 6- to 7-year old children with poor vocabulary. The main goals...

  7. Text-to-speech technology effects on reading rate and comprehension by adults with traumatic brain injury.

    Science.gov (United States)

    Harvey, Judy; Hux, Karen; Scott, Nikki; Snell, Jeffry

    2013-01-01

    This study's purpose was to examine the comprehension, rate and perceptions and reading preferences of adults with severe traumatic brain injury (TBI) when reading passages with and without computerized text-to-speech (TTS) support. Nine adults with severe TBI read 24 passages in two conditions: with and without TTS support. The researchers compared reading rate and comprehension accuracy across conditions. Also, participants rated their perceived performance and reading preferences via a follow-up questionnaire. Comparison to normative data revealed that all nine participants read slower than average neurotypical readers. As a group, participants read significantly faster with TTS support than without such support, even though the TTS reading rate was roughly comparable to the oral rather than silent reading rate of neurotypical adults. No significant differences in comprehension resulted between the two conditions. Over half of the participants preferred the TTS condition over the no-TTS condition. In general, participants were inaccurate in judging their relative reading rates and comprehension accuracy across conditions. TTS may improve reading efficiency without compromising reading comprehension accuracy for adults with TBI. Given this finding, some survivors may find use of TTS technology contributes to increased participation in and efficiency when performing reading activities.

  8. Trigonometry, A Tentative Guide Prepared for Use with the Text Plane Trigonometry with Tables.

    Science.gov (United States)

    Brant, Vincent

    This teacher's guide for a semester course in trigonometry is prepared for use with the text "Plane Trigonometry with Tables" by E. R. Heineman. Included is a daily schedule of topics for discussion and homework assignments. The scope of each lesson and teaching suggestions are provided. The content for the course includes trigonometric functions,…

  9. Auditory Speech Recognition and Visual Text Recognition in Younger and Older Adults: Similarities and Differences between Modalities and the Effects of Presentation Rate

    Science.gov (United States)

    Humes, Larry E.; Burk, Matthew H.; Coughlin, Maureen P.; Busey, Thomas A.; Strauser, Lauren E.

    2007-01-01

    Purpose: To examine age-related differences in auditory speech recognition and visual text recognition performance for parallel sets of stimulus materials in the auditory and visual modalities. In addition, the effects of variation in rate of presentation of stimuli in each modality were investigated in each age group. Method: A mixed-model design…

  10. Investigating Applications of Speech-to-Text Recognition Technology for a Face-to-Face Seminar to Assist Learning of Non-Native English-Speaking Participants

    Science.gov (United States)

    Shadiev, Rustam; Hwang, Wu-Yuin; Huang, Yueh-Min; Liu, Chia-Ju

    2016-01-01

    This study applied speech-to-text recognition (STR) technology to assist non-native English-speaking participants to learn at a seminar given in English. How participants used transcripts generated by the STR technology for learning and their perceptions toward the STR were explored. Three main findings are presented in this study. Most…

  11. Preparing technical text for translation: A comparison between International English and methods for simplifying language

    Energy Technology Data Exchange (ETDEWEB)

    Buican, I.; Hriscu, V.; Amador, M.

    1994-07-01

    For the past four and a half years, the International Communication Committee at Los Alamos National Laboratory has been working to develop a set of guidelines for writing technical and scientific documents in International English, that is, English for those whose native language is not English. Originally designed for documents intended for presentation in English to an international audience of technical experts, the International English guidelines apply equally well to the preparation of English text for translation. This is the second workshop in a series devoted to the topic of translation. The authors focus on the advantages of using International English, rather than various methods of simplifying language, to prepare scientific and technical text for translation.

  12. Syllables in speech production : effects of syllable preparation and syllable frequency

    NARCIS (Netherlands)

    Cholin, J.

    2004-01-01

    The fluent production of speech is a very complex human skill. It requires the coordination of several articulatory subsystems. The instructions that lead articulatory movements to execution are the result of the interplay of speech production levels that operate above the articulatory network. Duri

  13. REPORTED SPEECH IN FICTIONAL NARRATIVE TEXTS IN TERMS OF SPEECH ACTS THEORY SÖZ EDİMLERİ KURAMI AÇISINDAN KURGUSAL ANLATI METİNLERİNDE SÖZ AKTARIMI

    Directory of Open Access Journals (Sweden)

    Soner AKŞEHİRLİ

    2011-06-01

    Full Text Available Speech or discourse reporting (speech representation is a linguistic phenomenon which is seen both in ordinary communication and fictional narrative texts. In linguistics, speech reporting is differentiated as direct, indirect and free-indirect speech. On the other and, speech acts theory, suggested by J.L.Auistin, can provide a new perspective for speech reporting. According to theory, to say or to produce a statement (locutionary act is to perform an act (illocutionary act.Moreover, one can performed an act ifluenced by an locutionary act. In ordinary communication, reporter and in fictional texts narrator may report one, two or all of the locutionary act, illocutionary act and perlocutionary act of reported statement. At the same time, these processes must considered in determining point of view that governing narrative texts. So that, we can develop a new typology of speech reporting for fictional texts Söz ya da söylem aktarımı hem günlük iletişimde hem de kurgusal anlatı metinlerinde sıkça görülen dilbilimsel bir olgudur. Dilbilim açısından söz aktarımı doğrudan, dolaylı ve serbest dolaylı olmak üzere üç temel biçimde ele alınır. J.L.Austin tarafından geliştiren söz edimleri kuramı ise, söz aktarımına farklı bir açıdan bakmamızı sağlayabilir. Kurama göre bir söz söylemek (düzsöz, bir iş yapmaktır (edimsöz. Ayrıca söylenen sözün etkisiyle yapılan bir iş de olabilir (etkisöz. Günlük iletişimde aktarıcı, kurgusal metinlerde ise anlatıcı söz aktarımını gerçekleştirirken, aktardığı sözün düzsöz, edimsöz ve etkisöz bileşenlerinden herhangi birini, ikisini ya da üçünü birden aktarabilir. Bu aynı zamanda anlatısal metinleri yöneten bakış açısının belirlenmesinde de dikkate alınması gereken bir süreçtir. Böylece kurgusal metinler için söz edimleri kuramına dayanan yeni bir söz aktarım tipolojisi oluşturulabilir.

  14. Study on automatic prediction of sentential stress for Chinese Putonghua Text-to-Speech system with natural style

    Institute of Scientific and Technical Information of China (English)

    SHAO Yanqiu; HAN Jiqing; ZHAO Yongzhen; LIU Ting

    2007-01-01

    Stress is an important parameter for prosody processing in speech synthesis. In this paper, we compare the acoustic features of neutral tone syllables and strong stress syllables with moderate stress syllables, including pitch, syllable duration, intensity and pause length after syllable. The relation between duration and pitch, as well as the Third Tone (T3) and pitch are also studied. Three stress prediction models based on ANN, i.e. the acoustic model,the linguistic model and the mixed model, are presented for predicting Chinese sentential stress.The results show that the mixed model performs better than the other two models. In order to solve the problem of the diversity of manual labeling, an evaluation index of support ratio is proposed.

  15. Amharic Speech Recognition for Speech Translation

    OpenAIRE

    Melese, Michael; Besacier, Laurent; Meshesha, Million

    2016-01-01

    International audience; The state-of-the-art speech translation can be seen as a cascade of Automatic Speech Recognition, Statistical Machine Translation and Text-To-Speech synthesis. In this study an attempt is made to experiment on Amharic speech recognition for Amharic-English speech translation in tourism domain. Since there is no Amharic speech corpus, we developed a read-speech corpus of 7.43hr in tourism domain. The Amharic speech corpus has been recorded after translating standard Bas...

  16. Event-related brain potential investigation of preparation for speech production in late bilinguals

    Directory of Open Access Journals (Sweden)

    Yan Jing eWu

    2011-05-01

    Full Text Available It has been debated how bilinguals select the intended language and prevent interference from the unintended language when speaking. Here, we studied the nature of the mental representations accessed by late fluent bilinguals during a rhyming judgment task relying on covert speech production. We recorded event-related brain potentials in Chinese-English bilinguals and monolingual speakers of English while they indicated whether the names of pictures presented on a screen rhymed. Whether bilingual participants focussed on rhyming selectively in English or Chinese, we found a significant priming effect of language-specific sound repetition. Surprisingly, however, sound repetitions in Chinese elicited significant priming effects even when the rhyming task was performed in English. This cross-language priming effect was delayed by ~200 ms as compared to the within-language effect and was asymmetric, since there was no priming effect of sound repetitions in English when participants were asked to make rhyming judgements in Chinese. These results demonstrate that second language production hinders, but does not seal off, activation of the first language, whereas native language production appears immune to competition from the second language.

  17. Speech Therapy Prevention in Kindergarten

    Directory of Open Access Journals (Sweden)

    Vašíková Jana

    2017-08-01

    Full Text Available Introduction: This contribution presents the results of a research focused on speech therapy in kindergartens. This research was realized in Zlín Region. It explains how speech therapy prevention is realized in kindergartens, determines the educational qualifications of teachers for this activity and verifies the quality of the applied methodologies in the daily program of kindergartens. Methods: The empirical part of the study was conducted through a qualitative research. For data collection, we used participant observation. We analyzed the research data and presented them verbally, using frequency tables and graphs, which were subsequently interpreted. Results: In this research, 71% of the teachers completed a course of speech therapy prevention, 28% of the teachers received pedagogical training and just 1% of the teachers are clinical speech pathologists. In spite of this, the research data show that, in most of kindergartens, the aim of speech therapy prevention is performed in order to correct deficiencies in speech and voice. The content of speech therapy prevention is implemented in this direction. Discussion: Awareness of the teachers’/parents’ regarding speech therapy prevention in kindergartens. Limitations: This research was implemented in autumn of 2016 in Zlín Region. Research data cannot be generalized to the entire population. We have the ambition to expand this research to other regions next year. Conclusions: Results show that both forms of speech therapy prevention - individual and group - are used. It is also often a combination of both. The aim of the individual forms is, in most cases, to prepare a child for cooperation during voice correction. The research also confirmed that most teachers do not have sufficient education in speech therapy. Most of them completed a course of speech therapy as primary prevention educators. The results also show that teachers spend a lot of time by speech therapy prevention in

  18. The software for automatic creation of the formal grammars used by speech recognition, computer vision, editable text conversion systems, and some new functions

    Science.gov (United States)

    Kardava, Irakli; Tadyszak, Krzysztof; Gulua, Nana; Jurga, Stefan

    2017-02-01

    For more flexibility of environmental perception by artificial intelligence it is needed to exist the supporting software modules, which will be able to automate the creation of specific language syntax and to make a further analysis for relevant decisions based on semantic functions. According of our proposed approach, of which implementation it is possible to create the couples of formal rules of given sentences (in case of natural languages) or statements (in case of special languages) by helping of computer vision, speech recognition or editable text conversion system for further automatic improvement. In other words, we have developed an approach, by which it can be achieved to significantly improve the training process automation of artificial intelligence, which as a result will give us a higher level of self-developing skills independently from us (from users). At the base of our approach we have developed a software demo version, which includes the algorithm and software code for the entire above mentioned component's implementation (computer vision, speech recognition and editable text conversion system). The program has the ability to work in a multi - stream mode and simultaneously create a syntax based on receiving information from several sources.

  19. The Analysis of Speech Reporting in Political News Texts%政治新闻语篇言语转述分析

    Institute of Scientific and Technical Information of China (English)

    许华蓉

    2014-01-01

    The phenomenon of speech reporting is currently receiving more attention .Based on the in‐tertextuality theory ,this thesis studies the phenomenon in political news texts .With the appliance and a‐nalysis of the corpus of more than 30 pieces of news texts from China Daily and Washington Post ,the thesis trys to figure out their features in speech reporting ,explore the realization of new s value .%言语转述现象近年来引起人们越来越多的关注。本文主要以互文性理论为基础,研究政治新闻语篇的言语转述现象。以《中国日报》(China Daily)和《华盛顿邮报》(The Washington Post)30多篇新闻语篇为语料,分析其言语转述特点,探究其新闻价值的实现。

  20. 基于HCSIPA的中英文混合语音合成%Mandarin-English Mixed Text to Speech Based on HCSIPA

    Institute of Scientific and Technical Information of China (English)

    徐英进; 蔡莲红

    2013-01-01

    基于双语说话人的中英文混合合成,提出一种中英文通用音标符号——HCSIPA,采用发音方式和部位为构造标准.在中英文混合聚类中,构造针对HCSIPA的中英文共用问题集,以提高中英文在决策树结构上的区分度.实验结果表明,HCSIPA能提高中文和英文的发音单元混合度,减少语言差距带来的音色差距,基于HCSIPA的混合合成系统可以合成较高质量的中英文语音,且中英文混合对单种语言合成的质量下降不明显.%A Mandarin-English general phonetic alphabet Human Computer Speech Interaction Phonetic Alphabet(HCSIPA),in which the articulation manner and place are applied for construction criterion,is proposed in bilingual speaker based Mandarin-English mixed Text to Speech(TTS).A Mandarin-English common question set suitable for HCSIPA is constructed in Mandarin-English mix clustering.Its construction target is to increase the distinguishing between Mandarin and English in the decision tree structure.The result verifies the HCSIPA can increase the mixing with pronunciation unit of Mandarin and English,to reduce the timbre gap caused by language gap.The mixed TTS based on HCSIPA can synthesize the speech of Mandarin and English with a high quality,and the quality decline to single language synthesis caused by mix of Mandarin and English is not obvious.

  1. Measuring cognitive factors in speech comprehension: the value of using the text reception threshold test as a visual equivalent of the SRT test

    NARCIS (Netherlands)

    Kramer, S.E.; Zekveld, A.A.; Houtgast, T.

    2009-01-01

    The ability to comprehend speech in noise is influenced by bottom-up auditory and top-down cognitive capacities. Separate examination of these capacities is relevant for various purposes. Speech-Reception-Threshold (SRT) tests measure an individual's ability to comprehend speech. This paper addresse

  2. Distracted While Reading? Changing to A Hard-to-read Font Shields against the Effects of Environmental Noise and Speech on Text Memory

    Directory of Open Access Journals (Sweden)

    Niklas Halin

    2016-08-01

    Full Text Available The purpose of this study was to investigate the distractive effects of background speech, aircraft noise and road traffic noise on text memory and particularly to examine if displaying the texts in a hard-to-read font can shield against the detrimental effects of these types of background sounds. This issue was addressed in an experiment where 56 students read shorter texts about different classes of fictitious creatures (i.e., animals, fishes, birds, and dinosaurs against a background of the aforementioned background sounds respectively and silence. For half of the participants the texts were displayed in an easy-to-read font (i.e., Times New Roman and for the other half in a hard-to-read font (i.e., Haettenschweiler. The dependent measure was the proportion correct answers on the multiple-choice tests that followed each sound condition. Participants’ performance in the easy-to-read font condition was significantly impaired by all three background sound conditions compared to silence. In contrast, there were no effects of the three background sound conditions compared to silence in the hard-to-read font condition. These results suggest that an increase in task demand—by displaying the text in a hard-to-read font—shields against various types of distracting background sounds by promoting a more steadfast locus-of-attention and by reducing the processing of background sound.

  3. Enhancing Learning Performance, Attention, and Meditation Using a Speech-to-Text Recognition Application: Evidence from Multiple Data Sources

    Science.gov (United States)

    Shadiev, Rustam; Wu, Ting-Ting; Huang, Yueh-Min

    2017-01-01

    In this study, we provide STR-texts to non-native English speaking students during English lectures to facilitate learning, attention, and meditation. We carry out an experiment to test the feasibility of our approach. Our results show that the participants in the experimental group both outperform those in the control group on the post-tests and…

  4. Negative constructions in memoir texts by A.Konchalovsky as indices of speech characteristics and author’s personality characteristics

    Directory of Open Access Journals (Sweden)

    Elena A. Bolotova

    2016-09-01

    Full Text Available The aim of the article is to examine the relevant problem of negation in linguistics, as well as the possible ways of use by the author the memoirs texts by A. Konchalovsky («Low truths» and «Exalts deception» the negative constructions, their pragmatic orientation. It becomes clear how and what certain fiction-figurative and expressive means, based on the use of linguistic resources at various levels, it is possible to characterize the individual memoirist using negative semantic proposals. We consider the specifics of the use of various methods of influence by the author of memoirs on the reader by the means of representation categories of negopositivism. It reveals that any information that is submitted by the author in the text of memoirs, refers to the identity of the author, memoirist, and entering into a dialogue with their readers, the author establishes contacts with them. In the course of this dialogue there is the exchange of information about people, about oneself, about different events and facts in the different periods of the life of the author and the country.

  5. "What Is the Text Doing?" Preparing Pre-Service Teachers to Teach Primary Sources Effectively

    Science.gov (United States)

    Neumann, David J.

    2010-01-01

    Recent education research strongly endorses the notion that knowledge is discipline-based, dependent upon the ground rules and procedures of a particular field of study. Since reading primary sources is central to the discipline of history, many teacher preparation programs in history-social science devote substantial instructional time to…

  6. Text Linguistics in Research Papers Prepared by University Students: Teaching through Lesson Plans and Textbooks

    Directory of Open Access Journals (Sweden)

    Manuel Albarrán-Santiago

    2015-01-01

    Full Text Available This research project revolves around the properties of text linguistics under a qualitative approach.  The author analyzed drafts of a research paper by two university students as well as lesson plans and textbooks of high school Spanish Language and Literature courses and lesson plans of courses from the Licentiate degree in Education.  According to the information from the drafts, students struggle with coherence and cohesion in writing; however, they succeed in choosing the correct language for the type of writing.  Difficulties are most likely due to fact that this topic is not included in secondary education plans and is not commonly addressed in textbooks or university classes.  In conclusion, teachers should include the properties of text linguistics in their lesson plans in order to help students overcome these difficulties.

  7. Product of Likelihood Ratio Scores Fusion of Dynamic Face, Text Independent Speech and On-line Signature Based Biometrics Verification Application Systems

    Directory of Open Access Journals (Sweden)

    Mohamed SOLTANE

    2015-09-01

    Full Text Available In this paper, the use of finite Gaussian mixture modal (GMM tuned using Expectation Maximization (EM estimating algorithms for score level data fusion is proposed. Automated biometric systems for human identification measure a “signature” of the human body, compare the resulting characteristic to a database, and render an application dependent decision. These biometric systems for personal authentication and identification are based upon physiological or behavioral features which are typically distinctive, Multi-biometric systems, which consolidate information from multiple biometric sources, are gaining popularity because they are able to overcome limitations such as non-universality, noisy sensor data, large intra-user variations and susceptibility to spoof attacks that are commonly encountered in mono modal biometric systems. Simulation result show that finite mixture modal (GMM is quite effective in modelling the genuine and impostor score densities, fusion based the product of Likelihood Ratio achieves a significant performance on eNTERFACE 2005 multi-biometric database based on dynamic face, on-line signature and text independent speech modalities.

  8. Preparing College Students To Search Full-Text Databases: Is Instruction Necessary?

    Science.gov (United States)

    Riley, Cheryl; Wales, Barbara

    Full-text databases allow Central Missouri State University's clients to access some of the serials that libraries have had to cancel due to escalating subscription costs; EbscoHost, the subject of this study, is one such database. The database is available free to all Missouri residents. A survey was designed consisting of 21 questions intended…

  9. A Comparison of Inter-Professional Education Programs in Preparing Prospective Teachers and Speech and Language Pathologists for Collaborative Language-Literacy Instruction

    Science.gov (United States)

    Wilson, Leanne; McNeill, Brigid; Gillon, Gail T.

    2016-01-01

    Ensuring teacher and speech and language pathology graduates are prepared to work collaboratively together to meet the diverse language literacy learning needs of children is an important goal. This study investigated the efficacy of a 3-h inter-professional education program focused on explicit instruction in the language skills that underpin…

  10. Sensorimotor oscillations prior to speech onset reflect altered motor networks in adults who stutter

    Directory of Open Access Journals (Sweden)

    Anna-Maria Mersov

    2016-09-01

    Full Text Available Adults who stutter (AWS have demonstrated atypical coordination of motor and sensory regions during speech production. Yet little is known of the speech-motor network in AWS in the brief time window preceding audible speech onset. The purpose of the current study was to characterize neural oscillations in the speech-motor network during preparation for and execution of overt speech production in AWS using magnetoencephalography (MEG. Twelve AWS and twelve age-matched controls were presented with 220 words, each word embedded in a carrier phrase. Controls were presented with the same word list as their matched AWS participant. Neural oscillatory activity was localized using minimum-variance beamforming during two time periods of interest: speech preparation (prior to speech onset and speech execution (following speech onset. Compared to controls, AWS showed stronger beta (15-25Hz suppression in the speech preparation stage, followed by stronger beta synchronization in the bilateral mouth motor cortex. AWS also recruited the right mouth motor cortex significantly earlier in the speech preparation stage compared to controls. Exaggerated motor preparation is discussed in the context of reduced coordination in the speech-motor network of AWS. It is further proposed that exaggerated beta synchronization may reflect a more strongly inhibited motor system that requires a stronger beta suppression to disengage prior to speech initiation. These novel findings highlight critical differences in the speech-motor network of AWS that occur prior to speech onset and emphasize the need to investigate further the speech-motor assembly in the stuttering population.

  11. 基于CART技术的汉语韵律短语分析%A CART-based Prosodic Phrasing Method for Chinese Text-to-Speech

    Institute of Scientific and Technical Information of China (English)

    谌卫军; 林福宗; 李建民; 张钹

    2002-01-01

    @@ 韵律短语分析是文语转换(Text-to-Speech,TTS)系统中的一项重要内容.它负责在输入文本的不同位置上,标记上不同级别的符号形式的韵律短语边界,从而把一个句子分隔成不同层次的短语.韵律短语分析的结果直接影响着TTS系统的后继步骤,如韵律参数的生成.杨玉芳[1]研究了不同级别的韵律短语边界与附近音节的时长、停顿、基频等韵律参数之间的关系,考察了这些参数随边界等级的系统变化.此外,在口语理解等应用系统中,韵律短语分析也有可用之处.

  12. Preparing for reading comprehension: Fostering text comprehension skills in preschool and early elementary school children

    Directory of Open Access Journals (Sweden)

    Paul van den BROEK

    2011-11-01

    Full Text Available To understand what they read or hear, children and adults must create a coherent mental representation of presented information. Recent research suggests that the ability to do so starts to develop early –well before reading age- and that early individual differences are predictive of later reading-comprehension performance. In this paper, we review this research and discuss potential applications to early intervention. We then present two exploratory studies in which we examine whether it is feasible to design interventions with early readers (3rd grade and even toddlers (2-3 years old. The interventions employed causal questioning techniques as children listen to orally presented,age-appropriate narratives. Afterwards, comprehension was tested through question answering and recall tasks. Results indicate that such interventions are indeed feasible. Moreover, they suggest thatfor both toddlers and early readers questions during comprehension are more effective than questions after comprehension. Finally, for both groups higher working memory capacity was related to bettercomprehension.

  13. On the Effective Language Writing of English Prepared Speeches%英文命题演讲写作的语言问题探讨

    Institute of Scientific and Technical Information of China (English)

    王琦

    2012-01-01

    英文命题演讲的写作比其他演讲写作对语言表达的要求更高,容错度更低。词汇使用不精确、句子结构不合理、语篇衔接不自然、缺乏修辞手段的润色是学生演讲稿常见的语言问题。发现并解决这些语言问题是提升演讲稿的语言面貌和整体质量的关键。%Unlike other types of speech writing, English prepared speeches are expected to be highly-polished and error-free. Inaccurate words, illogical sentence structure, abrupt cohesion between paragraphs and no employment of rhetorical devices-all these are detriments to effective and appropriate speech language. It is vital for the speakers to address these problems in speech language so as to create memorable speeches.

  14. Analysis of Nominalization Phenomenon and Function of Nominalization in the Text of Political Speech%政治演说类语篇中的名物化现象及功能分析

    Institute of Scientific and Technical Information of China (English)

    孙英林

    2012-01-01

    名物化是语法隐喻的主要形式,政治演说类语篇是兼有口头和书面特征的正式的成人语篇,其名物化现象更有其复杂性和特殊性,主要体现在增加语篇的信息量和客观性等方面。为此,有必要对于这种语篇进行专门的研究,以揭示这种兼有两面特征的语篇中的名物化的分布情况及功能特征,从而更好的把握该类语篇的内在规律,更好地理解其语言的意义和功能。%The nominalization is the main form of grammatical metaphor,a political speech text is the formal adult discourse combining both verbal and written features,and its nominalization phenomenon has its complexity and particularity, mainly reflected in the aspects of increasing the speech information and objective.Therefore,it is necessary to dedicated study the speech,to reveal the characteristics of both sides of the nominalization distribution and functional characteristics of speech,so as to better grasp of such inherent laws of speech and better understanding of its meaning and function.

  15. Increase in Beta-Band Activity during Preparation for Overt Speech in Patients with Parkinson’s Disease

    OpenAIRE

    Sörös, Peter; Doñamayor, Nuria; Wittke, Catharina; Al-Khaled, Mohamed; Brüggemann, Norbert; Thomas F Münte

    2017-01-01

    Speech impairment is a frequent and often serious symptom of Parkinson’s disease (PD), characterized by a disorder of phonation, articulation and prosody. While research on the pathogenesis of the prominent limb motor symptoms has made considerable progress in recent years, the pathophysiology of PD speech impairment is still incompletely understood. To investigate the neural correlates of speech production in PD, EEG was recorded in 14 non-demented patients with idiopathic PD and preserved v...

  16. Blending the principles of Suggestopedia and thetheory of Speech Acts in writing suggestopedicdidactic texts, with reference to German andZulu scripts

    Directory of Open Access Journals (Sweden)

    R.H. Bodenstein

    2013-02-01

    Full Text Available This paper suggests that language teachers who use the suggestopedic method should write their own texts that comply with suggestopedic principles. This is imperative because of the lack of material that can be acquired and used in such courses. Writing their own scripts also enables teachers to identify with their materials and brings much reward and personal growth. Guidelines for the writing and setting up of these texts are provided The text should embody the philosophic and didactic .framework of suggestopedia. It should also be presented as a didactic play wherein the language components to be learned are presented in the form of new scenes in a continuous drama text. Traditional beliefs about level of complexity of the language suitable for beginners' courses are considered unfounded Suggestopedic scripts therefore contain complex, 'reallife' language .from the outset, starting with the language needed to make contact with native target language speakers. The main guideline for the organisation and structuring of the text is that it should mirror authentic communicative situations. The paper therefore argues that suggestopedic scripts should be written according to the lists of language jUnctions (or speech acts and topic areas required for the so-called 'threshold level' of language competence. The paper concludes with examples .from a German and a Zulu text to illustrate the didactic and structural principles and guidelines that were outlined in the article.Die artikel voer aan dat taalonderwysers wat die suggestopediese metode gebruik, self tekste behoort te skryf wat strook met die suggestopediese beginsels. Die gebrek aan geskikte materiaal op die mark noodsaak hulle om dit te doen. Wanneer onderwysers hulle eie tekste skryj beteken dit egter ook dat hulle met die onderrigmateriaal kan identifiseer. Dit kan professioneellonend wees en persoonlike groei teweeg bring. Riglyne vir die skryf en ontwerp van sulke tekste word verskaf Die

  17. Speech to Text: Today and Tomorrow. Proceedings of a Conference at Gallaudet University (Washington, D.C., September, 1988). GRI Monograh Series B, No. 2.

    Science.gov (United States)

    Harkins, Judith E., Ed.; Virvan, Barbara M., Ed.

    The conference proceedings contains 23 papers on telephone relay service, real-time captioning, and automatic speech recognition, and a glossary. The keynote address, by Representative Major R. Owens, examines current issues in federal legislation. Other papers have the following titles and authors: "Telephone Relay Service: Rationale and…

  18. Passion and Preparation in the Basic Course: The Influence of Students' Ego-Involvement with Speech Topics and Preparation Time on Public-Speaking Grades

    Science.gov (United States)

    Mazer, Joseph P.; Titsworth, Scott

    2012-01-01

    Authors of basic public-speaking course textbooks frequently encourage students to select speech topics in which they have vested interest, care deeply about, and hold strong opinions and beliefs. This study explores students' level of ego-involvement with informative and persuasive speech topics, examines possible ego-involvement predictors of…

  19. Passion and Preparation in the Basic Course: The Influence of Students' Ego-Involvement with Speech Topics and Preparation Time on Public-Speaking Grades

    Science.gov (United States)

    Mazer, Joseph P.; Titsworth, Scott

    2012-01-01

    Authors of basic public-speaking course textbooks frequently encourage students to select speech topics in which they have vested interest, care deeply about, and hold strong opinions and beliefs. This study explores students' level of ego-involvement with informative and persuasive speech topics, examines possible ego-involvement predictors of…

  20. A prepared speech in front of a pre-recorded audience: subjective, physiological, and neuroendocrine responses to the Leiden Public Speaking Task.

    Science.gov (United States)

    Westenberg, P Michiel; Bokhorst, Caroline L; Miers, Anne C; Sumter, Sindy R; Kallen, Victor L; van Pelt, Johannes; Blöte, Anke W

    2009-10-01

    This study describes a new public speaking protocol for youth. The main question asked whether a speech prepared at home and given in front of a pre-recorded audience creates a condition of social-evaluative threat. Findings showed that, on average, this task elicits a moderate stress response in a community sample of 83 12- to 15-year-old adolescents. During the speech, participants reported feeling more nervous and having higher heart rate and sweatiness of the hands than at baseline or recovery. Likewise, physiological (heart rate and skin conductance) and neuroendocrine (cortisol) activity were higher during the speech than at baseline or recovery. Additionally, an anticipation effect was observed: baseline levels were higher than recovery levels for most variables. Taking the anticipation and speech response together, a substantial cortisol response was observed for 55% of participants. The findings indicate that the Leiden Public Speaking Task might be particularly suited to investigate individual differences in sensitivity to social-evaluative situations.

  1. Speech dynamics

    NARCIS (Netherlands)

    Pols, L.C.W.

    2011-01-01

    In order for speech to be informative and communicative, segmental and suprasegmental variation is mandatory. Only this leads to meaningful words and sentences. The building blocks are no stable entities put next to each other (like beads on a string or like printed text), but there are gradual tran

  2. Speech Correction in the Schools.

    Science.gov (United States)

    Eisenson, Jon; Ogilvie, Mardel

    An introduction to the problems and therapeutic needs of school age children whose speech requires remedial attention, the text is intended for both the classroom teacher and the speech correctionist. General considerations include classification and incidence of speech defects, speech correction services, the teacher as a speaker, the mechanism…

  3. Under-resourced speech recognition based on the speech manifold

    CSIR Research Space (South Africa)

    Sahraeian, R

    2015-09-01

    Full Text Available Conventional acoustic modeling involves estimating many parameters to effectively model feature distributions. The sparseness of speech and text data, however, degrades the reliability of the estimation process and makes speech recognition a...

  4. Hate speech

    Directory of Open Access Journals (Sweden)

    Anne Birgitta Nilsen

    2014-12-01

    Full Text Available The manifesto of the Norwegian terrorist Anders Behring Breivik is based on the “Eurabia” conspiracy theory. This theory is a key starting point for hate speech amongst many right-wing extremists in Europe, but also has ramifications beyond these environments. In brief, proponents of the Eurabia theory claim that Muslims are occupying Europe and destroying Western culture, with the assistance of the EU and European governments. By contrast, members of Al-Qaeda and other extreme Islamists promote the conspiracy theory “the Crusade” in their hate speech directed against the West. Proponents of the latter theory argue that the West is leading a crusade to eradicate Islam and Muslims, a crusade that is similarly facilitated by their governments. This article presents analyses of texts written by right-wing extremists and Muslim extremists in an effort to shed light on how hate speech promulgates conspiracy theories in order to spread hatred and intolerance.The aim of the article is to contribute to a more thorough understanding of hate speech’s nature by applying rhetorical analysis. Rhetorical analysis is chosen because it offers a means of understanding the persuasive power of speech. It is thus a suitable tool to describe how hate speech works to convince and persuade. The concepts from rhetorical theory used in this article are ethos, logos and pathos. The concept of ethos is used to pinpoint factors that contributed to Osama bin Laden's impact, namely factors that lent credibility to his promotion of the conspiracy theory of the Crusade. In particular, Bin Laden projected common sense, good morals and good will towards his audience. He seemed to have coherent and relevant arguments; he appeared to possess moral credibility; and his use of language demonstrated that he wanted the best for his audience.The concept of pathos is used to define hate speech, since hate speech targets its audience's emotions. In hate speech it is the

  5. Speech Compression for Noise-Corrupted Thai Expressive Speech

    Directory of Open Access Journals (Sweden)

    Suphattharachai Chomphan

    2011-01-01

    Full Text Available Problem statement: In speech communication, speech coding aims at preserving the speech quality with lower coding bitrate. When considering the communication environment, various types of noises deteriorates the speech quality. The expressive speech with different speaking styles may cause different speech quality with the same coding method. Approach: This research proposed a study of speech compression for noise-corrupted Thai expressive speech by using two coding methods of CS-ACELP and MP-CELP. The speech material included a hundredmale speech utterances and a hundred female speech utterances. Four speaking styles included enjoyable, sad, angry and reading styles. Five sentences of Thai speech were chosen. Three types of noises were included (train, car and air conditioner. Five levels of each type of noise were varied from 0-20 dB. The subjective test of mean opinion score was exploited in the evaluation process. Results: The experimental results showed that CS-ACELP gave the better speech quality than that of MP-CELP at all three bitrates of 6000, 8600-12600 bps. When considering the levels of noise, the 20-dB noise gave the best speech quality, while 0-dB noise gave the worst speech quality. When considering the speech gender, female speech gave the better results than that of male speech. When considering the types of noise, the air-conditioner noise gave the best speech quality, while the train noise gave the worst speech quality. Conclusion: From the study, it can be seen that coding methods, types of noise, levels of noise, speech gender influence on the coding speech quality.

  6. 嗓音障碍听感知评估汉语普通话朗读文本的设计%Designing A New Chinese Mandarin Version of Speech Sample Text for Perceptual Assessment of Voice Disorders

    Institute of Scientific and Technical Information of China (English)

    赵逸; 王伟; 张秋韵; 吕琨; 李孟; 陈世彩; 郑宏良

    2014-01-01

    目的:设计一篇可用于嗓音障碍听感知评估的汉语普通话朗读文本。方法自行设计一篇涵盖汉语普通话所有音素朗读文本《生日礼物》,采用Pearson’s相关系数和Spearman相关系数检验其与中国科学院声学研究所统计的汉语声母、韵母和声调构成比的相关性,并采用配对样本 t检验和Wilcoxon符号秩检验来检验两者间声母、韵母和声调构成比的差异性,使用易读性公式来检验该短文的易读性。结果朗读文本《生日礼物》含21个声母,38个韵母,4种声调,其声母、韵母和声调构成比与中国科学院声学研究所提出的声母、韵母及声调构成比高度相关,其中,声母构成比的Pearson’s相关系数和Spearman相关系数分别为0.802、0.808,均为 P<0.01;韵母构成比的Pearson’s相关系数和Spearman相关系数分别为0.908、0.775,均为 P<0.01;声调构成比的Pearson’s相关系数 r=0.967,P<0.01,Spearman相关系数 r>0.999,P<0.01。两者间声母、韵母和声调构成比差异无统计学意义(均为 P>0.2)。该短文适度年级值为2,易读性良好。结论短文《生日礼物》可作为嗓音听感知评估的汉语普通话朗读文本,其应用价值有待进一步临床验证。%Objective To design a new Chinese mandarin version of speech sample text for perceptual assessment .Meth-ods The speech sample text included all phonemes and limit number of words .The correlation analysis of the constituent rati-os of all phonemes was performed between the speech text and the ones in Chinese language reported by the institute of Acous-tics Chinese Academy of Sciences using Pearson's correlation coefficient and Spearman correlation coefficient .And their differ-ences were tested by paired sample t test and Wilcoxon signed rank test .Legibility was alsotested .Results Our speech sarrple text contained 21 vowel ,38 consonant

  7. Text-To-Visual Speech in Chinese Based on Data-Driven Approach%基于数据驱动方法的汉语文本-可视语音合成

    Institute of Scientific and Technical Information of China (English)

    王志明; 蔡莲红; 艾海舟

    2005-01-01

    Text-To-Visual speech (TTVS) synthesis by computer can increase the speech intelligibility and make the human-computer interaction interfaces more friendly. This paper describes a Chinese text-to-visual speech synthesis system based on data-driven (sample based) approach, which is realized by short video segments concatenation. An effective method to construct two visual confusion trees for Chinese initials and finals is developed. A co-articulation model based on visual distance and hardness factor is proposed, which can be used in the recording corpus sentence selection in analysis phase and the unit selection in synthesis phase. The obvious difference between bound ary images of the concatenation video segments is smoothed by image morphing technique. By combining with the acoustic Text-To-Speech (TTS) synthesis, a Chinese text-to-visual speech synthesis system is realized.%计算机文本-可视语音合成系统(TTVS)可以增强语音的可懂度,并使人机交互界面变得更为友好.给出一个基于数据驱动方法(基于样本方法)的汉语文本-可视语音合成系统,通过将小段视频拼接生成新的可视语音.给出一种构造汉语声韵母视觉混淆树的有效方法,并提出了一个基于视觉混淆树和硬度因子的协同发音模型,模型可用于分析阶段的语料库选取和合成阶段的基元选取.对于拼接边界处两帧图像的明显差别,采用图像变形技术进行平滑并.结合已有的文本-语音合成系统(TTS),实现了一个中文文本视觉语音合成系统.

  8. Priorities of Dialogic Speech Teaching Methodology at Higher Non-Linguistic School

    Directory of Open Access Journals (Sweden)

    Vida Asanavičienė

    2011-04-01

    Full Text Available The article deals with a number of relevant methodological issues. First of all, the author analyses psychological peculiarities of dialogic speech and states that the dialogue is the product of at least two persons. Therefore, in this view, dialogic speech, unlike monologic speech, happens impromptu and is not prepared in advance. Dialogic speech is mainly of situational character. The linguistic nature of dialogic speech, in the author’s opinion, lies in the process of exchanging replications, which are coherent in structural and functional character. The author classifies dialogue groups by the number of replications and communicative parameters. The basic goal of dialogic speech teaching is developing the abilities and skills which enable to exchange replications. The author distinguishes two basic stages of dialogic speech teaching: 1. Training of abilities to exchange replications during communicative exercises. 2. Development of skills by training the capability to perform exercises of creative nature during a group dialogue, conversation or debate.

  9. Preparation

    Directory of Open Access Journals (Sweden)

    M.M. Dardir

    2014-03-01

    Full Text Available Some hexanamide-mono and di-linoleniate esters were prepared by the reaction of linolenic acid and hexanamide (derived from the reaction of hexanoic acid and diethanolamine. The chemical structure for the newly prepared hexanamide-mono and di-linoleniate esters were elucidated using elemental analysis, (FTIR, H 1NMR and chemical ionization mass spectra (CI/Ms spectroscopic techniques. The results of the spectroscopic analysis indicated that they were prepared through the right method and they have high purity. The new prepared esters have high biodegradability and lower toxicity (environmentally friendly so they were evaluated as a synthetic-based mud (ester-based mud for oil-well drilling fluids. The evaluation included study of the rheological properties, filtration and thermal properties of the ester based-muds formulated with the newly prepared esters compared to the reference commercial synthetic-based mud.

  10. ALinguisticandStylisticAnalysisoftheChineseTranslationofObama’sSpeechattheFirstMeetingoftheStrategicandEconomicDialoguebetweentheUnitedStatesandChina--ThroughtheLensofReiss’sTextTypeTheory

    Institute of Scientific and Technical Information of China (English)

    付端凌

    2014-01-01

    AccordingtoReiss’sTextTypetheory,akeypartofthefunctionalistapproachintranslationstudies,thesourcetextcanbeassignedtoatexttypeandtoagenre.Inmakingthisassignment,thetranslatorcandecideonthehierarchyofpostulateswhichhastobeobservedduringtarget-textproduction(Mona,2005).ThisessayintendstoconductalinguisticandstylisticanalysisoftheChinesetranslationofObama’sspeechtoexplorethegeneralapproachofthetranslatorifthereisone),bycomparingtherespectiveresultsofthetwoanalysesfromtheperspectiveofKatharinaReiss’sTextTypetheory.Indoingso,criticaljudgmentswillaccordinglybemadeastowhethersuchanapproachisjustifiableornot.

  11. Public Speech.

    Science.gov (United States)

    Green, Thomas F.

    1994-01-01

    Discusses the importance of public speech in society, noting the power of public speech to create a world and a public. The paper offers a theory of public speech, identifies types of public speech, and types of public speech fallacies. Two ways of speaking of the public and of public life are distinguished. (SM)

  12. Problems in Preparing for the English Impromptu Speech Contest: The Case of Yuanpei Institute of Science and Technology in Taiwan

    Science.gov (United States)

    Hsieh, Shu-min

    2006-01-01

    Entering an "English Impromptu Speech Contest" intimidates many students who do not have a good command of the English language. Some choose to give up before the contest date while others stand speechless on the stage. This paper identifies a range of problems confronted by contestants from my college, the Yuanpei Institute of Science…

  13. Preparation, Clinical Support, and Confidence of Speech-Language Therapists Managing Clients with a Tracheostomy in the UK

    Science.gov (United States)

    Ward, Elizabeth; Morgan, Tessa; McGowan, Sue; Spurgin, Ann-Louise; Solley, Maura

    2012-01-01

    Background: Literature regarding the education, training, clinical support and confidence of speech-language therapists (SLTs) working with patients with a tracheostomy is limited; however, it suggests that many clinicians have reduced clinical confidence when managing this complex population, many face role and team challenges practising in this…

  14. The stylistic idiosyncrasy of Almodóvar’s speech and its influence on the target text. Analysis of some translation strategies for rendering social dialects in the Polish version of the film "Volver" (2006

    Directory of Open Access Journals (Sweden)

    Leticia Santamaría Ciordia

    2012-01-01

    Full Text Available The identification of an original and idiosyncratic style in the work of the Spanish filmmaker Pedro Almodóvar provides the tone of the speech with a semantic function: linguistic variation and specific registers help to transmit the message and intention of the author. As a consequence of this, the possibilities for the translation being carried away in similar linguistic terms to those of the original text will be higher than in texts without linguistic peculiarities or stylistically unmarked. The purpose of this paper is to reflect on the influence exerted by the original author in the reception of his work abroad and, consequently, in its translation. In this sense, we will analyse the informative value of certain linguistic uses as well as the importance of enriching stylistically the subtitles in these connotative texts through the example of Almodóvar’s film "Volver" and its translation into Polish (2006.

  15. Application of Speech-to-Text in the Development of Mobile Software%语音-文本转换技术在手机软件开发中的应用

    Institute of Scientific and Technical Information of China (English)

    杨颖; 曹红兵; 吴方; 杨晴龙

    2016-01-01

    Speech-to-Text aims at converting the lexical content of the speaker’ s speech into the computer-readable text in-put, and making computer capable of “understanding” the human voice.Based on Android platform, the paper designed and de-veloped two mobile applications, namely order inquiry and voice notes, by the domestically advanced Speech-to-Text technology. The voice function of the commodity order inquiry provides a more convenient personalized operation .The main functions of the voice notes include single and multi-person cooperative recording, especially the latter function achieves the recording and sharing of multi-person voice communication.In conclusion, these two mobile applications have some practical values.%语音-文本转换旨在将说话人语音中的词汇内容转换为计算机可读的文本输入,使计算机具有能够“听懂”人类语音的能力。基于 Android 智能手机平台,利用国内先进的语音-文本转换技术设计并实现了两款手机软件———商品订单查询和手机语音笔记。商品订单查询的语音功能提供了更为便捷的个性化操作。手机语音笔记的主要功能包括单人录音和多人协作录音,尤其是多人协作录音实现了多人语音交流的记录和分享。这两款手机软件具有一定的实际应用价值。

  16. Speech Problems

    Science.gov (United States)

    ... of your treatment plan may include seeing a speech therapist , a person who is trained to treat speech disorders. How often you have to see the speech therapist will vary — you'll probably start out seeing ...

  17. Reaction time analysis of two types of motor preparation for speech articulation: action as a sequence of chunks.

    Science.gov (United States)

    Klapp, Stuart T

    2003-06-01

    Reaction time (RT) prior to speech articulation increased as a function of response complexity. The RT findings formed 2 patterns, each of which was a different Response Complexity x Paradigm (choice RT vs. simple RT) interaction. That result extends previous findings from manual button-pressing tasks (S. T. Klapp, 1995) to a different action modality. Two different types of response programming, INT and SEQ, are assumed in the interpretation. Whereas INT can be identified with response programming within a word, SEQ fits a different interpretation related to timing of onsets of speech units. A critical assumption is that a long response is represented as a sequence of chunks; that organization is subject to manipulation. New findings suggest some modifications of the previous theory.

  18. RESEARCH ON FRONT PART TEXT PROCESSING TECHNOLOGY FOR UYGHUR LANGUAGE SPEECH SYNTHESIS%维吾尔语语言合成系统前端文本处理技术研究

    Institute of Scientific and Technical Information of China (English)

    帕丽旦·木合塔尔; 吾守尔·斯拉木

    2012-01-01

    文本分析、特殊符号处理、韵律短语划分是语音合成系统前端处理模块的核心内容.根据维吾尔语的语言和语音特征,提出一套准确可行的前端文本处理策略,很好地分析与描述了维语语言层的信息,并将之映射到语音层上.研究的完成为进一步实现维吾尔语语音合成系统打下坚实的基础.%Text analysis, special symbols processing and rhythm phrases division are the core content of the front part processing module of the speech synthesis systems. This paper proposes a set of accurate and feasible front part text processing strategy according to Uighur language features, makes a good analyses and description of the information of Uighur language layer and mapped the information layer to voice layer. The completion of this research establishes a solid foundation for the Uighur speech synthesis system further.

  19. Illustrated Speech Anatomy.

    Science.gov (United States)

    Shearer, William M.

    Written for students in the fields of speech correction and audiology, the text deals with the following: structures involved in respiration; the skeleton and the processes of inhalation and exhalation; phonation and pitch, the larynx, and esophageal speech; muscles involved in articulation; muscles involved in resonance; and the anatomy of the…

  20. A study of the concession speech by President Goodluck Jonathan ...

    African Journals Online (AJOL)

    A study of the concession speech by President Goodluck Jonathan. ... It adopts the speech act theory in the classification of the illocutionary acts which are contained in the speech. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  1. SPECTRAL METHODS IN POLISH EMOTIONAL SPEECH RECOGNITION

    Directory of Open Access Journals (Sweden)

    Paweł Powroźnik

    2016-12-01

    Full Text Available In this article the issue of emotion recognition based on Polish emotional speech signal analysis was presented. The Polish database of emotional speech, prepared and shared by the Medical Electronics Division of the Lodz University of Technology, has been used for research. Speech signal has been processed by Artificial Neural Networks (ANN. The inputs for ANN were information obtained from signal spectrogram. Researches were conducted for three different spectrogram divisions. The ANN consists of four layers but the number of neurons in each layer depends of spectrogram division. Conducted researches focused on six emotional states: a neutral state, sadness, joy, anger, fear and boredom. The averange effectiveness of emotions recognition was about 80%.

  2. SPEECH DISORDERS ENCOUNTERED DURING SPEECH THERAPY AND THERAPY TECHNIQUES

    Directory of Open Access Journals (Sweden)

    İlhan ERDEM

    2013-06-01

    Full Text Available Speech which is a physical and mental process, agreed signs and sounds to create a sense of mind to the message that change . Process to identify the sounds of speech it is essential to know the structure and function of various organs which allows to happen the conversation. Speech is a physical and mental process so many factors can lead to speech disorders. Speech disorder can be about language acquisitions as well as it can be caused medical and psychological many factors. Disordered speech, language, medical and psychological conditions as well as acquisitions also be caused by many factors. Speaking, is the collective work of many organs, such as an orchestra. Mental dimension of the speech disorder which is a very complex skill so it must be found which of these obstacles inhibit conversation. Speech disorder is a defect in speech flow, rhythm, tizliğinde, beats, the composition and vocalization. In this study, speech disorders such as articulation disorders, stuttering, aphasia, dysarthria, a local dialect speech, , language and lip-laziness, rapid speech peech defects in a term of language skills. This causes of speech disorders were investigated and presented suggestions for remedy was discussed.

  3. Denial Denied: Freedom of Speech

    Directory of Open Access Journals (Sweden)

    Glen Newey

    2009-12-01

    Full Text Available Free speech is a widely held principle. This is in some ways surprising, since formal and informal censorship of speech is widespread, and rather different issues seem to arise depending on whether the censorship concerns who speaks, what content is spoken or how it is spoken. I argue that despite these facts, free speech can indeed be seen as a unitary principle. On my analysis, the core of the free speech principle is the denial of the denial of speech, whether to a speaker, to a proposition, or to a mode of expression. Underlying free speech is the principle of freedom of association, according to which speech is both a precondition of future association (e.g. as a medium for negotiation and a mode of association in its own right. I conclude by applying this account briefly to two contentious issues: hate speech and pornography.

  4. F. M. Dostoevsky, V. P. Meschersky Correspondence (1872–1880. Text is prepared by E. Vial, V. Zinkova

    Directory of Open Access Journals (Sweden)

    Elena Vial

    2017-04-01

    Full Text Available Correspondence between F. M. Dostoevsky and V. P. Meshchersky covers the period of nine years from 1872 till 1880. Their letters contain valuable biographical and literary information on the history of Russian literature and journalism of the 1870's.

  5. Preparing for the Worst--and Striving for the Best: Training University Employees To Respond Clearly, Constructively, and Constitutionally to Hateful Speech on Campus.

    Science.gov (United States)

    Calleros, Charles R.

    1997-01-01

    University administrators are challenged to develop policies that protect the educational atmosphere for marginalized groups while also protecting freedom of speech. The appropriate administrative response to outrageous or hateful speech on campus may be constructive speech and action, not discipline of speakers. This article summarizes relevant…

  6. Spectral Psychoanalysis of Speech under Strain | Sharma ...

    African Journals Online (AJOL)

    Spectral Psychoanalysis of Speech under Strain. ... Different voice features from the speech signal to be influenced by strain are: loudness, fundamental frequency, jitter, zero-crossing rate, ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  7. INTEGRATING MACHINE TRANSLATION AND SPEECH SYNTHESIS COMPONENT FOR ENGLISH TO DRAVIDIAN LANGUAGE SPEECH TO SPEECH TRANSLATION SYSTEM

    Directory of Open Access Journals (Sweden)

    J. SANGEETHA

    2015-02-01

    Full Text Available This paper provides an interface between the machine translation and speech synthesis system for converting English speech to Tamil text in English to Tamil speech to speech translation system. The speech translation system consists of three modules: automatic speech recognition, machine translation and text to speech synthesis. Many procedures for incorporation of speech recognition and machine translation have been projected. Still speech synthesis system has not yet been measured. In this paper, we focus on integration of machine translation and speech synthesis, and report a subjective evaluation to investigate the impact of speech synthesis, machine translation and the integration of machine translation and speech synthesis components. Here we implement a hybrid machine translation (combination of rule based and statistical machine translation and concatenative syllable based speech synthesis technique. In order to retain the naturalness and intelligibility of synthesized speech Auto Associative Neural Network (AANN prosody prediction is used in this work. The results of this system investigation demonstrate that the naturalness and intelligibility of the synthesized speech are strongly influenced by the fluency and correctness of the translated text.

  8. Speech Matters

    DEFF Research Database (Denmark)

    Hasse Jørgensen, Stina

    2011-01-01

    About Speech Matters - Katarina Gregos, the Greek curator's exhibition at the Danish Pavillion, the Venice Biannual 2011.......About Speech Matters - Katarina Gregos, the Greek curator's exhibition at the Danish Pavillion, the Venice Biannual 2011....

  9. Speech Development

    Science.gov (United States)

    ... The speech-language pathologist should consistently assess your child’s speech and language development, as well as screen for hearing problems (with ... and caregivers play a vital role in a child’s speech and language development. It is important that you talk to your ...

  10. Interventions for Speech Sound Disorders in Children

    Science.gov (United States)

    Williams, A. Lynn, Ed.; McLeod, Sharynne, Ed.; McCauley, Rebecca J., Ed.

    2010-01-01

    With detailed discussion and invaluable video footage of 23 treatment interventions for speech sound disorders (SSDs) in children, this textbook and DVD set should be part of every speech-language pathologist's professional preparation. Focusing on children with functional or motor-based speech disorders from early childhood through the early…

  11. DIFFICULTIES EMERGING IN THE PROCESS OF TEACHING RUSSIAN STUDENTS TO MAKE A SPEECH IN JAPANESE AT THE STAGE OF THE SPEECH ORAL PRESENTATION

    Directory of Open Access Journals (Sweden)

    N. L. Maksimenko

    2014-09-01

    Full Text Available Acquiring skills of making a speech in the Japanese language as a kind of prepared oral monologue speech implies not only an ability to collect and select material, to find arguments for issue revealing, to expound the topic logically and to arrange the utterance in accordance with the main idea, but also to arrange all the material and give the compositional ending of the text, to intone the text correctly, to answer questions etc. The article gives a detailed description of difficulties that emerge at the moment of oral speech presentation made by Russian students in the Japanese language at the initial stage of education. These difficulties include the following: the difficulty of report articulation and intonation with the tempo which is normal for Japanese monologue passages; the difficulty of contact establishment with the audience during the delivery of speech in the Japanese language; the difficulty of selecting polite expressions used in answering questions in Japanese.

  12. A Survey on Speech Enhancement Methodologies

    Directory of Open Access Journals (Sweden)

    Ravi Kumar. K

    2016-12-01

    Full Text Available Speech enhancement is a technique which processes the noisy speech signal. The aim of speech enhancement is to improve the perceived quality of speech and/or to improve its intelligibility. Due to its vast applications in mobile telephony, VOIP, hearing aids, Skype and speaker recognition, the challenges in speech enhancement have grown over the years. It is more challenging to suppress back ground noise that effects human communication in noisy environments like airports, road works, traffic, and cars. The objective of this survey paper is to outline the single channel speech enhancement methodologies used for enhancing the speech signal which is corrupted with additive background noise and also discuss the challenges and opportunities of single channel speech enhancement. This paper mainly focuses on transform domain techniques and supervised (NMF, HMM speech enhancement techniques. This paper gives frame work for developments in speech enhancement methodologies

  13. Speech Recognition on Mobile Devices

    DEFF Research Database (Denmark)

    Tan, Zheng-Hua; Lindberg, Børge

    2010-01-01

    The enthusiasm of deploying automatic speech recognition (ASR) on mobile devices is driven both by remarkable advances in ASR technology and by the demand for efficient user interfaces on such devices as mobile phones and personal digital assistants (PDAs). This chapter presents an overview of ASR...... in the mobile context covering motivations, challenges, fundamental techniques and applications. Three ASR architectures are introduced: embedded speech recognition, distributed speech recognition and network speech recognition. Their pros and cons and implementation issues are discussed. Applications within...... command and control, text entry and search are presented with an emphasis on mobile text entry....

  14. Full text

    African Journals Online (AJOL)

    IndexCopernicus Portal System

    Regenerative Modeling, McGowan Institute for Regenerative Medicine, ... MODS: Multiple Organ Dysfunction Syndrome; NO: Nitric Oxide; PAMP: ... motor, behavioural or speech-language ..... The model that best reflects the clinical setting ... allows for assessment of compensatory ..... The epidemiology and modern.

  15. Emotion Detection from Text

    CERN Document Server

    Shivhare, Shiv Naresh

    2012-01-01

    Emotion can be expressed in many ways that can be seen such as facial expression and gestures, speech and by written text. Emotion Detection in text documents is essentially a content - based classification problem involving concepts from the domains of Natural Language Processing as well as Machine Learning. In this paper emotion recognition based on textual data and the techniques used in emotion detection are discussed.

  16. Speech recognition in university classrooms

    OpenAIRE

    Wald, Mike; Bain, Keith; Basson, Sara H

    2002-01-01

    The LIBERATED LEARNING PROJECT (LLP) is an applied research project studying two core questions: 1) Can speech recognition (SR) technology successfully digitize lectures to display spoken words as text in university classrooms? 2) Can speech recognition technology be used successfully as an alternative to traditional classroom notetaking for persons with disabilities? This paper addresses these intriguing questions and explores the underlying complex relationship between speech recognition te...

  17. Noise Reduction in Car Speech

    Directory of Open Access Journals (Sweden)

    V. Bolom

    2009-01-01

    Full Text Available This paper presents properties of chosen multichannel algorithms for speech enhancement in a noisy environment. These methods are suitable for hands-free communication in a car cabin. Criteria for evaluation of these systems are also presented. The criteria consider both the level of noise suppression and the level of speech distortion. The performance of multichannel algorithms is investigated for a mixed model of speech signals and car noise and for real signals recorded in a car. 

  18. Speech disfluencies in Parkinson’s disease

    Directory of Open Access Journals (Sweden)

    Paweł J. Półrola

    2016-01-01

    Full Text Available Introduction : Even though speech disfluency is listed in the clinical description of Parkinson’s disease (PD, its nature, intensity, symptomatology, and the effect on verbal communication have not hitherto been defined. Aim of the research: The research paper presents the results of studies aimed at the description of speech disfluencies in PD and the influence on verbal communication. Material and methods : The tests involved 10 patients from 54 to 72 years of age with documented PD, responsive to L-dopa preparations. The principal method of the study was based on the linguistic analysis of the utterances produced by the people with PD. Results: The intensity of the speech disfluency observed in the utterances of persons with PD ranged from 6.6% to 23.0%, so it was significantly higher than that which is assumed as acceptable (3–5%; the speaking rate of the examined persons ranged from 0.7 syllabes (syl./s to 4.0 syl./s, and only 2 examined persons spoke with a rate considered to be correct (4–6 syl./s. This demonstrates that speech disfluency is a communication barrier in PD. Conclusions : The absence of differentiation in the speech disfluency (SD severity between different types of verbal utterances (difference not statistically significant and a specified hierarchy of SD symptoms indicate that the speech disfluency in PD has an essentially organic background and is generated by cognitive, linguistic, and motor deficits resulting from the damage to the central nervous system. This is also confirmed by the established hierarchy of utterances with respect to the SD intensity, not excluding the simultaneous participation of the emotional factor.

  19. A Blueprint for a Comprehensive Australian English Auditory-Visual Speech Corpus

    NARCIS (Netherlands)

    Burnham, D.; Ambikairajah, E.; Arciuli, J.; Bennamoun, M.; Best, C.T.; Bird, S.; Butcher, A.R.; Cassidy, S.; Chetty, G.; Cox, F.M.; Cutler, A.; Dale, R.; Epps, J.R.; Fletcher, J.M.; Goecke, R.; Grayden, D.B.; Hajek, J.T.; Ingram, J.C.; Ishihara, S.; Kemp, N.; Kinoshita, Y.; Kuratate, T.; Lewis, T.W.; Loakes, D.E.; Onslow, M.; Powers, D.M.; Rose, P.; Togneri, R.; Tran, D.; Wagner, M.

    2009-01-01

    Large auditory-visual (AV) speech corpora are the grist of modern research in speech science, but no such corpus exists for Australian English. This is unfortunate, for speech science is the brains behind speech technology and applications such as text-to-speech (TTS) synthesis, automatic speech rec

  20. Sensorimotor Interactions in Speech Learning

    Directory of Open Access Journals (Sweden)

    Douglas M Shiller

    2011-10-01

    Full Text Available Auditory input is essential for normal speech development and plays a key role in speech production throughout the life span. In traditional models, auditory input plays two critical roles: 1 establishing the acoustic correlates of speech sounds that serve, in part, as the targets of speech production, and 2 as a source of feedback about a talker's own speech outcomes. This talk will focus on both of these roles, describing a series of studies that examine the capacity of children and adults to adapt to real-time manipulations of auditory feedback during speech production. In one study, we examined sensory and motor adaptation to a manipulation of auditory feedback during production of the fricative “s”. In contrast to prior accounts, adaptive changes were observed not only in speech motor output but also in subjects' perception of the sound. In a second study, speech adaptation was examined following a period of auditory–perceptual training targeting the perception of vowels. The perceptual training was found to systematically improve subjects' motor adaptation response to altered auditory feedback during speech production. The results of both studies support the idea that perceptual and motor processes are tightly coupled in speech production learning, and that the degree and nature of this coupling may change with development.

  1. Speech Indexing

    NARCIS (Netherlands)

    Ordelman, R.J.F.; Jong, de F.M.G.; Leeuwen, van D.A.; Blanken, H.M.; de Vries, A.P.; Blok, H.E.; Feng, L.

    2007-01-01

    This chapter will focus on the automatic extraction of information from the speech in multimedia documents. This approach is often referred to as speech indexing and it can be regarded as a subfield of audio indexing that also incorporates for example the analysis of music and sounds. If the objecti

  2. Plowing Speech

    OpenAIRE

    Zla ba sgrol ma

    2009-01-01

    This file contains a plowing speech and a discussion about the speech This collection presents forty-nine audio files including: several folk song genres; folktales and; local history from the Sman shad Valley of Sde dge county World Oral Literature Project

  3. Speech coding

    Energy Technology Data Exchange (ETDEWEB)

    Ravishankar, C., Hughes Network Systems, Germantown, MD

    1998-05-08

    Speech is the predominant means of communication between human beings and since the invention of the telephone by Alexander Graham Bell in 1876, speech services have remained to be the core service in almost all telecommunication systems. Original analog methods of telephony had the disadvantage of speech signal getting corrupted by noise, cross-talk and distortion Long haul transmissions which use repeaters to compensate for the loss in signal strength on transmission links also increase the associated noise and distortion. On the other hand digital transmission is relatively immune to noise, cross-talk and distortion primarily because of the capability to faithfully regenerate digital signal at each repeater purely based on a binary decision. Hence end-to-end performance of the digital link essentially becomes independent of the length and operating frequency bands of the link Hence from a transmission point of view digital transmission has been the preferred approach due to its higher immunity to noise. The need to carry digital speech became extremely important from a service provision point of view as well. Modem requirements have introduced the need for robust, flexible and secure services that can carry a multitude of signal types (such as voice, data and video) without a fundamental change in infrastructure. Such a requirement could not have been easily met without the advent of digital transmission systems, thereby requiring speech to be coded digitally. The term Speech Coding is often referred to techniques that represent or code speech signals either directly as a waveform or as a set of parameters by analyzing the speech signal. In either case, the codes are transmitted to the distant end where speech is reconstructed or synthesized using the received set of codes. A more generic term that is applicable to these techniques that is often interchangeably used with speech coding is the term voice coding. This term is more generic in the sense that the

  4. Tactile Modulation of Emotional Speech Samples

    Directory of Open Access Journals (Sweden)

    Katri Salminen

    2012-01-01

    Full Text Available Traditionally only speech communicates emotions via mobile phone. However, in daily communication the sense of touch mediates emotional information during conversation. The present aim was to study if tactile stimulation affects emotional ratings of speech when measured with scales of pleasantness, arousal, approachability, and dominance. In the Experiment 1 participants rated speech-only and speech-tactile stimuli. The tactile signal mimicked the amplitude changes of the speech. In the Experiment 2 the aim was to study whether the way the tactile signal was produced affected the ratings. The tactile signal either mimicked the amplitude changes of the speech sample in question, or the amplitude changes of another speech sample. Also, concurrent static vibration was included. The results showed that the speech-tactile stimuli were rated as more arousing and dominant than the speech-only stimuli. The speech-only stimuli were rated as more approachable than the speech-tactile stimuli, but only in the Experiment 1. Variations in tactile stimulation also affected the ratings. When the tactile stimulation was static vibration the speech-tactile stimuli were rated as more arousing than when the concurrent tactile stimulation was mimicking speech samples. The results suggest that tactile stimulation offers new ways of modulating and enriching the interpretation of speech.

  5. Effect of speech rate variation on acoustic phone stability in Afrikaans speech recognition

    CSIR Research Space (South Africa)

    Badenhorst, JAC

    2007-11-01

    Full Text Available The authors analyse the effect of speech rate variation on Afrikaans phone stability from an acoustic perspective. Specifically they introduce two techniques for the acoustic analysis of speech rate variation, apply these techniques to an Afrikaans...

  6. Statistical Parametric Speech Synthesis of Malay Language using Found Training Data

    Directory of Open Access Journals (Sweden)

    Lau Chee Yong

    2014-06-01

    Full Text Available The preparation of training data for statistical parametric speech synthesis can be sophisticated. To ensure the good quality of synthetic speech, high quality low noise recording must be prepared. The preparation of recording script can be also tremendous from words collection, words selection and sentences design. It requires tremendous human effort and takes a lot of time. In this study, we used alternative free source of recording and text such as audio-book, clean speech and so on as the training data. Some of the free source can provide high quality recording with low noise which is suitable to become training data. Statistical parametric speech synthesis method applying Hidden Markov Model (HMM has been used. To test the reliability of synthetic speech, perceptual test has been conducted. The result of naturalness test is fairly reasonable. The intelligibility test showed encouraging result. The Word Error Rate (WER for normal synthetic sentences is below 15% while for Semantically Unpredictable Sentences (SUS is averagely in 30%. In short, using free and ready source as training data can leverage the process of preparing training data while obtaining motivating synthetic result.

  7. Neural pathways for visual speech perception

    Directory of Open Access Journals (Sweden)

    Lynne E Bernstein

    2014-12-01

    Full Text Available This paper examines the questions, what levels of speech can be perceived visually, and how is visual speech represented by the brain? Review of the literature leads to the conclusions that every level of psycholinguistic speech structure (i.e., phonetic features, phonemes, syllables, words, and prosody can be perceived visually, although individuals differ in their abilities to do so; and that there are visual modality-specific representations of speech qua speech in higher-level vision brain areas. That is, the visual system represents the modal patterns of visual speech. The suggestion that the auditory speech pathway receives and represents visual speech is examined in light of neuroimaging evidence on the auditory speech pathways. We outline the generally agreed-upon organization of the visual ventral and dorsal pathways and examine several types of visual processing that might be related to speech through those pathways, specifically, face and body, orthography, and sign language processing. In this context, we examine the visual speech processing literature, which reveals widespread diverse patterns activity in posterior temporal cortices in response to visual speech stimuli. We outline a model of the visual and auditory speech pathways and make several suggestions: (1 The visual perception of speech relies on visual pathway representations of speech qua speech. (2 A proposed site of these representations, the temporal visual speech area (TVSA has been demonstrated in posterior temporal cortex, ventral and posterior to multisensory posterior superior temporal sulcus (pSTS. (3 Given that visual speech has dynamic and configural features, its representations in feedforward visual pathways are expected to integrate these features, possibly in TVSA.

  8. Speech Genres in Writing Cognitive Artifacts.

    Science.gov (United States)

    Shambaugh, R. Neal

    This paper reports on the analysis of an instructional text on the basis of M. Bakhtin's (1986) notion of speech genres, which is used to theorize the different influences on the writing of an instructional text. Speech genres are used to reveal the multiple voices inherent in any text: the writer's, the reader's, and the text's. The…

  9. A NOVEL APPROACH TO STUTTERED SPEECH CORRECTION

    Directory of Open Access Journals (Sweden)

    Alim Sabur Ajibola

    2016-06-01

    Full Text Available Stuttered speech is a dysfluency rich speech, more prevalent in males than females. It has been associated with insufficient air pressure or poor articulation, even though the root causes are more complex. The primary features include prolonged speech and repetitive speech, while some of its secondary features include, anxiety, fear, and shame. This study used LPC analysis and synthesis algorithms to reconstruct the stuttered speech. The results were evaluated using cepstral distance, Itakura-Saito distance, mean square error, and likelihood ratio. These measures implied perfect speech reconstruction quality. ASR was used for further testing, and the results showed that all the reconstructed speech samples were perfectly recognized while only three samples of the original speech were perfectly recognized.

  10. Speech Enhancement with Natural Sounding Residual Noise Based on Connected Time-Frequency Speech Presence Regions

    Directory of Open Access Journals (Sweden)

    Sørensen Karsten Vandborg

    2005-01-01

    Full Text Available We propose time-frequency domain methods for noise estimation and speech enhancement. A speech presence detection method is used to find connected time-frequency regions of speech presence. These regions are used by a noise estimation method and both the speech presence decisions and the noise estimate are used in the speech enhancement method. Different attenuation rules are applied to regions with and without speech presence to achieve enhanced speech with natural sounding attenuated background noise. The proposed speech enhancement method has a computational complexity, which makes it feasible for application in hearing aids. An informal listening test shows that the proposed speech enhancement method has significantly higher mean opinion scores than minimum mean-square error log-spectral amplitude (MMSE-LSA and decision-directed MMSE-LSA.

  11. An Approach to Intelligent Speech Production System

    Institute of Scientific and Technical Information of China (English)

    陈芳; 袁保宗

    1997-01-01

    In the paper an intelligent speech production system is established by using language information processing technology.The concept of bi-directional grammar is proposed in Chinese language information processing and a corresponding Chinese characteristic network is completed.Correct text can be generated through grammar parsing and some additional rules.According to the generated text the system generates speech which has good quality in naturalness and intelligibility using Chinese Text-to-Speech Conversion System.

  12. Current trends in multilingual speech processing

    Indian Academy of Sciences (India)

    Hervé Bourlard; John Dines; Mathew Magimai-Doss; Philip N Garner; David Imseng; Petr Motlicek; Hui Liang; Lakshmi Saheer; Fabio Valente

    2011-10-01

    In this paper, we describe recent work at Idiap Research Institute in the domain of multilingual speech processing and provide some insights into emerging challenges for the research community. Multilingual speech processing has been a topic of ongoing interest to the research community for many years and the field is now receiving renewed interest owing to two strong driving forces. Firstly, technical advances in speech recognition and synthesis are posing new challenges and opportunities to researchers. For example, discriminative features are seeing wide application by the speech recognition community, but additional issues arise when using such features in a multilingual setting. Another example is the apparent convergence of speech recognition and speech synthesis technologies in the form of statistical parametric methodologies. This convergence enables the investigation of new approaches to unified modelling for automatic speech recognition and text-to-speech synthesis (TTS) as well as cross-lingual speaker adaptation for TTS. The second driving force is the impetus being provided by both government and industry for technologies to help break down domestic and international language barriers, these also being barriers to the expansion of policy and commerce. Speech-to-speech and speech-to-text translation are thus emerging as key technologies at the heart of which lies multilingual speech processing.

  13. RECOGNISING SPEECH ACTS

    Directory of Open Access Journals (Sweden)

    Phyllis Kaburise

    2012-09-01

    Full Text Available Speech Act Theory (SAT, a theory in pragmatics, is an attempt to describe what happens during linguistic interactions. Inherent within SAT is the idea that language forms and intentions are relatively formulaic and that there is a direct correspondence between sentence forms (for example, in terms of structure and lexicon and the function or meaning of an utterance. The contention offered in this paper is that when such a correspondence does not exist, as in indirect speech utterances, this creates challenges for English second language speakers and may result in miscommunication. This arises because indirect speech acts allow speakers to employ various pragmatic devices such as inference, implicature, presuppositions and context clues to transmit their messages. Such devices, operating within the non-literal level of language competence, may pose challenges for ESL learners.

  14. Contesting the existence of male speech stereotypes within the ...

    African Journals Online (AJOL)

    Contesting the existence of male speech stereotypes within the communicative context of nursing: a case of Bonda mission ... speech stereotypes to the communicative context of nursing. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  15. Constructing an inclusive speech community from two mutually ...

    African Journals Online (AJOL)

    Constructing an inclusive speech community from two mutually excluding ones: The third ... which meaningfully includes all its speakers as members, referred to as a “speech community”. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  16. A Method for Wavefront Curvature Ranging of Speech Sources ...

    African Journals Online (AJOL)

    A Method for Wavefront Curvature Ranging of Speech Sources. ... A new approach for estimating the location of a speech source in a reverberant environment is presented. The approach ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  17. Perverts and sodomites: homophobia as hate speech in Africa | Reddy

    African Journals Online (AJOL)

    Perverts and sodomites: homophobia as hate speech in Africa. ... African context by locating the issue of homophobia and hate speech primarily within the context of language and gender. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  18. Bilateral, posterior parietal polymicrogyria as part of speech therapy ...

    African Journals Online (AJOL)

    Bilateral, posterior parietal polymicrogyria as part of speech therapy work-up. ... units to make the diagnosis of bilateral posterior parietal polymicrogyria in a child with speech pathology. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  19. Discourse, Statement and Speech Act

    Directory of Open Access Journals (Sweden)

    Елена Александровна Красина

    2016-12-01

    Full Text Available Being a component of socio-cultural interaction discourse constitutes a sophisticated cohesion of language form, meaning and performance, i.e. communicative event or act. Cohesion with event and performance let us treat discourse as a certain lifeform, appealing both to communicative interaction and pragmatic environment using the methodology of studies of E. Benveniste, M. Foucault, I. Kecskes, J.R. Searle et al. In linguistics and other fields of humanitarian knowledge the notion of discourse facilitates the integration of studies in humanities. Principles of integration, incorporation into broad humanitarian context reveal some topics of discourse-speech act-utterance interaction which leads to substantive solutions of a number of linguistic topics, in particular, that of an utterance. Logicians determine utterance through proposition; linguists - through sentence, while speech act theory does it by means of illocutionary act. Integrated in a discourse or its part, utterance makes up their integral constituents although not unique ones. In relation to speech acts, utterance happens to be the unique definitional domain synchronically modelling and denoting speech act by means of propositional content. The goal of the research is to show the conditions of interaction and correlation of discourse, speech act and utterance as linguistic constructions, reveal some similarities and differences of their characteristics and prove the importance of the constructive role of utterance as a minimal unit of speech production. Discourse-speech act-utterance correlation supports the utterance role of a discrete unit within syntactic continuum, facing both language and speech: still, it belongs exclusively neither to language nor speech, but specifies their interaction in course of speech activity exposing simultaneously its nature of an ‘atom of discourse’ and creating the definitional domain of a speech act.

  20. Speech Compression Using Multecirculerletet Transform

    Directory of Open Access Journals (Sweden)

    Sulaiman Murtadha

    2012-01-01

    Full Text Available Compressing the speech reduces the data storage requirements, leading to reducing the time of transmitting the digitized speech over long-haul links like internet. To obtain best performance in speech compression, wavelet transforms require filters that combine a number of desirable properties, such as orthogonality and symmetry.The MCT bases functions are derived from GHM bases function using 2D linear convolution .The fast computation algorithm methods introduced here added desirable features to the current transform. We further assess the performance of the MCT in speech compression application. This paper discusses the effect of using DWT and MCT (one and two dimension on speech compression. DWT and MCT performances in terms of compression ratio (CR, mean square error (MSE and peak signal to noise ratio (PSNR are assessed. Computer simulation results indicate that the two dimensions MCT offer a better compression ratio, MSE and PSNR than DWT.

  1. Text Maps: Helping Students Navigate Informational Texts.

    Science.gov (United States)

    Spencer, Brenda H.

    2003-01-01

    Notes that a text map is an instructional approach designed to help students gain fluency in reading content area materials. Discusses how the goal is to teach students about the important features of the material and how the maps can be used to build new understandings. Presents the procedures for preparing and using a text map. (SG)

  2. Speech Evaluation with Special Focus on Children Suffering from Apraxia of Speech

    Directory of Open Access Journals (Sweden)

    Manasi Dixit

    2013-07-01

    Full Text Available Speech disorders are very complicated in individuals suffering from Apraxia of Speech-AOS. In this paper ,the pathological cases of speech disabled children affected with AOS are analyzed. The speech signalsamples of childrenSpeech disorders are very complicated in individuals suffering from Apraxia of Speech-AOS. In this paper ,the pathological cases of speech disabled children affected with AOS are analyzed. The speech signalsamples of children of age between three to eight years are considered for the present study. These speechsignals are digitized and enhanced using the using the Speech Pause Index, Jitter,Skew ,Kurtosis analysisThis analysis is conducted on speech data samples which are concerned with both place of articulation andmanner of articulation. The speech disability of pathological subjects was estimated using results of aboveanalysis. of age between three to eight years are considered for the present study. These speechsignals are digitized and enhanced using the using the Speech Pause Index, Jitter,Skew ,Kurtosis analysisThis analysis is conducted on speech data samples which are concerned with both place of articulation andmanner of articulation. The speech disability of pathological subjects was estimated using results of aboveanalysis.

  3. Detection and Separation of Speech Event Using Audio and Video Information Fusion and Its Application to Robust Speech Interface

    Directory of Open Access Journals (Sweden)

    Futoshi Asano

    2004-09-01

    Full Text Available A method of detecting speech events in a multiple-sound-source condition using audio and video information is proposed. For detecting speech events, sound localization using a microphone array and human tracking by stereo vision is combined by a Bayesian network. From the inference results of the Bayesian network, information on the time and location of speech events can be known. The information on the detected speech events is then utilized in the robust speech interface. A maximum likelihood adaptive beamformer is employed as a preprocessor of the speech recognizer to separate the speech signal from environmental noise. The coefficients of the beamformer are kept updated based on the information of the speech events. The information on the speech events is also used by the speech recognizer for extracting the speech segment.

  4. 77 FR 43538 - Misuse of Internet Protocol (IP) Relay Service; Telecommunications Relay Services and Speech-to...

    Science.gov (United States)

    2012-07-25

    ... and Speech-to-Speech Services for Individuals With Hearing and Speech Disabilities AGENCY: Federal... Services and Speech-to-Speech Services for Individuals with Hearing and Speech Disabilities, First Report... Docket Nos. 12-38 and 03-123. The full text of document FCC 12-71 will be available for public inspection...

  5. Perceived Speech Quality Estimation Using DTW Algorithm

    Directory of Open Access Journals (Sweden)

    S. Arsenovski

    2009-06-01

    Full Text Available In this paper a method for speech quality estimation is evaluated by simulating the transfer of speech over packet switched and mobile networks. The proposed system uses Dynamic Time Warping algorithm for test and received speech comparison. Several tests have been made on a test speech sample of a single speaker with simulated packet (frame loss effects on the perceived speech. The achieved results have been compared with measured PESQ values on the used transmission channel and their correlation has been observed.

  6. SPEECH MANUAL. RHETORIC CURRICULUM V, STUDENT VERSION.

    Science.gov (United States)

    KITZHABER, ALBERT R.

    THIS MANUAL IS A REFERENCE AID FOR 11TH-GRADE STUDENTS PREPARING SPEAKING ASSIGNMENTS. CHAPTER 1, "THE PHYSIOLOGY OF SPEECH," CONTAINS INFORMATION ON THE SPEECH ORGANS AND THEIR FUNCTIONS IN THE PRODUCTION OF SOUNDS. THE MAIN POINTS OF "ROBERT'S RULES OF ORDER" ARE OUTLINED IN CHAPTER 2. CHAPTER 3 GIVES ATTENTION TO OUTLINING…

  7. Speech enhancement

    CERN Document Server

    Benesty, Jacob; Chen, Jingdong

    2006-01-01

    We live in a noisy world! In all applications (telecommunications, hands-free communications, recording, human-machine interfaces, etc.) that require at least one microphone, the signal of interest is usually contaminated by noise and reverberation. As a result, the microphone signal has to be ""cleaned"" with digital signal processing tools before it is played out, transmitted, or stored.This book is about speech enhancement. Different well-known and state-of-the-art methods for noise reduction, with one or multiple microphones, are discussed. By speech enhancement, we mean not only noise red

  8. Speech Intelligibility

    Science.gov (United States)

    Brand, Thomas

    Speech intelligibility (SI) is important for different fields of research, engineering and diagnostics in order to quantify very different phenomena like the quality of recordings, communication and playback devices, the reverberation of auditoria, characteristics of hearing impairment, benefit using hearing aids or combinations of these things.

  9. Text Mining.

    Science.gov (United States)

    Trybula, Walter J.

    1999-01-01

    Reviews the state of research in text mining, focusing on newer developments. The intent is to describe the disparate investigations currently included under the term text mining and provide a cohesive structure for these efforts. A summary of research identifies key organizations responsible for pushing the development of text mining. A section…

  10. Text Mining.

    Science.gov (United States)

    Trybula, Walter J.

    1999-01-01

    Reviews the state of research in text mining, focusing on newer developments. The intent is to describe the disparate investigations currently included under the term text mining and provide a cohesive structure for these efforts. A summary of research identifies key organizations responsible for pushing the development of text mining. A section…

  11. Collecting and evaluating speech recognition corpora for 11 South African languages

    CSIR Research Space (South Africa)

    Badenhorst, J

    2011-08-01

    Full Text Available The authors describe the Lwazi corpus for automatic speech recognition (ASR), a new telephone speech corpus which contains data from the eleven official languages of South Africa. Because of practical constraints, the amount of speech per language...

  12. Collecting and evaluating speech recognition corpora for nine Southern Bantu languages

    CSIR Research Space (South Africa)

    Badenhorst, JAC

    2009-03-01

    Full Text Available The authors describes the Lwazi corpus for automatic speech recognition (ASR), a new telephone speech corpus which includes data from nine Southern Bantu languages. Because of practical constraints, the amount of speech per language is relatively...

  13. Speech Segregation based on Pitch Track Correction and Music-Speech Classification

    Directory of Open Access Journals (Sweden)

    KIM, H.-G.

    2012-05-01

    Full Text Available A novel approach for pitch track correction and music-speech classification is proposed in order to improve the performance of the speech segregation system. The proposed pitch track correction method adjusts unreliable pitch estimates from adjacent reliable pitch streaks, in contrast to the previous approach using a single pitch streak which is the longest among the reliable pitch streaks in a sentence. The proposed music and speech classification method finds continuous pitch streaks of the mixture, and labels each streak as music-dominant or speech-dominant based on the observation that music pitch seldom changes in a short-time period whereas speech pitch fluctuates a lot. The speech segregation results for mixtures of speech and various competing sound sources demonstrated that the proposed methods are superior to the conventional method, especially for mixtures of speech and music signals.

  14. PCA-Based Speech Enhancement for Distorted Speech Recognition

    Directory of Open Access Journals (Sweden)

    Tetsuya Takiguchi

    2007-09-01

    Full Text Available We investigated a robust speech feature extraction method using kernel PCA (Principal Component Analysis for distorted speech recognition. Kernel PCA has been suggested for various image processing tasks requiring an image model, such as denoising, where a noise-free image is constructed from a noisy input image. Much research for robust speech feature extraction has been done, but it remains difficult to completely remove additive or convolution noise (distortion. The most commonly used noise-removal techniques are based on the spectraldomain operation, and then for speech recognition, the MFCC (Mel Frequency Cepstral Coefficient is computed, where DCT (Discrete Cosine Transform is applied to the mel-scale filter bank output. This paper describes a new PCA-based speech enhancement algorithm using kernel PCA instead of DCT, where the main speech element is projected onto low-order features, while the noise or distortion element is projected onto high-order features. Its effectiveness is confirmed by word recognition experiments on distorted speech.

  15. Basic Chad Arabic: Comprehension Texts.

    Science.gov (United States)

    Absi, Samir Abu; Sinaud, Andre

    This text, principally designed for use in a three-volume course on Chad Arabic, complements the pre-speech and active phases of the course in that it provides the answers to comprehension exercises students are required to complete during the course. The comprehension exercises require that students listen to an instructor or tape and write…

  16. Speech communications in noise

    Science.gov (United States)

    1984-07-01

    The physical characteristics of speech, the methods of speech masking measurement, and the effects of noise on speech communication are investigated. Topics include the speech signal and intelligibility, the effects of noise on intelligibility, the articulation index, and various devices for evaluating speech systems.

  17. Commencement Speech as a Hybrid Polydiscursive Practice

    Directory of Open Access Journals (Sweden)

    Светлана Викторовна Иванова

    2017-12-01

    Full Text Available Discourse and media communication researchers pay attention to the fact that popular discursive and communicative practices have a tendency to hybridization and convergence. Discourse which is understood as language in use is flexible. Consequently, it turns out that one and the same text can represent several types of discourses. A vivid example of this tendency is revealed in American commencement speech / commencement address / graduation speech. A commencement speech is a speech university graduates are addressed with which in compliance with the modern trend is delivered by outstanding media personalities (politicians, athletes, actors, etc.. The objective of this study is to define the specificity of the realization of polydiscursive practices within commencement speech. The research involves discursive, contextual, stylistic and definitive analyses. Methodologically the study is based on the discourse analysis theory, in particular the notion of a discursive practice as a verbalized social practice makes up the conceptual basis of the research. This research draws upon a hundred commencement speeches delivered by prominent representatives of American society since 1980s till now. In brief, commencement speech belongs to institutional discourse public speech embodies. Commencement speech institutional parameters are well represented in speeches delivered by people in power like American and university presidents. Nevertheless, as the results of the research indicate commencement speech institutional character is not its only feature. Conceptual information analysis enables to refer commencement speech to didactic discourse as it is aimed at teaching university graduates how to deal with challenges life is rich in. Discursive practices of personal discourse are also actively integrated into the commencement speech discourse. More than that, existential discursive practices also find their way into the discourse under study. Commencement

  18. Tone realisation in a Yoruba speech recognition corpus

    CSIR Research Space (South Africa)

    Van Niekerk, D

    2012-05-01

    Full Text Available The authors investigate the acoustic realisation of tone in short continuous utterances in Yoruba. Fundamental frequency contours are extracted for automatically aligned syllables from a speech corpus of 33 speakers collected for speech recognition...

  19. 新修辞框架下中外领导致辞文本的对比及对翻译的启示%Under the framework of the new rhetoric the comparison of Chinese and foreign leadership speech text and its implications for translation

    Institute of Scientific and Technical Information of China (English)

    肖丽

    2015-01-01

    领导致辞是一种带有互动性、针对性、权威性的讲话文体。一般言语凝练,主题鲜明,态度明确。领导者通过言语来表达思想,体现政策和措施。领导致辞可直接体现领导者的个人魅力和政策水平。中外领导致辞有共性有差异,本文将在新修辞的理论框架下,以受众为中心,从话语表现形式、诉求手段、美学手段等方面对中外领导致辞文本进行对比分析,探讨共性和差异,并总结其对汉译英翻译的启示。%The leadership speech is a kind of interaction and pertinence, with authoritative speech style. Generaly concise words, distinct themes, attitude is clear. Leaders use words to express ideas, policies and measures. Leadership speech can directly reflect the leader's personal charm and policy level. Leadership speech have common differences between China and foreign countries, this article wil under the new rhetoric theory framework, with the audience as the center, from the discourse forms, methods, aesthetic appeal means of contrast analysis of Chinese and foreign leaders addressed the text, discuss commonalities and differences, and summarizes its enlightenment to translation from Chinese into English translation.

  20. A Review on Speech Corpus Development for Automatic Speech Recognition in Indian Languages

    Directory of Open Access Journals (Sweden)

    Cini kurian

    2015-05-01

    Full Text Available Corpus development gained much attention due to recent statistics based natural language processing. It has new applications in Language Technology, linguistic research, language education and information exchange. Corpus based Language research has an innovative outlook which will discard the aged linguistic theories. Speech corpus is the essential resources for building a speech recognizer. One of the main challenges faced by speech scientist is the unavailability of these resources. Very fewer efforts have been made in Indian languages to make these resources available to public compared to English. In this paper we review the efforts made in Indian languages for developing speech corpus for automatic speech recognition.

  1. Reflection and Optimization of Primary English Teachers’Speech Acts Based on Speech Act Theory

    Institute of Scientific and Technical Information of China (English)

    HU Qi-hai

    2015-01-01

    The primary English teacher's speech acts have major impact on foreign language teaching and learning in primary school. Application of teacher,s speech acts in the classroom is actually a kind of selective process. From the perspective of Speech Act Theory, primary English teachers can optimize their speech acts with the strategies of activating the greetings with proper con⁃text information, standardizing teacher talk, choosing suitable questions,providing appropriate feedback for pupils ’classroom per⁃formances in order to improve the effectiveness of primary teachers,classroom speech acts.

  2. Automatic speech recognition (ASR) based approach for speech therapy of aphasic patients: A review

    Science.gov (United States)

    Jamal, Norezmi; Shanta, Shahnoor; Mahmud, Farhanahani; Sha'abani, MNAH

    2017-09-01

    This paper reviews the state-of-the-art an automatic speech recognition (ASR) based approach for speech therapy of aphasic patients. Aphasia is a condition in which the affected person suffers from speech and language disorder resulting from a stroke or brain injury. Since there is a growing body of evidence indicating the possibility of improving the symptoms at an early stage, ASR based solutions are increasingly being researched for speech and language therapy. ASR is a technology that transfers human speech into transcript text by matching with the system's library. This is particularly useful in speech rehabilitation therapies as they provide accurate, real-time evaluation for speech input from an individual with speech disorder. ASR based approaches for speech therapy recognize the speech input from the aphasic patient and provide real-time feedback response to their mistakes. However, the accuracy of ASR is dependent on many factors such as, phoneme recognition, speech continuity, speaker and environmental differences as well as our depth of knowledge on human language understanding. Hence, the review examines recent development of ASR technologies and its performance for individuals with speech and language disorders.

  3. Speech and non-speech audio-visual illusions: a developmental study.

    Directory of Open Access Journals (Sweden)

    Corinne Tremblay

    Full Text Available It is well known that simultaneous presentation of incongruent audio and visual stimuli can lead to illusory percepts. Recent data suggest that distinct processes underlie non-specific intersensory speech as opposed to non-speech perception. However, the development of both speech and non-speech intersensory perception across childhood and adolescence remains poorly defined. Thirty-eight observers aged 5 to 19 were tested on the McGurk effect (an audio-visual illusion involving speech, the Illusory Flash effect and the Fusion effect (two audio-visual illusions not involving speech to investigate the development of audio-visual interactions and contrast speech vs. non-speech developmental patterns. Whereas the strength of audio-visual speech illusions varied as a direct function of maturational level, performance on non-speech illusory tasks appeared to be homogeneous across all ages. These data support the existence of independent maturational processes underlying speech and non-speech audio-visual illusory effects.

  4. Exploring the role of brain oscillations in speech perception in noise: Intelligibility of isochronously retimed speech

    Directory of Open Access Journals (Sweden)

    Vincent Aubanel

    2016-08-01

    Full Text Available A growing body of evidence shows that brain oscillations track speech. This mechanism is thought to maximise processing efficiency by allocating resources to important speech information, effectively parsing speech into units of appropriate granularity for further decoding. However, some aspects of this mechanism remain unclear. First, while periodicity is an intrinsic property of this physiological mechanism, speech is only quasi-periodic, so it is not clear whether periodicity would present an advantage in processing. Second, it is still a matter of debate which aspect of speech triggers or maintains cortical entrainment, from bottom-up cues such as fluctuations of the amplitude envelope of speech to higher level linguistic cues such as syntactic structure. We present data from a behavioural experiment assessing the effect of isochronous retiming of speech on speech perception in noise. Two types of anchor points were defined for retiming speech, namely syllable onsets and amplitude envelope peaks. For each anchor point type, retiming was implemented at two hierarchical levels, a slow time scale around 2.5 Hz and a fast time scale around 4 Hz. Results show that while any temporal distortion resulted in reduced speech intelligibility, isochronous speech anchored to P-centers (approximated by stressed syllable vowel onsets was significantly more intelligible than a matched anisochronous retiming, suggesting a facilitative role of periodicity defined on linguistically motivated units in processing speech in noise.

  5. Intelligibility Enhancement of Speech in Noise

    OpenAIRE

    Valentini-Botinhao, Cassia; Yamagishi, Junichi; King, Simon

    2014-01-01

    Speech technology can facilitate human-machine interaction and create new communication interfaces. Text-To-Speech (TTS) systems provide speech output for dialogue, notification and reading applications as well as personalized voices for people that have lost the use of their own. TTS systems are built to produce synthetic voices that should sound as natural, expressive and intelligible as possible and if necessary be similar to a particular speaker. Although naturalness is an important requi...

  6. CAR2 - Czech Database of Car Speech

    Directory of Open Access Journals (Sweden)

    P. Sovka

    1999-12-01

    Full Text Available This paper presents new Czech language two-channel (stereo speech database recorded in car environment. The created database was designed for experiments with speech enhancement for communication purposes and for the study and the design of a robust speech recognition systems. Tools for automated phoneme labelling based on Baum-Welch re-estimation were realised. The noise analysis of the car background environment was done.

  7. Perceptual learning of interrupted speech.

    Directory of Open Access Journals (Sweden)

    Michel Ruben Benard

    Full Text Available The intelligibility of periodically interrupted speech improves once the silent gaps are filled with noise bursts. This improvement has been attributed to phonemic restoration, a top-down repair mechanism that helps intelligibility of degraded speech in daily life. Two hypotheses were investigated using perceptual learning of interrupted speech. If different cognitive processes played a role in restoring interrupted speech with and without filler noise, the two forms of speech would be learned at different rates and with different perceived mental effort. If the restoration benefit were an artificial outcome of using the ecologically invalid stimulus of speech with silent gaps, this benefit would diminish with training. Two groups of normal-hearing listeners were trained, one with interrupted sentences with the filler noise, and the other without. Feedback was provided with the auditory playback of the unprocessed and processed sentences, as well as the visual display of the sentence text. Training increased the overall performance significantly, however restoration benefit did not diminish. The increase in intelligibility and the decrease in perceived mental effort were relatively similar between the groups, implying similar cognitive mechanisms for the restoration of the two types of interruptions. Training effects were generalizable, as both groups improved their performance also with the other form of speech than that they were trained with, and retainable. Due to null results and relatively small number of participants (10 per group, further research is needed to more confidently draw conclusions. Nevertheless, training with interrupted speech seems to be effective, stimulating participants to more actively and efficiently use the top-down restoration. This finding further implies the potential of this training approach as a rehabilitative tool for hearing-impaired/elderly populations.

  8. Text Illustrations.

    Science.gov (United States)

    Duchastel, Philippe C.

    1983-01-01

    Discusses three roles of textbook illustrations--to arrest the reader's attention and arouse interest, to provide explanation and clarification of complex verbal descriptions, and to aid retention of the information presented in the text. It is recommended that illustrations be designed with their specific role(s) in mind. (EAO)

  9. Going to a Speech Therapist

    Science.gov (United States)

    ... Video: Getting an X-ray Going to a Speech Therapist KidsHealth > For Kids > Going to a Speech Therapist ... therapists (also called speech-language pathologists ). What Do Speech Therapists Help With? Speech therapists help people of all ...

  10. PSYCHOLOGICAL-PEDAGOGICAL ASPECT OF WORKING WITH CHILDREN WITH SEVERE SPEECH DISORDERS UNDER CONDITIONS OF INCLUSIVE EDUCATION

    Directory of Open Access Journals (Sweden)

    S. N. Phanaeva

    2016-01-01

    Full Text Available Modern science, including special education, is characterized by a change of paradigms. Special education system has undergone many changes came as a result of the need to give children not only knowledge in various fields, but also to prepare them for independent life in society. One of the priority directions of education of children with severe speech disorders is their social integration, as reflected in the basic regulatory documents. Speech disorder prevents the full communication, the formation of psychological and social readiness to live and work in a new environment, meet unusual and complex situations, moral stability, assimilation of social norms and values, development of social behavior. In the article the problem of development of children with severe speech disorders under conditions of inclusive education. This concept describes the types of abnormalities of speech development.

  11. Speech research

    Science.gov (United States)

    1992-06-01

    Phonology is traditionally seen as the discipline that concerns itself with the building blocks of linguistic messages. It is the study of the structure of sound inventories of languages and of the participation of sounds in rules or processes. Phonetics, in contrast, concerns speech sounds as produced and perceived. Two extreme positions on the relationship between phonological messages and phonetic realizations are represented in the literature. One holds that the primary home for linguistic symbols, including phonological ones, is the human mind, itself housed in the human brain. The second holds that their primary home is the human vocal tract.

  12. A Survey on Statistical Based Single Channel Speech Enhancement Techniques

    Directory of Open Access Journals (Sweden)

    Sunnydayal. V

    2014-11-01

    Full Text Available Speech enhancement is a long standing problem with various applications like hearing aids, automatic recognition and coding of speech signals. Single channel speech enhancement technique is used for enhancement of the speech degraded by additive background noises. The background noise can have an adverse impact on our ability to converse without hindrance or smoothly in very noisy environments, such as busy streets, in a car or cockpit of an airplane. Such type of noises can affect quality and intelligibility of speech. This is a survey paper and its object is to provide an overview of speech enhancement algorithms so that enhance the noisy speech signal which is corrupted by additive noise. The algorithms are mainly based on statistical based approaches. Different estimators are compared. Challenges and Opportunities of speech enhancement are also discussed. This paper helps in choosing the best statistical based technique for speech enhancement

  13. The Phase Spectra Based Feature for Robust Speech Recognition

    Directory of Open Access Journals (Sweden)

    Abbasian ALI

    2009-07-01

    Full Text Available Speech recognition in adverse environment is one of the major issue in automatic speech recognition nowadays. While most current speech recognition system show to be highly efficient for ideal environment but their performance go down extremely when they are applied in real environment because of noise effected speech. In this paper a new feature representation based on phase spectra and Perceptual Linear Prediction (PLP has been suggested which can be used for robust speech recognition. It is shown that this new features can improve the performance of speech recognition not only in clean condition but also in various levels of noise condition when it is compared to PLP features.

  14. From Persuasive to Authoritative Speech Genres

    DEFF Research Database (Denmark)

    Nørreklit, Hanne; Scapens, Robert

    2014-01-01

    by a professional editor in the USA before it was published. Design/methodology/approach: The paper analyses the "persuasive" speech genre of the original version and the "authoritative" speech genre of the published version. Findings: Although is was initially thought that the differences between the two versions......, the authors have focused on just one instance in which a text written by academics was re-written for publication in a practitioner journal. Originality/value: The paper contrasts the rationalism of the persuasive speech genre and the pragmatism of the authoritative speech genre. It cautions academic...

  15. Anticipatory Posturing of the Vocal Tract Reveals Dissociation of Speech Movement Plans from Linguistic Units.

    Directory of Open Access Journals (Sweden)

    Sam Tilsen

    Full Text Available Models of speech production typically assume that control over the timing of speech movements is governed by the selection of higher-level linguistic units, such as segments or syllables. This study used real-time magnetic resonance imaging of the vocal tract to investigate the anticipatory movements speakers make prior to producing a vocal response. Two factors were varied: preparation (whether or not speakers had foreknowledge of the target response and pre-response constraint (whether or not speakers were required to maintain a specific vocal tract posture prior to the response. In prepared responses, many speakers were observed to produce pre-response anticipatory movements with a variety of articulators, showing that that speech movements can be readily dissociated from higher-level linguistic units. Substantial variation was observed across speakers with regard to the articulators used for anticipatory posturing and the contexts in which anticipatory movements occurred. The findings of this study have important consequences for models of speech production and for our understanding of the normal range of variation in anticipatory speech behaviors.

  16. Studying the Speech Recognition Scores of Hearing Impaied Children by Using Nonesense Syllables

    Directory of Open Access Journals (Sweden)

    Mohammad Reza Keyhani

    1998-09-01

    Full Text Available Background: The current article is aimed at evaluating speech recognition scores in hearing aid wearers to determine whether nonsense syllables are suitable speech materials to evaluate the effectiveness of their hearing aids. Method: Subjects were 60 children (15 males and 15 females with bilateral moderate and moderately severe sensorineural hearing impairment who were aged between 7.7-14 years old. Gain prescription was fitted by NAL method. Then speech evaluation was performed in a quiet place with and without hearing aid by using a list of 25 monosyllable words recorded on a tape. A list was prepared for the subjects to check in the correct response. The same method was used to obtain results for normal subjects. Results: The results revealed that the subjects using hearing aids achieved significantly higher SRS in comparison of not wearing it. Although the speech recognition ability was not compensated completely (the maximum score obtained was 60% it was also revealed that the syllable recognition ability in the less amplified frequencies were decreased. the SRS was very higher in normal subjects (with an average of 88%. Conclusion: It seems that Speech recognition score can prepare Audiologist with a more comprehensive method to evaluate the hearing aid benefits.

  17. Building Searchable Collections of Enterprise Speech Data.

    Science.gov (United States)

    Cooper, James W.; Viswanathan, Mahesh; Byron, Donna; Chan, Margaret

    The study has applied speech recognition and text-mining technologies to a set of recorded outbound marketing calls and analyzed the results. Since speaker-independent speech recognition technology results in a significantly lower recognition rate than that found when the recognizer is trained for a particular speaker, a number of post-processing…

  18. HMM Adaptation for child speech synthesis

    CSIR Research Space (South Africa)

    Govender, Avashna

    2015-09-01

    Full Text Available Hidden Markov Model (HMM)-based synthesis in combination with speaker adaptation has proven to be an approach that is well-suited for child speech synthesis. This paper describes the development and evaluation of different HMM-based child speech...

  19. Analysis on Thematic Structure and Thematic Progression in Obama's weekly radio speech

    Institute of Scientific and Technical Information of China (English)

    李成烨

    2016-01-01

    Thematic structure and thematic progression are important concept of textual function. Based on the example of Obama's weekly radio speech, we can examine the practicality of thematic structure and thematic progression and give a guidance for preparation of political speeches.

  20. Utility of TMS to understand the neurobiology of speech

    Directory of Open Access Journals (Sweden)

    Takenobu eMurakami

    2013-07-01

    Full Text Available According to a traditional view, speech perception and production are processed largely separately in sensory and motor brain areas. Recent psycholinguistic and neuroimaging studies provide novel evidence that the sensory and motor systems dynamically interact in speech processing, by demonstrating that speech perception and imitation share regional brain activations. However, the exact nature and mechanisms of these sensorimotor interactions are not completely understood yet.Transcranial magnetic stimulation (TMS has often been used in the cognitive neurosciences, including speech research, as a complementary technique to behavioral and neuroimaging studies. Here we provide an up-to-date review focusing on TMS studies that explored speech perception and imitation.Single-pulse TMS of the primary motor cortex (M1 demonstrated a speech specific and somatotopically specific increase of excitability of the M1 lip area during speech perception (listening to speech or lip reading. A paired-coil TMS approach showed increases in effective connectivity from brain regions that are involved in speech processing to the M1 lip area when listening to speech. TMS in virtual lesion mode applied to speech processing areas modulated performance of phonological recognition and imitation of perceived speech.In summary, TMS is an innovative tool to investigate processing of speech perception and imitation. TMS studies have provided strong evidence that the sensory system is critically involved in mapping sensory input onto motor output and that the motor system plays an important role in speech perception.

  1. Speech production, Psychology of

    NARCIS (Netherlands)

    Schriefers, H.J.; Vigliocco, G.

    2015-01-01

    Research on speech production investigates the cognitive processes involved in transforming thoughts into speech. This article starts with a discussion of the methodological issues inherent to research in speech production that illustrates how empirical approaches to speech production must differ fr

  2. Automatic Speech Recognition from Neural Signals: A Focused Review

    Directory of Open Access Journals (Sweden)

    Christian Herff

    2016-09-01

    Full Text Available Speech interfaces have become widely accepted and are nowadays integrated in various real-life applications and devices. They have become a part of our daily life. However, speech interfaces presume the ability to produce intelligible speech, which might be impossible due to either loud environments, bothering bystanders or incapabilities to produce speech (i.e.~patients suffering from locked-in syndrome. For these reasons it would be highly desirable to not speak but to simply envision oneself to say words or sentences. Interfaces based on imagined speech would enable fast and natural communication without the need for audible speech and would give a voice to otherwise mute people.This focused review analyzes the potential of different brain imaging techniques to recognize speech from neural signals by applying Automatic Speech Recognition technology. We argue that modalities based on metabolic processes, such as functional Near Infrared Spectroscopy and functional Magnetic Resonance Imaging, are less suited for Automatic Speech Recognition from neural signals due to low temporal resolution but are very useful for the investigation of the underlying neural mechanisms involved in speech processes. In contrast, electrophysiologic activity is fast enough to capture speech processes and is therefor better suited for ASR. Our experimental results indicate the potential of these signals for speech recognition from neural data with a focus on invasively measured brain activity (electrocorticography. As a first example of Automatic Speech Recognition techniques used from neural signals, we discuss the emph{Brain-to-text} system.

  3. 78 FR 49717 - Speech-to-Speech and Internet Protocol (IP) Speech-to-Speech Telecommunications Relay Services...

    Science.gov (United States)

    2013-08-15

    ... reasons that STS ] has not been more widely utilized. Are people with speech disabilities not connected to... COMMISSION 47 CFR Part 64 Speech-to-Speech and Internet Protocol (IP) Speech-to-Speech Telecommunications Relay Services; Telecommunications Relay Services and Speech-to-Speech Services for Individuals With...

  4. Speech Enhancement

    DEFF Research Database (Denmark)

    Benesty, Jacob; Jensen, Jesper Rindom; Christensen, Mads Græsbøll;

    of methods and have been introduced in somewhat different contexts. Linear filtering methods originate in stochastic processes, while subspace methods have largely been based on developments in numerical linear algebra and matrix approximation theory. This book bridges the gap between these two classes......Speech enhancement is a classical problem in signal processing, yet still largely unsolved. Two of the conventional approaches for solving this problem are linear filtering, like the classical Wiener filter, and subspace methods. These approaches have traditionally been treated as different classes...... of methods by showing how the ideas behind subspace methods can be incorporated into traditional linear filtering. In the context of subspace methods, the enhancement problem can then be seen as a classical linear filter design problem. This means that various solutions can more easily be compared...

  5. Speech therapy with obturator.

    Science.gov (United States)

    Shyammohan, A; Sreenivasulu, D

    2010-12-01

    Rehabilitation of speech is tantamount to closure of defect in cases with velopharyngeal insufficiency. Often the importance of speech therapy is sidelined during the fabrication of obturators. Usually the speech part is taken up only at a later stage and is relegated entirely to a speech therapist without the active involvement of the prosthodontist. The article suggests a protocol for speech therapy in such cases to be done in unison with a prosthodontist.

  6. A Software Agent for Speech Abiding Systems

    Directory of Open Access Journals (Sweden)

    R. Manoharan

    2009-01-01

    Full Text Available Problem statement: In order to bring speech into the mainstream of business process an efficient digital signal processor is necessary. The Fast Fourier Transform (FFT and the butter fly structure symmetry will enable the harwaring easier. With the DSP and software proposed, togetherly established by means of a system, named here as “Speech Abiding System (SAS”, a software agent, which involves the digital representation of speech signals and the use of digital processors to analyze, synthesize, or modify such signals. The proposed SAS addresses the issues in two parts. Part I: Capturing the Speaker and the Language independent error free Speech Content for speech applications processing and Part II: To accomplish the speech content as an input to the Speech User Applications/Interface (SUI. Approach: Discrete Fourier Transform (DFT of the speech signal is the essential ingredient to evolve this SAS and Discrete-Time Fourier Transform (DTFT links the discrete-time domain to the continuous-frequency domain. The direct computation of DFT is prohibitively expensive in terms of the required computer operations. Fortunately, a number of “fast” transforms have been developed that are mathematically equivalent to the DFT, but which require significantly a fewer computer operations for their implementation. Results: From Part-I, the SAS able to capture an error free Speech content to facilitate the speech as a good input in the main stream of business processing. Part-II provides an environment to implement the speech user applications at a primitive level. Conclusion/Recommendations: The SAS agent along with the required hardware architecture, a Finite State Automata (FSA machine can be created to develop global oriented domain specific speech user applications easily. It will have a major impact on interoperability and disintermediation in the Information Technology Cycle (ITC for computer program generating.

  7. LITURGICAL TEXT IN ANTON CHEKHOV'S NOVELLA "THE DUEL"

    Directory of Open Access Journals (Sweden)

    Syzranov S. V.

    2008-11-01

    Full Text Available The article examines the principle of interaction between the sacred speech, embodied in liturgical texts, and the literary text, typical for Anton Chekhov's works, by the example of his novella "The Duel".

  8. Text Categorization with Latent Dirichlet Allocation

    Directory of Open Access Journals (Sweden)

    ZLACKÝ Daniel

    2014-05-01

    Full Text Available This paper focuses on the text categorization of Slovak text corpora using latent Dirichlet allocation. Our goal is to build text subcorpora that contain similar text documents. We want to use these better organized text subcorpora to build more robust language models that can be used in the area of speech recognition systems. Our previous research in the area of text categorization showed that we can achieve better results with categorized text corpora. In this paper we used latent Dirichlet allocation for text categorization. We divided initial text corpus into 2, 5, 10, 20 or 100 subcorpora with various iterations and save steps. Language models were built on these subcorpora and adapted with linear interpolation to judicial domain. The experiment results showed that text categorization using latent Dirichlet allocation can improve the system for automatic speech recognition by creating the language models from organized text corpora.

  9. Teaching Speech Acts

    Directory of Open Access Journals (Sweden)

    Teaching Speech Acts

    2007-01-01

    Full Text Available In this paper I argue that pragmatic ability must become part of what we teach in the classroom if we are to realize the goals of communicative competence for our students. I review the research on pragmatics, especially those articles that point to the effectiveness of teaching pragmatics in an explicit manner, and those that posit methods for teaching. I also note two areas of scholarship that address classroom needs—the use of authentic data and appropriate assessment tools. The essay concludes with a summary of my own experience teaching speech acts in an advanced-level Portuguese class.

  10. Gaucho Gazette: Speech and Sensationalism

    Directory of Open Access Journals (Sweden)

    Roberto José Ramos

    2013-07-01

    Full Text Available The Gaucho Gazette presents itself as a “popular newspaper”. Attempts to produce a denial about his aesthetic tabloid. Search only say that discloses what happens, as if the media were merely a reflection of society. This paper will seek to understand and explain your Sensationalism, through their speeches. Use for both, semiology, Roland Barthes, in their possibilities transdisciplinary.

  11. Une Analyse d'un discours politique (An Analysis of a Type of Political Speech)

    Science.gov (United States)

    Collin-Platini, Muriel

    1978-01-01

    A linguistic analysis of 30 political speeches by de Gaulle is presented. The speeches are compared with regard to pronouns, verb tense, ways of referring to France, length of phrase, and total length of speech. The speeches are then characterized as electoral/non-electoral and evolution/crisis. (Text is in French.) (MLA)

  12. The Effect of English Verbal Songs on Connected Speech Aspects of Adult English Learners’ Speech Production

    Directory of Open Access Journals (Sweden)

    Farshid Tayari Ashtiani

    2015-02-01

    Full Text Available The present study was an attempt to investigate the impact of English verbal songs on connected speech aspects of adult English learners’ speech production. 40 participants were selected based on the results of their performance in a piloted and validated version of NELSON test given to 60 intermediate English learners in a language institute in Tehran. Then they were equally distributed in two control and experimental groups and received a validated pretest of reading aloud and speaking in English. Afterward, the treatment was performed in 18 sessions by singing preselected songs culled based on some criteria such as popularity, familiarity, amount, and speed of speech delivery, etc. In the end, the posttests of reading aloud and speaking in English were administered. The results revealed that the treatment had statistically positive effects on the connected speech aspects of English learners’ speech production at statistical .05 level of significance. Meanwhile, the results represented that there was not any significant difference between the experimental group’s mean scores on the posttests of reading aloud and speaking. It was thus concluded that providing the EFL learners with English verbal songs could positively affect connected speech aspects of both modes of speech production, reading aloud and speaking. The Findings of this study have pedagogical implications for language teachers to be more aware and knowledgeable of the benefits of verbal songs to promote speech production of language learners in terms of naturalness and fluency. Keywords: English Verbal Songs, Connected Speech, Speech Production, Reading Aloud, Speaking

  13. Gesture & Speech Based Appliance Control

    Directory of Open Access Journals (Sweden)

    Dr. Sayleegharge,

    2014-01-01

    Full Text Available This document explores the use of speech & gestures to control home appliances. Aiming at the aging population of the world and relieving them from their dependencies. The two approaches used to sail through the target are the MFCC approach for speech processing and the Identification of Characteristic Point Algorithm for gesture recognition. A barrier preventing wide adoption is that this audience can find controlling assistive technology difficult, as they are less dexterous and computer literate. Our results hope to provide a more natural and intuitive interface to help bridge the gap between technology and elderly users.

  14. Reproducible Research in Speech Sciences

    Directory of Open Access Journals (Sweden)

    Kandaacute;lmandaacute;n Abari

    2012-11-01

    Full Text Available Reproducible research is the minimum standard of scientific claims in cases when independent replication proves to be difficult. With the special combination of available software tools, we provide a reproducibility recipe for the experimental research conducted in some fields of speech sciences. We have based our model on the triad of the R environment, the EMU-format speech database, and the executable publication. We present the use of three typesetting systems (LaTeX, Markdown, Org, with the help of a mini research.

  15. NEOLOGISMS IN CHILD'S SPEECH CAUSED BY THE DISSEMINATION OF VIDEOS

    Directory of Open Access Journals (Sweden)

    Sidorenko A. V.

    2015-11-01

    Full Text Available The article studies words and set expressions of children's speech appearance of which are directly linked to the dissemination of movies in the late 1980s in the Soviet Union. These words and set expressions are analyzed on the basis of the material collected by the author for «The Dictionary of Soviet Childhood» being prepared for printing. As a result of exposure to foreign video products, some new expressions and words entered children's speech and have become an important part of the linguistic landscape of the Soviet childhood. Many of them have been stored in conversations so far: invective vocabulary, translation loanwords, barbarisms, exotic vocabulary, proper names. Many words have developed new meanings and become part of new Russian lexical sets. These neologisms in vocabulary are reflected in children's folklore; some of them have become part of the literary language and have been fixed in modern dictionaries of the Russian language. Many of them are still used in Russian informal speech. The author gives examples of the use of these lexical units in fiction and publicist texts paying attention to their semantic, phonetic, spelling and grammatical features

  16. Delayed Speech or Language Development

    Science.gov (United States)

    ... to 2-Year-Old Delayed Speech or Language Development KidsHealth > For Parents > Delayed Speech or Language Development ... child is right on schedule. Normal Speech & Language Development It's important to discuss early speech and language ...

  17. International Conference on Harmonisation; Guidance on Q4B Evaluation and Recommendation of Pharmacopoeial Texts for Use in the International Conference on Harmonisation Regions; Annex on Test for Extractable Volume of Parenteral Preparations General Chapter; availability. Notice.

    Science.gov (United States)

    2009-01-09

    The Food and Drug Administration (FDA) is announcing the availability of a guidance entitled "Q4B Evaluation and Recommendation of Pharmacopoeial Texts for Use in the ICH Regions; Annex 2: Test for Extractable Volume of Parenteral Preparations General Chapter." The guidance was prepared under the auspices of the International Conference on Harmonisation of Technical Requirements for Registration of Pharmaceuticals for Human Use (ICH). The guidance provides the results of the ICH Q4B evaluation of the Test for Extractable Volume of Parenteral Preparations General Chapter harmonized text from each of the three pharmacopoeias (United States, European, and Japanese) represented by the Pharmacopoeial Discussion Group (PDG). The guidance conveys recognition of the three pharmacopoeial methods by the three ICH regulatory regions and provides specific information regarding the recognition. The guidance is intended to recognize the interchangeability between the local regional pharmacopoeias, thus avoiding redundant testing in favor of a common testing strategy in each regulatory region. In the Federal Register of February 21, 2008 (73 FR 9575), FDA made available a guidance on the Q4B process entitled "Q4B Evaluation and Recommendation of Pharmacopoeial Texts for Use in the ICH Regions."

  18. Impact of enhanced input, pushed output, and inner speech on the ...

    African Journals Online (AJOL)

    Impact of enhanced input, pushed output, and inner speech on the accuracy of EFL Learners under ... The results demonstrates that learner inner speech helps to draw the attention of ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  19. Genesis 2–3 and Alcibiades's speech in Plato's Symposium : A ...

    African Journals Online (AJOL)

    Genesis 2–3 and Alcibiades's speech in Plato's Symposium : A cultural critical reading. ... interpretation can arise from the analysis of Alcibiades's speech compared to M- and LXX-Genesis ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT

  20. The development of sensorimotor influences in the audiovisual speech domain: Some critical questions

    Directory of Open Access Journals (Sweden)

    Bahia eGuellaï

    2014-08-01

    Full Text Available Speech researchers have long been interested in how auditory and visual speech signals are integrated, and recent work has revived interest in the role of speech production with respect to this process. Here we discuss these issues from a developmental perspective. Because speech perception abilities typically outstrip speech production abilities in infancy and childhood, it is unclear how speech-like movements could influence audiovisual speech perception in development. While work on this question is still in its preliminary stages, there is nevertheless increasing evidence that sensorimotor processes (defined here as any motor or proprioceptive process related to orofacial movements affect developmental audiovisual speech processing. We suggest three areas on which to focus in future research: i the relation between audiovisual speech perception and sensorimotor processes at birth, ii the pathways through which sensorimotor processes interact with audiovisual speech processing in infancy, and iii developmental change in sensorimotor pathways as speech production emerges in childhood.

  1. Can you hear my age? Influences of speech rate and speech spontaneity on estimation of speaker age

    Directory of Open Access Journals (Sweden)

    Sara eWaller Skoog

    2015-07-01

    Full Text Available Cognitive hearing science is mainly about the study of how cognitive factors contribute to speech comprehension, but cognitive factors also partake in speech processing to infer non-linguistic information from speech signals, such as the intentions of the talker and the speaker’s age. Here, we report two experiments on age estimation by naïve listeners. The aim was to study how speech rate influences estimation of speaker age by comparing the speakers’ natural speech rate with increased or decreased speech rate. In Experiment 1, listeners were presented with audio samples of read speech from three different speaker age groups (young, middle aged and old adults. They estimated the speakers as younger when speech rate was faster than normal and as older when speech rate was slower than normal. This speech rate effect was slightly greater in magnitude for older (60-65 years speakers in comparison with younger (20-25 years speakers, suggesting that speech rate may gain greater importance as a perceptual age cue with increased speaker age. This pattern was more pronounced in Experiment 2, in which listeners estimated age from spontaneous speech. Faster speech rate was associated with lower age estimates, but only for older and middle aged (40-45 years speakers. Taken together, speakers of all age groups were estimated as older when speech rate decreased, except for the youngest speakers in Experiment 2. The absence of a linear speech rate effect in estimates of younger speakers, for spontaneous speech, implies that listeners use different age estimation strategies or cues (possibly vocabulary depending on the age of the speaker and the spontaneity of the speech. Potential implications for forensic investigations and other applied domains are discussed.

  2. Speech perception as an active cognitive process

    Directory of Open Access Journals (Sweden)

    Shannon eHeald

    2014-03-01

    Full Text Available One view of speech perception is that acoustic signals are transformed into representations for pattern matching to determine linguistic structure. This process can be taken as a statistical pattern-matching problem, assuming realtively stable linguistic categories are characterized by neural representations related to auditory properties of speech that can be compared to speech input. This kind of pattern matching can be termed a passive process which implies rigidity of processingd with few demands on cognitive processing. An alternative view is that speech recognition, even in early stages, is an active process in which speech analysis is attentionally guided. Note that this does not mean consciously guided but that information-contingent changes in early auditory encoding can occur as a function of context and experience. Active processing assumes that attention, plasticity, and listening goals are important in considering how listeners cope with adverse circumstances that impair hearing by masking noise in the environment or hearing loss. Although theories of speech perception have begun to incorporate some active processing, they seldom treat early speech encoding as plastic and attentionally guided. Recent research has suggested that speech perception is the product of both feedforward and feedback interactions between a number of brain regions that include descending projections perhaps as far downstream as the cochlea. It is important to understand how the ambiguity of the speech signal and constraints of context dynamically determine cognitive resources recruited during perception including focused attention, learning, and working memory. Theories of speech perception need to go beyond the current corticocentric approach in order to account for the intrinsic dynamics of the auditory encoding of speech. In doing so, this may provide new insights into ways in which hearing disorders and loss may be treated either through augementation or

  3. Corpus Design for Malay Corpus-based Speech Synthesis System

    Directory of Open Access Journals (Sweden)

    Tian-Swee Tan

    2009-01-01

    Full Text Available Problem statement: Speech corpus is one of the major components in corpus-based synthesis. The quality and coverage in speech corpus will affect the quality of synthesis speech sound. Approach: This study proposes a corpus design for Malay corpus-based speech synthesis system. This includes the study of design criteria in corpus-based speech synthesis, Malay corpus based database design and the concatenation engine in Malay corpus-based synthesis system. A set of 10 millions digital text corpuses for Malay language has been collected from Malay internet news. This text corpus had been analyzed using word frequency count to find out all high frequency words to be used for designing the sentences for speech corpus. Results: Altogether 381 sentences for speech corpus had been designed using 70% of high frequency words from 10 million text corpus. It consists of 16826 phoneme units and the total storage size is 37.6Mb. All the phone units are phonetically transcribed to preserve the phonetic context of its origin that will be used for phonetic context unit. This speech corpus had been labeled at phoneme level and used for variable length continuous phoneme based concatenation. Speech corpus is one of the major components in corpus-based synthesis. The quality and coverage in speech corpus will affect the quality of synthesized speech sound. Conclusion/Recommendation: This study has proposed a platform for designing speech corpus especially for Malay Text to Speech which can be further enhanced to support more coverage and higher naturalness of synthetic speech.

  4. Speech Emotion Recognition Using Fuzzy Logic Classifier

    Directory of Open Access Journals (Sweden)

    Daniar aghsanavard

    2016-01-01

    Full Text Available Over the last two decades, emotions, speech recognition and signal processing have been one of the most significant issues in the adoption of techniques to detect them. Each method has advantages and disadvantages. This paper tries to suggest fuzzy speech emotion recognition based on the classification of speech's signals in order to better recognition along with a higher speed. In this system, the use of fuzzy logic system with 5 layers, which is the combination of neural progressive network and algorithm optimization of firefly, first, speech samples have been given to input of fuzzy orbit and then, signals will be investigated and primary classified in a fuzzy framework. In this model, a pattern of signals will be created for each class of signals, which results in reduction of signal data dimension as well as easier speech recognition. The obtained experimental results show that our proposed method (categorized by firefly, improves recognition of utterances.

  5. 78 FR 49693 - Speech-to-Speech and Internet Protocol (IP) Speech-to-Speech Telecommunications Relay Services...

    Science.gov (United States)

    2013-08-15

    ... COMMISSION 47 CFR Part 64 Speech-to-Speech and Internet Protocol (IP) Speech-to-Speech Telecommunications Relay Services; Telecommunications Relay Services and Speech-to-Speech Services for Individuals With... this document, the Commission amends telecommunications relay services (TRS) mandatory...

  6. Speech and Language Impairments

    Science.gov (United States)

    ... impairment. Many children are identified as having a speech or language impairment after they enter the public school system. A teacher may notice difficulties in a child’s speech or communication skills and refer the child for ...

  7. Speech versus singing: Infants choose happier sounds

    Directory of Open Access Journals (Sweden)

    Marieve eCorbeil

    2013-06-01

    Full Text Available Infants prefer speech to non-vocal sounds and to non-human vocalizations, and they prefer happy-sounding speech to neutral speech. They also exhibit an interest in singing, but there is little knowledge of their relative interest in speech and singing. The present study explored infants’ attention to unfamiliar audio samples of speech and singing. In Experiment 1, infants 4-13 months of age were exposed to happy-sounding infant-directed speech versus hummed lullabies by the same woman. They listened significantly longer to the speech, which had considerably greater acoustic variability and expressiveness, than to the lullabies. In Experiment 2, infants of comparable age who heard the lyrics of a Turkish children’s song spoken versus sung in a joyful/happy manner did not exhibit differential listening. Infants in Experiment 3 heard the happily sung lyrics of the Turkish children’s song versus a version that was spoken in an adult-directed or affectively neutral manner. They listened significantly longer to the sung version. Overall, happy voice quality rather than vocal mode (speech or singing was the principal contributor to infant attention, regardless of age.

  8. Industrial Applications of Automatic Speech Recognition Systems

    Directory of Open Access Journals (Sweden)

    Dr. Jayashri Vajpai

    2016-03-01

    Full Text Available Current trends in developing technologies form important bridges to the future, fortified by the early and productive use of technology for enriching the human life. Speech signal processing, which includes automatic speech recognition, synthetic speech, and natural language processing, is beginning to have a significant impact on business, industry and ease of operation of personal computers. Apart from this, it facilitates the deeper understanding of complex mechanism of functioning of human brain. Advances in speech recognition technology, over the past five decades, have enabled a wide range of industrial applications. Yet today's applications provide a small preview of a rich future for speech and voice interface technology that will eventually replace keyboards with microphones for designing human machine interface for providing easy access to increasingly intelligent machines. It also shows how the capabilities of speech recognition systems in industrial applications are evolving over time to usher in the next generation of voice-enabled services. This paper aims to present an effective survey of the speech recognition technology described in the available literature and integrate the insights gained during the process of study of individual research and developments. The current applications of speech recognition for real world and industry have also been outlined with special reference to applications in the areas of medical, industrial robotics, forensic, defence and aviation

  9. Disordered Speech Assessment Using Automatic Methods Based on Quantitative Measures

    Directory of Open Access Journals (Sweden)

    Christine Sapienza

    2005-06-01

    Full Text Available Speech quality assessment methods are necessary for evaluating and documenting treatment outcomes of patients suffering from degraded speech due to Parkinson's disease, stroke, or other disease processes. Subjective methods of speech quality assessment are more accurate and more robust than objective methods but are time-consuming and costly. We propose a novel objective measure of speech quality assessment that builds on traditional speech processing techniques such as dynamic time warping (DTW and the Itakura-Saito (IS distortion measure. Initial results show that our objective measure correlates well with the more expensive subjective methods.

  10. Texts, Transmissions, Receptions. Modern Approaches to Narratives

    NARCIS (Netherlands)

    Lardinois, A.P.M.H.; Levie, S.A.; Hoeken, H.; Lüthy, C.H.

    2015-01-01

    The papers collected in this volume study the function and meaning of narrative texts from a variety of perspectives. The word 'text' is used here in the broadest sense of the term: it denotes literary books, but also oral tales, speeches, newspaper articles and comics. One of the purposes of this v

  11. Texts, Transmissions, Receptions. Modern Approaches to Narratives

    NARCIS (Netherlands)

    Lardinois, A.P.M.H.; Levie, S.A.; Hoeken, H.; Lüthy, C.H.

    2015-01-01

    The papers collected in this volume study the function and meaning of narrative texts from a variety of perspectives. The word 'text' is used here in the broadest sense of the term: it denotes literary books, but also oral tales, speeches, newspaper articles and comics. One of the purposes of this

  12. Bakhtin’s Theory of Speech Genres and the Addressivity in Cultural Pragmatics of Genres in Literature: “Genus” vs. “Genre”

    OpenAIRE

    Grübel, Rainer

    2016-01-01

    This paper suggests a new approach to literary genres and modes via the addressivity of texts and performances, which takes it starting point from Bakhtin’s concept of speech genres. First it deals with Bakhtin’s publications on literary genres, which are interpreted differently by literary theoreticians and critics, and the problematic status of his texts on speech genres, which were not prepared for publication by the author. It then takes Jan-Luc Nancy’s concept of addressing – with its co...

  13. Speech 7 through 12.

    Science.gov (United States)

    Nederland Independent School District, TX.

    GRADES OR AGES: Grades 7 through 12. SUBJECT MATTER: Speech. ORGANIZATION AND PHYSICAL APPEARANCE: Following the foreward, philosophy and objectives, this guide presents a speech curriculum. The curriculum covers junior high and Speech I, II, III (senior high). Thirteen units of study are presented for junior high, each unit is divided into…

  14. A Rule Based System for Speech Language Context Understanding

    Institute of Scientific and Technical Information of China (English)

    Imran Sarwar Bajwa; Muhammad Abbas Choudhary

    2006-01-01

    Speech or Natural language contents are major tools of communication. This research paper presents a natural language processing based automated system for understanding speech language text. A new rule based model has been presented for analyzing the natural languages and extracting the relative meanings from the given text. User writes the natural language text in simple English in a few paragraphs and the designed system has a sound ability of analyzing the given script by the user. After composite analysis and extraction of associated information, the designed system gives particular meanings to an assortment of speech language text on the basis of its context. The designed system uses standard speech language rules that are clearly defined for all speech languages as English,Urdu, Chinese, Arabic, French, etc. The designed system provides a quick and reliable way to comprehend speech language context and generate respective meanings.

  15. Causes of Speech Disorders in Primary School Students of Zahedan

    Directory of Open Access Journals (Sweden)

    Saeed Fakhrerahimi

    2013-02-01

    Full Text Available Background: Since making communication with others is the most important function of speech, undoubtedly, any type of disorder in speech will affect the human communicability with others. The objective of the study was to investigate reasons behind the [high] prevalence rate of stammer, producing disorders and aglossia.Materials and Methods: This descriptive-analytical study was conducted on 118 male and female students, who were studying in a primary school in Zahedan; they had referred to the Speech Therapy Centers of Zahedan University of Medical Sciences in a period of seven months. The speech therapist examinations, diagnosis tools common in speech therapy, Spielberg Children Trait and also patients' cases were used to find the reasons behind the [high] prevalence rate of speech disorders. Results: Psychological causes had the highest rate of correlation with the speech disorders among the other factors affecting the speech disorders. After psychological causes, family history and age of the subjects are the other factors which may bring about the speech disorders (P<0.05. Bilingualism and birth order has a negative relationship with the speech disorders. Likewise, another result of this study shows that only psychological causes, social causes, hereditary causes and age of subjects can predict the speech disorders (P<0.05.Conclusion: The present study shows that the speech disorders have a strong and close relationship with the psychological causes at the first step and also history of family and age of individuals at the next steps.

  16. DIRECTIVE SPEECH ACT IN THE MOVIE SLEEPING BEAUTY

    Directory of Open Access Journals (Sweden)

    Muhartoyo

    2013-09-01

    Full Text Available Pragmatics is one of linguistics studies that is quite attractive to learn more about. There are many aspects of pragmatics; one of them is dealing with speech acts. Speech acts consist of many categories; one of them is directive speech act. This study aims to identify the directive speech act performed in Sleeping Beauty movie. Likewise, it will find out how often the directive speech act performed and which type of directive speech act that are most frequently used in the movie. This study used qualitative method in which data collection is done by watching the movie, analyzing the body movement and the dialogues of each character, reading the script and library research. A total of 139 directive speech acts were successfully identified. The result of analysis showed that the directive speech act of ordering is the most frequently used in the movie (21,6%. The least frequently used directive speech act is inviting directive speech act (0,7%. The study also revealed the importance of directive speech act in keeping the flow of storyline of the movie. This study is expected to give some useful insights in understanding what directive speech acts is.

  17. Sensorimotor Oscillations Prior to Speech Onset Reflect Altered Motor Networks in Adults Who Stutter

    Science.gov (United States)

    Mersov, Anna-Maria; Jobst, Cecilia; Cheyne, Douglas O.; De Nil, Luc

    2016-01-01

    Adults who stutter (AWS) have demonstrated atypical coordination of motor and sensory regions during speech production. Yet little is known of the speech-motor network in AWS in the brief time window preceding audible speech onset. The purpose of the current study was to characterize neural oscillations in the speech-motor network during preparation for and execution of overt speech production in AWS using magnetoencephalography (MEG). Twelve AWS and 12 age-matched controls were presented with 220 words, each word embedded in a carrier phrase. Controls were presented with the same word list as their matched AWS participant. Neural oscillatory activity was localized using minimum-variance beamforming during two time periods of interest: speech preparation (prior to speech onset) and speech execution (following speech onset). Compared to controls, AWS showed stronger beta (15–25 Hz) suppression in the speech preparation stage, followed by stronger beta synchronization in the bilateral mouth motor cortex. AWS also recruited the right mouth motor cortex significantly earlier in the speech preparation stage compared to controls. Exaggerated motor preparation is discussed in the context of reduced coordination in the speech-motor network of AWS. It is further proposed that exaggerated beta synchronization may reflect a more strongly inhibited motor system that requires a stronger beta suppression to disengage prior to speech initiation. These novel findings highlight critical differences in the speech-motor network of AWS that occur prior to speech onset and emphasize the need to investigate further the speech-motor assembly in the stuttering population. PMID:27642279

  18. THE ONTOGENESIS OF SPEECH DEVELOPMENT

    Directory of Open Access Journals (Sweden)

    T. E. Braudo

    2017-01-01

    Full Text Available The purpose of this article is to acquaint the specialists, working with children having developmental disorders, with age-related norms for speech development. Many well-known linguists and psychologists studied speech ontogenesis (logogenesis. Speech is a higher mental function, which integrates many functional systems. Speech development in infants during the first months after birth is ensured by the innate hearing and emerging ability to fix the gaze on the face of an adult. Innate emotional reactions are also being developed during this period, turning into nonverbal forms of communication. At about 6 months a baby starts to pronounce some syllables; at 7–9 months – repeats various sounds combinations, pronounced by adults. At 10–11 months a baby begins to react on the words, referred to him/her. The first words usually appear at an age of 1 year; this is the start of the stage of active speech development. At this time it is acceptable, if a child confuses or rearranges sounds, distorts or misses them. By the age of 1.5 years a child begins to understand abstract explanations of adults. Significant vocabulary enlargement occurs between 2 and 3 years; grammatical structures of the language are being formed during this period (a child starts to use phrases and sentences. Preschool age (3–7 y. o. is characterized by incorrect, but steadily improving pronunciation of sounds and phonemic perception. The vocabulary increases; abstract speech and retelling are being formed. Children over 7 y. o. continue to improve grammar, writing and reading skills. The described stages may not have strict age boundaries, as soon as they are dependent not only on environment, but also on the child’s mental constitution, heredity and character.

  19. An audio-visual corpus for multimodal speech recognition in Dutch language

    NARCIS (Netherlands)

    Wojdel, J.; Wiggers, P.; Rothkrantz, L.J.M.

    2002-01-01

    This paper describes the gathering and availability of an audio-visual speech corpus for Dutch language. The corpus was prepared with the multi-modal speech recognition in mind and it is currently used in our research on lip-reading and bimodal speech recognition. It contains the prompts used also i

  20. An audio-visual corpus for multimodal speech recognition in Dutch language

    NARCIS (Netherlands)

    Wojdel, J.; Wiggers, P.; Rothkrantz, L.J.M.

    2002-01-01

    This paper describes the gathering and availability of an audio-visual speech corpus for Dutch language. The corpus was prepared with the multi-modal speech recognition in mind and it is currently used in our research on lip-reading and bimodal speech recognition. It contains the prompts used also

  1. An audio-visual corpus for multimodal speech recognition in Dutch language

    NARCIS (Netherlands)

    Wojdel, J.; Wiggers, P.; Rothkrantz, L.J.M.

    2002-01-01

    This paper describes the gathering and availability of an audio-visual speech corpus for Dutch language. The corpus was prepared with the multi-modal speech recognition in mind and it is currently used in our research on lip-reading and bimodal speech recognition. It contains the prompts used also i

  2. Relations between affective music and speech: Evidence from dynamics of affective piano performance and speech production

    Directory of Open Access Journals (Sweden)

    Xiaoluan eLiu

    2015-07-01

    Full Text Available This study compares affective piano performance with speech production from the perspective of dynamics: unlike previous research, this study uses finger force and articulatory effort as indexes reflecting the dynamics of affective piano performance and speech production respectively. Moreover, for the first time physical constraints such as piano fingerings and speech articulatory distance are included due to their potential contribution to different patterns of dynamics. A piano performance experiment and speech production experiment were conducted in four emotions: anger, fear, happiness and sadness. The results show that in both piano performance and speech production, anger and happiness generally have high dynamics while sadness has the lowest dynamics, with fear in the middle. Fingerings interact with fear in the piano experiment and articulatory distance interacts with anger in the speech experiment, i.e., large physical constraints produce significantly higher dynamics than small physical constraints in piano performance under the condition of fear and in speech production under the condition of anger. Using production experiments, this study firstly supports previous perception studies on relations between affective music and speech. Moreover, this is the first study to show quantitative evidence for the importance of considering motor aspects such as dynamics in comparing music performance and speech production in which motor mechanisms play a crucial role.

  3. CONVERGING TOWARDS A COMMON SPEECH CODE: IMITATIVE AND PERCEPTUO-MOTOR RECALIBRATION PROCESSES IN SPEECH PRODUCTION

    Directory of Open Access Journals (Sweden)

    Marc eSato

    2013-07-01

    Full Text Available Auditory and somatosensory systems play a key role in speech motor control. In the act of speaking, segmental speech movements are programmed to reach phonemic sensory goals, which in turn are used to estimate actual sensory feedback in order to further control production. The adult’s tendency to automatically imitate a number of acoustic-phonetic characteristics in another speaker's speech however suggests that speech production not only relies on the intended phonemic sensory goals and actual sensory feedback but also on the processing of external speech inputs. These online adaptive changes in speech production, or phonetic convergence effects, are thought to facilitate conversational exchange by contributing to setting a common perceptuo-motor ground between the speaker and the listener. In line with previous studies on phonetic convergence, we here demonstrate, in a non-interactive situation of communication, online unintentional and voluntary imitative changes in relevant acoustic features of acoustic vowel targets (fundamental and first formant frequencies during speech production and imitation. In addition, perceptuo-motor recalibration processes, or after-effects, occurred not only after vowel production and imitation but also after auditory categorization of the acoustic vowel targets. Altogether, these findings demonstrate adaptive plasticity of phonemic sensory-motor goals and suggest that, apart from sensory-motor knowledge, speech production continuously draws on perceptual learning from the external speech environment.

  4. From Gesture to Speech

    Directory of Open Access Journals (Sweden)

    Maurizio Gentilucci

    2012-11-01

    Full Text Available One of the major problems concerning the evolution of human language is to understand how sounds became associated to meaningful gestures. It has been proposed that the circuit controlling gestures and speech evolved from a circuit involved in the control of arm and mouth movements related to ingestion. This circuit contributed to the evolution of spoken language, moving from a system of communication based on arm gestures. The discovery of the mirror neurons has provided strong support for the gestural theory of speech origin because they offer a natural substrate for the embodiment of language and create a direct link between sender and receiver of a message. Behavioural studies indicate that manual gestures are linked to mouth movements used for syllable emission. Grasping with the hand selectively affected movement of inner or outer parts of the mouth according to syllable pronunciation and hand postures, in addition to hand actions, influenced the control of mouth grasp and vocalization. Gestures and words are also related to each other. It was found that when producing communicative gestures (emblems the intention to interact directly with a conspecific was transferred from gestures to words, inducing modification in voice parameters. Transfer effects of the meaning of representational gestures were found on both vocalizations and meaningful words. It has been concluded that the results of our studies suggest the existence of a system relating gesture to vocalization which was precursor of a more general system reciprocally relating gesture to word.

  5. Auto Spell Suggestion for High Quality Speech Synthesis in Hindi

    Science.gov (United States)

    Kabra, Shikha; Agarwal, Ritika

    2014-02-01

    The goal of Text-to-Speech (TTS) synthesis in a particular language is to convert arbitrary input text to intelligible and natural sounding speech. However, for a particular language like Hindi, which is a highly confusing language (due to very close spellings), it is not an easy task to identify errors/mistakes in input text and an incorrect text degrade the quality of output speech hence this paper is a contribution to the development of high quality speech synthesis with the involvement of Spellchecker which generates spell suggestions for misspelled words automatically. Involvement of spellchecker would increase the efficiency of speech synthesis by providing spell suggestions for incorrect input text. Furthermore, we have provided the comparative study for evaluating the resultant effect on to phonetic text by adding spellchecker on to input text.

  6. Hate Speech Revisited: The "Toon" Controversy

    Directory of Open Access Journals (Sweden)

    Rajeev Dhavan

    2010-01-01

    Full Text Available Examining the cartoon controversy which ignited violent protests and ban in various countries, this article examines the contours of "hate speech" in various legal systems. While broadly supporting the case of free speech the authors remind users of free speech to exercise self-restraint. Absolute bans should not be made, but time, person and place constraints may be essential. Ironically, the toon controversy also reveals the silence of the sympathetic majority. Similarly, there is a duty to speak. Even though not enforceable, it remains a duty to democracy.

  7. The Functional Connectome of Speech Control.

    Directory of Open Access Journals (Sweden)

    Stefan Fuertinger

    2015-07-01

    Full Text Available In the past few years, several studies have been directed to understanding the complexity of functional interactions between different brain regions during various human behaviors. Among these, neuroimaging research installed the notion that speech and language require an orchestration of brain regions for comprehension, planning, and integration of a heard sound with a spoken word. However, these studies have been largely limited to mapping the neural correlates of separate speech elements and examining distinct cortical or subcortical circuits involved in different aspects of speech control. As a result, the complexity of the brain network machinery controlling speech and language remained largely unknown. Using graph theoretical analysis of functional MRI (fMRI data in healthy subjects, we quantified the large-scale speech network topology by constructing functional brain networks of increasing hierarchy from the resting state to motor output of meaningless syllables to complex production of real-life speech as well as compared to non-speech-related sequential finger tapping and pure tone discrimination networks. We identified a segregated network of highly connected local neural communities (hubs in the primary sensorimotor and parietal regions, which formed a commonly shared core hub network across the examined conditions, with the left area 4p playing an important role in speech network organization. These sensorimotor core hubs exhibited features of flexible hubs based on their participation in several functional domains across different networks and ability to adaptively switch long-range functional connectivity depending on task content, resulting in a distinct community structure of each examined network. Specifically, compared to other tasks, speech production was characterized by the formation of six distinct neural communities with specialized recruitment of the prefrontal cortex, insula, putamen, and thalamus, which collectively

  8. Speech in spinocerebellar ataxia.

    Science.gov (United States)

    Schalling, Ellika; Hartelius, Lena

    2013-12-01

    Spinocerebellar ataxias (SCAs) are a heterogeneous group of autosomal dominant cerebellar ataxias clinically characterized by progressive ataxia, dysarthria and a range of other concomitant neurological symptoms. Only a few studies include detailed characterization of speech symptoms in SCA. Speech symptoms in SCA resemble ataxic dysarthria but symptoms related to phonation may be more prominent. One study to date has shown an association between differences in speech and voice symptoms related to genotype. More studies of speech and voice phenotypes are motivated, to possibly aid in clinical diagnosis. In addition, instrumental speech analysis has been demonstrated to be a reliable measure that may be used to monitor disease progression or therapy outcomes in possible future pharmacological treatments. Intervention by speech and language pathologists should go beyond assessment. Clinical guidelines for management of speech, communication and swallowing need to be developed for individuals with progressive cerebellar ataxia.

  9. Digital speech processing using Matlab

    CERN Document Server

    Gopi, E S

    2014-01-01

    Digital Speech Processing Using Matlab deals with digital speech pattern recognition, speech production model, speech feature extraction, and speech compression. The book is written in a manner that is suitable for beginners pursuing basic research in digital speech processing. Matlab illustrations are provided for most topics to enable better understanding of concepts. This book also deals with the basic pattern recognition techniques (illustrated with speech signals using Matlab) such as PCA, LDA, ICA, SVM, HMM, GMM, BPN, and KSOM.

  10. The rhetoric of public speech in a digital era

    NARCIS (Netherlands)

    Hoven, P.J. van den

    2012-01-01

    In the early Chinese tradition as well as in the Western tradition, rhetorical studies predominantly address one prototypical communicative setting; an orator delivers a carefully prepared speech to a well-defined audience on a well-defined occasion.

  11. Malay Speech Intelligibility Test (MSIT for Deaf Malaysian Children

    Directory of Open Access Journals (Sweden)

    Zulkhairi Md. Yusof

    2014-01-01

    Full Text Available This paper presents a novel approach to objectively measure speech intelligibility of deaf Malaysian children using nonsense syllables as test words. In this study, speech intelligibility of deaf children is measured through the ability to say simple syllables (consisting of a consonant and a vowel for all the 22 Malay consonants. The MSIT score should indicate how well these children can produce speech; the higher the score, the better their speech intelligibility. The MSIT have been tested with twenty deaf and ten normal hearning Malaysian children and has been verified by five naïve and two expert listeners. The application of the test is presented as speech intelligibility index for deaf children in a deaf school. The successful development of the MSIT system serves to assist speech pathologist, therapist, teachers and parents of deaf children to determine the level of Malay speech deficiencies and the effectiveness of corrective measures taken.

  12. Speech Databases of Typical Children and Children with SLI.

    Directory of Open Access Journals (Sweden)

    Pavel Grill

    Full Text Available The extent of research on children's speech in general and on disordered speech specifically is very limited. In this article, we describe the process of creating databases of children's speech and the possibilities for using such databases, which have been created by the LANNA research group in the Faculty of Electrical Engineering at Czech Technical University in Prague. These databases have been principally compiled for medical research but also for use in other areas, such as linguistics. Two databases were recorded: one for healthy children's speech (recorded in kindergarten and in the first level of elementary school and the other for pathological speech of children with a Specific Language Impairment (recorded at a surgery of speech and language therapists and at the hospital. Both databases were sub-divided according to specific demands of medical research. Their utilization can be exoteric, specifically for linguistic research and pedagogical use as well as for studies of speech-signal processing.

  13. [Restoration of speech function in oncological patients with maxillary defects].

    Science.gov (United States)

    Matiakin, E G; Chuchkov, V M; Akhundov, A A; Azizian, R I; Romanov, I S; Chuchkov, M V; Agapov, V V

    2009-01-01

    Speech quality was evaluated in 188 patients with acquired maxillary defects. Prosthetic treatment of 29 patients was preceded by pharmacopsychotherapy. Sixty three patients had lessons with a logopedist and 66 practiced self-tuition based on the specially developed test. Thirty patients were examined for the quality of speech without preliminary preparation. Speech quality was assessed by auditory and spectral analysis. The main forms of impaired speech quality in the patients with maxillary defects were marked rhinophonia and impaired articulation. The proposed analytical tests were based on a combination of "difficult" vowels and consonants. The use of a removable prostheses with an obturator failed to correct the affected speech function but created prerequisites for the formation of the correct speech stereotype. Results of the study suggest the relationship between the quality of speech in subjects with maxillary defects and their intellectual faculties as well as the desire to overcome this drawback. The proposed tests are designed to activate the neuromuscular apparatus responsible for the generation of the speech. Lessons with a speech therapist give a powerful emotional incentive to the patients and promote their efforts toward restoration of speaking ability. Pharmacopsychotherapy and self-control are another efficacious tools for the improvement of speech quality in patients with maxillary defects.

  14. Towards Quranic reader controlled by speech

    Directory of Open Access Journals (Sweden)

    Yacine Yekache

    2011-11-01

    Full Text Available In this paper we describe the process of designing a task-oriented continuous speech recognition system for Arabic, based on CMU Sphinx4, to be used in the voice interface of Quranic reader. The concept of the Quranic reader controlled by speech is presented, the collection of the corpus and creation of acoustic model are described in detail taking into account a specificities of Arabic language and the desired application.

  15. Acoustic Analysis of PD Speech

    Directory of Open Access Journals (Sweden)

    Karen Chenausky

    2011-01-01

    Full Text Available According to the U.S. National Institutes of Health, approximately 500,000 Americans have Parkinson's disease (PD, with roughly another 50,000 receiving new diagnoses each year. 70%–90% of these people also have the hypokinetic dysarthria associated with PD. Deep brain stimulation (DBS substantially relieves motor symptoms in advanced-stage patients for whom medication produces disabling dyskinesias. This study investigated speech changes as a result of DBS settings chosen to maximize motor performance. The speech of 10 PD patients and 12 normal controls was analyzed for syllable rate and variability, syllable length patterning, vowel fraction, voice-onset time variability, and spirantization. These were normalized by the controls' standard deviation to represent distance from normal and combined into a composite measure. Results show that DBS settings relieving motor symptoms can improve speech, making it up to three standard deviations closer to normal. However, the clinically motivated settings evaluated here show greater capacity to impair, rather than improve, speech. A feedback device developed from these findings could be useful to clinicians adjusting DBS parameters, as a means for ensuring they do not unwittingly choose DBS settings which impair patients' communication.

  16. General-purpose isiZulu speech synthesiser

    CSIR Research Space (South Africa)

    Louw, A

    2005-07-01

    Full Text Available A general-purpose isiZulu text-to-speech (TTS) system was developed, based on the “Multisyn” unit-selection approach supported by the Festival TTS toolkit. The development involved a number of challenges related to the interface between speech...

  17. Speech disorders encountered in routine ENT practice and the role of speech therapy in its effective management

    Directory of Open Access Journals (Sweden)

    Jay Dave

    2013-01-01

    Full Text Available Aim: To study speech disorders encountered in routine ENT practice and the role of speech therapy in its effective management. Material and Methods: The present study was carried out on 60 patients, who were presented to the out patient department of E.N.T., M.P.SHAH MEDICAL COLEGE, JAMNAGAR (Gujarat,for speech disorders. The OPD and Indoor patients of the E.N.T. department, who had a speech defect prior to or following surgical interference for organic Otorhinolaryngologic disorders were also included in this study. Results: Speech disorders like Dyslalia, Dysglossia, Stuttering were prevalent in both the sexes and in every level of society, though the disorders were more prevalent in males as compared to females. Stuttering was the most common disorder encountered and speech therapy had encouraging results. All the ten patients of Dyslalia were subjected to appropriate articulation training, following which eight patients improved satisfactorily whereas two patients missed follow-up. All the eight Deaf-Mute patients were given auditory training and speech and language stimulation therapy, following which, two patients acquired satisfactory speech level. All seven patients of Dysglossia improved satisfactorily after tongue tie release surgery and appropriate speech therapy. Conclusion: With the advent of newer treatment modalities and improved coordination between ENT surgeons and speech-pathologists for correct diagnosis of speech disorders, speech therapy is suitably individualized according to the expectation of the patients and limitation of the speech pathologists. A fuller understanding of the effectiveness of speech therapy amongst the patients has developed and it has emerged as an acceptable modality of treatment of various speech disorders.

  18. [Speech characteristics of persons wearing full upper and lower prostheses].

    Science.gov (United States)

    Kaán, M; Bolla, K; Keszler, B

    1993-02-01

    We've taken under comparative examination the phonation of people having full upper and lower dentures and the ones having normal teeth. We've tested twenty 51 to 65 years old persons having their dentures at least for a year being pleased with them (well speaking and chewing normally) and twenty persons (without distinction as to sex) having normal teeth normal occlusion. The investigation of the articulation was compiled a lingual corpus consisted of texts of different length (words, word-groups, sentences and text-pieces). This material was suitable for the examination of disorders of the articulation, the pronunciation of the sound combinations, the articulatory problems, and the suprasegmental components (e.g. melody, speech tempo, dynamics of speech). The experimental material was prepared by the following instruments: an VII (Voice Identification, Inc.) 700 series sound spectrograph, an IBM compatible computer with acoustic analytical programs suitable for the investigation of selected sound samples by an A/D (analog/digital) converter, the number of sounds in a second was measured with a 8SO-4 type oscillograph. The results can be summarized as follows: 1. various differences appear among the speech structures of persons having full upper and lower dentures: the labial [b p m], labiodental [v f ], dental, dentialveolar [d t n z s dz ts l r], alveolar [symbol: see text] and prepalatal [symbol: see text] consonants distort principally. 2. From the long consonants the voicelles stops (p: t: k:) are short-ended especially.(ABSTRACT TRUNCATED AT 250 WORDS)

  19. Speech-to-Speech Relay Service

    Science.gov (United States)

    ... to make an STS call. You are then connected to an STS CA who will repeat your spoken words, making the spoken words clear to the other party. Persons with speech disabilities may also receive STS calls. The calling ...

  20. Opening speech

    Directory of Open Access Journals (Sweden)

    Anders Hallberg

    2006-10-01

    Full Text Available Mr President of LIBER, Mr Chancellor of Justice (JK Göran Lambertz finns med i publiken, and Distinguished Heads and Representatives of important European Libraries, First, a warm welcome to all of you in Uppsala and Sweden. It is, for reasons unknown to me, actually the first time in your Association’s 35 years proud history that you convene in this country.

  1. Theater, Speech, Light

    Directory of Open Access Journals (Sweden)

    Primož Vitez

    2011-07-01

    Full Text Available This paper considers a medium as a substantial translator: an intermediary between the producers and receivers of a communicational act. A medium is a material support to the spiritual potential of human sources. If the medium is a support to meaning, then the relations between different media can be interpreted as a space for making sense of these meanings, a generator of sense: it means that the interaction of substances creates an intermedial space that conceives of a contextualization of specific meaningful elements in order to combine them into the sense of a communicational intervention. The theater itself is multimedia. A theatrical event is a communicational act based on a combination of several autonomous structures: text, scenography, light design, sound, directing, literary interpretation, speech, and, of course, the one that contains all of these: the actor in a human body. The actor is a physical and symbolic, anatomic, and emblematic figure in the synesthetic theatrical act because he reunites in his body all the essential principles and components of theater itself. The actor is an audio-visual being, made of kinetic energy, speech, and human spirit. The actor’s body, as a source, instrument, and goal of the theater, becomes an intersection of sound and light. However, theater as intermedial art is no intermediate practice; it must be seen as interposing bodies between conceivers and receivers, between authors and auditors. The body is not self-evident; the body in contemporary art forms is being redefined as a privilege. The art needs bodily dimensions to explore the medial qualities of substances: because it is alive, it returns to studying biology. The fact that theater is an archaic art form is also the purest promise of its future.

  2. Exploration of Speech Planning and Producing by Speech Error Analysis

    Institute of Scientific and Technical Information of China (English)

    冷卉

    2012-01-01

    Speech error analysis is an indirect way to discover speech planning and producing processes. From some speech errors made by people in their daily life, linguists and learners can reveal the planning and producing processes more easily and clearly.

  3. Perceptual evaluation of corpus-based speech synthesis techniques in under-resourced environments

    CSIR Research Space (South Africa)

    Van Niekerk, DR

    2009-11-01

    Full Text Available With the increasing prominence and maturity of corpus-based techniques for speech synthesis, the process of system development has in some ways been simplified considerably. However, the dependence on sufficient amounts of relevant speech data...

  4. Indirect Speech Acts

    Institute of Scientific and Technical Information of China (English)

    李威

    2001-01-01

    Indirect speech acts are frequently used in verbal communication, the interpretation of them is of great importance in order to meet the demands of the development of students' communicative competence. This paper, therefore, intends to present Searle' s indirect speech acts and explore the way how indirect speech acts are interpreted in accordance with two influential theories. It consists of four parts. Part one gives a general introduction to the notion of speech acts theory. Part two makes an elaboration upon the conception of indirect speech act theory proposed by Searle and his supplement and development of illocutionary acts. Part three deals with the interpretation of indirect speech acts. Part four draws implication from the previous study and also serves as the conclusion of the dissertation.

  5. Esophageal speeches modified by the Speech Enhancer Program®

    OpenAIRE

    Manochiopinig, Sriwimon; Boonpramuk, Panuthat

    2014-01-01

    Esophageal speech appears to be the first choice of speech treatment for a laryngectomy. However, many laryngectomy people are unable to speak well. The aim of this study was to evaluate post-modified speech quality of Thai esophageal speakers using the Speech Enhancer Program®. The method adopted was to approach five speech–language pathologists to assess the speech accuracy and intelligibility of the words and continuing speech of the seven laryngectomy people. A comparison study was conduc...

  6. Texts, Transmissions, Receptions. Modern Approaches to Narratives

    NARCIS (Netherlands)

    Lardinois, A.P.M.H.; Levie, S.A.; Hoeken, H.; Lüthy, C.H.

    2015-01-01

    The papers collected in this volume study the function and meaning of narrative texts from a variety of perspectives. The word “text” is used here in the broadest sense of the term: it denotes literary books, but also oral tales, speeches, newspaper articles and comics. One of the purposes of this v

  7. Texts, Transmissions, Receptions. Modern Approaches to Narratives

    NARCIS (Netherlands)

    Lardinois, A.P.M.H.; Levie, S.A.; Hoeken, H.; Lüthy, C.H.

    2015-01-01

    The papers collected in this volume study the function and meaning of narrative texts from a variety of perspectives. The word “text” is used here in the broadest sense of the term: it denotes literary books, but also oral tales, speeches, newspaper articles and comics. One of the purposes of this

  8. Robust digital processing of speech signals

    CERN Document Server

    Kovacevic, Branko; Veinović, Mladen; Marković, Milan

    2017-01-01

    This book focuses on speech signal phenomena, presenting a robustification of the usual speech generation models with regard to the presumed types of excitation signals, which is equivalent to the introduction of a class of nonlinear models and the corresponding criterion functions for parameter estimation. Compared to the general class of nonlinear models, such as various neural networks, these models possess good properties of controlled complexity, the option of working in “online” mode, as well as a low information volume for efficient speech encoding and transmission. Providing comprehensive insights, the book is based on the authors’ research, which has already been published, supplemented by additional texts discussing general considerations of speech modeling, linear predictive analysis and robust parameter estimation.

  9. Principles of speech coding

    CERN Document Server

    Ogunfunmi, Tokunbo

    2010-01-01

    It is becoming increasingly apparent that all forms of communication-including voice-will be transmitted through packet-switched networks based on the Internet Protocol (IP). Therefore, the design of modern devices that rely on speech interfaces, such as cell phones and PDAs, requires a complete and up-to-date understanding of the basics of speech coding. Outlines key signal processing algorithms used to mitigate impairments to speech quality in VoIP networksOffering a detailed yet easily accessible introduction to the field, Principles of Speech Coding provides an in-depth examination of the

  10. Ear, Hearing and Speech

    DEFF Research Database (Denmark)

    Poulsen, Torben

    2000-01-01

    An introduction is given to the the anatomy and the function of the ear, basic psychoacoustic matters (hearing threshold, loudness, masking), the speech signal and speech intelligibility. The lecture note is written for the course: Fundamentals of Acoustics and Noise Control (51001)......An introduction is given to the the anatomy and the function of the ear, basic psychoacoustic matters (hearing threshold, loudness, masking), the speech signal and speech intelligibility. The lecture note is written for the course: Fundamentals of Acoustics and Noise Control (51001)...

  11. Advances in Speech Recognition

    CERN Document Server

    Neustein, Amy

    2010-01-01

    This volume is comprised of contributions from eminent leaders in the speech industry, and presents a comprehensive and in depth analysis of the progress of speech technology in the topical areas of mobile settings, healthcare and call centers. The material addresses the technical aspects of voice technology within the framework of societal needs, such as the use of speech recognition software to produce up-to-date electronic health records, not withstanding patients making changes to health plans and physicians. Included will be discussion of speech engineering, linguistics, human factors ana

  12. IMPLEMENTING SPEECH COMMUNITY STRATEGY TO ENHANCE STUDENTS’ ENGLISH SPEAKING ABILITY

    Directory of Open Access Journals (Sweden)

    Huriyah Huriyah

    2015-12-01

    Full Text Available Implementing speech community strategy to enhance students English speaking ability. This classroom action research describes how the implementation of speech community strategy increases the students’ English speaking ability. The research stages consist of planning, implementing, observing, evaluating and reflecting. The study indicates that the providing of speech community can increase English speaking ability at students of SMA Sekar Kemuning Islamic Boarding School Cirebon

  13. The Unsupervised Acquisition of a Lexicon from Continuous Speech.

    Science.gov (United States)

    1995-11-01

    COGNITIVE SCIENCES A.I. Memo No. 1558 November, 1996 C.B.C.L. Memo No. 129 The Unsupervised Acquisition of a Lexicon from Continuous Speech Carl de...tion of the input. Thus, it has diverse application in speech recognition, lexicography , text and speech compression, machine translation, and the...the Cognitive Science Society, pages 28{36, 1993. [5] Michael R. Brent, Andrew Lundberg, and Sreerama Murthy. Discovering morphemic suxes: A case

  14. THE MEANING OF THE PREVENTION WITH SPEECH THERAPY AS A IMPORTANT FAC-TOR FOR THE PROPER DEVELOPMENT OF THE CHILDREN SPEECH

    Directory of Open Access Journals (Sweden)

    S. FILIPOVA

    1999-11-01

    Full Text Available The paper presented some conscientious and results from the finished research which showing the meaning of the prevention with speech therapy in the development of the speech. The research was done at Negotino and with that are shown the most frequent speech deficiency of the children at preschool age.

  15. Formation of intonational party of speech in the structure of the correction letter at younger schoolboys with mild expressed general underdevelopment of speech with erased form of dysarthria

    Directory of Open Access Journals (Sweden)

    Zoya Kyrbanova

    2015-05-01

    Full Text Available The problem of reading and writing disorders in children with speech underdevelopment with dizartricheskim component. The methodology of the survey speech capabilities for this category of children. The directions of correctional and developmental work on the formation of intonational party of speech in order to prevent violations of reading and writing.

  16. Speech-Language Therapy (For Parents)

    Science.gov (United States)

    ... Feeding Your 1- to 2-Year-Old Speech-Language Therapy KidsHealth > For Parents > Speech-Language Therapy A ... with speech and/or language disorders. Speech Disorders, Language Disorders, and Feeding Disorders A speech disorder refers ...

  17. Acoustic cues identifying phonetic transitions for speech segmentation

    CSIR Research Space (South Africa)

    Van Niekerk, DR

    2008-11-01

    Full Text Available The quality of corpus-based text-to-speech (TTS) systems depends strongly on the consistency of boundary placements during phonetic alignments. Expert human transcribers use visually represented acoustic cues in order to consistently place...

  18. Auto Spell Suggestion for High Quality Speech Synthesis in Hindi

    National Research Council Canada - National Science Library

    Shikha Kabra; Ritika Agarwal

    2014-01-01

    ...), it is not an easy task to identify errors/mistakes in input text and an incorrect text degrade the quality of output speech hence this paper is a contribution to the development of high quality...

  19. Speech parts as Poisson processes.

    Science.gov (United States)

    Badalamenti, A F

    2001-09-01

    This paper presents evidence that six of the seven parts of speech occur in written text as Poisson processes, simple or recurring. The six major parts are nouns, verbs, adjectives, adverbs, prepositions, and conjunctions, with the interjection occurring too infrequently to support a model. The data consist of more than the first 5000 words of works by four major authors coded to label the parts of speech, as well as periods (sentence terminators). Sentence length is measured via the period and found to be normally distributed with no stochastic model identified for its occurrence. The models for all six speech parts but the noun significantly distinguish some pairs of authors and likewise for the joint use of all words types. Any one author is significantly distinguished from any other by at least one word type and sentence length very significantly distinguishes each from all others. The variety of word type use, measured by Shannon entropy, builds to about 90% of its maximum possible value. The rate constants for nouns are close to the fractions of maximum entropy achieved. This finding together with the stochastic models and the relations among them suggest that the noun may be a primitive organizer of written text.

  20. THE BASIS FOR SPEECH PREVENTION

    Directory of Open Access Journals (Sweden)

    Jordan JORDANOVSKI

    1997-06-01

    Full Text Available The speech is a tool for accurate communication of ideas. When we talk about speech prevention as a practical realization of the language, we are referring to the fact that it should be comprised of the elements of the criteria as viewed from the perspective of the standards. This criteria, in the broad sense of the word, presupposes an exact realization of the thought expressed between the speaker and the recipient.The absence of this criterion catches the eye through the practical realization of the language and brings forth consequences, often hidden very deeply in the human psyche. Their outer manifestation already represents a delayed reaction of the social environment. The foundation for overcoming and standardization of this phenomenon must be the anatomy-physiological patterns of the body, accomplished through methods in concordance with the nature of the body.

  1. Cognitive Complexity and Second Language Speech Production.

    Science.gov (United States)

    Appel, Gabriela; Lantolf, James P.

    A study compared the effects of cognitive complexity on the speech production of 14 advanced non-native speakers of English and 14 native English-speakers. Cognitively simple and complex tasks were distinguished based on text type (narrative versus expository). Subjects read one narrative and one expository text in separate sessions, then wrote…

  2. Processing of Speech Signals for Physical and Sensory Disabilities

    Science.gov (United States)

    Levitt, Harry

    1995-10-01

    Assistive technology involving voice communication is used primarily by people who are deaf, hard of hearing, or who have speech and/or language disabilities. It is also used to a lesser extent by people with visual or motor disabilities. A very wide range of devices has been developed for people with hearing loss. These devices can be categorized not only by the modality of stimulation [i.e., auditory, visual, tactile, or direct electrical stimulation of the auditory nerve (auditory-neural)] but also in terms of the degree of speech processing that is used. At least four such categories can be distinguished: assistive devices (a) that are not designed specifically for speech, (b) that take the average characteristics of speech into account, (c) that process articulatory or phonetic characteristics of speech, and (d) that embody some degree of automatic speech recognition. Assistive devices for people with speech and/or language disabilities typically involve some form of speech synthesis or symbol generation for severe forms of language disability. Speech synthesis is also used in text-to-speech systems for sightless persons. Other applications of assistive technology involving voice communication include voice control of wheelchairs and other devices for people with mobility disabilities.

  3. Tracking Speech Sound Acquisition

    Science.gov (United States)

    Powell, Thomas W.

    2011-01-01

    This article describes a procedure to aid in the clinical appraisal of child speech. The approach, based on the work by Dinnsen, Chin, Elbert, and Powell (1990; Some constraints on functionally disordered phonologies: Phonetic inventories and phonotactics. "Journal of Speech and Hearing Research", 33, 28-37), uses a railway idiom to track gains in…

  4. Preschool Connected Speech Inventory.

    Science.gov (United States)

    DiJohnson, Albert; And Others

    This speech inventory developed for a study of aurally handicapped preschool children (see TM 001 129) provides information on intonation patterns in connected speech. The inventory consists of a list of phrases and simple sentences accompanied by pictorial clues. The test is individually administered by a teacher-examiner who presents the spoken…

  5. Private Speech in Ballet

    Science.gov (United States)

    Johnston, Dale

    2006-01-01

    Authoritarian teaching practices in ballet inhibit the use of private speech. This paper highlights the critical importance of private speech in the cognitive development of young ballet students, within what is largely a non-verbal art form. It draws upon research by Russian psychologist Lev Vygotsky and contemporary socioculturalists, to…

  6. Private Speech in Ballet

    Science.gov (United States)

    Johnston, Dale

    2006-01-01

    Authoritarian teaching practices in ballet inhibit the use of private speech. This paper highlights the critical importance of private speech in the cognitive development of young ballet students, within what is largely a non-verbal art form. It draws upon research by Russian psychologist Lev Vygotsky and contemporary socioculturalists, to…

  7. Tracking Speech Sound Acquisition

    Science.gov (United States)

    Powell, Thomas W.

    2011-01-01

    This article describes a procedure to aid in the clinical appraisal of child speech. The approach, based on the work by Dinnsen, Chin, Elbert, and Powell (1990; Some constraints on functionally disordered phonologies: Phonetic inventories and phonotactics. "Journal of Speech and Hearing Research", 33, 28-37), uses a railway idiom to track gains in…

  8. Free Speech Yearbook 1976.

    Science.gov (United States)

    Phifer, Gregg, Ed.

    The articles collected in this annual address several aspects of First Amendment Law. The following titles are included: "Freedom of Speech As an Academic Discipline" (Franklyn S. Haiman), "Free Speech and Foreign-Policy Decision Making" (Douglas N. Freeman), "The Supreme Court and the First Amendment: 1975-1976"…

  9. Preschool Connected Speech Inventory.

    Science.gov (United States)

    DiJohnson, Albert; And Others

    This speech inventory developed for a study of aurally handicapped preschool children (see TM 001 129) provides information on intonation patterns in connected speech. The inventory consists of a list of phrases and simple sentences accompanied by pictorial clues. The test is individually administered by a teacher-examiner who presents the spoken…

  10. Advertising and Free Speech.

    Science.gov (United States)

    Hyman, Allen, Ed.; Johnson, M. Bruce, Ed.

    The articles collected in this book originated at a conference at which legal and economic scholars discussed the issue of First Amendment protection for commercial speech. The first article, in arguing for freedom for commercial speech, finds inconsistent and untenable the arguments of those who advocate freedom from regulation for political…

  11. Free Speech. No. 38.

    Science.gov (United States)

    Kane, Peter E., Ed.

    This issue of "Free Speech" contains the following articles: "Daniel Schoor Relieved of Reporting Duties" by Laurence Stern, "The Sellout at CBS" by Michael Harrington, "Defending Dan Schorr" by Tome Wicker, "Speech to the Washington Press Club, February 25, 1976" by Daniel Schorr, "Funds Voted For Schorr Inquiry" by Richard Lyons, "Erosion of the…

  12. THE USE OF EXPRESSIVE SPEECH ACTS IN HANNAH MONTANA SESSION 1

    Directory of Open Access Journals (Sweden)

    Nur Vita Handayani

    2015-07-01

    Full Text Available This study aims to describe kinds and forms of expressive speech act in Hannah Montana Session 1. It belongs to descriptive qualitative method. The research object was expressive speech act. The data source was utterances which contain expressive speech acts in the film Hannah Montana Session 1. The researcher used observation method and noting technique in collecting the data. In analyzing the data, descriptive qualitative method was used. The research findings show that there are ten kinds of expressive speech act found in Hannah Montana Session 1, namely expressing apology, expressing thanks, expressing sympathy, expressing attitudes, expressing greeting, expressing wishes, expressing joy, expressing pain, expressing likes, and expressing dislikes. The forms of expressive speech act are direct literal expressive speech act, direct non-literal expressive speech act, indirect literal expressive speech act, and indirect non-literal expressive speech act.

  13. A Research of Speech Emotion Recognition Based on Deep Belief Network and SVM

    Directory of Open Access Journals (Sweden)

    Chenchen Huang

    2014-01-01

    Full Text Available Feature extraction is a very important part in speech emotion recognition, and in allusion to feature extraction in speech emotion recognition problems, this paper proposed a new method of feature extraction, using DBNs in DNN to extract emotional features in speech signal automatically. By training a 5 layers depth DBNs, to extract speech emotion feature and incorporate multiple consecutive frames to form a high dimensional feature. The features after training in DBNs were the input of nonlinear SVM classifier, and finally speech emotion recognition multiple classifier system was achieved. The speech emotion recognition rate of the system reached 86.5%, which was 7% higher than the original method.

  14. Polish Phoneme Statistics Obtained On Large Set Of Written Texts

    Directory of Open Access Journals (Sweden)

    Bartosz Ziółko

    2009-01-01

    Full Text Available The phonetical statistics were collected from several Polish corpora. The paper is a summaryof the data which are phoneme n-grams and some phenomena in the statistics. Triphonestatistics apply context-dependent speech units which have an important role in speech recognitionsystems and were never calculated for a large set of Polish written texts. The standardphonetic alphabet for Polish, SAMPA, and methods of providing phonetic transcriptions are described.

  15. Ensemble Feature Extraction Modules for Improved Hindi Speech Recognition System

    Directory of Open Access Journals (Sweden)

    Malay Kumar

    2012-05-01

    Full Text Available Speech is the most natural way of communication between human beings. The field of speech recognition generates intrigues of man - machine conversation and due to its versatile applications; automatic speech recognition systems have been designed. In this paper we are presenting a novel approach for Hindi speech recognition by ensemble feature extraction modules of ASR systems and their outputs have been combined using voting technique ROVER. Experimental results have been shown that proposed system will produce better result than traditional ASR systems.

  16. Charisma in business speeches

    DEFF Research Database (Denmark)

    Niebuhr, Oliver; Brem, Alexander; Novák-Tót, Eszter

    2016-01-01

    to business speeches. Consistent with the public opinion, our findings are indicative of Steve Jobs being a more charismatic speaker than Mark Zuckerberg. Beyond previous studies, our data suggest that rhythm and emphatic accentuation are also involved in conveying charisma. Furthermore, the differences......Charisma is a key component of spoken language interaction; and it is probably for this reason that charismatic speech has been the subject of intensive research for centuries. However, what is still largely missing is a quantitative and objective line of research that, firstly, involves analyses...... of the acoustic-prosodic signal, secondly, focuses on business speeches like product presentations, and, thirdly, in doing so, advances the still fairly fragmentary evidence on the prosodic correlates of charismatic speech. We show that the prosodic features of charisma in political speeches also apply...

  17. Multiple Transcoding Impact on Speech Quality in Ideal Network Conditions

    Directory of Open Access Journals (Sweden)

    Martin Mikulec

    2015-01-01

    Full Text Available This paper deals with the impact of transcoding on the speech quality. We have focused mainly on the transcoding between codecs without the negative influence of the network parameters such as packet loss and delay. It has ensured objective and repeatable results from our measurement. The measurement was performed on the Transcoding Measuring System developed especially for this purpose. The system is based on the open source projects and is useful as a design tool for VoIP system administrators. The paper compares the most used codecs from the transcoding perspective. The multiple transcoding between G711, GSM and G729 codecs were performed and the speech quality of these calls was evaluated. The speech quality was measured by Perceptual Evaluation of Speech Quality method, which provides results in Mean Opinion Score used to describe the speech quality on a scale from 1 to 5. The obtained results indicate periodical speech quality degradation on every transcoding between two codecs.

  18. Role of neural network models for developing speech systems

    Indian Academy of Sciences (India)

    K Sreenivasa Rao

    2011-10-01

    This paper discusses the application of neural networks for developing different speech systems. Prosodic parameters of speech at syllable level depend on positional, contextual and phonological features of the syllables. In this paper, neural networks are explored to model the prosodic parameters of the syllables from their positional, contextual and phonological features. The prosodic parameters considered in this work are duration and sequence of pitch $(F_0)$ values of the syllables. These prosody models are further examined for applications such as text to speech synthesis, speech recognition, speaker recognition and language identification. Neural network models in voice conversion system are explored for capturing the mapping functions between source and target speakers at source, system and prosodic levels. We have also used neural network models for characterizing the emotions present in speech. For identification of dialects in Hindi, neural network models are used to capture the dialect specific information from spectral and prosodic features of speech.

  19. Multi-thread Parallel Speech Recognition for Mobile Applications

    Directory of Open Access Journals (Sweden)

    LOJKA Martin

    2014-05-01

    Full Text Available In this paper, the server based solution of the multi-thread large vocabulary automatic speech recognition engine is described along with the Android OS and HTML5 practical application examples. The basic idea was to bring speech recognition available for full variety of applications for computers and especially for mobile devices. The speech recognition engine should be independent of commercial products and services (where the dictionary could not be modified. Using of third-party services could be also a security and privacy problem in specific applications, when the unsecured audio data could not be sent to uncontrolled environments (voice data transferred to servers around the globe. Using our experience with speech recognition applications, we have been able to construct a multi-thread speech recognition serverbased solution designed for simple applications interface (API to speech recognition engine modified to specific needs of particular application.

  20. An Embedded Application for Degraded Text Recognition

    Directory of Open Access Journals (Sweden)

    Thillou Céline

    2005-01-01

    Full Text Available This paper describes a mobile device which tries to give the blind or visually impaired access to text information. Three key technologies are required for this system: text detection, optical character recognition, and speech synthesis. Blind users and the mobile environment imply two strong constraints. First, pictures will be taken without control on camera settings and a priori information on text (font or size and background. The second issue is to link several techniques together with an optimal compromise between computational constraints and recognition efficiency. We will present the overall description of the system from text detection to OCR error correction.

  1. SUSTAINABILITY IN THE BOWELS OF SPEECHES

    Directory of Open Access Journals (Sweden)

    Jadir Mauro Galvao

    2012-10-01

    Full Text Available The theme of sustainability has not yet achieved the feat of make up as an integral part the theoretical medley that brings out our most everyday actions, often visits some of our thoughts and permeates many of our speeches. The big event of 2012, the meeting gathered Rio +20 glances from all corners of the planet around that theme as burning, but we still see forward timidly. Although we have no very clear what the term sustainability closes it does not sound quite strange. Associate with things like ecology, planet, wastes emitted by smokestacks of factories, deforestation, recycling and global warming must be related, but our goal in this article is the least of clarifying the term conceptually and more try to observe as it appears in speeches of such conference. When the competent authorities talk about sustainability relate to what? We intend to investigate the lines and between the lines of these speeches, any assumptions associated with the term. Therefore we will analyze the speech of the People´s Summit, the opening speech of President Dilma and emblematic speech of the President of Uruguay, José Pepe Mujica.

  2. Wideband Speech Recovery Using Psychoacoustic Criteria

    Directory of Open Access Journals (Sweden)

    Visar Berisha

    2007-08-01

    Full Text Available Many modern speech bandwidth extension techniques predict the high-frequency band based on features extracted from the lower band. While this method works for certain types of speech, problems arise when the correlation between the low and the high bands is not sufficient for adequate prediction. These situations require that additional high-band information is sent to the decoder. This overhead information, however, can be cleverly quantized using human auditory system models. In this paper, we propose a novel speech compression method that relies on bandwidth extension. The novelty of the technique lies in an elaborate perceptual model that determines a quantization scheme for wideband recovery and synthesis. Furthermore, a source/filter bandwidth extension algorithm based on spectral spline fitting is proposed. Results reveal that the proposed system improves the quality of narrowband speech while performing at a lower bitrate. When compared to other wideband speech coding schemes, the proposed algorithms provide comparable speech quality at a lower bitrate.

  3. Wideband Speech Recovery Using Psychoacoustic Criteria

    Directory of Open Access Journals (Sweden)

    Berisha Visar

    2007-01-01

    Full Text Available Many modern speech bandwidth extension techniques predict the high-frequency band based on features extracted from the lower band. While this method works for certain types of speech, problems arise when the correlation between the low and the high bands is not sufficient for adequate prediction. These situations require that additional high-band information is sent to the decoder. This overhead information, however, can be cleverly quantized using human auditory system models. In this paper, we propose a novel speech compression method that relies on bandwidth extension. The novelty of the technique lies in an elaborate perceptual model that determines a quantization scheme for wideband recovery and synthesis. Furthermore, a source/filter bandwidth extension algorithm based on spectral spline fitting is proposed. Results reveal that the proposed system improves the quality of narrowband speech while performing at a lower bitrate. When compared to other wideband speech coding schemes, the proposed algorithms provide comparable speech quality at a lower bitrate.

  4. Expression of future prospective in indirect speech

    Directory of Open Access Journals (Sweden)

    Bodnaruk Elena Vladimirovna

    2015-03-01

    Full Text Available The article analyzes the characteristics and use of grammatical semantics and lexical and grammatical means used to create future prospects in double indirect discourse. The material for the study were epic works by contemporary German writers. In the analysis of the empirical material it has been pointed out that indirect discourse has preterial basis and is the kind of most frequent inner speech of characters. The most widely used form with future semantics in preterial indirect speech is conditional I, formally having a conjunctive basis, but is mostly used with the indicative semantics. Competitive to conditional I in indirect speech is preterial indicative. A characteristic feature of the indirect speech is the use of modal verbs, which, thanks to its semantics is usually referred as an action at a later term, creating the prospect of future statements. The most frequent were modal verbs wollen and sollen in the form of the preterite, more rare verbs were m ssen and k nnen. German indirect speech distinguishes the ability to use forms on the basis of conjunctive: preterite and plusquamperfect of conjunctive. Both forms express values similar to those of the indicative. However, conjunctive forms the basis of the data shown in a slightly more pronounced seme of uncertainty that accompanies future uses of these forms in indirect speech. In addition, plusquamperfect conjunctive differs from others by the presence of the seme of completeness.

  5. Service Robot SCORPIO with Robust Speech Interface

    Directory of Open Access Journals (Sweden)

    Stanislav Ondas

    2013-01-01

    Full Text Available The SCORPIO is a small‐size mini‐teleoperator mobile service robot for booby‐trap disposal. It can be manually controlled by an operator through a portable briefcase remote control device using joystick, keyboard and buttons. In this paper, the speech interface is described. As an auxiliary function, the remote interface allows a human operator to concentrate sight and/or hands on other operation activities that are more important. The developed speech interface is based on HMM‐based acoustic models trained using the SpeechDatE‐SK database, a small‐vocabulary language model based on fixed connected words, grammar, and the speech recognition setup adapted for low‐resource devices. To improve the robustness of the speech interface in an outdoor environment, which is the working area of the SCORPIO service robot, a speech enhancement based on the spectral subtraction method, as well as a unique combination of an iterative approach and a modified LIMA framework, were researched, developed and tested on simulated and real outdoor recordings.

  6. The First Amendment in a Time of Media Proliferation: Does Freedom of Speech Entail a Private Right to Censor?

    Directory of Open Access Journals (Sweden)

    Patrick M. Garry

    2003-04-01

    Full Text Available Modern information technology offers not only more speech, but more ways to deliver that speech. Through the media, every kind of speech is readily available, sometimes with no more than the push of a finger. Many Internet users spend as much time avoiding speech as retrieving it. It is not unfeasible, therefore, that in an age of over-abundant information, freedom of speech may not have the same connotations as it did fifty years ago.

  7. Competency-Based Teacher Education; Group 1 Report of the Education Priorities Division of the Speech Communication Association.

    Science.gov (United States)

    Friedrich, Gustav W.; And Others

    This paper focuses on performance-based teacher education in speech communication as an alternative to the traditional method of preparing speech teachers. Following an introductory argument stating the case for performance-based teacher education in speech communication, contents include: "Competency Based Teacher Training: A Perspective on a Set…

  8. Speech rate normalization used to improve speaker verification

    CSIR Research Space (South Africa)

    Van Heerden, CJ

    2006-11-01

    Full Text Available A novel approach to speech rate normalization is presented. Models are constructed to model the way in which speech rate variation of a specific speaker influences the duration of phonemes. The models are evaluated in two ways. Firstly, the mean...

  9. The NCHLT speech corpus of the South African languages

    CSIR Research Space (South Africa)

    Barnard, E

    2014-05-01

    Full Text Available The NCHLT speech corpus contains wide-band speech from approximately 200 speakers per language, in each of the eleven of cial languages of South Africa. We describe the design and development processes that were undertaken in order to develop...

  10. Speech at the Inaugural Meeting of CDSET-CSA

    Institute of Scientific and Technical Information of China (English)

    ZhangQingwei

    2004-01-01

    The inaugural meeting of the Committee of Deep Space Exploration Technology, Chinese Society of Astronautics (CDSET-CSA) was held in Beijing on June 15, 2004. Mr. Zhang Qingwei, President of China Aerospace Science and Technology Corporation (CASC) and President of Chinese Society of Astronautics (CSA) attended the meeting and made an important speech. The following is the full text of the speech.

  11. Developing a broadband automatic speech recognition system for Afrikaans

    CSIR Research Space (South Africa)

    De Wet, Febe

    2011-08-01

    Full Text Available Afrikaans is one of the eleven official languages of South Africa. It is classified as an under-resourced language. No annotated broadband speech corpora currently exist for Afrikaans. This article reports on the development of speech resources...

  12. HMM adaptation for child speech synthesis using ASR data

    CSIR Research Space (South Africa)

    Govender, N

    2015-11-01

    Full Text Available Acquiring large amounts of child speech data is a particularly difficult task. One could therefore consider the possibility to add existing corpora of child speech data to the severely limited resources that are available for developing child voices...

  13. Predicting speech intelligibility in conditions with nonlinearly processed noisy speech

    DEFF Research Database (Denmark)

    Jørgensen, Søren; Dau, Torsten

    2013-01-01

    The speech-based envelope power spectrum model (sEPSM; [1]) was proposed in order to overcome the limitations of the classical speech transmission index (STI) and speech intelligibility index (SII). The sEPSM applies the signal-tonoise ratio in the envelope domain (SNRenv), which was demonstrated...... to successfully predict speech intelligibility in conditions with nonlinearly processed noisy speech, such as processing with spectral subtraction. Moreover, a multiresolution version (mr-sEPSM) was demonstrated to account for speech intelligibility in various conditions with stationary and fluctuating...... from computational auditory scene analysis and further support the hypothesis that the SNRenv is a powerful metric for speech intelligibility prediction....

  14. Phonetic Alphabet for Speech Recognition of Czech

    Directory of Open Access Journals (Sweden)

    J. Uhlir

    1997-12-01

    Full Text Available In the paper we introduce and discuss an alphabet that has been proposed for phonemicly oriented automatic speech recognition. The alphabet, denoted as a PAC (Phonetic Alphabet for Czech consists of 48 basic symbols that allow for distinguishing all major events occurring in spoken Czech language. The symbols can be used both for phonetic transcription of Czech texts as well as for labeling recorded speech signals. From practical reasons, the alphabet occurs in two versions; one utilizes Czech native characters and the other employs symbols similar to those used for English in the DARPA and NIST alphabets.

  15. PRACTICING SPEECH THERAPY INTERVENTION FOR SOCIAL INTEGRATION OF CHILDREN WITH SPEECH DISORDERS

    Directory of Open Access Journals (Sweden)

    Martin Ofelia POPESCU

    2016-11-01

    Full Text Available The article presents a concise speech correction intervention program in of dyslalia in conjunction with capacity development of intra, interpersonal and social integration of children with speech disorders. The program main objectives represent: the potential increasing of individual social integration by correcting speech disorders in conjunction with intra- and interpersonal capacity, the potential growth of children and community groups for social integration by optimizing the socio-relational context of children with speech disorder. In the program were included 60 children / students with dyslalia speech disorders (monomorphic and polymorphic dyslalia, from 11 educational institutions - 6 kindergartens and 5 schools / secondary schools, joined with inter-school logopedic centre (CLI from Targu Jiu city and areas of Gorj district. The program was implemented under the assumption that therapeutic-formative intervention to correct speech disorders and facilitate the social integration will lead, in combination with correct pronunciation disorders, to social integration optimization of children with speech disorders. The results conirm the hypothesis and gives facts about the intervention program eficiency.

  16. Speech Prosody in Persian Language

    Directory of Open Access Journals (Sweden)

    Maryam Nikravesh

    2014-05-01

    Full Text Available Background: In verbal communication in addition of semantic and grammatical aspects, includes: vocabulary, syntax and phoneme, some special voice characteristics were use that called speech prosody. Speech prosody is one of the important factors of communication which includes: intonation, duration, pitch, loudness, stress, rhythm and etc. The aim of this survey is studying some factors of prosody as duration, fundamental frequency range and intonation contour. Materials and Methods: This study is performed with cross-sectional and descriptive-analytic approach. The participants include 134 male and female between 18-30 years old who normally speak Persian. Two sentences include: an interrogative and one declarative sentence were studied. Voice samples were analyzed by Dr. Speech software (real analysis software and data were analyzed by statistical test of unilateral variance analysis and in depended T test, and intonation contour was drawn for sentences. Results: Mean of duration between kinds of sentences had a significant difference. Mean of duration had significant difference between female and male. Fundamental frequency range between kinds of sentences had not significant difference. Fundamental frequency range in female is higher than male. Conclusion: Duration is an affective factor in Persian prosody. The higher fundamental frequency range in female is because of different anatomical and physiological mechanisms in phonation system. In addition higher fundamental frequency range in female is the result of an authority of language use in Farsi female. The end part of intonation contour in yes/no question is rising, in declarative sentence is falling.

  17. Sperry Univac speech communications technology

    Science.gov (United States)

    Medress, Mark F.

    1977-01-01

    Technology and systems for effective verbal communication with computers were developed. A continuous speech recognition system for verbal input, a word spotting system to locate key words in conversational speech, prosodic tools to aid speech analysis, and a prerecorded voice response system for speech output are described.

  18. Voice and Speech after Laryngectomy

    Science.gov (United States)

    Stajner-Katusic, Smiljka; Horga, Damir; Musura, Maja; Globlek, Dubravka

    2006-01-01

    The aim of the investigation is to compare voice and speech quality in alaryngeal patients using esophageal speech (ESOP, eight subjects), electroacoustical speech aid (EACA, six subjects) and tracheoesophageal voice prosthesis (TEVP, three subjects). The subjects reading a short story were recorded in the sound-proof booth and the speech samples…

  19. Environmental Contamination of Normal Speech.

    Science.gov (United States)

    Harley, Trevor A.

    1990-01-01

    Environmentally contaminated speech errors (irrelevant words or phrases derived from the speaker's environment and erroneously incorporated into speech) are hypothesized to occur at a high level of speech processing, but with a relatively late insertion point. The data indicate that speech production processes are not independent of other…

  20. A Pedagogical Treatment of King's "I Have a Dream" Speech: Toward Incorporating Orality in Rhetorical Criticism.

    Science.gov (United States)

    Weitzel, Al R.

    This paper first discusses the ways in which Martin Luther King's "I Have a Dream" speech demonstrates some of the fundamental principles that should be used to teach rhetorical criticism, and then offers a microscopic, intrinsic criticism of the speech. The paper's four sections (1) review the nature of written speech texts; (2) discuss…

  1. The Inaccuracies in the Reprintings of Martin Luther King's "I Have a Dream" Speech.

    Science.gov (United States)

    Bosmajian, Haig

    1982-01-01

    Identifies the additions, deletions, and changes in the reprinted versions of King's classic speech. (Serves both as a warning that published speeches cannot be depended upon to preserve the original text and as a source of reference for teachers using Dr. King.s speech in classroom study.) (PD)

  2. AHP 28: Too Much Loving-Kindness to Repay: Funeral Speeches of the Wenquan Pumi

    Directory of Open Access Journals (Sweden)

    Gerong Pincuo (kɛ́izoŋ pʰiŋtsʰu

    2013-12-01

    Full Text Available Two Pumi funeral speech rituals of the Wenquan Pumi area in northwestern Yunnan Province illustrate the traditional genre of speeches through their use of metaphor and parallellism. The speeches express the central concept of giving and repaying that plays an important role in strengthening social cohesion among Pumi relatives.

  3. Affective Speech for Social Communication: Implementation Challenges in Text-to-Speech for Short Messages

    NARCIS (Netherlands)

    Amin, A.K.; Wang, J.

    2006-01-01

    The flexibility to choose from different modal content presentation will be an important feature in future ubiquitous application. Currently, short messages (e.g. SMS/MMS) are only available in visual form. However, in certain situations, users may like to have these messages presented in audio form

  4. Speech processing in mobile environments

    CERN Document Server

    Rao, K Sreenivasa

    2014-01-01

    This book focuses on speech processing in the presence of low-bit rate coding and varying background environments. The methods presented in the book exploit the speech events which are robust in noisy environments. Accurate estimation of these crucial events will be useful for carrying out various speech tasks such as speech recognition, speaker recognition and speech rate modification in mobile environments. The authors provide insights into designing and developing robust methods to process the speech in mobile environments. Covering temporal and spectral enhancement methods to minimize the effect of noise and examining methods and models on speech and speaker recognition applications in mobile environments.

  5. A Customizable Text Classifier for Text Mining

    Directory of Open Access Journals (Sweden)

    Yun-liang Zhang

    2007-12-01

    Full Text Available Text mining deals with complex and unstructured texts. Usually a particular collection of texts that is specified to one or more domains is necessary. We have developed a customizable text classifier for users to mine the collection automatically. It derives from the sentence category of the HNC theory and corresponding techniques. It can start with a few texts, and it can adjust automatically or be adjusted by user. The user can also control the number of domains chosen and decide the standard with which to choose the texts based on demand and abundance of materials. The performance of the classifier varies with the user's choice.

  6. Speech Recognition Technology for Hearing Disabled Community

    Directory of Open Access Journals (Sweden)

    Tanvi Dua

    2014-09-01

    Full Text Available As the number of people with hearing disabilities are increasing significantly in the world, it is always required to use technology for filling the gap of communication between Deaf and Hearing communities. To fill this gap and to allow people with hearing disabilities to communicate this paper suggests a framework that contributes to the efficient integration of people with hearing disabilities. This paper presents a robust speech recognition system, which converts the continuous speech into text and image. The results are obtained with an accuracy of 95% with the small size vocabulary of 20 greeting sentences of continuous speech form tested in a speaker independent mode. In this testing phase all these continuous sentences were given as live input to the proposed system.

  7. Global Freedom of Speech

    DEFF Research Database (Denmark)

    Binderup, Lars Grassme

    2007-01-01

    , as opposed to a legal norm, that curbs exercises of the right to free speech that offend the feelings or beliefs of members from other cultural groups. The paper rejects the suggestion that acceptance of such a norm is in line with liberal egalitarian thinking. Following a review of the classical liberal...... egalitarian reasons for free speech - reasons from overall welfare, from autonomy and from respect for the equality of citizens - it is argued that these reasons outweigh the proposed reasons for curbing culturally offensive speech. Currently controversial cases such as that of the Danish Cartoon Controversy...

  8. Changes in breathing while listening to read speech: the effect of reader and speech mode

    Science.gov (United States)

    Rochet-Capellan, Amélie; Fuchs, Susanne

    2013-01-01

    The current paper extends previous work on breathing during speech perception and provides supplementary material regarding the hypothesis that adaptation of breathing during perception “could be a basis for understanding and imitating actions performed by other people” (Paccalin and Jeannerod, 2000). The experiments were designed to test how the differences in reader breathing due to speaker-specific characteristics, or differences induced by changes in loudness level or speech rate influence the listener breathing. Two readers (a male and a female) were pre-recorded while reading short texts with normal and then loud speech (both readers) or slow speech (female only). These recordings were then played back to 48 female listeners. The movements of the rib cage and abdomen were analyzed for both the readers and the listeners. Breathing profiles were characterized by the movement expansion due to inhalation and the duration of the breathing cycle. We found that both loudness and speech rate affected each reader’s breathing in different ways. Listener breathing was different when listening to the male or the female reader and to the different speech modes. However, differences in listener breathing were not systematically in the same direction as reader differences. The breathing of listeners was strongly sensitive to the order of presentation of speech mode and displayed some adaptation in the time course of the experiment in some conditions. In contrast to specific alignments of breathing previously observed in face-to-face dialog, no clear evidence for a listener–reader alignment in breathing was found in this purely auditory speech perception task. The results and methods are relevant to the question of the involvement of physiological adaptations in speech perception and to the basic mechanisms of listener–speaker coupling. PMID:24367344

  9. Changes in breathing while listening to read speech: the effect of reader and speech mode.

    Science.gov (United States)

    Rochet-Capellan, Amélie; Fuchs, Susanne

    2013-01-01

    The current paper extends previous work on breathing during speech perception and provides supplementary material regarding the hypothesis that adaptation of breathing during perception "could be a basis for understanding and imitating actions performed by other people" (Paccalin and Jeannerod, 2000). The experiments were designed to test how the differences in reader breathing due to speaker-specific characteristics, or differences induced by changes in loudness level or speech rate influence the listener breathing. Two readers (a male and a female) were pre-recorded while reading short texts with normal and then loud speech (both readers) or slow speech (female only). These recordings were then played back to 48 female listeners. The movements of the rib cage and abdomen were analyzed for both the readers and the listeners. Breathing profiles were characterized by the movement expansion due to inhalation and the duration of the breathing cycle. We found that both loudness and speech rate affected each reader's breathing in different ways. Listener breathing was different when listening to the male or the female reader and to the different speech modes. However, differences in listener breathing were not systematically in the same direction as reader differences. The breathing of listeners was strongly sensitive to the order of presentation of speech mode and displayed some adaptation in the time course of the experiment in some conditions. In contrast to specific alignments of breathing previously observed in face-to-face dialog, no clear evidence for a listener-reader alignment in breathing was found in this purely auditory speech perception task. The results and methods are relevant to the question of the involvement of physiological adaptations in speech perception and to the basic mechanisms of listener-speaker coupling.

  10. The Rhetoric in English Speech

    Institute of Scientific and Technical Information of China (English)

    马鑫

    2014-01-01

    English speech has a very long history and always attached importance of people highly. People usually give a speech in economic activities, political forums and academic reports to express their opinions to investigate or persuade others. English speech plays a rather important role in English literature. The distinct theme of speech should attribute to the rhetoric. It discusses parallelism, repetition and rhetorical question in English speech, aiming to help people appreciate better the charm of them.

  11. From hate speech to respectful speech in the preschool education system

    Directory of Open Access Journals (Sweden)

    Karmen Chakir

    2016-12-01

    Full Text Available In the present article we discuss the issue of hate speech in the society and in educational institutions, as well as the necessity of institutional preschool education that stimulates the development and the use of respectful speech among children. There are many appeals for tolerance and for observance of norms aimed at limiting the hate speech, but the current events in the society and various researches show that those appeals do not always have the desired effects. Therefore, in the article we stress the role of educators who educate children to respect the other with his differences and similarities and to use respectful speech, by pointing out especially the role of an adequate model of educational method, precisely the inductive educational approach.

  12. Effective Presentation Speech Support System for Representing Emphasis-Intention

    Directory of Open Access Journals (Sweden)

    Tomoko Kojiri

    2015-12-01

    Full Text Available A research presentation integrates slides and speech. If these two aspects do not represent the same intention, the presentation will probably fail to effectively explain the presenter’s intention. This paper focuses on the representation of the critical contents in a presentation. In an effective speech, the speaker adds more intonation and stress to emphasize the importance of the slide contents. Audiences recognize that important contents are those that are explained in a stronger voice or that are said after a short pause. However, in ineffective speeches, such voice effects do not always correspond to the important contents that are indicated by slides. On slides, the important contents are represented by levels of text indentation and size, color, and animation. This research develops a presentation speech support system that estimates important contents from slides and voices that might be recognized by audiences and extracts numerical differences. In addition, the system provides comments and feedback to improve speeches.

  13. Cross-modal retrieval of scripted speech audio

    Science.gov (United States)

    Owen, Charles B.; Makedon, Fillia

    1997-12-01

    This paper describes an approach to the problem of searching speech-based digital audio using cross-modal information retrieval. Audio containing speech (speech-based audio) is difficult to search. Open vocabulary speech recognition is advancing rapidly, but cannot yield high accuracy in either search or transcription modalities. However, text can be searched quickly and efficiently with high accuracy. Script- light digital audio is audio that has an available transcription. This is a surprisingly large class of content including legal testimony, broadcasting, dramatic productions and political meetings and speeches. An automatic mechanism for deriving the synchronization between the transcription and the audio allows for very accurate retrieval of segments of that audio. The mechanism described in this paper is based on building a transcription graph from the text and computing biphone probabilities for the audio. A modified beam search algorithm is presented to compute the alignment.

  14. How Prepared is Prepared Enough?

    Science.gov (United States)

    Porter-Levy; Macleod; Rickert

    1996-10-01

    A 17-year-old female was in the final stage in treatment of right unilateral cleft lip and palate. She had undergone a number of previous surgeries. Hearing and speech were good on evaluation, and her social and family situation were deemed excellent. After preparatory orthodontics she underwent a Lefort I maxillary advancement. Surgery was successful and she was admitted into postoperative recovery. However, the lack of adequate preoperative preparation caused traumatic reaction from the patient and her parents: anxiety over appearance, crying, refusal of oral fluids and oral care, refusal of analgesia, and refusal to mobilize. The patience and persistence of hospital staff slowly overcame all adversities and the patient moved on to full and successful recovery, but this case prompted changes in preoperative procedures and involvement of patients and their families in postoperative meal selection, planing, and preparation.

  15. Objective voice and speech analysis of persons with chronic hoarseness by prosodic analysis of speech samples.

    Science.gov (United States)

    Haderlein, Tino; Döllinger, Michael; Matoušek, Václav; Nöth, Elmar

    2016-10-01

    Automatic voice assessment is often performed using sustained vowels. In contrast, speech analysis of read-out texts can be applied to voice and speech assessment. Automatic speech recognition and prosodic analysis were used to find regression formulae between automatic and perceptual assessment of four voice and four speech criteria. The regression was trained with 21 men and 62 women (average age 49.2 years) and tested with another set of 24 men and 49 women (48.3 years), all suffering from chronic hoarseness. They read the text 'Der Nordwind und die Sonne' ('The North Wind and the Sun'). Five voice and speech therapists evaluated the data on 5-point Likert scales. Ten prosodic and recognition accuracy measures (features) were identified which describe all the examined criteria. Inter-rater correlation within the expert group was between r = 0.63 for the criterion 'match of breath and sense units' and r = 0.87 for the overall voice quality. Human-machine correlation was between r = 0.40 for the match of breath and sense units and r = 0.82 for intelligibility. The perceptual ratings of different criteria were highly correlated with each other. Likewise, the feature sets modeling the criteria were very similar. The automatic method is suitable for assessing chronic hoarseness in general and for subgroups of functional and organic dysphonia. In its current version, it is almost as reliable as a randomly picked rater from a group of voice and speech therapists.

  16. Source Separation via Spectral Masking for Speech Recognition Systems

    Directory of Open Access Journals (Sweden)

    Gustavo Fernandes Rodrigues

    2012-12-01

    Full Text Available In this paper we present an insight into the use of spectral masking techniques in time-frequency domain, as a preprocessing step for the speech signal recognition. Speech recognition systems have their performance negatively affected in noisy environments or in the presence of other speech signals. The limits of these masking techniques for different levels of the signal-to-noise ratio are discussed. We show the robustness of the spectral masking techniques against four types of noise: white, pink, brown and human speech noise (bubble noise. The main contribution of this work is to analyze the performance limits of recognition systems  using spectral masking. We obtain an increase of 18% on the speech hit rate, when the speech signals were corrupted by other speech signals or bubble noise, with different signal-to-noise ratio of approximately 1, 10 and 20 dB. On the other hand, applying the ideal binary masks to mixtures corrupted by white, pink and brown noise, results an average growth of 9% on the speech hit rate, with the same different signal-to-noise ratio. The experimental results suggest that the masking spectral techniques are more suitable for the case when it is applied a bubble noise, which is produced by human speech, than for the case of applying white, pink and brown noise.

  17. Speech misperception: speaking and seeing interfere differently with hearing.

    Directory of Open Access Journals (Sweden)

    Takemi Mochida

    Full Text Available Speech perception is thought to be linked to speech motor production. This linkage is considered to mediate multimodal aspects of speech perception, such as audio-visual and audio-tactile integration. However, direct coupling between articulatory movement and auditory perception has been little studied. The present study reveals a clear dissociation between the effects of a listener's own speech action and the effects of viewing another's speech movements on the perception of auditory phonemes. We assessed the intelligibility of the syllables [pa], [ta], and [ka] when listeners silently and simultaneously articulated syllables that were congruent/incongruent with the syllables they heard. The intelligibility was compared with a condition where the listeners simultaneously watched another's mouth producing congruent/incongruent syllables, but did not articulate. The intelligibility of [ta] and [ka] were degraded by articulating [ka] and [ta] respectively, which are associated with the same primary articulator (tongue as the heard syllables. But they were not affected by articulating [pa], which is associated with a different primary articulator (lips from the heard syllables. In contrast, the intelligibility of [ta] and [ka] was degraded by watching the production of [pa]. These results indicate that the articulatory-induced distortion of speech perception occurs in an articulator-specific manner while visually induced distortion does not. The articulator-specific nature of the auditory-motor interaction in speech perception suggests that speech motor processing directly contributes to our ability to hear speech.

  18. Bandwidth Extension of Telephone Speech Aided by Data Embedding

    Directory of Open Access Journals (Sweden)

    David Malah

    2007-01-01

    Full Text Available A system for bandwidth extension of telephone speech, aided by data embedding, is presented. The proposed system uses the transmitted analog narrowband speech signal as a carrier of the side information needed to carry out the bandwidth extension. The upper band of the wideband speech is reconstructed at the receiving end from two components: a synthetic wideband excitation signal, generated from the narrowband telephone speech and a wideband spectral envelope, parametrically represented and transmitted as embedded data in the telephone speech. We propose a novel data embedding scheme, in which the scalar Costa scheme is combined with an auditory masking model allowing high rate transparent embedding, while maintaining a low bit error rate. The signal is transformed to the frequency domain via the discrete Hartley transform (DHT and is partitioned into subbands. Data is embedded in an adaptively chosen subset of subbands by modifying the DHT coefficients. In our simulations, high quality wideband speech was obtained from speech transmitted over a telephone line (characterized by spectral magnitude distortion, dispersion, and noise, in which side information data is transparently embedded at the rate of 600 information bits/second and with a bit error rate of approximately 3⋅10−4. In a listening test, the reconstructed wideband speech was preferred (at different degrees over conventional telephone speech in 92.5% of the test utterances.

  19. Bandwidth Extension of Telephone Speech Aided by Data Embedding

    Directory of Open Access Journals (Sweden)

    Sagi Ariel

    2007-01-01

    Full Text Available A system for bandwidth extension of telephone speech, aided by data embedding, is presented. The proposed system uses the transmitted analog narrowband speech signal as a carrier of the side information needed to carry out the bandwidth extension. The upper band of the wideband speech is reconstructed at the receiving end from two components: a synthetic wideband excitation signal, generated from the narrowband telephone speech and a wideband spectral envelope, parametrically represented and transmitted as embedded data in the telephone speech. We propose a novel data embedding scheme, in which the scalar Costa scheme is combined with an auditory masking model allowing high rate transparent embedding, while maintaining a low bit error rate. The signal is transformed to the frequency domain via the discrete Hartley transform (DHT and is partitioned into subbands. Data is embedded in an adaptively chosen subset of subbands by modifying the DHT coefficients. In our simulations, high quality wideband speech was obtained from speech transmitted over a telephone line (characterized by spectral magnitude distortion, dispersion, and noise, in which side information data is transparently embedded at the rate of 600 information bits/second and with a bit error rate of approximately . In a listening test, the reconstructed wideband speech was preferred (at different degrees over conventional telephone speech in of the test utterances.

  20. Prevalence of speech disorders among Kermanshah primary schoolchildren

    Directory of Open Access Journals (Sweden)

    Akram Soleimani

    2011-07-01

    Full Text Available Background: Speech is the most important aspect of development that related to central nervous system maturation in cognitive sensory motor domain. Disorders of speech may disrupt other aspects of development. The study aimed to investigate the prevalence of speech disorders among primary schoolchildren in Kermanshah during 2008-9. Methods: Six hundred schoolchildren from three educational regions of the city were selected using multi-stage cluster sampling method. In an hour assessment session, an expert Speech-Language pathologist evaluated the spontaneous speech of children by phonetic test and specialist judgment. Results: 39.5% of cases were female and 60.5% male. Mean age of students was 9.12±1.52 years. 67 out of 600 studied schoolchildren (11.2 % were diagnosed as speech disordered. Prevalence of speech disorders among boys and girls was 11 and 11.3 percent respectively. There was no significant difference in speech disorders prevalence according to age and sex.Conclusion: It’s concluded that prevalence of speech disorders among primary schoolchildren in Kermanshah was lower than similar reports in other studies.

  1. SPEECH ACT OF ILTIFAT AND ITS INDONESIAN TRANSLATION PROBLEMS

    Directory of Open Access Journals (Sweden)

    Zaka Al Farisi

    2015-01-01

    Full Text Available Abstract: Iltifat (shifting speech act is distinctive and considered unique style of Arabic. It has potential errors when it is translated into Indonesian. Therefore, translation of iltifat speech act into another language can be an important issue. The objective of the study is to know translation procedures/techniques and ideology required in dealing with iltifat speech act. This research is directed at translation as a cognitive product of a translator. The data used in the present study were the corpus of Koranic verses that contain iltifat speech act along with their translation. Data analysis typically used descriptive-evaluative method with content analysis model. The data source of this research consisted of the Koran and its translation. The purposive sampling technique was employed, with the sample of the iltifat speech act contained in the Koran. The results showed that more than 60% of iltifat speech act were translated by using literal procedure. The significant number of literal translation of the verses asserts that the Ministry of Religious Affairs tended to use literal method of translation. In other words, the Koran translation made by the Ministry of Religious Affairs tended to be oriented to the source language in dealing with iltifat speech act. The number of the literal procedure used shows a tendency of foreignization ideology. Transitional pronouns contained in the iltifat speech act can be clearly translated when thick translations were used in the form of description in parentheses. In this case, explanation can be a choice in translating iltifat speech act.

  2. Sound frequency affects speech emotion perception: Results from congenital amusia

    Directory of Open Access Journals (Sweden)

    Sydney eLolli

    2015-09-01

    Full Text Available Congenital amusics, or tone-deaf individuals, show difficulty in perceiving and producing small pitch differences. While amusia has marked effects on music perception, its impact on speech perception is less clear. Here we test the hypothesis that individual differences in pitch perception affect judgment of emotion in speech, by applying band-pass filters to spoken statements of emotional speech. A norming study was first conducted on Mechanical Turk to ensure that the intended emotions from the Macquarie Battery for Evaluation of Prosody (MBEP were reliably identifiable by US English speakers. The most reliably identified emotional speech samples were used in in Experiment 1, in which subjects performed a psychophysical pitch discrimination task, and an emotion identification task under band-pass and unfiltered speech conditions. Results showed a significant correlation between pitch discrimination threshold and emotion identification accuracy for band-pass filtered speech, with amusics (defined here as those with a pitch discrimination threshold > 16 Hz performing worse than controls. This relationship with pitch discrimination was not seen in unfiltered speech conditions. Given the dissociation between band-pass filtered and unfiltered speech conditions, we inferred that amusics may be compensating for poorer pitch perception by using speech cues that are filtered out in this manipulation.

  3. Speech nasality and nasometry in cleft lip and palate

    Directory of Open Access Journals (Sweden)

    Fabiane Rodrigues Larangeira

    Full Text Available ABSTRACT INTRODUCTION: Perceptual evaluation is considered the gold standard to evaluate speech nasality. Several procedures are used to collect and analyze perceptual data, which makes it susceptible to errors. Therefore, there has been an increasing desire to find methods that can improve the assessment. OBJECTIVE: To describe and compare the results of speech nasality obtained by assessments of live speech, the Test of Hypernasality (THYPER, assessments of audio recorded speech, and nasometry. METHODS: A retrospective study consisting of 331 patients with operated unilateral cleft lip and palate. Speech nasality was assessed by four methods of assessment: live perceptual judgement, THYPER, audio-recorded speech sample judgement by multiple judges, and nasometry. All data were collected from medical records of patients, with the exception of the speech sample recording assessment, which was carried out by multiple judges. RESULTS: The results showed that the highest percentages of absence of hypernasality were obtained from judgements performed live and from the THYPER, with equal results between them (79%. Lower percentages were obtained from the recordings by judges (66% and from nasometry (57%. CONCLUSION: The best results among the four speech nasality evaluation methods were obtained for the ones performed live (live nasality judgement by a speech pathologist and THYPER.

  4. Neural oscillations carry speech rhythm through to comprehension

    Directory of Open Access Journals (Sweden)

    Jonathan E Peelle

    2012-09-01

    Full Text Available A key feature of speech is the quasi-regular rhythmic information contained in its slow amplitude modulations. In this article we review the information conveyed by speech rhythm, and the role of ongoing brain oscillations in listeners’ processing of this content. Our starting point is the fact that speech is inherently temporal, and that rhythmic information conveyed by the amplitude envelope contains important markers for place and manner of articulation, segmental information, and speech rate. Behavioral studies demonstrate that amplitude envelope information is relied upon by listeners and plays a key role in speech intelligibility. Extending behavioral findings, data from neuroimaging—particularly electroencephalography (EEG and magnetoencephalography (MEG—point to phase locking by ongoing cortical oscillations to low-frequency information (~4–8 Hz in the speech envelope. This phase modulation effectively encodes a prediction of when important events (such as stressed syllables are likely to occur, and acts to increase sensitivity to these relevant acoustic cues. We suggest a framework through which such neural entrainment to speech rhythm can explain effects of speech rate on word and on segment perception (i.e., that the perception of phonemes and words in connected speech are influenced by preceding speech rate. Neuroanatomically, acoustic amplitude modulations are processed largely bilaterally in auditory cortex, with intelligible speech resulting in additional recruitment of left hemisphere regions. Notable among these is lateral anterior temporal cortex, which we propose functions in a domain-general fashion to support ongoing memory and integration of meaningful input. Together, the reviewed evidence suggests that low frequency oscillations in the acoustic speech signal form the foundation of a rhythmic hierarchy supporting spoken language, mirrored by phase-locked oscillations in the human brain.

  5. Visual speech form influences the speed of auditory speech processing.

    Science.gov (United States)

    Paris, Tim; Kim, Jeesun; Davis, Chris

    2013-09-01

    An important property of visual speech (movements of the lips and mouth) is that it generally begins before auditory speech. Research using brain-based paradigms has demonstrated that seeing visual speech speeds up the activation of the listener's auditory cortex but it is not clear whether these observed neural processes link to behaviour. It was hypothesized that the very early portion of visual speech (occurring before auditory speech) will allow listeners to predict the following auditory event and so facilitate the speed of speech perception. This was tested in the current behavioural experiments. Further, we tested whether the salience of the visual speech played a role in this speech facilitation effect (Experiment 1). We also determined the relative contributions that visual form (what) and temporal (when) cues made (Experiment 2). The results showed that visual speech cues facilitated response times and that this was based on form rather than temporal cues. Copyright © 2013 Elsevier Inc. All rights reserved.

  6. Anxiety and ritualized speech

    Science.gov (United States)

    Lalljee, Mansur; Cook, Mark

    1975-01-01

    The experiment examines the effects on a number of words that seem irrelevant to semantic communication. The Units of Ritualized Speech (URSs) considered are: 'I mean', 'in fact', 'really', 'sort of', 'well' and 'you know'. (Editor)

  7. Anxiety and ritualized speech

    Science.gov (United States)

    Lalljee, Mansur; Cook, Mark

    1975-01-01

    The experiment examines the effects on a number of words that seem irrelevant to semantic communication. The Units of Ritualized Speech (URSs) considered are: 'I mean', 'in fact', 'really', 'sort of', 'well' and 'you know'. (Editor)

  8. HATE SPEECH AS COMMUNICATION

    National Research Council Canada - National Science Library

    Gladilin Aleksey Vladimirovich

    2012-01-01

    The purpose of the paper is a theoretical comprehension of hate speech from communication point of view, on the one hand, and from the point of view of prejudice, stereotypes and discrimination on the other...

  9. Speech intelligibility in hospitals.

    Science.gov (United States)

    Ryherd, Erica E; Moeller, Michael; Hsu, Timothy

    2013-07-01

    Effective communication between staff members is key to patient safety in hospitals. A variety of patient care activities including admittance, evaluation, and treatment rely on oral communication. Surprisingly, published information on speech intelligibility in hospitals is extremely limited. In this study, speech intelligibility measurements and occupant evaluations were conducted in 20 units of five different U.S. hospitals. A variety of unit types and locations were studied. Results show that overall, no unit had "good" intelligibility based on the speech intelligibility index (SII > 0.75) and several locations found to have "poor" intelligibility (SII speech intelligibility across a variety of hospitals and unit types, offers some evidence of the positive impact of absorption on intelligibility, and identifies areas for future research.

  10. Speech disorders - children

    Science.gov (United States)

    ... this page: //medlineplus.gov/ency/article/001430.htm Speech disorders - children To use the sharing features on ... 2017, A.D.A.M., Inc. Duplication for commercial use must be authorized in writing by ADAM ...

  11. Speech impairment (adult)

    Science.gov (United States)

    ... this page: //medlineplus.gov/ency/article/003204.htm Speech impairment (adult) To use the sharing features on ... 2017, A.D.A.M., Inc. Duplication for commercial use must be authorized in writing by ADAM ...

  12. Filled pause refinement based on the pronunciation probability for lecture speech.

    Science.gov (United States)

    Long, Yan-Hua; Ye, Hong

    2015-01-01

    Nowadays, although automatic speech recognition has become quite proficient in recognizing or transcribing well-prepared fluent speech, the transcription of speech that contains many disfluencies remains problematic, such as spontaneous conversational and lecture speech. Filled pauses (FPs) are the most frequently occurring disfluencies in this type of speech. Most recent studies have shown that FPs are widely believed to increase the error rates for state-of-the-art speech transcription, primarily because most FPs are not well annotated or provided in training data transcriptions and because of the similarities in acoustic characteristics between FPs and some common non-content words. To enhance the speech transcription system, we propose a new automatic refinement approach to detect FPs in British English lecture speech transcription. This approach combines the pronunciation probabilities for each word in the dictionary and acoustic language model scores for FP refinement through a modified speech recognition forced-alignment framework. We evaluate the proposed approach on the Reith Lectures speech transcription task, in which only imperfect training transcriptions are available. Successful results are achieved for both the development and evaluation datasets. Acoustic models trained on different styles of speech genres have been investigated with respect to FP refinement. To further validate the effectiveness of the proposed approach, speech transcription performance has also been examined using systems built on training data transcriptions with and without FP refinement.

  13. Syntactic error modeling and scoring normalization in speech recognition: Error modeling and scoring normalization in the speech recognition task for adult literacy training

    Science.gov (United States)

    Olorenshaw, Lex; Trawick, David

    1991-01-01

    The purpose was to develop a speech recognition system to be able to detect speech which is pronounced incorrectly, given that the text of the spoken speech is known to the recognizer. Better mechanisms are provided for using speech recognition in a literacy tutor application. Using a combination of scoring normalization techniques and cheater-mode decoding, a reasonable acceptance/rejection threshold was provided. In continuous speech, the system was tested to be able to provide above 80 pct. correct acceptance of words, while correctly rejecting over 80 pct. of incorrectly pronounced words.

  14. Recognizing GSM Digital Speech

    OpenAIRE

    Gallardo-Antolín, Ascensión; Peláez-Moreno, Carmen; Díaz-de-María, Fernando

    2005-01-01

    The Global System for Mobile (GSM) environment encompasses three main problems for automatic speech recognition (ASR) systems: noisy scenarios, source coding distortion, and transmission errors. The first one has already received much attention; however, source coding distortion and transmission errors must be explicitly addressed. In this paper, we propose an alternative front-end for speech recognition over GSM networks. This front-end is specially conceived to be effective against source c...

  15. Speech Compression and Synthesis

    Science.gov (United States)

    1980-10-01

    phonological rules combined with diphone improved the algorithms used by the phonetic synthesis prog?Im for gain normalization and time... phonetic vocoder, spectral template. i0^Th^TreprtTörc"u’d1sTuV^ork for the past two years on speech compression’and synthesis. Since there was an...from Block 19: speech recognition, pnoneme recogmtion. initial design for a phonetic recognition program. We also recorded ana partially labeled a

  16. Recognizing GSM Digital Speech

    OpenAIRE

    2005-01-01

    The Global System for Mobile (GSM) environment encompasses three main problems for automatic speech recognition (ASR) systems: noisy scenarios, source coding distortion, and transmission errors. The first one has already received much attention; however, source coding distortion and transmission errors must be explicitly addressed. In this paper, we propose an alternative front-end for speech recognition over GSM networks. This front-end is specially conceived to be effective against source c...

  17. A Case with Partial 9p Trisomy and Speech Impairment

    Directory of Open Access Journals (Sweden)

    Mehmet Elbistan

    2014-03-01

    Full Text Available In this study, we aimed to discuss the relationships between his phenotypic anomalies and der(15, ish t(9;15(p12;q10 balanced reciprocal translocation entity that determined from 5 year old male child who referred to our laboratory for cytogenetic analysis. He has complaints speech impairment and growth retardation. And whose family has no phenotypic anomalies. After assessing the case’s pedigree, the preparations obtained from case and relatives using peripheric blood culture method and evaluated with GTG banding. Cytogenetic analyses of her father and mother revealed normal karyotypes. We convinced that the balanced reciprocal translocation may be as a result of de novo mechanism. So, we discussed relationship between phenotypic anomalies and genetic characteristics of case in the light of the literature.

  18. LIBERDADE DE EXPRESSÃO E DISCURSO DO ÓDIO NO BRASIL / FREE SPEECH AND HATE SPEECH IN BRAZIL

    Directory of Open Access Journals (Sweden)

    Nevita Maria Pessoa de Aquino Franca Luna

    2014-12-01

    Full Text Available The purpose of this article is to analyze the restriction of free speech when it comes close to hate speech. In this perspective, the aim of this study is to answer the question: what is the understanding adopted by the Brazilian Supreme Court in cases involving the conflict between free speech and hate speech? The methodology combines a bibliographic review on the theoretical assumptions of the research (concept of free speech and hate speech, and understanding of the rights of defense of traditionally discriminated minorities and empirical research (documental and jurisprudential analysis of judged cases of American Court, German Court and Brazilian Court. Firstly, free speech is discussed, defining its meaning, content and purpose. Then, the hate speech is pointed as an inhibitor element of free speech for offending members of traditionally discriminated minorities, who are outnumbered or in a situation of cultural, socioeconomic or political subordination. Subsequently, are discussed some aspects of American (negative freedom and German models (positive freedom, to demonstrate that different cultures adopt different legal solutions. At the end, it is concluded that there is an approximation of the Brazilian understanding with the German doctrine, from the analysis of landmark cases as the publisher Siegfried Ellwanger (2003 and the Samba School Unidos do Viradouro (2008. The Brazilian comprehension, a multicultural country made up of different ethnicities, leads to a new process of defending minorities who, despite of involving the collision of fundamental rights (dignity, equality and freedom, is still restrained by incompatible barriers of a contemporary pluralistic democracy.

  19. Speech evaluation in children with temporomandibular disorders

    Directory of Open Access Journals (Sweden)

    Raquel Aparecida Pizolato

    2011-10-01

    Full Text Available OBJECTIVE: The aims of this study were to evaluate the influence of temporomandibular disorders (TMD on speech in children, and to verify the influence of occlusal characteristics. MATERIAL AND METHODS: Speech and dental occlusal characteristics were assessed in 152 Brazilian children (78 boys and 74 girls, aged 8 to 12 (mean age 10.05 ± 1.39 years with or without TMD signs and symptoms. The clinical signs were evaluated using the Research Diagnostic Criteria for TMD (RDC/TMD (axis I and the symptoms were evaluated using a questionnaire. The following groups were formed: Group TMD (n=40, TMD signs and symptoms (Group S and S, n=68, TMD signs or symptoms (Group S or S, n=33, and without signs and symptoms (Group N, n=11. Articulatory speech disorders were diagnosed during spontaneous speech and repetition of the words using the "Phonological Assessment of Child Speech" for the Portuguese language. It was also applied a list of 40 phonological balanced words, read by the speech pathologist and repeated by the children. Data were analyzed by descriptive statistics, Fisher's exact or Chi-square tests (α=0.05. RESULTS: A slight prevalence of articulatory disturbances, such as substitutions, omissions and distortions of the sibilants /s/ and /z/, and no deviations in jaw lateral movements were observed. Reduction of vertical amplitude was found in 10 children, the prevalence being greater in TMD signs and symptoms children than in the normal children. The tongue protrusion in phonemes /t/, /d/, /n/, /l/ and frontal lips in phonemes /s/ and /z/ were the most prevalent visual alterations. There was a high percentage of dental occlusal alterations. CONCLUSIONS: There was no association between TMD and speech disorders. Occlusal alterations may be factors of influence, allowing distortions and frontal lisp in phonemes /s/ and /z/ and inadequate tongue position in phonemes /t/; /d/; /n/; /l/.

  20. Kannada Phonemes to Speech Dictionary: Statistical Approach

    Directory of Open Access Journals (Sweden)

    Mallamma V. Reddy

    2017-01-01

    Full Text Available The input or output of a natural Language processing system can be either written text or speech. To process written text we need to analyze: lexical, syntactic, semantic knowledge about the language, discourse information, real world knowledge to process spoken language, we need to analyze everything required to process written text, along with the challenges of speech recognition and speech synthesis. This paper describes how articulatory phonetics of Kannada is used to generate the phoneme to speech dictionary for Kannada; the statistical computational approach is used to map the elements which are taken from input query or documents. The articulatory phonetics is the place of articulation of a consonant. It is the point of contact where an obstruction occurs in the vocal tract between an articulatory gesture, an active articulator, typically some part of the tongue, and a passive location, typically some part of the roof of the mouth. Along with the manner of articulation and the phonation, this gives the consonant its distinctive sound. The results are presented for the same.

  1. Thematic Progression and Textual Coherence in Speech

    Institute of Scientific and Technical Information of China (English)

    周云鹤

    2014-01-01

    Thematic progression can affect the flow of information and directly affect the discourse coherence. This paper analyzes thematic progression patterns of a speech about“people and nature” in the “CCTV Cup”English Speaking Contest and finds that there are three progression patterns in this text, which are parallel progression, continuous progression, and crossing progression.

  2. A distributed approach to speech resource collection

    CSIR Research Space (South Africa)

    Molapo, R

    2013-12-01

    Full Text Available The authors describe the integration of several tools to enable the end-to-end development of an Automatic Speech Recognition system in a typical under-resourced language. The authors analyse the data acquired by each of the tools and develop an ASR...

  3. CREATIVE STYLISTICS AND CREATIVE SPEECH TECHNOLOGIES

    Directory of Open Access Journals (Sweden)

    Natalia A. Kupina

    2016-01-01

    Full Text Available The article develops linguo-aesthetic ideas of Professor V.P. Grigoriev that are connected to the wider interpretation of poetic language and isolation of a ‘creatheme’ as a unit of poetic language. Analyzing data from “Integrum” database, the article discovers specific characteristics of the lexical compatibility of the popular word kreativny (‘creative’. The study is reconstructing a fragment of the worldview that reflects the current conventional understanding of the subjects and areas of creative activity, including creative speech, as well as products of this activity prevailing at the present time. The article raises a question about the formation and development of creative stylistics; of its object, subject, goals, and vector of development. For the purposes of specific stylistical analysis, the article brings up isolated creathemes from colloquial speech of factory workers, newspaper writing, official/business speech, advertising texts, applied poetry texts, and mass literature. The study includes analysis of newspaper headlines bringing under scrutiny the phenomenon of paronymic attraction. In the process of interpretation of the creative speech technologies in women prose, we prove, basing on ‘packaging material’ notion coined by V.P. Grigoriev, that creathemes with aesthetic meanings which strengthen axiologic function of sentence and/or text serve as this sort of ‘packaging’ in writing. 

  4. Basic speech recognition for spoken dialogues

    CSIR Research Space (South Africa)

    Van Heerden, C

    2009-09-01

    Full Text Available speech recognisers for a diverse multitude of languages. The paper investigates the feasibility of developing small-vocabulary speaker-independent ASR systems designed for use in a telephone-based information system, using ten resource-scarce languages...

  5. SPEECH DELAY IN THE PRACTICE OF A PAEDIATRICIAN AND CHILD’S NEUROLOGIST

    Directory of Open Access Journals (Sweden)

    N. N. Zavadenko

    2015-01-01

    Full Text Available The article describes the main clinical forms and causes of speech delay in children. It presents modern data on the role of neurobiological factors in the speech delay pathogenesis, including early organic damage to the central nervous system due to the pregnancy and childbirth pathology, as well as genetic mechanisms. For early and accurate diagnosis of speech disorders in children, you need to consider normal patterns of speech development. The article presents indicators of pre-speech and speech development in children and describes the screening method for determining the speech delay. The main areas of complex correction are speech therapy, psycho-pedagogical and psychotherapeutic assistance, as well as pharmaceutical treatment. The capabilities of drug therapy for dysphasia (alalia are shown. 

  6. Mapping Speech Spectra from Throat Microphone to Close-Speaking Microphone: A Neural Network Approach

    Directory of Open Access Journals (Sweden)

    B. Yegnanarayana

    2007-01-01

    Full Text Available Speech recorded from a throat microphone is robust to the surrounding noise, but sounds unnatural unlike the speech recorded from a close-speaking microphone. This paper addresses the issue of improving the perceptual quality of the throat microphone speech by mapping the speech spectra from the throat microphone to the close-speaking microphone. A neural network model is used to capture the speaker-dependent functional relationship between the feature vectors (cepstral coefficients of the two speech signals. A method is proposed to ensure the stability of the all-pole synthesis filter. Objective evaluations indicate the effectiveness of the proposed mapping scheme. The advantage of this method is that the model gives a smooth estimate of the spectra of the close-speaking microphone speech. No distortions are perceived in the reconstructed speech. This mapping technique is also used for bandwidth extension of telephone speech.

  7. Lope and the Battle-Speech

    Directory of Open Access Journals (Sweden)

    Juan Carlos Iglesias-Zoido

    2013-05-01

    Full Text Available This article analyzes the way in which Lope de Vega conceives in his theater the pre-battle harangue, the most characteristic speech in ancient and renaissance historiography. Having this aim in mind, I have analyzed the role played by this type of speech in a group of plays dealing with historical and military subjects. These plays were written in a period when Lope was particularly interested in historical issues: La Santa Liga (1598-1603, Arauco domado (1599, El asalto de Mastrique (1595-1606 and Los Guanches de Tenerife (1604-1606.

  8. Sentence Clustering Using Parts-of-Speech

    Directory of Open Access Journals (Sweden)

    Richard Khoury

    2012-02-01

    Full Text Available Clustering algorithms are used in many Natural Language Processing (NLP tasks. They have proven to be popular and effective tools to use to discover groups of similar linguistic items. In this exploratory paper, we propose a new clustering algorithm to automatically cluster together similar sentences based on the sentences’ part-of-speech syntax. The algorithm generates and merges together the clusters using a syntactic similarity metric based on a hierarchical organization of the parts-of-speech. We demonstrate the features of this algorithm by implementing it in a question type classification system, in order to determine the positive or negative impact of different changes to the algorithm.

  9. Cohesive and coherent connected speech deficits in mild stroke.

    Science.gov (United States)

    Barker, Megan S; Young, Breanne; Robinson, Gail A

    2017-05-01

    Spoken language production theories and lesion studies highlight several important prelinguistic conceptual preparation processes involved in the production of cohesive and coherent connected speech. Cohesion and coherence broadly connect sentences with preceding ideas and the overall topic. Broader cognitive mechanisms may mediate these processes. This study aims to investigate (1) whether stroke patients without aphasia exhibit impairments in cohesion and coherence in connected speech, and (2) the role of attention and executive functions in the production of connected speech. Eighteen stroke patients (8 right hemisphere stroke [RHS]; 6 left [LHS]) and 21 healthy controls completed two self-generated narrative tasks to elicit connected speech. A multi-level analysis of within and between-sentence processing ability was conducted. Cohesion and coherence impairments were found in the stroke group, particularly RHS patients, relative to controls. In the whole stroke group, better performance on the Hayling Test of executive function, which taps verbal initiation/suppression, was related to fewer propositional repetitions and global coherence errors. Better performance on attention tasks was related to fewer propositional repetitions, and decreased global coherence errors. In the RHS group, aspects of cohesive and coherent speech were associated with better performance on attention tasks. Better Hayling Test scores were related to more cohesive and coherent speech in RHS patients, and more coherent speech in LHS patients. Thus, we documented connected speech deficits in a heterogeneous stroke group without prominent aphasia. Our results suggest that broader cognitive processes may play a role in producing connected speech at the early conceptual preparation stage. Copyright © 2017 Elsevier Inc. All rights reserved.

  10. Computer-based speech therapy for childhood speech sound disorders.

    Science.gov (United States)

    Furlong, Lisa; Erickson, Shane; Morris, Meg E

    2017-07-01

    With the current worldwide workforce shortage of Speech-Language Pathologists, new and innovative ways of delivering therapy to children with speech sound disorders are needed. Computer-based speech therapy may be an effective and viable means of addressing service access issues for children with speech sound disorders. To evaluate the efficacy of computer-based speech therapy programs for children with speech sound disorders. Studies reporting the efficacy of computer-based speech therapy programs were identified via a systematic, computerised database search. Key study characteristics, results, main findings and details of computer-based speech therapy programs were extracted. The methodological quality was evaluated using a structured critical appraisal tool. 14 studies were identified and a total of 11 computer-based speech therapy programs were evaluated. The results showed that computer-based speech therapy is associated with positive clinical changes for some children with speech sound disorders. There is a need for collaborative research between computer engineers and clinicians, particularly during the design and development of computer-based speech therapy programs. Evaluation using rigorous experimental designs is required to understand the benefits of computer-based speech therapy. The reader will be able to 1) discuss how computerbased speech therapy has the potential to improve service access for children with speech sound disorders, 2) explain the ways in which computer-based speech therapy programs may enhance traditional tabletop therapy and 3) compare the features of computer-based speech therapy programs designed for different client populations. Copyright © 2017 Elsevier Inc. All rights reserved.

  11. Automatic Speech Recognition Systems for the Evaluation of Voice and Speech Disorders in Head and Neck Cancer

    Directory of Open Access Journals (Sweden)

    Andreas Maier

    2010-01-01

    Full Text Available In patients suffering from head and neck cancer, speech intelligibility is often restricted. For assessment and outcome measurements, automatic speech recognition systems have previously been shown to be appropriate for objective and quick evaluation of intelligibility. In this study we investigate the applicability of the method to speech disorders caused by head and neck cancer. Intelligibility was quantified by speech recognition on recordings of a standard text read by 41 German laryngectomized patients with cancer of the larynx or hypopharynx and 49 German patients who had suffered from oral cancer. The speech recognition provides the percentage of correctly recognized words of a sequence, that is, the word recognition rate. Automatic evaluation was compared to perceptual ratings by a panel of experts and to an age-matched control group. Both patient groups showed significantly lower word recognition rates than the control group. Automatic speech recognition yielded word recognition rates which complied with experts' evaluation of intelligibility on a significant level. Automatic speech recognition serves as a good means with low effort to objectify and quantify the most important aspect of pathologic speech—the intelligibility. The system was successfully applied to voice and speech disorders.

  12. A Study of Rate of Speech and Intelligibility of Speech in Hearing impaired Pupils of Baghcheban School of Zahedan, Iran

    Directory of Open Access Journals (Sweden)

    Mahla SAEDI

    2013-12-01

    Full Text Available Hearing loss results in disruption of coordination in muscles and structures engaged in speech timing tasks and consequently acquisition of speech sounds is affected. The purpose of this study is to investigate and compare the speed and intelligibility of speech in hearing impaired and normal hearing Persian speaking elementary school boys and girls. 33 hearing impaired students are randomly selected from three hearing impaired groups (profound, severe and moderately-severe and compared with 60 normal hearing students. The speed of speech was measured according to reading Words Per Minute (WPM, and speech intelligibility was tested by a 5-rank scale method. As expected, the normal hearing had more speed and intelligibility of speech in contrast to the hearing impaired group. Also hearing impaired boys had a better speed and intelligibility of speech compared to hearing impaired girls but in normal hearing group, girls had better speed. The amount of P-value for moderately-severe and sever hearing impaired was 0.006 and this amount for profound and moderately-severe hearing impaired was 0.002, so p-value is <0.05 and significantly important. Profound hearing impaired group read the text more slowly and had lower speech intelligibility than other two hearing impaired groups.

  13. Freedom of Speech: A Selected, Annotated Basic Bibliography.

    Science.gov (United States)

    Tedford, Thomas L.

    This bibliography lists 36 books related to problems of freedom of speech. General sources (history, analyses, texts, and anthologies) are listed separately from those dealing with censorship of obscenity and pornography. Each entry is briefly annotated. (AA)

  14. Phonetic alignment for speech synthesis in under-resourced languages

    CSIR Research Space (South Africa)

    Van Niekerk, DR

    2009-09-01

    Full Text Available The rapid development of concatenative speech synthesis systems in resource scarce languages requires an efficient and accurate solution with regard to automated phonetic alignment. However, in this context corpora are often minimally designed due...

  15. Developing speech resources from parliamentary data for South African english

    CSIR Research Space (South Africa)

    De Wet, Febe

    2016-05-01

    Full Text Available Workshop on Spoken Language Technology for Under-resourced Languages, SLTU 2016, 9-12 May 2016, Yogyakarta, Indonesia Developing Speech Resources from Parliamentary Data for South African English Febe de Wet*, Jaco Badenhorst, Thipe Modipa Human...

  16. Markers of Deception in Italian Speech

    Directory of Open Access Journals (Sweden)

    Katelyn eSpence

    2012-10-01

    Full Text Available Lying is a universal activity and the detection of lying a universal concern. Presently, there is great interest in determining objective measures of deception. The examination of speech, in particular, holds promise in this regard; yet, most of what we know about the relationship between speech and lying is based on the assessment of English-speaking participants. Few studies have examined indicators of deception in languages other than English. The world’s languages differ in significant ways, and cross-linguistic studies of deceptive communications are a research imperative. Here we review some of these differences amongst the world’s languages, and provide an overview of a number of recent studies demonstrating that cross-linguistic research is a worthwhile endeavour. In addition, we report the results of an empirical investigation of pitch, response latency, and speech rate as cues to deception in Italian speech. True and false opinions were elicited in an audio-taped interview. A within subjects analysis revealed no significant difference between the average pitch of the two conditions; however, speech rate was significantly slower, while response latency was longer, during deception compared with truth-telling. We explore the implications of these findings and propose directions for future research, with the aim of expanding the cross-linguistic branch of research on markers of deception.

  17. Music and speech prosody: A common rhythm

    Directory of Open Access Journals (Sweden)

    Maija eHausen

    2013-09-01

    Full Text Available Disorders of music and speech perception, known as amusia and aphasia, have traditionally been regarded as dissociated deficits based on studies of brain damaged patients. This has been taken as evidence that music and speech are perceived by largely separate and independent networks in the brain. However, recent studies of congenital amusia have broadened this view by showing that the deficit is associated with problems in perceiving speech prosody, especially intonation and emotional prosody. In the present study the association between the perception of music and speech prosody was investigated with healthy Finnish adults (n = 61 using an on-line music perception test including the Scale subtest of Montreal Battery of Evaluation of Amusia (MBEA and Off-Beat and Out-of-key tasks as well as a prosodic verbal task that measures the perception of word stress. Regression analyses showed that there was a clear association between prosody perception and music perception, especially in the domain of rhythm perception. This association was evident after controlling for music education, age, pitch perception, visuospatial perception and working memory. Pitch perception was significantly associated with music perception but not with prosody perception. The association between music perception and visuospatial perception (measured using analogous tasks was less clear. Overall, the pattern of results indicates that there is a robust link between music and speech perception and that this link can be mediated by rhythmic cues (time and stress.

  18. Speech production as state feedback control

    Directory of Open Access Journals (Sweden)

    John F Houde

    2011-10-01

    Full Text Available Spoken language exists because of a remarkable neural process. Inside a speaker’s brain, an intended message gives rise to neural signals activating the muscles of the vocal tract. The process is remarkable because these muscles are activated in just the right way that the vocal tract produces sounds a listener understands as the intended message. What is the best approach to understanding the neural substrate of this crucial motor control process? One of the key recent modeling developments in neuroscience has been the use of state feedback control (SFC theory to explain the role of the CNS in motor control. SFC postulates that the CNS controls motor output by (1 estimating the current dynamic state of the thing (e.g., arm being controlled, and (2 generating controls based on this estimated state. SFC has successfully predicted a great range of non-speech motor phenomena, but as yet has not received attention in the speech motor control community. Here, we review some of the key characteristics of speech motor control and what they say about the role of the CNS in the process. We then discuss prior efforts to model the role of CNS in speech motor control, and argue that these models have inherent limitations – limitations that are overcome by an SFC model of speech motor control which we describe. We conclude by discussing a plausible neural substrate of our model.

  19. PERSON DEIXIS IN USA PRESIDENTIAL CAMPAIGN SPEECHES

    Directory of Open Access Journals (Sweden)

    Nanda Anggarani Putri

    2015-06-01

    Full Text Available This study investigates the use of person deixis in presidential campaign speeches. This study is important because the use of person deixis in political speeches has been proved by many studies to give significant effects to the audience. The study largely employs a descriptive qualitative method. However, it also employs a simple quantitative method in calculating the number of personal pronouns used in the speeches and their percentages. The data for the study were collected from the transcriptions of six presidential campaign speeches of Barack Obama and Mitt Romney during the campaign rally in various places across the United States of America in July, September, and November 2012. The results of this study show that the presidential candidates make the best use of pronouns as a way to promote themselves and to attack their opponents. The results also suggest that the use of pronouns in the speeches enables the candidates to construct positive identity and reality, which are favorable to them and make them appear more eligible for the position.

  20. Speech Evoked Auditory Brainstem Response in Stuttering

    Directory of Open Access Journals (Sweden)

    Ali Akbar Tahaei

    2014-01-01

    Full Text Available Auditory processing deficits have been hypothesized as an underlying mechanism for stuttering. Previous studies have demonstrated abnormal responses in subjects with persistent developmental stuttering (PDS at the higher level of the central auditory system using speech stimuli. Recently, the potential usefulness of speech evoked auditory brainstem responses in central auditory processing disorders has been emphasized. The current study used the speech evoked ABR to investigate the hypothesis that subjects with PDS have specific auditory perceptual dysfunction. Objectives. To determine whether brainstem responses to speech stimuli differ between PDS subjects and normal fluent speakers. Methods. Twenty-five subjects with PDS participated in this study. The speech-ABRs were elicited by the 5-formant synthesized syllable/da/, with duration of 40 ms. Results. There were significant group differences for the onset and offset transient peaks. Subjects with PDS had longer latencies for the onset and offset peaks relative to the control group. Conclusions. Subjects with PDS showed a deficient neural timing in the early stages of the auditory pathway consistent with temporal processing deficits and their abnormal timing may underlie to their disfluency.

  1. A Statistical Approach to Automatic Speech Summarization

    Directory of Open Access Journals (Sweden)

    Chiori Hori

    2003-02-01

    Full Text Available This paper proposes a statistical approach to automatic speech summarization. In our method, a set of words maximizing a summarization score indicating the appropriateness of summarization is extracted from automatically transcribed speech and then concatenated to create a summary. The extraction process is performed using a dynamic programming (DP technique based on a target compression ratio. In this paper, we demonstrate how an English news broadcast transcribed by a speech recognizer is automatically summarized. We adapted our method, which was originally proposed for Japanese, to English by modifying the model for estimating word concatenation probabilities based on a dependency structure in the original speech given by a stochastic dependency context free grammar (SDCFG. We also propose a method of summarizing multiple utterances using a two-level DP technique. The automatically summarized sentences are evaluated by summarization accuracy based on a comparison with a manual summary of speech that has been correctly transcribed by human subjects. Our experimental results indicate that the method we propose can effectively extract relatively important information and remove redundant and irrelevant information from English news broadcasts.

  2. Gesture facilitates the syntactic analysis of speech

    Directory of Open Access Journals (Sweden)

    Henning eHolle

    2012-03-01

    Full Text Available Recent research suggests that the brain routinely binds together information from gesture and speech. However, most of this research focused on the integration of representational gestures with the semantic content of speech. Much less is known about how other aspects of gesture, such as emphasis, influence the interpretation of the syntactic relations in a spoken message. Here, we investigated whether beat gestures alter which syntactic structure is assigned to ambiguous spoken German sentences. The P600 component of the Event Related Brain Potential indicated that the more complex syntactic structure is easier to process when the speaker emphasizes the subject of a sentence with a beat. Thus, a simple flick of the hand can change our interpretation of who has been doing what to whom in a spoken sentence. We conclude that gestures and speech are an integrated system. Unlike previous studies, which have shown that the brain effortlessly integrates semantic information from gesture and speech, our study is the first to demonstrate that this integration also occurs for syntactic information. Moreover, the effect appears to be gesture-specific and was not found for other stimuli that draw attention to certain parts of speech, including prosodic emphasis, or a moving visual stimulus with the same trajectory as the gesture. This suggests that only visual emphasis produced with a communicative intention in mind (that is, beat gestures influences language comprehension, but not a simple visual movement lacking such an intention.

  3. Practical speech user interface design

    CERN Document Server

    Lewis, James R

    2010-01-01

    Although speech is the most natural form of communication between humans, most people find using speech to communicate with machines anything but natural. Drawing from psychology, human-computer interaction, linguistics, and communication theory, Practical Speech User Interface Design provides a comprehensive yet concise survey of practical speech user interface (SUI) design. It offers practice-based and research-based guidance on how to design effective, efficient, and pleasant speech applications that people can really use. Focusing on the design of speech user interfaces for IVR application

  4. Post-editing through Speech Recognition

    DEFF Research Database (Denmark)

    Mesa-Lao, Bartolomé

    In the past couple of years automatic speech recognition (ASR) software has quietly created a niche for itself in many situations of our lives. Nowadays it can be found at the other end of customer-support hotlines, it is built into operating systems and it is offered as an alternative text......-input method for smartphones. On another front, given the significant improvements in Machine Translation (MT) quality and the increasing demand for translations, post-editing of MT is becoming a popular practice in the translation industry, since it has been shown to allow for larger volumes of translations...... to be produced saving time and costs. The translation industry is at a deeply transformative point in its evolution and the coming years herald an era of converge where speech technology could make a difference. As post-editing services are becoming a common practice among language service providers and speech...

  5. Fighting Hate Speech through EU Law

    Directory of Open Access Journals (Sweden)

    Uladzislau Belavusau

    2012-02-01

    Full Text Available

    This article explores the rise of the European ‘First Amendment’ beyond national and Strasbourg law, offering a fresh look into the previously under-theorised issue of hate speech in EU law. Building its argument on (1 the scrutiny of fundamental rights protection, (2 the distinction between commercial and non-commercial speech, and, finally, (3 the looking glass of critical race theory, the paper demonstrates how the judgment of the ECJ in the Feryn case implicitly consolidated legal narratives on hate speech in Europe. In this way, the paper reconstructs the dominant European theory of freedom of expression via rhetorical and victim-centered constitutional analysis, bearing important ethical implications for European integration.

     

  6. Automatic Speech Segmentation Based on HMM

    Directory of Open Access Journals (Sweden)

    M. Kroul

    2007-06-01

    Full Text Available This contribution deals with the problem of automatic phoneme segmentation using HMMs. Automatization of speech segmentation task is important for applications, where large amount of data is needed to process, so manual segmentation is out of the question. In this paper we focus on automatic segmentation of recordings, which will be used for triphone synthesis unit database creation. For speech synthesis, the speech unit quality is a crucial aspect, so the maximal accuracy in segmentation is needed here. In this work, different kinds of HMMs with various parameters have been trained and their usefulness for automatic segmentation is discussed. At the end of this work, some segmentation accuracy tests of all models are presented.

  7. Resourcing speech-language pathologists to work with multilingual children.

    Science.gov (United States)

    McLeod, Sharynne

    2014-06-01

    Speech-language pathologists play important roles in supporting people to be competent communicators in the languages of their communities. However, with over 7000 languages spoken throughout the world and the majority of the global population being multilingual, there is often a mismatch between the languages spoken by children and families and their speech-language pathologists. This paper provides insights into service provision for multilingual children within an English-dominant country by viewing Australia's multilingual population as a microcosm of ethnolinguistic minorities. Recent population studies of Australian pre-school children show that their most common languages other than English are: Arabic, Cantonese, Vietnamese, Italian, Mandarin, Spanish, and Greek. Although 20.2% of services by Speech Pathology Australia members are offered in languages other than English, there is a mismatch between the language of the services and the languages of children within similar geographical communities. Australian speech-language pathologists typically use informal or English-based assessments and intervention tools with multilingual children. Thus, there is a need for accessible culturally and linguistically appropriate resources for working with multilingual children. Recent international collaborations have resulted in practical strategies to support speech-language pathologists during assessment, intervention, and collaboration with families, communities, and other professionals. The International Expert Panel on Multilingual Children's Speech was assembled to prepare a position paper to address issues faced by speech-language pathologists when working with multilingual populations. The Multilingual Children's Speech website ( http://www.csu.edu.au/research/multilingual-speech ) addresses one of the aims of the position paper by providing free resources and information for speech-language pathologists about more than 45 languages. These international

  8. Speech-Language Therapy (For Parents)

    Science.gov (United States)

    ... Speech-language pathologists (SLPs), often informally known as speech therapists, are professionals educated in the study of human ... Palate Hearing Evaluation in Children Going to a Speech Therapist Stuttering Hearing Impairment Speech Problems Cleft Lip and ...

  9. Thinking soap But Speaking ‘oaps’. The Sound Preparation Period: Backward Calculation From Utterance to Muscle Innervation

    Directory of Open Access Journals (Sweden)

    Nora Wiedenmann

    2010-04-01

    Full Text Available

    In this article’s model—on speech and on speech errors, dyscoordinations, and disorders—, the time-course from the muscle innervation impetuses to the utterance of sounds as intended for canonical speech sound sequences is calculated backward. This time-course is shown as the sum of all the known physiological durations of speech sounds and speech gestures that are necessary to produce an utterance. The model introduces two internal clocks, based on positive or negative factors, representing certain physiologically-based time-courses during the sound preparation period (Lautvorspann. The use of these internal clocks show that speech gestures—like other motor activities—work according to a simple serialization principle: Under non-default conditions,
    alterations of the time-courses may cause speech errors of sound serialization, dyscoordinations of sounds as observed during first language acquisition, or speech disorders as pathological cases. These alterations of the time-course are modelled by varying the two internal-clock factors. The calculation of time-courses uses as default values the sound durations of the context-dependent Munich PHONDAT Database of Spoken German (see Appendix 4. As a new, human approach, this calculation agrees mathematically with the approach of Linear Programming / Operations Research. This work gives strong support to the fairly old suspicion (of 1908 of the famous Austrian speech error scientist Meringer [15], namely that one mostly thinks and articulates in a different serialization than is audible from one’s uttered sound sequences.

  10. Speech Compression of Thai Dialects with Low-Bit-Rate Speech Coders

    Directory of Open Access Journals (Sweden)

    Suphattharachai Chomphan

    2012-01-01

    Full Text Available Problem statement: In modern speech communication at low bit rate, speech coding deteriorates the characteristics of the coded speech significantly. Considering the dialects in Thai, the coding quality of four main dialects spoken by Thai people residing in four core region including central, north, northeast and south regions has not been studied. Approach: This study presents a comparative study of the coding quality of four main Thai dialects by using different low-bit-rate speech coders including the Conjugate Structure Algebraic Code Excited Linear Predictive (CS-ACELP coder and the Multi-Pulse based Code Excited Linear Predictive (MP-CELP coder. The objective and subjective tests have been conducted to evaluate the coding quality of four main dialects. Results: From the experimental results, both tests show that the coding quality of North dialect is highest, meanwhile the coding quality of Northeast dialect is lowest. Moreover, the coding quality of male speech is mostly higher than that of female speech. Conclusion: From the study, it can be obviously seen that the coding quality of all Thai dialects are different.

  11. Statistical speech segmentation and word learning in parallel: scaffolding from child-directed speech

    Directory of Open Access Journals (Sweden)

    Daniel eYurovsky

    2012-10-01

    Full Text Available In order to acquire their native languages, children must learn richly structured systems with regularities at multiple levels. While structure at different levels could be learned serially, e.g. speech segmentation coming before word-object mapping, redundancies across levels make parallel learning more efficient. For instance, a series of syllables is likely to be a word not only because of high transitional probabilities, but also because of a consistently co-occurring object. But additional statistics require additional processing, and thus might not be useful to cognitively constrained learners. We show that the structure of child-directed speech makes this problem solvable for human learners. First, a corpus of child-directed speech was recorded from parents and children engaged in a naturalistic free-play task. Analyses revealed two consistent regularities in the sentence structure of naming events. These regularities were subsequently encoded in an artificial language to which adult participants were exposed in the context of simultaneous statistical speech segmentation and word learning. Either regularity was sufficient to support successful learning, but no learning occurred in the absence of both regularities. Thus, the structure of child-directed speech plays an important role in scaffolding speech segmentation and word learning in parallel.

  12. Speech processing using maximum likelihood continuity mapping

    Energy Technology Data Exchange (ETDEWEB)

    Hogden, John E. (Santa Fe, NM)

    2000-01-01

    Speech processing is obtained that, given a probabilistic mapping between static speech sounds and pseudo-articulator positions, allows sequences of speech sounds to be mapped to smooth sequences of pseudo-articulator positions. In addition, a method for learning a probabilistic mapping between static speech sounds and pseudo-articulator position is described. The method for learning the mapping between static speech sounds and pseudo-articulator position uses a set of training data composed only of speech sounds. The said speech processing can be applied to various speech analysis tasks, including speech recognition, speaker recognition, speech coding, speech synthesis, and voice mimicry.

  13. Speech processing using maximum likelihood continuity mapping

    Energy Technology Data Exchange (ETDEWEB)

    Hogden, J.E.

    2000-04-18

    Speech processing is obtained that, given a probabilistic mapping between static speech sounds and pseudo-articulator positions, allows sequences of speech sounds to be mapped to smooth sequences of pseudo-articulator positions. In addition, a method for learning a probabilistic mapping between static speech sounds and pseudo-articulator position is described. The method for learning the mapping between static speech sounds and pseudo-articulator position uses a set of training data composed only of speech sounds. The said speech processing can be applied to various speech analysis tasks, including speech recognition, speaker recognition, speech coding, speech synthesis, and voice mimicry.

  14. Managing the reaction effects of speech disorders on speech ...

    African Journals Online (AJOL)

    Speech disorders is responsible for defective speaking. It is usually ... They occur as a result of persistent frustrations which speech defectives usually encounter for speaking defectively. This paper ... AJOL African Journals Online. HOW TO ...

  15. Song and speech: examining the link between singing talent and speech imitation ability

    Directory of Open Access Journals (Sweden)

    Markus eChristiner

    2013-11-01

    Full Text Available In previous research on speech imitation, musicality and an ability to sing were isolated as the strongest indicators of good pronunciation skills in foreign languages. We, therefore, wanted to take a closer look at the nature of the ability to sing, which shares a common ground with the ability to imitate speech. This study focuses on whether good singing performance predicts good speech imitation. Fourty-one singers of different levels of proficiency were selected for the study and their ability to sing, to imitate speech, their musical talent and working memory were tested. Results indicated that singing performance is a better indicator of the ability to imitate speech than the playing of a musical instrument. A multiple regression revealed that 64 % of the speech imitation score variance could be explained by working memory together with educational background and singing performance. A second multiple regression showed that 66 % of the speech imitation variance of completely unintelligible and unfamiliar language stimuli (Hindi could be explained by working memory together with a singer’s sense of rhythm and quality of voice. This supports the idea that both vocal behaviors have a common grounding in terms of vocal and motor flexibility, ontogenetic and phylogenetic development, neural orchestration and sound memory with singing fitting better into the category of "speech" on the productive level and "music" on the acoustic level. As a result, good singers benefit from vocal and motor flexibility, productively and cognitively, in three ways. 1. Motor flexibility and the ability to sing improve language and musical function. 2. Good singers retain a certain plasticity and are open to new and unusual sound combinations during adulthood both perceptually and productively. 3. The ability to sing improves the memory span of the auditory short term memory.

  16. Speech Acquisition and Automatic Speech Recognition for Integrated Spacesuit Audio Systems

    Science.gov (United States)

    Huang, Yiteng; Chen, Jingdong; Chen, Shaoyan

    2010-01-01

    A voice-command human-machine interface system has been developed for spacesuit extravehicular activity (EVA) missions. A multichannel acoustic signal processing method has been created for distant speech acquisition in noisy and reverberant environments. This technology reduces noise by exploiting differences in the statistical nature of signal (i.e., speech) and noise that exists in the spatial and temporal domains. As a result, the automatic speech recognition (ASR) accuracy can be improved to the level at which crewmembers would find the speech interface useful. The developed speech human/machine interface will enable both crewmember usability and operational efficiency. It can enjoy a fast rate of data/text entry, small overall size, and can be lightweight. In addition, this design will free the hands and eyes of a suited crewmember. The system components and steps include beam forming/multi-channel noise reduction, single-channel noise reduction, speech feature extraction, feature transformation and normalization, feature compression, model adaption, ASR HMM (Hidden Markov Model) training, and ASR decoding. A state-of-the-art phoneme recognizer can obtain an accuracy rate of 65 percent when the training and testing data are free of noise. When it is used in spacesuits, the rate drops to about 33 percent. With the developed microphone array speech-processing technologies, the performance is improved and the phoneme recognition accuracy rate rises to 44 percent. The recognizer can be further improved by combining the microphone array and HMM model adaptation techniques and using speech samples collected from inside spacesuits. In addition, arithmetic complexity models for the major HMMbased ASR components were developed. They can help real-time ASR system designers select proper tasks when in the face of constraints in computational resources.

  17. Implement of Speech Application Program Based on Speech SDK%基于Speech SDK的语音应用程序实现

    Institute of Scientific and Technical Information of China (English)

    高敬惠; 姜子敬; 胡金铭

    2005-01-01

    利用Microsoft Speech SDK的API for Text-to-Speech 和API for Speech Recognition ,采用Visual Basic 6.0语言,建立文本语音转换应用程序和实现语音识别程序,简单地实现了语音识别的功能,识别出来的内容即可保存为文件,也可作为命令使用,让计算机执行某项操作.

  18. Intelligibility of speech of children with speech and sound disorders

    OpenAIRE

    Ivetac, Tina

    2014-01-01

    The purpose of this study is to examine speech intelligibility of children with primary speech and sound disorders aged 3 to 6 years in everyday life. The research problem is based on the degree to which parents or guardians, immediate family members (sister, brother, grandparents), extended family members (aunt, uncle, cousin), child's friends, other acquaintances, child's teachers and strangers understand the speech of children with speech sound disorders. We examined whether the level ...

  19. Automatic speech recognition An evaluation of Google Speech

    OpenAIRE

    Stenman, Magnus

    2015-01-01

    The use of speech recognition is increasing rapidly and is now available in smart TVs, desktop computers, every new smart phone, etc. allowing us to talk to computers naturally. With the use in home appliances, education and even in surgical procedures accuracy and speed becomes very important. This thesis aims to give an introduction to speech recognition and discuss its use in robotics. An evaluation of Google Speech, using Google’s speech API, in regards to word error rate and translation ...

  20. The politeness prosody of the Javanese directive speech

    Directory of Open Access Journals (Sweden)

    F.X. Rahyono

    2009-10-01

    Full Text Available This experimental phonetic research deals with the prosodies of directive speech in Javanese. The research procedures were: (1 speech production, (2 acoustic analysis, and (3 perception test. The data investigated are three directive utterances, in the form of statements, commands, and questions. The data were obtained by recording dialogues that present polite as well as impolite speech. Three acoustic experiments were conducted for statements, commands, and questions in directive speech: (1 modifications of duration, (2 modifications of contour, and (3 modifications of fundamental frequency. The result of the subsequent perception tests to 90 stimuli with 24 subjects were analysed statistically with ANOVA (Analysis of Variant. Based on this statistic analysis, the prosodic characteristics of polite and impolite speech were identified.

  1. Bandwidth Extension of Speech Signals: A Comprehensive Review

    Directory of Open Access Journals (Sweden)

    N.Prasad

    2016-02-01

    Full Text Available Telephone systems commonly transmit narrowband (NB speech with an audio bandwidth limited to the traditional telephone band of 300-3400 Hz. To improve the quality and intelligibility of speech degraded by narrow bandwidth, researchers have tried to standardize the telephonic networks by introducing wideband (50-7000 Hz speech codecs. Wideband (WB speech transmission requires the transmission network and terminal devices at both ends to be upgraded to the wideband that turns out to be time-consuming. In this situation, novel Bandwidth extension (BWE techniques have been developed to overcome the limitations of NB speech. This paper discusses the basic principles, realization, and applications of BWE. Challenges and limitations of BWE are also addressed.

  2. Objects Control through Speech Recognition Using LabVIEW

    Directory of Open Access Journals (Sweden)

    Ankush Sharma

    2013-01-01

    Full Text Available Speech is the natural form of human communication and the speech processing is the one of the most stimulating area of the signal processing. Speech recognition technology has made it possible for computer to follow the human voice command and understand the human languages. The objects (LED, Toggle switch etc. control through human speech is designed in this paper. By combine the virtual instrumentation technology and speech recognition techniques. And also provided password authentication. This can be done with the help of LabVIEW programming concepts. The microphone is using to take voice commands from Human. This microphone signals interface with LabVIEW code. The LabVIEW code will generate appropriate control signal to control the objects. The entire work done on the LabVIEW platform.

  3. Dual Key Speech Encryption Algorithm Based Underdetermined BSS

    Directory of Open Access Journals (Sweden)

    Huan Zhao

    2014-01-01

    Full Text Available When the number of the mixed signals is less than that of the source signals, the underdetermined blind source separation (BSS is a significant difficult problem. Due to the fact that the great amount data of speech communications and real-time communication has been required, we utilize the intractability of the underdetermined BSS problem to present a dual key speech encryption method. The original speech is mixed with dual key signals which consist of random key signals (one-time pad generated by secret seed and chaotic signals generated from chaotic system. In the decryption process, approximate calculation is used to recover the original speech signals. The proposed algorithm for speech signals encryption can resist traditional attacks against the encryption system, and owing to approximate calculation, decryption becomes faster and more accurate. It is demonstrated that the proposed method has high level of security and can recover the original signals quickly and efficiently yet maintaining excellent audio quality.

  4. Pragmatic context of B.Obama’s public speeches

    Directory of Open Access Journals (Sweden)

    Yailokhanova Alexandra Sergeevna

    2016-03-01

    Full Text Available The article highlights the approaches to the definition of «discourse» given in the fundamental works of Western and Russian scientists-linguists. It also discusses the conditions of implementation of pragmatic attitudes in the speeches of public speeches of US President B. Obama. We present the definition of «pragmatic context factor». As the main factors of pragmatic context it is allocated space factor, the time factor, the factor of events, phenomena and factors, recipient factor. The interaction of these factors in specific contexts of speech are to achieve the intended objectives of a politician, an analysis of possible situations that affect the organization of discourse. Examples of linguistic means of expression of these factors in the context of pragmatic speech are given. The result of this study is the analysis of given public speeches on the subject of the content and methods of implementation of a pragmatic context factors.

  5. Infant rule learning: advantage language, or advantage speech?

    Directory of Open Access Journals (Sweden)

    Hugh Rabagliati

    Full Text Available Infants appear to learn abstract rule-like regularities (e.g., la la da follows an AAB pattern more easily from speech than from a variety of other auditory and visual stimuli (Marcus et al., 2007. We test if that facilitation reflects a specialization to learn from speech alone, or from modality-independent communicative stimuli more generally, by measuring 7.5-month-old infants' ability to learn abstract rules from sign language-like gestures. Whereas infants appear to easily learn many different rules from speech, we found that with sign-like stimuli, and under circumstances comparable to those of Marcus et al. (1999, hearing infants were able to learn an ABB rule, but not an AAB rule. This is consistent with results of studies that demonstrate lower levels of infant rule learning from a variety of other non-speech stimuli, and we discuss implications for accounts of speech-facilitation.

  6. Histogram Equalization to Model Adaptation for Robust Speech Recognition

    Directory of Open Access Journals (Sweden)

    Hoirin Kim

    2010-01-01

    Full Text Available We propose a new model adaptation method based on the histogram equalization technique for providing robustness in noisy environments. The trained acoustic mean models of a speech recognizer are adapted into environmentally matched conditions by using the histogram equalization algorithm on a single utterance basis. For more robust speech recognition in the heavily noisy conditions, trained acoustic covariance models are efficiently adapted by the signal-to-noise ratio-dependent linear interpolation between trained covariance models and utterance-level sample covariance models. Speech recognition experiments on both the digit-based Aurora2 task and the large vocabulary-based task showed that the proposed model adaptation approach provides significant performance improvements compared to the baseline speech recognizer trained on the clean speech data.

  7. Speech Outcomes after Tonsillectomy in Patients with Known Velopharyngeal Insufficiency

    Directory of Open Access Journals (Sweden)

    L. M. Paulson

    2012-01-01

    Full Text Available Introduction. Controversy exists over whether tonsillectomy will affect speech in patients with known velopharyngeal insufficiency (VPI, particularly in those with cleft palate. Methods. All patients seen at the OHSU Doernbecher Children's Hospital VPI clinic between 1997 and 2010 with VPI who underwent tonsillectomy were reviewed. Speech parameters were assessed before and after tonsillectomy. Wilcoxon rank-sum testing was used to evaluate for significance. Results. A total of 46 patients with VPI underwent tonsillectomy during this period. Twenty-three had pre- and postoperative speech evaluation sufficient for analysis. The majority (87% had a history of cleft palate. Indications for tonsillectomy included obstructive sleep apnea in 11 (48% and staged tonsillectomy prior to pharyngoplasty in 10 (43%. There was no significant difference between pre- and postoperative speech intelligibility or velopharyngeal competency in this population. Conclusion. In this study, tonsillectomy in patients with VPI did not significantly alter speech intelligibility or velopharyngeal competence.

  8. Aspects of Connected Speech Processes in Nigerian English

    Directory of Open Access Journals (Sweden)

    Rotimi Olanrele Oladipupo

    2014-12-01

    Full Text Available Nigerian English (NigE, like other new Englishes, possesses its unique features at various domains of phonology. This article examined aspects of connected speech processes (CSPs, the phenomena that account for sound modifications and simplifications in speech, with a view to establishing features that characterize Standard NigE connected speech. Natural phonology (NP, which provides explanations for substitutions, alternations, and variations in the speech of second language speakers, was adopted as theoretical framework. The subjects of the study were 360 educated NigE speakers, accidentally sampled from different language groups in Nigeria. The CSPs found in their semi-spontaneous speeches were transcribed perceptually and analyzed statistically, by allotting marks to instances of occurrence and converting such to percentages. Three categories of CSPs were identified in the data: dominant, minor, and idiosyncratic processes. The study affirms that only the dominant CSPs, typical of NigE speakers, are acceptable as Standard Nigerian spoken English.

  9. Differential Diagnosis of Severe Speech Disorders Using Speech Gestures

    Science.gov (United States)

    Bahr, Ruth Huntley

    2005-01-01

    The differentiation of childhood apraxia of speech from severe phonological disorder is a common clinical problem. This article reports on an attempt to describe speech errors in children with childhood apraxia of speech on the basis of gesture use and acoustic analyses of articulatory gestures. The focus was on the movement of articulators and…

  10. Type of speech material affects Acceptable Noise Level test outcome

    Directory of Open Access Journals (Sweden)

    Xaver eKoch

    2016-02-01

    Full Text Available The Acceptable Noise Level (ANL test, in which individuals indicate what level of noise they are willing to put up with while following speech, has been used to guide hearing aid fitting decisions and has been found to relate to prospective hearing aid use. Unlike objective measures of speech perception ability, ANL outcome is not related to individual hearing loss or age, but rather reflects an individual's inherent acceptance of competing noise while listening to speech. As such, the measure may predict aspects of hearing aid success. Crucially, however, recent studies have questioned its repeatability (test-retest reliability. The first question for this study was whether the inconsistent results regarding the repeatability of the ANL test may be due to differences in speech material types used in previous studies. Second, it is unclear whether meaningfulness and semantic coherence of the speech modify ANL outcome. To investigate these questions, we compared ANLs obtained with three types of materials: the International Speech Test Signal (ISTS, which is non-meaningful and semantically non-coherent by definition, passages consisting of concatenated meaningful standard audiology sentences, and longer fragments taken from conversational speech. We included conversational speech as this type of speech material is most representative of everyday listening. Additionally, we investigated whether ANL outcomes, obtained with these three different speech materials, were associated with self-reported limitations due to hearing problems and listening effort in everyday life, as assessed by a questionnaire. ANL data were collected for 57 relatively good-hearing adult participants with an age range representative for hearing aid users. Results showed that meaningfulness, but not semantic coherence of the speech material affected ANL. Less noise was accepted for the non-meaningful ISTS signal than for the meaningful speech materials. ANL repeatability was

  11. Tackling the complexity in speech

    DEFF Research Database (Denmark)

    section includes four carefully selected chapters. They deal with facets of speech production, speech acoustics, and/or speech perception or recognition, place them in an integrated phonetic-phonological perspective, and relate them in more or less explicit ways to aspects of speech technology. Therefore......, we hope that this volume can help speech scientists with traditional training in phonetics and phonology to keep up with the latest developments in speech technology. In the opposite direction, speech researchers starting from a technological perspective will hopefully get inspired by reading about...... the questions, phenomena, and communicative functions that are currently addressed in phonetics and phonology. Either way, the future of speech research lies in international, interdisciplinary collaborations, and our volume is meant to reflect and facilitate such collaborations...

  12. Tackling the complexity in speech

    DEFF Research Database (Denmark)

    section includes four carefully selected chapters. They deal with facets of speech production, speech acoustics, and/or speech perception or recognition, place them in an integrated phonetic-phonological perspective, and relate them in more or less explicit ways to aspects of speech technology. Therefore......, we hope that this volume can help speech scientists with traditional training in phonetics and phonology to keep up with the latest developments in speech technology. In the opposite direction, speech researchers starting from a technological perspective will hopefully get inspired by reading about...... the questions, phenomena, and communicative functions that are currently addressed in phonetics and phonology. Either way, the future of speech research lies in international, interdisciplinary collaborations, and our volume is meant to reflect and facilitate such collaborations...

  13. Pathomechanisms and compensatory efforts related to Parkinsonian speech.

    Science.gov (United States)

    Arnold, Christiane; Gehrig, Johannes; Gispert, Suzana; Seifried, Carola; Kell, Christian A

    2014-01-01

    Voice and speech in Parkinson's disease (PD) patients are classically affected by a hypophonia, dysprosody, and dysarthria. The underlying pathomechanisms of these disabling symptoms are not well understood. To identify functional anomalies related to pathophysiology and compensation we compared speech-related brain activity and effective connectivity in early PD patients who did not yet develop voice or speech symptoms and matched controls. During fMRI 20 PD patients ON and OFF levodopa and 20 control participants read 75 sentences covertly, overtly with neutral, or with happy intonation. A cue-target reading paradigm allowed for dissociating task preparation from execution. We found pathologically reduced striato-prefrontal preparatory effective connectivity in early PD patients associated with subcortical (OFF state) or cortical (ON state) compensatory networks. While speaking, PD patients showed signs of diminished monitoring of external auditory feedback. During generation of affective prosody, a reduced functional coupling between the ventral and dorsal striatum was observed. Our results suggest three pathomechanisms affecting speech in PD: While diminished energization on the basis of striato-prefrontal hypo-connectivity together with dysfunctional self-monitoring mechanisms could underlie hypophonia, dysarthria may result from fading speech motor representations given that they are not sufficiently well updated by external auditory feedback. A pathological interplay between the limbic and sensorimotor striatum could interfere with affective modulation of speech routines, which affects emotional prosody generation. However, early PD patients show compensatory mechanisms that could help improve future speech therapies.

  14. Presentation video retrieval using automatically recovered slide and spoken text

    Science.gov (United States)

    Cooper, Matthew

    2013-03-01

    Video is becoming a prevalent medium for e-learning. Lecture videos contain text information in both the presentation slides and lecturer's speech. This paper examines the relative utility of automatically recovered text from these sources for lecture video retrieval. To extract the visual information, we automatically detect slides within the videos and apply optical character recognition to obtain their text. Automatic speech recognition is used similarly to extract spoken text from the recorded audio. We perform controlled experiments with manually created ground truth for both the slide and spoken text from more than 60 hours of lecture video. We compare the automatically extracted slide and spoken text in terms of accuracy relative to ground truth, overlap with one another, and utility for video retrieval. Results reveal that automatically recovered slide text and spoken text contain different content with varying error profiles. Experiments demonstrate that automatically extracted slide text enables higher precision video retrieval than automatically recovered spoken text.

  15. Perception of words and pitch patterns in song and speech

    Directory of Open Access Journals (Sweden)

    Julia eMerrill

    2012-03-01

    Full Text Available This fMRI study examines shared and distinct cortical areas involved in the auditory perception of song and speech at the level of their underlying constituents: words, pitch and rhythm. Univariate and multivariate analyses were performed on the brain activity patterns of six conditions, arranged in a subtractive hierarchy: sung sentences including words, pitch and rhythm; hummed speech prosody and song melody containing only pitch patterns and rhythm; as well as the pure musical or speech rhythm.Systematic contrasts between these balanced conditions following their hierarchical organization showed a great overlap between song and speech at all levels in the bilateral temporal lobe, but suggested a differential role of the inferior frontal gyrus (IFG and intraparietal sulcus (IPS in processing song and speech. The left IFG was involved in word- and pitch-related processing in speech, the right IFG in processing pitch in song.Furthermore, the IPS showed sensitivity to discrete pitch relations in song as opposed to the gliding pitch in speech. Finally, the superior temporal gyrus and premotor cortex coded for general differences between words and pitch patterns, irrespective of whether they were sung or spoken. Thus, song and speech share many features which are reflected in a fundamental similarity of brain areas involved in their perception. However, fine-grained acoustic differences on word and pitch level are reflected in the activity of IFG and IPS.

  16. Speech Enhancement Using Kernel and Normalized Kernel Affine Projection Algorithm

    Directory of Open Access Journals (Sweden)

    Bolimera Ravi

    2013-08-01

    Full Text Available The goal of this paper is to investigate the speech signal enhancement using Kernel Affine ProjectionAlgorithm (KAPA and Normalized KAPA. The removal of background noise is very important in manyapplications like speech recognition, telephone conversations, hearing aids, forensic, etc. Kernel adaptivefilters shown good performance for removal of noise. If the evaluation of background noise is more slowlythan the speech, i.e., noise signal is more stationary than the speech, we can easily estimate the noiseduring the pauses in speech. Otherwise it is more difficult to estimate the noise which results indegradation of speech. In order to improve the quality and intelligibility of speech, unlike time andfrequency domains, we can process the signal in new domain like Reproducing Kernel Hilbert Space(RKHS for high dimensional to yield more powerful nonlinear extensions. For experiments, we have usedthe database of noisy speech corpus (NOIZEUS. From the results, we observed the removal noise in RKHShas great performance in signal to noise ratio values in comparison with conventional adaptive filters.

  17. Infants' preference for native audiovisual speech dissociated from congruency preference.

    Directory of Open Access Journals (Sweden)

    Kathleen Shaw

    Full Text Available Although infant speech perception in often studied in isolated modalities, infants' experience with speech is largely multimodal (i.e., speech sounds they hear are accompanied by articulating faces. Across two experiments, we tested infants' sensitivity to the relationship between the auditory and visual components of audiovisual speech in their native (English and non-native (Spanish language. In Experiment 1, infants' looking times were measured during a preferential looking task in which they saw two simultaneous visual speech streams articulating a story, one in English and the other in Spanish, while they heard either the English or the Spanish version of the story. In Experiment 2, looking times from another group of infants were measured as they watched single displays of congruent and incongruent combinations of English and Spanish audio and visual speech streams. Findings demonstrated an age-related increase in looking towards the native relative to non-native visual speech stream when accompanied by the corresponding (native auditory speech. This increase in native language preference did not appear to be driven by a difference in preference for native vs. non-native audiovisual congruence as we observed no difference in looking times at the audiovisual streams in Experiment 2.

  18. Processing changes when listening to foreign-accented speech

    Directory of Open Access Journals (Sweden)

    Carlos eRomero-Rivas

    2015-03-01

    Full Text Available This study investigates the mechanisms responsible for fast changes in processing foreign-accented speech. Event Related brain Potentials (ERPs were obtained while native speakers of Spanish listened to native and foreign-accented speakers of Spanish. We observed a less positive P200 component for foreign-accented speech relative to native speech comprehension. This suggests that the extraction of spectral information and other important acoustic features was hampered during foreign-accented speech comprehension. However, the amplitude of the N400 component for foreign-accented speech comprehension decreased across the experiment, suggesting the use of a higher level, lexical mechanism. Furthermore, during native speech comprehension, semantic violations in the critical words elicited an N400 effect followed by a late positivity. During foreign-accented speech comprehension, semantic violations only elicited an N400 effect. Overall, our results suggest that, despite a lack of improvement in phonetic discrimination, native listeners experience changes at lexical-semantic levels of processing after brief exposure to foreign-accented speech. Moreover, these results suggest that lexical access, semantic integration and linguistic re-analysis processes are permeable to external factors, such as the accent of the speaker.

  19. Digitized Ethnic Hate Speech: Understanding Effects of Digital Media Hate Speech on Citizen Journalism in Kenya

    Directory of Open Access Journals (Sweden)

    Stephen Gichuhi Kimotho

    2016-06-01

    Full Text Available Ethnicity in Kenya permeates all spheres of life. However, it is in politics that ethnicity is most visible. Election time in Kenya often leads to ethnic competition and hatred, often expressed through various media. Ethnic hate speech characterized the 2007 general elections in party rallies and through text messages, emails, posters and leaflets. This resulted in widespread skirmishes that left over 1200 people dead, and many displaced (KNHRC, 2008. In 2013, however, the new battle zone was the war of words on social media platform. More than any other time in Kenyan history, Kenyans poured vitriolic ethnic hate speech through digital media like Facebook, tweeter and blogs. Although scholars have studied the role and effects of the mainstream media like television and radio in proliferating the ethnic hate speech in Kenya (Michael Chege, 2008; Goldstein & Rotich, 2008a; Ismail & Deane, 2008; Jacqueline Klopp & Prisca Kamungi, 2007, little has been done in regard to social media.  This paper investigated the nature of digitized hate speech by: describing the forms of ethnic hate speech on social media in Kenya; the effects of ethnic hate speech on Kenyan’s perception of ethnic entities; ethnic conflict and ethics of citizen journalism. This study adopted a descriptive interpretive design, and utilized Austin’s Speech Act Theory, which explains use of language to achieve desired purposes and direct behaviour (Tarhom & Miracle, 2013. Content published between January and April 2013 from six purposefully identified blogs was analysed. Questionnaires were used to collect data from university students as they form a good sample of Kenyan population, are most active on social media and are drawn from all parts of the country. Qualitative data were analysed using NVIVO 10 software, while responses from the questionnaire were analysed using IBM SPSS version 21. The findings indicated that Facebook and Twitter were the main platforms used to

  20. Detection and Separation of Speech Events in Meeting Recordings Using a Microphone Array

    Directory of Open Access Journals (Sweden)

    Yamada Miichi

    2007-01-01

    Full Text Available When applying automatic speech recognition (ASR to meeting recordings including spontaneous speech, the performance of ASR is greatly reduced by the overlap of speech events. In this paper, a method of separating the overlapping speech events by using an adaptive beamforming (ABF framework is proposed. The main feature of this method is that all the information necessary for the adaptation of ABF, including microphone calibration, is obtained from meeting recordings based on the results of speech-event detection. The performance of the separation is evaluated via ASR using real meeting recordings.

  1. Detection and Separation of Speech Events in Meeting Recordings Using a Microphone Array

    Directory of Open Access Journals (Sweden)

    Futoshi Asano

    2007-07-01

    Full Text Available When applying automatic speech recognition (ASR to meeting recordings including spontaneous speech, the performance of ASR is greatly reduced by the overlap of speech events. In this paper, a method of separating the overlapping speech events by using an adaptive beamforming (ABF framework is proposed. The main feature of this method is that all the information necessary for the adaptation of ABF, including microphone calibration, is obtained from meeting recordings based on the results of speech-event detection. The performance of the separation is evaluated via ASR using real meeting recordings.

  2. Comparing Speech Recognition Systems (Microsoft API, Google API And CMU Sphinx

    Directory of Open Access Journals (Sweden)

    Veton Këpuska

    2017-03-01

    Full Text Available The idea of this paper is to design a tool that will be used to test and compare commercial speech recognition systems, such as Microsoft Speech API and Google Speech API, with open-source speech recognition systems such as Sphinx-4. The best way to compare automatic speech recognition systems in different environments is by using some audio recordings that were selected from different sources and calculating the word error rate (WER. Although the WER of the three aforementioned systems were acceptable, it was observed that the Google API is superior.

  3. Speech spectrogram expert

    Energy Technology Data Exchange (ETDEWEB)

    Johannsen, J.; Macallister, J.; Michalek, T.; Ross, S.

    1983-01-01

    Various authors have pointed out that humans can become quite adept at deriving phonetic transcriptions from speech spectrograms (as good as 90percent accuracy at the phoneme level). The authors describe an expert system which attempts to simulate this performance. The speech spectrogram expert (spex) is actually a society made up of three experts: a 2-dimensional vision expert, an acoustic-phonetic expert, and a phonetics expert. The visual reasoning expert finds important visual features of the spectrogram. The acoustic-phonetic expert reasons about how visual features relates to phonemes, and about how phonemes change visually in different contexts. The phonetics expert reasons about allowable phoneme sequences and transformations, and deduces an english spelling for phoneme strings. The speech spectrogram expert is highly interactive, allowing users to investigate hypotheses and edit rules. 10 references.

  4. Indonesian Automatic Speech Recognition For Command Speech Controller Multimedia Player

    Directory of Open Access Journals (Sweden)

    Vivien Arief Wardhany

    2014-12-01

    Full Text Available The purpose of multimedia devices development is controlling through voice. Nowdays voice that can be recognized only in English. To overcome the issue, then recognition using Indonesian language model and accousticc model and dictionary. Automatic Speech Recognizier is build using engine CMU Sphinx with modified english language to Indonesian Language database and XBMC used as the multimedia player. The experiment is using 10 volunteers testing items based on 7 commands. The volunteers is classifiedd by the genders, 5 Male & 5 female. 10 samples is taken in each command, continue with each volunteer perform 10 testing command. Each volunteer also have to try all 7 command that already provided. Based on percentage clarification table, the word “Kanan” had the most recognize with percentage 83% while “pilih” is the lowest one. The word which had the most wrong clarification is “kembali” with percentagee 67%, while the word “kanan” is the lowest one. From the result of Recognition Rate by male there are several command such as “Kembali”, “Utama”, “Atas “ and “Bawah” has the low Recognition Rate. Especially for “kembali” cannot be recognized as the command in the female voices but in male voice that command has 4% of RR this is because the command doesn’t have similar word in english near to “kembali” so the system unrecognize the command. Also for the command “Pilih” using the female voice has 80% of RR but for the male voice has only 4% of RR. This problem is mostly because of the different voice characteristic between adult male and female which male has lower voice frequencies (from 85 to 180 Hz than woman (165 to 255 Hz.The result of the experiment showed that each man had different number of recognition rate caused by the difference tone, pronunciation, and speed of speech. For further work needs to be done in order to improving the accouracy of the Indonesian Automatic Speech Recognition system

  5. Protection limits on free speech

    Institute of Scientific and Technical Information of China (English)

    李敏

    2014-01-01

    Freedom of speech is one of the basic rights of citizens should receive broad protection, but in the real context of China under what kind of speech can be protected and be restricted, how to grasp between state power and free speech limit is a question worth considering. People tend to ignore the freedom of speech and its function, so that some of the rhetoric cannot be demonstrated in the open debates.

  6. The University and Free Speech

    OpenAIRE

    Grcic, Joseph

    2014-01-01

    Free speech is a necessary condition for the growth of knowledge and the implementation of real and rational democracy. Educational institutions play a central role in socializing individuals to function within their society. Academic freedom is the right to free speech in the context of the university and tenure, properly interpreted, is a necessary component of protecting academic freedom and free speech.

  7. Designing speech for a recipient

    DEFF Research Database (Denmark)

    Fischer, Kerstin

    is investigated on three candidates for so-called ‘simplified registers’: speech to children (also called motherese or baby talk), speech to foreigners (also called foreigner talk) and speech to robots. The volume integrates research from various disciplines, such as psychology, sociolinguistics...

  8. ADMINISTRATIVE GUIDE IN SPEECH CORRECTION.

    Science.gov (United States)

    HEALEY, WILLIAM C.

    WRITTEN PRIMARILY FOR SCHOOL SUPERINTENDENTS, PRINCIPALS, SPEECH CLINICIANS, AND SUPERVISORS, THIS GUIDE OUTLINES THE MECHANICS OF ORGANIZING AND CONDUCTING SPEECH CORRECTION ACTIVITIES IN THE PUBLIC SCHOOLS. IT INCLUDES THE REQUIREMENTS FOR CERTIFICATION OF A SPEECH CLINICIAN IN MISSOURI AND DESCRIBES ESSENTIAL STEPS FOR THE DEVELOPMENT OF A…

  9. A generative model of speech production in Broca’s and Wernicke’s areas

    Directory of Open Access Journals (Sweden)

    Cathy J Price

    2011-09-01

    Full Text Available Speech production involves the generation of an auditory signal from the articulators and vocal tract. When the intended auditory signal does not match the produced sounds, subsequent articulatory commands can be adjusted to reduce the difference between the intended and produced sounds. This requires an internal model of the intended speech output that can be compared to the produced speech. The aim of this functional imaging study was to identify brain activation related to the internal model of speech production after activation related to vocalisation, auditory feedback and movement in the articulators had been controlled. There were four conditions: silent articulation of speech, non-speech mouth movements, finger tapping and visual fixation. In the speech conditions, participants produced the mouth movements associated with the words one and three. We eliminated auditory feedback from the spoken output by instructing participants to articulate these words without producing any sound. The non-speech mouth movement conditions involved lip pursing and tongue protrusions to control for movement in the articulators. The main difference between our speech and non-speech mouth movement conditions is that prior experience producing speech sounds leads to the automatic and covert generation of auditory and phonological associations that may play a role in predicting auditory feedback. We found that, relative to non-speech mouth movements, silent speech activated Broca's area in the left dorsal pars opercularis and Wernicke's area in the left posterior superior temporal sulcus. We discuss these results in the context of a generative model of speech production and propose that Broca’s and Wernicke’s areas may be involved in predicting the speech output that follows articulation. These predictions could provide a mechanism by which rapid movement of the articulators is precisely matched to the intended speech outputs during future articulations.

  10. Evaluation of the speech perception in the noise in different positions in adults with cochlear implants

    Directory of Open Access Journals (Sweden)

    Santos, Karlos Thiago Pinheiro dos

    2009-03-01

    Full Text Available Introduction: The most frequent complaint of the cochlear implant users has been to recognize and understand the speech signal in the presence of noise. Researches have been developed on the speech perception of users of cochlear implant with focus on aspects such as the effect of the reduction to the signal/noise ratio in the speech perception, the speech recognition in the noise, with different types of cochlear implant and strategies of speech codification and the effects of the binaural stimulation in the speech perception in noise. Objective: 1-To assess the speech perception in cochlear implant adult users in different positions regarding the presentation of the stimulus, 2-to compare the index of speech recognition in the frontal, ipsilateral and contralateral positions and 3-to analyze the effect of monoaural adaptation in the speech perception with noise. Method: 22 cochlear implant adult users were evaluated regarding the speech perception. The individuals were submitted to sentences recognition evaluation, with competitive noise in the signal/noise ratio +10 decibels in three positions: frontal, ipsilateral and contralateral to the cochlear implant side. Results: The results demonstrated the largest index of speech recognition in the ipsilateral position (100% and the lowest index of speech recognition with sentences in the contralateral position (5%. Conclusion: The performance of speech perception in cochlear implant users is damaged when the competitive noise is introduced, the index of speech recognition is better when the speech is presented ipsilaterally, and it's consequently worse when presented contralaterally to the cochlear implant, and there are more damages in the speech intelligibility when there is only monoaural input.

  11. Speech rhythms and multiplexed oscillatory sensory coding in the human brain.

    Directory of Open Access Journals (Sweden)

    Joachim Gross

    2013-12-01

    Full Text Available Cortical oscillations are likely candidates for segmentation and coding of continuous speech. Here, we monitored continuous speech processing with magnetoencephalography (MEG to unravel the principles of speech segmentation and coding. We demonstrate that speech entrains the phase of low-frequency (delta, theta and the amplitude of high-frequency (gamma oscillations in the auditory cortex. Phase entrainment is stronger in the right and amplitude entrainment is stronger in the left auditory cortex. Furthermore, edges in the speech envelope phase reset auditory cortex oscillations thereby enhancing their entrainment to speech. This mechanism adapts to the changing physical features of the speech envelope and enables efficient, stimulus-specific speech sampling. Finally, we show that within the auditory cortex, coupling between delta, theta, and gamma oscillations increases following speech edges. Importantly, all couplings (i.e., brain-speech and also within the cortex attenuate for backward-presented speech, suggesting top-down control. We conclude that segmentation and coding of speech relies on a nested hierarchy of entrained cortical oscillations.

  12. Frequent word section extraction in a presentation speech by an effective dynamic programming algorithm.

    Science.gov (United States)

    Itoh, Yoshiaki; Tanaka, Kazuyo

    2004-08-01

    Word frequency in a document has often been utilized in text searching and summarization. Similarly, identifying frequent words or phrases in a speech data set for searching and summarization would also be meaningful. However, obtaining word frequency in a speech data set is difficult, because frequent words are often special terms in the speech and cannot be recognized by a general speech recognizer. This paper proposes another approach that is effective for automatic extraction of such frequent word sections in a speech data set. The proposed method is applicable to any domain of monologue speech, because no language models or specific terms are required in advance. The extracted sections can be regarded as speech labels of some kind or a digest of the speech presentation. The frequent word sections are determined by detecting similar sections, which are sections of audio data that represent the same word or phrase. The similar sections are detected by an efficient algorithm, called Shift Continuous Dynamic Programming (Shift CDP), which realizes fast matching between arbitrary sections in the reference speech pattern and those in the input speech, and enables frame-synchronous extraction of similar sections. In experiments, the algorithm is applied to extract the repeated sections in oral presentation speeches recorded in academic conferences in Japan. The results show that Shift CDP successfully detects similar sections and identifies the frequent word sections in individual presentation speeches, without prior domain knowledge, such as language models and terms.

  13. SPEECH DISORDERS ENCOUNTERED DURING SPEECH THERAPY AND THERAPY TECHNIQUES

    OpenAIRE

    2013-01-01

    Speech which is a physical and mental process, agreed signs and sounds to create a sense of mind to the message that change . Process to identify the sounds of speech it is essential to know the structure and function of various organs which allows to happen the conversation. Speech is a physical and mental process so many factors can lead to speech disorders. Speech disorder can be about language acquisitions as well as it can be caused medical and psychological many factors. Disordered sp...

  14. Freedom of Speech and Hate Speech: an analysis of possible limits for freedom of speech

    National Research Council Canada - National Science Library

    Riva Sobrado de Freitas; Matheus Felipe de Castro

    2013-01-01

      In a view to determining the outlines of the Freedom of Speech and to specify its contents, we face hate speech as an offensive and repulsive manifestation, particularly directed to minority groups...

  15. Speech transmission index from running speech: A neural network approach

    Science.gov (United States)

    Li, F. F.; Cox, T. J.

    2003-04-01

    Speech transmission index (STI) is an important objective parameter concerning speech intelligibility for sound transmission channels. It is normally measured with specific test signals to ensure high accuracy and good repeatability. Measurement with running speech was previously proposed, but accuracy is compromised and hence applications limited. A new approach that uses artificial neural networks to accurately extract the STI from received running speech is developed in this paper. Neural networks are trained on a large set of transmitted speech examples with prior knowledge of the transmission channels' STIs. The networks perform complicated nonlinear function mappings and spectral feature memorization to enable accurate objective parameter extraction from transmitted speech. Validations via simulations demonstrate the feasibility of this new method on a one-net-one-speech extract basis. In this case, accuracy is comparable with normal measurement methods. This provides an alternative to standard measurement techniques, and it is intended that the neural network method can facilitate occupied room acoustic measurements.

  16. Optical Character Recognition Based Speech Synthesis System Using LabVIEW

    Directory of Open Access Journals (Sweden)

    S.K. Singla

    2014-10-01

    Full Text Available Knowledge extraction by just listening to sounds is a distinctive property. Speech signal is more effective means of communication than text because blind and visually impaired persons can also respond to sounds. This paper aims to develop a cost effective, and user friendly optical character recognition (OCR based speech synthesis system. The OCR based speech synthesis system has been developed using Laboratory virtual instruments engineering workbench (LabVIEW 7.1.

  17. Brain-inspired speech segmentation for automatic speech recognition using the speech envelope as a temporal reference

    OpenAIRE

    Byeongwook Lee; Kwang-Hyun Cho

    2016-01-01

    Speech segmentation is a crucial step in automatic speech recognition because additional speech analyses are performed for each framed speech segment. Conventional segmentation techniques primarily segment speech using a fixed frame size for computational simplicity. However, this approach is insufficient for capturing the quasi-regular structure of speech, which causes substantial recognition failure in noisy environments. How does the brain handle quasi-regular structured speech and maintai...

  18. Human Emotion Recognition From Speech

    Directory of Open Access Journals (Sweden)

    Miss. Aparna P. Wanare

    2014-07-01

    Full Text Available Speech Emotion Recognition is a recent research topic in the Human Computer Interaction (HCI field. The need has risen for a more natural communication interface between humans and computer, as computers have become an integral part of our lives. A lot of work currently going on to improve the interaction between humans and computers. To achieve this goal, a computer would have to be able to distinguish its present situation and respond differently depending on that observation. Part of this process involves understanding a user‟s emotional state. To make the human computer interaction more natural, the objective is that computer should be able to recognize emotional states in the same as human does. The efficiency of emotion recognition system depends on type of features extracted and classifier used for detection of emotions. The proposed system aims at identification of basic emotional states such as anger, joy, neutral and sadness from human speech. While classifying different emotions, features like MFCC (Mel Frequency Cepstral Coefficient and Energy is used. In this paper, Standard Emotional Database i.e. English Database is used which gives the satisfactory detection of emotions than recorded samples of emotions. This methodology describes and compares the performances of Learning Vector Quantization Neural Network (LVQ NN, Multiclass Support Vector Machine (SVM and their combination for emotion recognition.

  19. Global Freedom of Speech

    DEFF Research Database (Denmark)

    Binderup, Lars Grassme

    2007-01-01

    , as opposed to a legal norm, that curbs exercises of the right to free speech that offend the feelings or beliefs of members from other cultural groups. The paper rejects the suggestion that acceptance of such a norm is in line with liberal egalitarian thinking. Following a review of the classical liberal...

  20. Speech and Hearing Therapy.

    Science.gov (United States)

    Sakata, Reiko; Sakata, Robert

    1978-01-01

    In the public school, the speech and hearing therapist attempts to foster child growth and development through the provision of services basic to awareness of self and others, management of personal and social interactions, and development of strategies for coping with the handicap. (MM)

  1. Perceptual learning in speech

    NARCIS (Netherlands)

    Norris, D.; McQueen, J.M.; Cutler, A.

    2003-01-01

    This study demonstrates that listeners use lexical knowledge in perceptual learning of speech sounds. Dutch listeners first made lexical decisions on Dutch words and nonwords. The final fricative of 20 critical words had been replaced by an ambiguous sound, between [f] and [s]. One group of listener

  2. Speech and Language Delay

    Science.gov (United States)

    ... home affect my child’s language and speech?The brain has to work harder to interpret and use 2 languages, so it may take longer for children to start using either one or both of the languages they’re learning. It’s not unusual for a bilingual child to ...

  3. Mandarin Visual Speech Information

    Science.gov (United States)

    Chen, Trevor H.

    2010-01-01

    While the auditory-only aspects of Mandarin speech are heavily-researched and well-known in the field, this dissertation addresses its lesser-known aspects: The visual and audio-visual perception of Mandarin segmental information and lexical-tone information. Chapter II of this dissertation focuses on the audiovisual perception of Mandarin…

  4. Speech After Banquet

    Science.gov (United States)

    Yang, Chen Ning

    2013-05-01

    I am usually not so short of words, but the previous speeches have rendered me really speechless. I have known and admired the eloquence of Freeman Dyson, but I did not know that there is a hidden eloquence in my colleague George Sterman...

  5. Speech disfluency in centenarians.

    Science.gov (United States)

    Searl, Jeffrey P; Gabel, Rodney M; Fulks, J Steven

    2002-01-01

    Other than a single case presentation of a 105-year-old female, no other studies have addressed the speech fluency characteristics of centenarians. The purpose of this study was to provide descriptive information on the fluency characteristics of speakers between the ages of 100-103 years. Conversational speech samples from seven speakers were evaluated for the frequency and types of disfluencies and speech rate. The centenarian speakers had a disfluency rate similar to that reported for 70-, 80-, and early 90-year-olds. The types of disfluencies observed also were similar to those reported for younger elderly speakers (primarily whole word/phrase, or formulative fluency breaks). Finally, the speech rate data for the current group of speakers supports prior literature reports of a slower rate with advancing age, but extends the finding to centenarians. As a result of this activity, participants will be able to: (1) describe the frequency of disfluency breaks and the types of disfluencies exhibited by centenarian speakers, (2) describe the mean and range of speaking rates in centenarians, and (3) compare the present findings for centenarians to the fluency and speaking rate characteristics reported in the literature.

  6. Mandarin Visual Speech Information

    Science.gov (United States)

    Chen, Trevor H.

    2010-01-01

    While the auditory-only aspects of Mandarin speech are heavily-researched and well-known in the field, this dissertation addresses its lesser-known aspects: The visual and audio-visual perception of Mandarin segmental information and lexical-tone information. Chapter II of this dissertation focuses on the audiovisual perception of Mandarin…

  7. The Commercial Speech Doctrine.

    Science.gov (United States)

    Luebke, Barbara F.

    In its 1942 ruling in the "Valentine vs. Christensen" case, the Supreme Court established the doctrine that commercial speech is not protected by the First Amendment. In 1975, in the "Bigelow vs. Virginia" case, the Supreme Court took a decisive step toward abrogating that doctrine, by ruling that advertising is not stripped of…

  8. Book Review - Telepractice in Speech-Language Pathology

    Directory of Open Access Journals (Sweden)

    Jean Blosser

    2015-01-01

    Full Text Available This article presents a review of the book: Telepractice in Speech-Language Pathology, authored by K. Todd Houston, PhD, CCC-SLP, LSLS Cert. AVT, and 20 contributing authors.   This is the first book entirely devoted to the delivery of speech-language pathology services at a distance. It provides practical information that includes: technical requirements, policy and regulatory issues, current applications in speech-language pathology, international perspectives on practice, and tele-supervision. Reviewer Dr. Jean Blosser highly recommends the work as a comprehensive resource on the topic of telepractice.  

  9. Combined Hand Gesture — Speech Model for Human Action Recognition

    Directory of Open Access Journals (Sweden)

    Sheng-Tzong Cheng

    2013-12-01

    Full Text Available This study proposes a dynamic hand gesture detection technology to effectively detect dynamic hand gesture areas, and a hand gesture recognition technology to improve the dynamic hand gesture recognition rate. Meanwhile, the corresponding relationship between state sequences in hand gesture and speech models is considered by integrating speech recognition technology with a multimodal model, thus improving the accuracy of human behavior recognition. The experimental results proved that the proposed method can effectively improve human behavior recognition accuracy and the feasibility of system applications. Experimental results verified that the multimodal gesture-speech model provided superior accuracy when compared to the single modal versions.

  10. The Unsupervised Acquisition of a Lexicon from Continuous Speech

    CERN Document Server

    De Marcken, C

    1995-01-01

    We present an unsupervised learning algorithm that acquires a natural-language lexicon from raw speech. The algorithm is based on the optimal encoding of symbol sequences in an MDL framework, and uses a hierarchical representation of language that overcomes many of the problems that have stymied previous grammar-induction procedures. The forward mapping from symbol sequences to the speech stream is modeled using features based on articulatory gestures. We present results on the acquisition of lexicons and language models from raw speech, text, and phonetic transcripts, and demonstrate that our algorithm compares very favorably to other reported results with respect to segmentation performance and statistical efficiency.

  11. Two speeches that changed the world: from Fulton to Zurich

    Directory of Open Access Journals (Sweden)

    Alan John Watson

    2016-12-01

    Full Text Available In this extract from his new book Churchill’s Legacy: Two Speeches to Save the World (Watson, 2016, Lord Watson of Richmond draws on his own experience of post war British politics, as a television presenter and media commentator and then as a Liberal Peer and Chairman of the English-Speaking Union, to analyse the significance of Churchill’s Zurich speech of 19 September 1946. He argues that, building on Churchill’s earlier speech at Fulton, Missouri, it helped change the perceptions of the West and alter their response to the emerging Cold War and the future of Europe.

  12. Conversation, speech acts, and memory.

    Science.gov (United States)

    Holtgraves, Thomas

    2008-03-01

    Speakers frequently have specific intentions that they want others to recognize (Grice, 1957). These specific intentions can be viewed as speech acts (Searle, 1969), and I argue that they play a role in long-term memory for conversation utterances. Five experiments were conducted to examine this idea. Participants in all experiments read scenarios ending with either a target utterance that performed a specific speech act (brag, beg, etc.) or a carefully matched control. Participants were more likely to falsely recall and recognize speech act verbs after having read the speech act version than after having read the control version, and the speech act verbs served as better recall cues for the speech act utterances than for the controls. Experiment 5 documented individual differences in the encoding of speech act verbs. The results suggest that people recognize and retain the actions that people perform with their utterances and that this is one of the organizing principles of conversation memory.

  13. Metaheuristic applications to speech enhancement

    CERN Document Server

    Kunche, Prajna

    2016-01-01

    This book serves as a basic reference for those interested in the application of metaheuristics to speech enhancement. The major goal of the book is to explain the basic concepts of optimization methods and their use in heuristic optimization in speech enhancement to scientists, practicing engineers, and academic researchers in speech processing. The authors discuss why it has been a challenging problem for researchers to develop new enhancement algorithms that aid in the quality and intelligibility of degraded speech. They present powerful optimization methods to speech enhancement that can help to solve the noise reduction problems. Readers will be able to understand the fundamentals of speech processing as well as the optimization techniques, how the speech enhancement algorithms are implemented by utilizing optimization methods, and will be given the tools to develop new algorithms. The authors also provide a comprehensive literature survey regarding the topic.

  14. Relationship between speech motor control and speech intelligibility in children with speech sound disorders.

    Science.gov (United States)

    Namasivayam, Aravind Kumar; Pukonen, Margit; Goshulak, Debra; Yu, Vickie Y; Kadis, Darren S; Kroll, Robert; Pang, Elizabeth W; De Nil, Luc F

    2013-01-01

    The current study was undertaken to investigate the impact of speech motor issues on the speech intelligibility of children with moderate to severe speech sound disorders (SSD) within the context of the PROMPT intervention approach. The word-level Children's Speech Intelligibility Measure (CSIM), the sentence-level Beginner's Intelligibility Test (BIT) and tests of speech motor control and articulation proficiency were administered to 12 children (3:11 to 6:7 years) before and after PROMPT therapy. PROMPT treatment was provided for 45 min twice a week for 8 weeks. Twenty-four naïve adult listeners aged 22-46 years judged the intelligibility of the words and sentences. For CSIM, each time a recorded word was played to the listeners they were asked to look at a list of 12 words (multiple-choice format) and circle the word while for BIT sentences, the listeners were asked to write down everything they heard. Words correctly circled (CSIM) or transcribed (BIT) were averaged across three naïve judges to calculate percentage speech intelligibility. Speech intelligibility at both the word and sentence level was significantly correlated with speech motor control, but not articulatory proficiency. Further, the severity of speech motor planning and sequencing issues may potentially be a limiting factor in connected speech intelligibility and highlights the need to target these issues early and directly in treatment. The reader will be able to: (1) outline the advantages and disadvantages of using word- and sentence-level speech intelligibility tests; (2) describe the impact of speech motor control and articulatory proficiency on speech intelligibility; and (3) describe how speech motor control and speech intelligibility data may provide critical information to aid treatment planning. Copyright © 2013 Elsevier Inc. All rights reserved.

  15. Multiresolution analysis applied to text-independent phone segmentation

    Energy Technology Data Exchange (ETDEWEB)

    Cherniz, AnalIa S [Laboratorio de Senales y Dinamicas no Lineales and Laboratorio de Cibernetica, Facultad de Ingenieria, Universidad Nacional de Entre Rios, C.C. 47 Suc. 3 - 3100 Parana (E.R.) (Argentina); Torres, MarIa E [Laboratorio de Senales y Dinamicas no Lineales and Laboratorio de Cibernetica, Facultad de Ingenieria, Universidad Nacional de Entre Rios, C.C. 47 Suc. 3 - 3100 Parana (E.R.) (Argentina); Rufiner, Hugo L [Laboratorio de Senales e Inteligencia Computacional, Universidad Nacional del Litoral, Santa Fe (Argentina); Esposito, Anna [Department of Psychology and IIASS, Second University of Naples, Caserta (Italy)

    2007-11-15

    Automatic speech segmentation is of fundamental importance in different speech applications. The most common implementations are based on hidden Markov models. They use a statistical modelling of the phonetic units to align the data along a known transcription. This is an expensive and time-consuming process, because of the huge amount of data needed to train the system. Text-independent speech segmentation procedures have been developed to overcome some of these problems. These methods detect transitions in the evolution of the time-varying features that represent the speech signal. Speech representation plays a central role is the segmentation task. In this work, two new speech parameterizations based on the continuous multiresolution entropy, using Shannon entropy, and the continuous multiresolution divergence, using Kullback-Leibler distance, are proposed. These approaches have been compared with the classical Melbank parameterization. The proposed encodings increase significantly the segmentation performance. Parameterization based on the continuous multiresolution divergence shows the best results, increasing the number of correctly detected boundaries and decreasing the amount of erroneously inserted points. This suggests that the parameterization based on multiresolution information measures provide information related to acoustic features that take into account phonemic transitions.

  16. A Mobile Phone based Speech Therapist

    OpenAIRE

    Pandey, Vinod K.; Pande, Arun; Kopparapu, Sunil Kumar

    2016-01-01

    Patients with articulatory disorders often have difficulty in speaking. These patients need several speech therapy sessions to enable them speak normally. These therapy sessions are conducted by a specialized speech therapist. The goal of speech therapy is to develop good speech habits as well as to teach how to articulate sounds the right way. Speech therapy is critical for continuous improvement to regain normal speech. Speech therapy sessions require a patient to travel to a hospital or a ...

  17. Digitized Ethnic Hate Speech: Understanding Effects of Digital Media Hate Speech on Citizen Journalism in Kenya

    Science.gov (United States)

    Kimotho, Stephen Gichuhi; Nyaga, Rahab Njeri

    2016-01-01

    Ethnicity in Kenya permeates all spheres of life. However, it is in politics that ethnicity is most visible. Election time in Kenya often leads to ethnic competition and hatred, often expressed through various media. Ethnic hate speech characterized the 2007 general elections in party rallies and through text messages, emails, posters and…

  18. Direct and indirect speech fusion in Charisius’ Ars grammatica

    Directory of Open Access Journals (Sweden)

    Marcos Antonio Pérez Alonso

    2012-06-01

    Full Text Available When transmitting the teaching of his predecessors, the grammarian Charisius usually makes use of either subordinating clauses or direct speech, but he hardly ever employs the so-called indirect speech, typically found in strictly literary works rather than in technical ones. However, we have encountered some passages in which portions in direct speech precede indirect speech clauses, a remarkable phenomenon which could be thought of as textual corruptions needing emendation. Nevertheless, from the fact that this phenomenon is mainly found in parts of Charisius’ work taken from Julius Romanus’ De analogía (where the latter quotes Pliny the Elder’s Dubius sermo, we conclude that the text is correct and that we are dealing with a type of citation characteristic of Julius Romanus, the description of which is not found in the literature on Latin Syntax.

  19. Analytical Study of High Pitch Delay Resolution Technique for Tonal Speech Coding

    Directory of Open Access Journals (Sweden)

    Suphattharachai Chomphan

    2012-01-01

    Full Text Available Problem statement: In tonal-language speech, since tone plays important role not only on the naturalness and also the intelligibility of the speech, it must be treated appropriately in a speech coder algorithm. Approach: This study proposes an analytical study of the technique of High Pitch Delay Resolutions (HPDR applied to the adaptive codebook of core coder of Multi-Pulse based Code Excited Linear Predictive (MP-CELP coder. Results: The experimental results show that the speech quality of the MP-CELP speech coder with HPDR technique is improved above the speech quality of the conventional coder. An optimum resolution of pitch delay is also presented. Conclusion: From the analytical study, it has been found that the proposed technique can improve the speech coding quality.

  20. Speech Recognition Using HMM with MFCC-An Analysis Using Frequency Specral Decomposion Technique

    Directory of Open Access Journals (Sweden)

    Ibrahim Patel

    2010-12-01

    Full Text Available This paper presents an approach to the recognition of speech signal using frequency spectral information with Mel frequency for the improvement of speech feature representation in a HMM based recognition approach. A frequency spectral information is incorporated to the conventional Mel spectrum base speech recognition approach. The Mel frequency approach exploits the frequency observation for speech signal in a given resolution which results in resolution feature overlapping resulting in recognition limit. Resolution decomposition with separating frequency is mapping approach for a HMM based speech recognition system. The Simulation results show an improvement in the quality metrics of speech recognition with respect to computational time, learning accuracy for a speech recognition system.

  1. Enhancement of Non-Air Conducted Speech Based on Wavelet-Packet Adaptive Threshold

    Directory of Open Access Journals (Sweden)

    Xijing Jing

    2013-01-01

    Full Text Available This study developed a new kind of speech detecting method by using millimeter wave. Because of the advantage of the millimeter wave, this speech detecting method has great potential application and may provide some exciting possibility for wide applications. However, the MMW conduct speech is in less intelligible and poor audibility since it is corrupted by additive combined noise. This paper, therefore, also developed an algorithm of wavelet packet threshold by using hard threshold and soft threshold for removing noise based on the good capability of wavelet packet for analyzing time-frequency signal. Comparing to traditional speech enhancement algorithm, the results from both simulation and listening evaluation suggest that the proposed algorithm takes on a better performance on noise removing while the distortion of MMW radar speech remains acceptable, the enhanced speech also sounds more pleasant to human listeners, resulting in improved results over classical speech enhancement algorithms.

  2. Local Linear Wavelet Neural Network and RLS for Usable Speech Classification

    Directory of Open Access Journals (Sweden)

    Suchismita Sahoo

    2011-07-01

    Full Text Available While operating in a co-channel environment, the accuracy of the speech processing technique degrades. When more than one person is talking at same time, then there occurs the co-channel speech. The objective of usable speech segmentation is identification and extraction of those portions of co-channel speech that are degraded in a negligible range but still needed for various speech processing application like speaker identification. Some features like usable speech measures are extracted from the co-channel signal to differentiate between usable and unusable types of speech. The features are extracted recursively by this new method and variable length segmentation is carried out by making sequential decision on class assignment of LLWNN pattern classifier. The correct classification using this technique is 84.5% whereas the false classification is 15.5%. The result shows that the proposed classifier gives better classification and is robust.

  3. Modern Tools in Patient-Centred Speech Therapy for Romanian Language

    Directory of Open Access Journals (Sweden)

    Mirela Danubianu

    2016-03-01

    Full Text Available The most common way to communicate with those around us is speech. Suffering from a speech disorder can have negative social effects: from leaving the individuals with low confidence and moral to problems with social interaction and the ability to live independently like adults. The speech therapy intervention is a complex process having particular objectives such as: discovery and identification of speech disorder and directing the therapy to correction, recovery, compensation, adaptation and social integration of patients. Computer-based Speech Therapy systems are a real help for therapists by creating a special learning environment. The Romanian language is a phonetic one, with special linguistic particularities. This paper aims to present a few computer-based speech therapy systems developed for the treatment of various speech disorders specific to Romanian language.

  4. Commercial applications of speech interface technology: an industry at the threshold.

    Science.gov (United States)

    Oberteuffer, J A

    1995-10-24

    Speech interface technology, which includes automatic speech recognition, synthetic speech, and natural language processing, is beginning to have a significant impact on business and personal computer use. Today, powerful and inexpensive microprocessors and improved algorithms are driving commercial applications in computer command, consumer, data entry, speech-to-text, telephone, and voice verification. Robust speaker-independent recognition systems for command and navigation in personal computers are now available; telephone-based transaction and database inquiry systems using both speech synthesis and recognition are coming into use. Large-vocabulary speech interface systems for document creation and read-aloud proofing are expanding beyond niche markets. Today's applications represent a small preview of a rich future for speech interface technology that will eventually replace keyboards with microphones and loud-speakers to give easy accessibility to increasingly intelligent machines.

  5. Contextual Text Mining

    Science.gov (United States)

    Mei, Qiaozhu

    2009-01-01

    With the dramatic growth of text information, there is an increasing need for powerful text mining systems that can automatically discover useful knowledge from text. Text is generally associated with all kinds of contextual information. Those contexts can be explicit, such as the time and the location where a blog article is written, and the…

  6. Text-Fabric

    NARCIS (Netherlands)

    Roorda, Dirk

    2016-01-01

    Text-Fabric is a Python3 package for Text plus Annotations. It provides a data model, a text file format, and a binary format for (ancient) text plus (linguistic) annotations. The emphasis of this all is on: data processing; sharing data; and contributing modules. A defining characteristic is that T

  7. Contextual Text Mining

    Science.gov (United States)

    Mei, Qiaozhu

    2009-01-01

    With the dramatic growth of text information, there is an increasing need for powerful text mining systems that can automatically discover useful knowledge from text. Text is generally associated with all kinds of contextual information. Those contexts can be explicit, such as the time and the location where a blog article is written, and the…

  8. A case of crossed aphasia with apraxia of speech

    Directory of Open Access Journals (Sweden)

    Yogesh Patidar

    2013-01-01

    Full Text Available Apraxia of speech (AOS is a rare, but well-defined motor speech disorder. It is characterized by irregular articulatory errors, attempts of self-correction and persistent prosodic abnormalities. Similar to aphasia, AOS is also localized to the dominant cerebral hemisphere. We report a case of Crossed Aphasia with AOS in a 48-year-old right-handed man due to an ischemic infarct in right cerebral hemisphere.

  9. ADAPTIVE LEARNING OF HIDDEN MARKOV MODELS FOR EMOTIONAL SPEECH

    Directory of Open Access Journals (Sweden)

    A. V. Tkachenia

    2014-01-01

    Full Text Available An on-line unsupervised algorithm for estimating the hidden Markov models (HMM parame-ters is presented. The problem of hidden Markov models adaptation to emotional speech is solved. To increase the reliability of estimated HMM parameters, a mechanism of forgetting and updating is proposed. A functional block diagram of the hidden Markov models adaptation algorithm is also provided with obtained results, which improve the efficiency of emotional speech recognition.

  10. Assigning phrase breaks from part-of-speech sequences.

    OpenAIRE

    Black, Alan W.; Taylor, Paul A.

    1997-01-01

    One of the important stages in the process of turning unmarked text into speech is the assignment of appropriate phrase break boundaries. Phrase break boundaries are important to later modules including accent assignment, duration control and pause insertion. A number of different algorithms have been proposed for such a task, ranging from the simple to the complex. These different algorithms require different information such as part of speech tags, syntax and even ...

  11. Speech Motor Control in Fluent and Dysfluent Speech Production of an Individual with Apraxia of Speech and Broca's Aphasia

    Science.gov (United States)

    van Lieshout, Pascal H. H. M.; Bose, Arpita; Square, Paula A.; Steele, Catriona M.

    2007-01-01

    Apraxia of speech (AOS) is typically described as a motor-speech disorder with clinically well-defined symptoms, but without a clear understanding of the underlying problems in motor control. A number of studies have compared the speech of subjects with AOS to the fluent speech of controls, but only a few have included speech movement data and if…

  12. SPEECH VISUALIZATION SISTEM AS A BASIS FOR SPEECH TRAINING AND COMMUNICATION AIDS

    Directory of Open Access Journals (Sweden)

    Oliana KRSTEVA

    1997-09-01

    Full Text Available One receives much more information through a visual sense than through a tactile one. However, most visual aids for hearing-impaired persons are not wearable because it is difficult to make them compact and it is not a best way to mask always their vision.Generally it is difficult to get the integrated patterns by a single mathematical transform of signals, such as a Foruier transform. In order to obtain the integrated pattern speech parameters should be carefully extracted by an analysis according as each parameter, and a visual pattern, which can intuitively be understood by anyone, must be synthesized from them. Successful integration of speech parameters will never disturb understanding of individual features, so that the system can be used for speech training and communication.

  13. Auditory and Cognitive Factors Underlying Individual Differences in Aided Speech-Understanding among Older Adults

    Directory of Open Access Journals (Sweden)

    Larry E. Humes

    2013-10-01

    Full Text Available This study was designed to address individual differences in aided speech understanding among a relatively large group of older adults. The group of older adults consisted of 98 adults (50 female and 48 male ranging in age from 60 to 86 (mean = 69.2. Hearing loss was typical for this age group and about 90% had not worn hearing aids. All subjects completed a battery of tests, including cognitive (6 measures, psychophysical (17 measures, and speech-understanding (9 measures, as well as the Speech, Spatial and Qualities of Hearing (SSQ self-report scale. Most of the speech-understanding measures made use of competing speech and the non-speech psychophysical measures were designed to tap phenomena thought to be relevant for the perception of speech in competing speech (e.g., stream segregation, modulation-detection interference. All measures of speech understanding were administered with spectral shaping applied to the speech stimuli to fully restore audibility through at least 4000 Hz. The measures used were demonstrated to be reliable in older adults and, when compared to a reference group of 28 young normal-hearing adults, age-group differences were observed on many of the measures. Principal-components factor analysis was applied successfully to reduce the number of independent and dependent (speech understanding measures for a multiple-regression analysis. Doing so yielded one global cognitive-processing factor and five non-speech psychoacoustic factors (hearing loss, dichotic signal detection, multi-burst masking, stream segregation, and modulation detection as potential predictors. To this set of six potential predictor variables were added subject age, Environmental Sound Identification (ESI, and performance on the text-recognition-threshold (TRT task (a visual analog of interrupted speech recognition. These variables were used to successfully predict one global aided speech-understanding factor, accounting for about 60% of the variance.

  14. Ascriptive speech act and legal language

    Directory of Open Access Journals (Sweden)

    Ogleznev Vitaly

    2016-01-01

    Full Text Available In this article the author explicates Herbert Hart’s theory of an ascriptive language as it has been developed in his influential early paper “The Ascription of Responsibility and Rights”. In the section ‘Discussion’ the author argues that the theory of ascriptive legal utterances, which is grounded on Austin’s and Searle’s theory of a speech act, provides the methodological basis for his analytical approach to philosophical and legal issues. In the section ‘Results’ the author justifies that an ascriptive is a specific speech (illocutionary act. In the section ‘Conclusion’ the matter concerns the original linguistic formula of an ascriptive that accurately reflects its nature. This article elaborates on the interpretation of ascriptive speech acts in legal language by evaluating the influence of philosophy of language on the formation of modern legal philosophy, along with evaluating the contribution of conceptual development of legal philosophy in the speech acts theory.

  15. The impact of voice on speech realization

    Directory of Open Access Journals (Sweden)

    Jelka Breznik

    2014-12-01

    Full Text Available The study discusses spoken literary language and the impact of voice on speech realization. The voice consists of a sound made by a human being using the vocal folds for talking, singing, laughing, crying, screaming… The human voice is specifically the part of human sound production in which the vocal folds (vocal cords are the primary sound source. Our voice is our instrument and identity card. How does the voice (voice tone affect others and how do they respond, positively or negatively? How important is voice (voice tone in communication process? The study presents how certain individuals perceive voice. The results of the research on the relationships between the spoken word, excellent speaker, voice and description / definition / identification of specific voices done by experts in the field of speech and voice as well as non-professionals are presented. The study encompasses two focus groups. One consists of amateurs (non-specialists in the field of speech or voice who have no knowledge in this field and the other consists of professionals who work with speech or language or voice. The questions were intensified from general to specific, directly related to the topic. The purpose of such a method of questioning was to create relaxed atmosphere, promote discussion, allow participants to interact, complement, and to set up self-listening and additional comments.

  16. Speech level shift in Japanese and Slovene

    Directory of Open Access Journals (Sweden)

    Jasmina BAJRAMI

    2016-12-01

    Full Text Available In verbal communication, we always aim to establish and maintain harmonious relations with others. Proper use of expressions and the choice of the way we speak are closely connected with politeness. In Japanese speech level is a level of formality or politeness in conversation, which is expressed by the use of linguistic forms (formal vs. informal within and at the end of an utterance and the use of honorific expressions. In Slovene the level of formality or politeness in conversation is mainly expressed by the use of formal language and general colloquial language. Speech level shift is a shift from one speech level to another – e.g. from a formal style to an informal, etc. According to previous research, these shifts express speaker's psychological distance and a change of attitude towards a hearer. In this paper I will first briefly present the theoretical framework of politeness and an outline of speech levels in Japanese and Slovene. I will then present the data and the method used in this study. Finally, I will present and discuss the results of the analysis of both Japanese and Slovene conversation.

  17. Quality text editing

    Directory of Open Access Journals (Sweden)

    Gyöngyi Bujdosó

    2009-10-01

    Full Text Available Text editing is more than the knowledge of word processing techniques. Originally typographers, printers, text editors were the ones qualified to edit texts, which were well structured, legible, easily understandable, clear, and were able to emphasize the coreof the text. Time has changed, and nowadays everyone has access to computers as well as to text editing software and most users believe that having these tools is enough to edit texts. However, text editing requires more skills. Texts appearing either in printed or inelectronic form reveal that most of the users do not realize that they are not qualified to edit and publish their works. Analyzing the ‘text-products’ of the last decade a tendency can clearly be drawn. More and more documents appear, which instead of emphasizingthe subject matter, are lost in the maze of unstructured text slices. Without further thoughts different font types, colors, sizes, strange arrangements of objects, etc. are applied. We present examples with the most common typographic and text editing errors. Our aim is to call the attention to these mistakes and persuadeusers to spend time to educate themselves in text editing. They have to realize that a well-structured text is able to strengthen the effect on the reader, thus the original message will reach the target group.

  18. Pragmatic Analyses of President Goodluck Jonathan’s Concession Speech and General Muhammadu Buhari’s Acceptance Speech: A Comparative Appraisal

    Directory of Open Access Journals (Sweden)

    Léonard A. Koussouhon

    2016-07-01

    Full Text Available Drawing on Austin’s (1962 Speech Act Theory, this paper investigates President Goodluck Jonathan’s Concession Speech and General Muhammadu Buhari’s Acceptance Speech for the purpose of examining the impacts of context and evaluating their effects on Nigerians. The application of Speech Act Theory to these political discourses has revealed valuable findings. To mention but a few, this study has shown a high proportion of claiming assertive speech acts in Jonathan’s speech indicating thus how unity, stability and progress of Nigeria depends on Jonathan who has excellently proved this by conceding victory to his opponent Buhari. This has been confirmed by the very low proportion of these acts in Buhari’s speech. Furthermore, Jonathan’s acts of thanking, congratulating and praising indicate not only his high degree of recognition, attachment to peace and democracy but also his magnanimity whereas those of Buhari indicate his degree of recognition. Through the use of directive speech acts both Jonathan and Buhari have proved to be law abiding and peaceful. Through the use of commissive speech acts Jonathan has proved to be democratic and patriotic whereas Buhari has proved to be open, cooperative and democratic.  The thoughtful performance of the different speech acts has enabled both speakers especially Jonathan to maintain peace and stability in Nigeria.

  19. Psychosocial Demands of Speech Therapy with Head-and-Neck Cancer Patients: Clinical Experiences, Communicative Skills and Need for Training of Speech Therapists in Oncology

    Directory of Open Access Journals (Sweden)

    Peter Ullrich

    2010-01-01

    Full Text Available Head and neck cancer patients suffer from high levels of psychological comorbidity and often do not receive adequate psychological assistance. Therefore, speech therapists are frequently confronted with specific psychosocial challenges beyond the scope of their standard professional education. The following aspects were examined via a mixed-methods approach to explore their own perceptions of these challenges and their skills to deal with them: 1 The experiences and problems of speech therapists working with head and neck cancer patients; 2 the self-assessment of speech therapists' skills in cancer patient care; 3 the experiences and observations of psychologists supervising case discussions in oncology with speech therapists; 4 the communication skills of speech therapists working with emotionally distressed patients in oncology Our results show that therapists are acutely aware of their patients' distress. They agree that there is a need for patients to receive psychosocial counseling and that a significant portion of their speech therapy sessions is often dedicated to talking about personal problems. This difficult situation leads to: 1 deficiencies in speech therapists' ability to address patients' demands and deficiencies in distancing themselves from their patients; 2 speech therapists feeling overwhelmed with excessive demands and experiencing mental distress themselves. The psycho-oncological knowledge and communication skills among speech therapists were often deficient in light of such demands. There is a need for further training in psychological and social skills for speech therapists working with head and neck cancer patients. URN: urn:nbn:de:0114-fqs1001220

  20. Semantic Text Indexing

    Directory of Open Access Journals (Sweden)

    Zbigniew Kaleta

    2014-01-01

    Full Text Available This article presents a specific issue of the semantic analysis of texts in natural language – text indexing and describes one field of its application (web browsing.The main part of this article describes the computer system assigning a set of semantic indexes (similar to keywords to a particular text. The indexing algorithm employs a semantic dictionary to find specific words in a text, that represent a text content. Furthermore it compares two given sets of semantic indexes to determine texts’ similarity (assigning numerical value. The article describes the semantic dictionary – a tool essentialto accomplish this task and its usefulness, main concepts of the algorithm and test results.