WorldWideScience

Sample records for synchronized auditory signals

  1. Efficient visual search from synchronized auditory signals requires transient audiovisual events.

    Directory of Open Access Journals (Sweden)

    Erik Van der Burg

    Full Text Available BACKGROUND: A prevailing view is that audiovisual integration requires temporally coincident signals. However, a recent study failed to find any evidence for audiovisual integration in visual search even when using synchronized audiovisual events. An important question is what information is critical to observe audiovisual integration. METHODOLOGY/PRINCIPAL FINDINGS: Here we demonstrate that temporal coincidence (i.e., synchrony of auditory and visual components can trigger audiovisual interaction in cluttered displays and consequently produce very fast and efficient target identification. In visual search experiments, subjects found a modulating visual target vastly more efficiently when it was paired with a synchronous auditory signal. By manipulating the kind of temporal modulation (sine wave vs. square wave vs. difference wave; harmonic sine-wave synthesis; gradient of onset/offset ramps we show that abrupt visual events are required for this search efficiency to occur, and that sinusoidal audiovisual modulations do not support efficient search. CONCLUSIONS/SIGNIFICANCE: Thus, audiovisual temporal alignment will only lead to benefits in visual search if the changes in the component signals are both synchronized and transient. We propose that transient signals are necessary in synchrony-driven binding to avoid spurious interactions with unrelated signals when these occur close together in time.

  2. Fronto-parietal and fronto-temporal theta phase synchronization for visual and auditory-verbal working memory.

    Science.gov (United States)

    Kawasaki, Masahiro; Kitajo, Keiichi; Yamaguchi, Yoko

    2014-01-01

    In humans, theta phase (4-8 Hz) synchronization observed on electroencephalography (EEG) plays an important role in the manipulation of mental representations during working memory (WM) tasks; fronto-temporal synchronization is involved in auditory-verbal WM tasks and fronto-parietal synchronization is involved in visual WM tasks. However, whether or not theta phase synchronization is able to select the to-be-manipulated modalities is uncertain. To address the issue, we recorded EEG data from subjects who were performing auditory-verbal and visual WM tasks; we compared the theta synchronizations when subjects performed either auditory-verbal or visual manipulations in separate WM tasks, or performed both two manipulations in the same WM task. The auditory-verbal WM task required subjects to calculate numbers presented by an auditory-verbal stimulus, whereas the visual WM task required subjects to move a spatial location in a mental representation in response to a visual stimulus. The dual WM task required subjects to manipulate auditory-verbal, visual, or both auditory-verbal and visual representations while maintaining auditory-verbal and visual representations. Our time-frequency EEG analyses revealed significant fronto-temporal theta phase synchronization during auditory-verbal manipulation in both auditory-verbal and auditory-verbal/visual WM tasks, but not during visual manipulation tasks. Similarly, we observed significant fronto-parietal theta phase synchronization during visual manipulation tasks, but not during auditory-verbal manipulation tasks. Moreover, we observed significant synchronization in both the fronto-temporal and fronto-parietal theta signals during simultaneous auditory-verbal/visual manipulations. These findings suggest that theta synchronization seems to flexibly connect the brain areas that manipulate WM.

  3. Fronto-parietal and fronto-temporal theta phase synchronization for visual and auditory-verbal working memory

    Directory of Open Access Journals (Sweden)

    Masahiro eKawasaki

    2014-03-01

    Full Text Available In humans, theta phase (4–8 Hz synchronization observed on electroencephalography (EEG plays an important role in the manipulation of mental representations during working memory (WM tasks; fronto-temporal synchronization is involved in auditory-verbal WM tasks and fronto-parietal synchronization is involved in visual WM tasks. However, whether or not theta phase synchronization is able to select the to-be-manipulated modalities is uncertain. To address the issue, we recorded EEG data from subjects who were performing auditory-verbal and visual WM tasks; we compared the theta synchronizations when subjects performed either auditory-verbal or visual manipulations in separate WM tasks, or performed both two manipulations in the same WM task. The auditory-verbal WM task required subjects to calculate numbers presented by an auditory-verbal stimulus, whereas the visual WM task required subjects to move a spatial location in a mental representation in response to a visual stimulus. The dual WM task required subjects to manipulate auditory-verbal, visual, or both auditory-verbal and visual representations while maintaining auditory-verbal and visual representations. Our time-frequency EEG analyses revealed significant fronto-temporal theta phase synchronization during auditory-verbal manipulation in both auditory-verbal and auditory-verbal/visual WM tasks, but not during visual manipulation tasks. Similarly, we observed significant fronto-parietal theta phase synchronization during visual manipulation tasks, but not during auditory-verbal manipulation tasks. Moreover, we observed significant synchronization in both the fronto-temporal and fronto-parietal theta signals during simultaneous auditory-verbal/visual manipulations. These findings suggest that theta synchronization seems to flexibly connect the brain areas that manipulate WM.

  4. Gymnasts utilize visual and auditory information for behavioural synchronization in trampolining.

    Science.gov (United States)

    Heinen, T; Koschnick, J; Schmidt-Maaß, D; Vinken, P M

    2014-08-01

    In synchronized trampolining, two gymnasts perform the same routine at the same time. While trained gymnasts are thought to coordinate their own movements with the movements of another gymnast by detecting relevant movement information, the question arises how visual and auditory information contribute to the emergence of synchronicity between both gymnasts. Therefore the aim of this study was to examine the role of visual and auditory information in the emergence of coordinated behaviour in synchronized trampolining. Twenty female gymnasts were asked to synchronize their leaps with the leaps of a model gymnast, while visual and auditory information was manipulated. The results revealed that gymnasts needed more leaps to reach synchronicity when only either auditory (12.9 leaps) or visual information (10.8 leaps) was available, as compared to when both auditory and visual information was available (8.1 leaps). It is concluded that visual and auditory information play significant roles in synchronized trampolining, whilst visual information seems to be the dominant source for emerging behavioural synchronization, and auditory information supports this emergence.

  5. Synchronization with competing visual and auditory rhythms: bouncing ball meets metronome.

    Science.gov (United States)

    Hove, Michael J; Iversen, John R; Zhang, Allen; Repp, Bruno H

    2013-07-01

    Synchronization of finger taps with periodically flashing visual stimuli is known to be much more variable than synchronization with an auditory metronome. When one of these rhythms is the synchronization target and the other serves as a distracter at various temporal offsets, strong auditory dominance is observed. However, it has recently been shown that visuomotor synchronization improves substantially with moving stimuli such as a continuously bouncing ball. The present study pitted a bouncing ball against an auditory metronome in a target-distracter synchronization paradigm, with the participants being auditory experts (musicians) and visual experts (video gamers and ball players). Synchronization was still less variable with auditory than with visual target stimuli in both groups. For musicians, auditory stimuli tended to be more distracting than visual stimuli, whereas the opposite was the case for the visual experts. Overall, there was no main effect of distracter modality. Thus, a distracting spatiotemporal visual rhythm can be as effective as a distracting auditory rhythm in its capacity to perturb synchronous movement, but its effectiveness also depends on modality-specific expertise.

  6. Synchronization to auditory and visual rhythms in hearing and deaf individuals

    Science.gov (United States)

    Iversen, John R.; Patel, Aniruddh D.; Nicodemus, Brenda; Emmorey, Karen

    2014-01-01

    A striking asymmetry in human sensorimotor processing is that humans synchronize movements to rhythmic sound with far greater precision than to temporally equivalent visual stimuli (e.g., to an auditory vs. a flashing visual metronome). Traditionally, this finding is thought to reflect a fundamental difference in auditory vs. visual processing, i.e., superior temporal processing by the auditory system and/or privileged coupling between the auditory and motor systems. It is unclear whether this asymmetry is an inevitable consequence of brain organization or whether it can be modified (or even eliminated) by stimulus characteristics or by experience. With respect to stimulus characteristics, we found that a moving, colliding visual stimulus (a silent image of a bouncing ball with a distinct collision point on the floor) was able to drive synchronization nearly as accurately as sound in hearing participants. To study the role of experience, we compared synchronization to flashing metronomes in hearing and profoundly deaf individuals. Deaf individuals performed better than hearing individuals when synchronizing with visual flashes, suggesting that cross-modal plasticity enhances the ability to synchronize with temporally discrete visual stimuli. Furthermore, when deaf (but not hearing) individuals synchronized with the bouncing ball, their tapping patterns suggest that visual timing may access higher-order beat perception mechanisms for deaf individuals. These results indicate that the auditory advantage in rhythmic synchronization is more experience- and stimulus-dependent than has been previously reported. PMID:25460395

  7. Subdividing the beat: auditory and motor contributions to synchronization

    NARCIS (Netherlands)

    Loehr, J.D.; Palmer, C.

    2009-01-01

    THE CURRENT STUDY EXAMINED HOW AUDITORY AND kinematic information influenced pianists' ability to synchronize musical sequences with a metronome. Pianists performed melodies in which quarter-note beats were subdivided by intervening eighth notes that resulted from auditory information (heard tones),

  8. Left auditory cortex gamma synchronization and auditory hallucination symptoms in schizophrenia

    Directory of Open Access Journals (Sweden)

    Shenton Martha E

    2009-07-01

    Full Text Available Abstract Background Oscillatory electroencephalogram (EEG abnormalities may reflect neural circuit dysfunction in neuropsychiatric disorders. Previously we have found positive correlations between the phase synchronization of beta and gamma oscillations and hallucination symptoms in schizophrenia patients. These findings suggest that the propensity for hallucinations is associated with an increased tendency for neural circuits in sensory cortex to enter states of oscillatory synchrony. Here we tested this hypothesis by examining whether the 40 Hz auditory steady-state response (ASSR generated in the left primary auditory cortex is positively correlated with auditory hallucination symptoms in schizophrenia. We also examined whether the 40 Hz ASSR deficit in schizophrenia was associated with cross-frequency interactions. Sixteen healthy control subjects (HC and 18 chronic schizophrenia patients (SZ listened to 40 Hz binaural click trains. The EEG was recorded from 60 electrodes and average-referenced offline. A 5-dipole model was fit from the HC grand average ASSR, with 2 pairs of superior temporal dipoles and a deep midline dipole. Time-frequency decomposition was performed on the scalp EEG and source data. Results Phase locking factor (PLF and evoked power were reduced in SZ at fronto-central electrodes, replicating prior findings. PLF was reduced in SZ for non-homologous right and left hemisphere sources. Left hemisphere source PLF in SZ was positively correlated with auditory hallucination symptoms, and was modulated by delta phase. Furthermore, the correlations between source evoked power and PLF found in HC was reduced in SZ for the LH sources. Conclusion These findings suggest that differential neural circuit abnormalities may be present in the left and right auditory cortices in schizophrenia. In addition, they provide further support for the hypothesis that hallucinations are related to cortical hyperexcitability, which is manifested by

  9. Auditory presentation and synchronization in Adobe Flash and HTML5/JavaScript Web experiments.

    Science.gov (United States)

    Reimers, Stian; Stewart, Neil

    2016-09-01

    Substantial recent research has examined the accuracy of presentation durations and response time measurements for visually presented stimuli in Web-based experiments, with a general conclusion that accuracy is acceptable for most kinds of experiments. However, many areas of behavioral research use auditory stimuli instead of, or in addition to, visual stimuli. Much less is known about auditory accuracy using standard Web-based testing procedures. We used a millisecond-accurate Black Box Toolkit to measure the actual durations of auditory stimuli and the synchronization of auditory and visual presentation onsets. We examined the distribution of timings for 100 presentations of auditory and visual stimuli across two computers with difference specs, three commonly used browsers, and code written in either Adobe Flash or JavaScript. We also examined different coding options for attempting to synchronize the auditory and visual onsets. Overall, we found that auditory durations were very consistent, but that the lags between visual and auditory onsets varied substantially across browsers and computer systems.

  10. Neural correlates of auditory temporal predictions during sensorimotor synchronization

    Directory of Open Access Journals (Sweden)

    Nadine ePecenka

    2013-08-01

    Full Text Available Musical ensemble performance requires temporally precise interpersonal action coordination. To play in synchrony, ensemble musicians presumably rely on anticipatory mechanisms that enable them to predict the timing of sounds produced by co-performers. Previous studies have shown that individuals differ in their ability to predict upcoming tempo changes in paced finger-tapping tasks (indexed by cross-correlations between tap timing and pacing events and that the degree of such prediction influences the accuracy of sensorimotor synchronization (SMS and interpersonal coordination in dyadic tapping tasks. The current functional magnetic resonance imaging study investigated the neural correlates of auditory temporal predictions during SMS in a within-subject design. Hemodynamic responses were recorded from 18 musicians while they tapped in synchrony with auditory sequences containing gradual tempo changes under conditions of varying cognitive load (achieved by a simultaneous visual n-back working-memory task comprising three levels of difficulty: observation only, 1-back, and 2-back object comparisons. Prediction ability during SMS decreased with increasing cognitive load. Results of a parametric analysis revealed that the generation of auditory temporal predictions during SMS recruits (1 a distributed network in cortico-cerebellar motor-related brain areas (left dorsal premotor and motor cortex, right lateral cerebellum, SMA proper and bilateral inferior parietal cortex and (2 medial cortical areas (medial prefrontal cortex, posterior cingulate cortex. While the first network is presumably involved in basic sensory prediction, sensorimotor integration, motor timing, and temporal adaptation, activation in the second set of areas may be related to higher-level social-cognitive processes elicited during action coordination with auditory signals that resemble music performed by human agents.

  11. Traffic signal synchronization.

    Science.gov (United States)

    Huang, Ding-wei; Huang, Wei-neng

    2003-05-01

    The benefits of traffic signal synchronization are examined within the cellular automata approach. The microsimulations of traffic flow are obtained with different settings of signal period T and time delay delta. Both numerical results and analytical approximations are presented. For undersaturated traffic, the green-light wave solutions can be realized. For saturated traffic, the correlation among the traffic signals has no effect on the throughput. For oversaturated traffic, the benefits of synchronization are manifest only when stochastic noise is suppressed.

  12. Working Memory and Auditory Imagery Predict Sensorimotor Synchronization with Expressively Timed Music.

    Science.gov (United States)

    Colley, Ian D; Keller, Peter E; Halpern, Andrea R

    2017-08-11

    Sensorimotor synchronization (SMS) is prevalent and readily studied in musical settings, as most people are able to perceive and synchronize with a beat (e.g. by finger tapping). We took an individual differences approach to understanding SMS to real music characterized by expressive timing (i.e. fluctuating beat regularity). Given the dynamic nature of SMS, we hypothesized that individual differences in working memory and auditory imagery-both fluid cognitive processes-would predict SMS at two levels: 1) mean absolute asynchrony (a measure of synchronization error), and 2) anticipatory timing (i.e. predicting, rather than reacting to beat intervals). In Experiment 1, participants completed two working memory tasks, four auditory imagery tasks, and an SMS-tapping task. Hierarchical regression models were used to predict SMS performance, with results showing dissociations among imagery types in relation to mean absolute asynchrony, and evidence of a role for working memory in anticipatory timing. In Experiment 2, a new sample of participants completed an expressive timing perception task to examine the role of imagery in perception without action. Results suggest that imagery vividness is important for perceiving and control is important for synchronizing with, irregular but ecologically valid musical time series. Working memory is implicated in synchronizing by anticipating events in the series.

  13. Synchronization and phonological skills: precise auditory timing hypothesis (PATH

    Directory of Open Access Journals (Sweden)

    Adam eTierney

    2014-11-01

    Full Text Available Phonological skills are enhanced by music training, but the mechanisms enabling this cross-domain enhancement remain unknown. To explain this cross-domain transfer, we propose a precise auditory timing hypothesis (PATH whereby entrainment practice is the core mechanism underlying enhanced phonological abilities in musicians. Both rhythmic synchronization and language skills such as consonant discrimination, detection of word and phrase boundaries, and conversational turn-taking rely on the perception of extremely fine-grained timing details in sound. Auditory-motor timing is an acoustic feature which meets all five of the pre-conditions necessary for cross-domain enhancement to occur (Patel 2011, 2012, 2014. There is overlap between the neural networks that process timing in the context of both music and language. Entrainment to music demands more precise timing sensitivity than does language processing. Moreover, auditory-motor timing integration captures the emotion of the trainee, is repeatedly practiced, and demands focused attention. The precise auditory timing hypothesis predicts that musical training emphasizing entrainment will be particularly effective in enhancing phonological skills.

  14. 40 CFR 93.128 - Traffic signal synchronization projects.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 20 2010-07-01 2010-07-01 false Traffic signal synchronization... synchronization projects. Traffic signal synchronization projects may be approved, funded, and implemented without... include such regionally significant traffic signal synchronization projects. ...

  15. Fronto-parietal and fronto-temporal theta phase synchronization for visual and auditory-verbal working memory

    OpenAIRE

    Masahiro eKawasaki; Masahiro eKawasaki; Masahiro eKawasaki; Keiichi eKitajo; Keiichi eKitajo; Yoko eYamaguchi

    2014-01-01

    In humans, theta phase (4–8 Hz) synchronization observed on electroencephalography (EEG) plays an important role in the manipulation of mental representations during working memory (WM) tasks; fronto-temporal synchronization is involved in auditory-verbal WM tasks and fronto-parietal synchronization is involved in visual WM tasks. However, whether or not theta phase synchronization is able to select the to-be-manipulated modalities is uncertain. To address the issue, we recorded EEG data from...

  16. An Association between Auditory-Visual Synchrony Processing and Reading Comprehension: Behavioral and Electrophysiological Evidence.

    Science.gov (United States)

    Mossbridge, Julia; Zweig, Jacob; Grabowecky, Marcia; Suzuki, Satoru

    2017-03-01

    The perceptual system integrates synchronized auditory-visual signals in part to promote individuation of objects in cluttered environments. The processing of auditory-visual synchrony may more generally contribute to cognition by synchronizing internally generated multimodal signals. Reading is a prime example because the ability to synchronize internal phonological and/or lexical processing with visual orthographic processing may facilitate encoding of words and meanings. Consistent with this possibility, developmental and clinical research has suggested a link between reading performance and the ability to compare visual spatial/temporal patterns with auditory temporal patterns. Here, we provide converging behavioral and electrophysiological evidence suggesting that greater behavioral ability to judge auditory-visual synchrony (Experiment 1) and greater sensitivity of an electrophysiological marker of auditory-visual synchrony processing (Experiment 2) both predict superior reading comprehension performance, accounting for 16% and 25% of the variance, respectively. These results support the idea that the mechanisms that detect auditory-visual synchrony contribute to reading comprehension.

  17. Auditory over visual advantage of sensorimotor synchronization in 6- to 7-year-old children but not in 12- to 15-year-old children and adults.

    Science.gov (United States)

    Mu, Yan; Huang, Yingyu; Ji, Chao; Gu, Li; Wu, Xiang

    2018-05-01

    The superiority of the auditory over visual modality in sensorimotor synchronization-a fundamental ability to coordinate movements with external rhythms-has long been established, whereas recent metronome synchronization work showed that synchronization of a visual bouncing ball was not less stable than synchronization of auditory tones in adults. The present study examined synchronization to isochronous sequences composed of auditory tones, visual flashes, or a bouncing ball in 6- to 7-year-old children, 12- to 15-year-old children, and 19- to 29-year-old adults. Consistent with previous reporting, the results showed that synchronization stability increased with age and synchronization was less stable for flashes than for tones and bouncing balls. As for the focus of the present study, the results revealed that synchronization of the bouncing ball was less stable than synchronization of tones for younger children, but not for teenagers and adults. The finding suggests the predisposition of the auditory advantage of sensorimotor synchronization in early childhood. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  18. The Power of Auditory-Motor Synchronization in Sports: Enhancing Running Performance by Coupling Cadence with the Right Beats

    Science.gov (United States)

    Bood, Robert Jan; Nijssen, Marijn; van der Kamp, John; Roerdink, Melvyn

    2013-01-01

    Acoustic stimuli, like music and metronomes, are often used in sports. Adjusting movement tempo to acoustic stimuli (i.e., auditory-motor synchronization) may be beneficial for sports performance. However, music also possesses motivational qualities that may further enhance performance. Our objective was to examine the relative effects of auditory-motor synchronization and the motivational impact of acoustic stimuli on running performance. To this end, 19 participants ran to exhaustion on a treadmill in 1) a control condition without acoustic stimuli, 2) a metronome condition with a sequence of beeps matching participants’ cadence (synchronization), and 3) a music condition with synchronous motivational music matched to participants’ cadence (synchronization+motivation). Conditions were counterbalanced and measurements were taken on separate days. As expected, time to exhaustion was significantly longer with acoustic stimuli than without. Unexpectedly, however, time to exhaustion did not differ between metronome and motivational music conditions, despite differences in motivational quality. Motivational music slightly reduced perceived exertion of sub-maximal running intensity and heart rates of (near-)maximal running intensity. The beat of the stimuli –which was most salient during the metronome condition– helped runners to maintain a consistent pace by coupling cadence to the prescribed tempo. Thus, acoustic stimuli may have enhanced running performance because runners worked harder as a result of motivational aspects (most pronounced with motivational music) and more efficiently as a result of auditory-motor synchronization (most notable with metronome beeps). These findings imply that running to motivational music with a very prominent and consistent beat matched to the runner’s cadence will likely yield optimal effects because it helps to elevate physiological effort at a high perceived exertion, whereas the consistent and correct cadence induced by

  19. The power of auditory-motor synchronization in sports: enhancing running performance by coupling cadence with the right beats.

    Directory of Open Access Journals (Sweden)

    Robert Jan Bood

    Full Text Available Acoustic stimuli, like music and metronomes, are often used in sports. Adjusting movement tempo to acoustic stimuli (i.e., auditory-motor synchronization may be beneficial for sports performance. However, music also possesses motivational qualities that may further enhance performance. Our objective was to examine the relative effects of auditory-motor synchronization and the motivational impact of acoustic stimuli on running performance. To this end, 19 participants ran to exhaustion on a treadmill in 1 a control condition without acoustic stimuli, 2 a metronome condition with a sequence of beeps matching participants' cadence (synchronization, and 3 a music condition with synchronous motivational music matched to participants' cadence (synchronization+motivation. Conditions were counterbalanced and measurements were taken on separate days. As expected, time to exhaustion was significantly longer with acoustic stimuli than without. Unexpectedly, however, time to exhaustion did not differ between metronome and motivational music conditions, despite differences in motivational quality. Motivational music slightly reduced perceived exertion of sub-maximal running intensity and heart rates of (near-maximal running intensity. The beat of the stimuli -which was most salient during the metronome condition- helped runners to maintain a consistent pace by coupling cadence to the prescribed tempo. Thus, acoustic stimuli may have enhanced running performance because runners worked harder as a result of motivational aspects (most pronounced with motivational music and more efficiently as a result of auditory-motor synchronization (most notable with metronome beeps. These findings imply that running to motivational music with a very prominent and consistent beat matched to the runner's cadence will likely yield optimal effects because it helps to elevate physiological effort at a high perceived exertion, whereas the consistent and correct cadence induced by

  20. The power of auditory-motor synchronization in sports: enhancing running performance by coupling cadence with the right beats.

    Science.gov (United States)

    Bood, Robert Jan; Nijssen, Marijn; van der Kamp, John; Roerdink, Melvyn

    2013-01-01

    Acoustic stimuli, like music and metronomes, are often used in sports. Adjusting movement tempo to acoustic stimuli (i.e., auditory-motor synchronization) may be beneficial for sports performance. However, music also possesses motivational qualities that may further enhance performance. Our objective was to examine the relative effects of auditory-motor synchronization and the motivational impact of acoustic stimuli on running performance. To this end, 19 participants ran to exhaustion on a treadmill in 1) a control condition without acoustic stimuli, 2) a metronome condition with a sequence of beeps matching participants' cadence (synchronization), and 3) a music condition with synchronous motivational music matched to participants' cadence (synchronization+motivation). Conditions were counterbalanced and measurements were taken on separate days. As expected, time to exhaustion was significantly longer with acoustic stimuli than without. Unexpectedly, however, time to exhaustion did not differ between metronome and motivational music conditions, despite differences in motivational quality. Motivational music slightly reduced perceived exertion of sub-maximal running intensity and heart rates of (near-)maximal running intensity. The beat of the stimuli -which was most salient during the metronome condition- helped runners to maintain a consistent pace by coupling cadence to the prescribed tempo. Thus, acoustic stimuli may have enhanced running performance because runners worked harder as a result of motivational aspects (most pronounced with motivational music) and more efficiently as a result of auditory-motor synchronization (most notable with metronome beeps). These findings imply that running to motivational music with a very prominent and consistent beat matched to the runner's cadence will likely yield optimal effects because it helps to elevate physiological effort at a high perceived exertion, whereas the consistent and correct cadence induced by auditory

  1. Objective quantification of the tinnitus decompensation by synchronization measures of auditory evoked single sweeps.

    Science.gov (United States)

    Strauss, Daniel J; Delb, Wolfgang; D'Amelio, Roberto; Low, Yin Fen; Falkai, Peter

    2008-02-01

    Large-scale neural correlates of the tinnitus decompensation might be used for an objective evaluation of therapies and neurofeedback based therapeutic approaches. In this study, we try to identify large-scale neural correlates of the tinnitus decompensation using wavelet phase stability criteria of single sweep sequences of late auditory evoked potentials as synchronization stability measure. The extracted measure provided an objective quantification of the tinnitus decompensation and allowed for a reliable discrimination between a group of compensated and decompensated tinnitus patients. We provide an interpretation for our results by a neural model of top-down projections based on the Jastreboff tinnitus model combined with the adaptive resonance theory which has not been applied to model tinnitus so far. Using this model, our stability measure of evoked potentials can be linked to the focus of attention on the tinnitus signal. It is concluded that the wavelet phase stability of late auditory evoked potential single sweeps might be used as objective tinnitus decompensation measure and can be interpreted in the framework of the Jastreboff tinnitus model and adaptive resonance theory.

  2. Audiovisual integration increases the intentional step synchronization of side-by-side walkers.

    Science.gov (United States)

    Noy, Dominic; Mouta, Sandra; Lamas, Joao; Basso, Daniel; Silva, Carlos; Santos, Jorge A

    2017-12-01

    When people walk side-by-side, they often synchronize their steps. To achieve this, individuals might cross-modally match audiovisual signals from the movements of the partner and kinesthetic, cutaneous, visual and auditory signals from their own movements. Because signals from different sensory systems are processed with noise and asynchronously, the challenge of the CNS is to derive the best estimate based on this conflicting information. This is currently thought to be done by a mechanism operating as a Maximum Likelihood Estimator (MLE). The present work investigated whether audiovisual signals from the partner are integrated according to MLE in order to synchronize steps during walking. Three experiments were conducted in which the sensory cues from a walking partner were virtually simulated. In Experiment 1 seven participants were instructed to synchronize with human-sized Point Light Walkers and/or footstep sounds. Results revealed highest synchronization performance with auditory and audiovisual cues. This was quantified by the time to achieve synchronization and by synchronization variability. However, this auditory dominance effect might have been due to artifacts of the setup. Therefore, in Experiment 2 human-sized virtual mannequins were implemented. Also, audiovisual stimuli were rendered in real-time and thus were synchronous and co-localized. All four participants synchronized best with audiovisual cues. For three of the four participants results point toward their optimal integration consistent with the MLE model. Experiment 3 yielded performance decrements for all three participants when the cues were incongruent. Overall, these findings suggest that individuals might optimally integrate audiovisual cues to synchronize steps during side-by-side walking. Copyright © 2017 Elsevier B.V. All rights reserved.

  3. Temporal expectation weights visual signals over auditory signals.

    Science.gov (United States)

    Menceloglu, Melisa; Grabowecky, Marcia; Suzuki, Satoru

    2017-04-01

    Temporal expectation is a process by which people use temporally structured sensory information to explicitly or implicitly predict the onset and/or the duration of future events. Because timing plays a critical role in crossmodal interactions, we investigated how temporal expectation influenced auditory-visual interaction, using an auditory-visual crossmodal congruity effect as a measure of crossmodal interaction. For auditory identification, an incongruent visual stimulus produced stronger interference when the crossmodal stimulus was presented with an expected rather than an unexpected timing. In contrast, for visual identification, an incongruent auditory stimulus produced weaker interference when the crossmodal stimulus was presented with an expected rather than an unexpected timing. The fact that temporal expectation made visual distractors more potent and visual targets less susceptible to auditory interference suggests that temporal expectation increases the perceptual weight of visual signals.

  4. A basic study on universal design of auditory signals in automobiles.

    Science.gov (United States)

    Yamauchi, Katsuya; Choi, Jong-dae; Maiguma, Ryo; Takada, Masayuki; Iwamiya, Shin-ichiro

    2004-11-01

    In this paper, the impression of various kinds of auditory signals currently used in automobiles and a comprehensive evaluation were measured by a semantic differential method. The desirable acoustic characteristic was examined for each type of auditory signal. Sharp sounds with dominant high-frequency components were not suitable for auditory signals in automobiles. This trend is expedient for the aged whose auditory sensitivity in the high frequency region is lower. When intermittent sounds were used, a longer OFF time was suitable. Generally, "dull (not sharp)" and "calm" sounds were appropriate for auditory signals. Furthermore, the comparison between the frequency spectrum of interior noise in automobiles and that of suitable sounds for various auditory signals indicates that the suitable sounds are not easily masked. The suitable auditory signals for various purposes is a good solution from the viewpoint of universal design.

  5. Large-scale synchronized activity during vocal deviance detection in the zebra finch auditory forebrain.

    Science.gov (United States)

    Beckers, Gabriël J L; Gahr, Manfred

    2012-08-01

    Auditory systems bias responses to sounds that are unexpected on the basis of recent stimulus history, a phenomenon that has been widely studied using sequences of unmodulated tones (mismatch negativity; stimulus-specific adaptation). Such a paradigm, however, does not directly reflect problems that neural systems normally solve for adaptive behavior. We recorded multiunit responses in the caudomedial auditory forebrain of anesthetized zebra finches (Taeniopygia guttata) at 32 sites simultaneously, to contact calls that recur probabilistically at a rate that is used in communication. Neurons in secondary, but not primary, auditory areas respond preferentially to calls when they are unexpected (deviant) compared with the same calls when they are expected (standard). This response bias is predominantly due to sites more often not responding to standard events than to deviant events. When two call stimuli alternate between standard and deviant roles, most sites exhibit a response bias to deviant events of both stimuli. This suggests that biases are not based on a use-dependent decrease in response strength but involve a more complex mechanism that is sensitive to auditory deviance per se. Furthermore, between many secondary sites, responses are tightly synchronized, a phenomenon that is driven by internal neuronal interactions rather than by the timing of stimulus acoustic features. We hypothesize that this deviance-sensitive, internally synchronized network of neurons is involved in the involuntary capturing of attention by unexpected and behaviorally potentially relevant events in natural auditory scenes.

  6. Motor-related signals in the auditory system for listening and learning.

    Science.gov (United States)

    Schneider, David M; Mooney, Richard

    2015-08-01

    In the auditory system, corollary discharge signals are theorized to facilitate normal hearing and the learning of acoustic behaviors, including speech and music. Despite clear evidence of corollary discharge signals in the auditory cortex and their presumed importance for hearing and auditory-guided motor learning, the circuitry and function of corollary discharge signals in the auditory cortex are not well described. In this review, we focus on recent developments in the mouse and songbird that provide insights into the circuitry that transmits corollary discharge signals to the auditory system and the function of these signals in the context of hearing and vocal learning. Copyright © 2015 Elsevier Ltd. All rights reserved.

  7. Phase synchronization of instrumental music signals

    Science.gov (United States)

    Mukherjee, Sayan; Palit, Sanjay Kumar; Banerjee, Santo; Ariffin, M. R. K.; Bhattacharya, D. K.

    2014-06-01

    Signal analysis is one of the finest scientific techniques in communication theory. Some quantitative and qualitative measures describe the pattern of a music signal, vary from one to another. Same musical recital, when played by different instrumentalists, generates different types of music patterns. The reason behind various patterns is the psycho-acoustic measures - Dynamics, Timber, Tonality and Rhythm, varies in each time. However, the psycho-acoustic study of the music signals does not reveal any idea about the similarity between the signals. For such cases, study of synchronization of long-term nonlinear dynamics may provide effective results. In this context, phase synchronization (PS) is one of the measures to show synchronization between two non-identical signals. In fact, it is very critical to investigate any other kind of synchronization for experimental condition, because those are completely non identical signals. Also, there exists equivalence between the phases and the distances of the diagonal line in Recurrence plot (RP) of the signals, which is quantifiable by the recurrence quantification measure τ-recurrence rate. This paper considers two nonlinear music signals based on same raga played by two eminent sitar instrumentalists as two non-identical sources. The psycho-acoustic study shows how the Dynamics, Timber, Tonality and Rhythm vary for the two music signals. Then, long term analysis in the form of phase space reconstruction is performed, which reveals the chaotic phase spaces for both the signals. From the RP of both the phase spaces, τ-recurrence rate is calculated. Finally by the correlation of normalized tau-recurrence rate of their 3D phase spaces and the PS of the two music signals has been established. The numerical results well support the analysis.

  8. Rhythm synchronization performance and auditory working memory in early- and late-trained musicians.

    Science.gov (United States)

    Bailey, Jennifer A; Penhune, Virginia B

    2010-07-01

    Behavioural and neuroimaging studies provide evidence for a possible "sensitive" period in childhood development during which musical training results in long-lasting changes in brain structure and auditory and motor performance. Previous work from our laboratory has shown that adult musicians who begin training before the age of 7 (early-trained; ET) perform better on a visuomotor task than those who begin after the age of 7 (late-trained; LT), even when matched on total years of musical training and experience. Two questions were raised regarding the findings from this experiment. First, would this group performance difference be observed using a more familiar, musically relevant task such as auditory rhythms? Second, would cognitive abilities mediate this difference in task performance? To address these questions, ET and LT musicians, matched on years of musical training, hours of current practice and experience, were tested on an auditory rhythm synchronization task. The task consisted of six woodblock rhythms of varying levels of metrical complexity. In addition, participants were tested on cognitive subtests measuring vocabulary, working memory and pattern recognition. The two groups of musicians differed in their performance of the rhythm task, such that the ET musicians were better at reproducing the temporal structure of the rhythms. There were no group differences on the cognitive measures. Interestingly, across both groups, individual task performance correlated with auditory working memory abilities and years of formal training. These results support the idea of a sensitive period during the early years of childhood for developing sensorimotor synchronization abilities via musical training.

  9. Electrotonic vascular signal conduction and nephron synchronization

    DEFF Research Database (Denmark)

    Marsh, D.J.; Toma, I.; Sosnovtseva, Olga

    2009-01-01

    Marsh DJ, Toma I, Sosnovtseva OV, Peti-Peterdi J, Holstein-Rathlou NH. Electrotonic vascular signal conduction and nephron synchronization. Am J Physiol Renal Physiol 296: F751-F761, 2009. First published December 30, 2008; doi:10.1152/ajprenal.90669.2008.-Tubuloglomerular feedback (TGF) and the ......Marsh DJ, Toma I, Sosnovtseva OV, Peti-Peterdi J, Holstein-Rathlou NH. Electrotonic vascular signal conduction and nephron synchronization. Am J Physiol Renal Physiol 296: F751-F761, 2009. First published December 30, 2008; doi:10.1152/ajprenal.90669.2008.-Tubuloglomerular feedback (TGF......) and the myogenic mechanism control afferent arteriolar diameter in each nephron and regulate blood flow. Both mechanisms generate self-sustained oscillations, the oscillations interact, TGF modulates the frequency and amplitude of the myogenic oscillation, and the oscillations synchronize; a 5: 1 frequency ratio...... is the most frequent. TGF oscillations synchronize in nephron pairs supplied from a common cortical radial artery, as do myogenic oscillations. We propose that electrotonic vascular signal propagation from one juxtaglomerular apparatus interacts with similar signals from other nephrons to produce...

  10. Auditory signal design for automatic number plate recognition system

    NARCIS (Netherlands)

    Heydra, C.G.; Jansen, R.J.; Van Egmond, R.

    2014-01-01

    This paper focuses on the design of an auditory signal for the Automatic Number Plate Recognition system of Dutch national police. The auditory signal is designed to alert police officers of suspicious cars in their proximity, communicating priority level and location of the suspicious car and

  11. A deafening flash! Visual interference of auditory signal detection.

    Science.gov (United States)

    Fassnidge, Christopher; Cecconi Marcotti, Claudia; Freeman, Elliot

    2017-03-01

    In some people, visual stimulation evokes auditory sensations. How prevalent and how perceptually real is this? 22% of our neurotypical adult participants responded 'Yes' when asked whether they heard faint sounds accompanying flash stimuli, and showed significantly better ability to discriminate visual 'Morse-code' sequences. This benefit might arise from an ability to recode visual signals as sounds, thus taking advantage of superior temporal acuity of audition. In support of this, those who showed better visual relative to auditory sequence discrimination also had poorer auditory detection in the presence of uninformative visual flashes, though this was independent of awareness of visually-evoked sounds. Thus a visually-evoked auditory representation may occur subliminally and disrupt detection of real auditory signals. The frequent natural correlation between visual and auditory stimuli might explain the surprising prevalence of this phenomenon. Overall, our results suggest that learned correspondences between strongly correlated modalities may provide a precursor for some synaesthetic abilities. Copyright © 2016 Elsevier Inc. All rights reserved.

  12. Selection of the signal synchronization method in software GPS receivers

    Directory of Open Access Journals (Sweden)

    Vlada S. Sokolović

    2011-04-01

    Full Text Available Introduction This paper presents a critical analysis of the signal processing flow carried out in a software GPS receiver and a critical comparison of different architectures for signal processing within the GPS receiver. A model of software receivers is shown. Based on the displayed model, a receiver has been realized in the MATLAB software package, in which the simulations of signal processing were carried out. The aim of this paper is to demonstrate the advantages and disadvantages of different methods of the synchronization of signals in the receiver, and to propose a solution acceptable for possible implementation. The signal processing flow was observed from the input circuit to the extraction of the bits of the navigation message. The entire signal processing was performed on the L1 signal and the data collected by the input circuit SE4110. A radio signal from the satellite was accepted with the input circuit, filtered and translated into a digital form. The input circuit ends with the hardware of the receiver. A digital signal from the input circuit is brought into the PC Pentium 4 (AMD 3000 + where the receiver is realized in Matlab. Model of software GPS receiver The first level of processing is signal acquisition. Signal acquisition was realized using the cyclic convolution. The acquisition process was carried out by measuring signals from satellites, and these parameters are passed to the next level of processing. The next level was done by monitoring the synchronization signal and extracting the navigation message bits. On the basis of the detection of the navigation message the receiver calculates the position of a satellite and then, based on the position of the satellite, its own position. Tracking of GPS signal synchronization In order to select the most acceptable method of signal synchronization in the receiver, different methods of signal synchronization are compared. The early-late-DLL (Delay Lock Loop, TDL (Tau Dither Loop

  13. Entrainment to an auditory signal: Is attention involved?

    NARCIS (Netherlands)

    Kunert, R.; Jongman, S.R.

    2017-01-01

    Many natural auditory signals, including music and language, change periodically. The effect of such auditory rhythms on the brain is unclear however. One widely held view, dynamic attending theory, proposes that the attentional system entrains to the rhythm and increases attention at moments of

  14. Sensory Entrainment Mechanisms in Auditory Perception: Neural Synchronization Cortico-Striatal Activation.

    Science.gov (United States)

    Sameiro-Barbosa, Catia M; Geiser, Eveline

    2016-01-01

    The auditory system displays modulations in sensitivity that can align with the temporal structure of the acoustic environment. This sensory entrainment can facilitate sensory perception and is particularly relevant for audition. Systems neuroscience is slowly uncovering the neural mechanisms underlying the behaviorally observed sensory entrainment effects in the human sensory system. The present article summarizes the prominent behavioral effects of sensory entrainment and reviews our current understanding of the neural basis of sensory entrainment, such as synchronized neural oscillations, and potentially, neural activation in the cortico-striatal system.

  15. Sensory Entrainment Mechanisms in Auditory Perception: Neural Synchronization Cortico-Striatal Activation

    Science.gov (United States)

    Sameiro-Barbosa, Catia M.; Geiser, Eveline

    2016-01-01

    The auditory system displays modulations in sensitivity that can align with the temporal structure of the acoustic environment. This sensory entrainment can facilitate sensory perception and is particularly relevant for audition. Systems neuroscience is slowly uncovering the neural mechanisms underlying the behaviorally observed sensory entrainment effects in the human sensory system. The present article summarizes the prominent behavioral effects of sensory entrainment and reviews our current understanding of the neural basis of sensory entrainment, such as synchronized neural oscillations, and potentially, neural activation in the cortico-striatal system. PMID:27559306

  16. Pip and pop : Non-spatial auditory signals improve spatial visual search

    NARCIS (Netherlands)

    Burg, E. van der; Olivers, C.N.L.; Bronkhorst, A.W.; Theeuwes, J.

    2008-01-01

    Searching for an object within a cluttered, continuously changing environment can be a very time-consuming process. The authors show that a simple auditory pip drastically decreases search times for a synchronized visual object that is normally very difficult to find. This effect occurs even though

  17. Synchronization transmission of laser pattern signal within uncertain switched network

    Science.gov (United States)

    Lü, Ling; Li, Chengren; Li, Gang; Sun, Ao; Yan, Zhe; Rong, Tingting; Gao, Yan

    2017-06-01

    We propose a new technology for synchronization transmission of laser pattern signal within uncertain network with controllable topology. In synchronization process, the connection of dynamic network can vary at all time according to different demands. Especially, we construct the Lyapunov function of network through designing a special semi-positive definite function, and the synchronization transmission of laser pattern signal within uncertain network with controllable topology can be realized perfectly, which effectively avoids the complicated calculation for solving the second largest eignvalue of the coupling matrix of the dynamic network in order to obtain the network synchronization condition. At the same time, the uncertain parameters in dynamic equations belonging to network nodes can also be identified accurately via designing the identification laws of uncertain parameters. In addition, there are not any limitations for the synchronization target of network in the new technology, in other words, the target can either be a state variable signal of an arbitrary node within the network or an exterior signal.

  18. The power of auditory-motor synchronization in sports: Enhancing running performance by coupling cadence with the right beats

    NARCIS (Netherlands)

    Bood, R.J.; Nijssen, M; van der Kamp, J.; Roerdink, M.

    2013-01-01

    Acoustic stimuli, like music and metronomes, are often used in sports. Adjusting movement tempo to acoustic stimuli (i.e., auditory-motor synchronization) may be beneficial for sports performance. However, music also possesses motivational qualities that may further enhance performance. Our

  19. Size and synchronization of auditory cortex promotes musical, literacy, and attentional skills in children.

    Science.gov (United States)

    Seither-Preisler, Annemarie; Parncutt, Richard; Schneider, Peter

    2014-08-13

    Playing a musical instrument is associated with numerous neural processes that continuously modify the human brain and may facilitate characteristic auditory skills. In a longitudinal study, we investigated the auditory and neural plasticity of musical learning in 111 young children (aged 7-9 y) as a function of the intensity of instrumental practice and musical aptitude. Because of the frequent co-occurrence of central auditory processing disorders and attentional deficits, we also tested 21 children with attention deficit (hyperactivity) disorder [AD(H)D]. Magnetic resonance imaging and magnetoencephalography revealed enlarged Heschl's gyri and enhanced right-left hemispheric synchronization of the primary evoked response (P1) to harmonic complex sounds in children who spent more time practicing a musical instrument. The anatomical characteristics were positively correlated with frequency discrimination, reading, and spelling skills. Conversely, AD(H)D children showed reduced volumes of Heschl's gyri and enhanced volumes of the plana temporalia that were associated with a distinct bilateral P1 asynchrony. This may indicate a risk for central auditory processing disorders that are often associated with attentional and literacy problems. The longitudinal comparisons revealed a very high stability of auditory cortex morphology and gray matter volumes, suggesting that the combined anatomical and functional parameters are neural markers of musicality and attention deficits. Educational and clinical implications are considered. Copyright © 2014 the authors 0270-6474/14/3410937-13$15.00/0.

  20. Vestibular hearing and neural synchronization.

    Science.gov (United States)

    Emami, Seyede Faranak; Daneshi, Ahmad

    2012-01-01

    Objectives. Vestibular hearing as an auditory sensitivity of the saccule in the human ear is revealed by cervical vestibular evoked myogenic potentials (cVEMPs). The range of the vestibular hearing lies in the low frequency. Also, the amplitude of an auditory brainstem response component depends on the amount of synchronized neural activity, and the auditory nerve fibers' responses have the best synchronization with the low frequency. Thus, the aim of this study was to investigate correlation between vestibular hearing using cVEMPs and neural synchronization via slow wave Auditory Brainstem Responses (sABR). Study Design. This case-control survey was consisted of twenty-two dizzy patients, compared to twenty healthy controls. Methods. Intervention comprised of Pure Tone Audiometry (PTA), Impedance acoustic metry (IA), Videonystagmography (VNG), fast wave ABR (fABR), sABR, and cVEMPs. Results. The affected ears of the dizzy patients had the abnormal findings of cVEMPs (insecure vestibular hearing) and the abnormal findings of sABR (decreased neural synchronization). Comparison of the cVEMPs at affected ears versus unaffected ears and the normal persons revealed significant differences (P < 0.05). Conclusion. Safe vestibular hearing was effective in the improvement of the neural synchronization.

  1. Increased BOLD Signals Elicited by High Gamma Auditory Stimulation of the Left Auditory Cortex in Acute State Schizophrenia

    Directory of Open Access Journals (Sweden)

    Hironori Kuga, M.D.

    2016-10-01

    We acquired BOLD responses elicited by click trains of 20, 30, 40 and 80-Hz frequencies from 15 patients with acute episode schizophrenia (AESZ, 14 symptom-severity-matched patients with non-acute episode schizophrenia (NASZ, and 24 healthy controls (HC, assessed via a standard general linear-model-based analysis. The AESZ group showed significantly increased ASSR-BOLD signals to 80-Hz stimuli in the left auditory cortex compared with the HC and NASZ groups. In addition, enhanced 80-Hz ASSR-BOLD signals were associated with more severe auditory hallucination experiences in AESZ participants. The present results indicate that neural over activation occurs during 80-Hz auditory stimulation of the left auditory cortex in individuals with acute state schizophrenia. Given the possible association between abnormal gamma activity and increased glutamate levels, our data may reflect glutamate toxicity in the auditory cortex in the acute state of schizophrenia, which might lead to progressive changes in the left transverse temporal gyrus.

  2. EEG synchronization to modulated auditory tones in schizophrenia, schizoaffective disorder, and schizotypal personality disorder.

    Science.gov (United States)

    Brenner, Colleen A; Sporns, Olaf; Lysaker, Paul H; O'Donnell, Brian F

    2003-12-01

    The authors tested whether neural synchronization deficits were present in subjects with schizophrenia and schizotypal personality disorder. Amplitude-modulated tones were used to evaluate auditory steady-state evoked potential entrainment in a combined group of 21 subjects with schizophrenia or schizoaffective disorder, 11 subjects with schizotypal personality disorder, and 22 nonpsychiatric comparison subjects. The schizophrenia or schizoaffective disorder group exhibited decreased power compared to the schizotypal personality disorder and nonpsychiatric comparison groups. There were no differences between groups in N100 amplitude. Subjects with schizophrenia but not subjects with schizotypal personality disorder have deficits in steady-state responses to periodic stimuli, despite an intact response to sensory-evoked potentials (N100). These deficits reflect aberrant neural synchronization or resolution and may contribute to disturbed perceptual and cognitive integration in schizophrenia.

  3. A virtual auditory environment for investigating the auditory signal processing of realistic sounds

    DEFF Research Database (Denmark)

    Favrot, Sylvain Emmanuel; Buchholz, Jörg

    2008-01-01

    In the present study, a novel multichannel loudspeaker-based virtual auditory environment (VAE) is introduced. The VAE aims at providing a versatile research environment for investigating the auditory signal processing in real environments, i.e., considering multiple sound sources and room...... reverberation. The environment is based on the ODEON room acoustic simulation software to render the acoustical scene. ODEON outputs are processed using a combination of different order Ambisonic techniques to calculate multichannel room impulse responses (mRIR). Auralization is then obtained by the convolution...... the VAE development, special care was taken in order to achieve a realistic auditory percept and to avoid “artifacts” such as unnatural coloration. The performance of the VAE has been evaluated and optimized on a 29 loudspeaker setup using both objective and subjective measurement techniques....

  4. Estimating parameters of chaotic systems synchronized by external driving signal

    International Nuclear Information System (INIS)

    Wu Xiaogang; Wang Zuxi

    2007-01-01

    Noise-induced synchronization (NIS) has evoked great research interests recently. Two uncoupled identical chaotic systems can achieve complete synchronization (CS) by feeding a common noise with appropriate intensity. Actually, NIS belongs to the category of external feedback control (EFC). The significance of applying EFC in secure communication lies in fact that the trajectory of chaotic systems is disturbed so strongly by external driving signal that phase space reconstruction attack fails. In this paper, however, we propose an approach that can accurately estimate the parameters of the chaotic systems synchronized by external driving signal through chaotic transmitted signal, driving signal and their derivatives. Numerical simulation indicates that this approach can estimate system parameters and external coupling strength under two driving modes in a very rapid manner, which implies that EFC is not superior to other methods in secure communication

  5. Sensorimotor synchronization with tempo-changing auditory sequences: Modeling temporal adaptation and anticipation.

    Science.gov (United States)

    van der Steen, M C Marieke; Jacoby, Nori; Fairhurst, Merle T; Keller, Peter E

    2015-11-11

    The current study investigated the human ability to synchronize movements with event sequences containing continuous tempo changes. This capacity is evident, for example, in ensemble musicians who maintain precise interpersonal coordination while modulating the performance tempo for expressive purposes. Here we tested an ADaptation and Anticipation Model (ADAM) that was developed to account for such behavior by combining error correction processes (adaptation) with a predictive temporal extrapolation process (anticipation). While previous computational models of synchronization incorporate error correction, they do not account for prediction during tempo-changing behavior. The fit between behavioral data and computer simulations based on four versions of ADAM was assessed. These versions included a model with adaptation only, one in which adaptation and anticipation act in combination (error correction is applied on the basis of predicted tempo changes), and two models in which adaptation and anticipation were linked in a joint module that corrects for predicted discrepancies between the outcomes of adaptive and anticipatory processes. The behavioral experiment required participants to tap their finger in time with three auditory pacing sequences containing tempo changes that differed in the rate of change and the number of turning points. Behavioral results indicated that sensorimotor synchronization accuracy and precision, while generally high, decreased with increases in the rate of tempo change and number of turning points. Simulations and model-based parameter estimates showed that adaptation mechanisms alone could not fully explain the observed precision of sensorimotor synchronization. Including anticipation in the model increased the precision of simulated sensorimotor synchronization and improved the fit of model to behavioral data, especially when adaptation and anticipation mechanisms were linked via a joint module based on the notion of joint internal

  6. A psychophysiological evaluation of the perceived urgency of auditory warning signals

    Science.gov (United States)

    Burt, J. L.; Bartolome, D. S.; Burdette, D. W.; Comstock, J. R. Jr

    1995-01-01

    One significant concern that pilots have about cockpit auditory warnings is that the signals presently used lack a sense of priority. The relationship between auditory warning sound parameters and perceived urgency is, therefore, an important topic of enquiry in aviation psychology. The present investigation examined the relationship among subjective assessments of urgency, reaction time, and brainwave activity with three auditory warning signals. Subjects performed a tracking task involving automated and manual conditions, and were presented with auditory warnings having various levels of perceived and situational urgency. Subjective assessments revealed that subjects were able to rank warnings on an urgency scale, but rankings were altered after warnings were mapped to a situational urgency scale. Reaction times differed between automated and manual tracking task conditions, and physiological data showed attentional differences in response to perceived and situational warning urgency levels. This study shows that the use of physiological measures sensitive to attention and arousal, in conjunction with behavioural and subjective measures, may lead to the design of auditory warnings that produce a sense of urgency in an operator that matches the urgency of the situation.

  7. A hardware model of the auditory periphery to transduce acoustic signals into neural activity

    Directory of Open Access Journals (Sweden)

    Takashi eTateno

    2013-11-01

    Full Text Available To improve the performance of cochlear implants, we have integrated a microdevice into a model of the auditory periphery with the goal of creating a microprocessor. We constructed an artificial peripheral auditory system using a hybrid model in which polyvinylidene difluoride was used as a piezoelectric sensor to convert mechanical stimuli into electric signals. To produce frequency selectivity, the slit on a stainless steel base plate was designed such that the local resonance frequency of the membrane over the slit reflected the transfer function. In the acoustic sensor, electric signals were generated based on the piezoelectric effect from local stress in the membrane. The electrodes on the resonating plate produced relatively large electric output signals. The signals were fed into a computer model that mimicked some functions of inner hair cells, inner hair cell–auditory nerve synapses, and auditory nerve fibers. In general, the responses of the model to pure-tone burst and complex stimuli accurately represented the discharge rates of high-spontaneous-rate auditory nerve fibers across a range of frequencies greater than 1 kHz and middle to high sound pressure levels. Thus, the model provides a tool to understand information processing in the peripheral auditory system and a basic design for connecting artificial acoustic sensors to the peripheral auditory nervous system. Finally, we discuss the need for stimulus control with an appropriate model of the auditory periphery based on auditory brainstem responses that were electrically evoked by different temporal pulse patterns with the same pulse number.

  8. Bouncing Ball with a Uniformly Varying Velocity in a Metronome Synchronization Task.

    Science.gov (United States)

    Huang, Yingyu; Gu, Li; Yang, Junkai; Wu, Xiang

    2017-09-21

    Sensorimotor synchronization (SMS), a fundamental human ability to coordinate movements with external rhythms, has long been thought to be modality specific. In the canonical metronome synchronization task that requires tapping a finger along with an isochronous sequence, a well-established finding is that synchronization is much more stable to an auditory sequence consisting of auditory tones than to a visual sequence consisting of visual flashes. However, recent studies have shown that periodically moving visual stimuli can substantially improve synchronization compared with visual flashes. In particular, synchronization of a visual bouncing ball that has a uniformly varying velocity was found to be not less stable than synchronization of auditory tones. Here, the current protocol describes the application of the bouncing ball with a uniformly varying velocity in a metronome synchronization task. The usage of the bouncing ball in sequences with different inter-onset intervals (IOI) is included. The representative results illustrate synchronization performance of the bouncing ball, as compared with the performances of auditory tones and visual flashes. Given its comparable synchronization performance to that of auditory tones, the bouncing ball is of particular importance for addressing the current research topic of whether modality-specific mechanisms underlay SMS.

  9. Traffic signal synchronization in the saturated high-density grid road network.

    Science.gov (United States)

    Hu, Xiaojian; Lu, Jian; Wang, Wei; Zhirui, Ye

    2015-01-01

    Most existing traffic signal synchronization strategies do not perform well in the saturated high-density grid road network (HGRN). Traffic congestion often occurs in the saturated HGRN, and the mobility of the network is difficult to restore. In order to alleviate traffic congestion and to improve traffic efficiency in the network, the study proposes a regional traffic signal synchronization strategy, named the long green and long red (LGLR) traffic signal synchronization strategy. The essence of the strategy is to control the formation and dissipation of queues and to maximize the efficiency of traffic flows at signalized intersections in the saturated HGRN. With this strategy, the same signal control timing plan is used at all signalized intersections in the HGRN, and the straight phase of the control timing plan has a long green time and a long red time. Therefore, continuous traffic flows can be maintained when vehicles travel, and traffic congestion can be alleviated when vehicles stop. Using the strategy, the LGLR traffic signal synchronization model is developed, with the objective of minimizing the number of stops. Finally, the simulation is executed to analyze the performance of the model by comparing it to other models, and the superiority of the LGLR model is evident in terms of delay, number of stops, queue length, and overall performance in the saturated HGRN.

  10. Identification of Auditory Object-Specific Attention from Single-Trial Electroencephalogram Signals via Entropy Measures and Machine Learning

    Directory of Open Access Journals (Sweden)

    Yun Lu

    2018-05-01

    Full Text Available Existing research has revealed that auditory attention can be tracked from ongoing electroencephalography (EEG signals. The aim of this novel study was to investigate the identification of peoples’ attention to a specific auditory object from single-trial EEG signals via entropy measures and machine learning. Approximate entropy (ApEn, sample entropy (SampEn, composite multiscale entropy (CmpMSE and fuzzy entropy (FuzzyEn were used to extract the informative features of EEG signals under three kinds of auditory object-specific attention (Rest, Auditory Object1 Attention (AOA1 and Auditory Object2 Attention (AOA2. The linear discriminant analysis and support vector machine (SVM, were used to construct two auditory attention classifiers. The statistical results of entropy measures indicated that there were significant differences in the values of ApEn, SampEn, CmpMSE and FuzzyEn between Rest, AOA1 and AOA2. For the SVM-based auditory attention classifier, the auditory object-specific attention of Rest, AOA1 and AOA2 could be identified from EEG signals using ApEn, SampEn, CmpMSE and FuzzyEn as features and the identification rates were significantly different from chance level. The optimal identification was achieved by the SVM-based auditory attention classifier using CmpMSE with the scale factor τ = 10. This study demonstrated a novel solution to identify the auditory object-specific attention from single-trial EEG signals without the need to access the auditory stimulus.

  11. On the nature of phase attraction in sensorimotor synchronization with interleaved auditory sequences.

    Science.gov (United States)

    Repp, Bruno H

    2004-10-01

    In a task that requires in-phase synchronization of finger taps with an isochronous sequence of target tones that is interleaved with a sequence of distractor tones at various fixed phase relationships, the taps tend to be attracted to the distractor tones, especially when the distractor tones closely precede the target tones [Repp, B. H. (2003a). Phase attraction in sensorimotor synchronization with auditory sequences: Effects of single and periodic distractors on synchronization accuracy. Journal of Experimental Psychology: Human Perception and Performance, 29, 290-309]. The present research addressed two related questions about this distractor effect: (1) Is it a function of the absolute temporal separation or of the relative phase of the two stimulus sequences? (2) Is it the result of perceptual grouping (integration) of target and distractor tones or of simultaneous attraction to two independent sequences? In three experiments, distractor effects were compared across two different sequence rates. The results suggest that absolute temporal separation, not relative phase, is the critical variable. Experiment 3 also included an anti-phase tapping task that addressed the second question directly. The results suggest that the attraction of taps to distractor tones is caused mainly by temporal integration of target and distractor tones within a fixed window of 100-150 ms duration, with the earlier-occurring tone being weighted more strongly than the later-occurring one.

  12. Contextual modulation of primary visual cortex by auditory signals.

    Science.gov (United States)

    Petro, L S; Paton, A T; Muckli, L

    2017-02-19

    Early visual cortex receives non-feedforward input from lateral and top-down connections (Muckli & Petro 2013 Curr. Opin. Neurobiol. 23, 195-201. (doi:10.1016/j.conb.2013.01.020)), including long-range projections from auditory areas. Early visual cortex can code for high-level auditory information, with neural patterns representing natural sound stimulation (Vetter et al. 2014 Curr. Biol. 24, 1256-1262. (doi:10.1016/j.cub.2014.04.020)). We discuss a number of questions arising from these findings. What is the adaptive function of bimodal representations in visual cortex? What type of information projects from auditory to visual cortex? What are the anatomical constraints of auditory information in V1, for example, periphery versus fovea, superficial versus deep cortical layers? Is there a putative neural mechanism we can infer from human neuroimaging data and recent theoretical accounts of cortex? We also present data showing we can read out high-level auditory information from the activation patterns of early visual cortex even when visual cortex receives simple visual stimulation, suggesting independent channels for visual and auditory signals in V1. We speculate which cellular mechanisms allow V1 to be contextually modulated by auditory input to facilitate perception, cognition and behaviour. Beyond cortical feedback that facilitates perception, we argue that there is also feedback serving counterfactual processing during imagery, dreaming and mind wandering, which is not relevant for immediate perception but for behaviour and cognition over a longer time frame.This article is part of the themed issue 'Auditory and visual scene analysis'. © 2017 The Authors.

  13. Development of a wireless system for auditory neuroscience.

    Science.gov (United States)

    Lukes, A J; Lear, A T; Snider, R K

    2001-01-01

    In order to study how the auditory cortex extracts communication sounds in a realistic acoustic environment, a wireless system is being developed that will transmit acoustic as well as neural signals. The miniature transmitter will be capable of transmitting two acoustic signals with 37.5 KHz bandwidths (75 KHz sample rate) and 56 neural signals with bandwidths of 9.375 KHz (18.75 KHz sample rate). These signals will be time-division multiplexed into one high bandwidth signal with a 1.2 MHz sample rate. This high bandwidth signal will then be frequency modulated onto a 2.4 GHz carrier, which resides in the industrial, scientic, and medical (ISM) band that is designed for low-power short-range wireless applications. On the receiver side, the signal will be demodulated from the 2.4 GHz carrier and then digitized by an analog-to-digital (A/D) converter. The acoustic and neural signals will be digitally demultiplexed from the multiplexed signal into their respective channels. Oversampling (20 MHz) will allow the reconstruction of the multiplexing clock by a digital signal processor (DSP) that will perform frame and bit synchronization. A frame is a subset of the signal that contains all the channels and several channels tied high and low will signal the start of a frame. This technological development will bring two benefits to auditory neuroscience. It will allow simultaneous recording of many neurons that will permit studies of population codes. It will also allow neural functions to be determined in higher auditory areas by correlating neural and acoustic signals without apriori knowledge of the necessary stimuli.

  14. Impulsive synchronization of Roessler systems with parameter driven by an external signal

    International Nuclear Information System (INIS)

    Zhang Rong; Hu Manfeng; Xu Zhenyuan

    2007-01-01

    In this Letter, an impulsive control scheme is presented to control two uncoupled identical Roessler systems. By driving the parameter of Roessler systems using external chaotic signal or periodic signal, chaotic synchronization and periodic synchronization can be implemented. This is a special impulsive control, but by using the existing results of impulsive control theory, a less conservative estimation of the upper bound of the impulse interval is given, which can guarantee the global asymptotical stability for the impulsive synchronization of Roessler systems. Numerical results are in accord with our estimation

  15. Lag synchronization of Rossler system and Chua circuit via a scalar signal

    International Nuclear Information System (INIS)

    Li Chuandong; Liao Xiaofeng

    2004-01-01

    In this Letter, a chaotic lag synchronization scheme is proposed based on combining a nonlinear with lag-in-time observer design. Our approach leads to a systematic methodology, which guarantees the synchronization of a wide class of chaotic systems via a scalar signal. The proposed technique has been applied to synchronize two well-known chaotic systems: Rossler's system and Chua circuit

  16. Influence of signal processing strategy in auditory abilities.

    Science.gov (United States)

    Melo, Tatiana Mendes de; Bevilacqua, Maria Cecília; Costa, Orozimbo Alves; Moret, Adriane Lima Mortari

    2013-01-01

    The signal processing strategy is a parameter that may influence the auditory performance of cochlear implant and is important to optimize this parameter to provide better speech perception, especially in difficult listening situations. To evaluate the individual's auditory performance using two different signal processing strategy. Prospective study with 11 prelingually deafened children with open-set speech recognition. A within-subjects design was used to compare performance with standard HiRes and HiRes 120 in three different moments. During test sessions, subject's performance was evaluated by warble-tone sound-field thresholds, speech perception evaluation, in quiet and in noise. In the silence, children S1, S4, S5, S7 showed better performance with the HiRes 120 strategy and children S2, S9, S11 showed better performance with the HiRes strategy. In the noise was also observed that some children performed better using the HiRes 120 strategy and other with HiRes. Not all children presented the same pattern of response to the different strategies used in this study, which reinforces the need to look at optimizing cochlear implant clinical programming.

  17. Joint Iterative Carrier Synchronization and Signal Detection for Dual Carrier 448 Gb/s PDM 16-QAM

    DEFF Research Database (Denmark)

    Zibar, Darko; Carvalho, Luis; Estaran Tolosa, Jose Manuel

    2013-01-01

    Soft decision driven joint carrier synchronization and signal detection, employing expectation maximization, is experimentally demonstrated. Employing soft decisions offers an improvement of 0.5 dB compared to hard decision digital PLL based carrier synchronization and demodulation.......Soft decision driven joint carrier synchronization and signal detection, employing expectation maximization, is experimentally demonstrated. Employing soft decisions offers an improvement of 0.5 dB compared to hard decision digital PLL based carrier synchronization and demodulation....

  18. Feeling the Beat: Bouncing Synchronization to Vibrotactile Music in Hearing and Early Deaf People

    Directory of Open Access Journals (Sweden)

    Pauline Tranchant

    2017-09-01

    Full Text Available The ability to dance relies on the ability to synchronize movements to a perceived musical beat. Typically, beat synchronization is studied with auditory stimuli. However, in many typical social dancing situations, music can also be perceived as vibrations when objects that generate sounds also generate vibrations. This vibrotactile musical perception is of particular relevance for deaf people, who rely on non-auditory sensory information for dancing. In the present study, we investigated beat synchronization to vibrotactile electronic dance music in hearing and deaf people. We tested seven deaf and 14 hearing individuals on their ability to bounce in time with the tempo of vibrotactile stimuli (no sound delivered through a vibrating platform. The corresponding auditory stimuli (no vibrations were used in an additional condition in the hearing group. We collected movement data using a camera-based motion capture system and subjected it to a phase-locking analysis to assess synchronization quality. The vast majority of participants were able to precisely time their bounces to the vibrations, with no difference in performance between the two groups. In addition, we found higher performance for the auditory condition compared to the vibrotactile condition in the hearing group. Our results thus show that accurate tactile-motor synchronization in a dance-like context occurs regardless of auditory experience, though auditory-motor synchronization is of superior quality.

  19. Cluster synchronization transmission of different external signals in discrete uncertain network

    Science.gov (United States)

    Li, Chengren; Lü, Ling; Chen, Liansong; Hong, Yixuan; Zhou, Shuang; Yang, Yiming

    2018-07-01

    We research cluster synchronization transmissions of different external signals in discrete uncertain network. Based on the Lyapunov theorem, the network controller and the identification law of uncertain adjustment parameter are designed, and they are efficiently used to achieve the cluster synchronization and the identification of uncertain adjustment parameter. In our technical scheme, the network nodes in each cluster and the transmitted external signal can be different, and they allow the presence of uncertain parameters in the network. Especially, we are free to choose the clustering topologies, the cluster number and the node number in each cluster.

  20. Synchronous Modeling of Modular Avionics Architectures using the SIGNAL Language

    OpenAIRE

    Gamatié , Abdoulaye; Gautier , Thierry

    2002-01-01

    This document presents a study on the modeling of architecture components for avionics applications. We consider the avionics standard ARINC 653 specifications as basis, as well as the synchronous language SIGNAL to describe the modeling. A library of APEX object models (partition, process, communication and synchronization services, etc.) has been implemented. This should allow to describe distributed real-time applications using POLYCHRONY, so as to access formal tools and techniques for ar...

  1. The influence of signal type on the internal auditory representation of a room

    Science.gov (United States)

    Teret, Elizabeth

    Currently, architectural acousticians make no real distinction between a room impulse response and the auditory system's internal representation of a room. With this lack of a good model for the auditory representation of a room, it is indirectly assumed that our internal representation of a room is independent of the sound source needed to make the room characteristics audible. The extent to which this assumption holds true is examined with perceptual tests. Listeners are presented with various pairs of signals (music, speech, and noise) convolved with synthesized impulse responses of different reverberation times. They are asked to adjust the reverberation of one of the signals to match the other. Analysis of the data show that the source signal significantly influences perceived reverberance. Listeners are less accurate when matching reverberation times of varied signals than they are with identical signals. Additional testing shows that perception of reverberation can be linked to the existence of transients in the signal.

  2. Inter-subject synchronization of brain responses during natural music listening

    Science.gov (United States)

    Abrams, Daniel A.; Ryali, Srikanth; Chen, Tianwen; Chordia, Parag; Khouzam, Amirah; Levitin, Daniel J.; Menon, Vinod

    2015-01-01

    Music is a cultural universal and a rich part of the human experience. However, little is known about common brain systems that support the processing and integration of extended, naturalistic ‘real-world’ music stimuli. We examined this question by presenting extended excerpts of symphonic music, and two pseudomusical stimuli in which the temporal and spectral structure of the Natural Music condition were disrupted, to non-musician participants undergoing functional brain imaging and analysing synchronized spatiotemporal activity patterns between listeners. We found that music synchronizes brain responses across listeners in bilateral auditory midbrain and thalamus, primary auditory and auditory association cortex, right-lateralized structures in frontal and parietal cortex, and motor planning regions of the brain. These effects were greater for natural music compared to the pseudo-musical control conditions. Remarkably, inter-subject synchronization in the inferior colliculus and medial geniculate nucleus was also greater for the natural music condition, indicating that synchronization at these early stages of auditory processing is not simply driven by spectro-temporal features of the stimulus. Increased synchronization during music listening was also evident in a right-hemisphere fronto-parietal attention network and bilateral cortical regions involved in motor planning. While these brain structures have previously been implicated in various aspects of musical processing, our results are the first to show that these regions track structural elements of a musical stimulus over extended time periods lasting minutes. Our results show that a hierarchical distributed network is synchronized between individuals during the processing of extended musical sequences, and provide new insight into the temporal integration of complex and biologically salient auditory sequences. PMID:23578016

  3. The speech signal segmentation algorithm using pitch synchronous analysis

    Directory of Open Access Journals (Sweden)

    Amirgaliyev Yedilkhan

    2017-03-01

    Full Text Available Parameterization of the speech signal using the algorithms of analysis synchronized with the pitch frequency is discussed. Speech parameterization is performed by the average number of zero transitions function and the signal energy function. Parameterization results are used to segment the speech signal and to isolate the segments with stable spectral characteristics. Segmentation results can be used to generate a digital voice pattern of a person or be applied in the automatic speech recognition. Stages needed for continuous speech segmentation are described.

  4. Video Synchronization With Bit-Rate Signals and Correntropy Function

    Directory of Open Access Journals (Sweden)

    Igor Pereira

    2017-09-01

    Full Text Available We propose an approach for the synchronization of video streams using correntropy. Essentially, the time offset is calculated on the basis of the instantaneous transfer rates of the video streams that are extracted in the form of a univariate signal known as variable bit-rate (VBR. The state-of-the-art approach uses a window segmentation strategy that is based on consensual zero-mean normalized cross-correlation (ZNCC. This strategy has an elevated computational complexity, making its application to synchronizing online data streaming difficult. Hence, our proposal uses a different window strategy that, together with the correntropy function, allows the synchronization to be performed for online applications. This provides equivalent synchronization scores with a rapid offset determination as the streams come into the system. The efficiency of our approach has been verified through experiments that demonstrate its viability with values that are as precise as those obtained by ZNCC. The proposed approach scored 81 % in time reference classification against the equivalent 81 % of the state-of-the-art approach, requiring much less computational power.

  5. Simulating GPS radio signal to synchronize network--a new technique for redundant timing.

    Science.gov (United States)

    Shan, Qingxiao; Jun, Yang; Le Floch, Jean-Michel; Fan, Yaohui; Ivanov, Eugene N; Tobar, Michael E

    2014-07-01

    Currently, many distributed systems such as 3G mobile communications and power systems are time synchronized with a Global Positioning System (GPS) signal. If there is a GPS failure, it is difficult to realize redundant timing, and thus time-synchronized devices may fail. In this work, we develop time transfer by simulating GPS signals, which promises no extra modification to original GPS-synchronized devices. This is achieved by applying a simplified GPS simulator for synchronization purposes only. Navigation data are calculated based on a pre-assigned time at a fixed position. Pseudo-range data which describes the distance change between the space vehicle (SV) and users are calculated. Because real-time simulation requires heavy-duty computations, we use self-developed software optimized on a PC to generate data, and save the data onto memory disks while the simulator is operating. The radio signal generation is similar to the SV at an initial position, and the frequency synthesis of the simulator is locked to a pre-assigned time. A filtering group technique is used to simulate the signal transmission delay corresponding to the SV displacement. Each SV generates a digital baseband signal, where a unique identifying code is added to the signal and up-converted to generate the output radio signal at the centered frequency of 1575.42 MHz (L1 band). A prototype with a field-programmable gate array (FPGA) has been built and experiments have been conducted to prove that we can realize time transfer. The prototype has been applied to the CDMA network for a three-month long experiment. Its precision has been verified and can meet the requirements of most telecommunication systems.

  6. Analysis of the influence of memory content of auditory stimuli on the memory content of EEG signal.

    Science.gov (United States)

    Namazi, Hamidreza; Khosrowabadi, Reza; Hussaini, Jamal; Habibi, Shaghayegh; Farid, Ali Akhavan; Kulish, Vladimir V

    2016-08-30

    One of the major challenges in brain research is to relate the structural features of the auditory stimulus to structural features of Electroencephalogram (EEG) signal. Memory content is an important feature of EEG signal and accordingly the brain. On the other hand, the memory content can also be considered in case of stimulus. Beside all works done on analysis of the effect of stimuli on human EEG and brain memory, no work discussed about the stimulus memory and also the relationship that may exist between the memory content of stimulus and the memory content of EEG signal. For this purpose we consider the Hurst exponent as the measure of memory. This study reveals the plasticity of human EEG signals in relation to the auditory stimuli. For the first time we demonstrated that the memory content of an EEG signal shifts towards the memory content of the auditory stimulus used. The results of this analysis showed that an auditory stimulus with higher memory content causes a larger increment in the memory content of an EEG signal. For the verification of this result, we benefit from approximate entropy as indicator of time series randomness. The capability, observed in this research, can be further investigated in relation to human memory.

  7. Intracranial electroencephalography power and phase synchronization changes during monaural and binaural beat stimulation.

    Science.gov (United States)

    Becher, Ann-Katrin; Höhne, Marlene; Axmacher, Nikolai; Chaieb, Leila; Elger, Christian E; Fell, Juergen

    2015-01-01

    Auditory stimulation with monaural or binaural auditory beats (i.e. sine waves with nearby frequencies presented either to both ears or to each ear separately) represents a non-invasive approach to influence electrical brain activity. It is still unclear exactly which brain sites are affected by beat stimulation. In particular, an impact of beat stimulation on mediotemporal brain areas could possibly provide new options for memory enhancement or seizure control. Therefore, we examined how electroencephalography (EEG) power and phase synchronization are modulated by auditory stimulation with beat frequencies corresponding to dominant EEG rhythms based on intracranial recordings in presurgical epilepsy patients. Monaural and binaural beat stimuli with beat frequencies of 5, 10, 40 and 80 Hz and non-superposed control signals were administered with low amplitudes (60 dB SPL) and for short durations (5 s). EEG power was intracranially recorded from mediotemporal, temporo-basal and temporo-lateral and surface sites. Evoked and total EEG power and phase synchronization during beat vs. control stimulation were compared by the use of Bonferroni-corrected non-parametric label-permutation tests. We found that power and phase synchronization were significantly modulated by beat stimulation not only at temporo-basal, temporo-lateral and surface sites, but also at mediotemporal sites. Generally, more significant decreases than increases were observed. The most prominent power increases were seen after stimulation with monaural 40-Hz beats. The most pronounced power and synchronization decreases resulted from stimulation with monaural 5-Hz and binaural 80-Hz beats. Our results suggest that beat stimulation offers a non-invasive approach for the modulation of intracranial EEG characteristics. © 2014 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  8. Sensorimotor synchronization with different metrical levels of point-light dance movements

    Directory of Open Access Journals (Sweden)

    Yi-Huang eSu

    2016-04-01

    Full Text Available Rhythm perception and synchronization have been extensively investigated in the auditory domain, as they underlie means of human communication such as music and speech. Although recent studies suggest comparable mechanisms for synchronizing with periodically moving visual objects, the extent to which it applies to ecologically relevant information, such as the rhythm of complex biological motion, remains unknown. The present study addressed this issue by linking rhythm of music and dance in the framework of action-perception coupling. As a previous study showed that observers perceived multiple metrical periodicities in dance movements that embodied this structure, the present study examined whether sensorimotor synchronization (SMS to dance movements resembles what is known of auditory SMS. Participants watched a point-light figure performing two basic steps of Swing dance cyclically, in which the trunk bounced at every beat and the limbs moved at every second beat, forming two metrical periodicities. Participants tapped synchronously to the bounce of the trunk with or without the limbs moving in the stimuli (Experiment 1, or tapped synchronously to the leg movements with or without the trunk bouncing simultaneously (Experiment 2. Results showed that, while synchronization with the bounce (lower-level pulse was not influenced by the presence or absence of limb movements (metrical accent, synchronization with the legs (beat was improved by the presence of the bounce (metrical subdivision across different movement types. The latter finding parallels the subdivision benefit often demonstrated in auditory tasks, suggesting common sensorimotor mechanisms for visual rhythms in dance and auditory rhythms in music.

  9. Sensorimotor Synchronization with Different Metrical Levels of Point-Light Dance Movements

    Science.gov (United States)

    Su, Yi-Huang

    2016-01-01

    Rhythm perception and synchronization have been extensively investigated in the auditory domain, as they underlie means of human communication such as music and speech. Although recent studies suggest comparable mechanisms for synchronizing with periodically moving visual objects, the extent to which it applies to ecologically relevant information, such as the rhythm of complex biological motion, remains unknown. The present study addressed this issue by linking rhythm of music and dance in the framework of action-perception coupling. As a previous study showed that observers perceived multiple metrical periodicities in dance movements that embodied this structure, the present study examined whether sensorimotor synchronization (SMS) to dance movements resembles what is known of auditory SMS. Participants watched a point-light figure performing two basic steps of Swing dance cyclically, in which the trunk bounced at every beat and the limbs moved at every second beat, forming two metrical periodicities. Participants tapped synchronously to the bounce of the trunk with or without the limbs moving in the stimuli (Experiment 1), or tapped synchronously to the leg movements with or without the trunk bouncing simultaneously (Experiment 2). Results showed that, while synchronization with the bounce (lower-level pulse) was not influenced by the presence or absence of limb movements (metrical accent), synchronization with the legs (beat) was improved by the presence of the bounce (metrical subdivision) across different movement types. The latter finding parallels the “subdivision benefit” often demonstrated in auditory tasks, suggesting common sensorimotor mechanisms for visual rhythms in dance and auditory rhythms in music. PMID:27199709

  10. Synchronization of femtosecond laser pulses and RF signal by using a Sagnac loop Mach-Zehnder interferometer

    International Nuclear Information System (INIS)

    Dai Hui; Hajima, Ryoichi

    2008-11-01

    For future advanced energy recovery linac to generate femtosecond X-ray pulses, precise synchronization between sub-systems is highly desired. Typical synchronization methods based on direct photo detection are limited by detector nonlinearities, which lead to amplitude-to-phase conversion and introduce excess timing jitter. In this paper, we experimentally demonstrate an optical-electronic mixed phase lock loop to synchronize the RF signal and laser pulses. In this synchronism setup, a Sagnac-loop Mach-Zehnder interferometer has been used to suppress the excess noise of direct photo detection. This scheme transfers the timing information into a intensity imbalance between the two output beams of the interferometer. As experimental demonstration, the single side-band phase noise of RF signal from the VCO is locked to the mode-locked Ti:Sapphire laser in the spectrum covering the range of 10 kHz to 1 MHz. This synchronization scheme greatly reduces the phase noise and timing jitter of the RF signal. (author)

  11. Detection of auditory signals in quiet and noisy backgrounds while performing a visuo-spatial task

    Directory of Open Access Journals (Sweden)

    Vishakha W Rawool

    2016-01-01

    Full Text Available Context: The ability to detect important auditory signals while performing visual tasks may be further compounded by background chatter. Thus, it is important to know how task performance may interact with background chatter to hinder signal detection. Aim: To examine any interactive effects of speech spectrum noise and task performance on the ability to detect signals. Settings and Design: The setting was a sound-treated booth. A repeated measures design was used. Materials and Methods: Auditory thresholds of 20 normal adults were determined at 0.5, 1, 2 and 4 kHz in the following conditions presented in a random order: (1 quiet with attention; (2 quiet with a visuo-spatial task or puzzle (distraction; (3 noise with attention and (4 noise with task. Statistical Analysis: Multivariate analyses of variance (MANOVA with three repeated factors (quiet versus noise, visuo-spatial task versus no task, signal frequency. Results: MANOVA revealed significant main effects for noise and signal frequency and significant noise–frequency and task–frequency interactions. Distraction caused by performing the task worsened the thresholds for tones presented at the beginning of the experiment and had no effect on tones presented in the middle. At the end of the experiment, thresholds (4 kHz were better while performing the task than those obtained without performing the task. These effects were similar across the quiet and noise conditions. Conclusion: Detection of auditory signals is difficult at the beginning of a distracting visuo-spatial task but over time, task learning and auditory training effects can nullify the effect of distraction and may improve detection of high frequency sounds.

  12. Auditory Warnings, Signal-Referent Relations, and Natural Indicators: Re-Thinking Theory and Application

    Science.gov (United States)

    Petocz, Agnes; Keller, Peter E.; Stevens, Catherine J.

    2008-01-01

    In auditory warning design the idea of the strength of the association between sound and referent has been pivotal. Research has proceeded via constructing classification systems of signal-referent associations and then testing predictions about ease of learning of different levels of signal-referent relation strength across and within different…

  13. Sustained Attention in Auditory and Visual Monitoring Tasks: Evaluation of the Administration of a Rest Break or Exogenous Vibrotactile Signals.

    Science.gov (United States)

    Arrabito, G Robert; Ho, Geoffrey; Aghaei, Behzad; Burns, Catherine; Hou, Ming

    2015-12-01

    Performance and mental workload were observed for the administration of a rest break or exogenous vibrotactile signals in auditory and visual monitoring tasks. Sustained attention is mentally demanding. Techniques are required to improve observer performance in vigilance tasks. Participants (N = 150) monitored an auditory or a visual display for changes in signal duration in a 40-min watch. During the watch, participants were administered a rest break or exogenous vibrotactile signals. Detection accuracy was significantly greater in the auditory than in the visual modality. A short rest break restored detection accuracy in both sensory modalities following deterioration in performance. Participants experienced significantly lower mental workload when monitoring auditory than visual signals, and a rest break significantly reduced mental workload in both sensory modalities. Exogenous vibrotactile signals had no beneficial effects on performance, or mental workload. A rest break can restore performance in auditory and visual vigilance tasks. Although sensory differences in vigilance tasks have been studied, this study is the initial effort to investigate the effects of a rest break countermeasure in both auditory and visual vigilance tasks, and it is also the initial effort to explore the effects of the intervention of a rest break on the perceived mental workload of auditory and visual vigilance tasks. Further research is warranted to determine exact characteristics of effective exogenous vibrotactile signals in vigilance tasks. Potential applications of this research include procedures for decreasing the temporal decline in observer performance and the high mental workload imposed by vigilance tasks. © 2015, Her Majesty the Queen in Right of Canada, as represented by the Minister of National Defence.

  14. Perceptual consequences of disrupted auditory nerve activity.

    Science.gov (United States)

    Zeng, Fan-Gang; Kong, Ying-Yee; Michalewski, Henry J; Starr, Arnold

    2005-06-01

    Perceptual consequences of disrupted auditory nerve activity were systematically studied in 21 subjects who had been clinically diagnosed with auditory neuropathy (AN), a recently defined disorder characterized by normal outer hair cell function but disrupted auditory nerve function. Neurological and electrophysical evidence suggests that disrupted auditory nerve activity is due to desynchronized or reduced neural activity or both. Psychophysical measures showed that the disrupted neural activity has minimal effects on intensity-related perception, such as loudness discrimination, pitch discrimination at high frequencies, and sound localization using interaural level differences. In contrast, the disrupted neural activity significantly impairs timing related perception, such as pitch discrimination at low frequencies, temporal integration, gap detection, temporal modulation detection, backward and forward masking, signal detection in noise, binaural beats, and sound localization using interaural time differences. These perceptual consequences are the opposite of what is typically observed in cochlear-impaired subjects who have impaired intensity perception but relatively normal temporal processing after taking their impaired intensity perception into account. These differences in perceptual consequences between auditory neuropathy and cochlear damage suggest the use of different neural codes in auditory perception: a suboptimal spike count code for intensity processing, a synchronized spike code for temporal processing, and a duplex code for frequency processing. We also proposed two underlying physiological models based on desynchronized and reduced discharge in the auditory nerve to successfully account for the observed neurological and behavioral data. These methods and measures cannot differentiate between these two AN models, but future studies using electric stimulation of the auditory nerve via a cochlear implant might. These results not only show the unique

  15. Synchronization of a class of chaotic signals via robust observer design

    Energy Technology Data Exchange (ETDEWEB)

    Aguilar-Lopez, Ricardo [Departamento de Energia, Universidad Autonoma Metropolitana - Azcapotzalco, San Pablo 180, Reynosa-Tamaulipas, Azcapotzalco 02200, Mexico, D.F. (Mexico)], E-mail: raguilar@correo.azc.uam.mx; Martinez-Guerra, Rafael [Departamento de Energia, Universidad Autonoma Metropolitana - Azcapotzalco, San Pablo 180, Reynosa-Tamaulipas, Azcapotzalco 02200, Mexico, D.F. (Mexico); Departamento de Control Automatico, CINVESTAV IPN, Apartado Postal 14-740, Mexico, D.F. C.P. 07360 (Mexico)], E-mail: rguerra@ctrl.cinvestav.mx

    2008-07-15

    In this paper the signal synchronization of a class of chaotic systems based on robust observer design is tackled. The task is the synchronization of the signals generated by two Chen oscillators with different initial condition. The proposed observer is robust against model uncertainties and noisy output measurements. An alternative system representation is proposed to transform the measured disturbance onto system disturbance, which leads a more adequate observer structure. The proposed methodology contains an uncertainty estimator based on the predictive contribution to infer the unobservable uncertainties and corrective contribution to estimate the observable uncertainties; which provides robustness against noisy measurements and model uncertainties. Convergence analysis of the proposed estimation methodology is realized, analyzing the dynamic equation of the estimation error, where asymptotic convergence is shown. Numerical experiments illustrate the good performance of the proposed methodology.

  16. Synchronization of a class of chaotic signals via robust observer design

    International Nuclear Information System (INIS)

    Aguilar-Lopez, Ricardo; Martinez-Guerra, Rafael

    2008-01-01

    In this paper the signal synchronization of a class of chaotic systems based on robust observer design is tackled. The task is the synchronization of the signals generated by two Chen oscillators with different initial condition. The proposed observer is robust against model uncertainties and noisy output measurements. An alternative system representation is proposed to transform the measured disturbance onto system disturbance, which leads a more adequate observer structure. The proposed methodology contains an uncertainty estimator based on the predictive contribution to infer the unobservable uncertainties and corrective contribution to estimate the observable uncertainties; which provides robustness against noisy measurements and model uncertainties. Convergence analysis of the proposed estimation methodology is realized, analyzing the dynamic equation of the estimation error, where asymptotic convergence is shown. Numerical experiments illustrate the good performance of the proposed methodology

  17. Uncovering beat deafness: detecting rhythm disorders with synchronized finger tapping and perceptual timing tasks.

    Science.gov (United States)

    Dalla Bella, Simone; Sowiński, Jakub

    2015-03-16

    A set of behavioral tasks for assessing perceptual and sensorimotor timing abilities in the general population (i.e., non-musicians) is presented here with the goal of uncovering rhythm disorders, such as beat deafness. Beat deafness is characterized by poor performance in perceiving durations in auditory rhythmic patterns or poor synchronization of movement with auditory rhythms (e.g., with musical beats). These tasks include the synchronization of finger tapping to the beat of simple and complex auditory stimuli and the detection of rhythmic irregularities (anisochrony detection task) embedded in the same stimuli. These tests, which are easy to administer, include an assessment of both perceptual and sensorimotor timing abilities under different conditions (e.g., beat rates and types of auditory material) and are based on the same auditory stimuli, ranging from a simple metronome to a complex musical excerpt. The analysis of synchronized tapping data is performed with circular statistics, which provide reliable measures of synchronization accuracy (e.g., the difference between the timing of the taps and the timing of the pacing stimuli) and consistency. Circular statistics on tapping data are particularly well-suited for detecting individual differences in the general population. Synchronized tapping and anisochrony detection are sensitive measures for identifying profiles of rhythm disorders and have been used with success to uncover cases of poor synchronization with spared perceptual timing. This systematic assessment of perceptual and sensorimotor timing can be extended to populations of patients with brain damage, neurodegenerative diseases (e.g., Parkinson's disease), and developmental disorders (e.g., Attention Deficit Hyperactivity Disorder).

  18. Moving Stimuli Facilitate Synchronization But Not Temporal Perception.

    Science.gov (United States)

    Silva, Susana; Castro, São Luís

    2016-01-01

    Recent studies have shown that a moving visual stimulus (e.g., a bouncing ball) facilitates synchronization compared to a static stimulus (e.g., a flashing light), and that it can even be as effective as an auditory beep. We asked a group of participants to perform different tasks with four stimulus types: beeps, siren-like sounds, visual flashes (static) and bouncing balls. First, participants performed synchronization with isochronous sequences (stimulus-guided synchronization), followed by a continuation phase in which the stimulus was internally generated (imagery-guided synchronization). Then they performed a perception task, in which they judged whether the final part of a temporal sequence was compatible with the previous beat structure (stimulus-guided perception). Similar to synchronization, an imagery-guided variant was added, in which sequences contained a gap in between (imagery-guided perception). Balls outperformed flashes and matched beeps (powerful ball effect) in stimulus-guided synchronization but not in perception (stimulus- or imagery-guided). In imagery-guided synchronization, performance accuracy decreased for beeps and balls, but not for flashes and sirens. Our findings suggest that the advantages of moving visual stimuli over static ones are grounded in action rather than perception, and they support the hypothesis that the sensorimotor coupling mechanisms for auditory (beeps) and moving visual stimuli (bouncing balls) overlap.

  19. Differential sensory cortical involvement in auditory and visual sensorimotor temporal recalibration: Evidence from transcranial direct current stimulation (tDCS).

    Science.gov (United States)

    Aytemür, Ali; Almeida, Nathalia; Lee, Kwang-Hyuk

    2017-02-01

    Adaptation to delayed sensory feedback following an action produces a subjective time compression between the action and the feedback (temporal recalibration effect, TRE). TRE is important for sensory delay compensation to maintain a relationship between causally related events. It is unclear whether TRE is a sensory modality-specific phenomenon. In 3 experiments employing a sensorimotor synchronization task, we investigated this question using cathodal transcranial direct-current stimulation (tDCS). We found that cathodal tDCS over the visual cortex, and to a lesser extent over the auditory cortex, produced decreased visual TRE. However, both auditory and visual cortex tDCS did not produce any measurable effects on auditory TRE. Our study revealed different nature of TRE in auditory and visual domains. Visual-motor TRE, which is more variable than auditory TRE, is a sensory modality-specific phenomenon, modulated by the auditory cortex. The robustness of auditory-motor TRE, unaffected by tDCS, suggests the dominance of the auditory system in temporal processing, by providing a frame of reference in the realignment of sensorimotor timing signals. Copyright © 2017 Elsevier Ltd. All rights reserved.

  20. A computational model of human auditory signal processing and perception

    DEFF Research Database (Denmark)

    Jepsen, Morten Løve; Ewert, Stephan D.; Dau, Torsten

    2008-01-01

    A model of computational auditory signal-processing and perception that accounts for various aspects of simultaneous and nonsimultaneous masking in human listeners is presented. The model is based on the modulation filterbank model described by Dau et al. [J. Acoust. Soc. Am. 102, 2892 (1997...... discrimination with pure tones and broadband noise, tone-in-noise detection, spectral masking with narrow-band signals and maskers, forward masking with tone signals and tone or noise maskers, and amplitude-modulation detection with narrow- and wideband noise carriers. The model can account for most of the key...... properties of the data and is more powerful than the original model. The model might be useful as a front end in technical applications....

  1. Shuttle bit rate synchronizer. [signal to noise ratios and error analysis

    Science.gov (United States)

    Huey, D. C.; Fultz, G. L.

    1974-01-01

    A shuttle bit rate synchronizer brassboard unit was designed, fabricated, and tested, which meets or exceeds the contractual specifications. The bit rate synchronizer operates at signal-to-noise ratios (in a bit rate bandwidth) down to -5 dB while exhibiting less than 0.6 dB bit error rate degradation. The mean acquisition time was measured to be less than 2 seconds. The synchronizer is designed around a digital data transition tracking loop whose phase and data detectors are integrate-and-dump filters matched to the Manchester encoded bits specified. It meets the reliability (no adjustments or tweaking) and versatility (multiple bit rates) of the shuttle S-band communication system through an implementation which is all digital after the initial stage of analog AGC and A/D conversion.

  2. Influence of different envelope maskers on signal recognition and neuronal representation in the auditory system of a grasshopper.

    Directory of Open Access Journals (Sweden)

    Daniela Neuhofer

    Full Text Available BACKGROUND: Animals that communicate by sound face the problem that the signals arriving at the receiver often are degraded and masked by noise. Frequency filters in the receiver's auditory system may improve the signal-to-noise ratio (SNR by excluding parts of the spectrum which are not occupied by the species-specific signals. This solution, however, is hardly amenable to species that produce broad band signals or have ears with broad frequency tuning. In mammals auditory filters exist that work in the temporal domain of amplitude modulations (AM. Do insects also use this type of filtering? PRINCIPAL FINDINGS: Combining behavioural and neurophysiological experiments we investigated whether AM filters may improve the recognition of masked communication signals in grasshoppers. The AM pattern of the sound, its envelope, is crucial for signal recognition in these animals. We degraded the species-specific song by adding random fluctuations to its envelope. Six noise bands were used that differed in their overlap with the spectral content of the song envelope. If AM filters contribute to reduced masking, signal recognition should depend on the degree of overlap between the song envelope spectrum and the noise spectra. Contrary to this prediction, the resistance against signal degradation was the same for five of six masker bands. Most remarkably, the band with the strongest frequency overlap to the natural song envelope (0-100 Hz impaired acceptance of degraded signals the least. To assess the noise filter capacities of single auditory neurons, the changes of spike trains as a function of the masking level were assessed. Increasing levels of signal degradation in different frequency bands led to similar changes in the spike trains in most neurones. CONCLUSIONS: There is no indication that auditory neurones of grasshoppers are specialized to improve the SNR with respect to the pattern of amplitude modulations.

  3. Performance of different synchronization measures in real data: A case study on electroencephalographic signals

    Science.gov (United States)

    Quian Quiroga, R.; Kraskov, A.; Kreuz, T.; Grassberger, P.

    2002-04-01

    We study the synchronization between left and right hemisphere rat electroencephalographic (EEG) channels by using various synchronization measures, namely nonlinear interdependences, phase synchronizations, mutual information, cross correlation, and the coherence function. In passing we show a close relation between two recently proposed phase synchronization measures and we extend the definition of one of them. In three typical examples we observe that except mutual information, all these measures give a useful quantification that is hard to be guessed beforehand from the raw data. Despite their differences, results are qualitatively the same. Therefore, we claim that the applied measures are valuable for the study of synchronization in real data. Moreover, in the particular case of EEG signals their use as complementary variables could be of clinical relevance.

  4. Instantaneous and Frequency-Warped Signal Processing Techniques for Auditory Source Separation.

    Science.gov (United States)

    Wang, Avery Li-Chun

    This thesis summarizes several contributions to the areas of signal processing and auditory source separation. The philosophy of Frequency-Warped Signal Processing is introduced as a means for separating the AM and FM contributions to the bandwidth of a complex-valued, frequency-varying sinusoid p (n), transforming it into a signal with slowly-varying parameters. This transformation facilitates the removal of p (n) from an additive mixture while minimizing the amount of damage done to other signal components. The average winding rate of a complex-valued phasor is explored as an estimate of the instantaneous frequency. Theorems are provided showing the robustness of this measure. To implement frequency tracking, a Frequency-Locked Loop algorithm is introduced which uses the complex winding error to update its frequency estimate. The input signal is dynamically demodulated and filtered to extract the envelope. This envelope may then be remodulated to reconstruct the target partial, which may be subtracted from the original signal mixture to yield a new, quickly-adapting form of notch filtering. Enhancements to the basic tracker are made which, under certain conditions, attain the Cramer -Rao bound for the instantaneous frequency estimate. To improve tracking, the novel idea of Harmonic -Locked Loop tracking, using N harmonically constrained trackers, is introduced for tracking signals, such as voices and certain musical instruments. The estimated fundamental frequency is computed from a maximum-likelihood weighting of the N tracking estimates, making it highly robust. The result is that harmonic signals, such as voices, can be isolated from complex mixtures in the presence of other spectrally overlapping signals. Additionally, since phase information is preserved, the resynthesized harmonic signals may be removed from the original mixtures with relatively little damage to the residual signal. Finally, a new methodology is given for designing linear-phase FIR filters

  5. Developmental changes of BOLD signal correlations with global human EEG power and synchronization during working memory.

    Directory of Open Access Journals (Sweden)

    Lars Michels

    Full Text Available In humans, theta band (5-7 Hz power typically increases when performing cognitively demanding working memory (WM tasks, and simultaneous EEG-fMRI recordings have revealed an inverse relationship between theta power and the BOLD (blood oxygen level dependent signal in the default mode network during WM. However, synchronization also plays a fundamental role in cognitive processing, and the level of theta and higher frequency band synchronization is modulated during WM. Yet, little is known about the link between BOLD, EEG power, and EEG synchronization during WM, and how these measures develop with human brain maturation or relate to behavioral changes. We examined EEG-BOLD signal correlations from 18 young adults and 15 school-aged children for age-dependent effects during a load-modulated Sternberg WM task. Frontal load (in-dependent EEG theta power was significantly enhanced in children compared to adults, while adults showed stronger fMRI load effects. Children demonstrated a stronger negative correlation between global theta power and the BOLD signal in the default mode network relative to adults. Therefore, we conclude that theta power mediates the suppression of a task-irrelevant network. We further conclude that children suppress this network even more than adults, probably from an increased level of task-preparedness to compensate for not fully mature cognitive functions, reflected in lower response accuracy and increased reaction time. In contrast to power, correlations between instantaneous theta global field synchronization and the BOLD signal were exclusively positive in both age groups but only significant in adults in the frontal-parietal and posterior cingulate cortices. Furthermore, theta synchronization was weaker in children and was--in contrast to EEG power--positively correlated with response accuracy in both age groups. In summary we conclude that theta EEG-BOLD signal correlations differ between spectral power and

  6. A signal normalization technique for illumination-based synchronization of 1,000-fps real-time vision sensors in dynamic scenes.

    Science.gov (United States)

    Hou, Lei; Kagami, Shingo; Hashimoto, Koichi

    2010-01-01

    To acquire images of dynamic scenes from multiple points of view simultaneously, the acquisition time of vision sensors should be synchronized. In this paper, an illumination-based synchronization derived from the phase-locked loop (PLL) mechanism based on the signal normalization method is proposed and evaluated. To eliminate the system dependency due to the amplitude fluctuation of the reference illumination, which may be caused by the moving objects or relative positional distance change between the light source and the observed objects, the fluctuant amplitude of the reference signal is normalized framely by the estimated maximum amplitude between the reference signal and its quadrature counterpart to generate a stable synchronization in highly dynamic scenes. Both simulated results and real world experimental results demonstrated successful synchronization result that 1,000-Hz frame rate vision sensors can be successfully synchronized to a LED illumination or its reflected light with satisfactory stability and only 28-μs jitters.

  7. A Signal Normalization Technique for Illumination-Based Synchronization of 1,000-fps Real-Time Vision Sensors in Dynamic Scenes

    Directory of Open Access Journals (Sweden)

    Koichi Hashimoto

    2010-09-01

    Full Text Available To acquire images of dynamic scenes from multiple points of view simultaneously, the acquisition time of vision sensors should be synchronized. In this paper, an illumination-based synchronization derived from the phase-locked loop (PLL mechanism based on the signal normalization method is proposed and evaluated. To eliminate the system dependency due to the amplitude fluctuation of the reference illumination, which may be caused by the moving objects or relative positional distance change between the light source and the observed objects, the fluctuant amplitude of the reference signal is normalized framely by the estimated maximum amplitude between the reference signal and its quadrature counterpart to generate a stable synchronization in highly dynamic scenes. Both simulated results and real world experimental results demonstrated successful synchronization result that 1,000-Hz frame rate vision sensors can be successfully synchronized to a LED illumination or its reflected light with satisfactory stability and only 28-μs jitters.

  8. Temporal factors affecting somatosensory-auditory interactions in speech processing

    Directory of Open Access Journals (Sweden)

    Takayuki eIto

    2014-11-01

    Full Text Available Speech perception is known to rely on both auditory and visual information. However, sound specific somatosensory input has been shown also to influence speech perceptual processing (Ito et al., 2009. In the present study we addressed further the relationship between somatosensory information and speech perceptual processing by addressing the hypothesis that the temporal relationship between orofacial movement and sound processing contributes to somatosensory-auditory interaction in speech perception. We examined the changes in event-related potentials in response to multisensory synchronous (simultaneous and asynchronous (90 ms lag and lead somatosensory and auditory stimulation compared to individual unisensory auditory and somatosensory stimulation alone. We used a robotic device to apply facial skin somatosensory deformations that were similar in timing and duration to those experienced in speech production. Following synchronous multisensory stimulation the amplitude of the event-related potential was reliably different from the two unisensory potentials. More importantly, the magnitude of the event-related potential difference varied as a function of the relative timing of the somatosensory-auditory stimulation. Event-related activity change due to stimulus timing was seen between 160-220 ms following somatosensory onset, mostly around the parietal area. The results demonstrate a dynamic modulation of somatosensory-auditory convergence and suggest the contribution of somatosensory information for speech processing process is dependent on the specific temporal order of sensory inputs in speech production.

  9. A Data Hiding Technique to Synchronously Embed Physiological Signals in H.264/AVC Encoded Video for Medicine Healthcare.

    Science.gov (United States)

    Peña, Raul; Ávila, Alfonso; Muñoz, David; Lavariega, Juan

    2015-01-01

    The recognition of clinical manifestations in both video images and physiological-signal waveforms is an important aid to improve the safety and effectiveness in medical care. Physicians can rely on video-waveform (VW) observations to recognize difficult-to-spot signs and symptoms. The VW observations can also reduce the number of false positive incidents and expand the recognition coverage to abnormal health conditions. The synchronization between the video images and the physiological-signal waveforms is fundamental for the successful recognition of the clinical manifestations. The use of conventional equipment to synchronously acquire and display the video-waveform information involves complex tasks such as the video capture/compression, the acquisition/compression of each physiological signal, and the video-waveform synchronization based on timestamps. This paper introduces a data hiding technique capable of both enabling embedding channels and synchronously hiding samples of physiological signals into encoded video sequences. Our data hiding technique offers large data capacity and simplifies the complexity of the video-waveform acquisition and reproduction. The experimental results revealed successful embedding and full restoration of signal's samples. Our results also demonstrated a small distortion in the video objective quality, a small increment in bit-rate, and embedded cost savings of -2.6196% for high and medium motion video sequences.

  10. Synchrony of auditory brain responses predicts behavioral ability to keep still in children with autism spectrum disorder

    Directory of Open Access Journals (Sweden)

    Yuko Yoshimura

    2016-01-01

    Full Text Available The auditory-evoked P1m, recorded by magnetoencephalography, reflects a central auditory processing ability in human children. One recent study revealed that asynchrony of P1m between the right and left hemispheres reflected a central auditory processing disorder (i.e., attention deficit hyperactivity disorder, ADHD in children. However, to date, the relationship between auditory P1m right-left hemispheric synchronization and the comorbidity of hyperactivity in children with autism spectrum disorder (ASD is unknown. In this study, based on a previous report of an asynchrony of P1m in children with ADHD, to clarify whether the P1m right-left hemispheric synchronization is related to the symptom of hyperactivity in children with ASD, we investigated the relationship between voice-evoked P1m right-left hemispheric synchronization and hyperactivity in children with ASD. In addition to synchronization, we investigated the right-left hemispheric lateralization. Our findings failed to demonstrate significant differences in these values between ASD children with and without the symptom of hyperactivity, which was evaluated using the Autism Diagnostic Observational Schedule, Generic (ADOS-G subscale. However, there was a significant correlation between the degrees of hemispheric synchronization and the ability to keep still during 12-minute MEG recording periods. Our results also suggested that asynchrony in the bilateral brain auditory processing system is associated with ADHD-like symptoms in children with ASD.

  11. Persistent fluctuations in stride intervals under fractal auditory stimulation.

    Science.gov (United States)

    Marmelat, Vivien; Torre, Kjerstin; Beek, Peter J; Daffertshofer, Andreas

    2014-01-01

    Stride sequences of healthy gait are characterized by persistent long-range correlations, which become anti-persistent in the presence of an isochronous metronome. The latter phenomenon is of particular interest because auditory cueing is generally considered to reduce stride variability and may hence be beneficial for stabilizing gait. Complex systems tend to match their correlation structure when synchronizing. In gait training, can one capitalize on this tendency by using a fractal metronome rather than an isochronous one? We examined whether auditory cues with fractal variations in inter-beat intervals yield similar fractal inter-stride interval variability as isochronous auditory cueing in two complementary experiments. In Experiment 1, participants walked on a treadmill while being paced by either an isochronous or a fractal metronome with different variation strengths between beats in order to test whether participants managed to synchronize with a fractal metronome and to determine the necessary amount of variability for participants to switch from anti-persistent to persistent inter-stride intervals. Participants did synchronize with the metronome despite its fractal randomness. The corresponding coefficient of variation of inter-beat intervals was fixed in Experiment 2, in which participants walked on a treadmill while being paced by non-isochronous metronomes with different scaling exponents. As expected, inter-stride intervals showed persistent correlations similar to self-paced walking only when cueing contained persistent correlations. Our results open up a new window to optimize rhythmic auditory cueing for gait stabilization by integrating fractal fluctuations in the inter-beat intervals.

  12. Evoked responses of the superior olive to amplitude-modulated signals.

    Science.gov (United States)

    Andreeva, N G; Lang, T T

    1977-01-01

    Evoked potentials of some auditory centers of Rhinolophidae bats to amplitude-modulated signals were studied. A synchronization response was found in the cochlear nuclei (with respect to the fast component of the response) and in the superior olivary complex (with respect to both fast and slow components of the response) within the range of frequency modulation from 50 to 2000 Hz. In the inferior colliculus a synchronized response was recorded at modulation frequencies below 150 Hz, but in the medial geniculate bodies no such response was found. Evoked responses of the superior olivary complex were investigated in detail. The lowest frequencies of synchronization were recorded within the carrier frequency range of 15-30 and 80-86 kHz. The amplitude of the synchronized response is a function of the frequency and coefficient of modulation and also of the angle of stimulus presentation.

  13. An Overview of a Class of Clock Synchronization Algorithms for Wireless Sensor Networks: A Statistical Signal Processing Perspective

    Directory of Open Access Journals (Sweden)

    Xu Wang

    2015-08-01

    Full Text Available Recently, wireless sensor networks (WSNs have drawn great interest due to their outstanding monitoring and management potential in medical, environmental and industrial applications. Most of the applications that employ WSNs demand all of the sensor nodes to run on a common time scale, a requirement that highlights the importance of clock synchronization. The clock synchronization problem in WSNs is inherently related to parameter estimation. The accuracy of clock synchronization algorithms depends essentially on the statistical properties of the parameter estimation algorithms. Recently, studies dedicated to the estimation of synchronization parameters, such as clock offset and skew, have begun to emerge in the literature. The aim of this article is to provide an overview of the state-of-the-art clock synchronization algorithms for WSNs from a statistical signal processing point of view. This article focuses on describing the key features of the class of clock synchronization algorithms that exploit the traditional two-way message (signal exchange mechanism. Upon introducing the two-way message exchange mechanism, the main clock offset estimation algorithms for pairwise synchronization of sensor nodes are first reviewed, and their performance is compared. The class of fully-distributed clock offset estimation algorithms for network-wide synchronization is then surveyed. The paper concludes with a list of open research problems pertaining to clock synchronization of WSNs.

  14. Diagnostics of synchronous motor based on analysis of acoustic signals with application of MFCC and Nearest Mean classifier

    OpenAIRE

    Adam Głowacz; Witold Głowacz; Andrzej Głowacz

    2010-01-01

    The paper presents method of diagnostics of imminent failure conditions of synchronous motor. This method is based on a study ofacoustic signals generated by synchronous motor. Sound recognition system is based on algorithms of data processing, such as MFCC andNearest Mean classifier with cosine distance. Software to recognize the sounds of synchronous motor was implemented. The studies werecarried out for four imminent failure conditions of synchronous motor. The results confirm that the sys...

  15. Absence of both auditory evoked potentials and auditory percepts dependent on timing cues.

    Science.gov (United States)

    Starr, A; McPherson, D; Patterson, J; Don, M; Luxford, W; Shannon, R; Sininger, Y; Tonakawa, L; Waring, M

    1991-06-01

    An 11-yr-old girl had an absence of sensory components of auditory evoked potentials (brainstem, middle and long-latency) to click and tone burst stimuli that she could clearly hear. Psychoacoustic tests revealed a marked impairment of those auditory perceptions dependent on temporal cues, that is, lateralization of binaural clicks, change of binaural masked threshold with changes in signal phase, binaural beats, detection of paired monaural clicks, monaural detection of a silent gap in a sound, and monaural threshold elevation for short duration tones. In contrast, auditory functions reflecting intensity or frequency discriminations (difference limens) were only minimally impaired. Pure tone audiometry showed a moderate (50 dB) bilateral hearing loss with a disproportionate severe loss of word intelligibility. Those auditory evoked potentials that were preserved included (1) cochlear microphonics reflecting hair cell activity; (2) cortical sustained potentials reflecting processing of slowly changing signals; and (3) long-latency cognitive components (P300, processing negativity) reflecting endogenous auditory cognitive processes. Both the evoked potential and perceptual deficits are attributed to changes in temporal encoding of acoustic signals perhaps occurring at the synapse between hair cell and eighth nerve dendrites. The results from this patient are discussed in relation to previously published cases with absent auditory evoked potentials and preserved hearing.

  16. Persistent fluctuations in stride intervals under fractal auditory stimulation.

    Directory of Open Access Journals (Sweden)

    Vivien Marmelat

    Full Text Available Stride sequences of healthy gait are characterized by persistent long-range correlations, which become anti-persistent in the presence of an isochronous metronome. The latter phenomenon is of particular interest because auditory cueing is generally considered to reduce stride variability and may hence be beneficial for stabilizing gait. Complex systems tend to match their correlation structure when synchronizing. In gait training, can one capitalize on this tendency by using a fractal metronome rather than an isochronous one? We examined whether auditory cues with fractal variations in inter-beat intervals yield similar fractal inter-stride interval variability as isochronous auditory cueing in two complementary experiments. In Experiment 1, participants walked on a treadmill while being paced by either an isochronous or a fractal metronome with different variation strengths between beats in order to test whether participants managed to synchronize with a fractal metronome and to determine the necessary amount of variability for participants to switch from anti-persistent to persistent inter-stride intervals. Participants did synchronize with the metronome despite its fractal randomness. The corresponding coefficient of variation of inter-beat intervals was fixed in Experiment 2, in which participants walked on a treadmill while being paced by non-isochronous metronomes with different scaling exponents. As expected, inter-stride intervals showed persistent correlations similar to self-paced walking only when cueing contained persistent correlations. Our results open up a new window to optimize rhythmic auditory cueing for gait stabilization by integrating fractal fluctuations in the inter-beat intervals.

  17. A Modified Differential Coherent Bit Synchronization Algorithm for BeiDou Weak Signals with Large Frequency Deviation.

    Science.gov (United States)

    Han, Zhifeng; Liu, Jianye; Li, Rongbing; Zeng, Qinghua; Wang, Yi

    2017-07-04

    BeiDou system navigation messages are modulated with a secondary NH (Neumann-Hoffman) code of 1 kbps, where frequent bit transitions limit the coherent integration time to 1 millisecond. Therefore, a bit synchronization algorithm is necessary to obtain bit edges and NH code phases. In order to realize bit synchronization for BeiDou weak signals with large frequency deviation, a bit synchronization algorithm based on differential coherent and maximum likelihood is proposed. Firstly, a differential coherent approach is used to remove the effect of frequency deviation, and the differential delay time is set to be a multiple of bit cycle to remove the influence of NH code. Secondly, the maximum likelihood function detection is used to improve the detection probability of weak signals. Finally, Monte Carlo simulations are conducted to analyze the detection performance of the proposed algorithm compared with a traditional algorithm under the CN0s of 20~40 dB-Hz and different frequency deviations. The results show that the proposed algorithm outperforms the traditional method with a frequency deviation of 50 Hz. This algorithm can remove the effect of BeiDou NH code effectively and weaken the influence of frequency deviation. To confirm the feasibility of the proposed algorithm, real data tests are conducted. The proposed algorithm is suitable for BeiDou weak signal bit synchronization with large frequency deviation.

  18. Compatibility of Motion Facilitates Visuomotor Synchronization

    Science.gov (United States)

    Hove, Michael J.; Spivey, Michael J.; Krumhansl, Carol L.

    2010-01-01

    Prior research indicates that synchronized tapping performance is very poor with flashing visual stimuli compared with auditory stimuli. Three finger-tapping experiments compared flashing visual metronomes with visual metronomes containing a spatial component, either compatible, incompatible, or orthogonal to the tapping action. In Experiment 1,…

  19. Listenmee and Listenmee smartphone application: synchronizing walking to rhythmic auditory cues to improve gait in Parkinson's disease.

    Science.gov (United States)

    Lopez, William Omar Contreras; Higuera, Carlos Andres Escalante; Fonoff, Erich Talamoni; Souza, Carolina de Oliveira; Albicker, Ulrich; Martinez, Jairo Alberto Espinoza

    2014-10-01

    Evidence supports the use of rhythmic external auditory signals to improve gait in PD patients (Arias & Cudeiro, 2008; Kenyon & Thaut, 2000; McIntosh, Rice & Thaut, 1994; McIntosh et al., 1997; Morris, Iansek, & Matyas, 1994; Thaut, McIntosh, & Rice, 1997; Suteerawattananon, Morris, Etnyre, Jankovic, & Protas , 2004; Willems, Nieuwboer, Chavert, & Desloovere, 2006). However, few prototypes are available for daily use, and to our knowledge, none utilize a smartphone application allowing individualized sounds and cadence. Therefore, we analyzed the effects on gait of Listenmee®, an intelligent glasses system with a portable auditory device, and present its smartphone application, the Listenmee app®, offering over 100 different sounds and an adjustable metronome to individualize the cueing rate as well as its smartwatch with accelerometer to detect magnitude and direction of the proper acceleration, track calorie count, sleep patterns, steps count and daily distances. The present study included patients with idiopathic PD presented gait disturbances including freezing. Auditory rhythmic cues were delivered through Listenmee®. Performance was analyzed in a motion and gait analysis laboratory. The results revealed significant improvements in gait performance over three major dependent variables: walking speed in 38.1%, cadence in 28.1% and stride length in 44.5%. Our findings suggest that auditory cueing through Listenmee® may significantly enhance gait performance. Further studies are needed to elucidate the potential role and maximize the benefits of these portable devices. Copyright © 2014 Elsevier B.V. All rights reserved.

  20. Perceptual Plasticity for Auditory Object Recognition

    Science.gov (United States)

    Heald, Shannon L. M.; Van Hedger, Stephen C.; Nusbaum, Howard C.

    2017-01-01

    In our auditory environment, we rarely experience the exact acoustic waveform twice. This is especially true for communicative signals that have meaning for listeners. In speech and music, the acoustic signal changes as a function of the talker (or instrument), speaking (or playing) rate, and room acoustics, to name a few factors. Yet, despite this acoustic variability, we are able to recognize a sentence or melody as the same across various kinds of acoustic inputs and determine meaning based on listening goals, expectations, context, and experience. The recognition process relates acoustic signals to prior experience despite variability in signal-relevant and signal-irrelevant acoustic properties, some of which could be considered as “noise” in service of a recognition goal. However, some acoustic variability, if systematic, is lawful and can be exploited by listeners to aid in recognition. Perceivable changes in systematic variability can herald a need for listeners to reorganize perception and reorient their attention to more immediately signal-relevant cues. This view is not incorporated currently in many extant theories of auditory perception, which traditionally reduce psychological or neural representations of perceptual objects and the processes that act on them to static entities. While this reduction is likely done for the sake of empirical tractability, such a reduction may seriously distort the perceptual process to be modeled. We argue that perceptual representations, as well as the processes underlying perception, are dynamically determined by an interaction between the uncertainty of the auditory signal and constraints of context. This suggests that the process of auditory recognition is highly context-dependent in that the identity of a given auditory object may be intrinsically tied to its preceding context. To argue for the flexible neural and psychological updating of sound-to-meaning mappings across speech and music, we draw upon examples

  1. Common-signal-induced synchronization in photonic integrated circuits and its application to secure key distribution.

    Science.gov (United States)

    Sasaki, Takuma; Kakesu, Izumi; Mitsui, Yusuke; Rontani, Damien; Uchida, Atsushi; Sunada, Satoshi; Yoshimura, Kazuyuki; Inubushi, Masanobu

    2017-10-16

    We experimentally achieve common-signal-induced synchronization in two photonic integrated circuits with short external cavities driven by a constant-amplitude random-phase light. The degree of synchronization can be controlled by changing the optical feedback phase of the two photonic integrated circuits. The change in the optical feedback phase leads to a significant redistribution of the spectral energy of optical and RF spectra, which is a unique characteristic of PICs with the short external cavity. The matching of the RF and optical spectra is necessary to achieve synchronization between the two PICs, and stable synchronization can be obtained over an hour in the presence of optical feedback. We succeed in generating information-theoretic secure keys and achieving the final key generation rate of 184 kb/s using the PICs.

  2. Sensitivity to external signals and synchronization properties of a non-isochronous auto-oscillator with delayed feedback

    Science.gov (United States)

    Tiberkevich, Vasil S.; Khymyn, Roman S.; Tang, Hong X.; Slavin, Andrei N.

    2014-01-01

    For auto-oscillators of different nature (e.g. active cells in a human heart under the action of a pacemaker, neurons in brain, spin-torque nano-oscillators, micro and nano-mechanical oscillators, or generating Josephson junctions) a critically important property is their ability to synchronize with each other. The synchronization properties of an auto oscillator are directly related to its sensitivity to external signals. Here we demonstrate that a non-isochronous (having generation frequency dependent on the amplitude) auto-oscillator with delayed feedback can have an extremely high sensitivity to external signals and unusually large width of the phase-locking band near the boundary of the stable auto-oscillation regime. This property could be used for the development of synchronized arrays of non-isochronous auto-oscillators in physics and engineering, and, for instance, might bring a better fundamental understanding of ways to control a heart arrythmia in medicine.

  3. Adaptive synchronization of fractional Lorenz systems using a reduced number of control signals and parameters

    International Nuclear Information System (INIS)

    Aguila-Camacho, Norelys; Duarte-Mermoud, Manuel A.; Delgado-Aguilera, Efredy

    2016-01-01

    This paper analyzes the synchronization of two fractional Lorenz systems in two cases: the first one considering fractional Lorenz systems with unknown parameters, and the second one considering known upper bounds on some of the fractional Lorenz systems parameters. The proposed control strategies use a reduced number of control signals and control parameters, employing mild assumptions. The stability of the synchronization errors is analytically demonstrated in all cases, and the convergence to zero of the synchronization errors is analytically proved in the case when the upper bounds on some system parameters are assumed to be known. Simulation studies are presented, which allows verifying the effectiveness of the proposed control strategies.

  4. Encoding of Sucrose's Palatability in the Nucleus Accumbens Shell and Its Modulation by Exteroceptive Auditory Cues

    Directory of Open Access Journals (Sweden)

    Miguel Villavicencio

    2018-05-01

    Full Text Available Although the palatability of sucrose is the primary reason for why it is over consumed, it is not well understood how it is encoded in the nucleus accumbens shell (NAcSh, a brain region involved in reward, feeding, and sensory/motor transformations. Similarly, untouched are issues regarding how an external auditory stimulus affects sucrose palatability and, in the NAcSh, the neuronal correlates of this behavior. To address these questions in behaving rats, we investigated how food-related auditory cues modulate sucrose's palatability. The goals are to determine whether NAcSh neuronal responses would track sucrose's palatability (as measured by the increase in hedonically positive oromotor responses lick rate, sucrose concentration, and how it processes auditory information. Using brief-access tests, we found that sucrose's palatability was enhanced by exteroceptive auditory cues that signal the start and the end of a reward epoch. With only the start cue the rejection of water was accelerated, and the sucrose/water ratio was enhanced, indicating greater palatability. However, the start cue also fragmented licking patterns and decreased caloric intake. In the presence of both start and stop cues, the animals fed continuously and increased their caloric intake. Analysis of the licking microstructure confirmed that auditory cues (either signaling the start alone or start/stop enhanced sucrose's oromotor-palatability responses. Recordings of extracellular single-unit activity identified several distinct populations of NAcSh responses that tracked either the sucrose palatability responses or the sucrose concentrations by increasing or decreasing their activity. Another neural population fired synchronously with licking and exhibited an enhancement in their coherence with increasing sucrose concentrations. The population of NAcSh's Palatability-related and Lick-Inactive neurons were the most important for decoding sucrose's palatability. Only the Lick

  5. Early musical training is linked to gray matter structure in the ventral premotor cortex and auditory-motor rhythm synchronization performance.

    Science.gov (United States)

    Bailey, Jennifer Anne; Zatorre, Robert J; Penhune, Virginia B

    2014-04-01

    Evidence in animals and humans indicates that there are sensitive periods during development, times when experience or stimulation has a greater influence on behavior and brain structure. Sensitive periods are the result of an interaction between maturational processes and experience-dependent plasticity mechanisms. Previous work from our laboratory has shown that adult musicians who begin training before the age of 7 show enhancements in behavior and white matter structure compared with those who begin later. Plastic changes in white matter and gray matter are hypothesized to co-occur; therefore, the current study investigated possible differences in gray matter structure between early-trained (ET; 7) musicians, matched for years of experience. Gray matter structure was assessed using voxel-wise analysis techniques (optimized voxel-based morphometry, traditional voxel-based morphometry, and deformation-based morphometry) and surface-based measures (cortical thickness, surface area and mean curvature). Deformation-based morphometry analyses identified group differences between ET and LT musicians in right ventral premotor cortex (vPMC), which correlated with performance on an auditory motor synchronization task and with age of onset of musical training. In addition, cortical surface area in vPMC was greater for ET musicians. These results are consistent with evidence that premotor cortex shows greatest maturational change between the ages of 6-9 years and that this region is important for integrating auditory and motor information. We propose that the auditory and motor interactions required by musical practice drive plasticity in vPMC and that this plasticity is greatest when maturation is near its peak.

  6. Data on the effect of conductive hearing loss on auditory and visual cortex activity revealed by intrinsic signal imaging.

    Science.gov (United States)

    Teichert, Manuel; Bolz, Jürgen

    2017-10-01

    This data article provides additional data related to the research article entitled "Simultaneous intrinsic signal imaging of auditory and visual cortex reveals profound effects of acute hearing loss on visual processing" (Teichert and Bolz, 2017) [1]. The primary auditory and visual cortex (A1 and V1) of adult male C57BL/6J mice (P120-P240) were mapped simultaneously using intrinsic signal imaging (Kalatsky and Stryker, 2003) [2]. A1 and V1 activity evoked by combined auditory and visual stimulation were measured before and after conductive hearing loss (CHL) induced by bilateral malleus removal. We provide data showing that A1 responsiveness evoked by sounds of different sound pressure levels (SPL) decreased after CHL whereas visually evoked V1 activity increased after this intervention. In addition, we also provide imaging data on percentage of V1 activity increases after CHL compared to pre-CHL.

  7. Coupling between Theta Oscillations and Cognitive Control Network during Cross-Modal Visual and Auditory Attention: Supramodal vs Modality-Specific Mechanisms.

    Science.gov (United States)

    Wang, Wuyi; Viswanathan, Shivakumar; Lee, Taraz; Grafton, Scott T

    2016-01-01

    Cortical theta band oscillations (4-8 Hz) in EEG signals have been shown to be important for a variety of different cognitive control operations in visual attention paradigms. However the synchronization source of these signals as defined by fMRI BOLD activity and the extent to which theta oscillations play a role in multimodal attention remains unknown. Here we investigated the extent to which cross-modal visual and auditory attention impacts theta oscillations. Using a simultaneous EEG-fMRI paradigm, healthy human participants performed an attentional vigilance task with six cross-modal conditions using naturalistic stimuli. To assess supramodal mechanisms, modulation of theta oscillation amplitude for attention to either visual or auditory stimuli was correlated with BOLD activity by conjunction analysis. Negative correlation was localized to cortical regions associated with the default mode network and positively with ventral premotor areas. Modality-associated attention to visual stimuli was marked by a positive correlation of theta and BOLD activity in fronto-parietal area that was not observed in the auditory condition. A positive correlation of theta and BOLD activity was observed in auditory cortex, while a negative correlation of theta and BOLD activity was observed in visual cortex during auditory attention. The data support a supramodal interaction of theta activity with of DMN function, and modality-associated processes within fronto-parietal networks related to top-down theta related cognitive control in cross-modal visual attention. On the other hand, in sensory cortices there are opposing effects of theta activity during cross-modal auditory attention.

  8. Integration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding.

    Science.gov (United States)

    Atilgan, Huriye; Town, Stephen M; Wood, Katherine C; Jones, Gareth P; Maddox, Ross K; Lee, Adrian K C; Bizley, Jennifer K

    2018-02-07

    How and where in the brain audio-visual signals are bound to create multimodal objects remains unknown. One hypothesis is that temporal coherence between dynamic multisensory signals provides a mechanism for binding stimulus features across sensory modalities. Here, we report that when the luminance of a visual stimulus is temporally coherent with the amplitude fluctuations of one sound in a mixture, the representation of that sound is enhanced in auditory cortex. Critically, this enhancement extends to include both binding and non-binding features of the sound. We demonstrate that visual information conveyed from visual cortex via the phase of the local field potential is combined with auditory information within auditory cortex. These data provide evidence that early cross-sensory binding provides a bottom-up mechanism for the formation of cross-sensory objects and that one role for multisensory binding in auditory cortex is to support auditory scene analysis. Copyright © 2018 The Author(s). Published by Elsevier Inc. All rights reserved.

  9. Representation of acoustic signals in the eighth nerve of the Tokay gecko: I. Pure tones.

    Science.gov (United States)

    Sams-Dodd, F; Capranica, R R

    1994-06-01

    A systematic study of the encoding properties of 146 auditory nerve fibers in the Tokay gecko (Gekko gecko, L) was conducted with respect to pure tones and two-tone rate suppression. Our aim was a comprehensive understanding of the peripheral encoding of simple tonal stimuli and their representation by temporal synchronization and spike rate codes as a prelude to subsequent studies of more complex signals. Auditory nerve fibers in the Tokay gecko have asymmetrical, V-shaped excitatory tuning curves with best excitatory frequencies that range from 200-5100 Hz and thresholds between 4-35 dB SPL. A low-frequency excitatory 'tail' extends far into the low-frequency range and two-tone suppression is present only on the high frequency side of the tuning curve. The response properties to pure tones at different loci within a tuning curve can differ greatly, due to evident interactions between the representations of temporal, spectral and intensity stimulus features. For frequencies below 1250 Hz, pure tones are encoded by both temporal synchronization and spike rate codes, whereas above this frequency a fiber's ability to encode the waveform periodicity is lost and only a rate code predominates. These complimentary representations within a tuning curve raise fundamental issues which need to be addressed in interpreting how more complex, bioacoustic communication signals are represented in the peripheral and central auditory system. And since auditory nerve fibers in the Tokay gecko exhibit tonal sensitivity, selective frequency tuning, and iso-intensity and iso-frequency contours that seem comparable to similar measures in birds and mammals, these issues likely apply to most higher vertebrates in general. The simpler wiring diagram of the reptilian auditory system, coupled with the Tokay gecko's remarkable vocalizations, make this animal a good evolutionary model in which to experimentally explore the encoding of more complex sounds of communicative significance.

  10. Predictive and tempo-flexible synchronization to a visual metronome in monkeys.

    Science.gov (United States)

    Takeya, Ryuji; Kameda, Masashi; Patel, Aniruddh D; Tanaka, Masaki

    2017-07-21

    Predictive and tempo-flexible synchronization to an auditory beat is a fundamental component of human music. To date, only certain vocal learning species show this behaviour spontaneously. Prior research training macaques (vocal non-learners) to tap to an auditory or visual metronome found their movements to be largely reactive, not predictive. Does this reflect the lack of capacity for predictive synchronization in monkeys, or lack of motivation to exhibit this behaviour? To discriminate these possibilities, we trained monkeys to make synchronized eye movements to a visual metronome. We found that monkeys could generate predictive saccades synchronized to periodic visual stimuli when an immediate reward was given for every predictive movement. This behaviour generalized to novel tempi, and the monkeys could maintain the tempo internally. Furthermore, monkeys could flexibly switch from predictive to reactive saccades when a reward was given for each reactive response. In contrast, when humans were asked to make a sequence of reactive saccades to a visual metronome, they often unintentionally generated predictive movements. These results suggest that even vocal non-learners may have the capacity for predictive and tempo-flexible synchronization to a beat, but that only certain vocal learning species are intrinsically motivated to do it.

  11. Systems and methods for self-synchronized digital sampling

    Science.gov (United States)

    Samson, Jr., John R. (Inventor)

    2008-01-01

    Systems and methods for self-synchronized data sampling are provided. In one embodiment, a system for capturing synchronous data samples is provided. The system includes an analog to digital converter adapted to capture signals from one or more sensors and convert the signals into a stream of digital data samples at a sampling frequency determined by a sampling control signal; and a synchronizer coupled to the analog to digital converter and adapted to receive a rotational frequency signal from a rotating machine, wherein the synchronizer is further adapted to generate the sampling control signal, and wherein the sampling control signal is based on the rotational frequency signal.

  12. Synchronization controller design of two coupling permanent magnet synchronous motors system with nonlinear constraints.

    Science.gov (United States)

    Deng, Zhenhua; Shang, Jing; Nian, Xiaohong

    2015-11-01

    In this paper, two coupling permanent magnet synchronous motors system with nonlinear constraints is studied. First of all, the mathematical model of the system is established according to the engineering practices, in which the dynamic model of motor and the nonlinear coupling effect between two motors are considered. In order to keep the two motors synchronization, a synchronization controller based on load observer is designed via cross-coupling idea and interval matrix. Moreover, speed, position and current signals of two motor all are taken as self-feedback signal as well as cross-feedback signal in the proposed controller, which is conducive to improving the dynamical performance and the synchronization performance of the system. The proposed control strategy is verified by simulation via Matlab/Simulink program. The simulation results show that the proposed control method has a better control performance, especially synchronization performance, than that of the conventional PI controller. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.

  13. Neural Entrainment and Sensorimotor Synchronization to the Beat in Children with Developmental Dyslexia: An EEG Study

    Directory of Open Access Journals (Sweden)

    Lincoln J. Colling

    2017-07-01

    Full Text Available Tapping in time to a metronome beat (hereafter beat synchronization shows considerable variability in child populations, and individual differences in beat synchronization are reliably related to reading development. Children with developmental dyslexia show impairments in beat synchronization. These impairments may reflect deficiencies in auditory perception of the beat which in turn affect auditory-motor mapping, or may reflect an independent motor deficit. Here, we used a new methodology in EEG based on measuring beat-related steady-state evoked potentials (SS-EPs, Nozaradan et al., 2015 in an attempt to disentangle neural sensory and motor contributions to behavioral beat synchronization in children with dyslexia. Children tapped with both their left and right hands to every second beat of a metronome pulse delivered at 2.4 Hz, or listened passively to the beat. Analyses of preferred phase in EEG showed that the children with dyslexia had a significantly different preferred phase compared to control children in all conditions. Regarding SS-EPs, the groups differed significantly for the passive Auditory listening condition at 2.4 Hz, and showed a trend toward a difference in the Right hand tapping condition at 3.6 Hz (sensorimotor integration measure. The data suggest that neural rhythmic entrainment is atypical in children with dyslexia for both an auditory beat and during sensorimotor coupling (tapping. The data are relevant to a growing literature suggesting that rhythm-based interventions may help language processing in children with developmental disorders of language learning.

  14. Synchronized Progression of Prestin Expression and Auditory Brainstem Response during Postnatal Development in Rats

    Directory of Open Access Journals (Sweden)

    Jianfeng Hang

    2016-01-01

    Full Text Available Prestin is the motor protein expressed in the cochlear outer hair cells (OHCs of mammalian inner ear. The electromotility of OHCs driven by prestin is responsible for the cochlear amplification which is required for normal hearing in adult animals. Postnatal expression of prestin and activity of OHCs may contribute to the maturation of hearing in rodents. However, the temporal and spatial expression of prestin in cochlea during the development is not well characterized. In the present study, we examined the expression and function of prestin from the OHCs in apical, middle, and basal turns of the cochleae of postnatal rats. Prestin first appeared at postnatal day 6 (P6 for basal turn, P7 in middle turn, and P9 for apical turn of cochlea. The expression level increased progressively over the next few days and by P14 reached the mature level for all three segments. By comparison with the time course of the development of auditory brainstem response for different frequencies, our data reveal that prestin expression synchronized with the hearing development. The present study suggests that the onset time of hearing may require the expression of prestin and is determined by the mature function of OHCs.

  15. Presbycusis and auditory brainstem responses: a review

    Directory of Open Access Journals (Sweden)

    Shilpa Khullar

    2011-06-01

    Full Text Available Age-related hearing loss or presbycusis is a complex phenomenon consisting of elevation of hearing levels as well as changes in the auditory processing. It is commonly classified into four categories depending on the cause. Auditory brainstem responses (ABRs are a type of early evoked potentials recorded within the first 10 ms of stimulation. They represent the synchronized activity of the auditory nerve and the brainstem. Some of the changes that occur in the aging auditory system may significantly influence the interpretation of the ABRs in comparison with the ABRs of the young adults. The waves of ABRs are described in terms of amplitude, latencies and interpeak latency of the different waves. There is a tendency of the amplitude to decrease and the absolute latencies to increase with advancing age but these trends are not always clear due to increase in threshold with advancing age that act a major confounding factor in the interpretation of ABRs.

  16. Motor Synchronization in Patients With Schizophrenia: Preserved Time Representation With Abnormalities in Predictive Timing

    Directory of Open Access Journals (Sweden)

    Hélène Wilquin

    2018-05-01

    Full Text Available Objective: Basic temporal dysfunctions have been described in patients with schizophrenia, which may impact their ability to connect and synchronize with the outer world. The present study was conducted with the aim to distinguish between interval timing and synchronization difficulties and more generally the spatial-temporal organization disturbances for voluntary actions. A new sensorimotor synchronization task was developed to test these abilities.Method: Twenty-four chronic schizophrenia patients matched with 27 controls performed a spatial-tapping task in which finger taps were to be produced in synchrony with a regular metronome to six visual targets presented around a virtual circle on a tactile screen. Isochronous (time intervals of 500 ms and non-isochronous auditory sequences (alternated time intervals of 300/600 ms were presented. The capacity to produce time intervals accurately versus the ability to synchronize own actions (tap with external events (tone were measured.Results: Patients with schizophrenia were able to produce the tapping patterns of both isochronous and non-isochronous auditory sequences as accurately as controls producing inter-response intervals close to the expected interval of 500 and 900 ms, respectively. However, the synchronization performances revealed significantly more positive asynchrony means (but similar variances in the patient group than in the control group for both types of auditory sequences.Conclusion: The patterns of results suggest that patients with schizophrenia are able to perceive and produce both simple and complex sequences of time intervals but are impaired in the ability to synchronize their actions with external events. These findings suggest a specific deficit in predictive timing, which may be at the core of early symptoms previously described in schizophrenia.

  17. Interactive rhythmic auditory stimulation reinstates natural 1/f timing in gait of Parkinson's patients.

    Directory of Open Access Journals (Sweden)

    Michael J Hove

    Full Text Available Parkinson's disease (PD and basal ganglia dysfunction impair movement timing, which leads to gait instability and falls. Parkinsonian gait consists of random, disconnected stride times--rather than the 1/f structure observed in healthy gait--and this randomness of stride times (low fractal scaling predicts falling. Walking with fixed-tempo Rhythmic Auditory Stimulation (RAS can improve many aspects of gait timing; however, it lowers fractal scaling (away from healthy 1/f structure and requires attention. Here we show that interactive rhythmic auditory stimulation reestablishes healthy gait dynamics in PD patients. In the experiment, PD patients and healthy participants walked with a no auditory stimulation, b fixed-tempo RAS, and c interactive rhythmic auditory stimulation. The interactive system used foot sensors and nonlinear oscillators to track and mutually entrain with the human's step timing. Patients consistently synchronized with the interactive system, their fractal scaling returned to levels of healthy participants, and their gait felt more stable to them. Patients and healthy participants rarely synchronized with fixed-tempo RAS, and when they did synchronize their fractal scaling declined from healthy 1/f levels. Five minutes after removing the interactive rhythmic stimulation, the PD patients' gait retained high fractal scaling, suggesting that the interaction stabilized the internal rhythm generating system and reintegrated timing networks. The experiment demonstrates that complex interaction is important in the (reemergence of 1/f structure in human behavior and that interactive rhythmic auditory stimulation is a promising therapeutic tool for improving gait of PD patients.

  18. OMEGA SYSTEM SYNCHRONIZATION.

    Science.gov (United States)

    TIME SIGNALS, * SYNCHRONIZATION (ELECTRONICS)), NETWORKS, FREQUENCY, STANDARDS, RADIO SIGNALS, ERRORS, VERY LOW FREQUENCY, PROPAGATION, ACCURACY, ATOMIC CLOCKS, CESIUM, RADIO STATIONS, NAVAL SHORE FACILITIES

  19. Stages of chaotic synchronization.

    Science.gov (United States)

    Tang, D. Y.; Dykstra, R.; Hamilton, M. W.; Heckenberg, N. R.

    1998-09-01

    In an experimental investigation of the response of a chaotic system to a chaotic driving force, we have observed synchronization of chaos of the response system in the forms of generalized synchronization, phase synchronization, and lag synchronization to the driving signal. In this paper we compare the features of these forms of synchronized chaos and study their relations and physical origins. We found that different forms of chaotic synchronization could be interpreted as different stages of nonlinear interaction between the coupled chaotic systems. (c) 1998 American Institute of Physics.

  20. AUX: a scripting language for auditory signal processing and software packages for psychoacoustic experiments and education.

    Science.gov (United States)

    Kwon, Bomjun J

    2012-06-01

    This article introduces AUX (AUditory syntaX), a scripting syntax specifically designed to describe auditory signals and processing, to the members of the behavioral research community. The syntax is based on descriptive function names and intuitive operators suitable for researchers and students without substantial training in programming, who wish to generate and examine sound signals using a written script. In this article, the essence of AUX is discussed and practical examples of AUX scripts specifying various signals are illustrated. Additionally, two accompanying Windows-based programs and development libraries are described. AUX Viewer is a program that generates, visualizes, and plays sounds specified in AUX. AUX Viewer can also be used for class demonstrations or presentations. Another program, Psycon, allows a wide range of sound signals to be used as stimuli in common psychophysical testing paradigms, such as the adaptive procedure, the method of constant stimuli, and the method of adjustment. AUX Library is also provided, so that researchers can develop their own programs utilizing AUX. The philosophical basis of AUX is to separate signal generation from the user interface needed for experiments. AUX scripts are portable and reusable; they can be shared by other researchers, regardless of differences in actual AUX-based programs, and reused for future experiments. In short, the use of AUX can be potentially beneficial to all members of the research community-both those with programming backgrounds and those without.

  1. Application of auditory signals to the operation of an agricultural vehicle: results of pilot testing.

    Science.gov (United States)

    Karimi, D; Mondor, T A; Mann, D D

    2008-01-01

    The operation of agricultural vehicles is a multitask activity that requires proper distribution of attentional resources. Human factors theories suggest that proper utilization of the operator's sensory capacities under such conditions can improve the operator's performance and reduce the operator's workload. Using a tractor driving simulator, this study investigated whether auditory cues can be used to improve performance of the operator of an agricultural vehicle. Steering of a vehicle was simulated in visual mode (where driving error was shown to the subject using a lightbar) and in auditory mode (where a pair of speakers were used to convey the driving error direction and/or magnitude). A secondary task was also introduced in order to simulate the monitoring of an attached machine. This task included monitoring of two identical displays, which were placed behind the simulator, and responding to them, when needed, using a joystick. This task was also implemented in auditory mode (in which a beep signaled the subject to push the proper button when a response was needed) and in visual mode (in which there was no beep and visual, monitoring of the displays was necessary). Two levels of difficulty of the monitoring task were used. Deviation of the simulated vehicle from a desired straight line was used as the measure of performance in the steering task, and reaction time to the displays was used as the measure of performance in the monitoring task. Results of the experiments showed that steering performance was significantly better when steering was a visual task (driving errors were 40% to 60% of the driving errors in auditory mode), although subjective evaluations showed that auditory steering could be easier, depending on the implementation. Performance in the monitoring task was significantly better for auditory implementation (reaction time was approximately 6 times shorter), and this result was strongly supported by subjective ratings. The majority of the

  2. A method to synchronize signals from multiple patient monitoring devices through a single input channel for inclusion in list-mode acquisitions

    International Nuclear Information System (INIS)

    O’Connor, J. Michael; Pretorius, P. Hendrik; Johnson, Karen; King, Michael A.

    2013-01-01

    Purpose: This technical note documents a method that the authors developed for combining a signal to synchronize a patient-monitoring device with a second physiological signal for inclusion into list-mode acquisition. Our specific application requires synchronizing an external patient motion-tracking system with a medical imaging system by multiplexing the tracking input with the ECG input. The authors believe that their methodology can be adapted for use in a variety of medical imaging modalities including single photon emission computed tomography (SPECT) and positron emission tomography (PET). Methods: The authors insert a unique pulse sequence into a single physiological input channel. This sequence is then recorded in the list-mode acquisition along with the R-wave pulse used for ECG gating. The specific form of our pulse sequence allows for recognition of the time point being synchronized even when portions of the pulse sequence are lost due to collisions with R-wave pulses. This was achieved by altering our software used in binning the list-mode data to recognize even a portion of our pulse sequence. Limitations on heart rates at which our pulse sequence could be reliably detected were investigated by simulating the mixing of the two signals as a function of heart rate and time point during the cardiac cycle at which our pulse sequence is mixed with the cardiac signal. Results: The authors have successfully achieved accurate temporal synchronization of our motion-tracking system with acquisition of SPECT projections used in 17 recent clinical research cases. In our simulation analysis the authors determined that synchronization to enable compensation for body and respiratory motion could be achieved for heart rates up to 125 beats-per-minute (bpm). Conclusions: Synchronization of list-mode acquisition with external patient monitoring devices such as those employed in motion-tracking can reliably be achieved using a simple method that can be implemented using

  3. A method to synchronize signals from multiple patient monitoring devices through a single input channel for inclusion in list-mode acquisitions

    Energy Technology Data Exchange (ETDEWEB)

    O’Connor, J. Michael; Pretorius, P. Hendrik; Johnson, Karen; King, Michael A., E-mail: Michael.King@umassmed.edu [Department of Radiology, University of Massachusetts Medical School, Worcester, Massachusetts 01655 (United States)

    2013-12-15

    Purpose: This technical note documents a method that the authors developed for combining a signal to synchronize a patient-monitoring device with a second physiological signal for inclusion into list-mode acquisition. Our specific application requires synchronizing an external patient motion-tracking system with a medical imaging system by multiplexing the tracking input with the ECG input. The authors believe that their methodology can be adapted for use in a variety of medical imaging modalities including single photon emission computed tomography (SPECT) and positron emission tomography (PET). Methods: The authors insert a unique pulse sequence into a single physiological input channel. This sequence is then recorded in the list-mode acquisition along with the R-wave pulse used for ECG gating. The specific form of our pulse sequence allows for recognition of the time point being synchronized even when portions of the pulse sequence are lost due to collisions with R-wave pulses. This was achieved by altering our software used in binning the list-mode data to recognize even a portion of our pulse sequence. Limitations on heart rates at which our pulse sequence could be reliably detected were investigated by simulating the mixing of the two signals as a function of heart rate and time point during the cardiac cycle at which our pulse sequence is mixed with the cardiac signal. Results: The authors have successfully achieved accurate temporal synchronization of our motion-tracking system with acquisition of SPECT projections used in 17 recent clinical research cases. In our simulation analysis the authors determined that synchronization to enable compensation for body and respiratory motion could be achieved for heart rates up to 125 beats-per-minute (bpm). Conclusions: Synchronization of list-mode acquisition with external patient monitoring devices such as those employed in motion-tracking can reliably be achieved using a simple method that can be implemented using

  4. Auditory evacuation beacons

    NARCIS (Netherlands)

    Wijngaarden, S.J. van; Bronkhorst, A.W.; Boer, L.C.

    2005-01-01

    Auditory evacuation beacons can be used to guide people to safe exits, even when vision is totally obscured by smoke. Conventional beacons make use of modulated noise signals. Controlled evacuation experiments show that such signals require explicit instructions and are often misunderstood. A new

  5. Amygdala and auditory cortex exhibit distinct sensitivity to relevant acoustic features of auditory emotions.

    Science.gov (United States)

    Pannese, Alessia; Grandjean, Didier; Frühholz, Sascha

    2016-12-01

    Discriminating between auditory signals of different affective value is critical to successful social interaction. It is commonly held that acoustic decoding of such signals occurs in the auditory system, whereas affective decoding occurs in the amygdala. However, given that the amygdala receives direct subcortical projections that bypass the auditory cortex, it is possible that some acoustic decoding occurs in the amygdala as well, when the acoustic features are relevant for affective discrimination. We tested this hypothesis by combining functional neuroimaging with the neurophysiological phenomena of repetition suppression (RS) and repetition enhancement (RE) in human listeners. Our results show that both amygdala and auditory cortex responded differentially to physical voice features, suggesting that the amygdala and auditory cortex decode the affective quality of the voice not only by processing the emotional content from previously processed acoustic features, but also by processing the acoustic features themselves, when these are relevant to the identification of the voice's affective value. Specifically, we found that the auditory cortex is sensitive to spectral high-frequency voice cues when discriminating vocal anger from vocal fear and joy, whereas the amygdala is sensitive to vocal pitch when discriminating between negative vocal emotions (i.e., anger and fear). Vocal pitch is an instantaneously recognized voice feature, which is potentially transferred to the amygdala by direct subcortical projections. These results together provide evidence that, besides the auditory cortex, the amygdala too processes acoustic information, when this is relevant to the discrimination of auditory emotions. Copyright © 2016 Elsevier Ltd. All rights reserved.

  6. Clock synchronization and dispersion

    International Nuclear Information System (INIS)

    Giovannetti, Vittorio; Lloyd, Seth; Maccone, Lorenzo; Wong, Franco N C

    2002-01-01

    We present a method to defeat effects of dispersion of timing signals when synchronizing clocks. It is based on the recently proposed 'conveyor belt synchronization' scheme and on the quantum dispersion cancellation effect

  7. Tap Arduino: An Arduino microcontroller for low-latency auditory feedback in sensorimotor synchronization experiments.

    Science.gov (United States)

    Schultz, Benjamin G; van Vugt, Floris T

    2016-12-01

    Timing abilities are often measured by having participants tap their finger along with a metronome and presenting tap-triggered auditory feedback. These experiments predominantly use electronic percussion pads combined with software (e.g., FTAP or Max/MSP) that records responses and delivers auditory feedback. However, these setups involve unknown latencies between tap onset and auditory feedback and can sometimes miss responses or record multiple, superfluous responses for a single tap. These issues may distort measurements of tapping performance or affect the performance of the individual. We present an alternative setup using an Arduino microcontroller that addresses these issues and delivers low-latency auditory feedback. We validated our setup by having participants (N = 6) tap on a force-sensitive resistor pad connected to the Arduino and on an electronic percussion pad with various levels of force and tempi. The Arduino delivered auditory feedback through a pulse-width modulation (PWM) pin connected to a headphone jack or a wave shield component. The Arduino's PWM (M = 0.6 ms, SD = 0.3) and wave shield (M = 2.6 ms, SD = 0.3) demonstrated significantly lower auditory feedback latencies than the percussion pad (M = 9.1 ms, SD = 2.0), FTAP (M = 14.6 ms, SD = 2.8), and Max/MSP (M = 15.8 ms, SD = 3.4). The PWM and wave shield latencies were also significantly less variable than those from FTAP and Max/MSP. The Arduino missed significantly fewer taps, and recorded fewer superfluous responses, than the percussion pad. The Arduino captured all responses, whereas at lower tapping forces, the percussion pad missed more taps. Regardless of tapping force, the Arduino outperformed the percussion pad. Overall, the Arduino is a high-precision, low-latency, portable, and affordable tool for auditory experiments.

  8. No Need for Templates in the Auditory Enhancement Effect.

    Science.gov (United States)

    Carcagno, Samuele; Semal, Catherine; Demany, Laurent

    2013-01-01

    The audibility of a target tone in a multitone background masker is enhanced by the presentation of a precursor sound consisting of the masker alone. There is evidence that precursor-induced neural adaptation plays a role in this perceptual enhancement. However, the precursor may also be strategically used by listeners as a spectral template of the following masker to better segregate it from the target. In the present study, we tested this hypothesis by measuring the audibility of a target tone in a multitone masker after the presentation of precursors which, in some conditions, were made dissimilar to the masker by gating their components asynchronously. The precursor and the following sound were presented either to the same ear or to opposite ears. In either case, we found no significant difference in the amount of enhancement produced by synchronous and asynchronous precursors. In a second experiment, listeners had to judge whether a synchronous multitone complex contained exactly the same tones as a preceding precursor complex or had one tone less. In this experiment, listeners performed significantly better with synchronous than with asynchronous precursors, showing that asynchronous precursors were poorer perceptual templates of the synchronous multitone complexes. Overall, our findings indicate that precursor-induced auditory enhancement cannot be fully explained by the strategic use of the precursor as a template of the following masker. Our results are consistent with an explanation of enhancement based on selective neural adaptation taking place at a central locus of the auditory system.

  9. Data-derived symbol synchronization of MASK and QASK signals. [Multilevel and Quadrature Amplitude Shift Keying

    Science.gov (United States)

    Simon, M. K.

    1975-01-01

    Much has been said in the literature regarding the problem of establishing symbol synchronization in binary baseband digital communication systems. By comparison, the literature contains little information relating to the extraction of symbol sync from multilevel baseband data. With the recent interest in multilevel amplitude-shift keying (MASK) and quadrature amplitude-shift keying (QASK) as signaling techniques for multilevel digital communications systems, the problem of providing symbol synchronization in the receivers of such systems becomes paramount. This paper presents a technique for extracting symbol sync from a MASK or QASK signal which has been transmitted over an infinite-bandwidth white Gaussian noise channel. The scheme is essentially a generalization of the data transition tracking loop (DTTL) which has heretofore been used in PSK systems. The performance of the loop is analyzed in terms of its mean-squared symbol sync jitter and its effects on the data detection process in MASK and QASK systems.

  10. Data-derived symbol synchronization of MASK and QASK signals. [for multilevel digital communication systems

    Science.gov (United States)

    Simon, M. K.

    1974-01-01

    Multilevel amplitude-shift-keying (MASK) and quadrature amplitude-shift-keying (QASK) as signaling techniques for multilevel digital communications systems, and the problem of providing symbol synchronization in the receivers of such systems are discussed. A technique is presented for extracting symbol sync from an MASK or QASK signal. The scheme is a generalization of the data transition tracking loop used in PSK systems. The performance of the loop was analyzed in terms of its mean-squared jitter and its effects on the data detection process in MASK and QASK systems.

  11. Auditory stimulus timing influences perceived duration of co-occurring visual stimuli

    Directory of Open Access Journals (Sweden)

    Vincenzo eRomei

    2011-09-01

    Full Text Available There is increasing interest in multisensory influences upon sensory-specific judgements, such as when auditory stimuli affect visual perception. Here we studied whether the duration of an auditory event can objectively affect the perceived duration of a co-occurring visual event. On each trial, participants were presented with a pair of successive flashes and had to judge whether the first or second was longer. Two beeps were presented with the flashes. The order of short and long stimuli could be the same across audition and vision (audiovisual congruent or reversed, so that the longer flash was accompanied by the shorter beep and vice versa (audiovisual incongruent; or the two beeps could have the same duration as each other. Beeps and flashes could onset synchronously or asynchronously. In a further control experiment, the beep durations were much longer (tripled than the flashes. Results showed that visual duration-discrimination sensitivity (d' was significantly higher for congruent (and significantly lower for incongruent audiovisual synchronous combinations, relative to the visual only presentation. This effect was abolished when auditory and visual stimuli were presented asynchronously, or when sound durations tripled those of flashes. We conclude that the temporal properties of co-occurring auditory stimuli influence the perceived duration of visual stimuli and that this can reflect genuine changes in visual sensitivity rather than mere response bias.

  12. Effect of delayed auditory feedback on stuttering with and without central auditory processing disorders.

    Science.gov (United States)

    Picoloto, Luana Altran; Cardoso, Ana Cláudia Vieira; Cerqueira, Amanda Venuti; Oliveira, Cristiane Moço Canhetti de

    2017-12-07

    To verify the effect of delayed auditory feedback on speech fluency of individuals who stutter with and without central auditory processing disorders. The participants were twenty individuals with stuttering from 7 to 17 years old and were divided into two groups: Stuttering Group with Auditory Processing Disorders (SGAPD): 10 individuals with central auditory processing disorders, and Stuttering Group (SG): 10 individuals without central auditory processing disorders. Procedures were: fluency assessment with non-altered auditory feedback (NAF) and delayed auditory feedback (DAF), assessment of the stuttering severity and central auditory processing (CAP). Phono Tools software was used to cause a delay of 100 milliseconds in the auditory feedback. The "Wilcoxon Signal Post" test was used in the intragroup analysis and "Mann-Whitney" test in the intergroup analysis. The DAF caused a statistically significant reduction in SG: in the frequency score of stuttering-like disfluencies in the analysis of the Stuttering Severity Instrument, in the amount of blocks and repetitions of monosyllabic words, and in the frequency of stuttering-like disfluencies of duration. Delayed auditory feedback did not cause statistically significant effects on SGAPD fluency, individuals with stuttering with auditory processing disorders. The effect of delayed auditory feedback in speech fluency of individuals who stutter was different in individuals of both groups, because there was an improvement in fluency only in individuals without auditory processing disorder.

  13. Neural synchronization via potassium signaling

    DEFF Research Database (Denmark)

    Postnov, Dmitry E; Ryazanova, Ludmila S; Mosekilde, Erik

    2006-01-01

    Using a relatively simple model we examine how variations of the extracellular potassium concentration can give rise to synchronization of two nearby pacemaker cells. With the volume of the extracellular space and the rate of potassium diffusion as control parameters, the dual nature of this reso...

  14. Recurrence of task set-related MEG signal patterns during auditory working memory.

    Science.gov (United States)

    Peters, Benjamin; Bledowski, Christoph; Rieder, Maria; Kaiser, Jochen

    2016-06-01

    Processing of auditory spatial and non-spatial information in working memory has been shown to rely on separate cortical systems. While previous studies have demonstrated differences in spatial versus non-spatial processing from the encoding of to-be-remembered stimuli onwards, here we investigated whether such differences would be detectable already prior to presentation of the sample stimulus. We analyzed broad-band magnetoencephalography data from 15 healthy adults during an auditory working memory paradigm starting with a visual cue indicating the task-relevant stimulus feature for a given trial (lateralization or pitch) and a subsequent 1.5-s pre-encoding phase. This was followed by a sample sound (0.2s), the delay phase (0.8s) and a test stimulus (0.2s) after which participants made a match/non-match decision. Linear discriminant functions were trained to decode task-specific signal patterns throughout the task, and temporal generalization was used to assess whether the neural codes discriminating between the tasks during the pre-encoding phase would recur during later task periods. The spatial versus non-spatial tasks could indeed be discriminated after the onset of the cue onwards, and decoders trained during the pre-encoding phase successfully discriminated the tasks during both sample stimulus encoding and during the delay phase. This demonstrates that task-specific neural codes are established already before the memorandum is presented and that the same patterns are reestablished during stimulus encoding and maintenance. This article is part of a Special Issue entitled SI: Auditory working memory. Copyright © 2015 Elsevier B.V. All rights reserved.

  15. Top-down (Prior Knowledge) and Bottom-up (Perceptual Modality) Influences on Spontaneous Interpersonal Synchronization.

    Science.gov (United States)

    Gipson, Christina L; Gorman, Jamie C; Hessler, Eric E

    2016-04-01

    Coordination with others is such a fundamental part of human activity that it can happen unintentionally. This unintentional coordination can manifest as synchronization and is observed in physical and human systems alike. We investigated the role of top-down influences (prior knowledge of the perceptual modality their partner is using) and bottom-up factors (perceptual modality combination) on spontaneous interpersonal synchronization. We examine this phenomena with respect to two different theoretical perspectives that differently emphasize top-down and bottom-up factors in interpersonal synchronization: joint-action/shared cognition theories and ecological-interactive theories. In an empirical study twelve dyads performed a finger oscillation task while attending to each other's movements through either visual, auditory, or visual and auditory perceptual modalities. Half of the participants were given prior knowledge of their partner's perceptual capabilities for coordinating across these different perceptual modality combinations. We found that the effect of top-down influence depends on the perceptual modality combination between two individuals. When people used the same perceptual modalities, top-down influence resulted in less synchronization and when people used different perceptual modalities, top-down influence resulted in more synchronization. Furthermore, persistence in the change in behavior as a result of having perceptual information about each other ('social memory') was stronger when this top-down influence was present.

  16. Validation of a raw data-based synchronization signal (kymogram) for phase-correlated cardiac image reconstruction

    International Nuclear Information System (INIS)

    Ertel, Dirk; Kachelriess, Marc; Kalender, Willi A.; Pflederer, Tobias; Achenbach, Stephan; Steffen, Peter

    2008-01-01

    Phase-correlated reconstruction is commonly used in computed tomography (CT)-based cardiac imaging. Alternatively to the commonly used ECG, the raw data-based kymogram function can be used as a synchronization signal. We used raw data of 100 consecutive patient exams to compare the performance of kymogram function to the ECG signal. For objective validation the correlation of the ECG and the kymogram was assessed. Additionally, we performed a double-blinded comparison of ECG-based and kymogram-based phase-correlated images. The two synchronization signals showed good correlation indicated by a mean difference in the detected heart rate of negligible 0.2 bpm. The mean image quality score was 2.0 points for kymogram-correlated images and 2.3 points for ECG-correlated images, respectively (3: best; 0: worst). The kymogram and the ECG provided images adequate for diagnosis for 93 and 97 patients, respectively. For 50% of the datasets the kymogram provided an equivalent or even higher image quality compared with the ECG signal. We conclude that an acceptable image quality can be assured in most cases by the kymogram. Improvements of image quality by the kymogram function were observed in a noticeable number of cases. The kymogram can serve as a backup solution when an ECG is not available or lacking in quality. (orig.)

  17. Auditory Pattern Memory and Group Signal Detection

    National Research Council Canada - National Science Library

    Sorkin, Robert

    1997-01-01

    .... The experiments with temporally-coded auditory patterns showed how listeners' attention is influenced by the position and the amount of information carried by different segments of the pattern...

  18. Auditory conflict and congruence in frontotemporal dementia.

    Science.gov (United States)

    Clark, Camilla N; Nicholas, Jennifer M; Agustus, Jennifer L; Hardy, Christopher J D; Russell, Lucy L; Brotherhood, Emilie V; Dick, Katrina M; Marshall, Charles R; Mummery, Catherine J; Rohrer, Jonathan D; Warren, Jason D

    2017-09-01

    Impaired analysis of signal conflict and congruence may contribute to diverse socio-emotional symptoms in frontotemporal dementias, however the underlying mechanisms have not been defined. Here we addressed this issue in patients with behavioural variant frontotemporal dementia (bvFTD; n = 19) and semantic dementia (SD; n = 10) relative to healthy older individuals (n = 20). We created auditory scenes in which semantic and emotional congruity of constituent sounds were independently probed; associated tasks controlled for auditory perceptual similarity, scene parsing and semantic competence. Neuroanatomical correlates of auditory congruity processing were assessed using voxel-based morphometry. Relative to healthy controls, both the bvFTD and SD groups had impaired semantic and emotional congruity processing (after taking auditory control task performance into account) and reduced affective integration of sounds into scenes. Grey matter correlates of auditory semantic congruity processing were identified in distributed regions encompassing prefrontal, parieto-temporal and insular areas and correlates of auditory emotional congruity in partly overlapping temporal, insular and striatal regions. Our findings suggest that decoding of auditory signal relatedness may probe a generic cognitive mechanism and neural architecture underpinning frontotemporal dementia syndromes. Copyright © 2017 The Author(s). Published by Elsevier Ltd.. All rights reserved.

  19. Main injector synchronous timing system

    International Nuclear Information System (INIS)

    Blokland, W.; Steimel, J.

    1998-01-01

    The Synchronous Timing System is designed to provide sub-nanosecond timing to instrumentation during the acceleration of particles in the Main Injector. Increased energy of the beam particles leads to a small but significant increase in speed, reducing the time it takes to complete a full turn of the ring by 61 nanoseconds (or more than 3 rf buckets). In contrast, the reference signal, used to trigger instrumentation and transmitted over a cable, has a constant group delay. This difference leads to a phase slip during the ramp and prevents instrumentation such as dampers from properly operating without additional measures. The Synchronous Timing System corrects for this phase slip as well as signal propagation time changes due to temperature variations. A module at the LLRF system uses a 1.2 Gbit/s G-Link chip to transmit the rf clock and digital data (e.g. the current frequency) over a single mode fiber around the ring. Fiber optic couplers at service buildings split off part of this signal for a local module which reconstructs a synchronous beam reference signal. This paper describes the background, design and expected performance of the Synchronous Timing System. copyright 1998 American Institute of Physics

  20. Main injector synchronous timing system

    International Nuclear Information System (INIS)

    Blokland, Willem; Steimel, James

    1998-01-01

    The Synchronous Timing System is designed to provide sub-nanosecond timing to instrumentation during the acceleration of particles in the Main Injector. Increased energy of the beam particles leads to a small but significant increase in speed, reducing the time it takes to complete a full turn of the ring by 61 nanoseconds (or more than 3 rf buckets). In contrast, the reference signal, used to trigger instrumentation and transmitted over a cable, has a constant group delay. This difference leads to a phase slip during the ramp and prevents instrumentation such as dampers from properly operating without additional measures. The Synchronous Timing System corrects for this phase slip as well as signal propagation time changes due to temperature variations. A module at the LLRF system uses a 1.2 Gbit/s G-Link chip to transmit the rf clock and digital data (e.g. the current frequency) over a single mode fiber around the ring. Fiber optic couplers at service buildings split off part of this signal for a local module which reconstructs a synchronous beam reference signal. This paper describes the background, design and expected performance of the Synchronous Timing System

  1. Pitch Synchronous Segmentation of Speech Signals

    Data.gov (United States)

    National Aeronautics and Space Administration — The Pitch Synchronous Segmentation (PSS) that accelerates speech without changing its fundamental frequency method could be applied and evaluated for use at NASA....

  2. The role of auditory cortices in the retrieval of single-trial auditory-visual object memories.

    Science.gov (United States)

    Matusz, Pawel J; Thelen, Antonia; Amrein, Sarah; Geiser, Eveline; Anken, Jacques; Murray, Micah M

    2015-03-01

    Single-trial encounters with multisensory stimuli affect both memory performance and early-latency brain responses to visual stimuli. Whether and how auditory cortices support memory processes based on single-trial multisensory learning is unknown and may differ qualitatively and quantitatively from comparable processes within visual cortices due to purported differences in memory capacities across the senses. We recorded event-related potentials (ERPs) as healthy adults (n = 18) performed a continuous recognition task in the auditory modality, discriminating initial (new) from repeated (old) sounds of environmental objects. Initial presentations were either unisensory or multisensory; the latter entailed synchronous presentation of a semantically congruent or a meaningless image. Repeated presentations were exclusively auditory, thus differing only according to the context in which the sound was initially encountered. Discrimination abilities (indexed by d') were increased for repeated sounds that were initially encountered with a semantically congruent image versus sounds initially encountered with either a meaningless or no image. Analyses of ERPs within an electrical neuroimaging framework revealed that early stages of auditory processing of repeated sounds were affected by prior single-trial multisensory contexts. These effects followed from significantly reduced activity within a distributed network, including the right superior temporal cortex, suggesting an inverse relationship between brain activity and behavioural outcome on this task. The present findings demonstrate how auditory cortices contribute to long-term effects of multisensory experiences on auditory object discrimination. We propose a new framework for the efficacy of multisensory processes to impact both current multisensory stimulus processing and unisensory discrimination abilities later in time. © 2015 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  3. Effects of sequential streaming on auditory masking using psychoacoustics and auditory evoked potentials.

    Science.gov (United States)

    Verhey, Jesko L; Ernst, Stephan M A; Yasin, Ifat

    2012-03-01

    The present study was aimed at investigating the relationship between the mismatch negativity (MMN) and psychoacoustical effects of sequential streaming on comodulation masking release (CMR). The influence of sequential streaming on CMR was investigated using a psychoacoustical alternative forced-choice procedure and electroencephalography (EEG) for the same group of subjects. The psychoacoustical data showed, that adding precursors comprising of only off-signal-frequency maskers abolished the CMR. Complementary EEG data showed an MMN irrespective of the masker envelope correlation across frequency when only the off-signal-frequency masker components were present. The addition of such precursors promotes a separation of the on- and off-frequency masker components into distinct auditory objects preventing the auditory system from using comodulation as an additional cue. A frequency-specific adaptation changing the representation of the flanking bands in the streaming conditions may also contribute to the reduction of CMR in the stream conditions, however, it is unlikely that adaptation is the primary reason for the streaming effect. A neurophysiological correlate of sequential streaming was found in EEG data using MMN, but the magnitude of the MMN was not correlated with the audibility of the signal in CMR experiments. Dipole source analysis indicated different cortical regions involved in processing auditory streaming and modulation detection. In particular, neural sources for processing auditory streaming include cortical regions involved in decision-making. Copyright © 2012 Elsevier B.V. All rights reserved.

  4. No Need for Templates in the Auditory Enhancement Effect.

    Directory of Open Access Journals (Sweden)

    Samuele Carcagno

    Full Text Available The audibility of a target tone in a multitone background masker is enhanced by the presentation of a precursor sound consisting of the masker alone. There is evidence that precursor-induced neural adaptation plays a role in this perceptual enhancement. However, the precursor may also be strategically used by listeners as a spectral template of the following masker to better segregate it from the target. In the present study, we tested this hypothesis by measuring the audibility of a target tone in a multitone masker after the presentation of precursors which, in some conditions, were made dissimilar to the masker by gating their components asynchronously. The precursor and the following sound were presented either to the same ear or to opposite ears. In either case, we found no significant difference in the amount of enhancement produced by synchronous and asynchronous precursors. In a second experiment, listeners had to judge whether a synchronous multitone complex contained exactly the same tones as a preceding precursor complex or had one tone less. In this experiment, listeners performed significantly better with synchronous than with asynchronous precursors, showing that asynchronous precursors were poorer perceptual templates of the synchronous multitone complexes. Overall, our findings indicate that precursor-induced auditory enhancement cannot be fully explained by the strategic use of the precursor as a template of the following masker. Our results are consistent with an explanation of enhancement based on selective neural adaptation taking place at a central locus of the auditory system.

  5. Development of a Stair-Step Multifrequency Synchronized Excitation Signal for Fast Bioimpedance Spectroscopy

    Science.gov (United States)

    Bian, He; Du, Fangling; Sun, Qiang

    2014-01-01

    Wideband excitation signal with finite prominent harmonic components is desirable for fast bioimpedance spectroscopy (BIS) measurements. This work introduces a simple method to synthesize and realize a type of periodical stair-step multifrequency synchronized (MFS) signal. The Fourier series analysis shows that the p-order MFS signal f(p, t) has constant 81.06% energy distributed equally on its p  2nth primary harmonics. The synthesis principle is described firstly and then two examples of the 4-order and 5-order MFS signals, f(4, t) and f(5, t), are synthesized. The method to implement the MFS waveform based on a field-programmable gate array (FPGA) and a digital to analog converter (DAC) is also presented. Both the number and the frequencies of the expected primary harmonics can be adjusted as needed. An impedance measurement experiment on a RC three-element equivalent model is performed, and results show acceptable precision, which validates the feasibility of the MFS excitation. PMID:24701563

  6. Development of a Stair-Step Multifrequency Synchronized Excitation Signal for Fast Bioimpedance Spectroscopy

    Directory of Open Access Journals (Sweden)

    Yuxiang Yang

    2014-01-01

    Full Text Available Wideband excitation signal with finite prominent harmonic components is desirable for fast bioimpedance spectroscopy (BIS measurements. This work introduces a simple method to synthesize and realize a type of periodical stair-step multifrequency synchronized (MFS signal. The Fourier series analysis shows that the p-order MFS signal f(p,t has constant 81.06% energy distributed equally on its p  2nth primary harmonics. The synthesis principle is described firstly and then two examples of the 4-order and 5-order MFS signals, f(4,t and f(5,t, are synthesized. The method to implement the MFS waveform based on a field-programmable gate array (FPGA and a digital to analog converter (DAC is also presented. Both the number and the frequencies of the expected primary harmonics can be adjusted as needed. An impedance measurement experiment on a RC three-element equivalent model is performed, and results show acceptable precision, which validates the feasibility of the MFS excitation.

  7. Amplitude-modulated stimuli reveal auditory-visual interactions in brain activity and brain connectivity

    Directory of Open Access Journals (Sweden)

    Mark eLaing

    2015-10-01

    Full Text Available The temporal congruence between auditory and visual signals coming from the same source can be a powerful means by which the brain integrates information from different senses. To investigate how the brain uses temporal information to integrate auditory and visual information from continuous yet unfamiliar stimuli, we use amplitude-modulated tones and size-modulated shapes with which we could manipulate the temporal congruence between the sensory signals. These signals were independently modulated at a slow or a fast rate. Participants were presented with auditory-only, visual-only or auditory-visual (AV trials in the scanner. On AV trials, the auditory and visual signal could have the same (AV congruent or different modulation rates (AV incongruent. Using psychophysiological interaction analyses, we found that auditory regions showed increased functional connectivity predominantly with frontal regions for AV incongruent relative to AV congruent stimuli. We further found that superior temporal regions, shown previously to integrate auditory and visual signals, showed increased connectivity with frontal and parietal regions for the same contrast. Our findings provide evidence that both activity in a network of brain regions and their connectivity are important for auditory-visual integration, and help to bridge the gap between transient and familiar AV stimuli used in previous studies.

  8. Development of the auditory system

    Science.gov (United States)

    Litovsky, Ruth

    2015-01-01

    Auditory development involves changes in the peripheral and central nervous system along the auditory pathways, and these occur naturally, and in response to stimulation. Human development occurs along a trajectory that can last decades, and is studied using behavioral psychophysics, as well as physiologic measurements with neural imaging. The auditory system constructs a perceptual space that takes information from objects and groups, segregates sounds, and provides meaning and access to communication tools such as language. Auditory signals are processed in a series of analysis stages, from peripheral to central. Coding of information has been studied for features of sound, including frequency, intensity, loudness, and location, in quiet and in the presence of maskers. In the latter case, the ability of the auditory system to perform an analysis of the scene becomes highly relevant. While some basic abilities are well developed at birth, there is a clear prolonged maturation of auditory development well into the teenage years. Maturation involves auditory pathways. However, non-auditory changes (attention, memory, cognition) play an important role in auditory development. The ability of the auditory system to adapt in response to novel stimuli is a key feature of development throughout the nervous system, known as neural plasticity. PMID:25726262

  9. Synchronized flow in oversaturated city traffic.

    Science.gov (United States)

    Kerner, Boris S; Klenov, Sergey L; Hermanns, Gerhard; Hemmerle, Peter; Rehborn, Hubert; Schreckenberg, Michael

    2013-11-01

    Based on numerical simulations with a stochastic three-phase traffic flow model, we reveal that moving queues (moving jams) in oversaturated city traffic dissolve at some distance upstream of the traffic signal while transforming into synchronized flow. It is found that, as in highway traffic [Kerner, Phys. Rev. E 85, 036110 (2012)], such a jam-absorption effect in city traffic is explained by a strong driver's speed adaptation: Time headways (space gaps) between vehicles increase upstream of a moving queue (moving jam), resulting in moving queue dissolution. It turns out that at given traffic signal parameters, the stronger the speed adaptation effect, the shorter the mean distance between the signal location and the road location at which moving queues dissolve fully and oversaturated traffic consists of synchronized flow only. A comparison of the synchronized flow in city traffic found in this Brief Report with synchronized flow in highway traffic is made.

  10. The human brain maintains contradictory and redundant auditory sensory predictions.

    Directory of Open Access Journals (Sweden)

    Marika Pieszek

    Full Text Available Computational and experimental research has revealed that auditory sensory predictions are derived from regularities of the current environment by using internal generative models. However, so far, what has not been addressed is how the auditory system handles situations giving rise to redundant or even contradictory predictions derived from different sources of information. To this end, we measured error signals in the event-related brain potentials (ERPs in response to violations of auditory predictions. Sounds could be predicted on the basis of overall probability, i.e., one sound was presented frequently and another sound rarely. Furthermore, each sound was predicted by an informative visual cue. Participants' task was to use the cue and to discriminate the two sounds as fast as possible. Violations of the probability based prediction (i.e., a rare sound as well as violations of the visual-auditory prediction (i.e., an incongruent sound elicited error signals in the ERPs (Mismatch Negativity [MMN] and Incongruency Response [IR]. Particular error signals were observed even in case the overall probability and the visual symbol predicted different sounds. That is, the auditory system concurrently maintains and tests contradictory predictions. Moreover, if the same sound was predicted, we observed an additive error signal (scalp potential and primary current density equaling the sum of the specific error signals. Thus, the auditory system maintains and tolerates functionally independently represented redundant and contradictory predictions. We argue that the auditory system exploits all currently active regularities in order to optimally prepare for future events.

  11. Synchronization ability of coupled cell-cycle oscillators in changing environments

    Science.gov (United States)

    2012-01-01

    Background The biochemical oscillator that controls periodic events during the Xenopus embryonic cell cycle is centered on the activity of CDKs, and the cell cycle is driven by a protein circuit that is centered on the cyclin-dependent protein kinase CDK1 and the anaphase-promoting complex (APC). Many studies have been conducted to confirm that the interactions in the cell cycle can produce oscillations and predict behaviors such as synchronization, but much less is known about how the various elaborations and collective behavior of the basic oscillators can affect the robustness of the system. Therefore, in this study, we investigate and model a multi-cell system of the Xenopus embryonic cell cycle oscillators that are coupled through a common complex protein, and then analyze their synchronization ability under four different external stimuli, including a constant input signal, a square-wave periodic signal, a sinusoidal signal and a noise signal. Results Through bifurcation analysis and numerical simulations, we obtain synchronization intervals of the sensitive parameters in the individual oscillator and the coupling parameters in the coupled oscillators. Then, we analyze the effects of these parameters on the synchronization period and amplitude, and find interesting phenomena, e.g., there are two synchronization intervals with activation coefficient in the Hill function of the activated CDK1 that activates the Plk1, and different synchronization intervals have distinct influences on the synchronization period and amplitude. To quantify the speediness and robustness of the synchronization, we use two quantities, the synchronization time and the robustness index, to evaluate the synchronization ability. More interestingly, we find that the coupled system has an optimal signal strength that maximizes the synchronization index under different external stimuli. Simulation results also show that the ability and robustness of the synchronization for the square

  12. Effect- and Performance-Based Auditory Feedback on Interpersonal Coordination

    Directory of Open Access Journals (Sweden)

    Tong-Hun Hwang

    2018-03-01

    Full Text Available When two individuals interact in a collaborative task, such as carrying a sofa or a table, usually spatiotemporal coordination of individual motor behavior will emerge. In many cases, interpersonal coordination can arise independently of verbal communication, based on the observation of the partners' movements and/or the object's movements. In this study, we investigate how social coupling between two individuals can emerge in a collaborative task under different modes of perceptual information. A visual reference condition was compared with three different conditions with new types of additional auditory feedback provided in real time: effect-based auditory feedback, performance-based auditory feedback, and combined effect/performance-based auditory feedback. We have developed a new paradigm in which the actions of both participants continuously result in a seamlessly merged effect on an object simulated by a tablet computer application. Here, participants should temporally synchronize their movements with a 90° phase difference and precisely adjust the finger dynamics in order to keep the object (a ball accurately rotating on a given circular trajectory on the tablet. Results demonstrate that interpersonal coordination in a joint task can be altered by different kinds of additional auditory information in various ways.

  13. Biological impact of music and software-based auditory training

    Science.gov (United States)

    Kraus, Nina

    2012-01-01

    Auditory-based communication skills are developed at a young age and are maintained throughout our lives. However, some individuals – both young and old – encounter difficulties in achieving or maintaining communication proficiency. Biological signals arising from hearing sounds relate to real-life communication skills such as listening to speech in noisy environments and reading, pointing to an intersection between hearing and cognition. Musical experience, amplification, and software-based training can improve these biological signals. These findings of biological plasticity, in a variety of subject populations, relate to attention and auditory memory, and represent an integrated auditory system influenced by both sensation and cognition. Learning outcomes The reader will (1) understand that the auditory system is malleable to experience and training, (2) learn the ingredients necessary for auditory learning to successfully be applied to communication, (3) learn that the auditory brainstem response to complex sounds (cABR) is a window into the integrated auditory system, and (4) see examples of how cABR can be used to track the outcome of experience and training. PMID:22789822

  14. Amplitude-modulated stimuli reveal auditory-visual interactions in brain activity and brain connectivity.

    Science.gov (United States)

    Laing, Mark; Rees, Adrian; Vuong, Quoc C

    2015-01-01

    The temporal congruence between auditory and visual signals coming from the same source can be a powerful means by which the brain integrates information from different senses. To investigate how the brain uses temporal information to integrate auditory and visual information from continuous yet unfamiliar stimuli, we used amplitude-modulated tones and size-modulated shapes with which we could manipulate the temporal congruence between the sensory signals. These signals were independently modulated at a slow or a fast rate. Participants were presented with auditory-only, visual-only, or auditory-visual (AV) trials in the fMRI scanner. On AV trials, the auditory and visual signal could have the same (AV congruent) or different modulation rates (AV incongruent). Using psychophysiological interaction analyses, we found that auditory regions showed increased functional connectivity predominantly with frontal regions for AV incongruent relative to AV congruent stimuli. We further found that superior temporal regions, shown previously to integrate auditory and visual signals, showed increased connectivity with frontal and parietal regions for the same contrast. Our findings provide evidence that both activity in a network of brain regions and their connectivity are important for AV integration, and help to bridge the gap between transient and familiar AV stimuli used in previous studies.

  15. Direct Contribution of Auditory Motion Information to Sound-Induced Visual Motion Perception

    Directory of Open Access Journals (Sweden)

    Souta Hidaka

    2011-10-01

    Full Text Available We have recently demonstrated that alternating left-right sound sources induce motion perception to static visual stimuli along the horizontal plane (SIVM: sound-induced visual motion perception, Hidaka et al., 2009. The aim of the current study was to elucidate whether auditory motion signals, rather than auditory positional signals, can directly contribute to the SIVM. We presented static visual flashes at retinal locations outside the fovea together with a lateral auditory motion provided by a virtual stereo noise source smoothly shifting in the horizontal plane. The flashes appeared to move in the situation where auditory positional information would have little influence on the perceived position of visual stimuli; the spatiotemporal position of the flashes was in the middle of the auditory motion trajectory. Furthermore, the auditory motion altered visual motion perception in a global motion display; in this display, different localized motion signals of multiple visual stimuli were combined to produce a coherent visual motion perception so that there was no clear one-to-one correspondence between the auditory stimuli and each visual stimulus. These findings suggest the existence of direct interactions between the auditory and visual modalities in motion processing and motion perception.

  16. Synchronization of chaotic systems

    International Nuclear Information System (INIS)

    Pecora, Louis M.; Carroll, Thomas L.

    2015-01-01

    We review some of the history and early work in the area of synchronization in chaotic systems. We start with our own discovery of the phenomenon, but go on to establish the historical timeline of this topic back to the earliest known paper. The topic of synchronization of chaotic systems has always been intriguing, since chaotic systems are known to resist synchronization because of their positive Lyapunov exponents. The convergence of the two systems to identical trajectories is a surprise. We show how people originally thought about this process and how the concept of synchronization changed over the years to a more geometric view using synchronization manifolds. We also show that building synchronizing systems leads naturally to engineering more complex systems whose constituents are chaotic, but which can be tuned to output various chaotic signals. We finally end up at a topic that is still in very active exploration today and that is synchronization of dynamical systems in networks of oscillators

  17. Auditory-visual integration of emotional signals in a virtual environment for cynophobia.

    Science.gov (United States)

    Taffou, Marine; Chapoulie, Emmanuelle; David, Adrien; Guerchouche, Rachid; Drettakis, George; Viaud-Delmon, Isabelle

    2012-01-01

    Cynophobia (dog phobia) has both visual and auditory relevant components. In order to investigate the efficacy of virtual reality (VR) exposure-based treatment for cynophobia, we studied the efficiency of auditory-visual environments in generating presence and emotion. We conducted an evaluation test with healthy participants sensitive to cynophobia in order to assess the capacity of auditory-visual virtual environments (VE) to generate fear reactions. Our application involves both high fidelity visual stimulation displayed in an immersive space and 3D sound. This specificity enables us to present and spatially manipulate fearful stimuli in the auditory modality, the visual modality and both. Our specific presentation of animated dog stimuli creates an environment that is highly arousing, suggesting that VR is a promising tool for cynophobia treatment and that manipulating auditory-visual integration might provide a way to modulate affect.

  18. Acoustic Trauma Changes the Parvalbumin-Positive Neurons in Rat Auditory Cortex

    Directory of Open Access Journals (Sweden)

    Congli Liu

    2018-01-01

    Full Text Available Acoustic trauma is being reported to damage the auditory periphery and central system, and the compromised cortical inhibition is involved in auditory disorders, such as hyperacusis and tinnitus. Parvalbumin-containing neurons (PV neurons, a subset of GABAergic neurons, greatly shape and synchronize neural network activities. However, the change of PV neurons following acoustic trauma remains to be elucidated. The present study investigated how auditory cortical PV neurons change following unilateral 1 hour noise exposure (left ear, one octave band noise centered at 16 kHz, 116 dB SPL. Noise exposure elevated the auditory brainstem response threshold of the exposed ear when examined 7 days later. More detectable PV neurons were observed in both sides of the auditory cortex of noise-exposed rats when compared to control. The detectable PV neurons of the left auditory cortex (ipsilateral to the exposed ear to noise exposure outnumbered those of the right auditory cortex (contralateral to the exposed ear. Quantification of Western blotted bands revealed higher expression level of PV protein in the left cortex. These findings of more active PV neurons in noise-exposed rats suggested that a compensatory mechanism might be initiated to maintain a stable state of the brain.

  19. Development of sub-100 femtosecond timing and synchronization system.

    Science.gov (United States)

    Lin, Zhenyang; Du, Yingchao; Yang, Jin; Xu, Yilun; Yan, Lixin; Huang, Wenhui; Tang, Chuanxiang; Huang, Gang; Du, Qiang; Doolittle, Lawrence; Wilcox, Russell; Byrd, John

    2018-01-01

    The precise timing and synchronization system is an essential part for the ultra-fast electron and X-ray sources based on the photocathode injector where strict synchronization among RF, laser, and beams are required. In this paper, we present an integrated sub-100 femtosecond timing and synchronization system developed and demonstrated recently in Tsinghua University based on the collaboration with Lawrence Berkeley National Lab. The timing and synchronization system includes the fiber-based CW carrier phase reference distribution system for delivering stabilized RF phase reference to multiple receiver clients, the Low Level RF (LLRF) control system to monitor and generate the phase and amplitude controllable pulse RF signal, and the laser-RF synchronization system for high precision synchronization between optical and RF signals. Each subsystem is characterized by its blocking structure and is also expansible. A novel asymmetric calibration sideband signal method was proposed for eliminating the non-linear distortion in the optical synchronization process. According to offline and online tests, the system can deliver a stable signal to each client and suppress the drift and jitter of the RF signal for the accelerator and the laser oscillator to less than 100 fs RMS (∼0.1° in 2856 MHz frequency). Moreover, a demo system with a LLRF client and a laser-RF synchronization client is deployed and operated successfully at the Tsinghua Thomson scattering X-ray source. The beam-based jitter measurement experiments have been conducted to evaluate the overall performance of the system, and the jitter sources are discussed.

  20. Multisensory stimuli improve relative localisation judgments compared to unisensory auditory or visual stimuli

    OpenAIRE

    Bizley, Jennifer; Wood, Katherine; Freeman, Laura

    2018-01-01

    Observers performed a relative localisation task in which they reported whether the second of two sequentially presented signals occurred to the left or right of the first. Stimuli were detectability-matched auditory, visual, or auditory-visual signals and the goal was to compare changes in performance with eccentricity across modalities. Visual performance was superior to auditory at the midline, but inferior in the periphery, while auditory-visual performance exceeded both at all locations....

  1. Auditory Tones and Foot-Shock Recapitulate Spontaneous Sub-Threshold Activity in Basolateral Amygdala Principal Neurons and Interneurons.

    Directory of Open Access Journals (Sweden)

    François Windels

    Full Text Available In quiescent states such as anesthesia and slow wave sleep, cortical networks show slow rhythmic synchronized activity. In sensory cortices this rhythmic activity shows a stereotypical pattern that is recapitulated by stimulation of the appropriate sensory modality. The amygdala receives sensory input from a variety of sources, and in anesthetized animals, neurons in the basolateral amygdala (BLA show slow rhythmic synchronized activity. Extracellular field potential recordings show that these oscillations are synchronized with sensory cortex and the thalamus, with both the thalamus and cortex leading the BLA. Using whole-cell recording in vivo we show that the membrane potential of principal neurons spontaneously oscillates between up- and down-states. Footshock and auditory stimulation delivered during down-states evokes an up-state that fully recapitulates those occurring spontaneously. These results suggest that neurons in the BLA receive convergent input from networks of cortical neurons with slow oscillatory activity and that somatosensory and auditory stimulation can trigger activity in these same networks.

  2. Observer-based hyperchaos synchronization in cascaded discrete-time systems

    International Nuclear Information System (INIS)

    Grassi, Giuseppe

    2009-01-01

    This paper deals with the observer-based synchronization in a cascade connection of hyperchaotic discrete-time systems. The paper demonstrates that exact synchronization in finite time is achievable between pairs of drive-response systems using only a scalar synchronizing signal. This 'propagated synchronization' starts from the innermost drive-response system pair and propagates toward the outermost drive-system pair. Choosing the drive-system input to be an information signal (encrypted via an arbitrary encryption function) yields a potential application of this architecture in chaos-based communications.

  3. A loudspeaker-based room auralization system for auditory perception research

    DEFF Research Database (Denmark)

    Buchholz, Jörg; Favrot, Sylvain Emmanuel

    2009-01-01

    Most research on basic auditory function has been conducted in anechoic or almost anechoic environments. The knowledge derived from these experiments cannot directly be transferred to reverberant environments. In order to investigate the auditory signal processing of reverberant sounds....... This system provides a flexible research platform for conducting auditory experiments with normal-hearing, hearing-impaired, and aided hearing-impaired listeners in a fully controlled and realistic environment. This includes measures of basic auditory function (e.g., signal detection, distance perception......) and measures of speech intelligibility. A battery of objective tests (e.g., reverberation time, clarity, interaural correlation coefficient) and subjective tests (e.g., speech reception thresholds) is presented that demonstrates the applicability of the LoRA system....

  4. Trigger delay compensation of beam synchronous sampling

    International Nuclear Information System (INIS)

    Steimel, J.

    1996-05-01

    One of the problems of providing beam feedback in a large accelerator is the lack of beam synchronous trigger signals far from the RF signal source. IF single bucket resolutions are required, a cable extending from the RF source to the other side of the accelerator will not provide a synchronous signal if the RF frequency changes significantly with respect to the cable delay. This paper offers a solution to this problem by locking to the RF, at the remote location, using a digital phase locked loop. Then, the digitized frequency value is used to calculate the phase shift required to remain synchronized to the beam. Results are shown for phase lock to the Fermilab Main Ring RF. 1 ref., 4 figs

  5. Synchronization of coupled nonidentical multidelay feedback systems

    International Nuclear Information System (INIS)

    Hoang, Thang Manh; Nakagawa, Masahiro

    2007-01-01

    We present the lag synchronization of coupled nonidentical multidelay feedback systems, in which the synchronization signal is the sum of nonlinearly transformed components of delayed state variable. The sufficient condition for synchronization is considered by the Krasovskii-Lyapunov theory. The specific examples will demonstrate and verify the effectiveness of the proposed model

  6. Facilitated auditory detection for speech sounds

    Directory of Open Access Journals (Sweden)

    Carine eSignoret

    2011-07-01

    Full Text Available If it is well known that knowledge facilitates higher cognitive functions, such as visual and auditory word recognition, little is known about the influence of knowledge on detection, particularly in the auditory modality. Our study tested the influence of phonological and lexical knowledge on auditory detection. Words, pseudo words and complex non phonological sounds, energetically matched as closely as possible, were presented at a range of presentation levels from sub threshold to clearly audible. The participants performed a detection task (Experiments 1 and 2 that was followed by a two alternative forced choice recognition task in Experiment 2. The results of this second task in Experiment 2 suggest a correct recognition of words in the absence of detection with a subjective threshold approach. In the detection task of both experiments, phonological stimuli (words and pseudo words were better detected than non phonological stimuli (complex sounds, presented close to the auditory threshold. This finding suggests an advantage of speech for signal detection. An additional advantage of words over pseudo words was observed in Experiment 2, suggesting that lexical knowledge could also improve auditory detection when listeners had to recognize the stimulus in a subsequent task. Two simulations of detection performance performed on the sound signals confirmed that the advantage of speech over non speech processing could not be attributed to energetic differences in the stimuli.

  7. Optical Synchronization of a 10-G Ethernet Packet and Time-Division Multiplexing to a 50-Gb/s Signal Using an Optical Time Lens

    DEFF Research Database (Denmark)

    Hu, Hao; Laguardia Areal, Janaina; Palushani, Evarist

    2010-01-01

    A 10-G Ethernet packet with maximum packet size of 1518 bytes is synchronized to a master clock with 200-kHz frequency offset using a time lens. The input 10-Gb/s non-return-to-zero packet is at the same time converted into a return-to-zero (RZ) packet with a pulsewidth of 10 ps and then time......-division multiplexed with four 10-Gb/s optical time-division-multiplexing (OTDM) channels, thus constituting a 50-Gb/s OTDM serial signal. Error-free performances of the synchronized RZ packet and demultiplexed packet from the aggregated 50-Gb/s OTDM signal are achieved....

  8. Detecting modulated signals in modulated noise: (II) neural thresholds in the songbird forebrain.

    Science.gov (United States)

    Bee, Mark A; Buschermöhle, Michael; Klump, Georg M

    2007-10-01

    Sounds in the real world fluctuate in amplitude. The vertebrate auditory system exploits patterns of amplitude fluctuations to improve signal detection in noise. One experimental paradigm demonstrating these general effects has been used in psychophysical studies of 'comodulation detection difference' (CDD). The CDD effect refers to the fact that thresholds for detecting a modulated, narrowband noise signal are lower when the envelopes of flanking bands of modulated noise are comodulated with each other, but fluctuate independently of the signal compared with conditions in which the envelopes of the signal and flanking bands are all comodulated. Here, we report results from a study of the neural correlates of CDD in European starlings (Sturnus vulgaris). We manipulated: (i) the envelope correlations between a narrowband noise signal and a masker comprised of six flanking bands of noise; (ii) the signal onset delay relative to masker onset; (iii) signal duration; and (iv) masker spectrum level. Masked detection thresholds were determined from neural responses using signal detection theory. Across conditions, the magnitude of neural CDD ranged between 2 and 8 dB, which is similar to that reported in a companion psychophysical study of starlings [U. Langemann & G.M. Klump (2007) Eur. J. Neurosci., 26, 1969-1978]. We found little evidence to suggest that neural CDD resulted from the across-channel processing of auditory grouping cues related to common envelope fluctuations and synchronous onsets between the signal and flanking bands. We discuss a within-channel model of peripheral processing that explains many of our results.

  9. Suppressed visual looming stimuli are not integrated with auditory looming signals: Evidence from continuous flash suppression.

    Science.gov (United States)

    Moors, Pieter; Huygelier, Hanne; Wagemans, Johan; de-Wit, Lee; van Ee, Raymond

    2015-01-01

    Previous studies using binocular rivalry have shown that signals in a modality other than the visual can bias dominance durations depending on their congruency with the rivaling stimuli. More recently, studies using continuous flash suppression (CFS) have reported that multisensory integration influences how long visual stimuli remain suppressed. In this study, using CFS, we examined whether the contrast thresholds for detecting visual looming stimuli are influenced by a congruent auditory stimulus. In Experiment 1, we show that a looming visual stimulus can result in lower detection thresholds compared to a static concentric grating, but that auditory tone pips congruent with the looming stimulus did not lower suppression thresholds any further. In Experiments 2, 3, and 4, we again observed no advantage for congruent multisensory stimuli. These results add to our understanding of the conditions under which multisensory integration is possible, and suggest that certain forms of multisensory integration are not evident when the visual stimulus is suppressed from awareness using CFS.

  10. Experiment on Synchronous Timing Signal Detection from ISDB-T Terrestrial Digital TV Signal with Application to Autonomous Distributed ITS-IVC Network

    Science.gov (United States)

    Karasawa, Yoshio; Kumagai, Taichi; Takemoto, Atsushi; Fujii, Takeo; Ito, Kenji; Suzuki, Noriyoshi

    A novel timing synchronizing scheme is proposed for use in inter-vehicle communication (IVC) with an autonomous distributed intelligent transport system (ITS). The scheme determines the timing of packet signal transmission in the IVC network and employs the guard interval (GI) timing in the orthogonal frequency divisional multiplexing (OFDM) signal currently used for terrestrial broadcasts in the Japanese digital television system (ISDB-T). This signal is used because it is expected that the automotive market will demand the capability for cars to receive terrestrial digital TV broadcasts in the near future. The use of broadcasts by automobiles presupposes that the on-board receivers are capable of accurately detecting the GI timing data in an extremely low carrier-to-noise ratio (CNR) condition regardless of a severe multipath environment which will introduce broad scatter in signal arrival times. Therefore, we analyzed actual broadcast signals received in a moving vehicle in a field experiment and showed that the GI timing signal is detected with the desired accuracy even in the case of extremely low-CNR environments. Some considerations were also given about how to use these findings.

  11. Combinatorial Modulation of Signaling Pathways Reveals Cell-Type-Specific Requirements for Highly Efficient and Synchronous iPSC Reprogramming

    Directory of Open Access Journals (Sweden)

    Simon E. Vidal

    2014-10-01

    Full Text Available The differentiated state of somatic cells provides barriers for the derivation of induced pluripotent stem cells (iPSCs. To address why some cell types reprogram more readily than others, we studied the effect of combined modulation of cellular signaling pathways. Surprisingly, inhibition of transforming growth factor β (TGF-β together with activation of Wnt signaling in the presence of ascorbic acid allows >80% of murine fibroblasts to acquire pluripotency after 1 week of reprogramming factor expression. In contrast, hepatic and blood progenitors predominantly required only TGF-β inhibition or canonical Wnt activation, respectively, to reprogram at efficiencies approaching 100%. Strikingly, blood progenitors reactivated endogenous pluripotency loci in a highly synchronous manner, and we demonstrate that expression of specific chromatin-modifying enzymes and reduced TGF-β/mitogen-activated protein (MAP kinase activity are intrinsic properties associated with the unique reprogramming response of these cells. Our observations define cell-type-specific requirements for the rapid and synchronous reprogramming of somatic cells.

  12. Second-Order Optimal Array Receivers for Synchronization of BPSK, MSK, and GMSK Signals Corrupted by Noncircular Interferences

    Directory of Open Access Journals (Sweden)

    Chevalier Pascal

    2007-01-01

    Full Text Available The synchronization and/or time acquisition problem in the presence of interferences has been strongly studied these last two decades, mainly to mitigate the multiple access interferences from other users in DS/CDMA systems. Among the available receivers, only some scarce receivers may also be used in other contexts such as F/TDMA systems. However, these receivers assume implicitly or explicitly circular (or proper interferences and become suboptimal for noncircular (or improper interferences. Such interferences are characteristic in particular of radio communication networks using either rectilinear (or monodimensional modulations such as BPSK modulation or modulation becoming quasirectilinear after a preprocessing such as MSK, GMSK, or OQAM modulations. For this reason, the purpose of this paper is to introduce and to analyze the performance of second-optimal array receivers for synchronization and/or time acquisition of BPSK, MSK, and GMSK signals corrupted by noncircular interferences. For given performances and in the presence of rectilinear signal and interferences, the proposed receiver allows a reduction of the number of sensors by a factor at least equal to two.

  13. You can't stop the music: reduced auditory alpha power and coupling between auditory and memory regions facilitate the illusory perception of music during noise.

    Science.gov (United States)

    Müller, Nadia; Keil, Julian; Obleser, Jonas; Schulz, Hannah; Grunwald, Thomas; Bernays, René-Ludwig; Huppertz, Hans-Jürgen; Weisz, Nathan

    2013-10-01

    Our brain has the capacity of providing an experience of hearing even in the absence of auditory stimulation. This can be seen as illusory conscious perception. While increasing evidence postulates that conscious perception requires specific brain states that systematically relate to specific patterns of oscillatory activity, the relationship between auditory illusions and oscillatory activity remains mostly unexplained. To investigate this we recorded brain activity with magnetoencephalography and collected intracranial data from epilepsy patients while participants listened to familiar as well as unknown music that was partly replaced by sections of pink noise. We hypothesized that participants have a stronger experience of hearing music throughout noise when the noise sections are embedded in familiar compared to unfamiliar music. This was supported by the behavioral results showing that participants rated the perception of music during noise as stronger when noise was presented in a familiar context. Time-frequency data show that the illusory perception of music is associated with a decrease in auditory alpha power pointing to increased auditory cortex excitability. Furthermore, the right auditory cortex is concurrently synchronized with the medial temporal lobe, putatively mediating memory aspects associated with the music illusion. We thus assume that neuronal activity in the highly excitable auditory cortex is shaped through extensive communication between the auditory cortex and the medial temporal lobe, thereby generating the illusion of hearing music during noise. Copyright © 2013 Elsevier Inc. All rights reserved.

  14. Auditory and Visual Sensations

    CERN Document Server

    Ando, Yoichi

    2010-01-01

    Professor Yoichi Ando, acoustic architectural designer of the Kirishima International Concert Hall in Japan, presents a comprehensive rational-scientific approach to designing performance spaces. His theory is based on systematic psychoacoustical observations of spatial hearing and listener preferences, whose neuronal correlates are observed in the neurophysiology of the human brain. A correlation-based model of neuronal signal processing in the central auditory system is proposed in which temporal sensations (pitch, timbre, loudness, duration) are represented by an internal autocorrelation representation, and spatial sensations (sound location, size, diffuseness related to envelopment) are represented by an internal interaural crosscorrelation function. Together these two internal central auditory representations account for the basic auditory qualities that are relevant for listening to music and speech in indoor performance spaces. Observed psychological and neurophysiological commonalities between auditor...

  15. Shaping the aging brain: Role of auditory input patterns in the emergence of auditory cortical impairments

    Directory of Open Access Journals (Sweden)

    Brishna Soraya Kamal

    2013-09-01

    Full Text Available Age-related impairments in the primary auditory cortex (A1 include poor tuning selectivity, neural desynchronization and degraded responses to low-probability sounds. These changes have been largely attributed to reduced inhibition in the aged brain, and are thought to contribute to substantial hearing impairment in both humans and animals. Since many of these changes can be partially reversed with auditory training, it has been speculated that they might not be purely degenerative, but might rather represent negative plastic adjustments to noisy or distorted auditory signals reaching the brain. To test this hypothesis, we examined the impact of exposing young adult rats to 8 weeks of low-grade broadband noise on several aspects of A1 function and structure. We then characterized the same A1 elements in aging rats for comparison. We found that the impact of noise exposure on A1 tuning selectivity, temporal processing of auditory signal and responses to oddball tones was almost indistinguishable from the effect of natural aging. Moreover, noise exposure resulted in a reduction in the population of parvalbumin inhibitory interneurons and cortical myelin as previously documented in the aged group. Most of these changes reversed after returning the rats to a quiet environment. These results support the hypothesis that age-related changes in A1 have a strong activity-dependent component and indicate that the presence or absence of clear auditory input patterns might be a key factor in sustaining adult A1 function.

  16. Demodulation Processes in Auditory Perception

    National Research Council Canada - National Science Library

    Feth, Lawrence

    1997-01-01

    The long range goal of this project was the understanding of human auditory processing of information conveyed by complex, time varying signals such as speech, music or important environmental sounds...

  17. Application of «Sensor signal analysis network» complex for distributed, time synchronized analysis of electromagnetic radiation

    Science.gov (United States)

    Mochalov, Vladimir; Mochalova, Anastasia

    2017-10-01

    The paper considers a developing software-hardware complex «Sensor signal analysis network» for distributed and time synchronized analysis of electromagnetic radiations. The areas of application and the main features of the complex are described. An example of application of the complex to monitor natural electromagnetic radiation sources is considered based on the data recorded in VLF range. A generalized functional scheme of stream analysis of signals by a complex functional node is suggested and its application for stream detection of atmospherics, whistlers and tweaks is considered.

  18. Analysis of the Auditory Feedback and Phonation in Normal Voices.

    Science.gov (United States)

    Arbeiter, Mareike; Petermann, Simon; Hoppe, Ulrich; Bohr, Christopher; Doellinger, Michael; Ziethe, Anke

    2018-02-01

    The aim of this study was to investigate the auditory feedback mechanisms and voice quality during phonation in response to a spontaneous pitch change in the auditory feedback. Does the pitch shift reflex (PSR) change voice pitch and voice quality? Quantitative and qualitative voice characteristics were analyzed during the PSR. Twenty-eight healthy subjects underwent transnasal high-speed video endoscopy (HSV) at 8000 fps during sustained phonation [a]. While phonating, the subjects heard their sound pitched up for 700 cents (interval of a fifth), lasting 300 milliseconds in their auditory feedback. The electroencephalography (EEG), acoustic voice signal, electroglottography (EGG), and high-speed-videoendoscopy (HSV) were analyzed to compare feedback mechanisms for the pitched and unpitched condition of the phonation paradigm statistically. Furthermore, quantitative and qualitative voice characteristics were analyzed. The PSR was successfully detected within all signals of the experimental tools (EEG, EGG, acoustic voice signal, HSV). A significant increase of the perturbation measures and an increase of the values of the acoustic parameters during the PSR were observed, especially for the audio signal. The auditory feedback mechanism seems not only to control for voice pitch but also for voice quality aspects.

  19. Magnetic resonance imaging of the internal auditory canal

    International Nuclear Information System (INIS)

    Daniels, D.L.; Herfkins, R.; Koehler, P.R.; Millen, S.J.; Shaffer, K.A.; Williams, A.L.; Haughton, V.M.

    1984-01-01

    Three patients with exclusively or predominantly intracanalicular neuromas and 5 with presumably normal internal auditory canals were examined with prototype 1.4- or 1.5-tesla magnetic resonance (MR) scanners. MR images showed the 7th and 8th cranial nerves in the internal auditory canal. The intracanalicular neuromas had larger diameter and slightly greater signal strength than the nerves. Early results suggest that minimal enlargement of the nerves can be detected even in the internal auditory canal

  20. Theta-alpha EEG phase distributions in the frontal area for dissociation of visual and auditory working memory.

    Science.gov (United States)

    Akiyama, Masakazu; Tero, Atsushi; Kawasaki, Masahiro; Nishiura, Yasumasa; Yamaguchi, Yoko

    2017-03-07

    Working memory (WM) is known to be associated with synchronization of the theta and alpha bands observed in electroencephalograms (EEGs). Although frontal-posterior global theta synchronization appears in modality-specific WM, local theta synchronization in frontal regions has been found in modality-independent WM. How frontal theta oscillations separately synchronize with task-relevant sensory brain areas remains an open question. Here, we focused on theta-alpha phase relationships in frontal areas using EEG, and then verified their functional roles with mathematical models. EEG data showed that the relationship between theta (6 Hz) and alpha (12 Hz) phases in the frontal areas was about 1:2 during both auditory and visual WM, and that the phase distributions between auditory and visual WM were different. Next, we used the differences in phase distributions to construct FitzHugh-Nagumo type mathematical models. The results replicated the modality-specific branching by orthogonally of the trigonometric functions for theta and alpha oscillations. Furthermore, mathematical and experimental results were consistent with regards to the phase relationships and amplitudes observed in frontal and sensory areas. These results indicate the important role that different phase distributions of theta and alpha oscillations have in modality-specific dissociation in the brain.

  1. Temporal Integration of Auditory Stimulation and Binocular Disparity Signals

    Directory of Open Access Journals (Sweden)

    Marina Zannoli

    2011-10-01

    Full Text Available Several studies using visual objects defined by luminance have reported that the auditory event must be presented 30 to 40 ms after the visual stimulus to perceive audiovisual synchrony. In the present study, we used visual objects defined only by their binocular disparity. We measured the optimal latency between visual and auditory stimuli for the perception of synchrony using a method introduced by Moutoussis & Zeki (1997. Visual stimuli were defined either by luminance and disparity or by disparity only. They moved either back and forth between 6 and 12 arcmin or from left to right at a constant disparity of 9 arcmin. This visual modulation was presented together with an amplitude-modulated 500 Hz tone. Both modulations were sinusoidal (frequency: 0.7 Hz. We found no difference between 2D and 3D motion for luminance stimuli: a 40 ms auditory lag was necessary for perceived synchrony. Surprisingly, even though stereopsis is often thought to be slow, we found a similar optimal latency in the disparity 3D motion condition (55 ms. However, when participants had to judge simultaneity for disparity 2D motion stimuli, it led to larger latencies (170 ms, suggesting that stereo motion detectors are poorly suited to track 2D motion.

  2. Joint Iterative Carrier Synchronization and Signal Detection Employing Expectation Maximization

    DEFF Research Database (Denmark)

    Zibar, Darko; de Carvalho, Luis Henrique Hecker; Estaran Tolosa, Jose Manuel

    2014-01-01

    and nonlinear phase noise, compared to digital phase-locked loop (PLL) followed by hard decisions. Additionally, soft decision driven joint carrier synchronization and detection offers an improvement of 0.5 dB in terms of input power compared to hard decision digital PLL based carrier synchronization...

  3. A loudspeaker-based room auralisation (LoRA) system for auditory perception research

    DEFF Research Database (Denmark)

    Buchholz, Jörg; Favrot, Sylvain Emmanuel

    Most research on understanding the signal processing of the auditory system has been realized in anechoic or almost anechoic environments. The knowledge derived from these experiments cannot be directly transferred to reverberant environments. In order to investigate the auditory signal processing...... are utilized to realise highly authentic room reverberation. This system aims at providing a flexible research platform for conducting auditory experiments with normal-hearing, hearing-impaired, and aided hearing-impaired listeners in a fully controlled and realistic environment. An overall description...

  4. Three types of generalized synchronization

    Energy Technology Data Exchange (ETDEWEB)

    Yang Junzhong [School of Science, Beijing University of Posts and Telecomunications, Beijing 100876 (China)]. E-mail: jzyang@bupt.edu.cn; Hu Gang [China Center for Advanced Science and Technology (CCAST) (World Laboratory), PO Box 8730, Beijing 100080 (China) and Department of Physics, Beijing Normal University, Beijing 100875 (China)]. E-mail: ganghu@bnu.edu.cn

    2007-02-05

    The roles played by drive and response systems on generalized chaos synchronization (GS) are studied. And the generalized synchronization is classified, based on these roles, to three distinctive types: the passive GS which is mainly determined by the response system and insensitive to the driving signal; the resonant GS where phase synchronization between the drive and response systems is preceding GS; and the interacting GS where both the drive and response have influences on the status of GS. The features of these GS types and the possible changes from one types to others are investigated.

  5. Three types of generalized synchronization

    International Nuclear Information System (INIS)

    Yang Junzhong; Hu Gang

    2007-01-01

    The roles played by drive and response systems on generalized chaos synchronization (GS) are studied. And the generalized synchronization is classified, based on these roles, to three distinctive types: the passive GS which is mainly determined by the response system and insensitive to the driving signal; the resonant GS where phase synchronization between the drive and response systems is preceding GS; and the interacting GS where both the drive and response have influences on the status of GS. The features of these GS types and the possible changes from one types to others are investigated

  6. Binaural beats increase interhemispheric alpha-band coherence between auditory cortices.

    Science.gov (United States)

    Solcà, Marco; Mottaz, Anaïs; Guggisberg, Adrian G

    2016-02-01

    Binaural beats (BBs) are an auditory illusion occurring when two tones of slightly different frequency are presented separately to each ear. BBs have been suggested to alter physiological and cognitive processes through synchronization of the brain hemispheres. To test this, we recorded electroencephalograms (EEG) at rest and while participants listened to BBs or a monaural control condition during which both tones were presented to both ears. We calculated for each condition the interhemispheric coherence, which expressed the synchrony between neural oscillations of both hemispheres. Compared to monaural beats and resting state, BBs enhanced interhemispheric coherence between the auditory cortices. Beat frequencies in the alpha (10 Hz) and theta (4 Hz) frequency range both increased interhemispheric coherence selectively at alpha frequencies. In a second experiment, we evaluated whether this coherence increase has a behavioral aftereffect on binaural listening. No effects were observed in a dichotic digit task performed immediately after BBs presentation. Our results suggest that BBs enhance alpha-band oscillation synchrony between the auditory cortices during auditory stimulation. This effect seems to reflect binaural integration rather than entrainment. Copyright © 2015 Elsevier B.V. All rights reserved.

  7. Neural oscillations in auditory working memory

    OpenAIRE

    Wilsch, A.

    2015-01-01

    The present thesis investigated memory load and memory decay in auditory working memory. Alpha power as a marker for memory load served as the primary indicator for load and decay fluctuations hypothetically reflecting functional inhibition of irrelevant information. Memory load was induced by presenting auditory signals (syllables and pure-tone sequences) in noise because speech-in-noise has been shown before to increase memory load. The aim of the thesis was to assess with magnetoencephalog...

  8. Robust Frame Synchronization for Low Signal-to-Noise Ratio Channels Using Energy-Corrected Differential Correlation

    Directory of Open Access Journals (Sweden)

    Kim Pansoo

    2009-01-01

    Full Text Available Recent standards for wireless transmission require reliable synchronization for channels with low signal-to-noise ratio (SNR as well as with a large amount of frequency offset, which necessitates a robust correlator structure for the initial frame synchronization process. In this paper, a new correlation strategy especially targeted for low SNR regions is proposed and its performance is analyzed. By utilizing a modified energy correction term, the proposed method effectively reduces the variance of the decision variable to enhance the detection performance. Most importantly, the method is demonstrated to outperform all previously reported schemes by a significant margin, for SNRs below 5 dB regardless of the existence of the frequency offsets. A variation of the proposed method is also presented for further enhancement over the channels with small frequency errors. The particular application considered for the performance verification is the second generation digital video broadcasting system for satellites (DVB-S2.

  9. Auditory prediction during speaking and listening.

    Science.gov (United States)

    Sato, Marc; Shiller, Douglas M

    2018-02-02

    In the present EEG study, the role of auditory prediction in speech was explored through the comparison of auditory cortical responses during active speaking and passive listening to the same acoustic speech signals. Two manipulations of sensory prediction accuracy were used during the speaking task: (1) a real-time change in vowel F1 feedback (reducing prediction accuracy relative to unaltered feedback) and (2) presenting a stable auditory target rather than a visual cue to speak (enhancing auditory prediction accuracy during baseline productions, and potentially enhancing the perturbing effect of altered feedback). While subjects compensated for the F1 manipulation, no difference between the auditory-cue and visual-cue conditions were found. Under visually-cued conditions, reduced N1/P2 amplitude was observed during speaking vs. listening, reflecting a motor-to-sensory prediction. In addition, a significant correlation was observed between the magnitude of behavioral compensatory F1 response and the magnitude of this speaking induced suppression (SIS) for P2 during the altered auditory feedback phase, where a stronger compensatory decrease in F1 was associated with a stronger the SIS effect. Finally, under the auditory-cued condition, an auditory repetition-suppression effect was observed in N1/P2 amplitude during the listening task but not active speaking, suggesting that auditory predictive processes during speaking and passive listening are functionally distinct. Copyright © 2018 Elsevier Inc. All rights reserved.

  10. Dependence of synchronization transitions on mean field approach in two-way coupled neural system

    Science.gov (United States)

    Shi, J. C.; Luo, M.; Huang, C. S.

    2018-03-01

    This work investigates the synchronization transitions in two-way coupled neural system by mean field approach. Results show that, there exists a critical noise intensity for the synchronization transitions, i.e., above (or below) the critical noise intensity, the synchronization transitions are decreased (or hardly change) with increasing the noise intensity. Meanwhile, the heterogeneity effect plays a negative role for the synchronization transitions, and above critical coupling strength, the heterogeneity effect on synchronization transitions can be negligible. Furthermore, when an external signal is introduced into the coupled system, the novel frequency-induced and amplitude-induced synchronization transitions are found, and there exist an optimal frequency and an optimal amplitude of external signal which makes the system to display the best synchronization transitions. In particular, it is observed that the synchronization transitions can not be further affected above critical frequency of external signal.

  11. Age-related changes in mitochondrial antioxidant enzyme Trx2 and TXNIP-Trx2-ASK1 signal pathways in the auditory cortex of a mimetic aging rat model: changes to Trx2 in the auditory cortex.

    Science.gov (United States)

    Sun, Hai-Ying; Hu, Yu-Juan; Zhao, Xue-Yan; Zhong, Yi; Zeng, Ling-Ling; Chen, Xu-Bo; Yuan, Jie; Wu, Jing; Sun, Yu; Kong, Wen; Kong, Wei-Jia

    2015-07-01

    Age-associated degeneration in the central auditory system, which is defined as central presbycusis, can impair sound localization and speech perception. Research has shown that oxidative stress plays a central role in the pathological process of central presbycusis. Thioredoxin 2 (Trx2), one member of thioredoxin family, plays a key role in regulating the homeostasis of cellular reactive oxygen species and anti-apoptosis. The purpose of this study was to explore the association between Trx2 and the phenotype of central presbycusis using a mimetic aging animal model induced by long-term exposure to d-galactose (d-Gal). We also explored changes in thioredoxin-interacting protein (TXNIP), apoptosis signal regulating kinase 1 (ASK1) and phosphorylated ASK1 (p-ASK1) expression, as well as the Trx2-TXNIP/Trx2-ASK1 binding complex in the auditory cortex of mimetic aging rats. Our results demonstrate that, compared with control groups, the levels of Trx2 and Trx2-ASK1 binding complex were significantly reduced, whereas TXNIP, ASK1 p-ASK1 expression, and Trx2-TXNIP binding complex were significantly increased in the auditory cortex of the mimetic aging groups. Our results indicated that changes in Trx2 and the TXNIP-Trx2-ASK1 signal pathway may participate in the pathogenesis of central presbycusis. © 2015 FEBS.

  12. Across frequency processes involved in auditory detection of coloration

    DEFF Research Database (Denmark)

    Buchholz, Jörg; Kerketsos, P

    2008-01-01

    filterbank was designed to approximate auditory filter-shapes measured by Oxenham and Shera [JARO, 2003, 541-554], derived from forward masking data. The results of the present study demonstrate that a “purely” spectrum-based model approach can successfully describe auditory coloration detection even at high......When an early wall reflection is added to a direct sound, a spectral modulation is introduced to the signal's power spectrum. This spectral modulation typically produces an auditory sensation of coloration or pitch. Throughout this study, auditory spectral-integration effects involved in coloration...... detection are investigated. Coloration detection thresholds were therefore measured as a function of reflection delay and stimulus bandwidth. In order to investigate the involved auditory mechanisms, an auditory model was employed that was conceptually similar to the peripheral weighting model [Yost, JASA...

  13. Towards an auditory account of speech rhythm: application of a model of the auditory 'primal sketch' to two multi-language corpora.

    Science.gov (United States)

    Lee, Christopher S; Todd, Neil P McAngus

    2004-10-01

    The world's languages display important differences in their rhythmic organization; most particularly, different languages seem to privilege different phonological units (mora, syllable, or stress foot) as their basic rhythmic unit. There is now considerable evidence that such differences have important consequences for crucial aspects of language acquisition and processing. Several questions remain, however, as to what exactly characterizes the rhythmic differences, how they are manifested at an auditory/acoustic level and how listeners, whether adult native speakers or young infants, process rhythmic information. In this paper it is proposed that the crucial determinant of rhythmic organization is the variability in the auditory prominence of phonetic events. In order to test this auditory prominence hypothesis, an auditory model is run on two multi-language data-sets, the first consisting of matched pairs of English and French sentences, and the second consisting of French, Italian, English and Dutch sentences. The model is based on a theory of the auditory primal sketch, and generates a primitive representation of an acoustic signal (the rhythmogram) which yields a crude segmentation of the speech signal and assigns prominence values to the obtained sequence of events. Its performance is compared with that of several recently proposed phonetic measures of vocalic and consonantal variability.

  14. Audio-visual synchronization in reading while listening to texts: Effects on visual behavior and verbal learning

    OpenAIRE

    Gerbier , Emilie; Bailly , Gérard; Bosse , Marie-Line

    2018-01-01

    International audience; Reading while listening to texts (RWL) is a promising way to improve the learning benefits provided by a reading experience. In an exploratory study, we investigated the effect of synchronizing the highlighting of words (visual) with their auditory (speech) counterpart during a RWL task. Forty French children from 3rd to 5th grade read short stories in their native language while hearing the story spoken by a narrator. In the non-synchronized (S-) condition the text wa...

  15. Visuo-perceptual capabilities predict sensitivity for coinciding auditory and visual transients in multi-element displays.

    Science.gov (United States)

    Meyerhoff, Hauke S; Gehrer, Nina A

    2017-01-01

    In order to obtain a coherent representation of the outside world, auditory and visual information are integrated during human information processing. There is remarkable variance among observers in the capability to integrate auditory and visual information. Here, we propose that visuo-perceptual capabilities predict detection performance for audiovisually coinciding transients in multi-element displays due to severe capacity limitations in audiovisual integration. In the reported experiment, we employed an individual differences approach in order to investigate this hypothesis. Therefore, we measured performance in a useful-field-of-view task that captures detection performance for briefly presented stimuli across a large perceptual field. Furthermore, we measured sensitivity for visual direction changes that coincide with tones within the same participants. Our results show that individual differences in visuo-perceptual capabilities predicted sensitivity for the presence of audiovisually synchronous events among competing visual stimuli. To ensure that this correlation does not stem from superordinate factors, we also tested performance in an unrelated working memory task. Performance in this task was independent of sensitivity for the presence of audiovisually synchronous events. Our findings strengthen the proposed link between visuo-perceptual capabilities and audiovisual integration. The results also suggest that basic visuo-perceptual capabilities provide the basis for the subsequent integration of auditory and visual information.

  16. Dynamic links between theta executive functions and alpha storage buffers in auditory and visual working memory.

    Science.gov (United States)

    Kawasaki, Masahiro; Kitajo, Keiichi; Yamaguchi, Yoko

    2010-05-01

    Working memory (WM) tasks require not only distinct functions such as a storage buffer and central executive functions, but also coordination among these functions. Neuroimaging studies have revealed the contributions of different brain regions to different functional roles in WM tasks; however, little is known about the neural mechanism governing their coordination. Electroencephalographic (EEG) rhythms, especially theta and alpha, are known to appear over distributed brain regions during WM tasks, but the rhythms associated with task-relevant regional coupling have not been obtained thus far. In this study, we conducted time-frequency analyses for EEG data in WM tasks that include manipulation periods and memory storage buffer periods. We used both auditory WM tasks and visual WM tasks. The results successfully demonstrated function-specific EEG activities. The frontal theta amplitudes increased during the manipulation periods of both tasks. The alpha amplitudes increased during not only the manipulation but also the maintenance periods in the temporal area for the auditory WM and the parietal area for the visual WM. The phase synchronization analyses indicated that, under the relevant task conditions, the temporal and parietal regions show enhanced phase synchronization in the theta bands with the frontal region, whereas phase synchronization between theta and alpha is significantly enhanced only within the individual areas. Our results suggest that WM task-relevant brain regions are coordinated by distant theta synchronization for central executive functions, by local alpha synchronization for the memory storage buffer, and by theta-alpha coupling for inter-functional integration.

  17. A Decline in Response Variability Improves Neural Signal Detection during Auditory Task Performance.

    Science.gov (United States)

    von Trapp, Gardiner; Buran, Bradley N; Sen, Kamal; Semple, Malcolm N; Sanes, Dan H

    2016-10-26

    The detection of a sensory stimulus arises from a significant change in neural activity, but a sensory neuron's response is rarely identical to successive presentations of the same stimulus. Large trial-to-trial variability would limit the central nervous system's ability to reliably detect a stimulus, presumably affecting perceptual performance. However, if response variability were to decrease while firing rate remained constant, then neural sensitivity could improve. Here, we asked whether engagement in an auditory detection task can modulate response variability, thereby increasing neural sensitivity. We recorded telemetrically from the core auditory cortex of gerbils, both while they engaged in an amplitude-modulation detection task and while they sat quietly listening to the identical stimuli. Using a signal detection theory framework, we found that neural sensitivity was improved during task performance, and this improvement was closely associated with a decrease in response variability. Moreover, units with the greatest change in response variability had absolute neural thresholds most closely aligned with simultaneously measured perceptual thresholds. Our findings suggest that the limitations imposed by response variability diminish during task performance, thereby improving the sensitivity of neural encoding and potentially leading to better perceptual sensitivity. The detection of a sensory stimulus arises from a significant change in neural activity. However, trial-to-trial variability of the neural response may limit perceptual performance. If the neural response to a stimulus is quite variable, then the response on a given trial could be confused with the pattern of neural activity generated when the stimulus is absent. Therefore, a neural mechanism that served to reduce response variability would allow for better stimulus detection. By recording from the cortex of freely moving animals engaged in an auditory detection task, we found that variability

  18. Control of phase synchronization of neuronal activity in the rat hippocampus.

    Science.gov (United States)

    Lian, Jun; Shuai, Jianwei; Durand, Dominique M

    2004-03-01

    Analysis of the synchronization mechanisms of neural activity is crucial to the understanding of the generation, propagation and control of epileptiform activity. Recently, phase synchronization (PS) analysis was applied to quantify the partial synchrony that exists in complex chaotic or noisy systems. In a previous study, we have shown that neural activity between two remotely located sites can be synchronized through a complete cut of the tissue by endogenous non-synaptic signals. Therefore, it should be possible to apply signals to control PS. In this study, we test the hypothesis that stimulation amplitudes below excitation level (sub-threshold) can be used to control phase synchronization of two neural signals and we investigate the underlying mechanisms. PS of neuronal activity is first analysed in two coupled Rossler neuron models. Both synchronization and desynchronization could be generated with sub-threshold sinusoidal stimulation. Phase synchronization was then studied in in vitro brain slices. Neuronal activity between two sites was modulated by the application of small sinusoidal electric fields. PS between two remote sites could be achieved by the application of two identical waveforms while phase desynchronization of two close sites was generated by the application of a stimulus at a single site. These results show that sub-threshold stimuli are able to phase synchronize or desynchronize two networks and suggest that small signals could play an important role in normal neural activity and epilepsy.

  19. Development of a Chirp Stimulus PC-Based Auditory Brainstem Response Audiometer

    Directory of Open Access Journals (Sweden)

    Ali AL-Afsaa

    2004-05-01

    Full Text Available Hearing losses during infancy and childhood have many negative future effects and impacts on the child life and productivity. The earlier detection of hearing losses, the earlier medical intervention and then the greater benefit of remediation will be. During this research a PC-based audiometer is designed and, currently, the audiometer prototype is in its final development steps. It is based on the auditory brainstem response (ABR method. Chirp stimuli instead of traditional click stimuli will be used to invoke the ABR signal. The stimulus is designed to synchronize the hair cells movement when it spreads out over the cochlea. In addition to the available hardware utilization (PC and PCI board, the efforts confined to design and implement a hardware prototype and to develop a software package that enables the system to behave as ABR audiometer. By using such a method and chirp stimulus, it is expected to be able to detect hearing impairment (sensorineural in the first few days of the life and conduct hearing test at low frequency of stimulus. Currently, the intended chirp stimulus has been successfully generated and the implemented module is able to amplify a signal (on the order of ABR signal to a recordable level. Moreover, a NI-DAQ data acquisition board has been chosen to implement the PC-prototype interface.

  20. Familiarity with music increases walking speed in rhythmic auditory cuing.

    Science.gov (United States)

    Leow, Li-Ann; Rinchon, Cricia; Grahn, Jessica

    2015-03-01

    Rhythmic auditory stimulation (RAS) is a gait rehabilitation method in which patients synchronize footsteps to a metronome or musical beats. Although RAS with music can ameliorate gait abnormalities, outcomes vary, possibly because music properties, such as groove or familiarity, differ across interventions. To optimize future interventions, we assessed how initially familiar and unfamiliar low-groove and high-groove music affected synchronization accuracy and gait in healthy individuals. We also experimentally increased music familiarity using repeated exposure to initially unfamiliar songs. Overall, familiar music elicited faster stride velocity and less variable strides, as well as better synchronization performance (matching of step tempo to beat tempo). High-groove music, as reported previously, led to faster stride velocity than low-groove music. We propose two mechanisms for familiarity's effects. First, familiarity with the beat structure reduces cognitive demands of synchronizing, leading to better synchronization performance and faster, less variable gait. Second, familiarity might have elicited faster gait by increasing enjoyment of the music, as enjoyment was higher after repeated exposure to initially low-enjoyment songs. Future studies are necessary to dissociate the contribution of these mechanisms to the observed RAS effects of familiar music on gait. © 2015 New York Academy of Sciences.

  1. Bursting synchronization in clustered neuronal networks

    International Nuclear Information System (INIS)

    Yu Hai-Tao; Wang Jiang; Deng Bin; Wei Xi-Le

    2013-01-01

    Neuronal networks in the brain exhibit the modular (clustered) property, i.e., they are composed of certain subnetworks with differential internal and external connectivity. We investigate bursting synchronization in a clustered neuronal network. A transition to mutual-phase synchronization takes place on the bursting time scale of coupled neurons, while on the spiking time scale, they behave asynchronously. This synchronization transition can be induced by the variations of inter- and intracoupling strengths, as well as the probability of random links between different subnetworks. Considering that some pathological conditions are related with the synchronization of bursting neurons in the brain, we analyze the control of bursting synchronization by using a time-periodic external signal in the clustered neuronal network. Simulation results show a frequency locking tongue in the driving parameter plane, where bursting synchronization is maintained, even in the presence of external driving. Hence, effective synchronization suppression can be realized with the driving parameters outside the frequency locking region. (interdisciplinary physics and related areas of science and technology)

  2. Bioacoustic Signal Classification in Cat Auditory Cortex

    Science.gov (United States)

    1994-01-01

    of the cat’s WINER. 1. A. Anatomy of layer IV in cat primary auditory cortex t4,1). J miedial geniculate body Ideintified by projections to binaural...34language" (see for example Tartter, 1986, chapter 8; and Lieberman, 1984). Attempts have been made to train animals (mainly apes, gorillas , _ _ ___I 3...gestures of a gorilla : Language acquisition in another Pongid. Brain and Language, 1978a, 5, 72-97. Patterson, F. Conversations with a gorilla

  3. Different auditory feedback control for echolocation and communication in horseshoe bats.

    Directory of Open Access Journals (Sweden)

    Ying Liu

    Full Text Available Auditory feedback from the animal's own voice is essential during bat echolocation: to optimize signal detection, bats continuously adjust various call parameters in response to changing echo signals. Auditory feedback seems also necessary for controlling many bat communication calls, although it remains unclear how auditory feedback control differs in echolocation and communication. We tackled this question by analyzing echolocation and communication in greater horseshoe bats, whose echolocation pulses are dominated by a constant frequency component that matches the frequency range they hear best. To maintain echoes within this "auditory fovea", horseshoe bats constantly adjust their echolocation call frequency depending on the frequency of the returning echo signal. This Doppler-shift compensation (DSC behavior represents one of the most precise forms of sensory-motor feedback known. We examined the variability of echolocation pulses emitted at rest (resting frequencies, RFs and one type of communication signal which resembles an echolocation pulse but is much shorter (short constant frequency communication calls, SCFs and produced only during social interactions. We found that while RFs varied from day to day, corroborating earlier studies in other constant frequency bats, SCF-frequencies remained unchanged. In addition, RFs overlapped for some bats whereas SCF-frequencies were always distinctly different. This indicates that auditory feedback during echolocation changed with varying RFs but remained constant or may have been absent during emission of SCF calls for communication. This fundamentally different feedback mechanism for echolocation and communication may have enabled these bats to use SCF calls for individual recognition whereas they adjusted RF calls to accommodate the daily shifts of their auditory fovea.

  4. Auditory N1 reveals planning and monitoring processes during music performance.

    Science.gov (United States)

    Mathias, Brian; Gehring, William J; Palmer, Caroline

    2017-02-01

    The current study investigated the relationship between planning processes and feedback monitoring during music performance, a complex task in which performers prepare upcoming events while monitoring their sensory outcomes. Theories of action planning in auditory-motor production tasks propose that the planning of future events co-occurs with the perception of auditory feedback. This study investigated the neural correlates of planning and feedback monitoring by manipulating the contents of auditory feedback during music performance. Pianists memorized and performed melodies at a cued tempo in a synchronization-continuation task while the EEG was recorded. During performance, auditory feedback associated with single melody tones was occasionally substituted with tones corresponding to future (next), present (current), or past (previous) melody tones. Only future-oriented altered feedback disrupted behavior: Future-oriented feedback caused pianists to slow down on the subsequent tone more than past-oriented feedback, and amplitudes of the auditory N1 potential elicited by the tone immediately following the altered feedback were larger for future-oriented than for past-oriented or noncontextual (unrelated) altered feedback; larger N1 amplitudes were associated with greater slowing following altered feedback in the future condition only. Feedback-related negativities were elicited in all altered feedback conditions. In sum, behavioral and neural evidence suggests that future-oriented feedback disrupts performance more than past-oriented feedback, consistent with planning theories that posit similarity-based interference between feedback and planning contents. Neural sensory processing of auditory feedback, reflected in the N1 ERP, may serve as a marker for temporal disruption caused by altered auditory feedback in auditory-motor production tasks. © 2016 Society for Psychophysiological Research.

  5. The tradeoff between signal detection and recognition rules auditory sensitivity under variable background noise conditions.

    Science.gov (United States)

    Lugli, Marco

    2015-12-07

    Animal acoustic communication commonly takes place under masked conditions. For instance, sound signals relevant for mating and survival are very often masked by background noise, which makes their detection and recognition by organisms difficult. Ambient noise (AN) varies in level and shape among different habitats, but also remarkable variations in time and space occurs within the same habitat. Variable AN conditions mask hearing thresholds of the receiver in complex and unpredictable ways, thereby causing distortions in sound perception. When communication takes place in a noisy environment, a highly sensitive system might confer no advantage to the receiver compared to a less sensitive one. The effects of noise masking on auditory thresholds and hearing-related functions are well known, and the potential role of AN in the evolution of the species' auditory sensitivity has been recognized by few authors. The mechanism of the underlying selection process has never been explored, however. Here I present a simple fitness model that seeks for the best sensitivity of a hearing system performing the detection and recognition of the sound under variable AN conditions. The model predicts higher sensitivity (i.e. lower hearing thresholds) as best strategy for species living in quiet habitats and lower sensitivity (i.e. higher hearing thresholds) as best strategy for those living in noisy habitats provided the cost of incorrect recognition is not low. The tradeoff between detection and recognition of acoustic signals appears to be a key factor determining the best level of hearing sensitivity of a species when acoustic communication is corrupted by noise. Copyright © 2015 Elsevier Ltd. All rights reserved.

  6. Synchronization of binocular motion parameters optoelectronic measurement system

    Science.gov (United States)

    Zhang, Lingfei; Ye, Dong; Che, Rensheng; Chen, Gang

    2008-10-01

    The synchronization between high-speed digital cameras and computers is very important for the binocular vision system based on light-weighted passive IR reflective markers and IR LED array PCB board, which is often used to measure the 3-D motion parameters of a rocket motor. In order to solve this problem, a comparison on the existing approaches to camera synchronization in the published literature was conducted. The advantages and disadvantages of the currently used methods were illustrated and their suitable applications were discussed. A new method, which uses self-made hardware resetting camera and software triggering image acquisition board, is provided. The self-made hardware is used to send TTL signal to two image acquisition boards one time per second. The TTL signal is used to reset two cameras and two image acquisition boards as PRIN signal, and then two image acquisition boards send same EXSYNC signal to two cameras. In this way, two cameras can be synchronized to exposure and capture images in the mean time. The test results indicated that the new approach designed in this paper can meet the demand of image acquisition at a speed of 200f/s, whose synchronization accuracy is up to micro second.

  7. Method for Converter Synchronization with RF Injection

    OpenAIRE

    Joshua P. Bruckmeyer; Ivica Kostanic

    2015-01-01

    This paper presents an injection method for synchronizing analog to digital converters (ADC). This approach can eliminate the need for precision routed discrete synchronization signals of current technologies, such as JESD204. By eliminating the setup and hold time requirements at the conversion (or near conversion) clock rate, higher sample rate systems can be synchronized. Measured data from an existing multiple ADC conversion system was used to evaluate the method. Coherent beams were simu...

  8. Auditory-motor learning influences auditory memory for music.

    Science.gov (United States)

    Brown, Rachel M; Palmer, Caroline

    2012-05-01

    In two experiments, we investigated how auditory-motor learning influences performers' memory for music. Skilled pianists learned novel melodies in four conditions: auditory only (listening), motor only (performing without sound), strongly coupled auditory-motor (normal performance), and weakly coupled auditory-motor (performing along with auditory recordings). Pianists' recognition of the learned melodies was better following auditory-only or auditory-motor (weakly coupled and strongly coupled) learning than following motor-only learning, and better following strongly coupled auditory-motor learning than following auditory-only learning. Auditory and motor imagery abilities modulated the learning effects: Pianists with high auditory imagery scores had better recognition following motor-only learning, suggesting that auditory imagery compensated for missing auditory feedback at the learning stage. Experiment 2 replicated the findings of Experiment 1 with melodies that contained greater variation in acoustic features. Melodies that were slower and less variable in tempo and intensity were remembered better following weakly coupled auditory-motor learning. These findings suggest that motor learning can aid performers' auditory recognition of music beyond auditory learning alone, and that motor learning is influenced by individual abilities in mental imagery and by variation in acoustic features.

  9. An adaptive synchronization strategy based on active control for demodulating message hidden in chaotic signals

    International Nuclear Information System (INIS)

    Tang Fang

    2008-01-01

    In the field of secure communication, it is very important to demodulate the message hidden in chaotic signals. In this paper, an adaptive synchronization strategy based on active control is proposed, which is used to design an active controller and an appropriate adaptive demodulator at the receiver to recover the transmitted message hidden in chaotic signals of a drive system. Based on Lyapunov stability theory, it is shown that the transmitted message can be theoretically recovered by using the proposed strategy. Numerical simulations based on the Chua's circuit are also presented to verify the effectiveness of the proposed strategy. In addition, it is shown via simulations that, by increasing the gain of the active controller the message error caused by the external noise and the discontinuous property of the message can be reduced

  10. Temporal envelope processing in the human auditory cortex: response and interconnections of auditory cortical areas.

    Science.gov (United States)

    Gourévitch, Boris; Le Bouquin Jeannès, Régine; Faucon, Gérard; Liégeois-Chauvel, Catherine

    2008-03-01

    Temporal envelope processing in the human auditory cortex has an important role in language analysis. In this paper, depth recordings of local field potentials in response to amplitude modulated white noises were used to design maps of activation in primary, secondary and associative auditory areas and to study the propagation of the cortical activity between them. The comparison of activations between auditory areas was based on a signal-to-noise ratio associated with the response to amplitude modulation (AM). The functional connectivity between cortical areas was quantified by the directed coherence (DCOH) applied to auditory evoked potentials. This study shows the following reproducible results on twenty subjects: (1) the primary auditory cortex (PAC), the secondary cortices (secondary auditory cortex (SAC) and planum temporale (PT)), the insular gyrus, the Brodmann area (BA) 22 and the posterior part of T1 gyrus (T1Post) respond to AM in both hemispheres. (2) A stronger response to AM was observed in SAC and T1Post of the left hemisphere independent of the modulation frequency (MF), and in the left BA22 for MFs 8 and 16Hz, compared to those in the right. (3) The activation and propagation features emphasized at least four different types of temporal processing. (4) A sequential activation of PAC, SAC and BA22 areas was clearly visible at all MFs, while other auditory areas may be more involved in parallel processing upon a stream originating from primary auditory area, which thus acts as a distribution hub. These results suggest that different psychological information is carried by the temporal envelope of sounds relative to the rate of amplitude modulation.

  11. Spatiotemporal Relationships among Audiovisual Stimuli Modulate Auditory Facilitation of Visual Target Discrimination.

    Science.gov (United States)

    Li, Qi; Yang, Huamin; Sun, Fang; Wu, Jinglong

    2015-03-01

    Sensory information is multimodal; through audiovisual interaction, task-irrelevant auditory stimuli tend to speed response times and increase visual perception accuracy. However, mechanisms underlying these performance enhancements have remained unclear. We hypothesize that task-irrelevant auditory stimuli might provide reliable temporal and spatial cues for visual target discrimination and behavioral response enhancement. Using signal detection theory, the present study investigated the effects of spatiotemporal relationships on auditory facilitation of visual target discrimination. Three experiments were conducted where an auditory stimulus maintained reliable temporal and/or spatial relationships with visual target stimuli. Results showed that perception sensitivity (d') to visual target stimuli was enhanced only when a task-irrelevant auditory stimulus maintained reliable spatiotemporal relationships with a visual target stimulus. When only reliable spatial or temporal information was contained, perception sensitivity was not enhanced. These results suggest that reliable spatiotemporal relationships between visual and auditory signals are required for audiovisual integration during a visual discrimination task, most likely due to a spread of attention. These results also indicate that auditory facilitation of visual target discrimination follows from late-stage cognitive processes rather than early stage sensory processes. © 2015 SAGE Publications.

  12. Stimulator with arbitrary waveform for auditory evoked potentials

    International Nuclear Information System (INIS)

    Martins, H R; Romao, M; Placido, D; Provenzano, F; Tierra-Criollo, C J

    2007-01-01

    The technological improvement helps many medical areas. The audiometric exams involving the auditory evoked potentials can make better diagnoses of auditory disorders. This paper proposes the development of a stimulator based on Digital Signal Processor. This stimulator is the first step of an auditory evoked potential system based on the ADSP-BF533 EZ KIT LITE (Analog Devices Company - USA). The stimulator can generate arbitrary waveform like Sine Waves, Modulated Amplitude, Pulses, Bursts and Pips. The waveforms are generated through a graphical interface programmed in C++ in which the user can define the parameters of the waveform. Furthermore, the user can set the exam parameters as number of stimuli, time with stimulation (Time ON) and time without stimulus (Time OFF). In future works will be implemented another parts of the system that includes the acquirement of electroencephalogram and signal processing to estimate and analyze the evoked potential

  13. Stimulator with arbitrary waveform for auditory evoked potentials

    Energy Technology Data Exchange (ETDEWEB)

    Martins, H R; Romao, M; Placido, D; Provenzano, F; Tierra-Criollo, C J [Universidade Federal de Minas Gerais (UFMG), Departamento de Engenharia Eletrica (DEE), Nucleo de Estudos e Pesquisa em Engenharia Biomedica NEPEB, Av. Ant. Carlos, 6627, sala 2206, Pampulha, Belo Horizonte, MG, 31.270-901 (Brazil)

    2007-11-15

    The technological improvement helps many medical areas. The audiometric exams involving the auditory evoked potentials can make better diagnoses of auditory disorders. This paper proposes the development of a stimulator based on Digital Signal Processor. This stimulator is the first step of an auditory evoked potential system based on the ADSP-BF533 EZ KIT LITE (Analog Devices Company - USA). The stimulator can generate arbitrary waveform like Sine Waves, Modulated Amplitude, Pulses, Bursts and Pips. The waveforms are generated through a graphical interface programmed in C++ in which the user can define the parameters of the waveform. Furthermore, the user can set the exam parameters as number of stimuli, time with stimulation (Time ON) and time without stimulus (Time OFF). In future works will be implemented another parts of the system that includes the acquirement of electroencephalogram and signal processing to estimate and analyze the evoked potential.

  14. Altered Sensory Feedbacks in Pianist's Dystonia: the altered auditory feedback paradigm and the glove effect

    Directory of Open Access Journals (Sweden)

    Felicia Pei-Hsin Cheng

    2013-12-01

    Full Text Available Background: This study investigates the effect of altered auditory feedback (AAF in musician's dystonia (MD and discusses whether altered auditory feedback can be considered as a sensory trick in MD. Furthermore, the effect of AAF is compared with altered tactile feedback, which can serve as a sensory trick in several other forms of focal dystonia. Methods: The method is based on scale analysis (Jabusch et al. 2004. Experiment 1 employs synchronization paradigm: 12 MD patients and 25 healthy pianists had to repeatedly play C-major scales in synchrony with a metronome on a MIDI-piano with 3 auditory feedback conditions: 1. normal feedback; 2. no feedback; 3. constant delayed feedback. Experiment 2 employs synchronization-continuation paradigm: 12 MD patients and 12 healthy pianists had to repeatedly play C-major scales in two phases: first in synchrony with a metronome, secondly continue the established tempo without the metronome. There are 4 experimental conditions, among them 3 are the same altered auditory feedback as in Experiment 1 and 1 is related to altered tactile sensory input. The coefficient of variation of inter-onset intervals of the key depressions was calculated to evaluate fine motor control. Results: In both experiments, the healthy controls and the patients behaved very similarly. There is no difference in the regularity of playing between the two groups under any condition, and neither did AAF nor did altered tactile feedback have a beneficial effect on patients’ fine motor control. Conclusions: The results of the two experiments suggest that in the context of our experimental designs, AAF and altered tactile feedback play a minor role in motor coordination in patients with musicians' dystonia. We propose that altered auditory and tactile feedback do not serve as effective sensory tricks and may not temporarily reduce the symptoms of patients suffering from MD in this experimental context.

  15. The importance of laughing in your face: influences of visual laughter on auditory laughter perception.

    Science.gov (United States)

    Jordan, Timothy R; Abedipour, Lily

    2010-01-01

    Hearing the sound of laughter is important for social communication, but processes contributing to the audibility of laughter remain to be determined. Production of laughter resembles production of speech in that both involve visible facial movements accompanying socially significant auditory signals. However, while it is known that speech is more audible when the facial movements producing the speech sound can be seen, similar visual enhancement of the audibility of laughter remains unknown. To address this issue, spontaneously occurring laughter was edited to produce stimuli comprising visual laughter, auditory laughter, visual and auditory laughter combined, and no laughter at all (either visual or auditory), all presented in four levels of background noise. Visual laughter and no-laughter stimuli produced very few reports of auditory laughter. However, visual laughter consistently made auditory laughter more audible, compared to the same auditory signal presented without visual laughter, resembling findings reported previously for speech.

  16. The synchronization of three fractional differential systems

    International Nuclear Information System (INIS)

    Li Changpin; Yan Jianping

    2007-01-01

    In this paper, a new method is proposed and applied to the synchronization of fractional differential systems (or 'differential systems with fractional orders'), where both drive and response systems have the same dimensionality and are coupled by the driving signal. The present technique is based on the stability criterion of linear fractional systems. This method is implemented in (chaos) synchronization of the fractional Lorenz system, Chen system and Chua circuit. Numerical simulations show the present synchronization method works well

  17. Visualization of synchronization of the uterine contraction signals: running cross-correlation and wavelet running cross-correlation methods.

    Science.gov (United States)

    Oczeretko, Edward; Swiatecka, Jolanta; Kitlas, Agnieszka; Laudanski, Tadeusz; Pierzynski, Piotr

    2006-01-01

    In physiological research, we often study multivariate data sets, containing two or more simultaneously recorded time series. The aim of this paper is to present the cross-correlation and the wavelet cross-correlation methods to assess synchronization between contractions in different topographic regions of the uterus. From a medical point of view, it is important to identify time delays between contractions, which may be of potential diagnostic significance in various pathologies. The cross-correlation was computed in a moving window with a width corresponding to approximately two or three contractions. As a result, the running cross-correlation function was obtained. The propagation% parameter assessed from this function allows quantitative description of synchronization in bivariate time series. In general, the uterine contraction signals are very complicated. Wavelet transforms provide insight into the structure of the time series at various frequencies (scales). To show the changes of the propagation% parameter along scales, a wavelet running cross-correlation was used. At first, the continuous wavelet transforms as the uterine contraction signals were received and afterwards, a running cross-correlation analysis was conducted for each pair of transformed time series. The findings show that running functions are very useful in the analysis of uterine contractions.

  18. Music and speech distractors disrupt sensorimotor synchronization: effects of musical training.

    Science.gov (United States)

    Białuńska, Anita; Dalla Bella, Simone

    2017-12-01

    Humans display a natural tendency to move to the beat of music, more than to the rhythm of any other auditory stimulus. We typically move with music, but rarely with speech. This proclivity is apparent early during development and can be further developed over the years via joint dancing, singing, or instrument playing. Synchronization of movement to the beat can thus improve with age, but also with musical experience. In a previous study, we found that music perturbed synchronization with a metronome more than speech fragments; music superiority disappeared when distractors shared isochrony and the same meter (Dalla Bella et al., PLoS One 8(8):e71945, 2013). Here, we examined if the interfering effect of music and speech distractors in a synchronization task is influenced by musical training. Musicians and non-musicians synchronized by producing finger force pulses to the sounds of a metronome while music and speech distractors were presented at one of various phase relationships with respect to the target. Distractors were familiar musical excerpts and fragments of children poetry comparable in terms of beat/stress isochrony. Music perturbed synchronization with the metronome more than speech did in both groups. However, the difference in synchronization error between music and speech distractors was smaller for musicians than for non-musicians, especially when the peak force of movement is reached. These findings point to a link between musical training and timing of sensorimotor synchronization when reacting to music and speech distractors.

  19. Category-specific responses to faces and objects in primate auditory cortex

    Directory of Open Access Journals (Sweden)

    Kari L Hoffman

    2008-03-01

    Full Text Available Auditory and visual signals often occur together, and the two sensory channels are known to infl uence each other to facilitate perception. The neural basis of this integration is not well understood, although other forms of multisensory infl uences have been shown to occur at surprisingly early stages of processing in cortex. Primary visual cortex neurons can show frequency-tuning to auditory stimuli, and auditory cortex responds selectively to certain somatosensory stimuli, supporting the possibility that complex visual signals may modulate early stages of auditory processing. To elucidate which auditory regions, if any, are responsive to complex visual stimuli, we recorded from auditory cortex and the superior temporal sulcus while presenting visual stimuli consisting of various objects, neutral faces, and facial expressions generated during vocalization. Both objects and conspecifi c faces elicited robust fi eld potential responses in auditory cortex sites, but the responses varied by category: both neutral and vocalizing faces had a highly consistent negative component (N100 followed by a broader positive component (P180 whereas object responses were more variable in time and shape, but could be discriminated consistently from the responses to faces. The face response did not vary within the face category, i.e., for expressive vs. neutral face stimuli. The presence of responses for both objects and neutral faces suggests that auditory cortex receives highly informative visual input that is not restricted to those stimuli associated with auditory components. These results reveal selectivity for complex visual stimuli in a brain region conventionally described as non-visual unisensory cortex.

  20. Methods for constructing time synchronous networks. Part 1. Consideration and experiment of time synchronizing functions and devices; Zenkei jikoku dokimo no kosei shuho. 1. Jikoku doki hoshiki no kino kento to jikkenteki kensho

    Energy Technology Data Exchange (ETDEWEB)

    Serizawa, Y.; Kitamura, K.; Myojin, M.; Shimizu, K. [Central Research Institute of Electric Power Industry, Tokyo (Japan); Nenohi, H.; Morimitsu, M.; Masui, O.; Matsushima, T. [NEC Corp., Tokyo (Japan)

    1996-03-01

    For the purpose of the back-up protection of regional power line system and the protection and control of accident influence, construction methods of time synchronizing network have been investigated. This paper proposes a method suitable for business power communication. For the synchronization between stations, transmission delay and its fluctuation have to be measured accurately by transmitting time signals between stations. Based on the observed delay values, the function of time correction (phase synchronizing control) of synchronized station is required. A mutual type synchronizing method was proposed, to which a sampling synchronization method of carrier relay system for transmission line protection was applied. In this method, transmission of time signals and measurement of delays are mutually conducted between stations, and phases of independent transmitters of the both stations are controlled so as to be equivalent. This method is characterized by the suppression of delays with fluctuation in the same direction of two-way transmission. For the transmission of time signals, two types of synchronizing devices were employed, i.e., synchronous and asynchronous manners. In the former, frequency synchronous network of SDH (synchronous digital hierarchy) was synchronized. In the latter, a carrier relay device was utilized independent of frequency synchronous network. It was expected that a synchronous network with an order of {mu}s can be constructed. 11 refs., 21 figs., 3 tabs.

  1. Synchronization of identical chaotic systems through external chaotic driving

    International Nuclear Information System (INIS)

    Patidar, V.; Sud, K.K.

    2005-11-01

    In recent years, the study of synchronization of identical chaotic systems subjected to a common fluctuating random driving signal has drawn considerable interest. In this communication, we report that it is possible to achieve synchronization between two identical chaotic systems, which are not coupled directly but subjected to an external chaotic signal. The external chaotic signal may be obtained from any chaotic system identical or non-identical to both identical chaotic systems. Results of numerical simulations on well known Roessler and jerk dynamical systems have been presented. (author)

  2. Division Multiplexing of 10 Gbit/s Ethernet Signals Synchronized by All-Optical Signal Processing Based on a Time-Lens

    DEFF Research Database (Denmark)

    Areal, Janaina Laguardia

    This Thesis presents 3 years work of an optical circuit that performs both pulse compression and frame synchronization and retiming. Our design aims at directly multiplexing several 10G Ethernet data packets (frames) to a high-speed OTDM link. This scheme is optically trans-parent and does not re...... coupler, completing the OTDM signal generation. We demonstrate the effectiveness of the design by laboratory experi-ments and simulations with VPI and MatLab....... not require clock recovery, resulting in a potentially very efficient solution. The scheme uses a time-lens, implemented through a sinusoidally driven optical phase modulation, combined with a linear dispersion element. As time-lenses are also used for pulse compression, we de-sign the circuit also to perform...

  3. Visual Information Present in Infragranular Layers of Mouse Auditory Cortex.

    Science.gov (United States)

    Morrill, Ryan J; Hasenstaub, Andrea R

    2018-03-14

    The cerebral cortex is a major hub for the convergence and integration of signals from across the sensory modalities; sensory cortices, including primary regions, are no exception. Here we show that visual stimuli influence neural firing in the auditory cortex of awake male and female mice, using multisite probes to sample single units across multiple cortical layers. We demonstrate that visual stimuli influence firing in both primary and secondary auditory cortex. We then determine the laminar location of recording sites through electrode track tracing with fluorescent dye and optogenetic identification using layer-specific markers. Spiking responses to visual stimulation occur deep in auditory cortex and are particularly prominent in layer 6. Visual modulation of firing rate occurs more frequently at areas with secondary-like auditory responses than those with primary-like responses. Auditory cortical responses to drifting visual gratings are not orientation-tuned, unlike visual cortex responses. The deepest cortical layers thus appear to be an important locus for cross-modal integration in auditory cortex. SIGNIFICANCE STATEMENT The deepest layers of the auditory cortex are often considered its most enigmatic, possessing a wide range of cell morphologies and atypical sensory responses. Here we show that, in mouse auditory cortex, these layers represent a locus of cross-modal convergence, containing many units responsive to visual stimuli. Our results suggest that this visual signal conveys the presence and timing of a stimulus rather than specifics about that stimulus, such as its orientation. These results shed light on both how and what types of cross-modal information is integrated at the earliest stages of sensory cortical processing. Copyright © 2018 the authors 0270-6474/18/382854-09$15.00/0.

  4. Hybrid electronic/optical synchronized chaos communication system.

    Science.gov (United States)

    Toomey, J P; Kane, D M; Davidović, A; Huntington, E H

    2009-04-27

    A hybrid electronic/optical system for synchronizing a chaotic receiver to a chaotic transmitter has been demonstrated. The chaotic signal is generated electronically and injected, in addition to a constant bias current, to a semiconductor laser to produce an optical carrier for transmission. The optical chaotic carrier is photodetected to regenerate an electronic signal for synchronization in a matched electronic receiver The system has been successfully used for the transmission and recovery of a chaos masked message that is added to the chaotic optical carrier. Past demonstrations of synchronized chaos based, secure communication systems have used either an electronic chaotic carrier or an optical chaotic carrier (such as the chaotic output of various nonlinear laser systems). This is the first electronic/optical hybrid system to be demonstrated. We call this generation of a chaotic optical carrier by electronic injection.

  5. Real-time synchronization of wireless sensor network by 1-PPS signal

    Science.gov (United States)

    Giammarini, Marco; Pieralisi, Marco; Isidori, Daniela; Concettoni, Enrico; Cristalli, Cristina; Fioravanti, Matteo

    2015-05-01

    The use of wireless sensor networks with different nodes is desirable in a smart environment, because the network setting up and installation on preexisting structures can be done without a fixed cabled infrastructure. The flexibility of the monitoring system is fundamental where the use of a considerable quantity of cables could compromise the normal exercise, could affect the quality of acquired signal and finally increase the cost of the materials and installation. The network is composed of several intelligent "nodes", which acquires data from different kind of sensors, and then store or transmit them to a central elaboration unit. The synchronization of data acquisition is the core of the real-time wireless sensor network (WSN). In this paper, we present a comparison between different methods proposed by literature for the real-time acquisition in a WSN and finally we present our solution based on 1-Pulse-Per-Second (1-PPS) signal generated by GPS systems. The sensor node developed is a small-embedded system based on ARM microcontroller that manages the acquisition, the timing and the post-processing of the data. The communications between the sensors and the master based on IEEE 802.15.4 protocol and managed by dedicated software. Finally, we present the preliminary results obtained on a 3 floor building simulator with the wireless sensors system developed.

  6. Synchronization System for Next Generation Light Sources

    Energy Technology Data Exchange (ETDEWEB)

    Zavriyev, Anton [MagiQ Technologies, Inc., Somerville, MA (United States)

    2014-03-27

    An alternative synchronization technique – one that would allow explicit control of the pulse train including its repetition rate and delay is clearly desired. We propose such a scheme. Our method is based on optical interferometry and permits synchronization of the pulse trains generated by two independent mode-locked lasers. As the next generation x-ray sources will be driven by a clock signal derived from a mode-locked optical source, our technique will provide a way to synchronize x-ray probe with the optical pump pulses.

  7. Neural responses in the primary auditory cortex of freely behaving cats while discriminating fast and slow click-trains.

    Science.gov (United States)

    Dong, Chao; Qin, Ling; Liu, Yongchun; Zhang, Xinan; Sato, Yu

    2011-01-01

    Repeated acoustic events are ubiquitous temporal features of natural sounds. To reveal the neural representation of the sound repetition rate, a number of electrophysiological studies have been conducted on various mammals and it has been proposed that both the spike-time and firing rate of primary auditory cortex (A1) neurons encode the repetition rate. However, previous studies rarely examined how the experimental animals perceive the difference in the sound repetition rate, and a caveat to these experiments is that they compared physiological data obtained from animals with psychophysical data obtained from humans. In this study, for the first time, we directly investigated acoustic perception and the underlying neural mechanisms in the same experimental animal by examining spike activities in the A1 of free-moving cats while performing a Go/No-go task to discriminate the click-trains at different repetition rates (12.5-200 Hz). As reported by previous studies on passively listening animals, A1 neurons showed both synchronized and non-synchronized responses to the click-trains. We further found that the neural performance estimated from the precise temporal information of synchronized units was good enough to distinguish all 16.7-200 Hz from the 12.5 Hz repetition rate; however, the cats showed declining behavioral performance with the decrease of the target repetition rate, indicating an increase of difficulty in discriminating two slower click-trains. Such behavioral performance was well explained by the firing rate of some synchronized and non-synchronized units. Trial-by-trial analysis indicated that A1 activity was not affected by the cat's judgment of behavioral response. Our results suggest that the main function of A1 is to effectively represent temporal signals using both spike timing and firing rate, while the cats may read out the rate-coding information to perform the task in this experiment.

  8. Neural responses in the primary auditory cortex of freely behaving cats while discriminating fast and slow click-trains.

    Directory of Open Access Journals (Sweden)

    Chao Dong

    Full Text Available Repeated acoustic events are ubiquitous temporal features of natural sounds. To reveal the neural representation of the sound repetition rate, a number of electrophysiological studies have been conducted on various mammals and it has been proposed that both the spike-time and firing rate of primary auditory cortex (A1 neurons encode the repetition rate. However, previous studies rarely examined how the experimental animals perceive the difference in the sound repetition rate, and a caveat to these experiments is that they compared physiological data obtained from animals with psychophysical data obtained from humans. In this study, for the first time, we directly investigated acoustic perception and the underlying neural mechanisms in the same experimental animal by examining spike activities in the A1 of free-moving cats while performing a Go/No-go task to discriminate the click-trains at different repetition rates (12.5-200 Hz. As reported by previous studies on passively listening animals, A1 neurons showed both synchronized and non-synchronized responses to the click-trains. We further found that the neural performance estimated from the precise temporal information of synchronized units was good enough to distinguish all 16.7-200 Hz from the 12.5 Hz repetition rate; however, the cats showed declining behavioral performance with the decrease of the target repetition rate, indicating an increase of difficulty in discriminating two slower click-trains. Such behavioral performance was well explained by the firing rate of some synchronized and non-synchronized units. Trial-by-trial analysis indicated that A1 activity was not affected by the cat's judgment of behavioral response. Our results suggest that the main function of A1 is to effectively represent temporal signals using both spike timing and firing rate, while the cats may read out the rate-coding information to perform the task in this experiment.

  9. A dominance hierarchy of auditory spatial cues in barn owls.

    Directory of Open Access Journals (Sweden)

    Ilana B Witten

    2010-04-01

    Full Text Available Barn owls integrate spatial information across frequency channels to localize sounds in space.We presented barn owls with synchronous sounds that contained different bands of frequencies (3-5 kHz and 7-9 kHz from different locations in space. When the owls were confronted with the conflicting localization cues from two synchronous sounds of equal level, their orienting responses were dominated by one of the sounds: they oriented toward the location of the low frequency sound when the sources were separated in azimuth; in contrast, they oriented toward the location of the high frequency sound when the sources were separated in elevation. We identified neural correlates of this behavioral effect in the optic tectum (OT, superior colliculus in mammals, which contains a map of auditory space and is involved in generating orienting movements to sounds. We found that low frequency cues dominate the representation of sound azimuth in the OT space map, whereas high frequency cues dominate the representation of sound elevation.We argue that the dominance hierarchy of localization cues reflects several factors: 1 the relative amplitude of the sound providing the cue, 2 the resolution with which the auditory system measures the value of a cue, and 3 the spatial ambiguity in interpreting the cue. These same factors may contribute to the relative weighting of sound localization cues in other species, including humans.

  10. Analysis and compensation of synchronous measurement error for multi-channel laser interferometer

    International Nuclear Information System (INIS)

    Du, Shengwu; Hu, Jinchun; Zhu, Yu; Hu, Chuxiong

    2017-01-01

    Dual-frequency laser interferometer has been widely used in precision motion system as a displacement sensor, to achieve nanoscale positioning or synchronization accuracy. In a multi-channel laser interferometer synchronous measurement system, signal delays are different in the different channels, which will cause asynchronous measurement, and then lead to measurement error, synchronous measurement error (SME). Based on signal delay analysis of the measurement system, this paper presents a multi-channel SME framework for synchronous measurement, and establishes the model between SME and motion velocity. Further, a real-time compensation method for SME is proposed. This method has been verified in a self-developed laser interferometer signal processing board (SPB). The experiment result showed that, using this compensation method, at a motion velocity 0.89 m s −1 , the max SME between two measuring channels in the SPB is 1.1 nm. This method is more easily implemented and applied to engineering than the method of directly testing smaller signal delay. (paper)

  11. Analysis and compensation of synchronous measurement error for multi-channel laser interferometer

    Science.gov (United States)

    Du, Shengwu; Hu, Jinchun; Zhu, Yu; Hu, Chuxiong

    2017-05-01

    Dual-frequency laser interferometer has been widely used in precision motion system as a displacement sensor, to achieve nanoscale positioning or synchronization accuracy. In a multi-channel laser interferometer synchronous measurement system, signal delays are different in the different channels, which will cause asynchronous measurement, and then lead to measurement error, synchronous measurement error (SME). Based on signal delay analysis of the measurement system, this paper presents a multi-channel SME framework for synchronous measurement, and establishes the model between SME and motion velocity. Further, a real-time compensation method for SME is proposed. This method has been verified in a self-developed laser interferometer signal processing board (SPB). The experiment result showed that, using this compensation method, at a motion velocity 0.89 m s-1, the max SME between two measuring channels in the SPB is 1.1 nm. This method is more easily implemented and applied to engineering than the method of directly testing smaller signal delay.

  12. Brain bases for auditory stimulus-driven figure-ground segregation.

    Science.gov (United States)

    Teki, Sundeep; Chait, Maria; Kumar, Sukhbinder; von Kriegstein, Katharina; Griffiths, Timothy D

    2011-01-05

    Auditory figure-ground segregation, listeners' ability to selectively hear out a sound of interest from a background of competing sounds, is a fundamental aspect of scene analysis. In contrast to the disordered acoustic environment we experience during everyday listening, most studies of auditory segregation have used relatively simple, temporally regular signals. We developed a new figure-ground stimulus that incorporates stochastic variation of the figure and background that captures the rich spectrotemporal complexity of natural acoustic scenes. Figure and background signals overlap in spectrotemporal space, but vary in the statistics of fluctuation, such that the only way to extract the figure is by integrating the patterns over time and frequency. Our behavioral results demonstrate that human listeners are remarkably sensitive to the appearance of such figures. In a functional magnetic resonance imaging experiment, aimed at investigating preattentive, stimulus-driven, auditory segregation mechanisms, naive subjects listened to these stimuli while performing an irrelevant task. Results demonstrate significant activations in the intraparietal sulcus (IPS) and the superior temporal sulcus related to bottom-up, stimulus-driven figure-ground decomposition. We did not observe any significant activation in the primary auditory cortex. Our results support a role for automatic, bottom-up mechanisms in the IPS in mediating stimulus-driven, auditory figure-ground segregation, which is consistent with accumulating evidence implicating the IPS in structuring sensory input and perceptual organization.

  13. Mixed synchronization in chaotic oscillators using scalar coupling

    Energy Technology Data Exchange (ETDEWEB)

    Bhowmick, Sourav K.; Hens, Chittaranjan [CSIR – Indian Institute of Chemical Biology, Jadavpur, Kolkata 700032 (India); Ghosh, Dibakar, E-mail: drghosh_math@yahoo.co.in [Department of Mathematics, University of Kalyani, West Bengal 741235 (India); Dana, Syamal K. [CSIR – Indian Institute of Chemical Biology, Jadavpur, Kolkata 700032 (India)

    2012-07-23

    We report experimental evidence of mixed synchronization in two unidirectionally coupled chaotic oscillators using a scalar coupling. In this synchronization regime, some of the state variables may be in complete synchronization while others may be in anti-synchronization state. We extended the theory by using an adaptive controller with an updating law based on Lyapunov function stability to include parameter fluctuation. Using the scheme, we implemented a cryptographic encoding for digital signal through parameter modulation. -- Highlights: ► We provided experimental evidence of the mixed synchronization scheme while other methods are mostly theoretical nature. ► We numerically studied adaptive mixed synchronization when the parameters are not completely known using scalar coupling. ► We proposed a secure communication system where three digital messages are transmitted using parameter modulation.

  14. Detection of generalized synchronization using echo state networks

    Science.gov (United States)

    Ibáñez-Soria, D.; Garcia-Ojalvo, J.; Soria-Frisch, A.; Ruffini, G.

    2018-03-01

    Generalized synchronization between coupled dynamical systems is a phenomenon of relevance in applications that range from secure communications to physiological modelling. Here, we test the capabilities of reservoir computing and, in particular, echo state networks for the detection of generalized synchronization. A nonlinear dynamical system consisting of two coupled Rössler chaotic attractors is used to generate temporal series consisting of time-locked generalized synchronized sequences interleaved with unsynchronized ones. Correctly tuned, echo state networks are able to efficiently discriminate between unsynchronized and synchronized sequences even in the presence of relatively high levels of noise. Compared to other state-of-the-art techniques of synchronization detection, the online capabilities of the proposed Echo State Network based methodology make it a promising choice for real-time applications aiming to monitor dynamical synchronization changes in continuous signals.

  15. Lag synchronization of hyperchaos with application to secure communications

    International Nuclear Information System (INIS)

    Li Chuandong; Liao Xiaofeng; Wong Kwokwo

    2005-01-01

    In this paper, hyperchaotic lag synchronization is restated as a nonlinear and lag-in-time observer design issue. This approach leads to a systematic tool, which guarantees the lag synchronization of a wide class of chaotic or hyperchaotic systems via a scalar signal. By exploiting this result, we propose a hyperchaos-based cryptosystem scheme that combines the conventional cryptographic methods and the lag synchronization of chaotic circuits. The computer simulation results show that the lag synchronization scheme and the cryptosystem proposed in this paper are both feasible

  16. Communication with spatial periodic chaos synchronization

    International Nuclear Information System (INIS)

    Zhou, J.; Huang, H.B.; Qi, G.X.; Yang, P.; Xie, X.

    2005-01-01

    Based on the spatial periodic chaos synchronization in coupled ring and linear arrays, we proposed a random high-dimensional chaotic encryption scheme. The transmitter can choose hyperchaotic signals randomly from the ring at any different time and simultaneously transmit the information of chaotic oscillators in the ring to receiver through public channel, so that the message can be masked by different hyperchaotic signals in different time intervals during communication, and the receiver can decode the message based on chaos synchronization but the attacker does not know the random hyperchaotic dynamics and cannot decode the message. Furthermore, the high sensitivity to the symmetry of the coupling structure makes the attacker very difficult to obtain any useful message from the channel

  17. Beat Synchronization across the Lifespan: Intersection of Development and Musical Experience.

    Directory of Open Access Journals (Sweden)

    Elaine C Thompson

    Full Text Available Rhythmic entrainment, or beat synchronization, provides an opportunity to understand how multiple systems operate together to integrate sensory-motor information. Also, synchronization is an essential component of musical performance that may be enhanced through musical training. Investigations of rhythmic entrainment have revealed a developmental trajectory across the lifespan, showing synchronization improves with age and musical experience. Here, we explore the development and maintenance of synchronization in childhood through older adulthood in a large cohort of participants (N = 145, and also ask how it may be altered by musical experience. We employed a uniform assessment of beat synchronization for all participants and compared performance developmentally and between individuals with and without musical experience. We show that the ability to consistently tap along to a beat improves with age into adulthood, yet in older adulthood tapping performance becomes more variable. Also, from childhood into young adulthood, individuals are able to tap increasingly close to the beat (i.e., asynchronies decline with age, however, this trend reverses from younger into older adulthood. There is a positive association between proportion of life spent playing music and tapping performance, which suggests a link between musical experience and auditory-motor integration. These results are broadly consistent with previous investigations into the development of beat synchronization across the lifespan, and thus complement existing studies and present new insights offered by a different, large cross-sectional sample.

  18. Increase in Synchronization of Autonomic Rhythms between Individuals When Listening to Music

    Science.gov (United States)

    Bernardi, Nicolò F.; Codrons, Erwan; di Leo, Rita; Vandoni, Matteo; Cavallaro, Filippo; Vita, Giuseppe; Bernardi, Luciano

    2017-01-01

    In light of theories postulating a role for music in forming emotional and social bonds, here we investigated whether endogenous rhythms synchronize between multiple individuals when listening to music. Cardiovascular and respiratory recordings were taken from multiple individuals (musically trained or music-naïve) simultaneously, at rest and during a live concert comprising music excerpts with varying degrees of complexity of the acoustic envelope. Inter-individual synchronization of cardiorespiratory rhythms showed a subtle but reliable increase during passively listening to music compared to baseline. The low-level auditory features of the music were largely responsible for creating or disrupting such synchronism, explaining ~80% of its variance, over and beyond subjective musical preferences and previous musical training. Listening to simple rhythms and melodies, which largely dominate the choice of music during rituals and mass events, brings individuals together in terms of their physiological rhythms, which could explain why music is widely used to favor social bonds. PMID:29089898

  19. Biomimetic Sonar for Electrical Activation of the Auditory Pathway

    Directory of Open Access Journals (Sweden)

    D. Menniti

    2017-01-01

    Full Text Available Relying on the mechanism of bat’s echolocation system, a bioinspired electronic device has been developed to investigate the cortical activity of mammals in response to auditory sensorial stimuli. By means of implanted electrodes, acoustical information about the external environment generated by a biomimetic system and converted in electrical signals was delivered to anatomically selected structures of the auditory pathway. Electrocorticographic recordings showed that cerebral activity response is highly dependent on the information carried out by ultrasounds and is frequency-locked with the signal repetition rate. Frequency analysis reveals that delta and beta rhythm content increases, suggesting that sensorial information is successfully transferred and integrated. In addition, principal component analysis highlights how all the stimuli generate patterns of neural activity which can be clearly classified. The results show that brain response is modulated by echo signal features suggesting that spatial information sent by biomimetic sonar is efficiently interpreted and encoded by the auditory system. Consequently, these results give new perspective in artificial environmental perception, which could be used for developing new techniques useful in treating pathological conditions or influencing our perception of the surroundings.

  20. Diffusion tractography of the subcortical auditory system in a postmortem human brain

    OpenAIRE

    Sitek, Kevin

    2017-01-01

    The subcortical auditory system is challenging to identify with standard human brain imaging techniques: MRI signal decreases toward the center of the brain as well as at higher resolution, both of which are necessary for imaging small brainstem auditory structures.Using high-resolution diffusion-weighted MRI, we asked:Can we identify auditory structures and connections in high-resolution ex vivo images?Which structures and connections can be mapped in vivo?

  1. Transmission delays in hardware clock synchronization

    Science.gov (United States)

    Shin, Kang G.; Ramanathan, P.

    1988-01-01

    Various methods, both with software and hardware, have been proposed to synchronize a set of physical clocks in a system. Software methods are very flexible and economical but suffer an excessive time overhead, whereas hardware methods require no time overhead but are unable to handle transmission delays in clock signals. The effects of nonzero transmission delays in synchronization have been studied extensively in the communication area in the absence of malicious or Byzantine faults. The authors show that it is easy to incorporate the ideas from the communication area into the existing hardware clock synchronization algorithms to take into account the presence of both malicious faults and nonzero transmission delays.

  2. Transponder-aided joint calibration and synchronization compensation for distributed radar systems.

    Science.gov (United States)

    Wang, Wen-Qin

    2015-01-01

    High-precision radiometric calibration and synchronization compensation must be provided for distributed radar system due to separate transmitters and receivers. This paper proposes a transponder-aided joint radiometric calibration, motion compensation and synchronization for distributed radar remote sensing. As the transponder signal can be separated from the normal radar returns, it is used to calibrate the distributed radar for radiometry. Meanwhile, the distributed radar motion compensation and synchronization compensation algorithms are presented by utilizing the transponder signals. This method requires no hardware modifications to both the normal radar transmitter and receiver and no change to the operating pulse repetition frequency (PRF). The distributed radar radiometric calibration and synchronization compensation require only one transponder, but the motion compensation requires six transponders because there are six independent variables in the distributed radar geometry. Furthermore, a maximum likelihood method is used to estimate the transponder signal parameters. The proposed methods are verified by simulation results.

  3. Synchronicity from Synchronized Chaos

    Directory of Open Access Journals (Sweden)

    Gregory S. Duane

    2015-03-01

    Full Text Available The synchronization of loosely-coupled chaotic oscillators, a phenomenon investigated intensively for the last two decades, may realize the philosophical concept of “synchronicity”—the commonplace notion that related eventsmysteriously occur at the same time. When extended to continuous media and/or large discrete arrays, and when general (non-identical correspondences are considered between states, intermittent synchronous relationships indeed become ubiquitous. Meaningful synchronicity follows naturally if meaningful events are identified with coherent structures, defined by internal synchronization between remote degrees of freedom; a condition that has been posited as necessary for synchronizability with an external system. The important case of synchronization between mind and matter is realized if mind is analogized to a computer model, synchronizing with a sporadically observed system, as in meteorological data assimilation. Evidence for the ubiquity of synchronization is reviewed along with recent proposals that: (1 synchronization of different models of the same objective process may be an expeditious route to improved computational modeling and may also describe the functioning of conscious brains; and (2 the nonlocality in quantum phenomena implied by Bell’s theorem may be explained in a variety of deterministic (hidden variable interpretations if the quantum world resides on a generalized synchronization “manifold”.

  4. Integrated Time and Phase Synchronization Strategy for a Multichannel Spaceborne-Stationary Bistatic SAR System

    Directory of Open Access Journals (Sweden)

    Feng Hong

    2016-07-01

    Full Text Available The spatial separation of the transmitter and receiver in Bistatic Synthetic Aperture Radar (BiSAR makes it a promising and useful supplement to a classical Monostatic SAR system (MonoSAR. This paper proposes a novel integrated time and phase synchronization strategy for a multichannel spaceborne-stationary BiSAR system. Firstly, the time synchronization strategy is proposed, which includes Pulse Repetition Frequency (PRF generation under noisy conditions, multichannel calibration and the alignment of the recorded data with the orbital data. Furthermore, the phase synchronization strategy, which fully considers the deteriorative factors in the BiSAR configuration, is well studied. The contribution of the phase synchronization strategy includes two aspects: it not only compensates the phase error, but also improves the Signal to Noise Ratio (SNR of the obtained signals. Specifically, all direct signals on different PRF time can be reconstructed with the shift and phase compensation operation using a reference signal. Besides, since the parameters of the reference signal can be estimated only once using the selected practical direct signal and a priori information, the processing complexity is well reduced. Final imaging results with and without compensation for real data are presented to validate the proposed synchronization strategy.

  5. Synchronization of Harb-Zohdy Chaotic System via Back-Stepping Design

    Directory of Open Access Journals (Sweden)

    M. R. Shamsyeh Zahedi∗

    2015-12-01

    Full Text Available This paper is concerned with the problem of synchronization of the Harb-Zohdy chaotic system using the back-stepping. Based on the stability theory, the control for the synchronization of chaotic systems Harb-Zohdy is considered without unknown parameters. Next, an adaptive back-stepping control law is derived to generate an error signal between the drive and response systems Harb-Zohdy with an uncertain parameter asymptotically synchronized. Finally, this method is extended to synchronize the system with two unknown parameters. Note that the method presented here needs only one controller to realize the synchronization. Numerical simulations indicate the effectiveness of the proposed chaos synchronization scheme

  6. Left hemispheric dominance during auditory processing in a noisy environment

    Directory of Open Access Journals (Sweden)

    Ross Bernhard

    2007-11-01

    Full Text Available Abstract Background In daily life, we are exposed to different sound inputs simultaneously. During neural encoding in the auditory pathway, neural activities elicited by these different sounds interact with each other. In the present study, we investigated neural interactions elicited by masker and amplitude-modulated test stimulus in primary and non-primary human auditory cortex during ipsi-lateral and contra-lateral masking by means of magnetoencephalography (MEG. Results We observed significant decrements of auditory evoked responses and a significant inter-hemispheric difference for the N1m response during both ipsi- and contra-lateral masking. Conclusion The decrements of auditory evoked neural activities during simultaneous masking can be explained by neural interactions evoked by masker and test stimulus in peripheral and central auditory systems. The inter-hemispheric differences of N1m decrements during ipsi- and contra-lateral masking reflect a basic hemispheric specialization contributing to the processing of complex auditory stimuli such as speech signals in noisy environments.

  7. Theta synchronization between medial prefrontal cortex and cerebellum is associated with adaptive performance of associative learning behavior

    Science.gov (United States)

    Chen, Hao; Wang, Yi-jie; Yang, Li; Sui, Jian-feng; Hu, Zhi-an; Hu, Bo

    2016-01-01

    Associative learning is thought to require coordinated activities among distributed brain regions. For example, to direct behavior appropriately, the medial prefrontal cortex (mPFC) must encode and maintain sensory information and then interact with the cerebellum during trace eyeblink conditioning (TEBC), a commonly-used associative learning model. However, the mechanisms by which these two distant areas interact remain elusive. By simultaneously recording local field potential (LFP) signals from the mPFC and the cerebellum in guinea pigs undergoing TEBC, we found that theta-frequency (5.0–12.0 Hz) oscillations in the mPFC and the cerebellum became strongly synchronized following presentation of auditory conditioned stimulus. Intriguingly, the conditioned eyeblink response (CR) with adaptive timing occurred preferentially in the trials where mPFC-cerebellum theta coherence was stronger. Moreover, both the mPFC-cerebellum theta coherence and the adaptive CR performance were impaired after the disruption of endogenous orexins in the cerebellum. Finally, association of the mPFC -cerebellum theta coherence with adaptive CR performance was time-limited occurring in the early stage of associative learning. These findings suggest that the mPFC and the cerebellum may act together to contribute to the adaptive performance of associative learning behavior by means of theta synchronization. PMID:26879632

  8. Fast Burst Synchronization for Power Line Communication Systems

    Directory of Open Access Journals (Sweden)

    Lampe Lutz

    2007-01-01

    Full Text Available Fast burst synchronization is an important requirement in asynchronous communication networks, where devices transmit short data packets in an unscheduled fashion. Such a synchronization is typically achieved by means of a preamble sent in front of the data packet. In this paper, we study fast burst synchronization for power line communication (PLC systems operating below 500 kHz and transmitting data rates of up to about 500 kbps as it is typical in various PLC network applications. In particular, we are concerned with the receiver processing of the preamble signal and the actual design of preambles suitable for fast burst synchronization in such PLC systems. Our approach is comprehensive in that it takes into account the most distinctive characteristics of the power line channel, which are multipath propagation, highly varying path loss, and disturbance by impulse noise, as well as important practical constraints, especially the need for spectral shaping of the preamble signal and fast adjustment of the automatic gain control (AGC. In fact, we regard the explicit incorporation of these various requirements into the preamble design as the main contribution of this work. We devise an optimization criterion and a stochastic algorithm to search for suitable preamble sequences. A comprehensive performance comparison of a designed and two conventional preambles shows that the designed sequence is superior in terms of (a fast burst synchronization in various transmission environments, (b fast AGC adjustment, and (c compliance of its spectrum with the spectral mask applied to the data transmit signal.

  9. Primate Auditory Recognition Memory Performance Varies With Sound Type

    OpenAIRE

    Chi-Wing, Ng; Bethany, Plakke; Amy, Poremba

    2009-01-01

    Neural correlates of auditory processing, including for species-specific vocalizations that convey biological and ethological significance (e.g. social status, kinship, environment),have been identified in a wide variety of areas including the temporal and frontal cortices. However, few studies elucidate how non-human primates interact with these vocalization signals when they are challenged by tasks requiring auditory discrimination, recognition, and/or memory. The present study employs a de...

  10. Optical Synchronization Systems for Femtosecond X-Ray Sources

    CERN Document Server

    Wilcox, Russell; Staples, John W

    2005-01-01

    In femtosecond pump/probe experiments using short x-ray and optical pulses, precise synchronization must be maintained between widely separated lasers in a synchrotron or FEL facility. We are developing synchronization systems using optical signals for applications requiring different ranges of timing error. For the sub-100fs range we use an amplitude modulated CW laser at 1GHz to transmit RF phase information, and control the delay through a 100m fiber by observing the retroreflected signal. Initial results show 40fs peak-to-peak error above 10Hz, and 200fs long term drift, mainly due to amplitude sensitivity in the analog mixers. For the sub-10fs range we will lock two single-frequency lasers separated by several teraHertz to a master modelocked fiber laser, transmit the two frequencies over fiber, and lock two comb lines of a slave laser to these frequencies, thus synchronizing the two modelocked laser envelopes. For attosecond synchronization we propose a stabilized, free space link using bulk lens wavegu...

  11. Operator auditory perception and spectral quantification of umbilical artery Doppler ultrasound signals.

    Directory of Open Access Journals (Sweden)

    Ann Thuring

    Full Text Available OBJECTIVE: An experienced sonographer can by listening to the Doppler audio signals perceive various timbres that distinguish different types of umbilical artery flow despite an unchanged pulsatility index (PI. Our aim was to develop an objective measure of the Doppler audio signals recorded from fetoplacental circulation in a sheep model. METHODS: Various degrees of pathological flow velocity waveforms in the umbilical artery, similar to those in human complicated pregnancies, were induced by microsphere embolization of the placental bed (embolization model, 7 lamb fetuses, 370 Doppler recordings or by fetal hemodilution (anemia model, 4 lamb fetuses, 184 recordings. A subjective 11-step operator auditory scale (OAS was related to conventional Doppler parameters, PI and time average mean velocity (TAM, and to sound frequency analysis of Doppler signals (sound frequency with the maximum energy content [MAXpeak] and frequency band at maximum level minus 15 dB [MAXpeak-15 dB] over several heart cycles. RESULTS: WE FOUND A NEGATIVE CORRELATION BETWEEN THE OAS AND PI: median Rho -0.73 (range -0.35- -0.94 and -0.68 (range -0.57- -0.78 in the two lamb models, respectively. There was a positive correlation between OAS and TAM in both models: median Rho 0.80 (range 0.58-0.95 and 0.90 (range 0.78-0.95, respectively. A strong correlation was found between TAM and the results of sound spectrum analysis; in the embolization model the median r was 0.91 (range 0.88-0.97 for MAXpeak and 0.91 (range 0.82-0.98 for MAXpeak-15 dB. In the anemia model, the corresponding values were 0.92 (range 0.78-0.96 and 0.96 (range 0.89-0.98, respectively. CONCLUSION: Audio-spectrum analysis reflects the subjective perception of Doppler sound signals in the umbilical artery and has a strong correlation to TAM-velocity. This information might be of importance for clinical management of complicated pregnancies as an addition to conventional Doppler parameters.

  12. Optical Time-Division Multiplexing of 10 Gbit/s Ethernet Signals Synchronized by All-Optical Signal Processing Based on a Time-Lens

    DEFF Research Database (Denmark)

    Areal, Janaina Laguardia

    This Thesis presents 3 years work of an optical circuit that performs both pulse compression and frame synchronization and retiming. Our design aims at directly multiplexing several 10G Ethernet data packets (frames) to a high-speed OTDM link. This scheme is optically transparent and does not req...... coupler, completing the OTDM signal generation. We demonstrate the effectiveness of the design by laboratory experiments and simulations with VPI and MatLab....... not require clock recovery, resulting in a potentially very efficient solution. The scheme uses a time-lens, implemented through a sinusoidally driven optical phase modulation, combined with a linear dispersion element. As time-lenses are also used for pulse compression, we design the circuit also to perform...

  13. Synchronous machine parameter identification in frequency and time domain

    Directory of Open Access Journals (Sweden)

    Hasni M.

    2007-01-01

    Full Text Available This paper presents the results of a frequency and time-domain identification procedure to estimate the linear parameters of a salient-pole synchronous machine at standstill. The objective of this study is to use several input signals to identify the model structure and parameters of a salient-pole synchronous machine from standstill test data. The procedure consists to define, to conduct the standstill tests and also to identify the model structure. The signals used for identification are the different excitation voltages at standstill and the flowing current in different windings. We estimate the parameters of operational impedances, or in other words the reactance and the time constants. The tests were carried out on synchronous machine of 1.5 kVA 380V 1500 rpm.

  14. Multivalued synchronization by Poincaré coupling

    Science.gov (United States)

    Ontañón-García, L. J.; Campos-Cantón, E.; Femat, R.; Campos-Cantón, I.; Bonilla-Marín, M.

    2013-10-01

    This work presents multivalued chaotic synchronization via coupling based on the Poincaré plane. The coupling is carried out by an underdamped signal, triggered every crossing event of the trajectory of the master system through a previously defined Poincaré plane. A master-slave system is explored, and the synchronization between the systems is detected via the auxiliary system approach and the maximum conditional Lyapunov exponent. Due to the response to specific conditions two phenomena may be obtained: univalued and multivalued synchronization. Since the Lyapunov exponent is not enough to detect these two phenomena, the distance between the pieces of trajectories of the slave and auxiliary systems with different initial conditions is also used as a tool for the detection of multivalued synchronization. Computer simulations using the benchmark chaotic systems of Lorenz and Rössler are used to exemplify the approach proposed.

  15. Instrument to synchronize Thomson scattering diagnostic measurements with MHD acitivity in a tokamak

    International Nuclear Information System (INIS)

    Wintenberg, A.L.

    1985-04-01

    An instrument to synchronize the firing of a ruby laser for a Thomson scattering diagnostic with plasma oscillations was designed, developed, and evaluated. The instrument will fire the laser at a user-selected phase of an input sine or sawtooth wave with an accuracy of +-15 0 . Allowable frequencies range from 20 to 500 Hz for a sawtooth and from 1 to 30 kHz for a sine wave. The instrument also allows synchronization with a sine wave to be enabled by a preselected sawtooth phase. The instrument uses analog signal processing circuits to separate the signal components, remove unwanted components, and produce zero-phase synchronization pulses. The instrument measures the period between zero-phase pulses in order to produce phase synchronization pulses delayed a fraction of the period from the zero-phase pulses. The laser is fired by the phase synchronization pulse. Unwanted signal components are attenuated by bandpass filters. A digitally controlled self-adjusting bandpass filter for sine processing. The instrument was used to investigate the variation of the electron temperature profile with the phase of the x-ray signal from an Impurity Studies Experiment (ISX-B) plasma exhibiting magnetohydrodynamic (MHD) activity

  16. The Role of Inhibition in a Computational Model of an Auditory Cortical Neuron during the Encoding of Temporal Information

    Science.gov (United States)

    Bendor, Daniel

    2015-01-01

    In auditory cortex, temporal information within a sound is represented by two complementary neural codes: a temporal representation based on stimulus-locked firing and a rate representation, where discharge rate co-varies with the timing between acoustic events but lacks a stimulus-synchronized response. Using a computational neuronal model, we find that stimulus-locked responses are generated when sound-evoked excitation is combined with strong, delayed inhibition. In contrast to this, a non-synchronized rate representation is generated when the net excitation evoked by the sound is weak, which occurs when excitation is coincident and balanced with inhibition. Using single-unit recordings from awake marmosets (Callithrix jacchus), we validate several model predictions, including differences in the temporal fidelity, discharge rates and temporal dynamics of stimulus-evoked responses between neurons with rate and temporal representations. Together these data suggest that feedforward inhibition provides a parsimonious explanation of the neural coding dichotomy observed in auditory cortex. PMID:25879843

  17. An all digital phase locked loop for synchronization of a sinusoidal signal embedded in white Gaussian noise

    Science.gov (United States)

    Reddy, C. P.; Gupta, S. C.

    1973-01-01

    An all digital phase locked loop which tracks the phase of the incoming sinusoidal signal once per carrier cycle is proposed. The different elements and their functions and the phase lock operation are explained in detail. The nonlinear difference equations which govern the operation of the digital loop when the incoming signal is embedded in white Gaussian noise are derived, and a suitable model is specified. The performance of the digital loop is considered for the synchronization of a sinusoidal signal. For this, the noise term is suitably modelled which allows specification of the output probabilities for the two level quantizer in the loop at any given phase error. The loop filter considered increases the probability of proper phase correction. The phase error states in modulo two-pi forms a finite state Markov chain which enables the calculation of steady state probabilities, RMS phase error, transient response and mean time for cycle skipping.

  18. Phase attraction in sensorimotor synchronization with auditory sequences: effects of single and periodic distractors on synchronization accuracy.

    Science.gov (United States)

    Repp, Bruno H

    2003-04-01

    Four experiments showed that both single and periodic distractor tones affected the timing of finger taps produced in synchrony with an isochronous auditory target sequence. Single distractors had only small effects, but periodic distractors occurring at various fixed or changing phase relationships exerted strong phase attraction. The attraction was asymmetric, being stronger when distractors preceded target tones than when they lagged behind. A large pitch difference between target and distractor tones (20 vs. 3 semitones) did not reduce phase attraction substantially, although in the case of continuously changing phase relationships it did prevent complete capture of the taps by the distractors. The results support the hypothesis that phase attraction is an automatic process that is sensitive primarily to event onsets.

  19. Modulation of isochronous movements in a flexible environment: links between motion and auditory experience.

    Science.gov (United States)

    Bravi, Riccardo; Del Tongo, Claudia; Cohen, Erez James; Dalle Mura, Gabriele; Tognetti, Alessandro; Minciacchi, Diego

    2014-06-01

    The ability to perform isochronous movements while listening to a rhythmic auditory stimulus requires a flexible process that integrates timing information with movement. Here, we explored how non-temporal and temporal characteristics of an auditory stimulus (presence, interval occupancy, and tempo) affect motor performance. These characteristics were chosen on the basis of their ability to modulate the precision and accuracy of synchronized movements. Subjects have participated in sessions in which they performed sets of repeated isochronous wrist's flexion-extensions under various conditions. The conditions were chosen on the basis of the defined characteristics. Kinematic parameters were evaluated during each session, and temporal parameters were analyzed. In order to study the effects of the auditory stimulus, we have minimized all other sensory information that could interfere with its perception or affect the performance of repeated isochronous movements. The present study shows that the distinct characteristics of an auditory stimulus significantly influence isochronous movements by altering their duration. Results provide evidence for an adaptable control of timing in the audio-motor coupling for isochronous movements. This flexibility would make plausible the use of different encoding strategies to adapt audio-motor coupling for specific tasks.

  20. Response properties of neurons in the cat's putamen during auditory discrimination.

    Science.gov (United States)

    Zhao, Zhenling; Sato, Yu; Qin, Ling

    2015-10-01

    The striatum integrates diverse convergent input and plays a critical role in the goal-directed behaviors. To date, the auditory functions of striatum are less studied. Recently, it was demonstrated that auditory cortico-striatal projections influence behavioral performance during a frequency discrimination task. To reveal the functions of striatal neurons in auditory discrimination, we recorded the single-unit spike activities in the putamen (dorsal striatum) of free-moving cats while performing a Go/No-go task to discriminate the sounds with different modulation rates (12.5 Hz vs. 50 Hz) or envelopes (damped vs. ramped). We found that the putamen neurons can be broadly divided into four groups according to their contributions to sound discrimination. First, 40% of neurons showed vigorous responses synchronized to the sound envelope, and could precisely discriminate different sounds. Second, 18% of neurons showed a high preference of ramped to damped sounds, but no preference for modulation rate. They could only discriminate the change of sound envelope. Third, 27% of neurons rapidly adapted to the sound stimuli, had no ability of sound discrimination. Fourth, 15% of neurons discriminated the sounds dependent on the reward-prediction. Comparing to passively listening condition, the activities of putamen neurons were significantly enhanced by the engagement of the auditory tasks, but not modulated by the cat's behavioral choice. The coexistence of multiple types of neurons suggests that the putamen is involved in the transformation from auditory representation to stimulus-reward association. Copyright © 2015 Elsevier B.V. All rights reserved.

  1. A secure communication scheme using projective chaos synchronization

    International Nuclear Information System (INIS)

    Li Zhigang; Xu Daolin

    2004-01-01

    Most secure communication schemes using chaotic dynamics are based on identical synchronization. In this paper, we show the possibility of secure communication using projective synchronization (PS). The unpredictability of the scaling factor in projective synchronization can additionally enhance the security of communication. It is also showed that the scaling factor can be employed to improve the robustness against noise contamination. The feasibility of the communication scheme in high-dimensional chaotic systems, such as the hyperchaotic Roessler system, is demonstrated. Numerical results show the success in transmitting a sound signal through chaotic systems

  2. Transition among synchronous schemes in coupled nonidentical multiple time delay systems

    International Nuclear Information System (INIS)

    Thang Manh Hoang

    2009-01-01

    We present the transition among possible synchronous schemes in coupled nonidentical multiple time delay systems, i.e., lag, projective-lag, complete, anticipating and projective-anticipating synchronization. The number of nonlinear transforms in the master's equation can be different from that in slave's, and nonlinear transforms can be in various forms. The driving signal is the sum of nonlinearly transformed components of delayed state variable. Moreover, the equation representing for driving signal is constructed exactly so that the difference between the master's and slave's structures is complemented. The sufficient condition for synchronization is considered by the Krasovskii-Lyapunov theory. The specific examples will demonstrate and verify the effectiveness of the proposed models.

  3. Auditory agnosia.

    Science.gov (United States)

    Slevc, L Robert; Shell, Alison R

    2015-01-01

    Auditory agnosia refers to impairments in sound perception and identification despite intact hearing, cognitive functioning, and language abilities (reading, writing, and speaking). Auditory agnosia can be general, affecting all types of sound perception, or can be (relatively) specific to a particular domain. Verbal auditory agnosia (also known as (pure) word deafness) refers to deficits specific to speech processing, environmental sound agnosia refers to difficulties confined to non-speech environmental sounds, and amusia refers to deficits confined to music. These deficits can be apperceptive, affecting basic perceptual processes, or associative, affecting the relation of a perceived auditory object to its meaning. This chapter discusses what is known about the behavioral symptoms and lesion correlates of these different types of auditory agnosia (focusing especially on verbal auditory agnosia), evidence for the role of a rapid temporal processing deficit in some aspects of auditory agnosia, and the few attempts to treat the perceptual deficits associated with auditory agnosia. A clear picture of auditory agnosia has been slow to emerge, hampered by the considerable heterogeneity in behavioral deficits, associated brain damage, and variable assessments across cases. Despite this lack of clarity, these striking deficits in complex sound processing continue to inform our understanding of auditory perception and cognition. © 2015 Elsevier B.V. All rights reserved.

  4. Synchronization of dynamic response measurements for the purpose of structural health monitoring

    International Nuclear Information System (INIS)

    Maes, K.; Reynders, E.; De Roeck, G.; Lombaert, G.; Rezayat, A.

    2016-01-01

    This paper presents a technique for offline time synchronization of data acquisition systems for linear structures with proportional damping. The technique can be applied when direct synchronization of data acquisition systems is impossible or not sufficiently accurate. The synchronization is based on the acquired dynamic response of the structure only, and does not require the acquisition of a shared sensor signal or a trigger signal. The time delay is identified from the spurious phase shift of the mode shape components that are obtained from system identification. A demonstration for a laboratory experiment on a cantilever steel beam shows that the proposed methodology can be used for accurate time synchronization, resulting in a significant improvement of the accuracy of the identified mode shapes. (paper)

  5. Auditory memory for temporal characteristics of sound.

    Science.gov (United States)

    Zokoll, Melanie A; Klump, Georg M; Langemann, Ulrike

    2008-05-01

    This study evaluates auditory memory for variations in the rate of sinusoidal amplitude modulation (SAM) of noise bursts in the European starling (Sturnus vulgaris). To estimate the extent of the starling's auditory short-term memory store, a delayed non-matching-to-sample paradigm was applied. The birds were trained to discriminate between a series of identical "sample stimuli" and a single "test stimulus". The birds classified SAM rates of sample and test stimuli as being either the same or different. Memory performance of the birds was measured as the percentage of correct classifications. Auditory memory persistence time was estimated as a function of the delay between sample and test stimuli. Memory performance was significantly affected by the delay between sample and test and by the number of sample stimuli presented before the test stimulus, but was not affected by the difference in SAM rate between sample and test stimuli. The individuals' auditory memory persistence times varied between 2 and 13 s. The starlings' auditory memory persistence in the present study for signals varying in the temporal domain was significantly shorter compared to that of a previous study (Zokoll et al. in J Acoust Soc Am 121:2842, 2007) applying tonal stimuli varying in the spectral domain.

  6. Auditory capture of visual motion: effects on perception and discrimination.

    Science.gov (United States)

    McCourt, Mark E; Leone, Lynnette M

    2016-09-28

    We asked whether the perceived direction of visual motion and contrast thresholds for motion discrimination are influenced by the concurrent motion of an auditory sound source. Visual motion stimuli were counterphasing Gabor patches, whose net motion energy was manipulated by adjusting the contrast of the leftward-moving and rightward-moving components. The presentation of these visual stimuli was paired with the simultaneous presentation of auditory stimuli, whose apparent motion in 3D auditory space (rightward, leftward, static, no sound) was manipulated using interaural time and intensity differences, and Doppler cues. In experiment 1, observers judged whether the Gabor visual stimulus appeared to move rightward or leftward. In experiment 2, contrast discrimination thresholds for detecting the interval containing unequal (rightward or leftward) visual motion energy were obtained under the same auditory conditions. Experiment 1 showed that the perceived direction of ambiguous visual motion is powerfully influenced by concurrent auditory motion, such that auditory motion 'captured' ambiguous visual motion. Experiment 2 showed that this interaction occurs at a sensory stage of processing as visual contrast discrimination thresholds (a criterion-free measure of sensitivity) were significantly elevated when paired with congruent auditory motion. These results suggest that auditory and visual motion signals are integrated and combined into a supramodal (audiovisual) representation of motion.

  7. Carrier phase synchronization system for improved amplitude modulation and television broadcast reception

    Science.gov (United States)

    Smith, Stephen F [Loudon, TN; Moore, James A [Powell, TN

    2009-09-08

    Systems and methods are described for carrier phase synchronization for improved AM and TV broadcast reception. A method includes synchronizing the phase of a carrier frequency of a broadcast signal with the phase of a remote reference frequency. An apparatus includes a receiver to detect the phase of a reference signal; a phase comparator coupled to the reference signal-phase receiver; a voltage controlled oscillator coupled to the phase comparator; and a phase-controlled radio frequency output coupled to the voltage controlled oscillator.

  8. Children's auditory working memory performance in degraded listening conditions.

    Science.gov (United States)

    Osman, Homira; Sullivan, Jessica R

    2014-08-01

    The objectives of this study were to determine (a) whether school-age children with typical hearing demonstrate poorer auditory working memory performance in multitalker babble at degraded signal-to-noise ratios than in quiet; and (b) whether the amount of cognitive demand of the task contributed to differences in performance in noise. It was hypothesized that stressing the working memory system with the presence of noise would impede working memory processes in real time and result in poorer working memory performance in degraded conditions. Twenty children with typical hearing between 8 and 10 years old were tested using 4 auditory working memory tasks (Forward Digit Recall, Backward Digit Recall, Listening Recall Primary, and Listening Recall Secondary). Stimuli were from the standardized Working Memory Test Battery for Children. Each task was administered in quiet and in 4-talker babble noise at 0 dB and -5 dB signal-to-noise ratios. Children's auditory working memory performance was systematically decreased in the presence of multitalker babble noise compared with quiet. Differences between low-complexity and high-complexity tasks were observed, with children performing more poorly on tasks with greater storage and processing demands. There was no interaction between noise and complexity of task. All tasks were negatively impacted similarly by the addition of noise. Auditory working memory performance was negatively impacted by the presence of multitalker babble noise. Regardless of complexity of task, noise had a similar effect on performance. These findings suggest that the addition of noise inhibits auditory working memory processes in real time for school-age children.

  9. The storage and recall of auditory memory.

    Science.gov (United States)

    Nebenzahl, I; Albeck, Y

    1990-01-01

    The architecture of the auditory memory is investigated. The auditory information is assumed to be represented by f-t patterns. With the help of a psycho-physical experiment it is demonstrated that the storage of these patterns is highly folded in the sense that a long signal is broken into many short stretches before being stored in the memory. Recognition takes place by correlating newly heard input in the short term memory to information previously stored in the long term memory. We show that this correlation is performed after the input is accumulated and held statically in the short term memory.

  10. Quantifying Neural Oscillatory Synchronization: A Comparison between Spectral Coherence and Phase-Locking Value Approaches

    Science.gov (United States)

    Lowet, Eric; Roberts, Mark J.; Bonizzi, Pietro; Karel, Joël; De Weerd, Peter

    2016-01-01

    Synchronization or phase-locking between oscillating neuronal groups is considered to be important for coordination of information among cortical networks. Spectral coherence is a commonly used approach to quantify phase locking between neural signals. We systematically explored the validity of spectral coherence measures for quantifying synchronization among neural oscillators. To that aim, we simulated coupled oscillatory signals that exhibited synchronization dynamics using an abstract phase-oscillator model as well as interacting gamma-generating spiking neural networks. We found that, within a large parameter range, the spectral coherence measure deviated substantially from the expected phase-locking. Moreover, spectral coherence did not converge to the expected value with increasing signal-to-noise ratio. We found that spectral coherence particularly failed when oscillators were in the partially (intermittent) synchronized state, which we expect to be the most likely state for neural synchronization. The failure was due to the fast frequency and amplitude changes induced by synchronization forces. We then investigated whether spectral coherence reflected the information flow among networks measured by transfer entropy (TE) of spike trains. We found that spectral coherence failed to robustly reflect changes in synchrony-mediated information flow between neural networks in many instances. As an alternative approach we explored a phase-locking value (PLV) method based on the reconstruction of the instantaneous phase. As one approach for reconstructing instantaneous phase, we used the Hilbert Transform (HT) preceded by Singular Spectrum Decomposition (SSD) of the signal. PLV estimates have broad applicability as they do not rely on stationarity, and, unlike spectral coherence, they enable more accurate estimations of oscillatory synchronization across a wide range of different synchronization regimes, and better tracking of synchronization-mediated information

  11. Quantifying Neural Oscillatory Synchronization: A Comparison between Spectral Coherence and Phase-Locking Value Approaches.

    Directory of Open Access Journals (Sweden)

    Eric Lowet

    Full Text Available Synchronization or phase-locking between oscillating neuronal groups is considered to be important for coordination of information among cortical networks. Spectral coherence is a commonly used approach to quantify phase locking between neural signals. We systematically explored the validity of spectral coherence measures for quantifying synchronization among neural oscillators. To that aim, we simulated coupled oscillatory signals that exhibited synchronization dynamics using an abstract phase-oscillator model as well as interacting gamma-generating spiking neural networks. We found that, within a large parameter range, the spectral coherence measure deviated substantially from the expected phase-locking. Moreover, spectral coherence did not converge to the expected value with increasing signal-to-noise ratio. We found that spectral coherence particularly failed when oscillators were in the partially (intermittent synchronized state, which we expect to be the most likely state for neural synchronization. The failure was due to the fast frequency and amplitude changes induced by synchronization forces. We then investigated whether spectral coherence reflected the information flow among networks measured by transfer entropy (TE of spike trains. We found that spectral coherence failed to robustly reflect changes in synchrony-mediated information flow between neural networks in many instances. As an alternative approach we explored a phase-locking value (PLV method based on the reconstruction of the instantaneous phase. As one approach for reconstructing instantaneous phase, we used the Hilbert Transform (HT preceded by Singular Spectrum Decomposition (SSD of the signal. PLV estimates have broad applicability as they do not rely on stationarity, and, unlike spectral coherence, they enable more accurate estimations of oscillatory synchronization across a wide range of different synchronization regimes, and better tracking of synchronization

  12. A Theoretical Framework for Soft-Information-Based Synchronization in Iterative (Turbo Receivers

    Directory of Open Access Journals (Sweden)

    Lottici Vincenzo

    2005-01-01

    Full Text Available This contribution considers turbo synchronization, that is to say, the use of soft data information to estimate parameters like carrier phase, frequency, or timing offsets of a modulated signal within an iterative data demodulator. In turbo synchronization, the receiver exploits the soft decisions computed at each turbo decoding iteration to provide a reliable estimate of some signal parameters. The aim of our paper is to show that such "turbo-estimation" approach can be regarded as a special case of the expectation-maximization (EM algorithm. This leads to a general theoretical framework for turbo synchronization that allows to derive parameter estimation procedures for carrier phase and frequency offset, as well as for timing offset and signal amplitude. The proposed mathematical framework is illustrated by simulation results reported for the particular case of carrier phase and frequency offsets estimation of a turbo-coded 16-QAM signal.

  13. Central auditory neurons have composite receptive fields.

    Science.gov (United States)

    Kozlov, Andrei S; Gentner, Timothy Q

    2016-02-02

    High-level neurons processing complex, behaviorally relevant signals are sensitive to conjunctions of features. Characterizing the receptive fields of such neurons is difficult with standard statistical tools, however, and the principles governing their organization remain poorly understood. Here, we demonstrate multiple distinct receptive-field features in individual high-level auditory neurons in a songbird, European starling, in response to natural vocal signals (songs). We then show that receptive fields with similar characteristics can be reproduced by an unsupervised neural network trained to represent starling songs with a single learning rule that enforces sparseness and divisive normalization. We conclude that central auditory neurons have composite receptive fields that can arise through a combination of sparseness and normalization in neural circuits. Our results, along with descriptions of random, discontinuous receptive fields in the central olfactory neurons in mammals and insects, suggest general principles of neural computation across sensory systems and animal classes.

  14. Individualization of music-based rhythmic auditory cueing in Parkinson's disease.

    Science.gov (United States)

    Bella, Simone Dalla; Dotov, Dobromir; Bardy, Benoît; de Cock, Valérie Cochen

    2018-06-04

    Gait dysfunctions in Parkinson's disease can be partly relieved by rhythmic auditory cueing. This consists in asking patients to walk with a rhythmic auditory stimulus such as a metronome or music. The effect on gait is visible immediately in terms of increased speed and stride length. Moreover, training programs based on rhythmic cueing can have long-term benefits. The effect of rhythmic cueing, however, varies from one patient to the other. Patients' response to the stimulation may depend on rhythmic abilities, often deteriorating with the disease. Relatively spared abilities to track the beat favor a positive response to rhythmic cueing. On the other hand, most patients with poor rhythmic abilities either do not respond to the cues or experience gait worsening when walking with cues. An individualized approach to rhythmic auditory cueing with music is proposed to cope with this variability in patients' response. This approach calls for using assistive mobile technologies capable of delivering cues that adapt in real time to patients' gait kinematics, thus affording step synchronization to the beat. Individualized rhythmic cueing can provide a safe and cost-effective alternative to standard cueing that patients may want to use in their everyday lives. © 2018 New York Academy of Sciences.

  15. Time signal distribution in communication networks based on synchronous digital hierarchy

    Science.gov (United States)

    Imaoka, Atsushi; Kihara, Masami

    1993-01-01

    A new method that uses round-trip paths to accurately measure transmission delay for time synchronization is proposed. The performance of the method in Synchronous Digital Hierarchy networks is discussed. The feature of this method is that it separately measures the initial round trip path delay and the variations in round-trip path delay. The delay generated in SDH equipment is determined by measuring the initial round-trip path delay. In an experiment with actual SDH equipment, the error of initial delay measurement was suppressed to 30ns.

  16. Selective Attention to Auditory Memory Neurally Enhances Perceptual Precision.

    Science.gov (United States)

    Lim, Sung-Joo; Wöstmann, Malte; Obleser, Jonas

    2015-12-09

    Selective attention to a task-relevant stimulus facilitates encoding of that stimulus into a working memory representation. It is less clear whether selective attention also improves the precision of a stimulus already represented in memory. Here, we investigate the behavioral and neural dynamics of selective attention to representations in auditory working memory (i.e., auditory objects) using psychophysical modeling and model-based analysis of electroencephalographic signals. Human listeners performed a syllable pitch discrimination task where two syllables served as to-be-encoded auditory objects. Valid (vs neutral) retroactive cues were presented during retention to allow listeners to selectively attend to the to-be-probed auditory object in memory. Behaviorally, listeners represented auditory objects in memory more precisely (expressed by steeper slopes of a psychometric curve) and made faster perceptual decisions when valid compared to neutral retrocues were presented. Neurally, valid compared to neutral retrocues elicited a larger frontocentral sustained negativity in the evoked potential as well as enhanced parietal alpha/low-beta oscillatory power (9-18 Hz) during memory retention. Critically, individual magnitudes of alpha oscillatory power (7-11 Hz) modulation predicted the degree to which valid retrocues benefitted individuals' behavior. Our results indicate that selective attention to a specific object in auditory memory does benefit human performance not by simply reducing memory load, but by actively engaging complementary neural resources to sharpen the precision of the task-relevant object in memory. Can selective attention improve the representational precision with which objects are held in memory? And if so, what are the neural mechanisms that support such improvement? These issues have been rarely examined within the auditory modality, in which acoustic signals change and vanish on a milliseconds time scale. Introducing a new auditory memory

  17. The multi-level impact of chronic intermittent hypoxia on central auditory processing.

    Science.gov (United States)

    Wong, Eddie; Yang, Bin; Du, Lida; Ho, Wai Hong; Lau, Condon; Ke, Ya; Chan, Ying Shing; Yung, Wing Ho; Wu, Ed X

    2017-08-01

    During hypoxia, the tissues do not obtain adequate oxygen. Chronic hypoxia can lead to many health problems. A relatively common cause of chronic hypoxia is sleep apnea. Sleep apnea is a sleep breathing disorder that affects 3-7% of the population. During sleep, the patient's breathing starts and stops. This can lead to hypertension, attention deficits, and hearing disorders. In this study, we apply an established chronic intermittent hypoxemia (CIH) model of sleep apnea to study its impact on auditory processing. Adult rats were reared for seven days during sleeping hours in a gas chamber with oxygen level cycled between 10% and 21% (normal atmosphere) every 90s. During awake hours, the subjects were housed in standard conditions with normal atmosphere. CIH treatment significantly reduces arterial oxygen partial pressure and oxygen saturation during sleeping hours (relative to controls). After treatment, subjects underwent functional magnetic resonance imaging (fMRI) with broadband sound stimulation. Responses are observed in major auditory centers in all subjects, including the auditory cortex (AC) and auditory midbrain. fMRI signals from the AC are statistically significantly increased after CIH by 0.13% in the contralateral hemisphere and 0.10% in the ipsilateral hemisphere. In contrast, signals from the lateral lemniscus of the midbrain are significantly reduced by 0.39%. Signals from the neighboring inferior colliculus of the midbrain are relatively unaffected. Chronic hypoxia affects multiple levels of the auditory system and these changes are likely related to hearing disorders associated with sleep apnea. Copyright © 2017 Elsevier Inc. All rights reserved.

  18. Detecting phase synchronization between coupled non-phase-coherent oscillators

    International Nuclear Information System (INIS)

    Follmann, Rosangela; Macau, Elbert E.N.; Rosa, Epaminondas

    2009-01-01

    We compare two methods for detecting phase synchronization in coupled non-phase-coherent oscillators. One method is based on the locking of self-sustained oscillators with an irregular signal. The other uses trajectory recurrences in phase space. We identify the pros and cons of both methods and propose guidelines to detect phase synchronization in data series.

  19. Effects of STN DBS and auditory cueing on the performance of sequential movements and the occurrence of action tremor in Parkinson’s disease

    NARCIS (Netherlands)

    Heida, Tjitske; Wentink, E.C.; Zhao, Yan; Marani, Enrico

    2014-01-01

    Background: Parkinson’s disease (PD) patients show a higher ability to perform repetitive movements when they are cued by external stimuli, suggesting that rhythmic synchronization with an auditory timekeeper can be achieved in the absence of intact basal ganglia function. Deep brain stimulation

  20. Synchronization Between Two Different Switched Chaotic Systems By Switching Control

    Directory of Open Access Journals (Sweden)

    Du Li Ming

    2016-01-01

    Full Text Available This paper is concerned with the synchronization problem of two different switched chaotic systems, considering the general case that the master-slave switched chaotic systems have uncertainties. Two basic problems are considered: one is projective synchronization of switched chaotic systems under arbitrary switching; the other is projective synchronization of switched chaotic systems by design of switching when synchronization cannot achieved by using any subsystems alone. For the two problems, common Lyapunov function method and multiple Lyapunov function method are used respectively, an adaptive control scheme has been presented, some sufficient synchronization conditions are attainted, and the switching signal is designed. Finally, the numerical simulation is provide to show the effectiveness of our method.

  1. A new type of chaotic synchronization with application to communication systems

    International Nuclear Information System (INIS)

    Kharel, Rupak; Busawon, Krishna

    2011-01-01

    In this paper, we propose a new methodology to synchronize a class of chaotic systems starting from different initial conditions under some given conditions. The method we propose is not based on the unidirectional synchronization method like the one proposed by Pecora-Caroll. The proposed method is unique in the sense that the chaotic oscillators to be synchronized have no direct connection between them; that is, there is no signal being sent from one to the other. Simulation result is presented to show the synchronization performance.

  2. Speed Sensorless Control of Permanent Magnet Synchronous Motors in Mine Electric Locomotive Drive

    Directory of Open Access Journals (Sweden)

    Yudong LI

    2014-04-01

    Full Text Available This paper presents a novel sensorless control method of permanent magnet synchronous motors a low speed based on a high-frequency voltage signal injection. The approach superimposes a persistent HF voltage signal into the estimated d-axis to get the rotor position error angle-related signal by detecting the corresponding voltage response and current response. Then the rotor position and motor speed are obtained. Theoretical analysis and simulation results demonstrate that the approach can achieve sensorless control of permanent magnet synchronous motors at zero and low speed, ensure good dynamic and static performances, and achieve effective control when applied to servo system. Finally, a test prototype system which used a digital signal processor and space vector pulse width modulation technology has been developed. Experimental results show that the system has better static, the effectiveness and dynamic performance of the adaptive test signals in a sensorless controlled surface-mounted permanent magnet synchronous machines.

  3. Hair cell regeneration in the avian auditory epithelium.

    Science.gov (United States)

    Stone, Jennifer S; Cotanche, Douglas A

    2007-01-01

    Regeneration of sensory hair cells in the mature avian inner ear was first described just over 20 years ago. Since then, it has been shown that many other non-mammalian species either continually produce new hair cells or regenerate them in response to trauma. However, mammals exhibit limited hair cell regeneration, particularly in the auditory epithelium. In birds and other non-mammals, regenerated hair cells arise from adjacent non-sensory (supporting) cells. Hair cell regeneration was initially described as a proliferative response whereby supporting cells re-enter the mitotic cycle, forming daughter cells that differentiate into either hair cells or supporting cells and thereby restore cytoarchitecture and function in the sensory epithelium. However, further analyses of the avian auditory epithelium (and amphibian vestibular epithelium) revealed a second regenerative mechanism, direct transdifferentiation, during which supporting cells change their gene expression and convert into hair cells without dividing. In the chicken auditory epithelium, these two distinct mechanisms show unique spatial and temporal patterns, suggesting they are differentially regulated. Current efforts are aimed at identifying signals that maintain supporting cells in a quiescent state or direct them to undergo direct transdifferentiation or cell division. Here, we review current knowledge about supporting cell properties and discuss candidate signaling molecules for regulating supporting cell behavior, in quiescence and after damage. While significant advances have been made in understanding regeneration in non-mammals over the last 20 years, we have yet to determine why the mammalian auditory epithelium lacks the ability to regenerate hair cells spontaneously and whether it is even capable of significant regeneration under additional circumstances. The continued study of mechanisms controlling regeneration in the avian auditory epithelium may lead to strategies for inducing

  4. Auditory and audio-visual processing in patients with cochlear, auditory brainstem, and auditory midbrain implants: An EEG study.

    Science.gov (United States)

    Schierholz, Irina; Finke, Mareike; Kral, Andrej; Büchner, Andreas; Rach, Stefan; Lenarz, Thomas; Dengler, Reinhard; Sandmann, Pascale

    2017-04-01

    There is substantial variability in speech recognition ability across patients with cochlear implants (CIs), auditory brainstem implants (ABIs), and auditory midbrain implants (AMIs). To better understand how this variability is related to central processing differences, the current electroencephalography (EEG) study compared hearing abilities and auditory-cortex activation in patients with electrical stimulation at different sites of the auditory pathway. Three different groups of patients with auditory implants (Hannover Medical School; ABI: n = 6, CI: n = 6; AMI: n = 2) performed a speeded response task and a speech recognition test with auditory, visual, and audio-visual stimuli. Behavioral performance and cortical processing of auditory and audio-visual stimuli were compared between groups. ABI and AMI patients showed prolonged response times on auditory and audio-visual stimuli compared with NH listeners and CI patients. This was confirmed by prolonged N1 latencies and reduced N1 amplitudes in ABI and AMI patients. However, patients with central auditory implants showed a remarkable gain in performance when visual and auditory input was combined, in both speech and non-speech conditions, which was reflected by a strong visual modulation of auditory-cortex activation in these individuals. In sum, the results suggest that the behavioral improvement for audio-visual conditions in central auditory implant patients is based on enhanced audio-visual interactions in the auditory cortex. Their findings may provide important implications for the optimization of electrical stimulation and rehabilitation strategies in patients with central auditory prostheses. Hum Brain Mapp 38:2206-2225, 2017. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  5. Synchronous monitoring of muscle dynamics and electromyogram

    Science.gov (United States)

    Zakir Hossain, M.; Grill, Wolfgang

    2011-04-01

    A non-intrusive novel detection scheme has been implemented to detect the lateral muscle extension, force of the skeletal muscle and the motor action potential (EMG) synchronously. This allows the comparison of muscle dynamics and EMG signals as a basis for modeling and further studies to determine which architectural parameters are most sensitive to changes in muscle activity. For this purpose the transmission time for ultrasonic chirp signal in the frequency range of 100 kHz to 2.5 MHz passing through the muscle under observation and respective motor action potentials are recorded synchronously to monitor and quantify biomechanical parameters related to muscle performance. Additionally an ultrasonic force sensor has been employed for monitoring. Ultrasonic traducers are placed on the skin to monitor muscle expansion. Surface electrodes are placed suitably to pick up the potential for activation of the monitored muscle. Isometric contraction of the monitored muscle is ensured by restricting the joint motion with the ultrasonic force sensor. Synchronous monitoring was initiated by a software activated audio beep starting at zero time of the subsequent data acquisition interval. Computer controlled electronics are used to generate and detect the ultrasonic signals and monitor the EMG signals. Custom developed software and data analysis is employed to analyze and quantify the monitored data. Reaction time, nerve conduction speed, latent period between the on-set of EMG signals and muscle response, degree of muscle activation and muscle fatigue development, rate of energy expenditure and motor neuron recruitment rate in isometric contraction, and other relevant parameters relating to muscle performance have been quantified with high spatial and temporal resolution.

  6. Public channel cryptography: chaos synchronization and Hilbert's tenth problem.

    Science.gov (United States)

    Kanter, Ido; Kopelowitz, Evi; Kinzel, Wolfgang

    2008-08-22

    The synchronization process of two mutually delayed coupled deterministic chaotic maps is demonstrated both analytically and numerically. The synchronization is preserved when the mutually transmitted signals are concealed by two commutative private filters, a convolution of the truncated time-delayed output signals or some powers of the delayed output signals. The task of a passive attacker is mapped onto Hilbert's tenth problem, solving a set of nonlinear Diophantine equations, which was proven to be in the class of NP-complete problems [problems that are both NP (verifiable in nondeterministic polynomial time) and NP-hard (any NP problem can be translated into this problem)]. This bridge between nonlinear dynamics and NP-complete problems opens a horizon for new types of secure public-channel protocols.

  7. Synchronization of bursting neurons with a slowly varying d. c. current

    International Nuclear Information System (INIS)

    Upadhyay, Ranjit Kumar; Mondal, Argha

    2017-01-01

    Highlights: • To examine synchronization, noisy chemical and electrical coupling have been considered for a coupled bursting M-L neurons. • Bursting presents the precursor to spike synchronization and coupling strength increases the locking between neurons (anti phase and in phase). • The stability of synchronization is established via similarity function. • The necessary condition to occur CS state is observed using master stability function. • A network of four M-L neurons is considered to observe the synchronization. - Abstract: Bursting of neuronal firing is an interesting dynamical consequences depending on fast/slow dynamics. Certain cells in different brain regions produce spike-burst activity. We study such firing activity and its transitions to synchronization using identical as well as non-identical coupled bursting Morris-Lecar (M-L) neurons. Synchronization of different firing activity is a multi-time-scale phenomenon and burst synchronization presents the precursor to spike synchronization. Chemical synapses are one of the dynamical means of information processing between neurons. Electrical synapses play a major role for synchronous activity in a certain network of neurons. Synaptically coupled neural cells exhibit different types of synchronization such as in phase or anti-phase depending on the nature and strength of coupling functions and the synchronization regimes are analyzed by similarity functions. The sequential transitions to synchronization regime are examined by the maximum transverse Lyapunov exponents. Synchronization of voltage traces of two types of planar bursting mechanisms is explored for both kind of synapses under realistic conditions. The noisy influence effects on the transmission of signals and strongly acts to the firing activity (such as periodic firing and bursting) and integration of signals for a network. It has been examined using the mean interspike interval analysis. The transition to synchronization states of

  8. Using auditory steady state responses to outline the functional connectivity in the tinnitus brain.

    Directory of Open Access Journals (Sweden)

    Winfried Schlee

    Full Text Available BACKGROUND: Tinnitus is an auditory phantom perception that is most likely generated in the central nervous system. Most of the tinnitus research has concentrated on the auditory system. However, it was suggested recently that also non-auditory structures are involved in a global network that encodes subjective tinnitus. We tested this assumption using auditory steady state responses to entrain the tinnitus network and investigated long-range functional connectivity across various non-auditory brain regions. METHODS AND FINDINGS: Using whole-head magnetoencephalography we investigated cortical connectivity by means of phase synchronization in tinnitus subjects and healthy controls. We found evidence for a deviating pattern of long-range functional connectivity in tinnitus that was strongly correlated with individual ratings of the tinnitus percept. Phase couplings between the anterior cingulum and the right frontal lobe and phase couplings between the anterior cingulum and the right parietal lobe showed significant condition x group interactions and were correlated with the individual tinnitus distress ratings only in the tinnitus condition and not in the control conditions. CONCLUSIONS: To the best of our knowledge this is the first study that demonstrates existence of a global tinnitus network of long-range cortical connections outside the central auditory system. This result extends the current knowledge of how tinnitus is generated in the brain. We propose that this global extend of the tinnitus network is crucial for the continuos perception of the tinnitus tone and a therapeutical intervention that is able to change this network should result in relief of tinnitus.

  9. Synchronized tapping facilitates learning sound sequences as indexed by the P300.

    Science.gov (United States)

    Kamiyama, Keiko S; Okanoya, Kazuo

    2014-01-01

    The purpose of the present study was to determine whether and how single finger tapping in synchrony with sound sequences contributed to the auditory processing of them. The participants learned two unfamiliar sound sequences via different methods. In the tapping condition, they learned an auditory sequence while they tapped in synchrony with each sound onset. In the no tapping condition, they learned another sequence while they kept pressing a key until the sequence ended. After these learning sessions, we presented the two melodies again and recorded event-related potentials (ERPs). During the ERP recordings, 10% of the tones within each melody deviated from the original tones. An analysis of the grand average ERPs showed that deviant stimuli elicited a significant P300 in the tapping but not in the no-tapping condition. In addition, the significance of the P300 effect in the tapping condition increased as the participants showed highly synchronized tapping behavior during the learning sessions. These results indicated that single finger tapping promoted the conscious detection and evaluation of deviants within the learned sequences. The effect was related to individuals' musical ability to coordinate their finger movements along with external auditory events.

  10. The Role of Head Movements and Signal Spectrum in an Auditory Front/Back Illusion

    Directory of Open Access Journals (Sweden)

    W Owen Brimijoin

    2012-04-01

    Full Text Available We used a dynamic auditory spatial illusion to investigate the role of self-motion and acoustics in shaping our spatial percept of the environment. Using motion capture, we smoothly moved a sound source around listeners as a function of their own head movements. A lowpass filtered sound behind a listener that moved in the direction it would have moved if it had been located in the front was perceived as statically located in front. The contrariwise effect occurred if the sound was in front but moved as if it were behind. The illusion was strongest for sounds lowpass filtered at 500 Hz and weakened as a function of increasing lowpass cutoff frequency. The signals with the most high frequency energy were often associated with an unstable location percept that flickered from front to back as self-motion cues and spectral cues for location came into conflict with one another.

  11. Perception of stochastically undersampled sound waveforms: A model of auditory deafferentation

    Directory of Open Access Journals (Sweden)

    Enrique A Lopez-Poveda

    2013-07-01

    Full Text Available Auditory deafferentation, or permanent loss of auditory nerve afferent terminals, occurs after noise overexposure and aging and may accompany many forms of hearing loss. It could cause significant auditory impairment but is undetected by regular clinical tests and so its effects on perception are poorly understood. Here, we hypothesize and test a neural mechanism by which deafferentation could deteriorate perception. The basic idea is that the spike train produced by each auditory afferent resembles a stochastically digitized version of the sound waveform and that the quality of the waveform representation in the whole nerve depends on the number of aggregated spike trains or auditory afferents. We reason that because spikes occur stochastically in time with a higher probability for high- than for low-intensity sounds, more afferents would be required for the nerve to faithfully encode high-frequency or low-intensity waveform features than low-frequency or high-intensity features. Deafferentation would thus degrade the encoding of these features. We further reason that due to the stochastic nature of nerve firing, the degradation would be greater in noise than in quiet. This hypothesis is tested using a vocoder. Sounds were filtered through ten adjacent frequency bands. For the signal in each band, multiple stochastically subsampled copies were obtained to roughly mimic different stochastic representations of that signal conveyed by different auditory afferents innervating a given cochlear region. These copies were then aggregated to obtain an acoustic stimulus. Tone detection and speech identification tests were performed by young, normal-hearing listeners using different numbers of stochastic samplers per frequency band in the vocoder. Results support the hypothesis that stochastic undersampling of the sound waveform, inspired by deafferentation, impairs speech perception in noise more than in quiet, consistent with auditory aging effects.

  12. Perception of stochastically undersampled sound waveforms: a model of auditory deafferentation

    Science.gov (United States)

    Lopez-Poveda, Enrique A.; Barrios, Pablo

    2013-01-01

    Auditory deafferentation, or permanent loss of auditory nerve afferent terminals, occurs after noise overexposure and aging and may accompany many forms of hearing loss. It could cause significant auditory impairment but is undetected by regular clinical tests and so its effects on perception are poorly understood. Here, we hypothesize and test a neural mechanism by which deafferentation could deteriorate perception. The basic idea is that the spike train produced by each auditory afferent resembles a stochastically digitized version of the sound waveform and that the quality of the waveform representation in the whole nerve depends on the number of aggregated spike trains or auditory afferents. We reason that because spikes occur stochastically in time with a higher probability for high- than for low-intensity sounds, more afferents would be required for the nerve to faithfully encode high-frequency or low-intensity waveform features than low-frequency or high-intensity features. Deafferentation would thus degrade the encoding of these features. We further reason that due to the stochastic nature of nerve firing, the degradation would be greater in noise than in quiet. This hypothesis is tested using a vocoder. Sounds were filtered through ten adjacent frequency bands. For the signal in each band, multiple stochastically subsampled copies were obtained to roughly mimic different stochastic representations of that signal conveyed by different auditory afferents innervating a given cochlear region. These copies were then aggregated to obtain an acoustic stimulus. Tone detection and speech identification tests were performed by young, normal-hearing listeners using different numbers of stochastic samplers per frequency band in the vocoder. Results support the hypothesis that stochastic undersampling of the sound waveform, inspired by deafferentation, impairs speech perception in noise more than in quiet, consistent with auditory aging effects. PMID:23882176

  13. Auditory Reserve and the Legacy of Auditory Experience

    Directory of Open Access Journals (Sweden)

    Erika Skoe

    2014-11-01

    Full Text Available Musical training during childhood has been linked to more robust encoding of sound later in life. We take this as evidence for an auditory reserve: a mechanism by which individuals capitalize on earlier life experiences to promote auditory processing. We assert that early auditory experiences guide how the reserve develops and is maintained over the lifetime. Experiences that occur after childhood, or which are limited in nature, are theorized to affect the reserve, although their influence on sensory processing may be less long-lasting and may potentially fade over time if not repeated. This auditory reserve may help to explain individual differences in how individuals cope with auditory impoverishment or loss of sensorineural function.

  14. A synchronization system to digitalize TJ-1 Tokamak data

    International Nuclear Information System (INIS)

    Guasp, J.; Perez-Navarro, A.; Pacios, L.

    1983-01-01

    At TJ-1 Tokamak signals are stored on a 60-channel magnetic memory. In this report, a system to address those channels and synchronize readout is presented. Digitalized signals are stored in structured files on PDP-11/34 magnetic disks. (author)

  15. Modeling auditory processing and speech perception in hearing-impaired listeners

    DEFF Research Database (Denmark)

    Jepsen, Morten Løve

    in a diagnostic rhyme test. The framework was constructed such that discrimination errors originating from the front-end and the back-end were separated. The front-end was fitted to individual listeners with cochlear hearing loss according to non-speech data, and speech data were obtained in the same listeners......A better understanding of how the human auditory system represents and analyzes sounds and how hearing impairment affects such processing is of great interest for researchers in the fields of auditory neuroscience, audiology, and speech communication as well as for applications in hearing......-instrument and speech technology. In this thesis, the primary focus was on the development and evaluation of a computational model of human auditory signal-processing and perception. The model was initially designed to simulate the normal-hearing auditory system with particular focus on the nonlinear processing...

  16. Multivariate sensitivity to voice during auditory categorization.

    Science.gov (United States)

    Lee, Yune Sang; Peelle, Jonathan E; Kraemer, David; Lloyd, Samuel; Granger, Richard

    2015-09-01

    Past neuroimaging studies have documented discrete regions of human temporal cortex that are more strongly activated by conspecific voice sounds than by nonvoice sounds. However, the mechanisms underlying this voice sensitivity remain unclear. In the present functional MRI study, we took a novel approach to examining voice sensitivity, in which we applied a signal detection paradigm to the assessment of multivariate pattern classification among several living and nonliving categories of auditory stimuli. Within this framework, voice sensitivity can be interpreted as a distinct neural representation of brain activity that correctly distinguishes human vocalizations from other auditory object categories. Across a series of auditory categorization tests, we found that bilateral superior and middle temporal cortex consistently exhibited robust sensitivity to human vocal sounds. Although the strongest categorization was in distinguishing human voice from other categories, subsets of these regions were also able to distinguish reliably between nonhuman categories, suggesting a general role in auditory object categorization. Our findings complement the current evidence of cortical sensitivity to human vocal sounds by revealing that the greatest sensitivity during categorization tasks is devoted to distinguishing voice from nonvoice categories within human temporal cortex. Copyright © 2015 the American Physiological Society.

  17. A Synchronous Multi-Body Sensor Platform in a Wireless Body Sensor Network: Design and Implementation

    Science.gov (United States)

    Gil, Yeongjoon; Wu, Wanqing; Lee, Jungtae

    2012-01-01

    Background Human life can be further improved if diseases and disorders can be predicted before they become dangerous, by correctly recognizing signals from the human body, so in order to make disease detection more precise, various body-signals need to be measured simultaneously in a synchronized manner. Object This research aims at developing an integrated system for measuring four signals (EEG, ECG, respiration, and PPG) and simultaneously producing synchronous signals on a Wireless Body Sensor Network. Design We designed and implemented a platform for multiple bio-signals using Bluetooth communication. Results First, we developed a prototype board and verified the signals from the sensor platform using frequency responses and quantities. Next, we designed and implemented a lightweight, ultra-compact, low cost, low power-consumption Printed Circuit Board. Conclusion A synchronous multi-body sensor platform is expected to be very useful in telemedicine and emergency rescue scenarios. Furthermore, this system is expected to be able to analyze the mutual effects among body signals. PMID:23112605

  18. A Synchronous Multi-Body Sensor Platform in a Wireless Body Sensor Network: Design and Implementation

    Directory of Open Access Journals (Sweden)

    Jungtae Lee

    2012-07-01

    Full Text Available Background: Human life can be further improved if diseases and disorders can be predicted before they become dangerous, by correctly recognizing signals from the human body, so in order to make disease detection more precise, various body-signals need to be measured simultaneously in a synchronized manner. Object: This research aims at developing an integrated system for measuring four signals (EEG, ECG, respiration, and PPG and simultaneously producing synchronous signals on a Wireless Body Sensor Network. Design: We designed and implemented a platform for multiple bio-signals using Bluetooth communication. Results: First, we developed a prototype board and verified the signals from the sensor platform using frequency responses and quantities. Next, we designed and implemented a lightweight, ultra-compact, low cost, low power-consumption Printed Circuit Board. Conclusion: A synchronous multi-body sensor platform is expected to be very useful in telemedicine and emergency rescue scenarios. Furthermore, this system is expected to be able to analyze the mutual effects among body signals.

  19. Neural Correlates of Auditory Processing, Learning and Memory Formation in Songbirds

    Science.gov (United States)

    Pinaud, R.; Terleph, T. A.; Wynne, R. D.; Tremere, L. A.

    Songbirds have emerged as powerful experimental models for the study of auditory processing of complex natural communication signals. Intact hearing is necessary for several behaviors in developing and adult animals including vocal learning, territorial defense, mate selection and individual recognition. These behaviors are thought to require the processing, discrimination and memorization of songs. Although much is known about the brain circuits that participate in sensorimotor (auditory-vocal) integration, especially the ``song-control" system, less is known about the anatomical and functional organization of central auditory pathways. Here we discuss findings associated with a telencephalic auditory area known as the caudomedial nidopallium (NCM). NCM has attracted significant interest as it exhibits functional properties that may support higher order auditory functions such as stimulus discrimination and the formation of auditory memories. NCM neurons are vigorously dr iven by auditory stimuli. Interestingly, these responses are selective to conspecific, relative to heterospecific songs and artificial stimuli. In addition, forms of experience-dependent plasticity occur in NCM and are song-specific. Finally, recent experiments employing high-throughput quantitative proteomics suggest that complex protein regulatory pathways are engaged in NCM as a result of auditory experience. These molecular cascades are likely central to experience-associated plasticity of NCM circuitry and may be part of a network of calcium-driven molecular events that support the formation of auditory memory traces.

  20. Distributed cooperative synchronization strategy for multi-bus microgrids

    DEFF Research Database (Denmark)

    Sun, Yao; Zhong, Chaolu; Hou, Xiaochao

    2017-01-01

    Microgrids can operate in both grid-connected mode and islanded mode. In order to smooth transfer from islanded mode to grid-connected mode, it is necessary to synchronize the point of common coupling (PCC) with main utility grid (UG) in voltage frequency, phase and amplitude. Conventional...... synchronization methods based on centralized communication are very costly and not suitable for multi-bus microgrids that have a large number of distributed generators (DGs). To address this concern, this study presents an active synchronization control strategy based on distributed cooperation technology...... for multi-bus microgrids. The proposed method can reconnect the microgrid in island to UG seamlessly with sparse communication channels. Synchronization correction signals are generated by a voltage controller, which are only transmitted to the leader DGs. Meanwhile, each DG exchanges information with its...

  1. Efficient synchronization of structurally adaptive coupled Hindmarsh-Rose neurons

    International Nuclear Information System (INIS)

    Moujahid, A.; D'Anjou, A.; Torrealdea, F.J.; Torrealdea, F.

    2011-01-01

    Highlights: → Neural activity might be constrained by a requirement of energy efficiency. → Signaling in synchrony is a normal way to propagate information between neurons. → Quality of synchrony affects the capacity to exchange information and the energy cost. → Adaptive mechanism leads to high degree of synchronization between nonidentical neurons. - Abstract: The use of spikes to carry information between brain areas implies complete or partial synchronization of the neurons involved. The degree of synchronization reached by two coupled systems and the energy cost of maintaining their synchronized behavior is highly dependent on the nature of the systems. For non-identical systems the maintenance of a synchronized regime is energetically a costly process. In this work, we study conditions under which two non-identical electrically coupled neurons can reach an efficient regime of synchronization at low energy cost. We show that the energy consumption required to keep the synchronized regime can be spontaneously reduced if the receiving neuron has adaptive mechanisms able to bring its biological parameters closer in value to the corresponding ones in the sending neuron.

  2. Synchronization of coupled metronomes on two layers

    Science.gov (United States)

    Zhang, Jing; Yu, Yi-Zhen; Wang, Xin-Gang

    2017-12-01

    Coupled metronomes serve as a paradigmatic model for exploring the collective behaviors of complex dynamical systems, as well as a classical setup for classroom demonstrations of synchronization phenomena. Whereas previous studies of metronome synchronization have been concentrating on symmetric coupling schemes, here we consider the asymmetric case by adopting the scheme of layered metronomes. Specifically, we place two metronomes on each layer, and couple two layers by placing one on top of the other. By varying the initial conditions of the metronomes and adjusting the friction between the two layers, a variety of synchronous patterns are observed in experiment, including the splay synchronization (SS) state, the generalized splay synchronization (GSS) state, the anti-phase synchronization (APS) state, the in-phase delay synchronization (IPDS) state, and the in-phase synchronization (IPS) state. In particular, the IPDS state, in which the metronomes on each layer are synchronized in phase but are of a constant phase delay to metronomes on the other layer, is observed for the first time. In addition, a new technique based on audio signals is proposed for pattern detection, which is more convenient and easier to apply than the existing acquisition techniques. Furthermore, a theoretical model is developed to explain the experimental observations, and is employed to explore the dynamical properties of the patterns, including the basin distributions and the pattern transitions. Our study sheds new lights on the collective behaviors of coupled metronomes, and the developed setup can be used in the classroom for demonstration purposes.

  3. Human pupillary dilation response to deviant auditory stimuli: Effects of stimulus properties and voluntary attention

    Directory of Open Access Journals (Sweden)

    Hsin-I eLiao

    2016-02-01

    Full Text Available A unique sound that deviates from a repetitive background sound induces signature neural responses, such as mismatch negativity and novelty P3 response in electro-encephalography studies. Here we show that a deviant auditory stimulus induces a human pupillary dilation response (PDR that is sensitive to the stimulus properties and irrespective whether attention is directed to the sounds or not. In an auditory oddball sequence, we used white noise and 2000-Hz tones as oddballs against repeated 1000-Hz tones. Participants’ pupillary responses were recorded while they listened to the auditory oddball sequence. In Experiment 1, they were not involved in any task. Results show that pupils dilated to the noise oddballs for approximately 4 s, but no such PDR was found for the 2000-Hz tone oddballs. In Experiments 2, two types of visual oddballs were presented synchronously with the auditory oddballs. Participants discriminated the auditory or visual oddballs while trying to ignore stimuli from the other modality. The purpose of this manipulation was to direct attention to or away from the auditory sequence. In Experiment 3, the visual oddballs and the auditory oddballs were always presented asynchronously to prevent residuals of attention on to-be-ignored oddballs due to the concurrence with the attended oddballs. Results show that pupils dilated to both the noise and 2000-Hz tone oddballs in all conditions. Most importantly, PDRs to noise were larger than those to the 2000-Hz tone oddballs regardless of the attention condition in both experiments. The overall results suggest that the stimulus-dependent factor of the PDR appears to be independent of attention.

  4. Human Pupillary Dilation Response to Deviant Auditory Stimuli: Effects of Stimulus Properties and Voluntary Attention.

    Science.gov (United States)

    Liao, Hsin-I; Yoneya, Makoto; Kidani, Shunsuke; Kashino, Makio; Furukawa, Shigeto

    2016-01-01

    A unique sound that deviates from a repetitive background sound induces signature neural responses, such as mismatch negativity and novelty P3 response in electro-encephalography studies. Here we show that a deviant auditory stimulus induces a human pupillary dilation response (PDR) that is sensitive to the stimulus properties and irrespective whether attention is directed to the sounds or not. In an auditory oddball sequence, we used white noise and 2000-Hz tones as oddballs against repeated 1000-Hz tones. Participants' pupillary responses were recorded while they listened to the auditory oddball sequence. In Experiment 1, they were not involved in any task. Results show that pupils dilated to the noise oddballs for approximately 4 s, but no such PDR was found for the 2000-Hz tone oddballs. In Experiments 2, two types of visual oddballs were presented synchronously with the auditory oddballs. Participants discriminated the auditory or visual oddballs while trying to ignore stimuli from the other modality. The purpose of this manipulation was to direct attention to or away from the auditory sequence. In Experiment 3, the visual oddballs and the auditory oddballs were always presented asynchronously to prevent residuals of attention on to-be-ignored oddballs due to the concurrence with the attended oddballs. Results show that pupils dilated to both the noise and 2000-Hz tone oddballs in all conditions. Most importantly, PDRs to noise were larger than those to the 2000-Hz tone oddballs regardless of the attention condition in both experiments. The overall results suggest that the stimulus-dependent factor of the PDR appears to be independent of attention.

  5. Large-Scale Analysis of Auditory Segregation Behavior Crowdsourced via a Smartphone App.

    Science.gov (United States)

    Teki, Sundeep; Kumar, Sukhbinder; Griffiths, Timothy D

    2016-01-01

    The human auditory system is adept at detecting sound sources of interest from a complex mixture of several other simultaneous sounds. The ability to selectively attend to the speech of one speaker whilst ignoring other speakers and background noise is of vital biological significance-the capacity to make sense of complex 'auditory scenes' is significantly impaired in aging populations as well as those with hearing loss. We investigated this problem by designing a synthetic signal, termed the 'stochastic figure-ground' stimulus that captures essential aspects of complex sounds in the natural environment. Previously, we showed that under controlled laboratory conditions, young listeners sampled from the university subject pool (n = 10) performed very well in detecting targets embedded in the stochastic figure-ground signal. Here, we presented a modified version of this cocktail party paradigm as a 'game' featured in a smartphone app (The Great Brain Experiment) and obtained data from a large population with diverse demographical patterns (n = 5148). Despite differences in paradigms and experimental settings, the observed target-detection performance by users of the app was robust and consistent with our previous results from the psychophysical study. Our results highlight the potential use of smartphone apps in capturing robust large-scale auditory behavioral data from normal healthy volunteers, which can also be extended to study auditory deficits in clinical populations with hearing impairments and central auditory disorders.

  6. Large-Scale Analysis of Auditory Segregation Behavior Crowdsourced via a Smartphone App.

    Directory of Open Access Journals (Sweden)

    Sundeep Teki

    Full Text Available The human auditory system is adept at detecting sound sources of interest from a complex mixture of several other simultaneous sounds. The ability to selectively attend to the speech of one speaker whilst ignoring other speakers and background noise is of vital biological significance-the capacity to make sense of complex 'auditory scenes' is significantly impaired in aging populations as well as those with hearing loss. We investigated this problem by designing a synthetic signal, termed the 'stochastic figure-ground' stimulus that captures essential aspects of complex sounds in the natural environment. Previously, we showed that under controlled laboratory conditions, young listeners sampled from the university subject pool (n = 10 performed very well in detecting targets embedded in the stochastic figure-ground signal. Here, we presented a modified version of this cocktail party paradigm as a 'game' featured in a smartphone app (The Great Brain Experiment and obtained data from a large population with diverse demographical patterns (n = 5148. Despite differences in paradigms and experimental settings, the observed target-detection performance by users of the app was robust and consistent with our previous results from the psychophysical study. Our results highlight the potential use of smartphone apps in capturing robust large-scale auditory behavioral data from normal healthy volunteers, which can also be extended to study auditory deficits in clinical populations with hearing impairments and central auditory disorders.

  7. Partial synchronization of different chaotic oscillators using robust PID feedback

    Energy Technology Data Exchange (ETDEWEB)

    Aguilar-Lopez, Ricardo [Departamento de Energia, Universidad Autonoma Metropolitana - Azcapotzalco, San Pablo 180, Reynosa-Tamaulipas, Azcapotzalco, 02200 Mexico, D.F. (Mexico)]. E-mail: raguilar@correo.azc.uam.mx; Martinez-Guerra, Rafael [Departamento de Control Automatico, CINVESTAV IPN, Apartado Postal 14-740, Mexico, D.F. C.P. 07360 (Mexico)]. E-mail: rguerra@ctrl.cinvestav.mx

    2007-07-15

    This work deals with the partial synchronization problem of two different chaotic oscillators considering model uncertainties in the slave system via control approach. The slave system is forced to follow the master signal via a linearizing controller based on model uncertainty reconstructor which leads to proportional-integral-derivative (PID) control structure. This reconstructor is related with a proportional-derivative (PD) reduced-order observer, it would be considered as a sub-slave system for the original slave of the synchronization procedure. The asymptotic performance of the synchronization methodology is proven via the dynamic of the synchronization error. Numerical experiment illustrates the closed-loop behavior of the proposed methodology.

  8. Partial synchronization of different chaotic oscillators using robust PID feedback

    International Nuclear Information System (INIS)

    Aguilar-Lopez, Ricardo; Martinez-Guerra, Rafael

    2007-01-01

    This work deals with the partial synchronization problem of two different chaotic oscillators considering model uncertainties in the slave system via control approach. The slave system is forced to follow the master signal via a linearizing controller based on model uncertainty reconstructor which leads to proportional-integral-derivative (PID) control structure. This reconstructor is related with a proportional-derivative (PD) reduced-order observer, it would be considered as a sub-slave system for the original slave of the synchronization procedure. The asymptotic performance of the synchronization methodology is proven via the dynamic of the synchronization error. Numerical experiment illustrates the closed-loop behavior of the proposed methodology

  9. Anticipating and projective-anticipating synchronization of coupled multidelay feedback systems

    International Nuclear Information System (INIS)

    Hoang, Thang Manh; Nakagawa, Masahiro

    2007-01-01

    In this Letter, the model of coupled multidelay feedback systems is investigated with the schemes of anticipating and projective-anticipating synchronizations. Under these synchronization schemes, the slave anticipates the master's trajectory. Moreover, with projective-anticipating synchronization there exists a scale factor in the amplitudes of the master's and slave's state variables. In the both cases, the driving signals are in the form of multiple nonlinear transformations of delayed state variable. The Krasovskii-Lyapunov theory is used to consider the sufficient condition for synchronization. The effectiveness of the proposed schemes is confirmed by the numerical simulation of specific examples with modified Ikeda and Mackey-Glass systems

  10. From sensation to percept: the neural signature of auditory event-related potentials.

    Science.gov (United States)

    Joos, Kathleen; Gilles, Annick; Van de Heyning, Paul; De Ridder, Dirk; Vanneste, Sven

    2014-05-01

    An external auditory stimulus induces an auditory sensation which may lead to a conscious auditory perception. Although the sensory aspect is well known, it is still a question how an auditory stimulus results in an individual's conscious percept. To unravel the uncertainties concerning the neural correlates of a conscious auditory percept, event-related potentials may serve as a useful tool. In the current review we mainly wanted to shed light on the perceptual aspects of auditory processing and therefore we mainly focused on the auditory late-latency responses. Moreover, there is increasing evidence that perception is an active process in which the brain searches for the information it expects to be present, suggesting that auditory perception requires the presence of both bottom-up, i.e. sensory and top-down, i.e. prediction-driven processing. Therefore, the auditory evoked potentials will be interpreted in the context of the Bayesian brain model, in which the brain predicts which information it expects and when this will happen. The internal representation of the auditory environment will be verified by sensation samples of the environment (P50, N100). When this incoming information violates the expectation, it will induce the emission of a prediction error signal (Mismatch Negativity), activating higher-order neural networks and inducing the update of prior internal representations of the environment (P300). Copyright © 2014 Elsevier Ltd. All rights reserved.

  11. Precise auditory-vocal mirroring in neurons for learned vocal communication.

    Science.gov (United States)

    Prather, J F; Peters, S; Nowicki, S; Mooney, R

    2008-01-17

    Brain mechanisms for communication must establish a correspondence between sensory and motor codes used to represent the signal. One idea is that this correspondence is established at the level of single neurons that are active when the individual performs a particular gesture or observes a similar gesture performed by another individual. Although neurons that display a precise auditory-vocal correspondence could facilitate vocal communication, they have yet to be identified. Here we report that a certain class of neurons in the swamp sparrow forebrain displays a precise auditory-vocal correspondence. We show that these neurons respond in a temporally precise fashion to auditory presentation of certain note sequences in this songbird's repertoire and to similar note sequences in other birds' songs. These neurons display nearly identical patterns of activity when the bird sings the same sequence, and disrupting auditory feedback does not alter this singing-related activity, indicating it is motor in nature. Furthermore, these neurons innervate striatal structures important for song learning, raising the possibility that singing-related activity in these cells is compared to auditory feedback to guide vocal learning.

  12. Effect of conductive hearing loss on central auditory function.

    Science.gov (United States)

    Bayat, Arash; Farhadi, Mohammad; Emamdjomeh, Hesam; Saki, Nader; Mirmomeni, Golshan; Rahim, Fakher

    It has been demonstrated that long-term Conductive Hearing Loss (CHL) may influence the precise detection of the temporal features of acoustic signals or Auditory Temporal Processing (ATP). It can be argued that ATP may be the underlying component of many central auditory processing capabilities such as speech comprehension or sound localization. Little is known about the consequences of CHL on temporal aspects of central auditory processing. This study was designed to assess auditory temporal processing ability in individuals with chronic CHL. During this analytical cross-sectional study, 52 patients with mild to moderate chronic CHL and 52 normal-hearing listeners (control), aged between 18 and 45 year-old, were recruited. In order to evaluate auditory temporal processing, the Gaps-in-Noise (GIN) test was used. The results obtained for each ear were analyzed based on the gap perception threshold and the percentage of correct responses. The average of GIN thresholds was significantly smaller for the control group than for the CHL group for both ears (right: p=0.004; left: phearing for both sides (phearing loss in either group (p>0.05). The results suggest reduced auditory temporal processing ability in adults with CHL compared to normal hearing subjects. Therefore, developing a clinical protocol to evaluate auditory temporal processing in this population is recommended. Copyright © 2017 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.

  13. Cortical layers, rhythms and BOLD signals.

    Science.gov (United States)

    Scheeringa, René; Fries, Pascal

    2017-11-03

    This review investigates how laminar fMRI can complement insights into brain function derived from the study of rhythmic neuronal synchronization. Neuronal synchronization in various frequency bands plays an important role in neuronal communication between brain areas, and it does so on the backbone of layer-specific interareal anatomical projections. Feedforward projections originate predominantly in supragranular cortical layers and terminate in layer 4, and this pattern is reflected in inter-laminar and interareal directed gamma-band influences. Thus, gamma-band synchronization likely subserves feedforward signaling. By contrast, anatomical feedback projections originate predominantly in infragranular layers and terminate outside layer 4, and this pattern is reflected in inter-laminar and interareal directed alpha- and/or beta-band influences. Thus, alpha-beta band synchronization likely subserves feedback signaling. Furthermore, these rhythms explain part of the BOLD signal, with independent contributions of alpha-beta and gamma. These findings suggest that laminar fMRI can provide us with a potentially useful method to test some of the predictions derived from the study of neuronal synchronization. We review central findings regarding the role of layer-specific neuronal synchronization for brain function, and regarding the link between neuronal synchronization and the BOLD signal. We discuss the role that laminar fMRI could play by comparing it to invasive and non-invasive electrophysiological recordings. Compared to direct electrophysiological recordings, this method provides a metric of neuronal activity that is slow and indirect, but that is uniquely non-invasive and layer-specific with potentially whole brain coverage. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.

  14. Robust chaos synchronization based on adaptive fuzzy delayed ...

    Indian Academy of Sciences (India)

    cations in secure communication, economics, signal generator design, chemical reaction, ... that may cause instability and poor performance. ... synchronization error system is asymptotically stable with a guaranteed H∞ norm bound.

  15. The Central Auditory Processing Kit[TM]. Book 1: Auditory Memory [and] Book 2: Auditory Discrimination, Auditory Closure, and Auditory Synthesis [and] Book 3: Auditory Figure-Ground, Auditory Cohesion, Auditory Binaural Integration, and Compensatory Strategies.

    Science.gov (United States)

    Mokhemar, Mary Ann

    This kit for assessing central auditory processing disorders (CAPD), in children in grades 1 through 8 includes 3 books, 14 full-color cards with picture scenes, and a card depicting a phone key pad, all contained in a sturdy carrying case. The units in each of the three books correspond with auditory skill areas most commonly addressed in…

  16. Auditory alert systems with enhanced detectability

    Science.gov (United States)

    Begault, Durand R. (Inventor)

    2008-01-01

    Methods and systems for distinguishing an auditory alert signal from a background of one or more non-alert signals. In a first embodiment, a prefix signal, associated with an existing alert signal, is provided that has a signal component in each of three or more selected frequency ranges, with each signal component in each of three or more selected level at least 3-10 dB above an estimated background (non-alert) level in that frequency range. The alert signal may be chirped within one or more frequency bands. In another embodiment, an alert signal moves, continuously or discontinuously, from one location to another over a short time interval, introducing a perceived spatial modulation or jitter. In another embodiment, a weighted sum of background signals adjacent to each ear is formed, and the weighted sum is delivered to each ear as a uniform background; a distinguishable alert signal is presented on top of this weighted sum signal at one ear, or distinguishable first and second alert signals are presented at two ears of a subject.

  17. Analysis of the time structure of synchronization in multidimensional chaotic systems

    Energy Technology Data Exchange (ETDEWEB)

    Makarenko, A. V., E-mail: avm.science@mail.ru [Constructive Cybernetics Research Group (Russian Federation)

    2015-05-15

    A new approach is proposed to the integrated analysis of the time structure of synchronization of multidimensional chaotic systems. The method allows one to diagnose and quantitatively evaluate the intermittency characteristics during synchronization of chaotic oscillations in the T-synchronization mode. A system of two identical logistic mappings with unidirectional coupling that operate in the developed chaos regime is analyzed. It is shown that the widely used approach, in which only synchronization patterns are subjected to analysis while desynchronization areas are considered as a background signal and removed from analysis, should be regarded as methodologically incomplete.

  18. Analysis of the time structure of synchronization in multidimensional chaotic systems

    International Nuclear Information System (INIS)

    Makarenko, A. V.

    2015-01-01

    A new approach is proposed to the integrated analysis of the time structure of synchronization of multidimensional chaotic systems. The method allows one to diagnose and quantitatively evaluate the intermittency characteristics during synchronization of chaotic oscillations in the T-synchronization mode. A system of two identical logistic mappings with unidirectional coupling that operate in the developed chaos regime is analyzed. It is shown that the widely used approach, in which only synchronization patterns are subjected to analysis while desynchronization areas are considered as a background signal and removed from analysis, should be regarded as methodologically incomplete

  19. Auditory Perceptual Abilities Are Associated with Specific Auditory Experience

    Directory of Open Access Journals (Sweden)

    Yael Zaltz

    2017-11-01

    Full Text Available The extent to which auditory experience can shape general auditory perceptual abilities is still under constant debate. Some studies show that specific auditory expertise may have a general effect on auditory perceptual abilities, while others show a more limited influence, exhibited only in a relatively narrow range associated with the area of expertise. The current study addresses this issue by examining experience-dependent enhancement in perceptual abilities in the auditory domain. Three experiments were performed. In the first experiment, 12 pop and rock musicians and 15 non-musicians were tested in frequency discrimination (DLF, intensity discrimination, spectrum discrimination (DLS, and time discrimination (DLT. Results showed significant superiority of the musician group only for the DLF and DLT tasks, illuminating enhanced perceptual skills in the key features of pop music, in which miniscule changes in amplitude and spectrum are not critical to performance. The next two experiments attempted to differentiate between generalization and specificity in the influence of auditory experience, by comparing subgroups of specialists. First, seven guitar players and eight percussionists were tested in the DLF and DLT tasks that were found superior for musicians. Results showed superior abilities on the DLF task for guitar players, though no difference between the groups in DLT, demonstrating some dependency of auditory learning on the specific area of expertise. Subsequently, a third experiment was conducted, testing a possible influence of vowel density in native language on auditory perceptual abilities. Ten native speakers of German (a language characterized by a dense vowel system of 14 vowels, and 10 native speakers of Hebrew (characterized by a sparse vowel system of five vowels, were tested in a formant discrimination task. This is the linguistic equivalent of a DLS task. Results showed that German speakers had superior formant

  20. Optical Synchronization Systems for Femtosecond X-raySources

    Energy Technology Data Exchange (ETDEWEB)

    Wilcox, Russell; Staples, John W.; Holzwarth, Ronald

    2004-05-09

    In femtosecond pump/probe experiments using short X-Ray and optical pulses, precise synchronization must be maintained between widely separated lasers in a synchrotron or FEL facility. We are developing synchronization systems using optical signals for applications requiring different ranges of timing error over 100 meter of glass fiber. For stabilization in the hundred femtosecond range a CW laser is amplitude modulated at 1 10 GHz, the signal retroreflected from the far end, and the relative phase used to correct the transit time with a piezoelectric phase modulator. For the sub-10 fsec range the laser frequency itself is upshifted 55 MHz with an acousto-optical modulator, retroreflected, upshifted again and phase compared at the sending end to a 110 MHz reference. Initial experiments indicate less than 1 fsec timing jitter. To lock lasers in the sub-10 fs range we will lock two single-frequency lasers separated by several tera Hertz to a master modelocked fiber laser, transmit the two frequencies over fiber, and lock two comb lines of a slave laser to these frequencies, thus synchronizing the two modelocked laser envelopes.

  1. Optical Synchronization Systems for Femtosecond X-ray Sources

    International Nuclear Information System (INIS)

    Wilcox, Russell; Staples, John W.; Holzwarth, Ronald

    2004-01-01

    In femtosecond pump/probe experiments using short X-Ray and optical pulses, precise synchronization must be maintained between widely separated lasers in a synchrotron or FEL facility. We are developing synchronization systems using optical signals for applications requiring different ranges of timing error over 100 meter of glass fiber. For stabilization in the hundred femtosecond range a CW laser is amplitude modulated at 1 10 GHz, the signal retroreflected from the far end, and the relative phase used to correct the transit time with a piezoelectric phase modulator. For the sub-10 fsec range the laser frequency itself is upshifted 55 MHz with an acousto-optical modulator, retroreflected, upshifted again and phase compared at the sending end to a 110 MHz reference. Initial experiments indicate less than 1 fsec timing jitter. To lock lasers in the sub-10 fs range we will lock two single-frequency lasers separated by several tera Hertz to a master modelocked fiber laser, transmit the two frequencies over fiber, and lock two comb lines of a slave laser to these frequencies, thus synchronizing the two modelocked laser envelopes

  2. Auditory midbrain processing is differentially modulated by auditory and visual cortices: An auditory fMRI study.

    Science.gov (United States)

    Gao, Patrick P; Zhang, Jevin W; Fan, Shu-Juan; Sanes, Dan H; Wu, Ed X

    2015-12-01

    The cortex contains extensive descending projections, yet the impact of cortical input on brainstem processing remains poorly understood. In the central auditory system, the auditory cortex contains direct and indirect pathways (via brainstem cholinergic cells) to nuclei of the auditory midbrain, called the inferior colliculus (IC). While these projections modulate auditory processing throughout the IC, single neuron recordings have samples from only a small fraction of cells during stimulation of the corticofugal pathway. Furthermore, assessments of cortical feedback have not been extended to sensory modalities other than audition. To address these issues, we devised blood-oxygen-level-dependent (BOLD) functional magnetic resonance imaging (fMRI) paradigms to measure the sound-evoked responses throughout the rat IC and investigated the effects of bilateral ablation of either auditory or visual cortices. Auditory cortex ablation increased the gain of IC responses to noise stimuli (primarily in the central nucleus of the IC) and decreased response selectivity to forward species-specific vocalizations (versus temporally reversed ones, most prominently in the external cortex of the IC). In contrast, visual cortex ablation decreased the gain and induced a much smaller effect on response selectivity. The results suggest that auditory cortical projections normally exert a large-scale and net suppressive influence on specific IC subnuclei, while visual cortical projections provide a facilitatory influence. Meanwhile, auditory cortical projections enhance the midbrain response selectivity to species-specific vocalizations. We also probed the role of the indirect cholinergic projections in the auditory system in the descending modulation process by pharmacologically blocking muscarinic cholinergic receptors. This manipulation did not affect the gain of IC responses but significantly reduced the response selectivity to vocalizations. The results imply that auditory cortical

  3. Psychophysical and Neural Correlates of Auditory Attraction and Aversion

    Science.gov (United States)

    Patten, Kristopher Jakob

    This study explores the psychophysical and neural processes associated with the perception of sounds as either pleasant or aversive. The underlying psychophysical theory is based on auditory scene analysis, the process through which listeners parse auditory signals into individual acoustic sources. The first experiment tests and confirms that a self-rated pleasantness continuum reliably exists for 20 various stimuli (r = .48). In addition, the pleasantness continuum correlated with the physical acoustic characteristics of consonance/dissonance (r = .78), which can facilitate auditory parsing processes. The second experiment uses an fMRI block design to test blood oxygen level dependent (BOLD) changes elicited by a subset of 5 exemplar stimuli chosen from Experiment 1 that are evenly distributed over the pleasantness continuum. Specifically, it tests and confirms that the pleasantness continuum produces systematic changes in brain activity for unpleasant acoustic stimuli beyond what occurs with pleasant auditory stimuli. Results revealed that the combination of two positively and two negatively valenced experimental sounds compared to one neutral baseline control elicited BOLD increases in the primary auditory cortex, specifically the bilateral superior temporal gyrus, and left dorsomedial prefrontal cortex; the latter being consistent with a frontal decision-making process common in identification tasks. The negatively-valenced stimuli yielded additional BOLD increases in the left insula, which typically indicates processing of visceral emotions. The positively-valenced stimuli did not yield any significant BOLD activation, consistent with consonant, harmonic stimuli being the prototypical acoustic pattern of auditory objects that is optimal for auditory scene analysis. Both the psychophysical findings of Experiment 1 and the neural processing findings of Experiment 2 support that consonance is an important dimension of sound that is processed in a manner that aids

  4. Synchronization of delayed systems in the presence of delay time modulation

    International Nuclear Information System (INIS)

    Kye, Won-Ho; Choi, Muhan; Kim, Myung-Woon; Lee, Soo-Young; Rim, Sunghwan; Kim, Chil-Min; Park, Young-Jai

    2004-01-01

    We investigate synchronization in the presence of delay time modulation for application to communication. We have observed that the robust synchronization is established by a common delay signal and its threshold is presented using Lyapunov exponents analysis. The influence of the delay time modulation in chaotic oscillators is also discussed

  5. Primate auditory recognition memory performance varies with sound type.

    Science.gov (United States)

    Ng, Chi-Wing; Plakke, Bethany; Poremba, Amy

    2009-10-01

    Neural correlates of auditory processing, including for species-specific vocalizations that convey biological and ethological significance (e.g., social status, kinship, environment), have been identified in a wide variety of areas including the temporal and frontal cortices. However, few studies elucidate how non-human primates interact with these vocalization signals when they are challenged by tasks requiring auditory discrimination, recognition and/or memory. The present study employs a delayed matching-to-sample task with auditory stimuli to examine auditory memory performance of rhesus macaques (Macaca mulatta), wherein two sounds are determined to be the same or different. Rhesus macaques seem to have relatively poor short-term memory with auditory stimuli, and we examine if particular sound types are more favorable for memory performance. Experiment 1 suggests memory performance with vocalization sound types (particularly monkey), are significantly better than when using non-vocalization sound types, and male monkeys outperform female monkeys overall. Experiment 2, controlling for number of sound exemplars and presentation pairings across types, replicates Experiment 1, demonstrating better performance or decreased response latencies, depending on trial type, to species-specific monkey vocalizations. The findings cannot be explained by acoustic differences between monkey vocalizations and the other sound types, suggesting the biological, and/or ethological meaning of these sounds are more effective for auditory memory. 2009 Elsevier B.V.

  6. How far away is plug 'n' play? Assessing the near-term potential of sonification and auditory display

    Science.gov (United States)

    Bargar, Robin

    1995-01-01

    The commercial music industry offers a broad range of plug 'n' play hardware and software scaled to music professionals and scaled to a broad consumer market. The principles of sound synthesis utilized in these products are relevant to application in virtual environments (VE). However, the closed architectures used in commercial music synthesizers are prohibitive to low-level control during real-time rendering, and the algorithms and sounds themselves are not standardized from product to product. To bring sound into VE requires a new generation of open architectures designed for human-controlled performance from interfaces embedded in immersive environments. This presentation addresses the state of the sonic arts in scientific computing and VE, analyzes research challenges facing sound computation, and offers suggestions regarding tools we might expect to become available during the next few years. A list of classes of audio functionality in VE includes sonification -- the use of sound to represent data from numerical models; 3D auditory display (spatialization and localization, also called externalization); navigation cues for positional orientation and for finding items or regions inside large spaces; voice recognition for controlling the computer; external communications between users in different spaces; and feedback to the user concerning his own actions or the state of the application interface. To effectively convey this considerable variety of signals, we apply principles of acoustic design to ensure the messages are neither confusing nor competing. We approach the design of auditory experience through a comprehensive structure for messages, and message interplay we refer to as an Automated Sound Environment. Our research addresses real-time sound synthesis, real-time signal processing and localization, interactive control of high-dimensional systems, and synchronization of sound and graphics.

  7. Bottom-up driven involuntary auditory evoked field change: constant sound sequencing amplifies but does not sharpen neural activity.

    Science.gov (United States)

    Okamoto, Hidehiko; Stracke, Henning; Lagemann, Lothar; Pantev, Christo

    2010-01-01

    The capability of involuntarily tracking certain sound signals during the simultaneous presence of noise is essential in human daily life. Previous studies have demonstrated that top-down auditory focused attention can enhance excitatory and inhibitory neural activity, resulting in sharpening of frequency tuning of auditory neurons. In the present study, we investigated bottom-up driven involuntary neural processing of sound signals in noisy environments by means of magnetoencephalography. We contrasted two sound signal sequencing conditions: "constant sequencing" versus "random sequencing." Based on a pool of 16 different frequencies, either identical (constant sequencing) or pseudorandomly chosen (random sequencing) test frequencies were presented blockwise together with band-eliminated noises to nonattending subjects. The results demonstrated that the auditory evoked fields elicited in the constant sequencing condition were significantly enhanced compared with the random sequencing condition. However, the enhancement was not significantly different between different band-eliminated noise conditions. Thus the present study confirms that by constant sound signal sequencing under nonattentive listening the neural activity in human auditory cortex can be enhanced, but not sharpened. Our results indicate that bottom-up driven involuntary neural processing may mainly amplify excitatory neural networks, but may not effectively enhance inhibitory neural circuits.

  8. Psychoacoustic and cognitive aspects of auditory roughness: definitions, models, and applications

    Science.gov (United States)

    Vassilakis, Pantelis N.; Kendall, Roger A.

    2010-02-01

    The term "auditory roughness" was first introduced in the 19th century to describe the buzzing, rattling auditory sensation accompanying narrow harmonic intervals (i.e. two tones with frequency difference in the range of ~15-150Hz, presented simultaneously). A broader definition and an overview of the psychoacoustic correlates of the auditory roughness sensation, also referred to as sensory dissonance, is followed by an examination of efforts to quantify it over the past one hundred and fifty years and leads to the introduction of a new roughness calculation model and an application that automates spectral and roughness analysis of sound signals. Implementation of spectral and roughness analysis is briefly discussed in the context of two pilot perceptual experiments, designed to assess the relationship among cultural background, music performance practice, and aesthetic attitudes towards the auditory roughness sensation.

  9. Digital-data receiver synchronization method and apparatus

    Science.gov (United States)

    Smith, Stephen F [Loudon, TN; Turner, Gary W [Clinton, TN

    2009-09-08

    Digital data receiver synchronization is provided with composite phase-frequency detectors, mutually cross-connected comparison feedback or both to provide robust reception of digital data signals. A single master clock can be used to provide frequency signals. Advantages can include fast lock-up time in moderately to severely noisy conditions, greater tolerance to noise and jitter when locked, and improved tolerance to clock asymmetries.

  10. Autofluorescence of pigmented skin lesions using a pulsed UV laser with synchronized detection: clinical results

    DEFF Research Database (Denmark)

    Cheng, Haynes Pak Hay; Svenmarker, Pontus; Tidemand-Lichtenberg, Peter

    2010-01-01

    signal, which may in turn produce high contrast images that improve diagnosis, even in the presence of ambient room light. The synchronized set-up utilizes a compact, diode pumped, pulsed UV laser at 355 nm which is coupled to a CCD camera and a liquid crystal tunable filter. The excitation and image......We report preliminary clinical results of autofluorescence imaging of malignant and benign skin lesions, using pulsed 355 nm laser excitation with synchronized detection. The novel synchronized detection system allows high signal-to-noise ratio to be achieved in the resulting autofluorescence...

  11. Auditory white noise reduces postural fluctuations even in the absence of vision.

    Science.gov (United States)

    Ross, Jessica Marie; Balasubramaniam, Ramesh

    2015-08-01

    The contributions of somatosensory, vestibular, and visual feedback to balance control are well documented, but the influence of auditory information, especially acoustic noise, on balance is less clear. Because somatosensory noise has been shown to reduce postural sway, we hypothesized that noise from the auditory modality might have a similar effect. Given that the nervous system uses noise to optimize signal transfer, adding mechanical or auditory noise should lead to increased feedback about sensory frames of reference used in balance control. In the present experiment, postural sway was analyzed in healthy young adults where they were presented with continuous white noise, in the presence and absence of visual information. Our results show reduced postural sway variability (as indexed by the body's center of pressure) in the presence of auditory noise, even when visual information was not present. Nonlinear time series analysis revealed that auditory noise has an additive effect, independent of vision, on postural stability. Further analysis revealed that auditory noise reduced postural sway variability in both low- and high-frequency regimes (> or noise. Our results support the idea that auditory white noise reduces postural sway, suggesting that auditory noise might be used for therapeutic and rehabilitation purposes in older individuals and those with balance disorders.

  12. Visual Input Enhances Selective Speech Envelope Tracking in Auditory Cortex at a ‘Cocktail Party’

    Science.gov (United States)

    Golumbic, Elana Zion; Cogan, Gregory B.; Schroeder, Charles E.; Poeppel, David

    2013-01-01

    Our ability to selectively attend to one auditory signal amidst competing input streams, epitomized by the ‘Cocktail Party’ problem, continues to stimulate research from various approaches. How this demanding perceptual feat is achieved from a neural systems perspective remains unclear and controversial. It is well established that neural responses to attended stimuli are enhanced compared to responses to ignored ones, but responses to ignored stimuli are nonetheless highly significant, leading to interference in performance. We investigated whether congruent visual input of an attended speaker enhances cortical selectivity in auditory cortex, leading to diminished representation of ignored stimuli. We recorded magnetoencephalographic (MEG) signals from human participants as they attended to segments of natural continuous speech. Using two complementary methods of quantifying the neural response to speech, we found that viewing a speaker’s face enhances the capacity of auditory cortex to track the temporal speech envelope of that speaker. This mechanism was most effective in a ‘Cocktail Party’ setting, promoting preferential tracking of the attended speaker, whereas without visual input no significant attentional modulation was observed. These neurophysiological results underscore the importance of visual input in resolving perceptual ambiguity in a noisy environment. Since visual cues in speech precede the associated auditory signals, they likely serve a predictive role in facilitating auditory processing of speech, perhaps by directing attentional resources to appropriate points in time when to-be-attended acoustic input is expected to arrive. PMID:23345218

  13. Synchronization and NRZ-to-RZ format conversion of 10 G Ethernet packet based on a time lens

    DEFF Research Database (Denmark)

    Hu, Hao; Laguardia Areal, Janaina; Palushani, Evarist

    2010-01-01

    10 G Ethernet packet with maximum frame size of 1518 bytes is synchronized to a global clock using a time lens. The 10 Gb/s NRZ signal is converted into RZ signal at the same time.......10 G Ethernet packet with maximum frame size of 1518 bytes is synchronized to a global clock using a time lens. The 10 Gb/s NRZ signal is converted into RZ signal at the same time....

  14. Representation of acoustic signals in the eighth nerve of the Tokay gecko. II. Masking of pure tones with noise.

    Science.gov (United States)

    Sams-Dodd, F; Capranica, R R

    1996-10-01

    Acoustic signals are generally encoded in the peripheral auditory system of vertebrates by a duality scheme. For frequency components that fall within the excitatory tuning curve, individual eighth nerve fibers can encode the effective spectral energy by a spike-rate code, while simultaneously preserving the signal waveform periodicity of lower frequency components by phase-locked spike-train discharges. To explore how robust this duality of representation may be in the presence of noise, we recorded the responses of auditory fibers in the eighth nerve of the Tokay gecko to tonal stimuli when masking noise was added simultaneously. We found that their spike-rate functions reached plateau levels fairly rapidly in the presence of noise, so the ability to signal the presence of a tone by a concomitant change in firing rate was quickly lost. On the other hand, their synchronization functions maintained a high degree of phase-locked firings to the tone even in the presence of high-intensity masking noise, thus enabling a robust detection of the tonal signal. Critical ratios (CR) and critical bandwidths showed that in the frequency range where units are able to phaselock to the tonal periodicity, the CR bands were relatively narrow and the bandwidths were independent of noise level. However, to higher frequency tones where phaselocking fails and only spike-rate codes apply, the CR bands were much wider and depended upon noise level, so that their ability to filter tones out of a noisy background degraded with increasing noise levels. The greater robustness of phase-locked temporal encoding contrasted with spike-rate coding verifies a important advantage in using lower frequency signals for communication in noisy environments.

  15. Modelling of intercellular synchronization in the Drosophila circadian clock

    International Nuclear Information System (INIS)

    Jun-Wei, Wang; Ai-Min, Chen; Jia-Jun, Zhang; Zhan-Jiang, Yuan; Tian-Shou, Zhou

    2009-01-01

    In circadian rhythm generation, intercellular signaling factors are shown to play a crucial role in both sustaining intrinsic cellular rhythmicity and acquiring collective behaviours across a population of circadian neurons. However, the physical mechanism behind their role remains to be fully understood. In this paper, we propose an indirectly coupled multicellular model for the synchronization of Drosophila circadian oscillators combining both intracellular and intercellular dynamics. By simulating different experimental conditions, we find that such an indirect coupling way can synchronize both heterogeneous self-sustained circadian neurons and heterogeneous mutational damped circadian neurons. Moreover, they can also be entrained to ambient light-dark (LD) cycles depending on intercellular signaling. (cross-disciplinary physics and related areas of science and technology)

  16. A Synchronization Method for Single-Phase Grid-Tied Inverters

    DEFF Research Database (Denmark)

    Hadjidemetriou, Lenos; Kyriakides, Elias; Yang, Yongheng

    2016-01-01

    The controllers of single-phase grid-tied inverters require improvements to enable distribution generation systems to meet the grid codes/standards with respect to power quality and the fault ride through capability. In that case, the response of the selected synchronization technique is crucial...... for the performance of the entire grid-tied inverter. In this paper, a new synchronization method with good dynamics and high accuracy under a highly distorted voltage is proposed. This method uses a Multi-Harmonic Decoupling Cell (MHDC), which thus can cancel out the oscillations on the synchronization signals due...... to the harmonic voltage distortion while maintaining the dynamic response of the synchronization. Therefore, the accurate and dynamic response of the proposed MHDC-PLL can be beneficial for the performance of the whole single-phase grid-tied inverter....

  17. Cardiorespiratory phase synchronization during normal rest and inward-attention meditation.

    Science.gov (United States)

    Wu, Shr-Da; Lo, Pei-Chen

    2010-06-11

    The cardiac and respiratory systems can be viewed as two self-sustained oscillators with various interactions between them. In this study, the cardiorespiratory phase synchronization (CRPS) quantified by synchrogram was investigated to explore the phase synchronization between these two systems. The synchrogram scheme was applied to electrocardiogram (ECG) and respiration signals. Particular focus was the distinct cardiac-respiratory regulation phenomena intervened by inward-attention meditation and normal relaxation. Four synchronization parameters were measured: frequency ratio, lasting length, number of epochs, and total length. The results showed that normal rest resulted in much weaker CRPS. Statistical analysis reveals that the number of synchronous epochs and the total synchronization length significantly increase (p=0.024 and 0.034 respectively) during meditation. Furthermore, a predominance of 4:1 and 5:1 rhythm-ratio synchronizations was observed during meditation. Consequently, this study concludes that CRPS can be enhanced during meditation, compared with normal relaxation, and reveals a predominance of specific frequency ratios. Copyright (c) 2008 Elsevier Ireland Ltd. All rights reserved.

  18. Auditory orientation in crickets: Pattern recognition controls reactive steering

    Science.gov (United States)

    Poulet, James F. A.; Hedwig, Berthold

    2005-10-01

    Many groups of insects are specialists in exploiting sensory cues to locate food resources or conspecifics. To achieve orientation, bees and ants analyze the polarization pattern of the sky, male moths orient along the females' odor plume, and cicadas, grasshoppers, and crickets use acoustic signals to locate singing conspecifics. In comparison with olfactory and visual orientation, where learning is involved, auditory processing underlying orientation in insects appears to be more hardwired and genetically determined. In each of these examples, however, orientation requires a recognition process identifying the crucial sensory pattern to interact with a localization process directing the animal's locomotor activity. Here, we characterize this interaction. Using a sensitive trackball system, we show that, during cricket auditory behavior, the recognition process that is tuned toward the species-specific song pattern controls the amplitude of auditory evoked steering responses. Females perform small reactive steering movements toward any sound patterns. Hearing the male's calling song increases the gain of auditory steering within 2-5 s, and the animals even steer toward nonattractive sound patterns inserted into the speciesspecific pattern. This gain control mechanism in the auditory-to-motor pathway allows crickets to pursue species-specific sound patterns temporarily corrupted by environmental factors and may reflect the organization of recognition and localization networks in insects. localization | phonotaxis

  19. Manipulation of BDNF signaling modifies the experience-dependent plasticity induced by pure tone exposure during the critical period in the primary auditory cortex.

    Science.gov (United States)

    Anomal, Renata; de Villers-Sidani, Etienne; Merzenich, Michael M; Panizzutti, Rogerio

    2013-01-01

    Sensory experience powerfully shapes cortical sensory representations during an early developmental "critical period" of plasticity. In the rat primary auditory cortex (A1), the experience-dependent plasticity is exemplified by significant, long-lasting distortions in frequency representation after mere exposure to repetitive frequencies during the second week of life. In the visual system, the normal unfolding of critical period plasticity is strongly dependent on the elaboration of brain-derived neurotrophic factor (BDNF), which promotes the establishment of inhibition. Here, we tested the hypothesis that BDNF signaling plays a role in the experience-dependent plasticity induced by pure tone exposure during the critical period in the primary auditory cortex. Elvax resin implants filled with either a blocking antibody against BDNF or the BDNF protein were placed on the A1 of rat pups throughout the critical period window. These pups were then exposed to 7 kHz pure tone for 7 consecutive days and their frequency representations were mapped. BDNF blockade completely prevented the shaping of cortical tuning by experience and resulted in poor overall frequency tuning in A1. By contrast, BDNF infusion on the developing A1 amplified the effect of 7 kHz tone exposure compared to control. These results indicate that BDNF signaling participates in the experience-dependent plasticity induced by pure tone exposure during the critical period in A1.

  20. Manipulation of BDNF signaling modifies the experience-dependent plasticity induced by pure tone exposure during the critical period in the primary auditory cortex.

    Directory of Open Access Journals (Sweden)

    Renata Anomal

    Full Text Available Sensory experience powerfully shapes cortical sensory representations during an early developmental "critical period" of plasticity. In the rat primary auditory cortex (A1, the experience-dependent plasticity is exemplified by significant, long-lasting distortions in frequency representation after mere exposure to repetitive frequencies during the second week of life. In the visual system, the normal unfolding of critical period plasticity is strongly dependent on the elaboration of brain-derived neurotrophic factor (BDNF, which promotes the establishment of inhibition. Here, we tested the hypothesis that BDNF signaling plays a role in the experience-dependent plasticity induced by pure tone exposure during the critical period in the primary auditory cortex. Elvax resin implants filled with either a blocking antibody against BDNF or the BDNF protein were placed on the A1 of rat pups throughout the critical period window. These pups were then exposed to 7 kHz pure tone for 7 consecutive days and their frequency representations were mapped. BDNF blockade completely prevented the shaping of cortical tuning by experience and resulted in poor overall frequency tuning in A1. By contrast, BDNF infusion on the developing A1 amplified the effect of 7 kHz tone exposure compared to control. These results indicate that BDNF signaling participates in the experience-dependent plasticity induced by pure tone exposure during the critical period in A1.

  1. Phase synchronization of neuronal noise in mouse hippocampal epileptiform dynamics.

    Science.gov (United States)

    Serletis, Demitre; Carlen, Peter L; Valiante, Taufik A; Bardakjian, Berj L

    2013-02-01

    Organized brain activity is the result of dynamical, segregated neuronal signals that may be used to investigate synchronization effects using sophisticated neuroengineering techniques. Phase synchrony analysis, in particular, has emerged as a promising methodology to study transient and frequency-specific coupling effects across multi-site signals. In this study, we investigated phase synchronization in intracellular recordings of interictal and ictal epileptiform events recorded from pairs of cells in the whole (intact) mouse hippocampus. In particular, we focused our analysis on the background noise-like activity (NLA), previously reported to exhibit complex neurodynamical properties. Our results show evidence for increased linear and nonlinear phase coupling in NLA across three frequency bands [theta (4-10 Hz), beta (12-30 Hz) and gamma (30-80 Hz)] in the ictal compared to interictal state dynamics. We also present qualitative and statistical evidence for increased phase synchronization in the theta, beta and gamma frequency bands from paired recordings of ictal NLA. Overall, our results validate the use of background NLA in the neurodynamical study of epileptiform transitions and suggest that what is considered "neuronal noise" is amenable to synchronization effects in the spatiotemporal domain.

  2. Synchronization and bidirectional communication without delay line using strong mutually coupled semiconductor lasers

    Science.gov (United States)

    Li, Guang-Hui; Wang, An-Bang; Feng, Ye; Wang, Yang

    2010-07-01

    This paper numerically demonstrates synchronization and bidirectional communication without delay line by using two semiconductor lasers with strong mutual injection in a face-to-face configuration. These results show that both of the two lasers' outputs synchronize with their input chaotic carriers. In addition, simulations demonstrate that this kind of synchronization can be used to realize bidirectional communications without delay line. Further studies indicate that within a small deviation in message amplitudes of two sides (±6%), the message can be extracted with signal-noise-ratio more than 10 dB; and the signal-noise-ratio is extremely sensitive to the message rates mismatch of two sides, which may be used as a key of bidirectional communication.

  3. Synchronization and bidirectional communication without delay line using strong mutually coupled semiconductor lasers

    International Nuclear Information System (INIS)

    Guang-Hui, Li; An-Bang, Wang; Ye, Feng; Yang, Wang

    2010-01-01

    This paper numerically demonstrates synchronization and bidirectional communication without delay line by using two semiconductor lasers with strong mutual injection in a face-to-face configuration. These results show that both of the two lasers' outputs synchronize with their input chaotic carriers. In addition, simulations demonstrate that this kind of synchronization can be used to realize bidirectional communications without delay line. Further studies indicate that within a small deviation in message amplitudes of two sides (±6%), the message can be extracted with signal-noise-ratio more than 10 dB; and the signal-noise-ratio is extremely sensitive to the message rates mismatch of two sides, which may be used as a key of bidirectional communication. (general)

  4. Time-sharing visual and auditory tracking tasks

    Science.gov (United States)

    Tsang, Pamela S.; Vidulich, Michael A.

    1987-01-01

    An experiment is described which examined the benefits of distributing the input demands of two tracking tasks as a function of task integrality. Visual and auditory compensatory tracking tasks were utilized. Results indicate that presenting the two tracking signals in two input modalities did not improve time-sharing efficiency. This was attributed to the difficulty insensitivity phenomenon.

  5. Synchronization and anti-synchronization coexist in Chen-Lee chaotic systems

    International Nuclear Information System (INIS)

    Chen, J.-H.; Chen, H.-K.; Lin, Y.-K.

    2009-01-01

    This study demonstrates that synchronization and anti-synchronization can coexist in Chen-Lee chaotic systems by direct linear coupling. Based on Lyapunov's direct method, a linear controller was designed to assure that two different types of synchronization can simultaneously be achieved. Further, the hybrid projective synchronization of Chen-Lee chaotic systems was studied using a nonlinear control scheme. The nonlinear controller was designed according to the Lyapunov stability theory to guarantee the hybrid projective synchronization, including synchronization, anti-synchronization, and projective synchronization. Finally, numerical examples are presented in order to illustrate the proposed synchronization approach.

  6. Optimal cellular mobility for synchronization arising from the gradual recovery of intercellular interactions

    International Nuclear Information System (INIS)

    Uriu, Koichiro; Ares, Saúl; Oates, Andrew C; Morelli, Luis G

    2012-01-01

    Cell movement and intercellular signaling occur simultaneously during the development of tissues, but little is known about how movement affects signaling. Previous theoretical studies have shown that faster moving cells favor synchronization across a population of locally coupled genetic oscillators. An important assumption in these studies is that cells can immediately interact with their new neighbors after arriving at a new location. However, intercellular interactions in cellular systems may need some time to become fully established. How movement affects synchronization in this situation has not been examined. Here, we develop a coupled phase oscillator model in which we consider cell movement and the gradual recovery of intercellular coupling experienced by a cell after movement, characterized by a moving rate and a coupling recovery rate, respectively. We find (1) an optimal moving rate for synchronization and (2) a critical moving rate above which achieving synchronization is not possible. These results indicate that the extent to which movement enhances synchrony is limited by a gradual recovery of coupling. These findings suggest that the ratio of time scales of movement and signaling recovery is critical for information transfer between moving cells. (paper)

  7. Auditory agnosia as a clinical symptom of childhood adrenoleukodystrophy.

    Science.gov (United States)

    Furushima, Wakana; Kaga, Makiko; Nakamura, Masako; Gunji, Atsuko; Inagaki, Masumi

    2015-08-01

    To investigate detailed auditory features in patients with auditory impairment as the first clinical symptoms of childhood adrenoleukodystrophy (CSALD). Three patients who had hearing difficulty as the first clinical signs and/or symptoms of ALD. Precise examination of the clinical characteristics of hearing and auditory function was performed, including assessments of pure tone audiometry, verbal sound discrimination, otoacoustic emission (OAE), and auditory brainstem response (ABR), as well as an environmental sound discrimination test, a sound lateralization test, and a dichotic listening test (DLT). The auditory pathway was evaluated by MRI in each patient. Poor response to calling was detected in all patients. Two patients were not aware of their hearing difficulty, and had been diagnosed with normal hearing by otolaryngologists at first. Pure-tone audiometry disclosed normal hearing in all patients. All patients showed a normal wave V ABR threshold. Three patients showed obvious difficulty in discriminating verbal sounds, environmental sounds, and sound lateralization and strong left-ear suppression in a dichotic listening test. However, once they discriminated verbal sounds, they correctly understood the meaning. Two patients showed elongation of the I-V and III-V interwave intervals in ABR, but one showed no abnormality. MRIs of these three patients revealed signal changes in auditory radiation including in other subcortical areas. The hearing features of these subjects were diagnosed as auditory agnosia and not aphasia. It should be emphasized that when patients are suspected to have hearing impairment but have no abnormalities in pure tone audiometry and/or ABR, this should not be diagnosed immediately as psychogenic response or pathomimesis, but auditory agnosia must also be considered. Copyright © 2014 The Japanese Society of Child Neurology. Published by Elsevier B.V. All rights reserved.

  8. Method and apparatus to debug an integrated circuit chip via synchronous clock stop and scan

    Science.gov (United States)

    Bellofatto, Ralph E [Ridgefield, CT; Ellavsky, Matthew R [Rochester, MN; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Gooding, Thomas M [Rochester, MN; Haring, Rudolf A [Cortlandt Manor, NY; Hehenberger, Lance G [Leander, TX; Ohmacht, Martin [Yorktown Heights, NY

    2012-03-20

    An apparatus and method for evaluating a state of an electronic or integrated circuit (IC), each IC including one or more processor elements for controlling operations of IC sub-units, and each the IC supporting multiple frequency clock domains. The method comprises: generating a synchronized set of enable signals in correspondence with one or more IC sub-units for starting operation of one or more IC sub-units according to a determined timing configuration; counting, in response to one signal of the synchronized set of enable signals, a number of main processor IC clock cycles; and, upon attaining a desired clock cycle number, generating a stop signal for each unique frequency clock domain to synchronously stop a functional clock for each respective frequency clock domain; and, upon synchronously stopping all on-chip functional clocks on all frequency clock domains in a deterministic fashion, scanning out data values at a desired IC chip state. The apparatus and methodology enables construction of a cycle-by-cycle view of any part of the state of a running IC chip, using a combination of on-chip circuitry and software.

  9. Auditory Association Cortex Lesions Impair Auditory Short-Term Memory in Monkeys

    Science.gov (United States)

    Colombo, Michael; D'Amato, Michael R.; Rodman, Hillary R.; Gross, Charles G.

    1990-01-01

    Monkeys that were trained to perform auditory and visual short-term memory tasks (delayed matching-to-sample) received lesions of the auditory association cortex in the superior temporal gyrus. Although visual memory was completely unaffected by the lesions, auditory memory was severely impaired. Despite this impairment, all monkeys could discriminate sounds closer in frequency than those used in the auditory memory task. This result suggests that the superior temporal cortex plays a role in auditory processing and retention similar to the role the inferior temporal cortex plays in visual processing and retention.

  10. Auditory hallucinations.

    Science.gov (United States)

    Blom, Jan Dirk

    2015-01-01

    Auditory hallucinations constitute a phenomenologically rich group of endogenously mediated percepts which are associated with psychiatric, neurologic, otologic, and other medical conditions, but which are also experienced by 10-15% of all healthy individuals in the general population. The group of phenomena is probably best known for its verbal auditory subtype, but it also includes musical hallucinations, echo of reading, exploding-head syndrome, and many other types. The subgroup of verbal auditory hallucinations has been studied extensively with the aid of neuroimaging techniques, and from those studies emerges an outline of a functional as well as a structural network of widely distributed brain areas involved in their mediation. The present chapter provides an overview of the various types of auditory hallucination described in the literature, summarizes our current knowledge of the auditory networks involved in their mediation, and draws on ideas from the philosophy of science and network science to reconceptualize the auditory hallucinatory experience, and point out directions for future research into its neurobiologic substrates. In addition, it provides an overview of known associations with various clinical conditions and of the existing evidence for pharmacologic and non-pharmacologic treatments. © 2015 Elsevier B.V. All rights reserved.

  11. Frequency-specific modulation of population-level frequency tuning in human auditory cortex

    Directory of Open Access Journals (Sweden)

    Roberts Larry E

    2009-01-01

    Full Text Available Abstract Background Under natural circumstances, attention plays an important role in extracting relevant auditory signals from simultaneously present, irrelevant noises. Excitatory and inhibitory neural activity, enhanced by attentional processes, seems to sharpen frequency tuning, contributing to improved auditory performance especially in noisy environments. In the present study, we investigated auditory magnetic fields in humans that were evoked by pure tones embedded in band-eliminated noises during two different stimulus sequencing conditions (constant vs. random under auditory focused attention by means of magnetoencephalography (MEG. Results In total, we used identical auditory stimuli between conditions, but presented them in a different order, thereby manipulating the neural processing and the auditory performance of the listeners. Constant stimulus sequencing blocks were characterized by the simultaneous presentation of pure tones of identical frequency with band-eliminated noises, whereas random sequencing blocks were characterized by the simultaneous presentation of pure tones of random frequencies and band-eliminated noises. We demonstrated that auditory evoked neural responses were larger in the constant sequencing compared to the random sequencing condition, particularly when the simultaneously presented noises contained narrow stop-bands. Conclusion The present study confirmed that population-level frequency tuning in human auditory cortex can be sharpened in a frequency-specific manner. This frequency-specific sharpening may contribute to improved auditory performance during detection and processing of relevant sound inputs characterized by specific frequency distributions in noisy environments.

  12. Power quality improvement of single-phase photovoltaic systems through a robust synchronization method

    DEFF Research Database (Denmark)

    Hadjidemetriou, Lenos; Kyriakides, Elias; Yang, Yongheng

    2014-01-01

    An increasing amount of single-phase photovoltaic (PV) systems on the distribution network requires more advanced synchronization methods in order to meet the grid codes with respect to power quality and fault ride through capability. The response of the synchronization technique selected...... is crucial for the performance of PV inverters. In this paper, a new synchronization method with good dynamics and accurate response under highly distorted voltage is proposed. This method uses a Multi-Harmonic Decoupling Cell (MHDC), which cancels out the oscillations on the synchronization signals due...

  13. The relationship between the age of onset of musical training and rhythm synchronization performance: Validation of sensitive period effects

    Directory of Open Access Journals (Sweden)

    Jennifer Anne Bailey

    2013-11-01

    Full Text Available A sensitive period associated with musical training has been proposed, suggesting the influence of musical training on the brain and behaviour is strongest during the early childhood years. Experiments from our laboratory have directly tested the sensitive period hypothesis for musical training by comparing musicians who began their training before age seven with those who began their training after age seven, while matching the two groups in terms of musical experience (Bailey & Penhune, 2010; 2012; Watanabe, Savion-Lemieux, & Penhune, 2007. Using this matching paradigm, the early-trained groups have demonstrated enhanced sensorimotor synchronization skills and associated differences in brain structure (Bailey, Zatorre, & Penhune, under review; Steele, Bailey, Zatorre, & Penhune, 2013. The current study takes a different approach to investigating the sensitive period hypothesis for musical training by examining a single large group of unmatched musicians (N=77 and exploring the relationship between age of onset of musical training as a continuous variable and performance on the Rhythm Synchronization Task (RST, a previously used auditory-motor rhythm synchronization task. Interestingly, age of onset was correlated with task performance for those who began training earlier; however, no such relationship was observed among those who began training in their later childhood years. In addition, years of formal training showed a similar pattern. However, individual working memory scores were predictive of task performance, regardless of age of onset of musical training. Overall, these results support the sensitive period hypothesis for musical training and suggest a non-linear relationship between age of onset of musical training and auditory-motor rhythm synchronization abilities, such that a relationship exists early in childhood but then plateaus later on in development, similar to maturational growth trajectories of brain regions implicated in

  14. Hearing after congenital deafness: central auditory plasticity and sensory deprivation.

    Science.gov (United States)

    Kral, A; Hartmann, R; Tillein, J; Heid, S; Klinke, R

    2002-08-01

    The congenitally deaf cat suffers from a degeneration of the inner ear. The organ of Corti bears no hair cells, yet the auditory afferents are preserved. Since these animals have no auditory experience, they were used as a model for congenital deafness. Kittens were equipped with a cochlear implant at different ages and electro-stimulated over a period of 2.0-5.5 months using a monopolar single-channel compressed analogue stimulation strategy (VIENNA-type signal processor). Following a period of auditory experience, we investigated cortical field potentials in response to electrical biphasic pulses applied by means of the cochlear implant. In comparison to naive unstimulated deaf cats and normal hearing cats, the chronically stimulated animals showed larger cortical regions producing middle-latency responses at or above 300 microV amplitude at the contralateral as well as the ipsilateral auditory cortex. The cortex ipsilateral to the chronically stimulated ear did not show any signs of reduced responsiveness when stimulating the 'untrained' ear through a second cochlear implant inserted in the final experiment. With comparable duration of auditory training, the activated cortical area was substantially smaller if implantation had been performed at an older age of 5-6 months. The data emphasize that young sensory systems in cats have a higher capacity for plasticity than older ones and that there is a sensitive period for the cat's auditory system.

  15. Impairments in musical abilities reflected in the auditory brainstem: evidence from congenital amusia.

    Science.gov (United States)

    Lehmann, Alexandre; Skoe, Erika; Moreau, Patricia; Peretz, Isabelle; Kraus, Nina

    2015-07-01

    Congenital amusia is a neurogenetic condition, characterized by a deficit in music perception and production, not explained by hearing loss, brain damage or lack of exposure to music. Despite inferior musical performance, amusics exhibit normal auditory cortical responses, with abnormal neural correlates suggested to lie beyond auditory cortices. Here we show, using auditory brainstem responses to complex sounds in humans, that fine-grained automatic processing of sounds is impoverished in amusia. Compared with matched non-musician controls, spectral amplitude was decreased in amusics for higher harmonic components of the auditory brainstem response. We also found a delayed response to the early transient aspects of the auditory stimulus in amusics. Neural measures of spectral amplitude and response timing correlated with participants' behavioral assessments of music processing. We demonstrate, for the first time, that amusia affects how complex acoustic signals are processed in the auditory brainstem. This neural signature of amusia mirrors what is observed in musicians, such that the aspects of the auditory brainstem responses that are enhanced in musicians are degraded in amusics. By showing that gradients of music abilities are reflected in the auditory brainstem, our findings have implications not only for current models of amusia but also for auditory functioning in general. © 2015 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  16. Rapid effects of hearing song on catecholaminergic activity in the songbird auditory pathway.

    Directory of Open Access Journals (Sweden)

    Lisa L Matragrano

    Full Text Available Catecholaminergic (CA neurons innervate sensory areas and affect the processing of sensory signals. For example, in birds, CA fibers innervate the auditory pathway at each level, including the midbrain, thalamus, and forebrain. We have shown previously that in female European starlings, CA activity in the auditory forebrain can be enhanced by exposure to attractive male song for one week. It is not known, however, whether hearing song can initiate that activity more rapidly. Here, we exposed estrogen-primed, female white-throated sparrows to conspecific male song and looked for evidence of rapid synthesis of catecholamines in auditory areas. In one hemisphere of the brain, we used immunohistochemistry to detect the phosphorylation of tyrosine hydroxylase (TH, a rate-limiting enzyme in the CA synthetic pathway. We found that immunoreactivity for TH phosphorylated at serine 40 increased dramatically in the auditory forebrain, but not the auditory thalamus and midbrain, after 15 min of song exposure. In the other hemisphere, we used high pressure liquid chromatography to measure catecholamines and their metabolites. We found that two dopamine metabolites, dihydroxyphenylacetic acid and homovanillic acid, increased in the auditory forebrain but not the auditory midbrain after 30 min of exposure to conspecific song. Our results are consistent with the hypothesis that exposure to a behaviorally relevant auditory stimulus rapidly induces CA activity, which may play a role in auditory responses.

  17. Auditory agnosia due to long-term severe hydrocephalus caused by spina bifida - specific auditory pathway versus nonspecific auditory pathway.

    Science.gov (United States)

    Zhang, Qing; Kaga, Kimitaka; Hayashi, Akimasa

    2011-07-01

    A 27-year-old female showed auditory agnosia after long-term severe hydrocephalus due to congenital spina bifida. After years of hydrocephalus, she gradually suffered from hearing loss in her right ear at 19 years of age, followed by her left ear. During the time when she retained some ability to hear, she experienced severe difficulty in distinguishing verbal, environmental, and musical instrumental sounds. However, her auditory brainstem response and distortion product otoacoustic emissions were largely intact in the left ear. Her bilateral auditory cortices were preserved, as shown by neuroimaging, whereas her auditory radiations were severely damaged owing to progressive hydrocephalus. Although she had a complete bilateral hearing loss, she felt great pleasure when exposed to music. After years of self-training to read lips, she regained fluent ability to communicate. Clinical manifestations of this patient indicate that auditory agnosia can occur after long-term hydrocephalus due to spina bifida; the secondary auditory pathway may play a role in both auditory perception and hearing rehabilitation.

  18. Linking topography to tonotopy in the mouse auditory thalamocortical circuit

    DEFF Research Database (Denmark)

    Hackett, Troy A; Rinaldi Barkat, Tania; O'Brien, Barbara M J

    2011-01-01

    The mouse sensory neocortex is reported to lack several hallmark features of topographic organization such as ocular dominance and orientation columns in primary visual cortex or fine-scale tonotopy in primary auditory cortex (AI). Here, we re-examined the question of auditory functional topography...... the tonotopic axis in the slice produced an orderly shift of voltage-sensitive dye (VSD) signals along the AI tonotopic axis, demonstrating topography in the mouse thalamocortical circuit that is preserved in the slice. However, compared with BF maps of neuronal spiking activity, the topographic order...... of subthreshold VSD maps was reduced in layer IV and even further degraded in layer II/III. Therefore, the precision of AI topography varies according to the source and layer of the mapping signal. Our findings further bridge the gap between in vivo and in vitro approaches for the detailed cellular study...

  19. MODELING SPECTRAL AND TEMPORAL MASKING IN THE HUMAN AUDITORY SYSTEM

    DEFF Research Database (Denmark)

    Dau, Torsten; Jepsen, Morten Løve; Ewert, Stephan D.

    2007-01-01

    An auditory signal processing model is presented that simulates psychoacoustical data from a large variety of experimental conditions related to spectral and temporal masking. The model is based on the modulation filterbank model by Dau et al. [J. Acoust. Soc. Am. 102, 2892-2905 (1997)] but inclu......An auditory signal processing model is presented that simulates psychoacoustical data from a large variety of experimental conditions related to spectral and temporal masking. The model is based on the modulation filterbank model by Dau et al. [J. Acoust. Soc. Am. 102, 2892-2905 (1997...... was tested in conditions of tone-in-noise masking, intensity discrimination, spectral masking with tones and narrowband noises, forward masking with (on- and off-frequency) noise- and pure-tone maskers, and amplitude modulation detection using different noise carrier bandwidths. One of the key properties...

  20. Auditory short-term memory in the primate auditory cortex.

    Science.gov (United States)

    Scott, Brian H; Mishkin, Mortimer

    2016-06-01

    Sounds are fleeting, and assembling the sequence of inputs at the ear into a coherent percept requires auditory memory across various time scales. Auditory short-term memory comprises at least two components: an active ׳working memory' bolstered by rehearsal, and a sensory trace that may be passively retained. Working memory relies on representations recalled from long-term memory, and their rehearsal may require phonological mechanisms unique to humans. The sensory component, passive short-term memory (pSTM), is tractable to study in nonhuman primates, whose brain architecture and behavioral repertoire are comparable to our own. This review discusses recent advances in the behavioral and neurophysiological study of auditory memory with a focus on single-unit recordings from macaque monkeys performing delayed-match-to-sample (DMS) tasks. Monkeys appear to employ pSTM to solve these tasks, as evidenced by the impact of interfering stimuli on memory performance. In several regards, pSTM in monkeys resembles pitch memory in humans, and may engage similar neural mechanisms. Neural correlates of DMS performance have been observed throughout the auditory and prefrontal cortex, defining a network of areas supporting auditory STM with parallels to that supporting visual STM. These correlates include persistent neural firing, or a suppression of firing, during the delay period of the memory task, as well as suppression or (less commonly) enhancement of sensory responses when a sound is repeated as a ׳match' stimulus. Auditory STM is supported by a distributed temporo-frontal network in which sensitivity to stimulus history is an intrinsic feature of auditory processing. This article is part of a Special Issue entitled SI: Auditory working memory. Published by Elsevier B.V.

  1. GTSO: Global Trace Synchronization and Ordering Mechanism for Wireless Sensor Network Monitoring Platforms.

    Science.gov (United States)

    Navia, Marlon; Campelo, José Carlos; Bonastre, Alberto; Ors, Rafael

    2017-12-23

    Monitoring is one of the best ways to evaluate the behavior of computer systems. When the monitored system is a distributed system-such as a wireless sensor network (WSN)-the monitoring operation must also be distributed, providing a distributed trace for further analysis. The temporal sequence of occurrence of the events registered by the distributed monitoring platform (DMP) must be correctly established to provide cause-effect relationships between them, so the logs obtained in different monitor nodes must be synchronized. Many of synchronization mechanisms applied to DMPs consist in adjusting the internal clocks of the nodes to the same value as a reference time. However, these mechanisms can create an incoherent event sequence. This article presents a new method to achieve global synchronization of the traces obtained in a DMP. It is based on periodic synchronization signals that are received by the monitor nodes and logged along with the recorded events. This mechanism processes all traces and generates a global post-synchronized trace by scaling all times registered proportionally according with the synchronization signals. It is intended to be a simple but efficient offline mechanism. Its application in a WSN-DMP demonstrates that it guarantees a correct ordering of the events, avoiding the aforementioned issues.

  2. GTSO: Global Trace Synchronization and Ordering Mechanism for Wireless Sensor Network Monitoring Platforms

    Science.gov (United States)

    Bonastre, Alberto; Ors, Rafael

    2017-01-01

    Monitoring is one of the best ways to evaluate the behavior of computer systems. When the monitored system is a distributed system—such as a wireless sensor network (WSN)—the monitoring operation must also be distributed, providing a distributed trace for further analysis. The temporal sequence of occurrence of the events registered by the distributed monitoring platform (DMP) must be correctly established to provide cause-effect relationships between them, so the logs obtained in different monitor nodes must be synchronized. Many of synchronization mechanisms applied to DMPs consist in adjusting the internal clocks of the nodes to the same value as a reference time. However, these mechanisms can create an incoherent event sequence. This article presents a new method to achieve global synchronization of the traces obtained in a DMP. It is based on periodic synchronization signals that are received by the monitor nodes and logged along with the recorded events. This mechanism processes all traces and generates a global post-synchronized trace by scaling all times registered proportionally according with the synchronization signals. It is intended to be a simple but efficient offline mechanism. Its application in a WSN-DMP demonstrates that it guarantees a correct ordering of the events, avoiding the aforementioned issues. PMID:29295494

  3. Keeping an eye on the conductor: neural correlates of visuo-motor synchronization and musical experience

    Science.gov (United States)

    Ono, Kentaro; Nakamura, Akinori; Maess, Burkhard

    2015-01-01

    For orchestra musicians, synchronized playing under a conductor’s direction is necessary to achieve optimal performance. Previous studies using simple auditory/visual stimuli have reported cortico-subcortical networks underlying synchronization and that training improves the accuracy of synchronization. However, it is unclear whether people who played regularly under a conductor and non-musicians activate the same networks when synchronizing with a conductor’s gestures. We conducted a functional magnetic resonance imaging (fMRI) experiment testing nonmusicians and musicians who regularly play music under a conductor. Participants were required to tap the rhythm they perceived from silent movies displaying either conductor’s gestures or a swinging metronome. Musicians performed tapping under a conductor with more precision than nonmusicians. Results from fMRI measurement showed greater activity in the anterior part of the left superior frontal gyrus (SFG) in musicians with more frequent practice under a conductor. Conversely, tapping with the metronome did not show any difference between musicians and nonmusicians, indicating that the expertize effect in tapping under the conductor does not result in a general increase in tapping performance for musicians. These results suggest that orchestra musicians have developed an advanced ability to predict conductor’s next action from the gestures. PMID:25883561

  4. Keeping an Eye on the Conductor: Neural Correlates of Visuo-motor Synchronization and Musical Experience

    Directory of Open Access Journals (Sweden)

    Kentaro eOno

    2015-04-01

    Full Text Available For orchestra musicians, synchronized playing under a conductor’s direction is necessary to achieve optimal performance. Previous studies using simple auditory/visual stimuli have reported cortico-subcortical networks underlying synchronization and that training improves the accuracy of synchronization. However, it is unclear whether people who played regularly under a conductor and non-musicians activate the same networks when synchronizing with a conductor’s gestures. We conducted a functional magnetic resonance imaging (fMRI experiment testing nonmusicians and musicians who regularly play music under a conductor. Participants were required to tap the rhythm they perceived from silent movies displaying either conductor’s gestures or a swinging metronome. Musicians performed tapping under a conductor with more precision than nonmusicians. Results from fMRI measurement showed greater activity in the anterior part of the left superior frontal gyrus (SFG in musicians with more frequent practice under a conductor. Conversely, tapping with the metronome did not show any difference between musicians and nonmusicians, indicating that the expertise effect in tapping under the conductor does not result in a general increase in tapping performance for musicians. These results suggest that orchestra musicians have developed an advanced ability to predict conductor’s next action from the gestures.

  5. Utility estimation of the application of auditory-visual-tactile sense feedback in respiratory gated radiation therapy

    Energy Technology Data Exchange (ETDEWEB)

    Jo, Jung Hun; KIm, Byeong Jin; Roh, Shi Won; Lee, Hyeon Chan; Jang, Hyeong Jun; Kim, Hoi Nam [Dept. of Radiation Oncology, Biomedical Engineering, Seoul St. Mary' s Hospital, Seoul (Korea, Republic of); Song, Jae Hoon [Dept. of Biomedical Engineering, Seoul St. Mary' s Hospital, Seoul (Korea, Republic of); Kim, Young Jae [Dept. of Radiological Technology, Gwang Yang Health Collage, Gwangyang (Korea, Republic of)

    2013-03-15

    The purpose of this study was to evaluate the possibility to optimize the gated treatment delivery time and maintenance of stable respiratory by the introduction of breath with the assistance of auditory-visual-tactile sense. The experimenter's respiration were measured by ANZAI 4D system. We obtained natural breathing signal, monitor-induced breathing signal, monitor and ventilator-induced breathing signal, and breath-hold signal using real time monitor during 10 minutes beam-on-time. In order to check the stability of respiratory signals distributed in each group were compared with means, standard deviation, variation value, beam{sub t}ime of the respiratory signal. The stability of each respiratory was measured in consideration of deviation change studied in each respiratory time lapse. As a result of an analysis of respiratory signal, all experimenters has showed that breathing signal used both Real time monitor and Ventilator was the most stable and shortest time. In this study, it was evaluated that respiratory gated radiation therapy with auditory-visual-tactual sense and without auditory-visual-tactual sense feedback. The study showed that respiratory gated radiation therapy delivery time could significantly be improved by the application of video feedback when this is combined with audio-tactual sense assistance. This delivery technique did prove its feasibility to limit the tumor motion during treatment delivery for all patients to a defined value while maintaining the accuracy and proved the applicability of the technique in a conventional clinical schedule.

  6. Utility estimation of the application of auditory-visual-tactile sense feedback in respiratory gated radiation therapy

    International Nuclear Information System (INIS)

    Jo, Jung Hun; KIm, Byeong Jin; Roh, Shi Won; Lee, Hyeon Chan; Jang, Hyeong Jun; Kim, Hoi Nam; Song, Jae Hoon; Kim, Young Jae

    2013-01-01

    The purpose of this study was to evaluate the possibility to optimize the gated treatment delivery time and maintenance of stable respiratory by the introduction of breath with the assistance of auditory-visual-tactile sense. The experimenter's respiration were measured by ANZAI 4D system. We obtained natural breathing signal, monitor-induced breathing signal, monitor and ventilator-induced breathing signal, and breath-hold signal using real time monitor during 10 minutes beam-on-time. In order to check the stability of respiratory signals distributed in each group were compared with means, standard deviation, variation value, beam t ime of the respiratory signal. The stability of each respiratory was measured in consideration of deviation change studied in each respiratory time lapse. As a result of an analysis of respiratory signal, all experimenters has showed that breathing signal used both Real time monitor and Ventilator was the most stable and shortest time. In this study, it was evaluated that respiratory gated radiation therapy with auditory-visual-tactual sense and without auditory-visual-tactual sense feedback. The study showed that respiratory gated radiation therapy delivery time could significantly be improved by the application of video feedback when this is combined with audio-tactual sense assistance. This delivery technique did prove its feasibility to limit the tumor motion during treatment delivery for all patients to a defined value while maintaining the accuracy and proved the applicability of the technique in a conventional clinical schedule

  7. Role of synchronized oscillatory brain activity for human pain perception.

    Science.gov (United States)

    Hauck, Michael; Lorenz, Jürgen; Engel, Andreas K

    2008-01-01

    The understanding of cortical pain processing in humans has significantly improved since the development of modern neuroimaging techniques. Non-invasive electrophysiological approaches such as electro- and magnetoencephalography have proven to be helpful tools for the real-time investigation of neuronal signals and synchronous communication between cortical areas. In particular, time-frequency decomposition of signals recorded with these techniques seems to be a promising approach because different pain-related oscillatory changes can be observed within different frequency bands, which are likely to be linked to specific sensory and motor functions. In this review we discuss the latest evidence on pain-induced time-frequency signals and propose that changes in oscillatory activity reflect an essential communication mechanism in the brain that is modulated during pain processing. The importance of synchronization processes for normal and pathological pain processing, such as chronic pain states, is discussed.

  8. EEG phase reset due to auditory attention: an inverse time-scale approach

    International Nuclear Information System (INIS)

    Low, Yin Fen; Strauss, Daniel J

    2009-01-01

    We propose a novel tool to evaluate the electroencephalograph (EEG) phase reset due to auditory attention by utilizing an inverse analysis of the instantaneous phase for the first time. EEGs were acquired through auditory attention experiments with a maximum entropy stimulation paradigm. We examined single sweeps of auditory late response (ALR) with the complex continuous wavelet transform. The phase in the frequency band that is associated with auditory attention (6–10 Hz, termed as theta–alpha border) was reset to the mean phase of the averaged EEGs. The inverse transform was applied to reconstruct the phase-modified signal. We found significant enhancement of the N100 wave in the reconstructed signal. Analysis of the phase noise shows the effects of phase jittering on the generation of the N100 wave implying that a preferred phase is necessary to generate the event-related potential (ERP). Power spectrum analysis shows a remarkable increase of evoked power but little change of total power after stabilizing the phase of EEGs. Furthermore, by resetting the phase only at the theta border of no attention data to the mean phase of attention data yields a result that resembles attention data. These results show strong connections between EEGs and ERP, in particular, we suggest that the presentation of an auditory stimulus triggers the phase reset process at the theta–alpha border which leads to the emergence of the N100 wave. It is concluded that our study reinforces other studies on the importance of the EEG in ERP genesis

  9. EEG phase reset due to auditory attention: an inverse time-scale approach.

    Science.gov (United States)

    Low, Yin Fen; Strauss, Daniel J

    2009-08-01

    We propose a novel tool to evaluate the electroencephalograph (EEG) phase reset due to auditory attention by utilizing an inverse analysis of the instantaneous phase for the first time. EEGs were acquired through auditory attention experiments with a maximum entropy stimulation paradigm. We examined single sweeps of auditory late response (ALR) with the complex continuous wavelet transform. The phase in the frequency band that is associated with auditory attention (6-10 Hz, termed as theta-alpha border) was reset to the mean phase of the averaged EEGs. The inverse transform was applied to reconstruct the phase-modified signal. We found significant enhancement of the N100 wave in the reconstructed signal. Analysis of the phase noise shows the effects of phase jittering on the generation of the N100 wave implying that a preferred phase is necessary to generate the event-related potential (ERP). Power spectrum analysis shows a remarkable increase of evoked power but little change of total power after stabilizing the phase of EEGs. Furthermore, by resetting the phase only at the theta border of no attention data to the mean phase of attention data yields a result that resembles attention data. These results show strong connections between EEGs and ERP, in particular, we suggest that the presentation of an auditory stimulus triggers the phase reset process at the theta-alpha border which leads to the emergence of the N100 wave. It is concluded that our study reinforces other studies on the importance of the EEG in ERP genesis.

  10. Directional Networking in GPS Denied Environments - Time Synchronization

    Science.gov (United States)

    2016-03-14

    RF-based measurements to synchronize time and measure node range.  Satellite Doppler: Using Doppler measurements from multiple satellites along...with satellite catalog data to determine time and position.  LTE : Use existing LTE base-stations for time and position.  Differential GPS: A...Opportunistic Signals: Opportunistically take advantage of existing RF signals (i.e., FM radio, DTV, LTE , etc.) transmitted from known locations

  11. Normal form and synchronization of strict-feedback chaotic systems

    International Nuclear Information System (INIS)

    Wang, Feng; Chen, Shihua; Yu Minghai; Wang Changping

    2004-01-01

    This study concerns the normal form and synchronization of strict-feedback chaotic systems. We prove that, any strict-feedback chaotic system can be rendered into a normal form with a invertible transform and then a design procedure to synchronize the normal form of a non-autonomous strict-feedback chaotic system is presented. This approach needs only a scalar driving signal to realize synchronization no matter how many dimensions the chaotic system contains. Furthermore, the Roessler chaotic system is taken as a concrete example to illustrate the procedure of designing without transforming a strict-feedback chaotic system into its normal form. Numerical simulations are also provided to show the effectiveness and feasibility of the developed methods

  12. Synchronization and matched filtering in time-frequency using the sunflower spiral

    NARCIS (Netherlands)

    Korevaar, C.W.; Kokkeler, Andre B.J.; de Boer, Pieter-Tjerk; Smit, Gerardus Johannes Maria

    2012-01-01

    Synchronization and matched filtering of signals in time dispersive, frequency dispersive and time-frequency dispersive channels are addressed in this paper. The ‘eigenfunctions’ of these channels form the signal sets under investigation. While using channel-eigenfunctions is a first requirement for

  13. Reliability-Weighted Integration of Audiovisual Signals Can Be Modulated by Top-down Attention

    Science.gov (United States)

    Noppeney, Uta

    2018-01-01

    Abstract Behaviorally, it is well established that human observers integrate signals near-optimally weighted in proportion to their reliabilities as predicted by maximum likelihood estimation. Yet, despite abundant behavioral evidence, it is unclear how the human brain accomplishes this feat. In a spatial ventriloquist paradigm, participants were presented with auditory, visual, and audiovisual signals and reported the location of the auditory or the visual signal. Combining psychophysics, multivariate functional MRI (fMRI) decoding, and models of maximum likelihood estimation (MLE), we characterized the computational operations underlying audiovisual integration at distinct cortical levels. We estimated observers’ behavioral weights by fitting psychometric functions to participants’ localization responses. Likewise, we estimated the neural weights by fitting neurometric functions to spatial locations decoded from regional fMRI activation patterns. Our results demonstrate that low-level auditory and visual areas encode predominantly the spatial location of the signal component of a region’s preferred auditory (or visual) modality. By contrast, intraparietal sulcus forms spatial representations by integrating auditory and visual signals weighted by their reliabilities. Critically, the neural and behavioral weights and the variance of the spatial representations depended not only on the sensory reliabilities as predicted by the MLE model but also on participants’ modality-specific attention and report (i.e., visual vs. auditory). These results suggest that audiovisual integration is not exclusively determined by bottom-up sensory reliabilities. Instead, modality-specific attention and report can flexibly modulate how intraparietal sulcus integrates sensory signals into spatial representations to guide behavioral responses (e.g., localization and orienting). PMID:29527567

  14. Auditory Sketches: Very Sparse Representations of Sounds Are Still Recognizable.

    Directory of Open Access Journals (Sweden)

    Vincent Isnard

    Full Text Available Sounds in our environment like voices, animal calls or musical instruments are easily recognized by human listeners. Understanding the key features underlying this robust sound recognition is an important question in auditory science. Here, we studied the recognition by human listeners of new classes of sounds: acoustic and auditory sketches, sounds that are severely impoverished but still recognizable. Starting from a time-frequency representation, a sketch is obtained by keeping only sparse elements of the original signal, here, by means of a simple peak-picking algorithm. Two time-frequency representations were compared: a biologically grounded one, the auditory spectrogram, which simulates peripheral auditory filtering, and a simple acoustic spectrogram, based on a Fourier transform. Three degrees of sparsity were also investigated. Listeners were asked to recognize the category to which a sketch sound belongs: singing voices, bird calls, musical instruments, and vehicle engine noises. Results showed that, with the exception of voice sounds, very sparse representations of sounds (10 features, or energy peaks, per second could be recognized above chance. No clear differences could be observed between the acoustic and the auditory sketches. For the voice sounds, however, a completely different pattern of results emerged, with at-chance or even below-chance recognition performances, suggesting that the important features of the voice, whatever they are, were removed by the sketch process. Overall, these perceptual results were well correlated with a model of auditory distances, based on spectro-temporal excitation patterns (STEPs. This study confirms the potential of these new classes of sounds, acoustic and auditory sketches, to study sound recognition.

  15. Changes in auditory perceptions and cortex resulting from hearing recovery after extended congenital unilateral hearing loss

    Directory of Open Access Journals (Sweden)

    Jill B Firszt

    2013-12-01

    Full Text Available Monaural hearing induces auditory system reorganization. Imbalanced input also degrades time-intensity cues for sound localization and signal segregation for listening in noise. While there have been studies of bilateral auditory deprivation and later hearing restoration (e.g. cochlear implants, less is known about unilateral auditory deprivation and subsequent hearing improvement. We investigated effects of long-term congenital unilateral hearing loss on localization, speech understanding, and cortical organization following hearing recovery. Hearing in the congenitally affected ear of a 41 year old female improved significantly after stapedotomy and reconstruction. Pre-operative hearing threshold levels showed unilateral, mixed, moderately-severe to profound hearing loss. The contralateral ear had hearing threshold levels within normal limits. Testing was completed prior to, and three and nine months after surgery. Measurements were of sound localization with intensity-roved stimuli and speech recognition in various noise conditions. We also evoked magnetic resonance signals with monaural stimulation to the unaffected ear. Activation magnitudes were determined in core, belt, and parabelt auditory cortex regions via an interrupted single event design. Hearing improvement following 40 years of congenital unilateral hearing loss resulted in substantially improved sound localization and speech recognition in noise. Auditory cortex also reorganized. Contralateral auditory cortex responses were increased after hearing recovery and the extent of activated cortex was bilateral, including a greater portion of the posterior superior temporal plane. Thus, prolonged predominant monaural stimulation did not prevent auditory system changes consequent to restored binaural hearing. Results support future research of unilateral auditory deprivation effects and plasticity, with consideration for length of deprivation, age at hearing correction, degree and type

  16. Auditory analysis for speech recognition based on physiological models

    Science.gov (United States)

    Jeon, Woojay; Juang, Biing-Hwang

    2004-05-01

    To address the limitations of traditional cepstrum or LPC based front-end processing methods for automatic speech recognition, more elaborate methods based on physiological models of the human auditory system may be used to achieve more robust speech recognition in adverse environments. For this purpose, a modified version of a model of the primary auditory cortex featuring a three dimensional mapping of auditory spectra [Wang and Shamma, IEEE Trans. Speech Audio Process. 3, 382-395 (1995)] is adopted and investigated for its use as an improved front-end processing method. The study is conducted in two ways: first, by relating the model's redundant representation to traditional spectral representations and showing that the former not only encompasses information provided by the latter, but also reveals more relevant information that makes it superior in describing the identifying features of speech signals; and second, by observing the statistical features of the representation for various classes of sound to show how different identifying features manifest themselves as specific patterns on the cortical map, thereby becoming a place-coded data set on which detection theory could be applied to simulate auditory perception and cognition.

  17. Amplification through chaotic synchronization in spatially extended beam-plasma systems

    Science.gov (United States)

    Moskalenko, Olga I.; Frolov, Nikita S.; Koronovskii, Alexey A.; Hramov, Alexander E.

    2017-12-01

    In this paper, we have studied the relationship between chaotic synchronization and microwave signal amplification in coupled beam-plasma systems. We have considered a 1D particle-in-cell numerical model of unidirectionally coupled beam-plasma oscillatory media being in the regime of electron pattern formation. We have shown the significant gain of microwave oscillation power in coupled beam-plasma media being in the different regimes of generation. The discovered effect has a close connection with the chaotic synchronization phenomenon, so we have observed that amplification appears after the onset of the complete time scale synchronization regime in the analyzed coupled spatially extended systems. We have also provided the numerical study of physical processes in the chain of beam-plasma systems leading to the chaotic synchronization and the amplification of microwave oscillations power, respectively.

  18. Synchronous Analysis for Diagnostics of Rolling Bearings in the Turbojet Engine

    Directory of Open Access Journals (Sweden)

    Gosiewski Zdzisław

    2005-01-01

    Full Text Available Designed in the Aviation Institute, the K-15 turbojet engine has got rolling bearings, which answers with frequency 5.87 × ω r the unbalance excitation. The signal with such frequency indicates a fault of the outer race of the rolling bearing. A set of the digital synchronous filters was used for the K-15 vibration spectrum analysis. A procedure of filtration was performed by the computer software. The sychronous summation of the measured signals was carried out before the spectrum analysis. Two cases were considered: the engine with a small force due to unbalance (a small angular velocity of the rotor, and the engine with a big unbalance force (high angular velocity. In the first case, the outer race frequency was not observed, despite the existence of the vibration amplitude (caused by unknown disturbances with such frequency before the synchronous summation. In the second case, the outer race frequency after synchronous summation has enlarged amplitude while other spectrum components in its vicinity have been damped. It underlines the usefulness of the synchronous analysis in the vibration diagnostics of the rotating machinery.

  19. Stuttering adults' lack of pre-speech auditory modulation normalizes when speaking with delayed auditory feedback.

    Science.gov (United States)

    Daliri, Ayoub; Max, Ludo

    2018-02-01

    Auditory modulation during speech movement planning is limited in adults who stutter (AWS), but the functional relevance of the phenomenon itself remains unknown. We investigated for AWS and adults who do not stutter (AWNS) (a) a potential relationship between pre-speech auditory modulation and auditory feedback contributions to speech motor learning and (b) the effect on pre-speech auditory modulation of real-time versus delayed auditory feedback. Experiment I used a sensorimotor adaptation paradigm to estimate auditory-motor speech learning. Using acoustic speech recordings, we quantified subjects' formant frequency adjustments across trials when continually exposed to formant-shifted auditory feedback. In Experiment II, we used electroencephalography to determine the same subjects' extent of pre-speech auditory modulation (reductions in auditory evoked potential N1 amplitude) when probe tones were delivered prior to speaking versus not speaking. To manipulate subjects' ability to monitor real-time feedback, we included speaking conditions with non-altered auditory feedback (NAF) and delayed auditory feedback (DAF). Experiment I showed that auditory-motor learning was limited for AWS versus AWNS, and the extent of learning was negatively correlated with stuttering frequency. Experiment II yielded several key findings: (a) our prior finding of limited pre-speech auditory modulation in AWS was replicated; (b) DAF caused a decrease in auditory modulation for most AWNS but an increase for most AWS; and (c) for AWS, the amount of auditory modulation when speaking with DAF was positively correlated with stuttering frequency. Lastly, AWNS showed no correlation between pre-speech auditory modulation (Experiment II) and extent of auditory-motor learning (Experiment I) whereas AWS showed a negative correlation between these measures. Thus, findings suggest that AWS show deficits in both pre-speech auditory modulation and auditory-motor learning; however, limited pre

  20. Comparison of Interpolation Methods as Applied to Time Synchronous Averaging

    National Research Council Canada - National Science Library

    Decker, Harry

    1999-01-01

    Several interpolation techniques were investigated to determine their effect on time synchronous averaging of gear vibration signals and also the effects on standard health monitoring diagnostic parameters...

  1. Evidence of functional connectivity between auditory cortical areas revealed by amplitude modulation sound processing.

    Science.gov (United States)

    Guéguin, Marie; Le Bouquin-Jeannès, Régine; Faucon, Gérard; Chauvel, Patrick; Liégeois-Chauvel, Catherine

    2007-02-01

    The human auditory cortex includes several interconnected areas. A better understanding of the mechanisms involved in auditory cortical functions requires a detailed knowledge of neuronal connectivity between functional cortical regions. In human, it is difficult to track in vivo neuronal connectivity. We investigated the interarea connection in vivo in the auditory cortex using a method of directed coherence (DCOH) applied to depth auditory evoked potentials (AEPs). This paper presents simultaneous AEPs recordings from insular gyrus (IG), primary and secondary cortices (Heschl's gyrus and planum temporale), and associative areas (Brodmann area [BA] 22) with multilead intracerebral electrodes in response to sinusoidal modulated white noises in 4 epileptic patients who underwent invasive monitoring with depth electrodes for epilepsy surgery. DCOH allowed estimation of the causality between 2 signals recorded from different cortical sites. The results showed 1) a predominant auditory stream within the primary auditory cortex from the most medial region to the most lateral one whatever the modulation frequency, 2) unidirectional functional connection from the primary to secondary auditory cortex, 3) a major auditory propagation from the posterior areas to the anterior ones, particularly at 8, 16, and 32 Hz, and 4) a particular role of Heschl's sulcus dispatching information to the different auditory areas. These findings suggest that cortical processing of auditory information is performed in serial and parallel streams. Our data showed that the auditory propagation could not be associated to a unidirectional traveling wave but to a constant interaction between these areas that could reflect the large adaptive and plastic capacities of auditory cortex. The role of the IG is discussed.

  2. Effects of Sleep Deprivation on Phase Synchronization as Assessed by Wavelet Phase Coherence Analysis of Prefrontal Tissue Oxyhemoglobin Signals.

    Science.gov (United States)

    Bu, Lingguo; Zhang, Ming; Li, Jianfeng; Li, Fangyi; Liu, Heshan; Li, Zengyong

    2017-01-01

    To reveal the physiological mechanism of the decline in cognitive function after sleep deprivation, a within-subject study was performed to assess sleep deprivation effects on phase synchronization, as revealed by wavelet phase coherence (WPCO) analysis of prefrontal tissue oxyhemoglobin signals. Twenty subjects (10 male and 10 female, 25.5 ± 3.5 years old) were recruited to participate in two tests: one without sleep deprivation (group A) and the other with 24 h of sleep deprivation (group B). Before the test, each subject underwent a subjective evaluation using visual analog scales. A cognitive task was performed by judging three random numbers. Continuous recordings of the near-infrared spectroscopy (NIRS) signals were obtained from both the left and right prefrontal lobes during rest, task, and post-task periods. The WPCO of cerebral Delta [HbO2] signals were analyzed for these three periods for both groups A and B. Six frequency intervals were defined: I: 0.6-2 Hz (cardiac activity), II: 0.145-0.6 Hz (respiratory activity), III: 0.052-0.145 Hz (myogenic activity), IV: 0.021-0.052 Hz (neurogenic activity), V: 0.0095-0.021 Hz (nitric oxide related endothelial activity) and VI: 0.005-0.0095 Hz (non-nitric oxide related endothelial activity). WPCO in intervals III (F = 5.955, p = 0.02) and V (F = 4.7, p = 0.037) was significantly lower in group B than in group A at rest. During the task period, WPCO in intervals III (F = 5.175, p = 0.029) and IV (F = 4.585, p = 0.039) was significantly lower in group B compared with group A. In the post-task recovery period, the WPCO in interval III (F = 6.125, p = 0.02) was significantly lower in group B compared with group A. Reaction time was significantly prolonged, and the accuracy rate and F1 score both declined after sleep deprivation. The decline in WPCO after sleep deprivation indicates reduced phase synchronization between left and right prefrontal oxyhemoglobin oscillations, which may contribute to the diminished

  3. Design of Wireless Automatic Synchronization for the Low-Frequency Coded Ground Penetrating Radar

    Directory of Open Access Journals (Sweden)

    Zhenghuan Xia

    2015-01-01

    Full Text Available Low-frequency coded ground penetrating radar (GPR with a pair of wire dipole antennas has some advantages for deep detection. Due to the large distance between the two antennas, the synchronization design is a major challenge of implementing the GPR system. This paper proposes a simple and stable wireless automatic synchronization method based on our developed GPR system, which does not need any synchronization chips or modules and reduces the cost of the hardware system. The transmitter omits the synchronization preamble and pseudorandom binary sequence (PRBS at an appropriate time interval, while receiver automatically estimates the synchronization time and receives the returned signal from the underground targets. All the processes are performed in a single FPGA. The performance of the proposed synchronization method is validated with experiment.

  4. Biological Impact of Music and Software-Based Auditory Training

    Science.gov (United States)

    Kraus, Nina

    2012-01-01

    Auditory-based communication skills are developed at a young age and are maintained throughout our lives. However, some individuals--both young and old--encounter difficulties in achieving or maintaining communication proficiency. Biological signals arising from hearing sounds relate to real-life communication skills such as listening to speech in…

  5. Auditory, visual and auditory-visual memory and sequencing performance in typically developing children.

    Science.gov (United States)

    Pillai, Roshni; Yathiraj, Asha

    2017-09-01

    The study evaluated whether there exists a difference/relation in the way four different memory skills (memory score, sequencing score, memory span, & sequencing span) are processed through the auditory modality, visual modality and combined modalities. Four memory skills were evaluated on 30 typically developing children aged 7 years and 8 years across three modality conditions (auditory, visual, & auditory-visual). Analogous auditory and visual stimuli were presented to evaluate the three modality conditions across the two age groups. The children obtained significantly higher memory scores through the auditory modality compared to the visual modality. Likewise, their memory scores were significantly higher through the auditory-visual modality condition than through the visual modality. However, no effect of modality was observed on the sequencing scores as well as for the memory and the sequencing span. A good agreement was seen between the different modality conditions that were studied (auditory, visual, & auditory-visual) for the different memory skills measures (memory scores, sequencing scores, memory span, & sequencing span). A relatively lower agreement was noted only between the auditory and visual modalities as well as between the visual and auditory-visual modality conditions for the memory scores, measured using Bland-Altman plots. The study highlights the efficacy of using analogous stimuli to assess the auditory, visual as well as combined modalities. The study supports the view that the performance of children on different memory skills was better through the auditory modality compared to the visual modality. Copyright © 2017 Elsevier B.V. All rights reserved.

  6. Can you hear me now? Musical training shapes functional brain networks for selective auditory attention and hearing speech in noise

    Directory of Open Access Journals (Sweden)

    Dana L Strait

    2011-06-01

    Full Text Available Even in the quietest of rooms, our senses are perpetually inundated by a barrage of sounds, requiring the auditory system to adapt to a variety of listening conditions in order to extract signals of interest (e.g., one speaker’s voice amidst others. Brain networks that promote selective attention are thought to sharpen the neural encoding of a target signal, suppressing competing sounds and enhancing perceptual performance. Here, we ask: does musical training benefit cortical mechanisms that underlie selective attention to speech? To answer this question, we assessed the impact of selective auditory attention on cortical auditory-evoked response variability in musicians and nonmusicians. Outcomes indicate strengthened brain networks for selective auditory attention in musicians in that musicians but not nonmusicians demonstrate decreased prefrontal response variability with auditory attention. Results are interpreted in the context of previous work from our laboratory documenting perceptual and subcortical advantages in musicians for the hearing and neural encoding of speech in background noise. Musicians’ neural proficiency for selectively engaging and sustaining auditory attention to language indicates a potential benefit of music for auditory training. Given the importance of auditory attention for the development of language-related skills, musical training may aid in the prevention, habilitation and remediation of children with a wide range of attention-based language and learning impairments.

  7. Time synchronization of a wired sensor network for structural health monitoring

    International Nuclear Information System (INIS)

    Ishikawa, Ken-ichiro; Mita, Akira

    2008-01-01

    This paper introduces a time synchronization system for wired smart sensor networks to be applied to the structural health monitoring of gigantic structures. The jitter of sensor nodes in the wired network depends on the wire length between the origin and the destination of the time synchronization signals. The proposed system can theoretically achieve the accuracy to limit the jitter of sensors within 34 ns by adjusting the timing depending on the wire length, and experimentally showed the jitter of 190 m separation to be within 25 ns. The proposed system uses local area network (LAN) cables and does not require additional cabling for synchronization. Thus the proposed synchronization system can be embedded in the sensor network with minimal cost

  8. Gene expression underlying enhanced, steroid-dependent auditory sensitivity of hair cell epithelium in a vocal fish.

    Science.gov (United States)

    Fergus, Daniel J; Feng, Ni Y; Bass, Andrew H

    2015-10-14

    Successful animal communication depends on a receiver's ability to detect a sender's signal. Exemplars of adaptive sender-receiver coupling include acoustic communication, often important in the context of seasonal reproduction. During the reproductive summer season, both male and female midshipman fish (Porichthys notatus) exhibit similar increases in the steroid-dependent frequency sensitivity of the saccule, the main auditory division of the inner ear. This form of auditory plasticity enhances detection of the higher frequency components of the multi-harmonic, long-duration advertisement calls produced repetitively by males during summer nights of peak vocal and spawning activity. The molecular basis of this seasonal auditory plasticity has not been fully resolved. Here, we utilize an unbiased transcriptomic RNA sequencing approach to identify differentially expressed transcripts within the saccule's hair cell epithelium of reproductive summer and non-reproductive winter fish. We assembled 74,027 unique transcripts from our saccular epithelial sequence reads. Of these, 6.4 % and 3.0 % were upregulated in the reproductive and non-reproductive saccular epithelium, respectively. Gene ontology (GO) term enrichment analyses of the differentially expressed transcripts showed that the reproductive saccular epithelium was transcriptionally, translationally, and metabolically more active than the non-reproductive epithelium. Furthermore, the expression of a specific suite of candidate genes, including ion channels and components of steroid-signaling pathways, was upregulated in the reproductive compared to the non-reproductive saccular epithelium. We found reported auditory functions for 14 candidate genes upregulated in the reproductive midshipman saccular epithelium, 8 of which are enriched in mouse hair cells, validating their hair cell-specific functions across vertebrates. We identified a suite of differentially expressed genes belonging to neurotransmission and

  9. Carrier and symbol synchronization system performance study

    Science.gov (United States)

    Lindsey, W. C.

    1976-01-01

    Results pertinent to predicting the performance of convolutionally encoded binary phase-shift keyed communication links were presented. The details of the development are provided in four sections. These sections are concerned with developing the bit error probability performance degradations due to PN despreading by a time-shared delay locked loop, the Costas demodulation process, symbol synchronization effects and cycle slipping phenomena in the Costas loop. In addition, Costas cycle slipping probabilities are studied as functions of Doppler count time and signal-to-noise conditions. The effect of cycle slipping in the symbol synchronizer is also studied as a function of channel Doppler and other frequency uncertainties.

  10. Low-cost synchronization of high-speed audio and video recordings in bio-acoustic experiments.

    Science.gov (United States)

    Laurijssen, Dennis; Verreycken, Erik; Geipel, Inga; Daems, Walter; Peremans, Herbert; Steckel, Jan

    2018-02-27

    In this paper, we present a method for synchronizing high-speed audio and video recordings of bio-acoustic experiments. By embedding a random signal into the recorded video and audio data, robust synchronization of a diverse set of sensor streams can be performed without the need to keep detailed records. The synchronization can be performed using recording devices without dedicated synchronization inputs. We demonstrate the efficacy of the approach in two sets of experiments: behavioral experiments on different species of echolocating bats and the recordings of field crickets. We present the general operating principle of the synchronization method, discuss its synchronization strength and provide insights into how to construct such a device using off-the-shelf components. © 2018. Published by The Company of Biologists Ltd.

  11. Blind I/Q Signal Separation-Based Solutions for Receiver Signal Processing

    Directory of Open Access Journals (Sweden)

    Visa Koivunen

    2005-09-01

    Full Text Available This paper introduces some novel digital signal processing (DSP-based approaches to some of the most fundamental tasks of radio receivers, namely, channel equalization, carrier synchronization, and I/Q mismatch compensation. The leading principle is to show that all these problems can be solved blindly (i.e., without training signals by forcing the I and Q components of the observed data as independent as possible. Blind signal separation (BSS is then introduced as an efficient tool to carry out these tasks, and simulation examples are used to illustrate the performance of the proposed approaches. The main application area of the presented carrier synchronization and I/Q mismatch compensation techniques is in direct-conversion type receivers, while the proposed channel equalization principles basically apply to any radio architecture.

  12. Modelling the Emergence and Dynamics of Perceptual Organisation in Auditory Streaming

    Science.gov (United States)

    Mill, Robert W.; Bőhm, Tamás M.; Bendixen, Alexandra; Winkler, István; Denham, Susan L.

    2013-01-01

    Many sound sources can only be recognised from the pattern of sounds they emit, and not from the individual sound events that make up their emission sequences. Auditory scene analysis addresses the difficult task of interpreting the sound world in terms of an unknown number of discrete sound sources (causes) with possibly overlapping signals, and therefore of associating each event with the appropriate source. There are potentially many different ways in which incoming events can be assigned to different causes, which means that the auditory system has to choose between them. This problem has been studied for many years using the auditory streaming paradigm, and recently it has become apparent that instead of making one fixed perceptual decision, given sufficient time, auditory perception switches back and forth between the alternatives—a phenomenon known as perceptual bi- or multi-stability. We propose a new model of auditory scene analysis at the core of which is a process that seeks to discover predictable patterns in the ongoing sound sequence. Representations of predictable fragments are created on the fly, and are maintained, strengthened or weakened on the basis of their predictive success, and conflict with other representations. Auditory perceptual organisation emerges spontaneously from the nature of the competition between these representations. We present detailed comparisons between the model simulations and data from an auditory streaming experiment, and show that the model accounts for many important findings, including: the emergence of, and switching between, alternative organisations; the influence of stimulus parameters on perceptual dominance, switching rate and perceptual phase durations; and the build-up of auditory streaming. The principal contribution of the model is to show that a two-stage process of pattern discovery and competition between incompatible patterns can account for both the contents (perceptual organisations) and the

  13. Global synchronization of parallel processors using clock pulse width modulation

    Science.gov (United States)

    Chen, Dong; Ellavsky, Matthew R.; Franke, Ross L.; Gara, Alan; Gooding, Thomas M.; Haring, Rudolf A.; Jeanson, Mark J.; Kopcsay, Gerard V.; Liebsch, Thomas A.; Littrell, Daniel; Ohmacht, Martin; Reed, Don D.; Schenck, Brandon E.; Swetz, Richard A.

    2013-04-02

    A circuit generates a global clock signal with a pulse width modification to synchronize processors in a parallel computing system. The circuit may include a hardware module and a clock splitter. The hardware module may generate a clock signal and performs a pulse width modification on the clock signal. The pulse width modification changes a pulse width within a clock period in the clock signal. The clock splitter may distribute the pulse width modified clock signal to a plurality of processors in the parallel computing system.

  14. Social interaction with a tutor modulates responsiveness of specific auditory neurons in juvenile zebra finches.

    Science.gov (United States)

    Yanagihara, Shin; Yazaki-Sugiyama, Yoko

    2018-04-12

    Behavioral states of animals, such as observing the behavior of a conspecific, modify signal perception and/or sensations that influence state-dependent higher cognitive behavior, such as learning. Recent studies have shown that neuronal responsiveness to sensory signals is modified when animals are engaged in social interactions with others or in locomotor activities. However, how these changes produce state-dependent differences in higher cognitive function is still largely unknown. Zebra finches, which have served as the premier songbird model, learn to sing from early auditory experiences with tutors. They also learn from playback of recorded songs however, learning can be greatly improved when song models are provided through social communication with tutors (Eales, 1989; Chen et al., 2016). Recently we found a subset of neurons in the higher-level auditory cortex of juvenile zebra finches that exhibit highly selective auditory responses to the tutor song after song learning, suggesting an auditory memory trace of the tutor song (Yanagihara and Yazaki-Sugiyama, 2016). Here we show that auditory responses of these selective neurons became greater when juveniles were paired with their tutors, while responses of non-selective neurons did not change. These results suggest that social interaction modulates cortical activity and might function in state-dependent song learning. Copyright © 2018 Elsevier B.V. All rights reserved.

  15. Multimodal signal variation in space and time : how important is matching a signal with its signaler?

    OpenAIRE

    Taylor, Ryan C.; Klein, Barrett; Stein, Joey; Ryan, Michael J.

    2011-01-01

    Multimodal signals (acoustic+visual) are known to be used by many anuran amphibians during courtship displays. The relative degree to which each signal component influences female mate choice, however, remains poorly understood. In this study we used a robotic frog with an inflating vocal sac and acoustic playbacks to document responses of female túngara frogs to unimodal signal components (acoustic and visual). We then tested female responses to a synchronous multimodal signal. Finally, we t...

  16. Auditory-like filterbank: An optimal speech processor for efficient ...

    Indian Academy of Sciences (India)

    The transmitter and the receiver in a communication system have to be designed optimally with respect to one another to ensure reliable and efficient communication. Following this principle, we derive an optimal filterbank for processing speech signal in the listener's auditory system (receiver), so that maximum information ...

  17. Memorable Audiovisual Narratives Synchronize Sensory and Supramodal Neural Responses

    Science.gov (United States)

    2016-01-01

    Abstract Our brains integrate information across sensory modalities to generate perceptual experiences and form memories. However, it is difficult to determine the conditions under which multisensory stimulation will benefit or hinder the retrieval of everyday experiences. We hypothesized that the determining factor is the reliability of information processing during stimulus presentation, which can be measured through intersubject correlation of stimulus-evoked activity. We therefore presented biographical auditory narratives and visual animations to 72 human subjects visually, auditorily, or combined, while neural activity was recorded using electroencephalography. Memory for the narrated information, contained in the auditory stream, was tested 3 weeks later. While the visual stimulus alone led to no meaningful retrieval, this related stimulus improved memory when it was combined with the story, even when it was temporally incongruent with the audio. Further, individuals with better subsequent memory elicited neural responses during encoding that were more correlated with their peers. Surprisingly, portions of this predictive synchronized activity were present regardless of the sensory modality of the stimulus. These data suggest that the strength of sensory and supramodal activity is predictive of memory performance after 3 weeks, and that neural synchrony may explain the mnemonic benefit of the functionally uninformative visual context observed for these real-world stimuli. PMID:27844062

  18. A novel joint timing/frequency synchronization scheme based on Radon-Wigner transform of LFM signals in CO-OFDM systems

    Science.gov (United States)

    Liu, Jianfei; Wei, Ying; Zeng, Xiangye; Lu, Jia; Zhang, Shuangxi; Wang, Mengjun

    2018-03-01

    A joint timing and frequency synchronization method has been proposed for coherent optical orthogonal frequency-division multiplexing (CO-OFDM) system in this paper. The timing offset (TO), integer frequency offset (FO) and the fractional FO can be realized by only one training symbol, which consists of two linear frequency modulation (LFM) signals with opposite chirp rates. By detecting the peak of LFM signals after Radon-Wigner transform (RWT), the TO and the integer FO can be estimated at the same time, moreover, the fractional FO can be acquired correspondingly through the self-correlation characteristic of the same training symbol. Simulation results show that the proposed method can give a more accurate TO estimation than the existing methods, especially at poor OSNR conditions; for the FO estimation, both the fractional and the integer FO can be estimated through the proposed training symbol with no extra overhead, a more accurate estimation and a large FO estimation range of [ - 5 GHz, 5GHz] can be acquired.

  19. Dynamic eccentricity fault diagnosis in round rotor synchronous motors

    International Nuclear Information System (INIS)

    Ebrahimi, Bashir Mahdi; Etemadrezaei, Mohammad; Faiz, Jawad

    2011-01-01

    Research highlights: → We have presented a novel approach to detect dynamic eccentricity in round rotor synchronous motors. → We have introduced an efficient index based on processing torque using time series data mining method. → The stator current spectrum of the motor under different levels of fault and load are computed. → Winding function method has been employed to model healthy and faulty synchronous motors. -- Abstract: In this paper, a novel approach is presented to detect dynamic eccentricity in round rotor synchronous motors. For this, an efficient index is introduced based on processing developed torque using time series data mining (TSDM) method. This index can be utilized to diagnose eccentricity fault and its degree. The capability of this index to predict dynamic eccentricity is illustrated by investigation of load variation impacts on the nominated index. Stator current spectrum of the faulty synchronous motor under different loads and dynamic eccentricity degrees are computed. Effects of the dynamic eccentricity and load variation simultaneously are scrutinized on the magnitude of 17th and 19th harmonic components as traditional indices for eccentricity fault diagnosis in synchronous motors. Necessity signals and parameters for processing and feature extraction are evaluated by winding function method which is employed to model healthy and faulty synchronous motors.

  20. The effect of auditory memory load on intensity resolution in individuals with Parkinson's disease

    Science.gov (United States)

    Richardson, Kelly C.

    Purpose: The purpose of the current study was to investigate the effect of auditory memory load on intensity resolution in individuals with Parkinson's disease (PD) as compared to two groups of listeners without PD. Methods: Nineteen individuals with Parkinson's disease, ten healthy age- and hearing-matched adults, and ten healthy young adults were studied. All listeners participated in two intensity discrimination tasks differing in auditory memory load; a lower memory load, 4IAX task and a higher memory load, ABX task. Intensity discrimination performance was assessed using a bias-free measurement of signal detectability known as d' (d-prime). Listeners further participated in a continuous loudness scaling task where they were instructed to rate the loudness level of each signal intensity using a computerized 150mm visual analogue scale. Results: Group discrimination functions indicated significantly lower intensity discrimination sensitivity (d') across tasks for the individuals with PD, as compared to the older and younger controls. No significant effect of aging on intensity discrimination was observed for either task. All three listeners groups demonstrated significantly lower intensity discrimination sensitivity for the higher auditory memory load, ABX task, compared to the lower auditory memory load, 4IAX task. Furthermore, a significant effect of aging was identified for the loudness scaling condition. The younger controls were found to rate most stimuli along the continuum as significantly louder than the older controls and the individuals with PD. Conclusions: The persons with PD showed evidence of impaired auditory perception for intensity information, as compared to the older and younger controls. The significant effect of aging on loudness perception may indicate peripheral and/or central auditory involvement.

  1. Popular song and lyrics synchronization and its application to music information retrieval

    Science.gov (United States)

    Chen, Kai; Gao, Sheng; Zhu, Yongwei; Sun, Qibin

    2006-01-01

    An automatic synchronization system of the popular song and its lyrics is presented in the paper. The system includes two main components: a) automatically detecting vocal/non-vocal in the audio signal and b) automatically aligning the acoustic signal of the song with its lyric using speech recognition techniques and positioning the boundaries of the lyrics in its acoustic realization at the multiple levels simultaneously (e.g. the word / syllable level and phrase level). The GMM models and a set of HMM-based acoustic model units are carefully designed and trained for the detection and alignment. To eliminate the severe mismatch due to the diversity of musical signal and sparse training data available, the unsupervised adaptation technique such as maximum likelihood linear regression (MLLR) is exploited for tailoring the models to the real environment, which improves robustness of the synchronization system. To further reduce the effect of the missed non-vocal music on alignment, a novel grammar net is build to direct the alignment. As we know, this is the first automatic synchronization system only based on the low-level acoustic feature such as MFCC. We evaluate the system on a Chinese song dataset collecting from 3 popular singers. We obtain 76.1% for the boundary accuracy at the syllable level (BAS) and 81.5% for the boundary accuracy at the phrase level (BAP) using fully automatic vocal/non-vocal detection and alignment. The synchronization system has many applications such as multi-modality (audio and textual) content-based popular song browsing and retrieval. Through the study, we would like to open up the discussion of some challenging problems when developing a robust synchronization system for largescale database.

  2. Auditory Spatial Layout

    Science.gov (United States)

    Wightman, Frederic L.; Jenison, Rick

    1995-01-01

    All auditory sensory information is packaged in a pair of acoustical pressure waveforms, one at each ear. While there is obvious structure in these waveforms, that structure (temporal and spectral patterns) bears no simple relationship to the structure of the environmental objects that produced them. The properties of auditory objects and their layout in space must be derived completely from higher level processing of the peripheral input. This chapter begins with a discussion of the peculiarities of acoustical stimuli and how they are received by the human auditory system. A distinction is made between the ambient sound field and the effective stimulus to differentiate the perceptual distinctions among various simple classes of sound sources (ambient field) from the known perceptual consequences of the linear transformations of the sound wave from source to receiver (effective stimulus). Next, the definition of an auditory object is dealt with, specifically the question of how the various components of a sound stream become segregated into distinct auditory objects. The remainder of the chapter focuses on issues related to the spatial layout of auditory objects, both stationary and moving.

  3. On the equivalence of electromagnetic and clock-transport synchronization in noninertial frames and gravitational fields

    International Nuclear Information System (INIS)

    Rumpf, H.

    1984-01-01

    Synchronization by slow clock transport is shown to be equivalent so that by electromagnetic signals for clocks moving along the trajectories of a timelike Killing vector field, provided the gravitational redshift is corrected for and the synchronization paths are the same. (Author)

  4. Multichannel auditory search: toward understanding control processes in polychotic auditory listening.

    Science.gov (United States)

    Lee, M D

    2001-01-01

    Two experiments are presented that serve as a framework for exploring auditory information processing. The framework is referred to as polychotic listening or auditory search, and it requires a listener to scan multiple simultaneous auditory streams for the appearance of a target word (the name of a letter such as A or M). Participants' ability to scan between two and six simultaneous auditory streams of letter and digit names for the name of a target letter was examined using six loudspeakers. The main independent variable was auditory load, or the number of active audio streams on a given trial. The primary dependent variables were target localization accuracy and reaction time. Results showed that as load increased, performance decreased. The performance decrease was evident in reaction time, accuracy, and sensitivity measures. The second study required participants to practice the same task for 10 sessions, for a total of 1800 trials. Results indicated that even with extensive practice, performance was still affected by auditory load. The present results are compared with findings in the visual search literature. The implications for the use of multiple auditory displays are discussed. Potential applications include cockpit and automobile warning displays, virtual reality systems, and training systems.

  5. Trait aspects of auditory mismatch negativity predict response to auditory training in individuals with early illness schizophrenia.

    Science.gov (United States)

    Biagianti, Bruno; Roach, Brian J; Fisher, Melissa; Loewy, Rachel; Ford, Judith M; Vinogradov, Sophia; Mathalon, Daniel H

    2017-01-01

    Individuals with schizophrenia have heterogeneous impairments of the auditory processing system that likely mediate differences in the cognitive gains induced by auditory training (AT). Mismatch negativity (MMN) is an event-related potential component reflecting auditory echoic memory, and its amplitude reduction in schizophrenia has been linked to cognitive deficits. Therefore, MMN may predict response to AT and identify individuals with schizophrenia who have the most to gain from AT. Furthermore, to the extent that AT strengthens auditory deviance processing, MMN may also serve as a readout of the underlying changes in the auditory system induced by AT. Fifty-six individuals early in the course of a schizophrenia-spectrum illness (ESZ) were randomly assigned to 40 h of AT or Computer Games (CG). Cognitive assessments and EEG recordings during a multi-deviant MMN paradigm were obtained before and after AT and CG. Changes in these measures were compared between the treatment groups. Baseline and trait-like MMN data were evaluated as predictors of treatment response. MMN data collected with the same paradigm from a sample of Healthy Controls (HC; n = 105) were compared to baseline MMN data from the ESZ group. Compared to HC, ESZ individuals showed significant MMN reductions at baseline ( p = .003). Reduced Double-Deviant MMN was associated with greater general cognitive impairment in ESZ individuals ( p = .020). Neither ESZ intervention group showed significant change in MMN. We found high correlations in all MMN deviant types (rs = .59-.68, all ps < .001) between baseline and post-intervention amplitudes irrespective of treatment group, suggesting trait-like stability of the MMN signal. Greater deficits in trait-like Double-Deviant MMN predicted greater cognitive improvements in the AT group ( p = .02), but not in the CG group. In this sample of ESZ individuals, AT had no effect on auditory deviance processing as assessed by MMN. In ESZ individuals, baseline MMN

  6. Prestimulus subsequent memory effects for auditory and visual events.

    Science.gov (United States)

    Otten, Leun J; Quayle, Angela H; Puvaneswaran, Bhamini

    2010-06-01

    It has been assumed that the effective encoding of information into memory primarily depends on neural activity elicited when an event is initially encountered. Recently, it has been shown that memory formation also relies on neural activity just before an event. The precise role of such activity in memory is currently unknown. Here, we address whether prestimulus activity affects the encoding of auditory and visual events, is set up on a trial-by-trial basis, and varies as a function of the type of recognition judgment an item later receives. Electrical brain activity was recorded from the scalps of 24 healthy young adults while they made semantic judgments on randomly intermixed series of visual and auditory words. Each word was preceded by a cue signaling the modality of the upcoming word. Auditory words were preceded by auditory cues and visual words by visual cues. A recognition memory test with remember/know judgments followed after a delay of about 45 min. As observed previously, a negative-going, frontally distributed modulation just before visual word onset predicted later recollection of the word. Crucially, the same effect was found for auditory words and observed on stay as well as switch trials. These findings emphasize the flexibility and general role of prestimulus activity in memory formation, and support a functional interpretation of the activity in terms of semantic preparation. At least with an unpredictable trial sequence, the activity is set up anew on each trial.

  7. Electrophysiological correlates of predictive coding of auditory location in the perception of natural audiovisual events.

    Science.gov (United States)

    Stekelenburg, Jeroen J; Vroomen, Jean

    2012-01-01

    In many natural audiovisual events (e.g., a clap of the two hands), the visual signal precedes the sound and thus allows observers to predict when, where, and which sound will occur. Previous studies have reported that there are distinct neural correlates of temporal (when) versus phonetic/semantic (which) content on audiovisual integration. Here we examined the effect of visual prediction of auditory location (where) in audiovisual biological motion stimuli by varying the spatial congruency between the auditory and visual parts. Visual stimuli were presented centrally, whereas auditory stimuli were presented either centrally or at 90° azimuth. Typical sub-additive amplitude reductions (AV - V audiovisual interaction was also found at 40-60 ms (P50) in the spatially congruent condition, while no effect of congruency was found on the suppression of the P2. This indicates that visual prediction of auditory location can be coded very early in auditory processing.

  8. Phenomena of synchronized response in biosystems and the possible mechanism.

    Science.gov (United States)

    Xu, Jingjing; Yang, Fan; Han, Danhong; Xu, Shengyong

    2018-02-05

    Phenomena of synchronized response is common among organs, tissues and cells in biosystems. We have analyzed and discussed three examples of synchronization in biosystems, including the direction-changing movement of paramecia, the prey behavior of flytraps, and the simultaneous discharge of electric eels. These phenomena and discussions support an electrical communication mechanism that in biosystems, the electrical signals are mainly soliton-like electromagnetic pulses, which are generated by the transient transmembrane ionic current through the ion channels and propagate along the dielectric membrane-based softmaterial waveguide network to complete synchronized responses. This transmission model implies that a uniform electrical communication mechanism might have been naturally developed in biosystem. Copyright © 2018 Elsevier Inc. All rights reserved.

  9. High-sweeping-speed optically synchronized dual-channel terahertz-signal generator for driving a superconducting tunneling mixer and its application to active gas sensing.

    Science.gov (United States)

    Oh, Kyoung-Hwan; Shimizu, Naofumi; Kohjiro, Satoshi; Kikuchi, Ken'ichi; Wakatsuki, Atsushi; Kukutsu, Naoya; Kado, Yuichi

    2009-10-12

    We propose a high-sweeping-speed optically synchronized dual-channel terahertz (THz) signal generator for an active gas-sensing system with a superconductor-insulator-superconductor (SIS) mixer. The generator can sweep a frequency range from 200 to 500 GHz at a speed of 375 GHz/s and a frequency resolution of 500 MHz. With the developed gas-sensing system, a gas-absorption-line measurement was successfully carried out with N(2)O gas in that frequency range.

  10. Measuring Auditory Selective Attention using Frequency Tagging

    Directory of Open Access Journals (Sweden)

    Hari M Bharadwaj

    2014-02-01

    Full Text Available Frequency tagging of sensory inputs (presenting stimuli that fluctuate periodically at rates to which the cortex can phase lock has been used to study attentional modulation of neural responses to inputs in different sensory modalities. For visual inputs, the visual steady-state response (VSSR at the frequency modulating an attended object is enhanced, while the VSSR to a distracting object is suppressed. In contrast, the effect of attention on the auditory steady-state response (ASSR is inconsistent across studies. However, most auditory studies analyzed results at the sensor level or used only a small number of equivalent current dipoles to fit cortical responses. In addition, most studies of auditory spatial attention used dichotic stimuli (independent signals at the ears rather than more natural, binaural stimuli. Here, we asked whether these methodological choices help explain discrepant results. Listeners attended to one of two competing speech streams, one simulated from the left and one from the right, that were modulated at different frequencies. Using distributed source modeling of magnetoencephalography results, we estimate how spatially directed attention modulates the ASSR in neural regions across the whole brain. Attention enhances the ASSR power at the frequency of the attended stream in the contralateral auditory cortex. The attended-stream modulation frequency also drives phase-locked responses in the left (but not right precentral sulcus (lPCS, a region implicated in control of eye gaze and visual spatial attention. Importantly, this region shows no phase locking to the distracting stream suggesting that the lPCS in engaged in an attention-specific manner. Modeling results that take account of the geometry and phases of the cortical sources phase locked to the two streams (including hemispheric asymmetry of lPCS activity help partly explain why past ASSR studies of auditory spatial attention yield seemingly contradictory

  11. Diminished auditory sensory gating during active auditory verbal hallucinations.

    Science.gov (United States)

    Thoma, Robert J; Meier, Andrew; Houck, Jon; Clark, Vincent P; Lewine, Jeffrey D; Turner, Jessica; Calhoun, Vince; Stephen, Julia

    2017-10-01

    Auditory sensory gating, assessed in a paired-click paradigm, indicates the extent to which incoming stimuli are filtered, or "gated", in auditory cortex. Gating is typically computed as the ratio of the peak amplitude of the event related potential (ERP) to a second click (S2) divided by the peak amplitude of the ERP to a first click (S1). Higher gating ratios are purportedly indicative of incomplete suppression of S2 and considered to represent sensory processing dysfunction. In schizophrenia, hallucination severity is positively correlated with gating ratios, and it was hypothesized that a failure of sensory control processes early in auditory sensation (gating) may represent a larger system failure within the auditory data stream; resulting in auditory verbal hallucinations (AVH). EEG data were collected while patients (N=12) with treatment-resistant AVH pressed a button to indicate the beginning (AVH-on) and end (AVH-off) of each AVH during a paired click protocol. For each participant, separate gating ratios were computed for the P50, N100, and P200 components for each of the AVH-off and AVH-on states. AVH trait severity was assessed using the Psychotic Symptoms Rating Scales AVH Total score (PSYRATS). The results of a mixed model ANOVA revealed an overall effect for AVH state, such that gating ratios were significantly higher during the AVH-on state than during AVH-off for all three components. PSYRATS score was significantly and negatively correlated with N100 gating ratio only in the AVH-off state. These findings link onset of AVH with a failure of an empirically-defined auditory inhibition system, auditory sensory gating, and pave the way for a sensory gating model of AVH. Copyright © 2017 Elsevier B.V. All rights reserved.

  12. Implementation of IEEE-1588 timing and synchronization for ATCA control and data acquisition systems

    International Nuclear Information System (INIS)

    Correia, Miguel; Sousa, Jorge; Combo, Álvaro; Rodrigues, António P.; Carvalho, Bernardo B.; Batista, António J.N.; Gonçalves, Bruno; Correia, Carlos M.B.A.; Varandas, Carlos A.F.

    2012-01-01

    Highlights: ► IEEE-1588 over Ethernet protocol is implemented for the synchronization of all clock signals of an ATCA AMC carrier module. ► The ATCA hardware consists of an AMC quad-carrier main-board with PCI Express switching. ► IEEE-1588 is to be implemented on a Virtex-6 FPGA. ► Timing signals on the ATX-AMC4-PTP are managed and routed by a crosspoint-switch implemented on a Virtex-6 FPGA. ► Each clock signal source may be independently located (on each of the AMC cards, RTM or ATCA backplane). - Abstract: Control and data acquisition (C and DA) systems for Fusion experiments are required to provide accurate timing and synchronization (T and S) signals to all of its components. IPFN adopted PICMG's Advanced Telecommunications Computing Architecture (ATCA) industry standard to develop C and DA instrumentation. ATCA was chosen not only for its high throughput characteristics but also for its high availability (HA) features which become of greater importance in steady-state operation scenarios. However, the specified ATCA clock and synchronization interface may be too limited for the timing and synchronization needs in advanced Physics experiments. Upcoming specification extensions, developed by the “xTCA for Physics” workgroups, will contemplate, among others, a complementary timing specification, developed by the PICMG xTCA for Physics IO, Timing and Synchronization Technical Committee. The IEEE-1588 Precision Time Protocol (PTP) over Ethernet is one of the protocols, proposed by the Committee, aiming for precise synchronization of clocks in measurement and control systems, based on low jitter and slave-to-slave skew criteria. The paper presents an implementation of IEEE-1588 over Ethernet, in an ATCA hardware platform. The ATCA hardware consists of an Advanced Mezzanine Card (AMC) quad-carrier front board with PCI Express switching. IEEE-1588 is to be implemented on a Virtex-6 FPGA. Ethernet connectivity with the remote master clock is located on

  13. Auditory and Visual Differences in Time Perception? An Investigation from a Developmental Perspective with Neuropsychological Tests

    Science.gov (United States)

    Zelanti, Pierre S.; Droit-Volet, Sylvie

    2012-01-01

    Adults and children (5- and 8-year-olds) performed a temporal bisection task with either auditory or visual signals and either a short (0.5-1.0s) or long (4.0-8.0s) duration range. Their working memory and attentional capacities were assessed by a series of neuropsychological tests administered in both the auditory and visual modalities. Results…

  14. Effectiveness of auditory and tactile crossmodal cues in a dual-task visual and auditory scenario.

    Science.gov (United States)

    Hopkins, Kevin; Kass, Steven J; Blalock, Lisa Durrance; Brill, J Christopher

    2017-05-01

    In this study, we examined how spatially informative auditory and tactile cues affected participants' performance on a visual search task while they simultaneously performed a secondary auditory task. Visual search task performance was assessed via reaction time and accuracy. Tactile and auditory cues provided the approximate location of the visual target within the search display. The inclusion of tactile and auditory cues improved performance in comparison to the no-cue baseline conditions. In comparison to the no-cue conditions, both tactile and auditory cues resulted in faster response times in the visual search only (single task) and visual-auditory (dual-task) conditions. However, the effectiveness of auditory and tactile cueing for visual task accuracy was shown to be dependent on task-type condition. Crossmodal cueing remains a viable strategy for improving task performance without increasing attentional load within a singular sensory modality. Practitioner Summary: Crossmodal cueing with dual-task performance has not been widely explored, yet has practical applications. We examined the effects of auditory and tactile crossmodal cues on visual search performance, with and without a secondary auditory task. Tactile cues aided visual search accuracy when also engaged in a secondary auditory task, whereas auditory cues did not.

  15. Activation of auditory white matter tracts as revealed by functional magnetic resonance imaging

    Energy Technology Data Exchange (ETDEWEB)

    Tae, Woo Suk [Kangwon National University, Neuroscience Research Institute, School of Medicine, Chuncheon (Korea, Republic of); Yakunina, Natalia; Nam, Eui-Cheol [Kangwon National University, Neuroscience Research Institute, School of Medicine, Chuncheon (Korea, Republic of); Kangwon National University, Department of Otolaryngology, School of Medicine, Chuncheon, Kangwon-do (Korea, Republic of); Kim, Tae Su [Kangwon National University Hospital, Department of Otolaryngology, Chuncheon (Korea, Republic of); Kim, Sam Soo [Kangwon National University, Neuroscience Research Institute, School of Medicine, Chuncheon (Korea, Republic of); Kangwon National University, Department of Radiology, School of Medicine, Chuncheon (Korea, Republic of)

    2014-07-15

    The ability of functional magnetic resonance imaging (fMRI) to detect activation in brain white matter (WM) is controversial. In particular, studies on the functional activation of WM tracts in the central auditory system are scarce. We utilized fMRI to assess and characterize the entire auditory WM pathway under robust experimental conditions involving the acquisition of a large number of functional volumes, the application of broadband auditory stimuli of high intensity, and the use of sparse temporal sampling to avoid scanner noise effects and increase signal-to-noise ratio. Nineteen healthy volunteers were subjected to broadband white noise in a block paradigm; each run had four sound-on/off alternations and was repeated nine times for each subject. Sparse sampling (TR = 8 s) was used. In addition to traditional gray matter (GM) auditory center activation, WM activation was detected in the isthmus and midbody of the corpus callosum (CC), tapetum, auditory radiation, lateral lemniscus, and decussation of the superior cerebellar peduncles. At the individual level, 13 of 19 subjects (68 %) had CC activation. Callosal WM exhibited a temporal delay of approximately 8 s in response to the stimulation compared with GM. These findings suggest that direct evaluation of the entire functional network of the central auditory system may be possible using fMRI, which may aid in understanding the neurophysiological basis of the central auditory system and in developing treatment strategies for various central auditory disorders. (orig.)

  16. Activation of auditory white matter tracts as revealed by functional magnetic resonance imaging

    International Nuclear Information System (INIS)

    Tae, Woo Suk; Yakunina, Natalia; Nam, Eui-Cheol; Kim, Tae Su; Kim, Sam Soo

    2014-01-01

    The ability of functional magnetic resonance imaging (fMRI) to detect activation in brain white matter (WM) is controversial. In particular, studies on the functional activation of WM tracts in the central auditory system are scarce. We utilized fMRI to assess and characterize the entire auditory WM pathway under robust experimental conditions involving the acquisition of a large number of functional volumes, the application of broadband auditory stimuli of high intensity, and the use of sparse temporal sampling to avoid scanner noise effects and increase signal-to-noise ratio. Nineteen healthy volunteers were subjected to broadband white noise in a block paradigm; each run had four sound-on/off alternations and was repeated nine times for each subject. Sparse sampling (TR = 8 s) was used. In addition to traditional gray matter (GM) auditory center activation, WM activation was detected in the isthmus and midbody of the corpus callosum (CC), tapetum, auditory radiation, lateral lemniscus, and decussation of the superior cerebellar peduncles. At the individual level, 13 of 19 subjects (68 %) had CC activation. Callosal WM exhibited a temporal delay of approximately 8 s in response to the stimulation compared with GM. These findings suggest that direct evaluation of the entire functional network of the central auditory system may be possible using fMRI, which may aid in understanding the neurophysiological basis of the central auditory system and in developing treatment strategies for various central auditory disorders. (orig.)

  17. Auditory processing efficiency deficits in children with developmental language impairments

    Science.gov (United States)

    Hartley, Douglas E. H.; Moore, David R.

    2002-12-01

    The ``temporal processing hypothesis'' suggests that individuals with specific language impairments (SLIs) and dyslexia have severe deficits in processing rapidly presented or brief sensory information, both within the auditory and visual domains. This hypothesis has been supported through evidence that language-impaired individuals have excess auditory backward masking. This paper presents an analysis of masking results from several studies in terms of a model of temporal resolution. Results from this modeling suggest that the masking results can be better explained by an ``auditory efficiency'' hypothesis. If impaired or immature listeners have a normal temporal window, but require a higher signal-to-noise level (poor processing efficiency), this hypothesis predicts the observed small deficits in the simultaneous masking task, and the much larger deficits in backward and forward masking tasks amongst those listeners. The difference in performance on these masking tasks is predictable from the compressive nonlinearity of the basilar membrane. The model also correctly predicts that backward masking (i) is more prone to training effects, (ii) has greater inter- and intrasubject variability, and (iii) increases less with masker level than do other masking tasks. These findings provide a new perspective on the mechanisms underlying communication disorders and auditory masking.

  18. Speech processing: from peripheral to hemispheric asymmetry of the auditory system.

    Science.gov (United States)

    Lazard, Diane S; Collette, Jean-Louis; Perrot, Xavier

    2012-01-01

    Language processing from the cochlea to auditory association cortices shows side-dependent specificities with an apparent left hemispheric dominance. The aim of this article was to propose to nonspeech specialists a didactic review of two complementary theories about hemispheric asymmetry in speech processing. Starting from anatomico-physiological and clinical observations of auditory asymmetry and interhemispheric connections, this review then exposes behavioral (dichotic listening paradigm) as well as functional (functional magnetic resonance imaging and positron emission tomography) experiments that assessed hemispheric specialization for speech processing. Even though speech at an early phonological level is regarded as being processed bilaterally, a left-hemispheric dominance exists for higher-level processing. This asymmetry may arise from a segregation of the speech signal, broken apart within nonprimary auditory areas in two distinct temporal integration windows--a fast one on the left and a slower one on the right--modeled through the asymmetric sampling in time theory or a spectro-temporal trade-off, with a higher temporal resolution in the left hemisphere and a higher spectral resolution in the right hemisphere, modeled through the spectral/temporal resolution trade-off theory. Both theories deal with the concept that lower-order tuning principles for acoustic signal might drive higher-order organization for speech processing. However, the precise nature, mechanisms, and origin of speech processing asymmetry are still being debated. Finally, an example of hemispheric asymmetry alteration, which has direct clinical implications, is given through the case of auditory aging that mixes peripheral disorder and modifications of central processing. Copyright © 2011 The American Laryngological, Rhinological, and Otological Society, Inc.

  19. A 100 MHz synchronized OEIC photoreceiver in n-well, CMOS technology

    DEFF Research Database (Denmark)

    Kamel, Ayadi; Danielsen, Per Lander

    1998-01-01

    We analyze and demonstrate a synchronized CMOS photoreceiver for the conversion of optical inputs of pulse-light to electronic digital signals. Small-signal and photonic analysis of the proposed circuit are detailed. The photoreceiver was operated at 100 MHz with only 13.3 fJ/pulse of 830-nm inpu...

  20. Fit for the frontline? A focus group exploration of auditory tasks carried out by infantry and combat support personnel

    OpenAIRE

    Bevis, Zoe L.; Semeraro, Hannah; van Besouw, R.M.; Rowan, D.; Lineton, B.; Allsopp, Adrian J.

    2014-01-01

    In order to preserve their operational effectiveness and ultimately their survival, military personnel must be able to detect important acoustic signals and maintain situational awareness. The possession of sufficient hearing ability to perform job-specific auditory tasks is defined as auditory fitness for duty (AFFD). Pure tone audiometry (PTA) is used to assess AFFD in the UK military; however, it is unclear whether PTA is able to accurately predict performance on job-specific auditory task...

  1. High Precision Clock Bias Prediction Model in Clock Synchronization System

    Directory of Open Access Journals (Sweden)

    Zan Liu

    2016-01-01

    Full Text Available Time synchronization is a fundamental requirement for many services provided by a distributed system. Clock calibration through the time signal is the usual way to realize the synchronization among the clocks used in the distributed system. The interference to time signal transmission or equipment failures may bring about failure to synchronize the time. To solve this problem, a clock bias prediction module is paralleled in the clock calibration system. And for improving the precision of clock bias prediction, the first-order grey model with one variable (GM(1,1 model is proposed. In the traditional GM(1,1 model, the combination of parameters determined by least squares criterion is not optimal; therefore, the particle swarm optimization (PSO is used to optimize GM(1,1 model. At the same time, in order to avoid PSO getting stuck at local optimization and improve its efficiency, the mechanisms that double subgroups and nonlinear decreasing inertia weight are proposed. In order to test the precision of the improved model, we design clock calibration experiments, where time signal is transferred via radio and wired channel, respectively. The improved model is built on the basis of clock bias acquired in the experiments. The results show that the improved model is superior to other models both in precision and in stability. The precision of improved model increased by 66.4%~76.7%.

  2. Synchronization of metronomes

    Science.gov (United States)

    Pantaleone, James

    2002-10-01

    Synchronization is a common phenomenon in physical and biological systems. We examine the synchronization of two (and more) metronomes placed on a freely moving base. The small motion of the base couples the pendulums causing synchronization. The synchronization is generally in-phase, with antiphase synchronization occurring only under special conditions. The metronome system provides a mechanical realization of the popular Kuramoto model for synchronization of biological oscillators, and is excellent for classroom demonstrations and an undergraduate physics lab.

  3. Synchronous phase and energy measurement system for a 6.7-MeV H- beam

    International Nuclear Information System (INIS)

    Gilpatrick, J.D.; Meyer, R.E.; Wells, F.D.; Power, J.F.; Shafer, R.E.

    1988-01-01

    A noninterceptive measurement system has been built to measure the energy and synchronous phase of a 6.7-MeV proton beam drifting from the ramped-gradient, drift-tube linac (RGDTL) in the accelerator test stand (ATS) facility. Axially-symmetric, capacitive probes used in these measurements produce signals that are proportional to the beam image current on their inner rings. Signals from two of these probes separated by 92.6 cm are down-converted from 425 to 20 MHz. The phase difference between these 20-MHz signals is then detected with an electronic, phase-comparator circuit. The phase-comparator signal output is a voltage that is related to momentum of the beam. A phase comparison is also provided between the 425-MHz fundamental rf field inside the RGDTL and the capacitive probe located nearest the RGDTL output. The total estimated error for the absolute and relative energy measurement is less than +- 12.2 and +- 3.1 keV, respectively. The total estimated error for the relative synchronous phase measurement is less than +-1/degree/. Beam energy versus synchronous phase experimental data agree with computer simulations. 3 refs., 3 figs., 1 tab

  4. Synchronization of uncertain chaotic systems using a single transmission channel

    International Nuclear Information System (INIS)

    Feng Yong; Yu Xinghuo; Sun Lixia

    2008-01-01

    This paper proposes a robust sliding mode observer for synchronization of uncertain chaotic systems with multi-nonlinearities. A new control strategy is proposed for the construction of the robust sliding mode observer, which can avoid the strict conditions in the design process of Walcott-Zak observer. A new method of multi-dimensional signal transmission via single transmission channel is proposed and applied to chaos synchronization of uncertain chaotic systems with multi-nonlinearities. The simulation results are presented to validate the method

  5. Analysis, Synthesis, and Classification of Nonlinear Systems Using Synchronized Swept-Sine Method for Audio Effects

    Directory of Open Access Journals (Sweden)

    Novak Antonin

    2010-01-01

    Full Text Available A new method of identification, based on an input synchronized exponential swept-sine signal, is used to analyze and synthesize nonlinear audio systems like overdrive pedals for guitar. Two different pedals are studied; the first one exhibiting a strong influence of the input signal level on its input/output law and the second one exhibiting a weak influence of this input signal level. The Synchronized Swept Sine method leads to a Generalized Polynomial Hammerstein model equivalent to the pedals under test. The behaviors of both pedals are illustrated through model-based resynthesized signals. Moreover, it is also shown that this method leads to a criterion allowing the classification of the nonlinear systems under test, according to the influence of the input signal levels on their input/output law.

  6. Objective measures of binaural masking level differences and comodulation masking release based on late auditory evoked potentials

    DEFF Research Database (Denmark)

    Epp, Bastian; Yasin, Ifat; Verhey, Jesko L.

    2013-01-01

    at a fixed physical intensity is varied by introducing auditory cues of (i) interaural target signal phase disparity and (ii) coherent masker level fluctuations in different frequency regions. In agreement with previous studies, psychoacoustical experiments showed that both stimulus manipulations result......The audibility of important sounds is often hampered due to the presence of other masking sounds. The present study investigates if a correlate of the audibility of a tone masked by noise is found in late auditory evoked potentials measured from human listeners. The audibility of the target sound...... in a masking release (i: binaural masking level difference; ii: comodulation masking release) compared to a condition where those cues are not present. Late auditory evoked potentials (N1, P2) were recorded for the stimuli at a constant masker level, but different signal levels within the same set of listeners...

  7. The Neurophysiology of Auditory Hallucinations – A Historic and Contemporary Review

    Directory of Open Access Journals (Sweden)

    Remko evan Lutterveld

    2011-05-01

    Full Text Available Electroencephalography (EEG and magnetoencephalography (MEG are two techniques that distinguish themselves from other neuroimaging methodologies through their ability to directly measure brain-related activity and their high temporal resolution. A large body of research has applied these techniques to study auditory hallucinations. Across a variety of approaches, the left superior temporal cortex is consistently reported to be involved in this symptom. Moreover, there is increasing evidence that a failure in corollary discharge, i.e. a neural signal originating in frontal speech areas that indicates to sensory areas that forthcoming thought is self-generated, may underlie the experience of auditory hallucinations

  8. Attending to auditory memory.

    Science.gov (United States)

    Zimmermann, Jacqueline F; Moscovitch, Morris; Alain, Claude

    2016-06-01

    Attention to memory describes the process of attending to memory traces when the object is no longer present. It has been studied primarily for representations of visual stimuli with only few studies examining attention to sound object representations in short-term memory. Here, we review the interplay of attention and auditory memory with an emphasis on 1) attending to auditory memory in the absence of related external stimuli (i.e., reflective attention) and 2) effects of existing memory on guiding attention. Attention to auditory memory is discussed in the context of change deafness, and we argue that failures to detect changes in our auditory environments are most likely the result of a faulty comparison system of incoming and stored information. Also, objects are the primary building blocks of auditory attention, but attention can also be directed to individual features (e.g., pitch). We review short-term and long-term memory guided modulation of attention based on characteristic features, location, and/or semantic properties of auditory objects, and propose that auditory attention to memory pathways emerge after sensory memory. A neural model for auditory attention to memory is developed, which comprises two separate pathways in the parietal cortex, one involved in attention to higher-order features and the other involved in attention to sensory information. This article is part of a Special Issue entitled SI: Auditory working memory. Copyright © 2015 Elsevier B.V. All rights reserved.

  9. Out-of-synchrony speech entrainment in developmental dyslexia.

    Science.gov (United States)

    Molinaro, Nicola; Lizarazu, Mikel; Lallier, Marie; Bourguignon, Mathieu; Carreiras, Manuel

    2016-08-01

    Developmental dyslexia is a reading disorder often characterized by reduced awareness of speech units. Whether the neural source of this phonological disorder in dyslexic readers results from the malfunctioning of the primary auditory system or damaged feedback communication between higher-order phonological regions (i.e., left inferior frontal regions) and the auditory cortex is still under dispute. Here we recorded magnetoencephalographic (MEG) signals from 20 dyslexic readers and 20 age-matched controls while they were listening to ∼10-s-long spoken sentences. Compared to controls, dyslexic readers had (1) an impaired neural entrainment to speech in the delta band (0.5-1 Hz); (2) a reduced delta synchronization in both the right auditory cortex and the left inferior frontal gyrus; and (3) an impaired feedforward functional coupling between neural oscillations in the right auditory cortex and the left inferior frontal regions. This shows that during speech listening, individuals with developmental dyslexia present reduced neural synchrony to low-frequency speech oscillations in primary auditory regions that hinders higher-order speech processing steps. The present findings, thus, strengthen proposals assuming that improper low-frequency acoustic entrainment affects speech sampling. This low speech-brain synchronization has the strong potential to cause severe consequences for both phonological and reading skills. Interestingly, the reduced speech-brain synchronization in dyslexic readers compared to normal readers (and its higher-order consequences across the speech processing network) appears preserved through the development from childhood to adulthood. Thus, the evaluation of speech-brain synchronization could possibly serve as a diagnostic tool for early detection of children at risk of dyslexia. Hum Brain Mapp 37:2767-2783, 2016. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.

  10. Bursting synchronization in scale-free networks

    International Nuclear Information System (INIS)

    Batista, C.A.S.; Batista, A.M.; Pontes, J.C.A. de; Lopes, S.R.; Viana, R.L.

    2009-01-01

    Neuronal networks in some areas of the brain cortex present the scale-free property, i.e., the neuron connectivity is distributed according to a power-law, such that neurons are more likely to couple with other already well-connected ones. Neuron activity presents two timescales, a fast one related to action-potential spiking, and a slow timescale in which bursting takes place. Some pathological conditions are related with the synchronization of the bursting activity in a weak sense, meaning the adjustment of the bursting phase due to coupling. Hence it has been proposed that an externally applied time-periodic signal be applied in order to control undesirable synchronized bursting rhythms. We investigated this kind of intervention using a two-dimensional map to describe neurons with spiking-bursting activity in a scale-free network.

  11. The role of the temporal pole in modulating primitive auditory memory.

    Science.gov (United States)

    Liu, Zhiliang; Wang, Qian; You, Yu; Yin, Peng; Ding, Hu; Bao, Xiaohan; Yang, Pengcheng; Lu, Hao; Gao, Yayue; Li, Liang

    2016-04-21

    Primitive auditory memory (PAM), which is recognized as the early point in the chain of the transient auditory memory system, faithfully maintains raw acoustic fine-structure signals for up to 20-30 milliseconds. The neural mechanisms underlying PAM have not been reported in the literature. Previous anatomical, brain-imaging, and neurophysiological studies have suggested that the temporal pole (TP), part of the parahippocampal region in the transitional area between perirhinal cortex and superior/inferior temporal gyri, is involved in auditory memories. This study investigated whether the TP plays a role in mediating/modulating PAM. The longest interaural interval (the interaural-delay threshold) for detecting a break in interaural correlation (BIC) embedded in interaurally correlated wideband noises was used to indicate the temporal preservation of PAM and examined in both healthy listeners and patients receiving unilateral anterior temporal lobectomy (ATL, centered on the TP) for treating their temporal lobe epilepsy (TLE). The results showed that patients with ATL were still able to detect the BIC even when an interaural interval was introduced, regardless of which ear was the leading one. However, in patient participants, the group-mean interaural-delay threshold for detecting the BIC under the contralateral-ear-leading (relative to the side of ATL) condition was significantly shorter than that under the ipsilateral-ear-leading condition. The results suggest that although the TP is not essential for integrating binaural signals and mediating the PAM, it plays a role in top-down modulating the PAM of raw acoustic fine-structure signals from the contralateral ear. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  12. Synchronizing spatiotemporal chaos by introducing a finite flat region in the local map

    Directory of Open Access Journals (Sweden)

    J. Y. Chen

    2001-01-01

    Full Text Available An approach to synchronize spatiotemporal chaos is proposed. It is achieved by introducing a finite flat region in the local map. By using this scheme, a number of orbits in both the drive and the response subsystems are forced to pass through a fixed point in every dimension. With only an arbitrary phase space variable as drive signal, synchronization of spatiotemporal chaos can be achieved rapidly in the response subsystem. This is an advantage when compared with other synchronization methods that require a linear combination of the original phase space variables.

  13. Predictive coding of visual-auditory and motor-auditory events: An electrophysiological study.

    Science.gov (United States)

    Stekelenburg, Jeroen J; Vroomen, Jean

    2015-11-11

    The amplitude of auditory components of the event-related potential (ERP) is attenuated when sounds are self-generated compared to externally generated sounds. This effect has been ascribed to internal forward modals predicting the sensory consequences of one's own motor actions. Auditory potentials are also attenuated when a sound is accompanied by a video of anticipatory visual motion that reliably predicts the sound. Here, we investigated whether the neural underpinnings of prediction of upcoming auditory stimuli are similar for motor-auditory (MA) and visual-auditory (VA) events using a stimulus omission paradigm. In the MA condition, a finger tap triggered the sound of a handclap whereas in the VA condition the same sound was accompanied by a video showing the handclap. In both conditions, the auditory stimulus was omitted in either 50% or 12% of the trials. These auditory omissions induced early and mid-latency ERP components (oN1 and oN2, presumably reflecting prediction and prediction error), and subsequent higher-order error evaluation processes. The oN1 and oN2 of MA and VA were alike in amplitude, topography, and neural sources despite that the origin of the prediction stems from different brain areas (motor versus visual cortex). This suggests that MA and VA predictions activate a sensory template of the sound in auditory cortex. This article is part of a Special Issue entitled SI: Prediction and Attention. Copyright © 2015 Elsevier B.V. All rights reserved.

  14. Learning sparse generative models of audiovisual signals

    OpenAIRE

    Monaci, Gianluca; Sommer, Friedrich T.; Vandergheynst, Pierre

    2008-01-01

    This paper presents a novel framework to learn sparse represen- tations for audiovisual signals. An audiovisual signal is modeled as a sparse sum of audiovisual kernels. The kernels are bimodal functions made of synchronous audio and video components that can be positioned independently and arbitrarily in space and time. We design an algorithm capable of learning sets of such audiovi- sual, synchronous, shift-invariant functions by alternatingly solving a coding and a learning pr...

  15. Synchronous atmospheric radiation correction of GF-2 satellite multispectral image

    Science.gov (United States)

    Bian, Fuqiang; Fan, Dongdong; Zhang, Yan; Wang, Dandan

    2018-02-01

    GF-2 remote sensing products have been widely used in many fields for its high-quality information, which provides technical support for the the macroeconomic decisions. Atmospheric correction is the necessary part in the data preprocessing of the quantitative high resolution remote sensing, which can eliminate the signal interference in the radiation path caused by atmospheric scattering and absorption, and reducting apparent reflectance into real reflectance of the surface targets. Aiming at the problem that current research lack of atmospheric date which are synchronization and region matching of the surface observation image, this research utilize the MODIS Level 1B synchronous data to simulate synchronized atmospheric condition, and write programs to implementation process of aerosol retrieval and atmospheric correction, then generate a lookup table of the remote sensing image based on the radioactive transfer model of 6S (second simulation of a satellite signal in the solar spectrum) to correct the atmospheric effect of multispectral image from GF-2 satellite PMS-1 payload. According to the correction results, this paper analyzes the pixel histogram of the reflectance spectrum of the 4 spectral bands of PMS-1, and evaluates the correction results of different spectral bands. Then conducted a comparison experiment on the same GF-2 image based on the QUAC. According to the different targets respectively statistics the average value of NDVI, implement a comparative study of NDVI from two different results. The degree of influence was discussed by whether to adopt synchronous atmospheric date. The study shows that the result of the synchronous atmospheric parameters have significantly improved the quantitative application of the GF-2 remote sensing data.

  16. Linear Approach for Synchronous State Stability in Fully Connected PLL Networks

    Directory of Open Access Journals (Sweden)

    José R. C. Piqueira

    2008-01-01

    Full Text Available Synchronization is an essential feature for the use of digital systems in telecommunication networks, integrated circuits, and manufacturing automation. Formerly, master-slave (MS architectures, with precise master clock generators sending signals to phase-locked loops (PLLs working as slave oscillators, were considered the best solution. Nowadays, the development of wireless networks with dynamical connectivity and the increase of the size and the operation frequency of integrated circuits suggest that the distribution of clock signals could be more efficient if distributed solutions with fully connected oscillators are used. Here, fully connected networks with second-order PLLs as nodes are considered. In previous work, how the synchronous state frequency for this type of network depends on the node parameters and delays was studied and an expression for the long-term frequency was derived (Piqueira, 2006. Here, by taking the first term of the Taylor series expansion for the dynamical system description, it is shown that for a generic network with N nodes, the synchronous state is locally asymptotically stable.

  17. Modelling and Simulation of a Synchronous Machine with Power Electronic Systems

    DEFF Research Database (Denmark)

    Chen, Zhe; Blaabjerg, Frede

    2005-01-01

    is modelled in SIMULINK as well. The resulting model can more accurately represent non-idea situations such as non-symmetrical parameters of the electrical machines and unbalance conditions. The model may be used for both steady state and large-signal dynamic analysis. This is particularly useful......This paper reports the modeling and simulation of a synchronous machine with a power electronic interface in direct phase model. The implementation of a direct phase model of synchronous machines in MATLAB/SIMULINK is presented .The power electronic system associated with the synchronous machine...... in the systems where a detailed study is needed in order to assess the overall system stability. Simulation studies are performed under various operation conditions. It is shown that the developed model could be used for studies of various applications of synchronous machines such as in renewable and DG...

  18. A note on chaotic synchronization of time-delay secure communication systems

    International Nuclear Information System (INIS)

    Li Demin; Wang Zidong; Zhou Jie; Fang Jianan; Ni Jinjin

    2008-01-01

    In a real world, the signals are often transmitted through a hostile environment, and therefore the secure communication system has attracted considerable research interests. In this paper, the observer-based chaotic synchronization problem is studied for a class of time-delay secure communication systems. The system under consideration is subject to delayed state and nonlinear disturbances. The time-delay is allowed to be time-varying, and the nonlinearities are assumed to satisfy global Lipschitz conditions. The problem addressed is the design of a synchronization scheme such that, for the admissible time-delay as well as nonlinear disturbances, the response system can globally synchronize the driving system. An effective algebraic matrix inequality approach is developed to solve the chaotic synchronization problem. A numerical example is presented to show the effectiveness and efficiency of the proposed secure communication scheme

  19. Stuttering Inhibition via Altered Auditory Feedback during Scripted Telephone Conversations

    Science.gov (United States)

    Hudock, Daniel; Kalinowski, Joseph

    2014-01-01

    Background: Overt stuttering is inhibited by approximately 80% when people who stutter read aloud as they hear an altered form of their speech feedback to them. However, levels of stuttering inhibition vary from 60% to 100% depending on speaking situation and signal presentation. For example, binaural presentations of delayed auditory feedback…

  20. Animal models for auditory streaming

    Science.gov (United States)

    Itatani, Naoya

    2017-01-01

    Sounds in the natural environment need to be assigned to acoustic sources to evaluate complex auditory scenes. Separating sources will affect the analysis of auditory features of sounds. As the benefits of assigning sounds to specific sources accrue to all species communicating acoustically, the ability for auditory scene analysis is widespread among different animals. Animal studies allow for a deeper insight into the neuronal mechanisms underlying auditory scene analysis. Here, we will review the paradigms applied in the study of auditory scene analysis and streaming of sequential sounds in animal models. We will compare the psychophysical results from the animal studies to the evidence obtained in human psychophysics of auditory streaming, i.e. in a task commonly used for measuring the capability for auditory scene analysis. Furthermore, the neuronal correlates of auditory streaming will be reviewed in different animal models and the observations of the neurons’ response measures will be related to perception. The across-species comparison will reveal whether similar demands in the analysis of acoustic scenes have resulted in similar perceptual and neuronal processing mechanisms in the wide range of species being capable of auditory scene analysis. This article is part of the themed issue ‘Auditory and visual scene analysis’. PMID:28044022

  1. Upgrade of the synchronous data management system of the EAST poloidal field power supply

    Energy Technology Data Exchange (ETDEWEB)

    Zhu, Lili; Huang, Liansheng, E-mail: huangls@ipp.ac.cn; Fu, Peng; Gao, Ge; He, Shiying

    2016-11-15

    Highlights: • The upgraded synchronous data management system of EAST poloidal field power supply supports long-pulse data storage. • Slice storage mechanism on MDSplus has been adopted for quasi real-time data storage. • The state machine has been adopted for managing the system sequencer. • IEEE-1588 protocol via Ethernet for the synchronization of clock signal was detailed described. - Abstract: Poloidal field (PF) power supply is an important subsystem of the Experimental Advanced Superconducting Tokamak (EAST). The upgrade of the PF control system of EAST is a great improvement over the original data management system which could not meet the requirements necessary for experiments on synchronization, modularity and sampling rate. In order to better analyze the power operation performance, the Synchronization Data Management System (SDMS) needs to be upgraded as well. This upgrade is based on distributed data acquisition and an MDSPLUS database. It consists of three data acquisition nodes synchronized by an reference clock from the EAST central timing system that also provides the start trigger of the EAST pulse. After being processed by a signal conditioning unit, experimental signals are digitized and written into the database in MDSPLUS format. Multi-channel, multi-tasking and continuous data storage have been achieved by using multi-threading technology on a Linux operation system. The SDMS has been used on the server in PF control system for the entire 2015 EAST campaign. The SDMS has had good performance during experiments and convenient human-machine interface to satisfy the requirements of all the experiments.

  2. Upgrade of the synchronous data management system of the EAST poloidal field power supply

    International Nuclear Information System (INIS)

    Zhu, Lili; Huang, Liansheng; Fu, Peng; Gao, Ge; He, Shiying

    2016-01-01

    Highlights: • The upgraded synchronous data management system of EAST poloidal field power supply supports long-pulse data storage. • Slice storage mechanism on MDSplus has been adopted for quasi real-time data storage. • The state machine has been adopted for managing the system sequencer. • IEEE-1588 protocol via Ethernet for the synchronization of clock signal was detailed described. - Abstract: Poloidal field (PF) power supply is an important subsystem of the Experimental Advanced Superconducting Tokamak (EAST). The upgrade of the PF control system of EAST is a great improvement over the original data management system which could not meet the requirements necessary for experiments on synchronization, modularity and sampling rate. In order to better analyze the power operation performance, the Synchronization Data Management System (SDMS) needs to be upgraded as well. This upgrade is based on distributed data acquisition and an MDSPLUS database. It consists of three data acquisition nodes synchronized by an reference clock from the EAST central timing system that also provides the start trigger of the EAST pulse. After being processed by a signal conditioning unit, experimental signals are digitized and written into the database in MDSPLUS format. Multi-channel, multi-tasking and continuous data storage have been achieved by using multi-threading technology on a Linux operation system. The SDMS has been used on the server in PF control system for the entire 2015 EAST campaign. The SDMS has had good performance during experiments and convenient human-machine interface to satisfy the requirements of all the experiments.

  3. Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus.

    Science.gov (United States)

    Venezia, Jonathan H; Vaden, Kenneth I; Rong, Feng; Maddox, Dale; Saberi, Kourosh; Hickok, Gregory

    2017-01-01

    The human superior temporal sulcus (STS) is responsive to visual and auditory information, including sounds and facial cues during speech recognition. We investigated the functional organization of STS with respect to modality-specific and multimodal speech representations. Twenty younger adult participants were instructed to perform an oddball detection task and were presented with auditory, visual, and audiovisual speech stimuli, as well as auditory and visual nonspeech control stimuli in a block fMRI design. Consistent with a hypothesized anterior-posterior processing gradient in STS, auditory, visual and audiovisual stimuli produced the largest BOLD effects in anterior, posterior and middle STS (mSTS), respectively, based on whole-brain, linear mixed effects and principal component analyses. Notably, the mSTS exhibited preferential responses to multisensory stimulation, as well as speech compared to nonspeech. Within the mid-posterior and mSTS regions, response preferences changed gradually from visual, to multisensory, to auditory moving posterior to anterior. Post hoc analysis of visual regions in the posterior STS revealed that a single subregion bordering the mSTS was insensitive to differences in low-level motion kinematics yet distinguished between visual speech and nonspeech based on multi-voxel activation patterns. These results suggest that auditory and visual speech representations are elaborated gradually within anterior and posterior processing streams, respectively, and may be integrated within the mSTS, which is sensitive to more abstract speech information within and across presentation modalities. The spatial organization of STS is consistent with processing streams that are hypothesized to synthesize perceptual speech representations from sensory signals that provide convergent information from visual and auditory modalities.

  4. Comparison of coherence and phase synchronization of the human sleep electroencephalogram

    Czech Academy of Sciences Publication Activity Database

    Mezeiová, K.; Paluš, Milan

    2012-01-01

    Roč. 123, č. 9 (2012), s. 1821-1830 ISSN 1388-2457 R&D Projects: GA MŠk 7E08027 EU Projects: European Commission(XE) 200728 - BRAINSYNC Grant - others:AV ČR - SAS(CZ-SK) Modern Methods for Analysis of Electrophysiological Signals Institutional research plan: CEZ:AV0Z10300504 Keywords : phase synchronization * complete synchronization * mean phase coherence * permutation surrogate data * coherence * human sleep EEG Subject RIV: FH - Neurology Impact factor: 3.144, year: 2012

  5. Spectral integration in binaural signal detection

    NARCIS (Netherlands)

    Breebaart, D.J.; Par, van de S.L.J.D.E.; Kohlrausch, A.G.

    1997-01-01

    For both monaural and binaural masking, the spectral content of the masker and of the signal to be detected are important stimulus properties influencing the detection process. It is generally accepted that the auditory system separates the incoming signals in several frequency bands. It is not

  6. Dynamic Synchronous Capture Algorithm for an Electromagnetic Flowmeter.

    Science.gov (United States)

    Fanjiang, Yong-Yi; Lu, Shih-Wei

    2017-04-10

    This paper proposes a dynamic synchronous capture (DSC) algorithm to calculate the flow rate for an electromagnetic flowmeter. The characteristics of the DSC algorithm can accurately calculate the flow rate signal and efficiently convert an analog signal to upgrade the execution performance of a microcontroller unit (MCU). Furthermore, it can reduce interference from abnormal noise. It is extremely steady and independent of fluctuations in the flow measurement. Moreover, it can calculate the current flow rate signal immediately (m/s). The DSC algorithm can be applied to the current general MCU firmware platform without using DSP (Digital Signal Processing) or a high-speed and high-end MCU platform, and signal amplification by hardware reduces the demand for ADC accuracy, which reduces the cost.

  7. Fiber laser master oscillators for optical synchronization systems

    International Nuclear Information System (INIS)

    Winter, A.

    2008-04-01

    New X-ray free electron lasers (e.g. the European XFEL) require a new generation of synchronization system to achieve a stability of the FEL pulse, such that pump-probe experiments can fully utilize the ultra-short pulse duration (50 fs). An optical synchronization system has been developed based on the distribution of sub-ps optical pulses in length-stabilized fiber links. The synchronization information is contained in the precise repetition frequency of the optical pulses. In this thesis, the design and characterization of the laser serving as laser master oscillator is presented. An erbium-doped mode-locked fiber laser was chosen. Amplitude and phase noise were measured and record-low values of 0.03 % and 10 fs for the frequency range of 1 kHz to the Nyquist frequency were obtained. Furthermore, an initial proof-of-principle experiment for the optical synchronization system was performed in an accelerator environment. In this experiment, the fiber laser wase phase-locked to a microwave reference oscillator and a 500 meter long fiber link was stabilized to 12 fs rms over a range of 0.1 Hz to 20 kHz. RF signals were obtained from a photodetector without significant degradation at the end of the link. Furthermore, the laser master oscillator for FLASH was designed and is presently in fabrication and the initial infrastructure for the optical synchronization system was setup. (orig.)

  8. Fiber laser master oscillators for optical synchronization systems

    Energy Technology Data Exchange (ETDEWEB)

    Winter, A.

    2008-04-15

    New X-ray free electron lasers (e.g. the European XFEL) require a new generation of synchronization system to achieve a stability of the FEL pulse, such that pump-probe experiments can fully utilize the ultra-short pulse duration (50 fs). An optical synchronization system has been developed based on the distribution of sub-ps optical pulses in length-stabilized fiber links. The synchronization information is contained in the precise repetition frequency of the optical pulses. In this thesis, the design and characterization of the laser serving as laser master oscillator is presented. An erbium-doped mode-locked fiber laser was chosen. Amplitude and phase noise were measured and record-low values of 0.03 % and 10 fs for the frequency range of 1 kHz to the Nyquist frequency were obtained. Furthermore, an initial proof-of-principle experiment for the optical synchronization system was performed in an accelerator environment. In this experiment, the fiber laser wase phase-locked to a microwave reference oscillator and a 500 meter long fiber link was stabilized to 12 fs rms over a range of 0.1 Hz to 20 kHz. RF signals were obtained from a photodetector without significant degradation at the end of the link. Furthermore, the laser master oscillator for FLASH was designed and is presently in fabrication and the initial infrastructure for the optical synchronization system was setup. (orig.)

  9. Laser Megajoule synchronization system

    International Nuclear Information System (INIS)

    Luttmann, M.; Pastor, J.F; Drouet, V.; Prat, M.; Raimbourg, J.; Adolf, A.

    2011-01-01

    This paper describes the synchronisation system under development on the Laser Megajoule (LMJ) in order to synchronize the laser quads on the target to better than 40 ps rms. Our architecture is based on a Timing System (TS) which delivers trigger signals with jitter down to 15 ps rms coupled with an ultra precision timing system with 5 ps rms jitter. In addition to TS, a sensor placed at the target chamber center measures the arrival times of the 3 omega nano joule laser pulses generated by front end shots. (authors)

  10. The effect of synesthetic associations between the visual and auditory modalities on the Colavita effect

    OpenAIRE

    Stekelenburg, Jeroen J.; Keetels, Mirjam

    2015-01-01

    The Colavita effect refers to the phenomenon that when confronted with an audiovisual stimulus, observers report more often to have perceived the visual than the auditory component. The Colavita effect depends on low-level stimulus factors such as spatial and temporal proximity between the unimodal signals. Here, we examined whether the Colavita effect is modulated by synesthetic congruency between visual size and auditory pitch. If the Colavita effect depends on synesthetic congruency, we ex...

  11. Dynamic Correlations between Intrinsic Connectivity and Extrinsic Connectivity of the Auditory Cortex in Humans.

    Science.gov (United States)

    Cui, Zhuang; Wang, Qian; Gao, Yayue; Wang, Jing; Wang, Mengyang; Teng, Pengfei; Guan, Yuguang; Zhou, Jian; Li, Tianfu; Luan, Guoming; Li, Liang

    2017-01-01

    The arrival of sound signals in the auditory cortex (AC) triggers both local and inter-regional signal propagations over time up to hundreds of milliseconds and builds up both intrinsic functional connectivity (iFC) and extrinsic functional connectivity (eFC) of the AC. However, interactions between iFC and eFC are largely unknown. Using intracranial stereo-electroencephalographic recordings in people with drug-refractory epilepsy, this study mainly investigated the temporal dynamic of the relationships between iFC and eFC of the AC. The results showed that a Gaussian wideband-noise burst markedly elicited potentials in both the AC and numerous higher-order cortical regions outside the AC (non-auditory cortices). Granger causality analyses revealed that in the earlier time window, iFC of the AC was positively correlated with both eFC from the AC to the inferior temporal gyrus and that to the inferior parietal lobule. While in later periods, the iFC of the AC was positively correlated with eFC from the precentral gyrus to the AC and that from the insula to the AC. In conclusion, dual-directional interactions occur between iFC and eFC of the AC at different time windows following the sound stimulation and may form the foundation underlying various central auditory processes, including auditory sensory memory, object formation, integrations between sensory, perceptional, attentional, motor, emotional, and executive processes.

  12. Selective attention modulates human auditory brainstem responses: relative contributions of frequency and spatial cues.

    Directory of Open Access Journals (Sweden)

    Alexandre Lehmann

    Full Text Available Selective attention is the mechanism that allows focusing one's attention on a particular stimulus while filtering out a range of other stimuli, for instance, on a single conversation in a noisy room. Attending to one sound source rather than another changes activity in the human auditory cortex, but it is unclear whether attention to different acoustic features, such as voice pitch and speaker location, modulates subcortical activity. Studies using a dichotic listening paradigm indicated that auditory brainstem processing may be modulated by the direction of attention. We investigated whether endogenous selective attention to one of two speech signals affects amplitude and phase locking in auditory brainstem responses when the signals were either discriminable by frequency content alone, or by frequency content and spatial location. Frequency-following responses to the speech sounds were significantly modulated in both conditions. The modulation was specific to the task-relevant frequency band. The effect was stronger when both frequency and spatial information were available. Patterns of response were variable between participants, and were correlated with psychophysical discriminability of the stimuli, suggesting that the modulation was biologically relevant. Our results demonstrate that auditory brainstem responses are susceptible to efferent modulation related to behavioral goals. Furthermore they suggest that mechanisms of selective attention actively shape activity at early subcortical processing stages according to task relevance and based on frequency and spatial cues.

  13. Investigation of phase synchronization of interictal EEG in right temporal lobe epilepsy

    Science.gov (United States)

    Yu, Haitao; Cai, Lihui; Wu, Xinyu; Song, Zhenxi; Wang, Jiang; Xia, Zijie; Liu, Jing; Cao, Yibin

    2018-02-01

    Epilepsy is commonly associated with abnormally synchronous activity of neurons located in epileptogenic zones. In this study, we investigated the synchronization characteristic of right temporal lobe epilepsy (RTLE). Multichannel electroencephalography (EEG) data were recorded from the RTLE patients during interictal period and normal controls. Power spectral density was first used to analyze the EEG power for two groups of subjects. It was found that the power of epileptics is increased in the whole brain compared with that of the control. We calculated phase lag index (PLI) to measure the phase synchronization between each pair of EEG signals. A higher degree of synchronization was observed in the epileptics especially between distant channels. In particular, the regional synchronization degree was negatively correlated with power spectral density and the correlation was weaker for epileptics. Moreover, the synchronization degree decayed with the increase of relative distance of channels for both the epilepsy and control, but the dependence was weakened in the former. The obtained results may provide new insights into the generation mechanism of epilepsy.

  14. Auditory imagery shapes movement timing and kinematics: evidence from a musical task.

    Science.gov (United States)

    Keller, Peter E; Dalla Bella, Simone; Koch, Iring

    2010-04-01

    The role of anticipatory auditory imagery in music-like sequential action was investigated by examining timing accuracy and kinematics using a motion capture system. Musicians responded to metronomic pacing signals by producing three unpaced taps on three vertically aligned keys at the given tempo. Taps triggered tones in two out of three blocked feedback conditions, where key-to-tone mappings were compatible or incompatible in terms of spatial and pitch height. Results indicate that, while timing was most accurate without tones, movements were smaller in amplitude and less forceful (i.e., acceleration prior to impact was lowest) when tones were present. Moreover, timing was more accurate and movements were less forceful with compatible than with incompatible auditory feedback. Observing these effects at the first tap (before tone onset) suggests that anticipatory auditory imagery modulates the temporal kinematics of regularly timed auditory action sequences, like those found in music. Such cross-modal ideomotor processes may function to facilitate planning efficiency and biomechanical economy in voluntary action. Copyright 2010 APA, all rights reserved.

  15. Improvement of auditory hallucinations and reduction of primary auditory area's activation following TMS

    International Nuclear Information System (INIS)

    Giesel, Frederik L.; Mehndiratta, Amit; Hempel, Albrecht; Hempel, Eckhard; Kress, Kai R.; Essig, Marco; Schröder, Johannes

    2012-01-01

    Background: In the present case study, improvement of auditory hallucinations following transcranial magnetic stimulation (TMS) therapy was investigated with respect to activation changes of the auditory cortices. Methods: Using functional magnetic resonance imaging (fMRI), activation of the auditory cortices was assessed prior to and after a 4-week TMS series of the left superior temporal gyrus in a schizophrenic patient with medication-resistant auditory hallucinations. Results: Hallucinations decreased slightly after the third and profoundly after the fourth week of TMS. Activation in the primary auditory area decreased, whereas activation in the operculum and insula remained stable. Conclusions: Combination of TMS and repetitive fMRI is promising to elucidate the physiological changes induced by TMS.

  16. Auditory and visual interactions between the superior and inferior colliculi in the ferret.

    Science.gov (United States)

    Stitt, Iain; Galindo-Leon, Edgar; Pieper, Florian; Hollensteiner, Karl J; Engler, Gerhard; Engel, Andreas K

    2015-05-01

    The integration of visual and auditory spatial information is important for building an accurate perception of the external world, but the fundamental mechanisms governing such audiovisual interaction have only partially been resolved. The earliest interface between auditory and visual processing pathways is in the midbrain, where the superior (SC) and inferior colliculi (IC) are reciprocally connected in an audiovisual loop. Here, we investigate the mechanisms of audiovisual interaction in the midbrain by recording neural signals from the SC and IC simultaneously in anesthetized ferrets. Visual stimuli reliably produced band-limited phase locking of IC local field potentials (LFPs) in two distinct frequency bands: 6-10 and 15-30 Hz. These visual LFP responses co-localized with robust auditory responses that were characteristic of the IC. Imaginary coherence analysis confirmed that visual responses in the IC were not volume-conducted signals from the neighboring SC. Visual responses in the IC occurred later than retinally driven superficial SC layers and earlier than deep SC layers that receive indirect visual inputs, suggesting that retinal inputs do not drive visually evoked responses in the IC. In addition, SC and IC recording sites with overlapping visual spatial receptive fields displayed stronger functional connectivity than sites with separate receptive fields, indicating that visual spatial maps are aligned across both midbrain structures. Reciprocal coupling between the IC and SC therefore probably serves the dynamic integration of visual and auditory representations of space. © 2015 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  17. Multielectrode recordings from auditory neurons in the brain of a small grasshopper.

    Science.gov (United States)

    Bhavsar, Mit Balvantray; Heinrich, Ralf; Stumpner, Andreas

    2015-12-30

    Grasshoppers have been used as a model system to study the neuronal basis of insect acoustic behavior. Auditory neurons have been described from intracellular recordings. The growing interest to study population activity of neurons has been satisfied so far with artificially combining data from different individuals. We for the first time used multielectrode recordings from a small grasshopper brain. We used three 12μm tungsten wires (combined in a multielectrode) to record from local brain neurons and from a population of auditory neurons entering the brain from the thorax. Spikes of the recorded units were separated by sorting algorithms and spike collision analysis. The tungsten wires enabled stable recordings with high signal to noise ratio. Due to the tight temporal coupling of auditory activity to the stimulus spike collisions were frequent and collision analysis retrieved 10-15% of additional spikes. Marking the electrode position was possible using a fluorescent dye or electrocoagulation with high current. Physiological identification of units described from intracellular recordings was hard to achieve. 12μm tungsten wires gave a better signal to noise ratio than 15μm copper wires previously used in recordings from bees' brains. Recording the population activity of auditory neurons in one individual prevents interindividual and trial-to-trial variability which otherwise reduce the validity of the analysis. Double intracellular recordings have quite low success rate and therefore are rarely achieved and their stability is much lower than that of multielectrode recordings which allows sampling of data for 30min or more. Copyright © 2015 Elsevier B.V. All rights reserved.

  18. Identification of neural structures involved in stuttering using vibrotactile feedback.

    Science.gov (United States)

    Cheadle, Oliver; Sorger, Clarissa; Howell, Peter

    Feedback delivered over auditory and vibratory afferent pathways has different effects on the fluency of people who stutter (PWS). These features were exploited to investigate the neural structures involved in stuttering. The speech signal vibrated locations on the body (vibrotactile feedback, VTF). Eleven PWS read passages under VTF and control (no-VTF) conditions. All combinations of vibration amplitude, synchronous or delayed VTF and vibrator position (hand, sternum or forehead) were presented. Control conditions were performed at the beginning, middle and end of test sessions. Stuttering rate, but not speaking rate, differed between the control and VTF conditions. Notably, speaking rate did not change between when VTF was delayed versus when it was synchronous in contrast with what happens with auditory feedback. This showed that cerebellar mechanisms, which are affected when auditory feedback is delayed, were not implicated in the fluency-enhancing effects of VTF, suggesting that there is a second fluency-enhancing mechanism. Copyright © 2018 Elsevier Inc. All rights reserved.

  19. The synchronous active neutron detection assay system

    International Nuclear Information System (INIS)

    Pickrell, M.M.; Kendall, P.K.

    1994-01-01

    We have begun to develop a novel technique for active neutron assay of fissile material in spent nuclear fuel. This approach will exploit a 14-MeV neutron generator developed by Schlumberger. The technique, termed synchronous active neutron detection (SAND), follows a method used routinely in other branches of physics to detect very small signals in presence of large backgrounds. Synchronous detection instruments are widely available commercially and are termed ''lock-in'' amplifiers. We have implemented a digital lock-in amplifier in conjunction with the Schlumberger neutron generator to explore the possibility of synchronous detection with active neutrons. The Schlumberger system can operate at up to a 50% duty factor, in effect, a square wave of neutron yield. Results are preliminary but promising. The system is capable of resolving the fissile material contained in a small fraction of the fuel rods in a cold fuel assembly; it also appears resilient to background neutron interference. The interrogating neutrons appear to be non-thermal and penetrating. Work remains to fully explore relevant physics and optimize instrument design

  20. Global chaos synchronization of three coupled nonlinear autonomous systems and a novel method of chaos encryption

    International Nuclear Information System (INIS)

    An Xinlei; Yu Jianning; Chu Yandong; Zhang Jiangang; Zhang Li

    2009-01-01

    In this paper, we discussed the fixed points and their linear stability of a new nonlinear autonomous system that introduced by J.C. Sprott. Based on Lyapunov stabilization theorem, a global chaos synchronization scheme of three coupled identical systems is investigated. By choosing proper coupling parameters, the states of all the three systems can be synchronized. Then this method was applied to secure communication through chaotic masking, used three coupled identical systems, propose a novel method of chaos encryption, after encrypting in the previous two transmitters, information signal can be recovered exactly at the receiver end. Simulation results show that the method can realize monotonous synchronization. Further more, the information signal can be recovered undistorted when applying this method to secure communication.

  1. Novel Blind Recognition Algorithm of Frame Synchronization Words Based on Soft-Decision in Digital Communication Systems.

    Science.gov (United States)

    Qin, Jiangyi; Huang, Zhiping; Liu, Chunwu; Su, Shaojing; Zhou, Jing

    2015-01-01

    A novel blind recognition algorithm of frame synchronization words is proposed to recognize the frame synchronization words parameters in digital communication systems. In this paper, a blind recognition method of frame synchronization words based on the hard-decision is deduced in detail. And the standards of parameter recognition are given. Comparing with the blind recognition based on the hard-decision, utilizing the soft-decision can improve the accuracy of blind recognition. Therefore, combining with the characteristics of Quadrature Phase Shift Keying (QPSK) signal, an improved blind recognition algorithm based on the soft-decision is proposed. Meanwhile, the improved algorithm can be extended to other signal modulation forms. Then, the complete blind recognition steps of the hard-decision algorithm and the soft-decision algorithm are given in detail. Finally, the simulation results show that both the hard-decision algorithm and the soft-decision algorithm can recognize the parameters of frame synchronization words blindly. What's more, the improved algorithm can enhance the accuracy of blind recognition obviously.

  2. Mapping the Information Trace in Local Field Potentials by a Computational Method of Two-Dimensional Time-Shifting Synchronization Likelihood Based on Graphic Processing Unit Acceleration.

    Science.gov (United States)

    Zhao, Zi-Fang; Li, Xue-Zhu; Wan, You

    2017-12-01

    The local field potential (LFP) is a signal reflecting the electrical activity of neurons surrounding the electrode tip. Synchronization between LFP signals provides important details about how neural networks are organized. Synchronization between two distant brain regions is hard to detect using linear synchronization algorithms like correlation and coherence. Synchronization likelihood (SL) is a non-linear synchronization-detecting algorithm widely used in studies of neural signals from two distant brain areas. One drawback of non-linear algorithms is the heavy computational burden. In the present study, we proposed a graphic processing unit (GPU)-accelerated implementation of an SL algorithm with optional 2-dimensional time-shifting. We tested the algorithm with both artificial data and raw LFP data. The results showed that this method revealed detailed information from original data with the synchronization values of two temporal axes, delay time and onset time, and thus can be used to reconstruct the temporal structure of a neural network. Our results suggest that this GPU-accelerated method can be extended to other algorithms for processing time-series signals (like EEG and fMRI) using similar recording techniques.

  3. Electrophysiological correlates of predictive coding of auditory location in the perception of natural audiovisual events

    Directory of Open Access Journals (Sweden)

    Jeroen eStekelenburg

    2012-05-01

    Full Text Available In many natural audiovisual events (e.g., a clap of the two hands, the visual signal precedes the sound and thus allows observers to predict when, where, and which sound will occur. Previous studies have already reported that there are distinct neural correlates of temporal (when versus phonetic/semantic (which content on audiovisual integration. Here we examined the effect of visual prediction of auditory location (where in audiovisual biological motion stimuli by varying the spatial congruency between the auditory and visual part of the audiovisual stimulus. Visual stimuli were presented centrally, whereas auditory stimuli were presented either centrally or at 90° azimuth. Typical subadditive amplitude reductions (AV – V < A were found for the auditory N1 and P2 for spatially congruent and incongruent conditions. The new finding is that the N1 suppression was larger for spatially congruent stimuli. A very early audiovisual interaction was also found at 30-50 ms in the spatially congruent condition, while no effect of congruency was found on the suppression of the P2. This indicates that visual prediction of auditory location can be coded very early in auditory processing.

  4. Auditory Perspective Taking

    National Research Council Canada - National Science Library

    Martinson, Eric; Brock, Derek

    2006-01-01

    .... From this knowledge of another's auditory perspective, a conversational partner can then adapt his or her auditory output to overcome a variety of environmental challenges and insure that what is said is intelligible...

  5. Hidden imperfect synchronization of wall turbulence.

    Science.gov (United States)

    Tardu, Sedat F

    2010-03-01

    Instantaneous amplitude and phase concept emerging from analytical signal formulation is applied to the wavelet coefficients of streamwise velocity fluctuations in the buffer layer of a near wall turbulent flow. Experiments and direct numerical simulations show both the existence of long periods of inert zones wherein the local phase is constant. These regions are separated by random phase jumps. The local amplitude is globally highly intermittent, but not in the phase locked regions wherein it varies smoothly. These behaviors are reminiscent of phase synchronization phenomena observed in stochastic chaotic systems. The lengths of the constant phase inert (laminar) zones reveal a type I intermittency behavior, in concordance with saddle-node bifurcation, and the periodic orbits of saddle nature recently identified in Couette turbulence. The imperfect synchronization is related to the footprint of coherent Reynolds shear stress producing eddies convecting in the low buffer.

  6. Multichannel Spatial Auditory Display for Speed Communications

    Science.gov (United States)

    Begault, Durand R.; Erbe, Tom

    1994-01-01

    A spatial auditory display for multiple speech communications was developed at NASA/Ames Research Center. Input is spatialized by the use of simplifiedhead-related transfer functions, adapted for FIR filtering on Motorola 56001 digital signal processors. Hardware and firmware design implementations are overviewed for the initial prototype developed for NASA-Kennedy Space Center. An adaptive staircase method was used to determine intelligibility levels of four-letter call signs used by launch personnel at NASA against diotic speech babble. Spatial positions at 30 degree azimuth increments were evaluated. The results from eight subjects showed a maximum intelligibility improvement of about 6-7 dB when the signal was spatialized to 60 or 90 degree azimuth positions.

  7. An interprojection sensor fusion approach to estimate blocked projection signal in synchronized moving grid-based CBCT system

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Hong; Kong, Vic [Department of Radiation Oncology, Georgia Regents University, Augusta, Georgia 30912 (United States); Ren, Lei; Giles, William; Zhang, You [Department of Radiation Oncology, Duke University, Durham, North Carolina 27710 (United States); Jin, Jian-Yue, E-mail: jjin@gru.edu [Department of Radiation Oncology, Georgia Regents University, Augusta, Georgia 30912 and Department of Radiology, Georgia Regents University, Augusta, Georgia 30912 (United States)

    2016-01-15

    Purpose: A preobject grid can reduce and correct scatter in cone beam computed tomography (CBCT). However, half of the signal in each projection is blocked by the grid. A synchronized moving grid (SMOG) has been proposed to acquire two complimentary projections at each gantry position and merge them into one complete projection. That approach, however, suffers from increased scanning time and the technical difficulty of accurately merging the two projections per gantry angle. Herein, the authors present a new SMOG approach which acquires a single projection per gantry angle, with complimentary grid patterns for any two adjacent projections, and use an interprojection sensor fusion (IPSF) technique to estimate the blocked signal in each projection. The method may have the additional benefit of reduced imaging dose due to the grid blocking half of the incident radiation. Methods: The IPSF considers multiple paired observations from two adjacent gantry angles as approximations of the blocked signal and uses a weighted least square regression of these observations to finally determine the blocked signal. The method was first tested with a simulated SMOG on a head phantom. The signal to noise ratio (SNR), which represents the difference of the recovered CBCT image to the original image without the SMOG, was used to evaluate the ability of the IPSF in recovering the missing signal. The IPSF approach was then tested using a Catphan phantom on a prototype SMOG assembly installed in a bench top CBCT system. Results: In the simulated SMOG experiment, the SNRs were increased from 15.1 and 12.7 dB to 35.6 and 28.9 dB comparing with a conventional interpolation method (inpainting method) for a projection and the reconstructed 3D image, respectively, suggesting that IPSF successfully recovered most of blocked signal. In the prototype SMOG experiment, the authors have successfully reconstructed a CBCT image using the IPSF-SMOG approach. The detailed geometric features in the

  8. An interprojection sensor fusion approach to estimate blocked projection signal in synchronized moving grid-based CBCT system

    International Nuclear Information System (INIS)

    Zhang, Hong; Kong, Vic; Ren, Lei; Giles, William; Zhang, You; Jin, Jian-Yue

    2016-01-01

    Purpose: A preobject grid can reduce and correct scatter in cone beam computed tomography (CBCT). However, half of the signal in each projection is blocked by the grid. A synchronized moving grid (SMOG) has been proposed to acquire two complimentary projections at each gantry position and merge them into one complete projection. That approach, however, suffers from increased scanning time and the technical difficulty of accurately merging the two projections per gantry angle. Herein, the authors present a new SMOG approach which acquires a single projection per gantry angle, with complimentary grid patterns for any two adjacent projections, and use an interprojection sensor fusion (IPSF) technique to estimate the blocked signal in each projection. The method may have the additional benefit of reduced imaging dose due to the grid blocking half of the incident radiation. Methods: The IPSF considers multiple paired observations from two adjacent gantry angles as approximations of the blocked signal and uses a weighted least square regression of these observations to finally determine the blocked signal. The method was first tested with a simulated SMOG on a head phantom. The signal to noise ratio (SNR), which represents the difference of the recovered CBCT image to the original image without the SMOG, was used to evaluate the ability of the IPSF in recovering the missing signal. The IPSF approach was then tested using a Catphan phantom on a prototype SMOG assembly installed in a bench top CBCT system. Results: In the simulated SMOG experiment, the SNRs were increased from 15.1 and 12.7 dB to 35.6 and 28.9 dB comparing with a conventional interpolation method (inpainting method) for a projection and the reconstructed 3D image, respectively, suggesting that IPSF successfully recovered most of blocked signal. In the prototype SMOG experiment, the authors have successfully reconstructed a CBCT image using the IPSF-SMOG approach. The detailed geometric features in the

  9. Synchronization control for ultrafast laser parallel microdrilling system

    Science.gov (United States)

    Zhai, Zhongsheng; Kuang, Zheng; Ouyang, Jinlei; Liu, Dun; Perrie, Walter; Edwardson, Stuart P.; Dearden, Geoff

    2014-11-01

    Ultrafast lasers, emitting ultra-short pulses of light, generally of the order of femtoseconds to ten picoseconds, are widely used in micro-processing with the advantage of very little thermal damage. Parallel micro-processing is seen significant developments in laser fabrication, thanking to the spatial light modulator (SLM) which can concert single beam to multiple beams through computer generate holograms (CGHs). However, without synchronization control, on the conditions of changing different holograms or processing on large area beyond scanning galvo's ability, the fabrication will be interrupted constantly for changing holograms and moving the stages. Therefore, synchronization control is very important to improve the convenience and application of parallel micro-processing. A synchronization control method, carried out through two application software: SAMLight (or WaveRunner) and Labview, is presented in this paper. SAMLight is used to control the laser and the scanning galvo to implement microprocessing, and the developed program with Labview is used to control the SLM and motion stages. The synchronization signals, transmitted between the two software, are utilized by a National Instruments (NI) device USB-6008. Using optimal control methods, the synchronized system can easily and automatically accomplish complicated fabrications with minimum time. A multi-drilling application is provided to verify the affectivity of the synchronized control method. It uses multiple annular beams, generated by superimposing multi-beam CGH onto a diffractive axicon CGH, to drill multiple holes at one time, and it can automatically finish different patterns based on synchronization control. This drilling way is an optical trepanning and it avoids huge laser energy waste with attenuation. The multi-beam CGHs, generated by the Grating and Lens algorithm, are different for different patterns. The processing is over 200 times faster than traditional mechanical trepanning

  10. Three-dimensional Acoustic Localisation via Directed Movements of a Two-dimensional Model of the Lizard Peripheral Auditory System

    DEFF Research Database (Denmark)

    Shaikh, Danish; Kjær Schmidt, Michael

    2017-01-01

    of the acoustic target with respect to one plane of rotation. A multi-layer perceptron neural network is trained via supervised learning to translate the combination of the two measurements into an estimate of the relative location of the acoustic target in terms of its azimuth and elevation. The acoustic...... localisation performance of the system is evaluated in simulation for noiseless as well as noisy sinusoidal auditory signals with a 20 dB signal-to-noise ratio for four different sound frequencies of 1450 Hz, 1650 Hz, 1850 Hz and 2050 Hz that span the response frequency range of the peripheral auditory model...

  11. The integration of temporally shifted visual feedback in a synchronization task: The role of perceptual stability in a visuo-proprioceptive conflict situation.

    Science.gov (United States)

    Ceux, Tanja; Montagne, Gilles; Buekers, Martinus J

    2010-12-01

    The present study examined whether the beneficial role of coherently grouped visual motion structures for performing complex (interlimb) coordination patterns can be generalized to synchronization behavior in a visuo-proprioceptive conflict situation. To achieve this goal, 17 participants had to synchronize a self-moved circle, representing the arm movement, with a visual target signal corresponding to five temporally shifted visual feedback conditions (0%, 25%, 50%, 75%, and 100% of the target cycle duration) in three synchronization modes (in-phase, anti-phase, and intermediate). The results showed that the perception of a newly generated perceptual Gestalt between the visual feedback of the arm and the target signal facilitated the synchronization performance in the preferred in-phase synchronization mode in contrast to the less stable anti-phase and intermediate mode. Our findings suggest that the complexity of the synchronization mode defines to what extent the visual and/or proprioceptive information source affects the synchronization performance in the present unimanual synchronization task. Copyright © 2010 Elsevier B.V. All rights reserved.

  12. Competition and convergence between auditory and cross-modal visual inputs to primary auditory cortical areas

    Science.gov (United States)

    Mao, Yu-Ting; Hua, Tian-Miao

    2011-01-01

    Sensory neocortex is capable of considerable plasticity after sensory deprivation or damage to input pathways, especially early in development. Although plasticity can often be restorative, sometimes novel, ectopic inputs invade the affected cortical area. Invading inputs from other sensory modalities may compromise the original function or even take over, imposing a new function and preventing recovery. Using ferrets whose retinal axons were rerouted into auditory thalamus at birth, we were able to examine the effect of varying the degree of ectopic, cross-modal input on reorganization of developing auditory cortex. In particular, we assayed whether the invading visual inputs and the existing auditory inputs competed for or shared postsynaptic targets and whether the convergence of input modalities would induce multisensory processing. We demonstrate that although the cross-modal inputs create new visual neurons in auditory cortex, some auditory processing remains. The degree of damage to auditory input to the medial geniculate nucleus was directly related to the proportion of visual neurons in auditory cortex, suggesting that the visual and residual auditory inputs compete for cortical territory. Visual neurons were not segregated from auditory neurons but shared target space even on individual target cells, substantially increasing the proportion of multisensory neurons. Thus spatial convergence of visual and auditory input modalities may be sufficient to expand multisensory representations. Together these findings argue that early, patterned visual activity does not drive segregation of visual and auditory afferents and suggest that auditory function might be compromised by converging visual inputs. These results indicate possible ways in which multisensory cortical areas may form during development and evolution. They also suggest that rehabilitative strategies designed to promote recovery of function after sensory deprivation or damage need to take into

  13. Differential Recruitment of Auditory Cortices in the Consolidation of Recent Auditory Fearful Memories.

    Science.gov (United States)

    Cambiaghi, Marco; Grosso, Anna; Renna, Annamaria; Sacchetti, Benedetto

    2016-08-17

    Memories of frightening events require a protracted consolidation process. Sensory cortex, such as the auditory cortex, is involved in the formation of fearful memories with a more complex sensory stimulus pattern. It remains controversial, however, whether the auditory cortex is also required for fearful memories related to simple sensory stimuli. In the present study, we found that, 1 d after training, the temporary inactivation of either the most anterior region of the auditory cortex, including the primary (Te1) cortex, or the most posterior region, which included the secondary (Te2) component, did not affect the retention of recent memories, which is consistent with the current literature. However, at this time point, the inactivation of the entire auditory cortices completely prevented the formation of new memories. Amnesia was site specific and was not due to auditory stimuli perception or processing and strictly related to the interference with memory consolidation processes. Strikingly, at a late time interval 4 d after training, blocking the posterior part (encompassing the Te2) alone impaired memory retention, whereas the inactivation of the anterior part (encompassing the Te1) left memory unaffected. Together, these data show that the auditory cortex is necessary for the consolidation of auditory fearful memories related to simple tones in rats. Moreover, these results suggest that, at early time intervals, memory information is processed in a distributed network composed of both the anterior and the posterior auditory cortical regions, whereas, at late time intervals, memory processing is concentrated in the most posterior part containing the Te2 region. Memories of threatening experiences undergo a prolonged process of "consolidation" to be maintained for a long time. The dynamic of fearful memory consolidation is poorly understood. Here, we show that 1 d after learning, memory is processed in a distributed network composed of both primary Te1 and

  14. The roles of superficial amygdala and auditory cortex in music-evoked fear and joy.

    Science.gov (United States)

    Koelsch, Stefan; Skouras, Stavros; Fritz, Thomas; Herrera, Perfecto; Bonhage, Corinna; Küssner, Mats B; Jacobs, Arthur M

    2013-11-01

    This study investigates neural correlates of music-evoked fear and joy with fMRI. Studies on neural correlates of music-evoked fear are scant, and there are only a few studies on neural correlates of joy in general. Eighteen individuals listened to excerpts of fear-evoking, joy-evoking, as well as neutral music and rated their own emotional state in terms of valence, arousal, fear, and joy. Results show that BOLD signal intensity increased during joy, and decreased during fear (compared to the neutral condition) in bilateral auditory cortex (AC) and bilateral superficial amygdala (SF). In the right primary somatosensory cortex (area 3b) BOLD signals increased during exposure to fear-evoking music. While emotion-specific activity in AC increased with increasing duration of each trial, SF responded phasically in the beginning of the stimulus, and then SF activity declined. Psychophysiological Interaction (PPI) analysis revealed extensive emotion-specific functional connectivity of AC with insula, cingulate cortex, as well as with visual, and parietal attentional structures. These findings show that the auditory cortex functions as a central hub of an affective-attentional network that is more extensive than previously believed. PPI analyses also showed functional connectivity of SF with AC during the joy condition, taken to reflect that SF is sensitive to social signals with positive valence. During fear music, SF showed functional connectivity with visual cortex and area 7 of the superior parietal lobule, taken to reflect increased visual alertness and an involuntary shift of attention during the perception of auditory signals of danger. Copyright © 2013 Elsevier Inc. All rights reserved.

  15. The algorithm of random length sequences synthesis for frame synchronization of digital television systems

    Directory of Open Access Journals (Sweden)

    Аndriy V. Sadchenko

    2015-12-01

    Full Text Available Digital television systems need to ensure that all digital signals processing operations are performed simultaneously and consistently. Frame synchronization dictated by the need to match phases of transmitter and receiver so that it would be possible to identify the start of a frame. As a frame synchronization signals are often used long length binary sequence with good aperiodic autocorrelation function. Aim: This work is dedicated to the development of the algorithm of random length sequences synthesis. Materials and Methods: The paper provides a comparative analysis of the known sequences, which can be used at present as synchronization ones, revealed their advantages and disadvantages. This work proposes the algorithm for the synthesis of binary synchronization sequences of random length with good autocorrelation properties based on noise generator with a uniform distribution law of probabilities. A "white noise" semiconductor generator is proposed to use as the initial material for the synthesis of binary sequences with desired properties. Results: The statistical analysis of the initial implementations of the "white noise" and synthesized sequences for frame synchronization of digital television is conducted. The comparative analysis of the synthesized sequences with known ones was carried out. The results show the benefits of obtained sequences in compare with known ones. The performed simulations confirm the obtained results. Conclusions: Thus, the search algorithm of binary synchronization sequences with desired autocorrelation properties received. According to this algorithm, the sequence can be longer in length and without length limitations. The received sync sequence can be used for frame synchronization in modern digital communication systems that will increase their efficiency and noise immunity.

  16. Overview of Cell Synchronization.

    Science.gov (United States)

    Banfalvi, Gaspar

    2017-01-01

    The widespread interest in cell synchronization is maintained by the studies of control mechanism involved in cell cycle regulation. During the synchronization distinct subpopulations of cells are obtained representing different stages of the cell cycle. These subpopulations are then used to study regulatory mechanisms of the cycle at the level of macromolecular biosynthesis (DNA synthesis, gene expression, protein synthesis), protein phosphorylation, development of new drugs, etc. Although several synchronization methods have been described, it is of general interest that scientists get a compilation and an updated view of these synchronization techniques. This introductory chapter summarizes: (1) the basic concepts and principal criteria of cell cycle synchronizations, (2) the most frequently used synchronization methods, such as physical fractionation (flow cytometry, dielectrophoresis, cytofluorometric purification), chemical blockade, (3) synchronization of embryonic cells, (4) synchronization at low temperature, (5) comparison of cell synchrony techniques, (6) synchronization of unicellular organisms, and (7) the effect of synchronization on transfection.

  17. A loudspeaker-based room auralization system for auditory research

    DEFF Research Database (Denmark)

    Favrot, Sylvain Emmanuel

    to systematically study the signal processing of realistic sounds by normal-hearing and hearing-impaired listeners, a flexible, reproducible and fully controllable auditory environment is needed. A loudspeaker-based room auralization (LoRA) system was developed in this thesis to provide virtual auditory...... in reverberant environments. Each part of the early incoming sound to the listener was auralized with either higher-order Ambisonic (HOA) or using a single loudspeaker. The late incoming sound was auralized with a specific algorithm in order to provide a diffuse reverberation with minimal coloration artifacts...... assessed the impact of the auralization technique used for the early incoming sound (HOA or single loudspeaker) on speech intelligibility. A listening test showed that speech intelligibility experiments can be reliably conducted with the LoRA system with both techniques. The second evaluation investigated...

  18. The relationship between the age of onset of musical training and rhythm synchronization performance: validation of sensitive period effects.

    Science.gov (United States)

    Bailey, Jennifer A; Penhune, Virginia B

    2013-01-01

    A sensitive period associated with musical training has been proposed, suggesting the influence of musical training on the brain and behavior is strongest during the early years of childhood. Experiments from our laboratory have directly tested the sensitive period hypothesis for musical training by comparing musicians who began their training prior to age seven with those who began their training after age seven, while matching the two groups in terms of musical experience (Watanabe et al., 2007; Bailey and Penhune, 2010, 2012). Using this matching paradigm, the early-trained groups have demonstrated enhanced sensorimotor synchronization skills and associated differences in brain structure (Bailey et al., 2013; Steele et al., 2013). The current study takes a different approach to investigating the sensitive period hypothesis for musical training by examining a single large group of unmatched musicians (N = 77) and exploring the relationship between age of onset of musical training as a continuous variable and performance on the Rhythm Synchronization Task (RST), a previously used auditory-motor RST. Interestingly, age of onset was correlated with task performance for those who began training earlier, however, no such relationship was observed among those who began training in their later childhood years. In addition, years of formal training showed a similar pattern. However, individual working memory scores were predictive of task performance, regardless of age of onset of musical training. Overall, these results support the sensitive period hypothesis for musical training and suggest a non-linear relationship between age of onset of musical training and auditory-motor rhythm synchronization abilities, such that a relationship exists early in childhood but then plateaus later on in development, similar to maturational growth trajectories of brain regions implicated in playing music.

  19. Method for traceable measurement of LTE signals

    Science.gov (United States)

    Sunder Dash, Soumya; Pythoud, Frederic; Leuchtmann, Pascal; Leuthold, Juerg

    2018-04-01

    This contribution presents a reference setup to measure the power of the cell-specific resource elements present in downlink long term evolution (LTE) signals in a way that the measurements are traceable to the international system of units. This setup can be used to calibrate the LTE code-selective field probes that are used to measure the radiation of base stations for mobile telephony. It can also be used to calibrate LTE signal generators and receivers. The method is based on traceable scope measurements performed directly at the output of a measuring antenna. It implements offline digital signal processing demodulation algorithms that consider the digital down-conversion, timing synchronization, frequency synchronization, phase synchronization and robust LTE cell identification to produce the downlink time-frequency LTE grid. Experimental results on conducted test scenarios, both single-input-single-output and multiple-input-multiple-output antenna configuration, show promising results confirming measurement uncertainties of the order of 0.05 dB with a coverage factor of 2.

  20. Manipulation of Auditory Inputs as Rehabilitation Therapy for Maladaptive Auditory Cortical Reorganization

    Directory of Open Access Journals (Sweden)

    Hidehiko Okamoto

    2018-01-01

    Full Text Available Neurophysiological and neuroimaging data suggest that the brains of not only children but also adults are reorganized based on sensory inputs and behaviors. Plastic changes in the brain are generally beneficial; however, maladaptive cortical reorganization in the auditory cortex may lead to hearing disorders such as tinnitus and hyperacusis. Recent studies attempted to noninvasively visualize pathological neural activity in the living human brain and reverse maladaptive cortical reorganization by the suitable manipulation of auditory inputs in order to alleviate detrimental auditory symptoms. The effects of the manipulation of auditory inputs on maladaptively reorganized brain were reviewed herein. The findings obtained indicate that rehabilitation therapy based on the manipulation of auditory inputs is an effective and safe approach for hearing disorders. The appropriate manipulation of sensory inputs guided by the visualization of pathological brain activities using recent neuroimaging techniques may contribute to the establishment of new clinical applications for affected individuals.

  1. The relationship between external auditory canal temperature and onset of estrus and ovulation in beef heifers.

    Science.gov (United States)

    Randi, Federico; McDonald, Michael; Duffy, Pat; Kelly, Alan K; Lonergan, Patrick

    2018-04-01

    The aim of this study was to evaluate the relationship of body temperature fluctuations, as measured by external auditory canal temperature, to the onset of estrus and ovulation. Beef heifers (n = 44, mean age 23.5 ± 0.4 months, mean weight 603.3 ± 5.7 kg) were fitted with a Boviminder ® ear tag 2 weeks before the start of the estrous synchronization protocol to allow acclimatization. The device recorded the temperature, accurate to 0.01° Fahrenheit, every 10 min and transmitted the data via a base station over the internet where it could be accessed remotely. The estrous cycles of all heifers were synchronized using an 8-day progesterone-based synchronization program; on day 0 a PRID was inserted in conjunction with an injection of GnRH, and PGF2α was administered the day before PRID removal. Heifers were checked for signs of estrus at 4-h intervals (i.e., 6 times per day) commencing 24 h after PRID withdrawal. Beginning 12 h after the onset of estrus, the ovaries were ultrasound scanned at 4-h intervals to determine the time of ovulation. Body temperature was recorded every 10 min and averaged to hourly means for the following 4 periods relative to the detected oestrus onset (=Time 0): Period I: -48 h to -7 h, Period II: -6 h to +6 h, Period III +7 h to ovulation, and Period IV: ovulation to 48 h post ovulation. Data were analysed using a Mixed Model ANOVA in SAS in a completely randomized design to observe effects of induced estrus on external auditory canal temperature. The mean (±SD) interval from removal of the PRID to onset of estrus activity was 46.6 ± 14.7 h. The mean duration of estrus was 16.0 ± 5.67 h and the mean interval from estrus onset to ovulation was 27.9 ± 7.68 h. Highest temperatures (100.95 ± 0.03 °F) were observed in Period II around estrus onset, whereas lowest temperatures were observed in the 48 h preceding estrus onset (100.28 ± 0.03 °F; Period I) and around ovulation (100.30

  2. Association of Concurrent fNIRS and EEG Signatures in Response to Auditory and Visual Stimuli.

    Science.gov (United States)

    Chen, Ling-Chia; Sandmann, Pascale; Thorne, Jeremy D; Herrmann, Christoph S; Debener, Stefan

    2015-09-01

    Functional near-infrared spectroscopy (fNIRS) has been proven reliable for investigation of low-level visual processing in both infants and adults. Similar investigation of fundamental auditory processes with fNIRS, however, remains only partially complete. Here we employed a systematic three-level validation approach to investigate whether fNIRS could capture fundamental aspects of bottom-up acoustic processing. We performed a simultaneous fNIRS-EEG experiment with visual and auditory stimulation in 24 participants, which allowed the relationship between changes in neural activity and hemoglobin concentrations to be studied. In the first level, the fNIRS results showed a clear distinction between visual and auditory sensory modalities. Specifically, the results demonstrated area specificity, that is, maximal fNIRS responses in visual and auditory areas for the visual and auditory stimuli respectively, and stimulus selectivity, whereby the visual and auditory areas responded mainly toward their respective stimuli. In the second level, a stimulus-dependent modulation of the fNIRS signal was observed in the visual area, as well as a loudness modulation in the auditory area. Finally in the last level, we observed significant correlations between simultaneously-recorded visual evoked potentials and deoxygenated hemoglobin (DeoxyHb) concentration, and between late auditory evoked potentials and oxygenated hemoglobin (OxyHb) concentration. In sum, these results suggest good sensitivity of fNIRS to low-level sensory processing in both the visual and the auditory domain, and provide further evidence of the neurovascular coupling between hemoglobin concentration changes and non-invasive brain electrical activity.

  3. Auditory temporal preparation induced by rhythmic cues during concurrent auditory working memory tasks.

    Science.gov (United States)

    Cutanda, Diana; Correa, Ángel; Sanabria, Daniel

    2015-06-01

    The present study investigated whether participants can develop temporal preparation driven by auditory isochronous rhythms when concurrently performing an auditory working memory (WM) task. In Experiment 1, participants had to respond to an auditory target presented after a regular or an irregular sequence of auditory stimuli while concurrently performing a Sternberg-type WM task. Results showed that participants responded faster after regular compared with irregular rhythms and that this effect was not affected by WM load; however, the lack of a significant main effect of WM load made it difficult to draw any conclusion regarding the influence of the dual-task manipulation in Experiment 1. In order to enhance dual-task interference, Experiment 2 combined the auditory rhythm procedure with an auditory N-Back task, which required WM updating (monitoring and coding of the information) and was presumably more demanding than the mere rehearsal of the WM task used in Experiment 1. Results now clearly showed dual-task interference effects (slower reaction times [RTs] in the high- vs. the low-load condition). However, such interference did not affect temporal preparation induced by rhythms, with faster RTs after regular than after irregular sequences in the high-load and low-load conditions. These results revealed that secondary tasks demanding memory updating, relative to tasks just demanding rehearsal, produced larger interference effects on overall RTs in the auditory rhythm task. Nevertheless, rhythm regularity exerted a strong temporal preparation effect that survived the interference of the WM task even when both tasks competed for processing resources within the auditory modality. (c) 2015 APA, all rights reserved).

  4. Adaptative synchronization in multi-output fractional-order complex dynamical networks and secure communications

    Science.gov (United States)

    Mata-Machuca, Juan L.; Aguilar-López, Ricardo

    2018-01-01

    This work deals with the adaptative synchronization of complex dynamical networks with fractional-order nodes and its application in secure communications employing chaotic parameter modulation. The complex network is composed of multiple fractional-order systems with mismatch parameters and the coupling functions are given to realize the network synchronization. We introduce a fractional algebraic synchronizability condition (FASC) and a fractional algebraic identifiability condition (FAIC) which are used to know if the synchronization and parameters estimation problems can be solved. To overcome these problems, an adaptative synchronization methodology is designed; the strategy consists in proposing multiple receiver systems which tend to follow asymptotically the uncertain transmitters systems. The coupling functions and parameters of the receiver systems are adjusted continually according to a convenient sigmoid-like adaptative controller (SLAC), until the measurable output errors converge to zero, hence, synchronization between transmitter and receivers is achieved and message signals are recovered. Indeed, the stability analysis of the synchronization error is based on the fractional Lyapunov direct method. Finally, numerical results corroborate the satisfactory performance of the proposed scheme by means of the synchronization of a complex network consisting of several fractional-order unified chaotic systems.

  5. New type of chaos synchronization in discrete-time systems: the F-M synchronization

    Directory of Open Access Journals (Sweden)

    Ouannas Adel

    2018-04-01

    Full Text Available In this paper, a new type of synchronization for chaotic (hyperchaotic maps with different dimensions is proposed. The novel scheme is called F – M synchronization, since it combines the inverse generalized synchronization (based on a functional relationship F with the matrix projective synchronization (based on a matrix M. In particular, the proposed approach enables F – M synchronization with index d to be achieved between n-dimensional drive system map and m-dimensional response system map, where the synchronization index d corresponds to the dimension of the synchronization error. The technique, which exploits nonlinear controllers and Lyapunov stability theory, proves to be effective in achieving the F – M synchronization not only when the synchronization index d equals n or m, but even if the synchronization index d is larger than the map dimensions n and m. Finally, simulation results are reported, with the aim to illustrate the capabilities of the novel scheme proposed herein.

  6. New type of chaos synchronization in discrete-time systems: the F-M synchronization

    Science.gov (United States)

    Ouannas, Adel; Grassi, Giuseppe; Karouma, Abdulrahman; Ziar, Toufik; Wang, Xiong; Pham, Viet-Thanh

    2018-04-01

    In this paper, a new type of synchronization for chaotic (hyperchaotic) maps with different dimensions is proposed. The novel scheme is called F - M synchronization, since it combines the inverse generalized synchronization (based on a functional relationship F) with the matrix projective synchronization (based on a matrix M). In particular, the proposed approach enables F - M synchronization with index d to be achieved between n-dimensional drive system map and m-dimensional response system map, where the synchronization index d corresponds to the dimension of the synchronization error. The technique, which exploits nonlinear controllers and Lyapunov stability theory, proves to be effective in achieving the F - M synchronization not only when the synchronization index d equals n or m, but even if the synchronization index d is larger than the map dimensions n and m. Finally, simulation results are reported, with the aim to illustrate the capabilities of the novel scheme proposed herein.

  7. Short-term plasticity in auditory cognition.

    Science.gov (United States)

    Jääskeläinen, Iiro P; Ahveninen, Jyrki; Belliveau, John W; Raij, Tommi; Sams, Mikko

    2007-12-01

    Converging lines of evidence suggest that auditory system short-term plasticity can enable several perceptual and cognitive functions that have been previously considered as relatively distinct phenomena. Here we review recent findings suggesting that auditory stimulation, auditory selective attention and cross-modal effects of visual stimulation each cause transient excitatory and (surround) inhibitory modulations in the auditory cortex. These modulations might adaptively tune hierarchically organized sound feature maps of the auditory cortex (e.g. tonotopy), thus filtering relevant sounds during rapidly changing environmental and task demands. This could support auditory sensory memory, pre-attentive detection of sound novelty, enhanced perception during selective attention, influence of visual processing on auditory perception and longer-term plastic changes associated with perceptual learning.

  8. Auditory Processing Disorder (For Parents)

    Science.gov (United States)

    ... role. Auditory cohesion problems: This is when higher-level listening tasks are difficult. Auditory cohesion skills — drawing inferences from conversations, understanding riddles, or comprehending verbal math problems — require heightened auditory processing and language levels. ...

  9. Analysis of electrically evoked compound action potential of the auditory nerve in children with bilateral cochlear implants.

    Science.gov (United States)

    Caldas, Fernanda Ferreira; Cardoso, Carolina Costa; Barreto, Monique Antunes de Souza Chelminski; Teixeira, Marina Santos; Hilgenberg, Anacléia Melo da Silva; Serra, Lucieny Silva Martins; Bahmad Junior, Fayez

    2016-01-01

    The cochlear implant device has the capacity to measure the electrically evoked compound action potential of the auditory nerve. The neural response telemetry is used in order to measure the electrically evoked compound action potential of the auditory nerve. To analyze the electrically evoked compound action potential, through the neural response telemetry, in children with bilateral cochlear implants. This is an analytical, prospective, longitudinal, historical cohort study. Six children, aged 1-4 years, with bilateral cochlear implant were assessed at five different intervals during their first year of cochlear implant use. There were significant differences in follow-up time (p=0.0082) and electrode position (p=0.0019) in the T-NRT measure. There was a significant difference in the interaction between time of follow-up and electrode position (p=0.0143) when measuring the N1-P1 wave amplitude between the three electrodes at each time of follow-up. The electrically evoked compound action potential measurement using neural response telemetry in children with bilateral cochlear implants during the first year of follow-up was effective in demonstrating the synchronized bilateral development of the peripheral auditory pathways in the studied population. Copyright © 2015 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.

  10. The Effect of Working Memory Training on Auditory Stream Segregation in Auditory Processing Disorders Children

    OpenAIRE

    Abdollah Moossavi; Saeideh Mehrkian; Yones Lotfi; Soghrat Faghih zadeh; Hamed Adjedi

    2015-01-01

    Objectives: This study investigated the efficacy of working memory training for improving working memory capacity and related auditory stream segregation in auditory processing disorders children. Methods: Fifteen subjects (9-11 years), clinically diagnosed with auditory processing disorder participated in this non-randomized case-controlled trial. Working memory abilities and auditory stream segregation were evaluated prior to beginning and six weeks after completing the training program...

  11. Dynamic Correlations between Intrinsic Connectivity and Extrinsic Connectivity of the Auditory Cortex in Humans

    Directory of Open Access Journals (Sweden)

    Zhuang Cui

    2017-08-01

    Full Text Available The arrival of sound signals in the auditory cortex (AC triggers both local and inter-regional signal propagations over time up to hundreds of milliseconds and builds up both intrinsic functional connectivity (iFC and extrinsic functional connectivity (eFC of the AC. However, interactions between iFC and eFC are largely unknown. Using intracranial stereo-electroencephalographic recordings in people with drug-refractory epilepsy, this study mainly investigated the temporal dynamic of the relationships between iFC and eFC of the AC. The results showed that a Gaussian wideband-noise burst markedly elicited potentials in both the AC and numerous higher-order cortical regions outside the AC (non-auditory cortices. Granger causality analyses revealed that in the earlier time window, iFC of the AC was positively correlated with both eFC from the AC to the inferior temporal gyrus and that to the inferior parietal lobule. While in later periods, the iFC of the AC was positively correlated with eFC from the precentral gyrus to the AC and that from the insula to the AC. In conclusion, dual-directional interactions occur between iFC and eFC of the AC at different time windows following the sound stimulation and may form the foundation underlying various central auditory processes, including auditory sensory memory, object formation, integrations between sensory, perceptional, attentional, motor, emotional, and executive processes.

  12. The effect of automatic blink correction on auditory evoked potentials.

    Science.gov (United States)

    Korpela, Jussi; Vigário, Ricardo; Huotilainen, Minna

    2012-01-01

    The effects of blink correction on auditory event-related potential (ERP) waveforms is assessed. Two blink correction strategies are compared. ICA-SSP combines independent component analysis (ICA) with signal space projection (SSP) and ICA-EMD uses empirical mode decomposition (EMD) to improve the performance of the standard ICA method. Five voluntary subjects performed an auditory oddball task. The resulting ERPs are used to compare the two blink correction methods to each other and against blink rejection. The results suggest that both methods qualitatively preserve the ERP waveform but that they underestimate some of the peak amplitudes. ICA-EMD performs slightly better than ICA-SSP. In conclusion, the use of blink correction is justified, especially if blink rejection leads to severe data loss.

  13. Lateralization of functional magnetic resonance imaging (fMRI) activation in the auditory pathway of patients with lateralized tinnitus

    Energy Technology Data Exchange (ETDEWEB)

    Smits, Marion [Erasmus MC - University Medical Center Rotterdam, Department of Radiology, Hs 224, Rotterdam (Netherlands); Kovacs, Silvia; Peeters, Ronald R; Hecke, Paul van; Sunaert, Stefan [University Hospitals of the Catholic University Leuven, Department of Radiology, Leuven (Belgium); Ridder, Dirk de [University of Antwerp, Department of Neurosurgery, Edegem (Belgium)

    2007-08-15

    Tinnitus is hypothesized to be an auditory phantom phenomenon resulting from spontaneous neuronal activity somewhere along the auditory pathway. We performed fMRI of the entire auditory pathway, including the inferior colliculus (IC), the medial geniculate body (MGB) and the auditory cortex (AC), in 42 patients with tinnitus and 10 healthy volunteers to assess lateralization of fMRI activation. Subjects were scanned on a 3T MRI scanner. A T2*-weighted EPI silent gap sequence was used during the stimulation paradigm, which consisted of a blocked design of 12 epochs in which music presented binaurally through headphones, which was switched on and off for periods of 50 s. Using SPM2 software, single subject and group statistical parametric maps were calculated. Lateralization of activation was assessed qualitatively and quantitatively. Tinnitus was lateralized in 35 patients (83%, 13 right-sided and 22 left-sided). Significant signal change (P{sub corrected} < 0.05) was found bilaterally in the primary and secondary AC, the IC and the MGB. Signal change was symmetrical in patients with bilateral tinnitus. In patients with lateralized tinnitus, fMRI activation was lateralized towards the side of perceived tinnitus in the primary AC and IC in patients with right-sided tinnitus, and in the MGB in patients with left-sided tinnitus. In healthy volunteers, activation in the primary AC was left-lateralized. Our paradigm adequately visualized the auditory pathways in tinnitus patients. In lateralized tinnitus fMRI activation was also lateralized, supporting the hypothesis that tinnitus is an auditory phantom phenomenon. (orig.)

  14. Lateralization of functional magnetic resonance imaging (fMRI) activation in the auditory pathway of patients with lateralized tinnitus

    International Nuclear Information System (INIS)

    Smits, Marion; Kovacs, Silvia; Peeters, Ronald R.; Hecke, Paul van; Sunaert, Stefan; Ridder, Dirk de

    2007-01-01

    Tinnitus is hypothesized to be an auditory phantom phenomenon resulting from spontaneous neuronal activity somewhere along the auditory pathway. We performed fMRI of the entire auditory pathway, including the inferior colliculus (IC), the medial geniculate body (MGB) and the auditory cortex (AC), in 42 patients with tinnitus and 10 healthy volunteers to assess lateralization of fMRI activation. Subjects were scanned on a 3T MRI scanner. A T2*-weighted EPI silent gap sequence was used during the stimulation paradigm, which consisted of a blocked design of 12 epochs in which music presented binaurally through headphones, which was switched on and off for periods of 50 s. Using SPM2 software, single subject and group statistical parametric maps were calculated. Lateralization of activation was assessed qualitatively and quantitatively. Tinnitus was lateralized in 35 patients (83%, 13 right-sided and 22 left-sided). Significant signal change (P corrected < 0.05) was found bilaterally in the primary and secondary AC, the IC and the MGB. Signal change was symmetrical in patients with bilateral tinnitus. In patients with lateralized tinnitus, fMRI activation was lateralized towards the side of perceived tinnitus in the primary AC and IC in patients with right-sided tinnitus, and in the MGB in patients with left-sided tinnitus. In healthy volunteers, activation in the primary AC was left-lateralized. Our paradigm adequately visualized the auditory pathways in tinnitus patients. In lateralized tinnitus fMRI activation was also lateralized, supporting the hypothesis that tinnitus is an auditory phantom phenomenon. (orig.)

  15. Synchronization in the Genesio Tesi and Coullet systems using the backstepping approach

    International Nuclear Information System (INIS)

    Hu, J-B; Han, Y; Zhao, L-D

    2008-01-01

    In this paper, the backstepping approach is proposed for synchronization in a pair of topologically inequivalent systems, the Genesio Tesi and Coullet systems. Firstly, the control problem for the chaos synchronization in the pair systems without unknown parameter is considered. Then an adaptive backstepping control law is designed to make the error signals between drive Genesio Tesi system and response Coullet system with three unknown parameters synchronized. The stability analysis in this article is proved according to a well-known Lyapunov stability theorem. These methods are applicable to a large class of topologically inequivalent systems where only a few algebraic inequalities are involved. Numerical simulation results are presented to show the effectiveness of the proposed scheme

  16. Synchronous Design and Test of Distributed Passive Radar Systems Based on Digital Broadcasting and Television

    Directory of Open Access Journals (Sweden)

    Wan Xianrong

    2017-02-01

    Full Text Available Digital broadcasting and television are important classes of illuminators of opportunity for passive radars. Distributed and multistatic structure are the development trends for passive radars. Most modern digital broadcasting and television systems work on a network, which not only provides a natural condition to distributed passive radar but also puts forward higher requirements on the design of passive radar systems. Among those requirements, precise synchronization among the receivers and transmitters as well as among multiple receiving stations, which mainly involves frequency and time synchronization, is the first to be solved. To satisfy the synchronization requirements of distributed passive radars, a synchronization scheme based on GPS is presented in this paper. Moreover, an effective scheme based on the China Mobile Multimedia Broadcasting signal is proposed to test the system synchronization performance. Finally, the reliability of the synchronization design is verified via the distributed multistatic passive radar experiments.

  17. On Synchronization Primitive Systems.

    Science.gov (United States)

    The report studies the question: what synchronization primitive should be used to handle inter-process communication. A formal model is presented...between these synchronization primitives. Although only four synchronization primitives are compared, the general methods can be used to compare other... synchronization primitives. Moreover, in the definitions of these synchronization primitives, conditional branches are explicitly allowed. In addition

  18. Objective measures of binaural masking level differences and comodulation masking release based on late auditory evoked potentials.

    Science.gov (United States)

    Epp, Bastian; Yasin, Ifat; Verhey, Jesko L

    2013-12-01

    The audibility of important sounds is often hampered due to the presence of other masking sounds. The present study investigates if a correlate of the audibility of a tone masked by noise is found in late auditory evoked potentials measured from human listeners. The audibility of the target sound at a fixed physical intensity is varied by introducing auditory cues of (i) interaural target signal phase disparity and (ii) coherent masker level fluctuations in different frequency regions. In agreement with previous studies, psychoacoustical experiments showed that both stimulus manipulations result in a masking release (i: binaural masking level difference; ii: comodulation masking release) compared to a condition where those cues are not present. Late auditory evoked potentials (N1, P2) were recorded for the stimuli at a constant masker level, but different signal levels within the same set of listeners who participated in the psychoacoustical experiment. The data indicate differences in N1 and P2 between stimuli with and without interaural phase disparities. However, differences for stimuli with and without coherent masker modulation were only found for P2, i.e., only P2 is sensitive to the increase in audibility, irrespective of the cue that caused the masking release. The amplitude of P2 is consistent with the psychoacoustical finding of an addition of the masking releases when both cues are present. Even though it cannot be concluded where along the auditory pathway the audibility is represented, the P2 component of auditory evoked potentials is a candidate for an objective measure of audibility in the human auditory system. Copyright © 2013 Elsevier B.V. All rights reserved.

  19. Synchronization of Switched Neural Networks With Communication Delays via the Event-Triggered Control.

    Science.gov (United States)

    Wen, Shiping; Zeng, Zhigang; Chen, Michael Z Q; Huang, Tingwen

    2017-10-01

    This paper addresses the issue of synchronization of switched delayed neural networks with communication delays via event-triggered control. For synchronizing coupled switched neural networks, we propose a novel event-triggered control law which could greatly reduce the number of control updates for synchronization tasks of coupled switched neural networks involving embedded microprocessors with limited on-board resources. The control signals are driven by properly defined events, which depend on the measurement errors and current-sampled states. By using a delay system method, a novel model of synchronization error system with delays is proposed with the communication delays and event-triggered control in the unified framework for coupled switched neural networks. The criteria are derived for the event-triggered synchronization analysis and control synthesis of switched neural networks via the Lyapunov-Krasovskii functional method and free weighting matrix approach. A numerical example is elaborated on to illustrate the effectiveness of the derived results.

  20. The role of the auditory brainstem in processing musically-relevant pitch

    Directory of Open Access Journals (Sweden)

    Gavin M. Bidelman

    2013-05-01

    Full Text Available Neuroimaging work has shed light on the cerebral architecture involved in processing the melodic and harmonic aspects of music. Here, recent evidence is reviewed illustrating that subcortical auditory structures contribute to the early formation and processing of musically-relevant pitch. Electrophysiological recordings from the human brainstem and population responses from the auditory nerve reveal that nascent features of tonal music (e.g., consonance/dissonance, pitch salience, harmonic sonority are evident at early, subcortical levels of the auditory pathway. The salience and harmonicity of brainstem activity is strongly correlated with listeners’ perceptual preferences and perceived consonance for the tonal relationships of music. Moreover, the hierarchical ordering of pitch intervals/chords described by the Western music practice and their perceptual consonance is well-predicted by the salience with which pitch combinations are encoded in subcortical auditory structures. While the neural correlates of consonance can be tuned and exaggerated with musical training, they persist even in the absence of musicianship or long-term enculturation. As such, it is posited that the structural foundations of musical pitch might result from innate processing performed by the central auditory system. A neurobiological predisposition for consonant, pleasant sounding pitch relationships may be one reason why these pitch combinations have been favored by composers and listeners for centuries. It is suggested that important perceptual dimensions of music emerge well before the auditory signal reaches cerebral cortex and prior to attentional engagement. While cortical mechanisms are no doubt critical to the perception, production, and enjoyment of music, the contribution of subcortical structures implicates a more integrated, hierarchically organized network underlying music processing within the brain.