WorldWideScience

Sample records for long-term asynchronous decoding

  1. Long-Term Asynchronous Decoding of Arm Motion Using Electrocorticographic Signals in Monkeys

    Science.gov (United States)

    Chao, Zenas C.; Nagasaka, Yasuo; Fujii, Naotaka

    2009-01-01

    Brain–machine interfaces (BMIs) employ the electrical activity generated by cortical neurons directly for controlling external devices and have been conceived as a means for restoring human cognitive or sensory-motor functions. The dominant approach in BMI research has been to decode motor variables based on single-unit activity (SUA). Unfortunately, this approach suffers from poor long-term stability and daily recalibration is normally required to maintain reliable performance. A possible alternative is BMIs based on electrocorticograms (ECoGs), which measure population activity and may provide more durable and stable recording. However, the level of long-term stability that ECoG-based decoding can offer remains unclear. Here we propose a novel ECoG-based decoding paradigm and show that we have successfully decoded hand positions and arm joint angles during an asynchronous food-reaching task in monkeys when explicit cues prompting the onset of movement were not required. Performance using our ECoG-based decoder was comparable to existing SUA-based systems while evincing far superior stability and durability. In addition, the same decoder could be used for months without any drift in accuracy or recalibration. These results were achieved by incorporating the spatio-spectro-temporal integration of activity across multiple cortical areas to compensate for the lower fidelity of ECoG signals. These results show the feasibility of high-performance, chronic and versatile ECoG-based neuroprosthetic devices for real-life applications. This new method provides a stable platform for investigating cortical correlates for understanding motor control, sensory perception, and high-level cognitive processes. PMID:20407639

  2. An Asynchronous Low Power and High Performance VLSI Architecture for Viterbi Decoder Implemented with Quasi Delay Insensitive Templates

    Directory of Open Access Journals (Sweden)

    T. Kalavathi Devi

    2015-01-01

    Full Text Available Convolutional codes are comprehensively used as Forward Error Correction (FEC codes in digital communication systems. For decoding of convolutional codes at the receiver end, Viterbi decoder is often used to have high priority. This decoder meets the demand of high speed and low power. At present, the design of a competent system in Very Large Scale Integration (VLSI technology requires these VLSI parameters to be finely defined. The proposed asynchronous method focuses on reducing the power consumption of Viterbi decoder for various constraint lengths using asynchronous modules. The asynchronous designs are based on commonly used Quasi Delay Insensitive (QDI templates, namely, Precharge Half Buffer (PCHB and Weak Conditioned Half Buffer (WCHB. The functionality of the proposed asynchronous design is simulated and verified using Tanner Spice (TSPICE in 0.25 µm, 65 nm, and 180 nm technologies of Taiwan Semiconductor Manufacture Company (TSMC. The simulation result illustrates that the asynchronous design techniques have 25.21% of power reduction compared to synchronous design and work at a speed of 475 MHz.

  3. Demonstration of asynchronous, 40 Gbps x 4-user DPSK-OCDMA transmission using a multi-port encoder/decoder.

    Science.gov (United States)

    Kataoka, Nobuyuki; Cincotti, Gabriella; Wada, Naoya; Kitayama, Ken-ichi

    2011-12-12

    We have developed a new 8-chip, 320 Gchip/s encoder/decoder with eight input/output ports, that can be used in 40-Gb/s PON networks. The device has been to multiplex four asynchronous 40 Gb/s users, using DPSK modulation. The transmission over 50 km has been successfully demonstrated for the first time. © 2011 Optical Society of America

  4. Algebraic Number Precoded OFDM Transmission for Asynchronous Cooperative Multirelay Networks

    Directory of Open Access Journals (Sweden)

    Hua Jiang

    2014-01-01

    Full Text Available This paper proposes a space-time block coding (STBC transmission scheme for asynchronous cooperative systems. By combination of rotated complex constellations and Hadamard transform, these constructed codes are capable of achieving full cooperative diversity with the analysis of the pairwise error probability (PEP. Due to the asynchronous characteristic of cooperative systems, orthogonal frequency division multiplexing (OFDM technique with cyclic prefix (CP is adopted for combating timing delays from relay nodes. The total transmit power across the entire network is fixed and appropriate power allocation can be implemented to optimize the network performance. The relay nodes do not require decoding and demodulation operation, resulting in a low complexity. Besides, there is no delay for forwarding the OFDM symbols to the destination node. At the destination node the received signals have the corresponding STBC structure on each subcarrier. In order to reduce the decoding complexity, the sphere decoder is implemented for fast data decoding. Bit error rate (BER performance demonstrates the effectiveness of the proposed scheme.

  5. High-throughput GPU-based LDPC decoding

    Science.gov (United States)

    Chang, Yang-Lang; Chang, Cheng-Chun; Huang, Min-Yu; Huang, Bormin

    2010-08-01

    Low-density parity-check (LDPC) code is a linear block code known to approach the Shannon limit via the iterative sum-product algorithm. LDPC codes have been adopted in most current communication systems such as DVB-S2, WiMAX, WI-FI and 10GBASE-T. LDPC for the needs of reliable and flexible communication links for a wide variety of communication standards and configurations have inspired the demand for high-performance and flexibility computing. Accordingly, finding a fast and reconfigurable developing platform for designing the high-throughput LDPC decoder has become important especially for rapidly changing communication standards and configurations. In this paper, a new graphic-processing-unit (GPU) LDPC decoding platform with the asynchronous data transfer is proposed to realize this practical implementation. Experimental results showed that the proposed GPU-based decoder achieved 271x speedup compared to its CPU-based counterpart. It can serve as a high-throughput LDPC decoder.

  6. Cellular automaton decoders of topological quantum memories in the fault tolerant setting

    International Nuclear Information System (INIS)

    Herold, Michael; Eisert, Jens; Kastoryano, Michael J; Campbell, Earl T

    2017-01-01

    Active error decoding and correction of topological quantum codes—in particular the toric code—remains one of the most viable routes to large scale quantum information processing. In contrast, passive error correction relies on the natural physical dynamics of a system to protect encoded quantum information. However, the search is ongoing for a completely satisfactory passive scheme applicable to locally interacting two-dimensional systems. Here, we investigate dynamical decoders that provide passive error correction by embedding the decoding process into local dynamics. We propose a specific discrete time cellular-automaton decoder in the fault tolerant setting and provide numerical evidence showing that the logical qubit has a survival time extended by several orders of magnitude over that of a bare unencoded qubit. We stress that (asynchronous) dynamical decoding gives rise to a Markovian dissipative process. We hence equate cellular-automaton decoding to a fully dissipative topological quantum memory, which removes errors continuously. In this sense, uncontrolled and unwanted local noise can be corrected for by a controlled local dissipative process. We analyze the required resources, commenting on additional polylogarithmic factors beyond those incurred by an ideal constant resource dynamical decoder. (paper)

  7. Harmonic Enhancement in Low Bitrate Audio Coding Using an Efficient Long-Term Predictor

    Directory of Open Access Journals (Sweden)

    Song Jeongook

    2010-01-01

    Full Text Available This paper proposes audio coding using an efficient long-term prediction method to enhance the perceptual quality of audio codecs to speech input signals at low bit-rates. The MPEG-4 AAC-LTP exploited a similar concept, but its improvement was not significant because of small prediction gain due to long prediction lags and aliased components caused by the transformation with a time-domain aliasing cancelation (TDAC technique. The proposed algorithm increases the prediction gain by employing a deharmonizing predictor and a long-term compensation filter. The look-back memory elements are first constructed by applying the de-harmonizing predictor to the input signal, then the prediction residual is encoded and decoded by transform audio coding. Finally, the long-term compensation filter is applied to the updated look-back memory of the decoded prediction residual to obtain synthesized signals. Experimental results show that the proposed algorithm has much lower spectral distortion and higher perceptual quality than conventional approaches especially for harmonic signals, such as voiced speech.

  8. Concatenated coding system with iterated sequential inner decoding

    DEFF Research Database (Denmark)

    Jensen, Ole Riis; Paaske, Erik

    1995-01-01

    We describe a concatenated coding system with iterated sequential inner decoding. The system uses convolutional codes of very long constraint length and operates on iterations between an inner Fano decoder and an outer Reed-Solomon decoder......We describe a concatenated coding system with iterated sequential inner decoding. The system uses convolutional codes of very long constraint length and operates on iterations between an inner Fano decoder and an outer Reed-Solomon decoder...

  9. The statistical significance of error probability as determined from decoding simulations for long codes

    Science.gov (United States)

    Massey, J. L.

    1976-01-01

    The very low error probability obtained with long error-correcting codes results in a very small number of observed errors in simulation studies of practical size and renders the usual confidence interval techniques inapplicable to the observed error probability. A natural extension of the notion of a 'confidence interval' is made and applied to such determinations of error probability by simulation. An example is included to show the surprisingly great significance of as few as two decoding errors in a very large number of decoding trials.

  10. Toric Codes, Multiplicative Structure and Decoding

    DEFF Research Database (Denmark)

    Hansen, Johan Peder

    2017-01-01

    Long linear codes constructed from toric varieties over finite fields, their multiplicative structure and decoding. The main theme is the inherent multiplicative structure on toric codes. The multiplicative structure allows for \\emph{decoding}, resembling the decoding of Reed-Solomon codes and al...

  11. Sequence-to-Sequence Prediction of Vehicle Trajectory via LSTM Encoder-Decoder Architecture

    OpenAIRE

    Park, Seong Hyeon; Kim, ByeongDo; Kang, Chang Mook; Chung, Chung Choo; Choi, Jun Won

    2018-01-01

    In this paper, we propose a deep learning based vehicle trajectory prediction technique which can generate the future trajectory sequence of surrounding vehicles in real time. We employ the encoder-decoder architecture which analyzes the pattern underlying in the past trajectory using the long short-term memory (LSTM) based encoder and generates the future trajectory sequence using the LSTM based decoder. This structure produces the $K$ most likely trajectory candidates over occupancy grid ma...

  12. CCS, locations and asynchronous transition systems

    DEFF Research Database (Denmark)

    Mukund, Madhavan; Nielsen, Mogens

    1992-01-01

    We provide a simple non-interleaved operational semantics for CCS in terms of asynchronous transition systems. We identify the concurrency present in the system in a natural way, in terms of events occurring at independent locations in the system. We extend the standard interleaving transition...... system for CCS by introducing labels on the transitions with information about the locations of events. We then show that the resulting transition system is an asynchronous transition system which has the additional property of being elementary, which means that it can also be represented by a 1-safe net....... We also introduce a notion of bisimulation on asynchronous transition systems which preserves independence. We conjecture that the induced equivalence on CCS processes coincides with the notion of location equivalence proposed by Boudol et al....

  13. Iterative Decoding of Concatenated Codes: A Tutorial

    Directory of Open Access Journals (Sweden)

    Phillip A. Regalia

    2005-05-01

    Full Text Available The turbo decoding algorithm of a decade ago constituted a milestone in error-correction coding for digital communications, and has inspired extensions to generalized receiver topologies, including turbo equalization, turbo synchronization, and turbo CDMA, among others. Despite an accrued understanding of iterative decoding over the years, the “turbo principle” remains elusive to master analytically, thereby inciting interest from researchers outside the communications domain. In this spirit, we develop a tutorial presentation of iterative decoding for parallel and serial concatenated codes, in terms hopefully accessible to a broader audience. We motivate iterative decoding as a computationally tractable attempt to approach maximum-likelihood decoding, and characterize fixed points in terms of a “consensus” property between constituent decoders. We review how the decoding algorithm for both parallel and serial concatenated codes coincides with an alternating projection algorithm, which allows one to identify conditions under which the algorithm indeed converges to a maximum-likelihood solution, in terms of particular likelihood functions factoring into the product of their marginals. The presentation emphasizes a common framework applicable to both parallel and serial concatenated codes.

  14. Modified Decoding Algorithm of LLR-SPA

    Directory of Open Access Journals (Sweden)

    Zhongxun Wang

    2014-09-01

    Full Text Available In wireless sensor networks, the energy consumption is mainly occurred in the stage of information transmission. The Low Density Parity Check code can make full use of the channel information to save energy. Because of the widely used decoding algorithm of the Low Density Parity Check code, this paper proposes a new decoding algorithm which is based on the LLR-SPA (Sum-Product Algorithm in Log-Likelihood-domain to improve the accuracy of the decoding algorithm. In the modified algorithm, a piecewise linear function is used to approximate the complicated Jacobi correction term in LLR-SPA decoding algorithm. Construct the tangent by the tangency point to the function of Jacobi correction term, which is based on the first order Taylor Series. In this way, the proposed piecewise linear approximation offers almost a perfect match to the function of Jacobi correction term. Meanwhile, the proposed piecewise linear approximation could avoid the operation of logarithmic which is more suitable for practical application. The simulation results show that the proposed algorithm could improve the decoding accuracy greatly without noticeable variation of the computational complexity.

  15. Generalized Asynchronous Systems

    Directory of Open Access Journals (Sweden)

    E. S. Kudryashova

    2012-01-01

    Full Text Available The paper consider a mathematical model of a concurrent system, the special case of which is an asynchronous system. Distributed asynchronous automata are introduced here. It is proved that Petri nets and transition systems with independence can be considered as distributed asynchronous automata. Time distributed asynchronous automata are defined in a standard way by correspondence which relates events with time intervals. It is proved that the time distributed asynchronous automata generalize time Petri nets and asynchronous systems.

  16. Decoding Illusory Self-location from Activity in the Human Hippocampus

    Directory of Open Access Journals (Sweden)

    Arvid eGuterstam

    2015-07-01

    Full Text Available Decades of research have demonstrated a role for the hippocampus in spatial navigation and episodic and spatial memory. However, empirical evidence linking hippocampal activity to the perceptual experience of being physically located at a particular place in the environment is lacking. In this study, we used a multisensory out-of-body illusion to perceptually ‘teleport’ six healthy participants between two different locations in the scanner room during high-resolution functional magnetic resonance imaging (fMRI. The participants were fitted with MRI-compatible head-mounted displays that changed their first-person visual perspective to that of a pair of cameras placed in one of two corners of the scanner room. To elicit the illusion of being physically located in this position, we delivered synchronous visuo-tactile stimulation in the form of an object moving towards the cameras coupled with touches applied to the participant’s chest. Asynchronous visuo-tactile stimulation did not induce the illusion and served as a control condition. We found that illusory self-location could be successfully decoded from patterns of activity in the hippocampus in all of the participants in the synchronous (P0.05. At the group-level, the decoding accuracy was significantly higher in the synchronous than in the asynchronous condition (P=0.012. These findings associate hippocampal activity with the perceived location of the bodily self in space, which suggests that the human hippocampus is involved not only in spatial navigation and memory but also in the construction of our sense of bodily self-location.

  17. Decoding the content of visual short-term memory under distraction in occipital and parietal areas.

    Science.gov (United States)

    Bettencourt, Katherine C; Xu, Yaoda

    2016-01-01

    Recent studies have provided conflicting accounts regarding where in the human brain visual short-term memory (VSTM) content is stored, with strong univariate fMRI responses being reported in superior intraparietal sulcus (IPS), but robust multivariate decoding being reported in occipital cortex. Given the continuous influx of information in everyday vision, VSTM storage under distraction is often required. We found that neither distractor presence nor predictability during the memory delay affected behavioral performance. Similarly, superior IPS exhibited consistent decoding of VSTM content across all distractor manipulations and had multivariate responses that closely tracked behavioral VSTM performance. However, occipital decoding of VSTM content was substantially modulated by distractor presence and predictability. Furthermore, we found no effect of target-distractor similarity on VSTM behavioral performance, further challenging the role of sensory regions in VSTM storage. Overall, consistent with previous univariate findings, our results indicate that superior IPS, but not occipital cortex, has a central role in VSTM storage.

  18. Decoding the content of visual short-term memory under distraction in occipital and parietal areas

    Science.gov (United States)

    Bettencourt, Katherine C.; Xu, Yaoda

    2015-01-01

    Recent studies have provided conflicting accounts regarding where in the human brain visual short-term memory (VSTM) content is stored, with strong univariate fMRI responses reported in superior intraparietal sulcus (IPS) but robust multivariate decoding reported in occipital cortex. Given the continuous influx of information in everyday vision, VSTM storage under distraction is often required. We found that neither distractor presence nor predictability during the memory delay affected behavioral performance. Similarly, superior IPS exhibited consistent decoding of VSTM content across all distractor manipulations and had multivariate responses that closely tracked behavioral VSTM performance. However, occipital decoding of VSTM content was significantly modulated by distractor presence and predictability. Furthermore, we found no effect of target-distractor similarity on VSTM behavioral performance, further challenging the role of sensory regions in VSTM storage. Overall, consistent with previous univariate findings, these results show that superior IPS, not occipital cortex, plays a central role in VSTM storage. PMID:26595654

  19. Decoding small surface codes with feedforward neural networks

    Science.gov (United States)

    Varsamopoulos, Savvas; Criger, Ben; Bertels, Koen

    2018-01-01

    Surface codes reach high error thresholds when decoded with known algorithms, but the decoding time will likely exceed the available time budget, especially for near-term implementations. To decrease the decoding time, we reduce the decoding problem to a classification problem that a feedforward neural network can solve. We investigate quantum error correction and fault tolerance at small code distances using neural network-based decoders, demonstrating that the neural network can generalize to inputs that were not provided during training and that they can reach similar or better decoding performance compared to previous algorithms. We conclude by discussing the time required by a feedforward neural network decoder in hardware.

  20. Asynchronous networks: modularization of dynamics theorem

    Science.gov (United States)

    Bick, Christian; Field, Michael

    2017-02-01

    Building on the first part of this paper, we develop the theory of functional asynchronous networks. We show that a large class of functional asynchronous networks can be (uniquely) represented as feedforward networks connecting events or dynamical modules. For these networks we can give a complete description of the network function in terms of the function of the events comprising the network: the modularization of dynamics theorem. We give examples to illustrate the main results.

  1. Full-asynchronous gigabit-symmetric DPSK downstream and OOK upstream OCDMA-PON with source-free ONUs employing all-optical self-clocked time gate.

    Science.gov (United States)

    Dai, Bo; Shimizu, Satoshi; Wang, Xu; Wada, Naoya

    2012-12-10

    We propose an asynchronous gigabit-symmetric optical code division multiplexing access passive optical network (OCDMA-PON) in which optical network units (ONUs) are source-free. In the experiment, we demonstrate a duplex OCDMA system with a 50 km 10 Gbit/s/user 4-user DPSK-OCDMA downlink and a 50 km 10 Gbit/s/user 4-user OOK-OCDMA uplink and error-free duplex transmissions are achieved. Besides, we investigate an all-optical self-clocked time gate, which is used for the signal regeneration of decoded signals and ensures asynchronization in the up/downstream transmissions. Furthermore, we evaluate the power budget of the proposed duplex transmission.

  2. SYMBOL LEVEL DECODING FOR DUO-BINARY TURBO CODES

    Directory of Open Access Journals (Sweden)

    Yogesh Beeharry

    2017-05-01

    Full Text Available This paper investigates the performance of three different symbol level decoding algorithms for Duo-Binary Turbo codes. Explicit details of the computations involved in the three decoding techniques, and a computational complexity analysis are given. Simulation results with different couple lengths, code-rates, and QPSK modulation reveal that the symbol level decoding with bit-level information outperforms the symbol level decoding by 0.1 dB on average in the error floor region. Moreover, a complexity analysis reveals that symbol level decoding with bit-level information reduces the decoding complexity by 19.6 % in terms of the total number of computations required for each half-iteration as compared to symbol level decoding.

  3. Relative phase asynchrony and long-range correlation of long-term solar magnetic activity

    Science.gov (United States)

    Deng, Linhua

    2017-07-01

    Statistical signal processing is one of the most important tasks in a large amount of areas of scientific studies, such as astrophysics, geophysics, and space physics. Phase recurrence analysis and long-range persistence are the two dynamical structures of the underlying processes for the given natural phenomenon. Linear and nonlinear time series analysis approaches (cross-correlation analysis, cross-recurrence plot, wavelet coherent transform, and Hurst analysis) are combined to investigate the relative phase interconnection and long-range correlation between solar activity and geomagnetic activity for the time interval from 1932 January to 2017 January. The following prominent results are found: (1) geomagnetic activity lags behind sunspot numbers with a phase shift of 21 months, and they have a high level of asynchronous behavior; (2) their relative phase interconnections are in phase for the periodic scales during 8-16 years, but have a mixing behavior for the periodic belts below 8 years; (3) both sunspot numbers and geomagnetic activity can not be regarded as a stochastic phenomenon because their dynamical behaviors display a long-term correlation and a fractal nature. We believe that the presented conclusions could provide further information on understanding the dynamical coupling of solar dynamo process with geomagnetic activity variation, and the crucial role of solar and geomagnetic activity in the long-term climate change.

  4. Functional asynchronous networks: Factorization of dynamics and function

    Directory of Open Access Journals (Sweden)

    Bick Christian

    2016-01-01

    Full Text Available In this note we describe the theory of functional asynchronous networks and one of the main results, the Modularization of Dynamics Theorem, which for a large class of functional asynchronous networks gives a factorization of dynamics in terms of constituent subnetworks. For these networks we can give a complete description of the network function in terms of the function of the events comprising the network and thereby answer a question originally raised by Alon in the context of biological networks.

  5. Automatic methods for long-term tracking and the detection and decoding of communication dances in honeybees

    Directory of Open Access Journals (Sweden)

    Fernando eWario

    2015-09-01

    Full Text Available The honeybee waggle dance communication system is an intriguing example of abstract animal communication and has been investigated thoroughly throughout the last seven decades. Typically, observables such as durations or angles are extracted manually directly from the observation hive or from video recordings to quantify dance properties, particularly to determine where bees have foraged. In recent years, biology has profited from automation, improving measurement precision, removing human bias, and accelerating data collection. As a further step, we have developed technologies to track all individuals of a honeybee colony and detect and decode communication dances automatically. In strong contrast to conventional approaches that focus on a small subset of the hive life, whether this regards time, space, or animal identity, our more inclusive system will help the understanding of the dance comprehensively in its spatial, temporal, and social context. In this contribution, we present full specifications of the recording setup and the software for automatic recognition and decoding of tags and dances, and we discuss potential research directions that may benefit from automation. Lastly, to exemplify the power of the methodology, we show experimental data and respective analyses for a continuous, experimental recording of nine weeks duration.

  6. Pseudo Asynchronous Level Crossing adc for ecg Signal Acquisition.

    Science.gov (United States)

    Marisa, T; Niederhauser, T; Haeberlin, A; Wildhaber, R A; Vogel, R; Goette, J; Jacomet, M

    2017-02-07

    A new pseudo asynchronous level crossing analogue-to-digital converter (adc) architecture targeted for low-power, implantable, long-term biomedical sensing applications is presented. In contrast to most of the existing asynchronous level crossing adc designs, the proposed design has no digital-to-analogue converter (dac) and no continuous time comparators. Instead, the proposed architecture uses an analogue memory cell and dynamic comparators. The architecture retains the signal activity dependent sampling operation by generating events only when the input signal is changing. The architecture offers the advantages of smaller chip area, energy saving and fewer analogue system components. Beside lower energy consumption the use of dynamic comparators results in a more robust performance in noise conditions. Moreover, dynamic comparators make interfacing the asynchronous level crossing system to synchronous processing blocks simpler. The proposed adc was implemented in [Formula: see text] complementary metal-oxide-semiconductor (cmos) technology, the hardware occupies a chip area of 0.0372 mm 2 and operates from a supply voltage of [Formula: see text] to [Formula: see text]. The adc's power consumption is as low as 0.6 μW with signal bandwidth from [Formula: see text] to [Formula: see text] and achieves an equivalent number of bits (enob) of up to 8 bits.

  7. Low latency asynchronous interface circuits

    Science.gov (United States)

    Sadowski, Greg

    2017-06-20

    In one form, a logic circuit includes an asynchronous logic circuit, a synchronous logic circuit, and an interface circuit coupled between the asynchronous logic circuit and the synchronous logic circuit. The asynchronous logic circuit has a plurality of asynchronous outputs for providing a corresponding plurality of asynchronous signals. The synchronous logic circuit has a plurality of synchronous inputs corresponding to the plurality of asynchronous outputs, a stretch input for receiving a stretch signal, and a clock output for providing a clock signal. The synchronous logic circuit provides the clock signal as a periodic signal but prolongs a predetermined state of the clock signal while the stretch signal is active. The asynchronous interface detects whether metastability could occur when latching any of the plurality of the asynchronous outputs of the asynchronous logic circuit using said clock signal, and activates the stretch signal while the metastability could occur.

  8. On Lattice Sequential Decoding for The Unconstrained AWGN Channel

    KAUST Repository

    Abediseid, Walid

    2012-10-01

    In this paper, the performance limits and the computational complexity of the lattice sequential decoder are analyzed for the unconstrained additive white Gaussian noise channel. The performance analysis available in the literature for such a channel has been studied only under the use of the minimum Euclidean distance decoder that is commonly referred to as the lattice decoder. Lattice decoders based on solutions to the NP-hard closest vector problem are very complex to implement, and the search for low complexity receivers for the detection of lattice codes is considered a challenging problem. However, the low computational complexity advantage that sequential decoding promises, makes it an alternative solution to the lattice decoder. In this work, we characterize the performance and complexity tradeoff via the error exponent and the decoding complexity, respectively, of such a decoder as a function of the decoding parameter --- the bias term. For the above channel, we derive the cut-off volume-to-noise ratio that is required to achieve a good error performance with low decoding complexity.

  9. On Lattice Sequential Decoding for The Unconstrained AWGN Channel

    KAUST Repository

    Abediseid, Walid

    2013-04-04

    In this paper, the performance limits and the computational complexity of the lattice sequential decoder are analyzed for the unconstrained additive white Gaussian noise channel. The performance analysis available in the literature for such a channel has been studied only under the use of the minimum Euclidean distance decoder that is commonly referred to as the \\\\textit{lattice decoder}. Lattice decoders based on solutions to the NP-hard closest vector problem are very complex to implement, and the search for low complexity receivers for the detection of lattice codes is considered a challenging problem. However, the low computational complexity advantage that sequential decoding promises, makes it an alternative solution to the lattice decoder. In this work, we characterize the performance and complexity tradeoff via the error exponent and the decoding complexity, respectively, of such a decoder as a function of the decoding parameter --- the bias term. For the above channel, we derive the cut-off volume-to-noise ratio that is required to achieve a good error performance with low decoding complexity.

  10. On Lattice Sequential Decoding for The Unconstrained AWGN Channel

    KAUST Repository

    Abediseid, Walid; Alouini, Mohamed-Slim

    2012-01-01

    In this paper, the performance limits and the computational complexity of the lattice sequential decoder are analyzed for the unconstrained additive white Gaussian noise channel. The performance analysis available in the literature for such a channel has been studied only under the use of the minimum Euclidean distance decoder that is commonly referred to as the lattice decoder. Lattice decoders based on solutions to the NP-hard closest vector problem are very complex to implement, and the search for low complexity receivers for the detection of lattice codes is considered a challenging problem. However, the low computational complexity advantage that sequential decoding promises, makes it an alternative solution to the lattice decoder. In this work, we characterize the performance and complexity tradeoff via the error exponent and the decoding complexity, respectively, of such a decoder as a function of the decoding parameter --- the bias term. For the above channel, we derive the cut-off volume-to-noise ratio that is required to achieve a good error performance with low decoding complexity.

  11. Relative Ease in Creating Detailed Orthographic Representations Contrasted with Severe Difficulties to Maintain Them in Long-term Memory Among Dyslexic Children.

    Science.gov (United States)

    Binamé, Florence; Danzio, Sophie; Poncelet, Martine

    2015-11-01

    Most research into orthographic learning abilities has been conducted in English with typically developing children using reading-based tasks. In the present study, we examined the abilities of French-speaking children with dyslexia to create novel orthographic representations for subsequent use in spelling and to maintain them in long-term memory. Their performance was compared with that of chronological age (CA)-matched and reading age (RA)-matched control children. We used an experimental task designed to provide optimal learning conditions (i.e. 10 spelling practice trials) ensuring the short-term acquisition of the spelling of the target orthographic word forms. After a 1-week delay, the long-term retention of the targets was assessed by a spelling post-test. Analysis of the results revealed that, in the short term, children with dyslexia learned the novel orthographic word forms well, only differing from both CA and RA controls on the initial decoding of the targets and from CA controls on the first two practice trials. In contrast, a dramatic drop was observed in their long-term retention relative to CA and RA controls. These results support the suggestion of the self-teaching hypothesis (Share, 1995) that initial errors in the decoding and spelling of unfamiliar words may hinder the establishment of fully specified novel orthographic representations. Copyright © 2015 John Wiley & Sons, Ltd.

  12. Encoder-decoder optimization for brain-computer interfaces.

    Science.gov (United States)

    Merel, Josh; Pianto, Donald M; Cunningham, John P; Paninski, Liam

    2015-06-01

    Neuroprosthetic brain-computer interfaces are systems that decode neural activity into useful control signals for effectors, such as a cursor on a computer screen. It has long been recognized that both the user and decoding system can adapt to increase the accuracy of the end effector. Co-adaptation is the process whereby a user learns to control the system in conjunction with the decoder adapting to learn the user's neural patterns. We provide a mathematical framework for co-adaptation and relate co-adaptation to the joint optimization of the user's control scheme ("encoding model") and the decoding algorithm's parameters. When the assumptions of that framework are respected, co-adaptation cannot yield better performance than that obtainable by an optimal initial choice of fixed decoder, coupled with optimal user learning. For a specific case, we provide numerical methods to obtain such an optimized decoder. We demonstrate our approach in a model brain-computer interface system using an online prosthesis simulator, a simple human-in-the-loop pyschophysics setup which provides a non-invasive simulation of the BCI setting. These experiments support two claims: that users can learn encoders matched to fixed, optimal decoders and that, once learned, our approach yields expected performance advantages.

  13. Encoder-decoder optimization for brain-computer interfaces.

    Directory of Open Access Journals (Sweden)

    Josh Merel

    2015-06-01

    Full Text Available Neuroprosthetic brain-computer interfaces are systems that decode neural activity into useful control signals for effectors, such as a cursor on a computer screen. It has long been recognized that both the user and decoding system can adapt to increase the accuracy of the end effector. Co-adaptation is the process whereby a user learns to control the system in conjunction with the decoder adapting to learn the user's neural patterns. We provide a mathematical framework for co-adaptation and relate co-adaptation to the joint optimization of the user's control scheme ("encoding model" and the decoding algorithm's parameters. When the assumptions of that framework are respected, co-adaptation cannot yield better performance than that obtainable by an optimal initial choice of fixed decoder, coupled with optimal user learning. For a specific case, we provide numerical methods to obtain such an optimized decoder. We demonstrate our approach in a model brain-computer interface system using an online prosthesis simulator, a simple human-in-the-loop pyschophysics setup which provides a non-invasive simulation of the BCI setting. These experiments support two claims: that users can learn encoders matched to fixed, optimal decoders and that, once learned, our approach yields expected performance advantages.

  14. Asynchronous variational integration using continuous assumed gradient elements.

    Science.gov (United States)

    Wolff, Sebastian; Bucher, Christian

    2013-03-01

    Asynchronous variational integration (AVI) is a tool which improves the numerical efficiency of explicit time stepping schemes when applied to finite element meshes with local spatial refinement. This is achieved by associating an individual time step length to each spatial domain. Furthermore, long-term stability is ensured by its variational structure. This article presents AVI in the context of finite elements based on a weakened weak form (W2) Liu (2009) [1], exemplified by continuous assumed gradient elements Wolff and Bucher (2011) [2]. The article presents the main ideas of the modified AVI, gives implementation notes and a recipe for estimating the critical time step.

  15. The role of short-term memory impairment in nonword repetition, real word repetition, and nonword decoding: A case study.

    Science.gov (United States)

    Peter, Beate

    2018-01-01

    In a companion study, adults with dyslexia and adults with a probable history of childhood apraxia of speech showed evidence of difficulty with processing sequential information during nonword repetition, multisyllabic real word repetition and nonword decoding. Results suggested that some errors arose in visual encoding during nonword reading, all levels of processing but especially short-term memory storage/retrieval during nonword repetition, and motor planning and programming during complex real word repetition. To further investigate the role of short-term memory, a participant with short-term memory impairment (MI) was recruited. MI was confirmed with poor performance during a sentence repetition and three nonword repetition tasks, all of which have a high short-term memory load, whereas typical performance was observed during tests of reading, spelling, and static verbal knowledge, all with low short-term memory loads. Experimental results show error-free performance during multisyllabic real word repetition but high counts of sequence errors, especially migrations and assimilations, during nonword repetition, supporting short-term memory as a locus of sequential processing deficit during nonword repetition. Results are also consistent with the hypothesis that during complex real word repetition, short-term memory is bypassed as the word is recognized and retrieved from long-term memory prior to producing the word.

  16. Decoding intention at sensorimotor timescales.

    Directory of Open Access Journals (Sweden)

    Mathew Salvaris

    Full Text Available The ability to decode an individual's intentions in real time has long been a 'holy grail' of research on human volition. For example, a reliable method could be used to improve scientific study of voluntary action by allowing external probe stimuli to be delivered at different moments during development of intention and action. Several Brain Computer Interface applications have used motor imagery of repetitive actions to achieve this goal. These systems are relatively successful, but only if the intention is sustained over a period of several seconds; much longer than the timescales identified in psychophysiological studies for normal preparation for voluntary action. We have used a combination of sensorimotor rhythms and motor imagery training to decode intentions in a single-trial cued-response paradigm similar to those used in human and non-human primate motor control research. Decoding accuracy of over 0.83 was achieved with twelve participants. With this approach, we could decode intentions to move the left or right hand at sub-second timescales, both for instructed choices instructed by an external stimulus and for free choices generated intentionally by the participant. The implications for volition are considered.

  17. Locally decodable codes and private information retrieval schemes

    CERN Document Server

    Yekhanin, Sergey

    2010-01-01

    Locally decodable codes (LDCs) are codes that simultaneously provide efficient random access retrieval and high noise resilience by allowing reliable reconstruction of an arbitrary bit of a message by looking at only a small number of randomly chosen codeword bits. Local decodability comes with a certain loss in terms of efficiency - specifically, locally decodable codes require longer codeword lengths than their classical counterparts. Private information retrieval (PIR) schemes are cryptographic protocols designed to safeguard the privacy of database users. They allow clients to retrieve rec

  18. The Effects of Musical Training on the Decoding Skills of German-Speaking Primary School Children

    Science.gov (United States)

    Rautenberg, Iris

    2015-01-01

    This paper outlines the results of a long-term study of 159 German-speaking primary school children. The correlations between musical skills (perception and differentiation of rhythmical and tonal/melodic patterns) and decoding skills, and the effects of musical training on word-level reading abilities were investigated. Cognitive skills and…

  19. Video encoder/decoder for encoding/decoding motion compensated images

    NARCIS (Netherlands)

    1996-01-01

    Video encoder and decoder, provided with a motion compensator for motion-compensated video coding or decoding in which a picture is coded or decoded in blocks in alternately horizontal and vertical steps. The motion compensator is provided with addressing means (160) and controlled multiplexers

  20. Firing rate estimation using infinite mixture models and its application to neural decoding.

    Science.gov (United States)

    Shibue, Ryohei; Komaki, Fumiyasu

    2017-11-01

    Neural decoding is a framework for reconstructing external stimuli from spike trains recorded by various neural recordings. Kloosterman et al. proposed a new decoding method using marked point processes (Kloosterman F, Layton SP, Chen Z, Wilson MA. J Neurophysiol 111: 217-227, 2014). This method does not require spike sorting and thereby improves decoding accuracy dramatically. In this method, they used kernel density estimation to estimate intensity functions of marked point processes. However, the use of kernel density estimation causes problems such as low decoding accuracy and high computational costs. To overcome these problems, we propose a new decoding method using infinite mixture models to estimate intensity. The proposed method improves decoding performance in terms of accuracy and computational speed. We apply the proposed method to simulation and experimental data to verify its performance. NEW & NOTEWORTHY We propose a new neural decoding method using infinite mixture models and nonparametric Bayesian statistics. The proposed method improves decoding performance in terms of accuracy and computation speed. We have successfully applied the proposed method to position decoding from spike trains recorded in a rat hippocampus. Copyright © 2017 the American Physiological Society.

  1. On minimizing the maximum broadcast decoding delay for instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.

    2014-09-01

    In this paper, we consider the problem of minimizing the maximum broadcast decoding delay experienced by all the receivers of generalized instantly decodable network coding (IDNC). Unlike the sum decoding delay, the maximum decoding delay as a definition of delay for IDNC allows a more equitable distribution of the delays between the different receivers and thus a better Quality of Service (QoS). In order to solve this problem, we first derive the expressions for the probability distributions of maximum decoding delay increments. Given these expressions, we formulate the problem as a maximum weight clique problem in the IDNC graph. Although this problem is known to be NP-hard, we design a greedy algorithm to perform effective packet selection. Through extensive simulations, we compare the sum decoding delay and the max decoding delay experienced when applying the policies to minimize the sum decoding delay and our policy to reduce the max decoding delay. Simulations results show that our policy gives a good agreement among all the delay aspects in all situations and outperforms the sum decoding delay policy to effectively minimize the sum decoding delay when the channel conditions become harsher. They also show that our definition of delay significantly improve the number of served receivers when they are subject to strict delay constraints.

  2. Dynamics of intracellular information decoding.

    Science.gov (United States)

    Kobayashi, Tetsuya J; Kamimura, Atsushi

    2011-10-01

    A variety of cellular functions are robust even to substantial intrinsic and extrinsic noise in intracellular reactions and the environment that could be strong enough to impair or limit them. In particular, of substantial importance is cellular decision-making in which a cell chooses a fate or behavior on the basis of information conveyed in noisy external signals. For robust decoding, the crucial step is filtering out the noise inevitably added during information transmission. As a minimal and optimal implementation of such an information decoding process, the autocatalytic phosphorylation and autocatalytic dephosphorylation (aPadP) cycle was recently proposed. Here, we analyze the dynamical properties of the aPadP cycle in detail. We describe the dynamical roles of the stationary and short-term responses in determining the efficiency of information decoding and clarify the optimality of the threshold value of the stationary response and its information-theoretical meaning. Furthermore, we investigate the robustness of the aPadP cycle against the receptor inactivation time and intrinsic noise. Finally, we discuss the relationship among information decoding with information-dependent actions, bet-hedging and network modularity.

  3. Dynamics of intracellular information decoding

    International Nuclear Information System (INIS)

    Kobayashi, Tetsuya J; Kamimura, Atsushi

    2011-01-01

    A variety of cellular functions are robust even to substantial intrinsic and extrinsic noise in intracellular reactions and the environment that could be strong enough to impair or limit them. In particular, of substantial importance is cellular decision-making in which a cell chooses a fate or behavior on the basis of information conveyed in noisy external signals. For robust decoding, the crucial step is filtering out the noise inevitably added during information transmission. As a minimal and optimal implementation of such an information decoding process, the autocatalytic phosphorylation and autocatalytic dephosphorylation (aPadP) cycle was recently proposed. Here, we analyze the dynamical properties of the aPadP cycle in detail. We describe the dynamical roles of the stationary and short-term responses in determining the efficiency of information decoding and clarify the optimality of the threshold value of the stationary response and its information-theoretical meaning. Furthermore, we investigate the robustness of the aPadP cycle against the receptor inactivation time and intrinsic noise. Finally, we discuss the relationship among information decoding with information-dependent actions, bet-hedging and network modularity

  4. Image transmission system using adaptive joint source and channel decoding

    Science.gov (United States)

    Liu, Weiliang; Daut, David G.

    2005-03-01

    In this paper, an adaptive joint source and channel decoding method is designed to accelerate the convergence of the iterative log-dimain sum-product decoding procedure of LDPC codes as well as to improve the reconstructed image quality. Error resilience modes are used in the JPEG2000 source codec, which makes it possible to provide useful source decoded information to the channel decoder. After each iteration, a tentative decoding is made and the channel decoded bits are then sent to the JPEG2000 decoder. Due to the error resilience modes, some bits are known to be either correct or in error. The positions of these bits are then fed back to the channel decoder. The log-likelihood ratios (LLR) of these bits are then modified by a weighting factor for the next iteration. By observing the statistics of the decoding procedure, the weighting factor is designed as a function of the channel condition. That is, for lower channel SNR, a larger factor is assigned, and vice versa. Results show that the proposed joint decoding methods can greatly reduce the number of iterations, and thereby reduce the decoding delay considerably. At the same time, this method always outperforms the non-source controlled decoding method up to 5dB in terms of PSNR for various reconstructed images.

  5. On minimizing the maximum broadcast decoding delay for instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.; Sorour, Sameh; Alouini, Mohamed-Slim; Ai-Naffouri, Tareq Y.

    2014-01-01

    In this paper, we consider the problem of minimizing the maximum broadcast decoding delay experienced by all the receivers of generalized instantly decodable network coding (IDNC). Unlike the sum decoding delay, the maximum decoding delay as a

  6. Performance-complexity tradeoff in sequential decoding for the unconstrained AWGN channel

    KAUST Repository

    Abediseid, Walid

    2013-06-01

    In this paper, the performance limits and the computational complexity of the lattice sequential decoder are analyzed for the unconstrained additive white Gaussian noise channel. The performance analysis available in the literature for such a channel has been studied only under the use of the minimum Euclidean distance decoder that is commonly referred to as the lattice decoder. Lattice decoders based on solutions to the NP-hard closest vector problem are very complex to implement, and the search for low complexity receivers for the detection of lattice codes is considered a challenging problem. However, the low computational complexity advantage that sequential decoding promises, makes it an alternative solution to the lattice decoder. In this work, we characterize the performance and complexity tradeoff via the error exponent and the decoding complexity, respectively, of such a decoder as a function of the decoding parameter - the bias term. For the above channel, we derive the cut-off volume-to-noise ratio that is required to achieve a good error performance with low decoding complexity. © 2013 IEEE.

  7. Decoding vigilance with NIRS.

    Science.gov (United States)

    Bogler, Carsten; Mehnert, Jan; Steinbrink, Jens; Haynes, John-Dylan

    2014-01-01

    Sustained, long-term cognitive workload is associated with variations and decrements in performance. Such fluctuations in vigilance can be a risk factor especially during dangerous attention demanding activities. Functional MRI studies have shown that attentional performance is correlated with BOLD-signals, especially in parietal and prefrontal cortical regions. An interesting question is whether these BOLD-signals could be measured in real-world scenarios, say to warn in a dangerous workplace whenever a subjects' vigilance is low. Because fMRI lacks the mobility needed for such applications, we tested whether the monitoring of vigilance might be possible using Near-Infrared Spectroscopy (NIRS). NIRS is a highly mobile technique that measures hemodynamics in the surface of the brain. We demonstrate that non-invasive NIRS signals correlate with vigilance. These signals carry enough information to decode subjects' reaction times at a single trial level.

  8. Serotonin and Dopamine Gene Variation and Theory of Mind Decoding Accuracy in Major Depression: A Preliminary Investigation.

    Science.gov (United States)

    Zahavi, Arielle Y; Sabbagh, Mark A; Washburn, Dustin; Mazurka, Raegan; Bagby, R Michael; Strauss, John; Kennedy, James L; Ravindran, Arun; Harkness, Kate L

    2016-01-01

    Theory of mind-the ability to decode and reason about others' mental states-is a universal human skill and forms the basis of social cognition. Theory of mind accuracy is impaired in clinical conditions evidencing social impairment, including major depressive disorder. The current study is a preliminary investigation of the association of polymorphisms of the serotonin transporter (SLC6A4), dopamine transporter (DAT1), dopamine receptor D4 (DRD4), and catechol-O-methyl transferase (COMT) genes with theory of mind decoding in a sample of adults with major depression. Ninety-six young adults (38 depressed, 58 non-depressed) completed the 'Reading the Mind in the Eyes task' and a non-mentalistic control task. Genetic associations were only found for the depressed group. Specifically, superior accuracy in decoding mental states of a positive valence was seen in those homozygous for the long allele of the serotonin transporter gene, 9-allele carriers of DAT1, and long-allele carriers of DRD4. In contrast, superior accuracy in decoding mental states of a negative valence was seen in short-allele carriers of the serotonin transporter gene and 10/10 homozygotes of DAT1. Results are discussed in terms of their implications for integrating social cognitive and neurobiological models of etiology in major depression.

  9. Serotonin and Dopamine Gene Variation and Theory of Mind Decoding Accuracy in Major Depression: A Preliminary Investigation.

    Directory of Open Access Journals (Sweden)

    Arielle Y Zahavi

    Full Text Available Theory of mind-the ability to decode and reason about others' mental states-is a universal human skill and forms the basis of social cognition. Theory of mind accuracy is impaired in clinical conditions evidencing social impairment, including major depressive disorder. The current study is a preliminary investigation of the association of polymorphisms of the serotonin transporter (SLC6A4, dopamine transporter (DAT1, dopamine receptor D4 (DRD4, and catechol-O-methyl transferase (COMT genes with theory of mind decoding in a sample of adults with major depression. Ninety-six young adults (38 depressed, 58 non-depressed completed the 'Reading the Mind in the Eyes task' and a non-mentalistic control task. Genetic associations were only found for the depressed group. Specifically, superior accuracy in decoding mental states of a positive valence was seen in those homozygous for the long allele of the serotonin transporter gene, 9-allele carriers of DAT1, and long-allele carriers of DRD4. In contrast, superior accuracy in decoding mental states of a negative valence was seen in short-allele carriers of the serotonin transporter gene and 10/10 homozygotes of DAT1. Results are discussed in terms of their implications for integrating social cognitive and neurobiological models of etiology in major depression.

  10. Performance breakdown in optimal stimulus decoding.

    Science.gov (United States)

    Lubomir Kostal; Lansky, Petr; Pilarski, Stevan

    2015-06-01

    One of the primary goals of neuroscience is to understand how neurons encode and process information about their environment. The problem is often approached indirectly by examining the degree to which the neuronal response reflects the stimulus feature of interest. In this context, the methods of signal estimation and detection theory provide the theoretical limits on the decoding accuracy with which the stimulus can be identified. The Cramér-Rao lower bound on the decoding precision is widely used, since it can be evaluated easily once the mathematical model of the stimulus-response relationship is determined. However, little is known about the behavior of different decoding schemes with respect to the bound if the neuronal population size is limited. We show that under broad conditions the optimal decoding displays a threshold-like shift in performance in dependence on the population size. The onset of the threshold determines a critical range where a small increment in size, signal-to-noise ratio or observation time yields a dramatic gain in the decoding precision. We demonstrate the existence of such threshold regions in early auditory and olfactory information coding. We discuss the origin of the threshold effect and its impact on the design of effective coding approaches in terms of relevant population size.

  11. Decoding the non-stationary neuron spike trains by dual Monte Carlo point process estimation in motor Brain Machine Interfaces.

    Science.gov (United States)

    Liao, Yuxi; Li, Hongbao; Zhang, Qiaosheng; Fan, Gong; Wang, Yiwen; Zheng, Xiaoxiang

    2014-01-01

    Decoding algorithm in motor Brain Machine Interfaces translates the neural signals to movement parameters. They usually assume the connection between the neural firings and movements to be stationary, which is not true according to the recent studies that observe the time-varying neuron tuning property. This property results from the neural plasticity and motor learning etc., which leads to the degeneration of the decoding performance when the model is fixed. To track the non-stationary neuron tuning during decoding, we propose a dual model approach based on Monte Carlo point process filtering method that enables the estimation also on the dynamic tuning parameters. When applied on both simulated neural signal and in vivo BMI data, the proposed adaptive method performs better than the one with static tuning parameters, which raises a promising way to design a long-term-performing model for Brain Machine Interfaces decoder.

  12. A method for the automated long-term monitoring of three-spined stickleback Gasterosteus aculeatus shoal dynamics.

    Science.gov (United States)

    Kleinhappel, T K; Al-Zoubi, A; Al-Diri, B; Burman, O; Dickinson, P; John, L; Wilkinson, A; Pike, T W

    2014-04-01

    This paper describes and evaluates a flexible, non-invasive tagging system for the automated identification and long-term monitoring of individual three-spined sticklebacks Gasterosteus aculeatus. The system is based on barcoded tags, which can be reliably and robustly detected and decoded to provide information on an individual's identity and location. Because large numbers of fish can be individually tagged, it can be used to monitor individual- and group-level dynamics within fish shoals. © 2014 The Fisheries Society of the British Isles.

  13. Fast Reed-Solomon Decoder

    Science.gov (United States)

    Liu, K. Y.

    1986-01-01

    High-speed decoder intended for use with Reed-Solomon (RS) codes of long code length and high error-correcting capability. Design based on algorithm that includes high-radix Fermat transform procedure, which is most efficient for high speeds. RS code in question has code-word length of 256 symbols, of which 224 are information symbols and 32 are redundant.

  14. Decoding of interleaved Reed-Solomon codes using improved power decoding

    DEFF Research Database (Denmark)

    Puchinger, Sven; Rosenkilde ne Nielsen, Johan

    2017-01-01

    We propose a new partial decoding algorithm for m-interleaved Reed-Solomon (IRS) codes that can decode, with high probability, a random error of relative weight 1 − Rm/m+1 at all code rates R, in time polynomial in the code length n. For m > 2, this is an asymptotic improvement over the previous...... state-of-the-art for all rates, and the first improvement for R > 1/3 in the last 20 years. The method combines collaborative decoding of IRS codes with power decoding up to the Johnson radius....

  15. Low-Power Bitstream-Residual Decoder for H.264/AVC Baseline Profile Decoding

    Directory of Open Access Journals (Sweden)

    Xu Ke

    2009-01-01

    Full Text Available Abstract We present the design and VLSI implementation of a novel low-power bitstream-residual decoder for H.264/AVC baseline profile. It comprises a syntax parser, a parameter decoder, and an Inverse Quantization Inverse Transform (IQIT decoder. The syntax parser detects and decodes each incoming codeword in the bitstream under the control of a hierarchical Finite State Machine (FSM; the IQIT decoder performs inverse transform and quantization with pipelining and parallelism. Various power reduction techniques, such as data-driven based on statistic results, nonuniform partition, precomputation, guarded evaluation, hierarchical FSM decomposition, TAG method, zero-block skipping, and clock gating , are adopted and integrated throughout the bitstream-residual decoder. With innovative architecture, the proposed design is able to decode QCIF video sequences of 30 fps at a clock rate as low as 1.5 MHz. A prototype H.264/AVC baseline decoding chip utilizing the proposed decoder is fabricated in UMC 0.18  m 1P6M CMOS technology. The proposed design is measured under 1 V 1.8 V supply with 0.1 V step. It dissipates 76  W at 1 V and 253  W at 1.8 V.

  16. Real-time SHVC software decoding with multi-threaded parallel processing

    Science.gov (United States)

    Gudumasu, Srinivas; He, Yuwen; Ye, Yan; He, Yong; Ryu, Eun-Seok; Dong, Jie; Xiu, Xiaoyu

    2014-09-01

    This paper proposes a parallel decoding framework for scalable HEVC (SHVC). Various optimization technologies are implemented on the basis of SHVC reference software SHM-2.0 to achieve real-time decoding speed for the two layer spatial scalability configuration. SHVC decoder complexity is analyzed with profiling information. The decoding process at each layer and the up-sampling process are designed in parallel and scheduled by a high level application task manager. Within each layer, multi-threaded decoding is applied to accelerate the layer decoding speed. Entropy decoding, reconstruction, and in-loop processing are pipeline designed with multiple threads based on groups of coding tree units (CTU). A group of CTUs is treated as a processing unit in each pipeline stage to achieve a better trade-off between parallelism and synchronization. Motion compensation, inverse quantization, and inverse transform modules are further optimized with SSE4 SIMD instructions. Simulations on a desktop with an Intel i7 processor 2600 running at 3.4 GHz show that the parallel SHVC software decoder is able to decode 1080p spatial 2x at up to 60 fps (frames per second) and 1080p spatial 1.5x at up to 50 fps for those bitstreams generated with SHVC common test conditions in the JCT-VC standardization group. The decoding performance at various bitrates with different optimization technologies and different numbers of threads are compared in terms of decoding speed and resource usage, including processor and memory.

  17. Decoding vigilance with NIRS.

    Directory of Open Access Journals (Sweden)

    Carsten Bogler

    Full Text Available Sustained, long-term cognitive workload is associated with variations and decrements in performance. Such fluctuations in vigilance can be a risk factor especially during dangerous attention demanding activities. Functional MRI studies have shown that attentional performance is correlated with BOLD-signals, especially in parietal and prefrontal cortical regions. An interesting question is whether these BOLD-signals could be measured in real-world scenarios, say to warn in a dangerous workplace whenever a subjects' vigilance is low. Because fMRI lacks the mobility needed for such applications, we tested whether the monitoring of vigilance might be possible using Near-Infrared Spectroscopy (NIRS. NIRS is a highly mobile technique that measures hemodynamics in the surface of the brain. We demonstrate that non-invasive NIRS signals correlate with vigilance. These signals carry enough information to decode subjects' reaction times at a single trial level.

  18. Novel Simplified Model for Asynchronous Machine with Consideration of Frequency Characteristic

    Directory of Open Access Journals (Sweden)

    Changchun Cai

    2014-01-01

    Full Text Available The frequency characteristic of electric equipment should be considered in the digital simulation of power systems. The traditional asynchronous machine third-order transient model excludes not only the stator transient but also the frequency characteristics, thus decreasing the application sphere of the model and resulting in a large error under some special conditions. Based on the physical equivalent circuit and Park model for asynchronous machines, this study proposes a novel asynchronous third-order transient machine model with consideration of the frequency characteristic. In the new definitions of variables, the voltages behind the reactance are redefined as the linear equation of flux linkage. In this way, the rotor voltage equation is not associated with the derivative terms of frequency. However, the derivative terms of frequency should not always be ignored in the application of the traditional third-order transient model. Compared with the traditional third-order transient model, the novel simplified third-order transient model with consideration of the frequency characteristic is more accurate without increasing the order and complexity. Simulation results show that the novel third-order transient model for the asynchronous machine is suitable and effective and is more accurate than the widely used traditional simplified third-order transient model under some special conditions with drastic frequency fluctuations.

  19. LONG-TERM EVOLUTION OF DOUBLE WHITE DWARF BINARIES ACCRETING THROUGH DIRECT IMPACT

    International Nuclear Information System (INIS)

    Kremer, Kyle; Kalogera, Vassiliki; Sepinsky, Jeremy

    2015-01-01

    We calculate the long-term evolution of angular momentum in double white dwarf binaries undergoing direct impact accretion over a broad range of parameter space. We allow the rotation rate of both components to vary and account for the exchange of angular momentum between the spins of the white dwarfs and the orbit, while conserving the total angular momentum. We include gravitational, tidal, and mass transfer effects in the orbital evolution, and allow the Roche radius of the donor star to vary with both the stellar mass and the rotation rate. We examine the long-term stability of these systems, focusing in particular on those systems that may be progenitors of AM CVn or SNe Ia. We find that our analysis yields an increase in the predicted number of stable systems compared to that in previous studies. Additionally, we find that by properly accounting for the effects of asynchronism between the donor and the orbit on the Roche-lobe size, we eliminate oscillations in the orbital parameters, which were found in previous studies. Removing these oscillations can reduce the peak mass transfer rate in some systems, keeping them from entering an unstable mass transfer phase

  20. Completion time reduction in instantly decodable network coding through decoding delay control

    KAUST Repository

    Douik, Ahmed S.; Sorour, Sameh; Alouini, Mohamed-Slim; Al-Naffouri, Tareq Y.

    2014-01-01

    For several years, the completion time and the decoding delay problems in Instantly Decodable Network Coding (IDNC) were considered separately and were thought to completely act against each other. Recently, some works aimed to balance the effects of these two important IDNC metrics but none of them studied a further optimization of one by controlling the other. In this paper, we study the effect of controlling the decoding delay to reduce the completion time below its currently best known solution. We first derive the decoding-delay-dependent expressions of the users' and their overall completion times. Although using such expressions to find the optimal overall completion time is NP-hard, we use a heuristic that minimizes the probability of increasing the maximum of these decoding-delay-dependent completion time expressions after each transmission through a layered control of their decoding delays. Simulation results show that this new algorithm achieves both a lower mean completion time and mean decoding delay compared to the best known heuristic for completion time reduction. The gap in performance becomes significant for harsh erasure scenarios.

  1. Completion time reduction in instantly decodable network coding through decoding delay control

    KAUST Repository

    Douik, Ahmed S.

    2014-12-01

    For several years, the completion time and the decoding delay problems in Instantly Decodable Network Coding (IDNC) were considered separately and were thought to completely act against each other. Recently, some works aimed to balance the effects of these two important IDNC metrics but none of them studied a further optimization of one by controlling the other. In this paper, we study the effect of controlling the decoding delay to reduce the completion time below its currently best known solution. We first derive the decoding-delay-dependent expressions of the users\\' and their overall completion times. Although using such expressions to find the optimal overall completion time is NP-hard, we use a heuristic that minimizes the probability of increasing the maximum of these decoding-delay-dependent completion time expressions after each transmission through a layered control of their decoding delays. Simulation results show that this new algorithm achieves both a lower mean completion time and mean decoding delay compared to the best known heuristic for completion time reduction. The gap in performance becomes significant for harsh erasure scenarios.

  2. Implementation of Layered Decoding Architecture for LDPC Code using Layered Min-Sum Algorithm

    OpenAIRE

    Sandeep Kakde; Atish Khobragade; Shrikant Ambatkar; Pranay Nandanwar

    2017-01-01

    For binary field and long code lengths, Low Density Parity Check (LDPC) code approaches Shannon limit performance. LDPC codes provide remarkable error correction performance and therefore enlarge the design space for communication systems.In this paper, we have compare different digital modulation techniques and found that BPSK modulation technique is better than other modulation techniques in terms of BER. It also gives error performance of LDPC decoder over AWGN channel using Min-Sum algori...

  3. Decoding Codes on Graphs

    Indian Academy of Sciences (India)

    Shannon limit of the channel. Among the earliest discovered codes that approach the. Shannon limit were the low density parity check (LDPC) codes. The term low density arises from the property of the parity check matrix defining the code. We will now define this matrix and the role that it plays in decoding. 2. Linear Codes.

  4. Within-Category Decoding of Information in Different Attentional States in Short-Term Memory.

    Science.gov (United States)

    LaRocque, Joshua J; Riggall, Adam C; Emrich, Stephen M; Postle, Bradley R

    2017-10-01

    A long-standing assumption of cognitive neuroscience has been that working memory (WM) is accomplished by sustained, elevated neural activity. More recently, theories of WM have expanded this view by describing different attentional states in WM with differing activation levels. Several studies have used multivariate pattern analysis (MVPA) of functional magnetic resonance imaging (fMRI) and electroencephalography (EEG) data to study neural activity corresponding to these WM states. Intriguingly, no evidence was found for active neural representations for information held in WM outside the focus of attention ("unattended memory items," UMIs), suggesting that only attended memory items (AMIs) are accompanied by an active trace. However, these results depended on category-level decoding, which lacks sensitivity to neural representations of individual items. Therefore, we employed a WM task in which subjects remembered the directions of motion of two dot arrays, with a retrocue indicating which was relevant for an imminent memory probe (the AMI). This design allowed MVPA decoding of delay-period fMRI signal at the stimulus-item level, affording a more sensitive test of the neural representation of UMIs. Whereas evidence for the AMI was reliably high, evidence for the UMI dropped to baseline, consistent with the notion that different WM attentional states may have qualitatively different mechanisms of retention. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  5. Progressive Image Transmission Based on Joint Source-Channel Decoding Using Adaptive Sum-Product Algorithm

    Directory of Open Access Journals (Sweden)

    David G. Daut

    2007-03-01

    Full Text Available A joint source-channel decoding method is designed to accelerate the iterative log-domain sum-product decoding procedure of LDPC codes as well as to improve the reconstructed image quality. Error resilience modes are used in the JPEG2000 source codec making it possible to provide useful source decoded information to the channel decoder. After each iteration, a tentative decoding is made and the channel decoded bits are then sent to the JPEG2000 decoder. The positions of bits belonging to error-free coding passes are then fed back to the channel decoder. The log-likelihood ratios (LLRs of these bits are then modified by a weighting factor for the next iteration. By observing the statistics of the decoding procedure, the weighting factor is designed as a function of the channel condition. Results show that the proposed joint decoding methods can greatly reduce the number of iterations, and thereby reduce the decoding delay considerably. At the same time, this method always outperforms the nonsource controlled decoding method by up to 3 dB in terms of PSNR.

  6. Progressive Image Transmission Based on Joint Source-Channel Decoding Using Adaptive Sum-Product Algorithm

    Directory of Open Access Journals (Sweden)

    Liu Weiliang

    2007-01-01

    Full Text Available A joint source-channel decoding method is designed to accelerate the iterative log-domain sum-product decoding procedure of LDPC codes as well as to improve the reconstructed image quality. Error resilience modes are used in the JPEG2000 source codec making it possible to provide useful source decoded information to the channel decoder. After each iteration, a tentative decoding is made and the channel decoded bits are then sent to the JPEG2000 decoder. The positions of bits belonging to error-free coding passes are then fed back to the channel decoder. The log-likelihood ratios (LLRs of these bits are then modified by a weighting factor for the next iteration. By observing the statistics of the decoding procedure, the weighting factor is designed as a function of the channel condition. Results show that the proposed joint decoding methods can greatly reduce the number of iterations, and thereby reduce the decoding delay considerably. At the same time, this method always outperforms the nonsource controlled decoding method by up to 3 dB in terms of PSNR.

  7. Iterative List Decoding

    DEFF Research Database (Denmark)

    Justesen, Jørn; Høholdt, Tom; Hjaltason, Johan

    2005-01-01

    We analyze the relation between iterative decoding and the extended parity check matrix. By considering a modified version of bit flipping, which produces a list of decoded words, we derive several relations between decodable error patterns and the parameters of the code. By developing a tree...... of codewords at minimal distance from the received vector, we also obtain new information about the code....

  8. Asynchronous LMS adaptive equalization

    NARCIS (Netherlands)

    Bergmans, J.W.M.; Lin, M.Y.; Modrie, D.; Otte, R.

    2005-01-01

    Digital data receivers often operate at a fixed sampling rate 1/Ts that is asynchronous to the baud rate 1/T. A digital equalizer that processes the incoming signal will also operate in the asynchronous clock domain. Existing adaptation techniques for this equalizer involve an error sequence ek that

  9. Adaptive decoding of convolutional codes

    Science.gov (United States)

    Hueske, K.; Geldmacher, J.; Götze, J.

    2007-06-01

    Convolutional codes, which are frequently used as error correction codes in digital transmission systems, are generally decoded using the Viterbi Decoder. On the one hand the Viterbi Decoder is an optimum maximum likelihood decoder, i.e. the most probable transmitted code sequence is obtained. On the other hand the mathematical complexity of the algorithm only depends on the used code, not on the number of transmission errors. To reduce the complexity of the decoding process for good transmission conditions, an alternative syndrome based decoder is presented. The reduction of complexity is realized by two different approaches, the syndrome zero sequence deactivation and the path metric equalization. The two approaches enable an easy adaptation of the decoding complexity for different transmission conditions, which results in a trade-off between decoding complexity and error correction performance.

  10. Asynchronous design of Networks-on-Chip

    DEFF Research Database (Denmark)

    Sparsø, Jens

    2007-01-01

    -synchronous, mesochronous, globally-asynchronous locally-synchronous and fully asynchronous), discusses the circuitry needed to implement these timing methodologies, and provides some implementation details for a couple of asynchronous NoCs designed at the Technical University of Denmark (DTU). The paper is written...... to support an invited talk at the NORCHIP’2007 conference....

  11. Decoding Xing-Ling codes

    DEFF Research Database (Denmark)

    Nielsen, Rasmus Refslund

    2002-01-01

    This paper describes an efficient decoding method for a recent construction of good linear codes as well as an extension to the construction. Furthermore, asymptotic properties and list decoding of the codes are discussed.......This paper describes an efficient decoding method for a recent construction of good linear codes as well as an extension to the construction. Furthermore, asymptotic properties and list decoding of the codes are discussed....

  12. Adaptive decoding of convolutional codes

    Directory of Open Access Journals (Sweden)

    K. Hueske

    2007-06-01

    Full Text Available Convolutional codes, which are frequently used as error correction codes in digital transmission systems, are generally decoded using the Viterbi Decoder. On the one hand the Viterbi Decoder is an optimum maximum likelihood decoder, i.e. the most probable transmitted code sequence is obtained. On the other hand the mathematical complexity of the algorithm only depends on the used code, not on the number of transmission errors. To reduce the complexity of the decoding process for good transmission conditions, an alternative syndrome based decoder is presented. The reduction of complexity is realized by two different approaches, the syndrome zero sequence deactivation and the path metric equalization. The two approaches enable an easy adaptation of the decoding complexity for different transmission conditions, which results in a trade-off between decoding complexity and error correction performance.

  13. Implementation of Layered Decoding Architecture for LDPC Code using Layered Min-Sum Algorithm

    Directory of Open Access Journals (Sweden)

    Sandeep Kakde

    2017-12-01

    Full Text Available For binary field and long code lengths, Low Density Parity Check (LDPC code approaches Shannon limit performance. LDPC codes provide remarkable error correction performance and therefore enlarge the design space for communication systems.In this paper, we have compare different digital modulation techniques and found that BPSK modulation technique is better than other modulation techniques in terms of BER. It also gives error performance of LDPC decoder over AWGN channel using Min-Sum algorithm. VLSI Architecture is proposed which uses the value re-use property of min-sum algorithm and gives high throughput. The proposed work has been implemented and tested on Xilinx Virtex 5 FPGA. The MATLAB result of LDPC decoder for low bit error rate (BER gives bit error rate in the range of 10-1 to 10-3.5 at SNR=1 to 2 for 20 no of iterations. So it gives good bit error rate performance. The latency of the parallel design of LDPC decoder has also reduced. It has accomplished 141.22 MHz maximum frequency and throughput of 2.02 Gbps while consuming less area of the design.

  14. Behavioral synthesis of asynchronous circuits

    DEFF Research Database (Denmark)

    Nielsen, Sune Fallgaard

    2005-01-01

    This thesis presents a method for behavioral synthesis of asynchronous circuits, which aims at providing a synthesis flow which uses and tranfers methods from synchronous circuits to asynchronous circuits. We move the synchronous behavioral synthesis abstraction into the asynchronous handshake...... is idle. This reduces unnecessary switching activity in the individual functional units and therefore the energy consumption of the entire circuit. A collection of behavioral synthesis algorithms have been developed allowing the designer to perform time and power constrained design space exploration...

  15. Decoding Facial Expressions: A New Test with Decoding Norms.

    Science.gov (United States)

    Leathers, Dale G.; Emigh, Ted H.

    1980-01-01

    Describes the development and testing of a new facial meaning sensitivity test designed to determine how specialized are the meanings that can be decoded from facial expressions. Demonstrates the use of the test to measure a receiver's current level of skill in decoding facial expressions. (JMF)

  16. Comparing face-to-face, synchronous, and asynchronous learning: postgraduate dental resident preferences.

    Science.gov (United States)

    Kunin, Marc; Julliard, Kell N; Rodriguez, Tobias E

    2014-06-01

    The Department of Dental Medicine of Lutheran Medical Center has developed an asynchronous online curriculum consisting of prerecorded PowerPoint presentations with audio explanations. The focus of this study was to evaluate if the new asynchronous format satisfied the educational needs of the residents compared to traditional lecture (face-to-face) and synchronous (distance learning) formats. Lectures were delivered to 219 dental residents employing face-to-face and synchronous formats, as well as the new asynchronous format; 169 (77 percent) participated in the study. Outcomes were assessed with pretests, posttests, and individual lecture surveys. Results found the residents preferred face-to-face and asynchronous formats to the synchronous format in terms of effectiveness and clarity of presentations. This preference was directly related to the residents' perception of how well the technology worked in each format. The residents also rated the quality of student-instructor and student-student interactions in the synchronous and asynchronous formats significantly higher after taking the lecture series than they did before taking it. However, they rated the face-to-face format as significantly more conducive to student-instructor and student-student interaction. While the study found technology had a major impact on the efficacy of this curricular model, the results suggest that the asynchronous format can be an effective way to teach a postgraduate course.

  17. Forced Sequence Sequential Decoding

    DEFF Research Database (Denmark)

    Jensen, Ole Riis; Paaske, Erik

    1998-01-01

    We describe a new concatenated decoding scheme based on iterations between an inner sequentially decoded convolutional code of rate R=1/4 and memory M=23, and block interleaved outer Reed-Solomon (RS) codes with nonuniform profile. With this scheme decoding with good performance is possible as low...... as Eb/N0=0.6 dB, which is about 1.25 dB below the signal-to-noise ratio (SNR) that marks the cutoff rate for the full system. Accounting for about 0.45 dB due to the outer codes, sequential decoding takes place at about 1.7 dB below the SNR cutoff rate for the convolutional code. This is possible since...... the iteration process provides the sequential decoders with side information that allows a smaller average load and minimizes the probability of computational overflow. Analytical results for the probability that the first RS word is decoded after C computations are presented. These results are supported...

  18. Delay Estimation in Long-Code Asynchronous DS/CDMA Systems Using Multiple Antennas

    Directory of Open Access Journals (Sweden)

    Sirbu Marius

    2004-01-01

    Full Text Available The problem of propagation delay estimation in asynchronous long-code DS-CDMA multiuser systems is addressed. Almost all the methods proposed so far in the literature for propagation delay estimation are derived for short codes and the knowledge of the codes is exploited by the estimators. In long-code CDMA, the spreading code is aperiodic and the methods developed for short codes may not be used or may increase the complexity significantly. For example, in the subspace-based estimators, the aperiodic nature of the code may require subspace tracking. In this paper we propose a novel method for simultaneous estimation of the propagation delays of several active users. A specific multiple-input multiple-output (MIMO system model is constructed in a multiuser scenario. In such model the channel matrix contains information about both the users propagation delays and channel impulse responses. Consequently, estimates of the delays are obtained as a by-product of the channel estimation task. The channel matrix has a special structure that is exploited in estimating the delays. The proposed delay estimation method lends itself to an adaptive implementation. Thus, it may be applied to joint channel and delay estimation in uplink DS-CDMA analogously to the method presented by the authors in 2003. The performance of the proposed method is studied in simulation using realistic time-varying channel model and different SNR levels in the face of near-far effects, and using low spreading factor (high data rates.

  19. Cascading Oscillators in Decoding Speech: Reflection of a Cortical Computation Principle

    Science.gov (United States)

    2016-09-06

    purely) auditory or articulatory model can explain this behavior. The insertion of gaps was interpreted as the act of providing extra decoding time...The windows are generated by a segmentation process, implemented by an array of cascaded oscillators. Correct segmentation is a critical...prerequisite for correct decoding, and segmentation is correct as long as the oscillators successfully track the input rhythms. Syllabic segmentation utilizes

  20. Performance analysis of multiple interference suppression over asynchronous/synchronous optical code-division multiple-access system based on complementary/prime/shifted coding scheme

    Science.gov (United States)

    Nieh, Ta-Chun; Yang, Chao-Chin; Huang, Jen-Fa

    2011-08-01

    A complete complementary/prime/shifted prime (CPS) code family for the optical code-division multiple-access (OCDMA) system is proposed. Based on the ability of complete complementary (CC) code, the multiple-access interference (MAI) can be suppressed and eliminated via spectral amplitude coding (SAC) OCDMA system under asynchronous/synchronous transmission. By utilizing the shifted prime (SP) code in the SAC scheme, the hardware implementation of encoder/decoder can be simplified with a reduced number of optical components, such as arrayed waveguide grating (AWG) and fiber Bragg grating (FBG). This system has a superior performance as compared to previous bipolar-bipolar coding OCDMA systems.

  1. Pro asynchronous programming with .NET

    CERN Document Server

    Blewett, Richard; Ltd, Rock Solid Knowledge

    2014-01-01

    Pro Asynchronous Programming with .NET teaches the essential skill of asynchronous programming in .NET. It answers critical questions in .NET application development, such as: how do I keep my program responding at all times to keep my users happy how do I make the most of the available hardware how can I improve performanceIn the modern world, users expect more and more from their applications and devices, and multi-core hardware has the potential to provide it. But it takes carefully crafted code to turn that potential into responsive, scalable applications.With Pro Asynchronous Programming

  2. Asynchronized synchronous machines

    CERN Document Server

    Botvinnik, M M

    1964-01-01

    Asynchronized Synchronous Machines focuses on the theoretical research on asynchronized synchronous (AS) machines, which are "hybrids” of synchronous and induction machines that can operate with slip. Topics covered in this book include the initial equations; vector diagram of an AS machine; regulation in cases of deviation from the law of full compensation; parameters of the excitation system; and schematic diagram of an excitation regulator. The possible applications of AS machines and its calculations in certain cases are also discussed. This publication is beneficial for students and indiv

  3. Error characterization for asynchronous computations: Proxy equation approach

    Science.gov (United States)

    Sallai, Gabriella; Mittal, Ankita; Girimaji, Sharath

    2017-11-01

    Numerical techniques for asynchronous fluid flow simulations are currently under development to enable efficient utilization of massively parallel computers. These numerical approaches attempt to accurately solve time evolution of transport equations using spatial information at different time levels. The truncation error of asynchronous methods can be divided into two parts: delay dependent (EA) or asynchronous error and delay independent (ES) or synchronous error. The focus of this study is a specific asynchronous error mitigation technique called proxy-equation approach. The aim of this study is to examine these errors as a function of the characteristic wavelength of the solution. Mitigation of asynchronous effects requires that the asynchronous error be smaller than synchronous truncation error. For a simple convection-diffusion equation, proxy-equation error analysis identifies critical initial wave-number, λc. At smaller wave numbers, synchronous error are larger than asynchronous errors. We examine various approaches to increase the value of λc in order to improve the range of applicability of proxy-equation approach.

  4. Forced Sequence Sequential Decoding

    DEFF Research Database (Denmark)

    Jensen, Ole Riis

    In this thesis we describe a new concatenated decoding scheme based on iterations between an inner sequentially decoded convolutional code of rate R=1/4 and memory M=23, and block interleaved outer Reed-Solomon codes with non-uniform profile. With this scheme decoding with good performance...... is possible as low as Eb/No=0.6 dB, which is about 1.7 dB below the signal-to-noise ratio that marks the cut-off rate for the convolutional code. This is possible since the iteration process provides the sequential decoders with side information that allows a smaller average load and minimizes the probability...... of computational overflow. Analytical results for the probability that the first Reed-Solomon word is decoded after C computations are presented. This is supported by simulation results that are also extended to other parameters....

  5. On Decoding Interleaved Chinese Remainder Codes

    DEFF Research Database (Denmark)

    Li, Wenhui; Sidorenko, Vladimir; Nielsen, Johan Sebastian Rosenkilde

    2013-01-01

    We model the decoding of Interleaved Chinese Remainder codes as that of finding a short vector in a Z-lattice. Using the LLL algorithm, we obtain an efficient decoding algorithm, correcting errors beyond the unique decoding bound and having nearly linear complexity. The algorithm can fail...... with a probability dependent on the number of errors, and we give an upper bound for this. Simulation results indicate that the bound is close to the truth. We apply the proposed decoding algorithm for decoding a single CR code using the idea of “Power” decoding, suggested for Reed-Solomon codes. A combination...... of these two methods can be used to decode low-rate Interleaved Chinese Remainder codes....

  6. Asynchronous zero-forcing adaptive equalization

    NARCIS (Netherlands)

    Bergmans, J.W.M.; Pozidis, H.; Lin, M.Y.

    2005-01-01

    Digital data receivers often operate at a fixed sampling rate 1/Ts that is asynchronous to the baud rate 1/T. A digital equalizer that processes the incoming signal will also be asynchronous, and its adaptation is commonly based on extensions of the LMS algorithm. In this paper, we develop and

  7. Trellises and Trellis-Based Decoding Algorithms for Linear Block Codes

    Science.gov (United States)

    Lin, Shu

    1998-01-01

    sectionalization of trellises. Chapter 7 discusses trellis decomposition and subtrellises for low-weight codewords. Chapter 8 first presents well known methods for constructing long powerful codes from short component codes or component codes of smaller dimensions, and then provides methods for constructing their trellises which include Shannon and Cartesian product techniques. Chapter 9 deals with convolutional codes, puncturing, zero-tail termination and tail-biting.Chapters 10 through 13 present various trellis-based decoding algorithms, old and new. Chapter 10 first discusses the application of the well known Viterbi decoding algorithm to linear block codes, optimum sectionalization of a code trellis to minimize computation complexity, and design issues for IC (integrated circuit) implementation of a Viterbi decoder. Then it presents a new decoding algorithm for convolutional codes, named Differential Trellis Decoding (DTD) algorithm. Chapter 12 presents a suboptimum reliability-based iterative decoding algorithm with a low-weight trellis search for the most likely codeword. This decoding algorithm provides a good trade-off between error performance and decoding complexity. All the decoding algorithms presented in Chapters 10 through 12 are devised to minimize word error probability. Chapter 13 presents decoding algorithms that minimize bit error probability and provide the corresponding soft (reliability) information at the output of the decoder. Decoding algorithms presented are the MAP (maximum a posteriori probability) decoding algorithm and the Soft-Output Viterbi Algorithm (SOVA) algorithm. Finally, the minimization of bit error probability in trellis-based MLD is discussed.

  8. List Decoding of Algebraic Codes

    DEFF Research Database (Denmark)

    Nielsen, Johan Sebastian Rosenkilde

    We investigate three paradigms for polynomial-time decoding of Reed–Solomon codes beyond half the minimum distance: the Guruswami–Sudan algorithm, Power decoding and the Wu algorithm. The main results concern shaping the computational core of all three methods to a problem solvable by module...... Hermitian codes using Guruswami–Sudan or Power decoding faster than previously known, and we show how to Wu list decode binary Goppa codes....... to solve such using module minimisation, or using our new Demand–Driven algorithm which is also based on module minimisation. The decoding paradigms are all derived and analysed in a self-contained manner, often in new ways or examined in greater depth than previously. Among a number of new results, we...

  9. Decoding Delay Controlled Completion Time Reduction in Instantly Decodable Network Coding

    KAUST Repository

    Douik, Ahmed

    2016-06-27

    For several years, the completion time and the decoding delay problems in Instantly Decodable Network Coding (IDNC) were considered separately and were thought to act completely against each other. Recently, some works aimed to balance the effects of these two important IDNC metrics but none of them studied a further optimization of one by controlling the other. This paper investigates the effect of controlling the decoding delay to reduce the completion time below its currently best-known solution in both perfect and imperfect feedback with persistent erasure channels. To solve the problem, the decodingdelay- dependent expressions of the users’ and overall completion times are derived in the complete feedback scenario. Although using such expressions to find the optimal overall completion time is NP-hard, the paper proposes two novel heuristics that minimizes the probability of increasing the maximum of these decoding-delay-dependent completion time expressions after each transmission through a layered control of their decoding delays. Afterward, the paper extends the study to the imperfect feedback scenario in which uncertainties at the sender affects its ability to anticipate accurately the decoding delay increase at each user. The paper formulates the problem in such environment and derives the expression of the minimum increase in the completion time. Simulation results show the performance of the proposed solutions and suggest that both heuristics achieves a lower mean completion time as compared to the best-known heuristics for the completion time reduction in perfect and imperfect feedback. The gap in performance becomes more significant as the erasure of the channel increases.

  10. Decoding Delay Controlled Completion Time Reduction in Instantly Decodable Network Coding

    KAUST Repository

    Douik, Ahmed S.; Sorour, Sameh; Al-Naffouri, Tareq Y.; Alouini, Mohamed-Slim

    2016-01-01

    For several years, the completion time and the decoding delay problems in Instantly Decodable Network Coding (IDNC) were considered separately and were thought to act completely against each other. Recently, some works aimed to balance the effects of these two important IDNC metrics but none of them studied a further optimization of one by controlling the other. This paper investigates the effect of controlling the decoding delay to reduce the completion time below its currently best-known solution in both perfect and imperfect feedback with persistent erasure channels. To solve the problem, the decodingdelay- dependent expressions of the users’ and overall completion times are derived in the complete feedback scenario. Although using such expressions to find the optimal overall completion time is NP-hard, the paper proposes two novel heuristics that minimizes the probability of increasing the maximum of these decoding-delay-dependent completion time expressions after each transmission through a layered control of their decoding delays. Afterward, the paper extends the study to the imperfect feedback scenario in which uncertainties at the sender affects its ability to anticipate accurately the decoding delay increase at each user. The paper formulates the problem in such environment and derives the expression of the minimum increase in the completion time. Simulation results show the performance of the proposed solutions and suggest that both heuristics achieves a lower mean completion time as compared to the best-known heuristics for the completion time reduction in perfect and imperfect feedback. The gap in performance becomes more significant as the erasure of the channel increases.

  11. The Aeolian Asynchronous Generator

    Directory of Open Access Journals (Sweden)

    Ionel Dragomirescu

    2008-10-01

    Full Text Available The production of the electric energy with lower costs could be realized with the help of the aeolian electric central. In these centrals we can use the squirrel cage asynchronous generators, because these machines are the most safety in function and easy exploited. This work show the function analyzing of the asynchronous generator having on involving torque depending on the square wind speed, the air-density and on the construction of the wing spiral.

  12. Neural decoding of visual imagery during sleep.

    Science.gov (United States)

    Horikawa, T; Tamaki, M; Miyawaki, Y; Kamitani, Y

    2013-05-03

    Visual imagery during sleep has long been a topic of persistent speculation, but its private nature has hampered objective analysis. Here we present a neural decoding approach in which machine-learning models predict the contents of visual imagery during the sleep-onset period, given measured brain activity, by discovering links between human functional magnetic resonance imaging patterns and verbal reports with the assistance of lexical and image databases. Decoding models trained on stimulus-induced brain activity in visual cortical areas showed accurate classification, detection, and identification of contents. Our findings demonstrate that specific visual experience during sleep is represented by brain activity patterns shared by stimulus perception, providing a means to uncover subjective contents of dreaming using objective neural measurement.

  13. Audit of long-term and short-term liabilities

    Directory of Open Access Journals (Sweden)

    Korinko M.D.

    2017-03-01

    Full Text Available The article determines the importance of long-term and short-term liabilities for the management of financial and material resources of an enterprise. It reviews the aim, objects and information generators for realization of audit of short-term and long-term obligations. The organizing and methodical providing of audit of long-term and short-term liabilities of an enterprise are generalized. The authors distinguish the stages of realization of audit of long-term and short-term liabilities, the aim of audit on each of the presented stages, and recommend methodical techniques. It is fixed that it is necessary to conduct the estimation of the systems of internal control and record-keeping of an enterprise by implementation of public accountant procedures for determination of volume and maintenance of selection realization. After estimating the indicated systems, a public accountant determines the methodology for realization of public accountant verification of long-term and short-term liabilities. The analytical procedures that public accountants are expedient to use for realization of audit of short-term and long-term obligations are determined. The authors suggest the classification of the educed defects on the results of the conducted public accountant verification of short-term and long-term obligations.

  14. Decoding of finger trajectory from ECoG using deep learning

    Science.gov (United States)

    Xie, Ziqian; Schwartz, Odelia; Prasad, Abhishek

    2018-06-01

    Objective. Conventional decoding pipeline for brain-machine interfaces (BMIs) consists of chained different stages of feature extraction, time-frequency analysis and statistical learning models. Each of these stages uses a different algorithm trained in a sequential manner, which makes it difficult to make the whole system adaptive. The goal was to create an adaptive online system with a single objective function and a single learning algorithm so that the whole system can be trained in parallel to increase the decoding performance. Here, we used deep neural networks consisting of convolutional neural networks (CNN) and a special kind of recurrent neural network (RNN) called long short term memory (LSTM) to address these needs. Approach. We used electrocorticography (ECoG) data collected by Kubanek et al. The task consisted of individual finger flexions upon a visual cue. Our model combined a hierarchical feature extractor CNN and a RNN that was able to process sequential data and recognize temporal dynamics in the neural data. CNN was used as the feature extractor and LSTM was used as the regression algorithm to capture the temporal dynamics of the signal. Main results. We predicted the finger trajectory using ECoG signals and compared results for the least angle regression (LARS), CNN-LSTM, random forest, LSTM model (LSTM_HC, for using hard-coded features) and a decoding pipeline consisting of band-pass filtering, energy extraction, feature selection and linear regression. The results showed that the deep learning models performed better than the commonly used linear model. The deep learning models not only gave smoother and more realistic trajectories but also learned the transition between movement and rest state. Significance. This study demonstrated a decoding network for BMI that involved a convolutional and recurrent neural network model. It integrated the feature extraction pipeline into the convolution and pooling layer and used LSTM layer to capture the

  15. Improved decoding for a concatenated coding system

    DEFF Research Database (Denmark)

    Paaske, Erik

    1990-01-01

    The concatenated coding system recommended by CCSDS (Consultative Committee for Space Data Systems) uses an outer (255,233) Reed-Solomon (RS) code based on 8-b symbols, followed by the block interleaver and an inner rate 1/2 convolutional code with memory 6. Viterbi decoding is assumed. Two new...... decoding procedures based on repeated decoding trials and exchange of information between the two decoders and the deinterleaver are proposed. In the first one, where the improvement is 0.3-0.4 dB, only the RS decoder performs repeated trials. In the second one, where the improvement is 0.5-0.6 dB, both...... decoders perform repeated decoding trials and decoding information is exchanged between them...

  16. Interference-Based Decode and Forward Scheme Using Relay Nodes in Heterogeneous Networks

    Directory of Open Access Journals (Sweden)

    Kentaro Nishimori

    2012-01-01

    Full Text Available This paper proposes interference-based decode and forward scheme that utilizes relay stations (RSs. In Long-Term-Evolution (LTE- Advanced, heterogeneous networks in which femto- and picocells are overlaid onto macrocells are extensively discussed. However, interference between macro- and pico(femtocells arises due to their different transmit power levels. Unlike conventional cooperative transmission schemes, the RS decodes interference in the first transmit timing period and forwards it to the user equipment (UE in the second period. Moreover, cooperative transmission can be achieved without stopping the transmission from the base station (BS to UE when forwarding the interference from the RS to the UE by utilizing the fact that signal-to-noise power ratio (SNR between the RS and UE is much greater than that between the BS and UE. The basic performance of the proposed method is shown based on computer simulation. Moreover, the interference temperature and shadowing effect are measured when considering the coexistence between macro- and femtocells, and the performance of the proposed method is verified using measured shadowing effect.

  17. Long-term follow-up study and long-term care of childhood cancer survivors

    Directory of Open Access Journals (Sweden)

    Hyeon Jin Park

    2010-04-01

    Full Text Available The number of long-term survivors is increasing in the western countries due to remarkable improvements in the treatment of childhood cancer. The long-term complications of childhood cancer survivors in these countries were brought to light by the childhood cancer survivor studies. In Korea, the 5-year survival rate of childhood cancer patients is approaching 70%; therefore, it is extremely important to undertake similar long-term follow-up studies and comprehensive long-term care for our population. On the basis of the experiences of childhood cancer survivorship care of the western countries and the current Korean status of childhood cancer survivors, long-term follow-up study and long-term care systems need to be established in Korea in the near future. This system might contribute to the improvement of the quality of life of childhood cancer survivors through effective intervention strategies.

  18. On Lattice Sequential Decoding for Large MIMO Systems

    KAUST Repository

    Ali, Konpal S.

    2014-04-01

    Due to their ability to provide high data rates, Multiple-Input Multiple-Output (MIMO) wireless communication systems have become increasingly popular. Decoding of these systems with acceptable error performance is computationally very demanding. In the case of large overdetermined MIMO systems, we employ the Sequential Decoder using the Fano Algorithm. A parameter called the bias is varied to attain different performance-complexity trade-offs. Low values of the bias result in excellent performance but at the expense of high complexity and vice versa for higher bias values. We attempt to bound the error by bounding the bias, using the minimum distance of a lattice. Also, a particular trend is observed with increasing SNR: a region of low complexity and high error, followed by a region of high complexity and error falling, and finally a region of low complexity and low error. For lower bias values, the stages of the trend are incurred at lower SNR than for higher bias values. This has the important implication that a low enough bias value, at low to moderate SNR, can result in low error and low complexity even for large MIMO systems. Our work is compared against Lattice Reduction (LR) aided Linear Decoders (LDs). Another impressive observation for low bias values that satisfy the error bound is that the Sequential Decoder\\'s error is seen to fall with increasing system size, while it grows for the LR-aided LDs. For the case of large underdetermined MIMO systems, Sequential Decoding with two preprocessing schemes is proposed – 1) Minimum Mean Square Error Generalized Decision Feedback Equalization (MMSE-GDFE) preprocessing 2) MMSE-GDFE preprocessing, followed by Lattice Reduction and Greedy Ordering. Our work is compared against previous work which employs Sphere Decoding preprocessed using MMSE-GDFE, Lattice Reduction and Greedy Ordering. For the case of large systems, this results in high complexity and difficulty in choosing the sphere radius. Our schemes

  19. Optimized Min-Sum Decoding Algorithm for Low Density Parity Check Codes

    OpenAIRE

    Mohammad Rakibul Islam; Dewan Siam Shafiullah; Muhammad Mostafa Amir Faisal; Imran Rahman

    2011-01-01

    Low Density Parity Check (LDPC) code approaches Shannon–limit performance for binary field and long code lengths. However, performance of binary LDPC code is degraded when the code word length is small. An optimized min-sum algorithm for LDPC code is proposed in this paper. In this algorithm unlike other decoding methods, an optimization factor has been introduced in both check node and bit node of the Min-sum algorithm. The optimization factor is obtained before decoding program, and the sam...

  20. Asynchronous hatching provides females with a means for increasing male care but incurs a cost by reducing offspring fitness.

    Science.gov (United States)

    Ford, L E; Smiseth, P T

    2016-02-01

    In species with biparental care, sexual conflict occurs because the benefit of care depends on the total amount of care provided by the two parents while the cost of care depends on each parent's own contribution. Asynchronous hatching may play a role in mediating the resolution of this conflict over parental care. The sexual conflict hypothesis for the evolution of asynchronous hatching suggests that females adjust hatching patterns in order to increase male parental effort relative to female effort. We tested this hypothesis in the burying beetle Nicrophorus vespilloides by setting up experimental broods with three different hatching patterns: synchronous, asynchronous and highly asynchronous broods. As predicted, we found that males provided care for longer in asynchronous broods whereas the opposite was true of females. However, we did not find any benefit to females of reducing their duration of care in terms of increased lifespan or reduced mass loss during breeding. We found substantial negative effects of hatching asynchrony on offspring fitness as larval mass was lower and fewer larvae survived to dispersal in highly asynchronous broods compared to synchronous or asynchronous broods. Our results suggest that, even though females can increase male parental effort by hatching their broods more asynchronously, females pay a substantial cost from doing so in terms of reducing offspring growth and survival. Thus, females should be under selection to produce a hatching pattern that provides the best possible trade-off between the benefits of increased male parental effort and the costs due to reduced offspring fitness. © 2015 European Society For Evolutionary Biology. Journal of Evolutionary Biology © 2015 European Society For Evolutionary Biology.

  1. Burst-Mode Asynchronous Controllers on FPGA

    Directory of Open Access Journals (Sweden)

    Duarte L. Oliveira

    2008-01-01

    Full Text Available FPGAs have been mainly used to design synchronous circuits. Asynchronous design on FPGAs is difficult because the resulting circuit may suffer from hazard problems. We propose a method that implements a popular class of asynchronous circuits, known as burst mode, on FPGAs based on look-up table architectures. We present two conditions that, if satisfied, guarantee essential hazard-free implementation on any LUT-based FPGA. By doing that, besides all the intrinsic advantages of asynchronous over synchronous circuits, they also take advantage of the shorter design time and lower cost associated with FPGA designs.

  2. FAST: A fully asynchronous and status-tracking pattern for geoprocessing services orchestration

    Science.gov (United States)

    Wu, Huayi; You, Lan; Gui, Zhipeng; Gao, Shuang; Li, Zhenqiang; Yu, Jingmin

    2014-09-01

    Geoprocessing service orchestration (GSO) provides a unified and flexible way to implement cross-application, long-lived, and multi-step geoprocessing service workflows by coordinating geoprocessing services collaboratively. Usually, geoprocessing services and geoprocessing service workflows are data and/or computing intensive. The intensity feature may make the execution process of a workflow time-consuming. Since it initials an execution request without blocking other interactions on the client side, an asynchronous mechanism is especially appropriate for GSO workflows. Many critical problems remain to be solved in existing asynchronous patterns for GSO including difficulties in improving performance, status tracking, and clarifying the workflow structure. These problems are a challenge when orchestrating performance efficiency, making statuses instantly available, and constructing clearly structured GSO workflows. A Fully Asynchronous and Status-Tracking (FAST) pattern that adopts asynchronous interactions throughout the whole communication tier of a workflow is proposed for GSO. The proposed FAST pattern includes a mechanism that actively pushes the latest status to clients instantly and economically. An independent proxy was designed to isolate the status tracking logic from the geoprocessing business logic, which assists the formation of a clear GSO workflow structure. A workflow was implemented in the FAST pattern to simulate the flooding process in the Poyang Lake region. Experimental results show that the proposed FAST pattern can efficiently tackle data/computing intensive geoprocessing tasks. The performance of all collaborative partners was improved due to the asynchronous mechanism throughout communication tier. A status-tracking mechanism helps users retrieve the latest running status of a GSO workflow in an efficient and instant way. The clear structure of the GSO workflow lowers the barriers for geospatial domain experts and model designers to

  3. Fast decoders for qudit topological codes

    International Nuclear Information System (INIS)

    Anwar, Hussain; Brown, Benjamin J; Campbell, Earl T; Browne, Dan E

    2014-01-01

    Qudit toric codes are a natural higher-dimensional generalization of the well-studied qubit toric code. However, standard methods for error correction of the qubit toric code are not applicable to them. Novel decoders are needed. In this paper we introduce two renormalization group decoders for qudit codes and analyse their error correction thresholds and efficiency. The first decoder is a generalization of a ‘hard-decisions’ decoder due to Bravyi and Haah (arXiv:1112.3252). We modify this decoder to overcome a percolation effect which limits its threshold performance for many-level quantum systems. The second decoder is a generalization of a ‘soft-decisions’ decoder due to Poulin and Duclos-Cianci (2010 Phys. Rev. Lett. 104 050504), with a small cell size to optimize the efficiency of implementation in the high dimensional case. In each case, we estimate thresholds for the uncorrelated bit-flip error model and provide a comparative analysis of the performance of both these approaches to error correction of qudit toric codes. (paper)

  4. Efficient decoding with steady-state Kalman filter in neural interface systems.

    Science.gov (United States)

    Malik, Wasim Q; Truccolo, Wilson; Brown, Emery N; Hochberg, Leigh R

    2011-02-01

    The Kalman filter is commonly used in neural interface systems to decode neural activity and estimate the desired movement kinematics. We analyze a low-complexity Kalman filter implementation in which the filter gain is approximated by its steady-state form, computed offline before real-time decoding commences. We evaluate its performance using human motor cortical spike train data obtained from an intracortical recording array as part of an ongoing pilot clinical trial. We demonstrate that the standard Kalman filter gain converges to within 95% of the steady-state filter gain in 1.5±0.5 s (mean ±s.d.). The difference in the intended movement velocity decoded by the two filters vanishes within 5 s, with a correlation coefficient of 0.99 between the two decoded velocities over the session length. We also find that the steady-state Kalman filter reduces the computational load (algorithm execution time) for decoding the firing rates of 25±3 single units by a factor of 7.0±0.9. We expect that the gain in computational efficiency will be much higher in systems with larger neural ensembles. The steady-state filter can thus provide substantial runtime efficiency at little cost in terms of estimation accuracy. This far more efficient neural decoding approach will facilitate the practical implementation of future large-dimensional, multisignal neural interface systems.

  5. Trellises and Trellis-Based Decoding Algorithms for Linear Block Codes. Part 3; The Map and Related Decoding Algirithms

    Science.gov (United States)

    Lin, Shu; Fossorier, Marc

    1998-01-01

    In a coded communication system with equiprobable signaling, MLD minimizes the word error probability and delivers the most likely codeword associated with the corresponding received sequence. This decoding has two drawbacks. First, minimization of the word error probability is not equivalent to minimization of the bit error probability. Therefore, MLD becomes suboptimum with respect to the bit error probability. Second, MLD delivers a hard-decision estimate of the received sequence, so that information is lost between the input and output of the ML decoder. This information is important in coded schemes where the decoded sequence is further processed, such as concatenated coding schemes, multi-stage and iterative decoding schemes. In this chapter, we first present a decoding algorithm which both minimizes bit error probability, and provides the corresponding soft information at the output of the decoder. This algorithm is referred to as the MAP (maximum aposteriori probability) decoding algorithm.

  6. Comparing the force ripple during asynchronous and conventional stimulation.

    Science.gov (United States)

    Downey, Ryan J; Tate, Mark; Kawai, Hiroyuki; Dixon, Warren E

    2014-10-01

    Asynchronous stimulation has been shown to reduce fatigue during electrical stimulation; however, it may also exhibit a force ripple. We quantified the ripple during asynchronous and conventional single-channel transcutaneous stimulation across a range of stimulation frequencies. The ripple was measured during 5 asynchronous stimulation protocols, 2 conventional stimulation protocols, and 3 volitional contractions in 12 healthy individuals. Conventional 40 Hz and asynchronous 16 Hz stimulation were found to induce contractions that were as smooth as volitional contractions. Asynchronous 8, 10, and 12 Hz stimulation induced contractions with significant ripple. Lower stimulation frequencies can reduce fatigue; however, they may also lead to increased ripple. Future efforts should study the relationship between force ripple and the smoothness of the evoked movements in addition to the relationship between stimulation frequency and NMES-induced fatigue to elucidate an optimal stimulation frequency for asynchronous stimulation. © 2014 Wiley Periodicals, Inc.

  7. Fast decoding algorithms for geometric coded apertures

    International Nuclear Information System (INIS)

    Byard, Kevin

    2015-01-01

    Fast decoding algorithms are described for the class of coded aperture designs known as geometric coded apertures which were introduced by Gourlay and Stephen. When compared to the direct decoding method, the algorithms significantly reduce the number of calculations required when performing the decoding for these apertures and hence speed up the decoding process. Experimental tests confirm the efficacy of these fast algorithms, demonstrating a speed up of approximately two to three orders of magnitude over direct decoding.

  8. [Long-term psychiatric hospitalizations].

    Science.gov (United States)

    Plancke, L; Amariei, A

    2017-02-01

    Long-term hospitalizations in psychiatry raise the question of desocialisation of the patients and the inherent costs. Individual indicators were extracted from a medical administrative database containing full-time psychiatric hospitalizations for the period 2011-2013 of people over 16 years old living in the French region of Nord-Pas-de-Calais. We calculated the proportion of people who had experienced a hospitalization with a duration of 292 days or more during the study period. A bivariate analysis was conducted, then ecological data (level of health-care offer, the deprivation index and the size of the municipalities of residence) were included into a multilevel regression model in order to identify the factors significantly related to variability of long-term hospitalization rates. Among hospitalized individuals in psychiatry, 2.6% had had at least one hospitalization of 292 days or more during the observation period; the number of days in long-term hospitalization represented 22.5% of the total of days of full-time hospitalization in psychiatry. The bivariate analysis revealed that seniority in the psychiatric system was strongly correlated with long hospitalization rates. In the multivariate analysis, the individual indicators the most related to an increased risk of long-term hospitalization were: total lack of autonomy (OR=9.0; 95% CI: 6.7-12.2; P<001); diagnoses of psychological development disorders (OR=9.7; CI95%: 4.5-20.6; P<.001); mental retardation (OR=4.5; CI95%: 2.5-8.2; P<.001): schizophrenia (OR=3.0; CI95%: 1.7-5.2; P<.001); compulsory hospitalization (OR=1.7; CI95%: 1.4-2.1; P<.001); having experienced therapeutic isolation (OR=1.8; CI95%: 1.5-2.1; P<.001). Variations of long-term hospitalization rates depending on the type of establishment were very high, but the density of hospital beds or intensity of ambulatory activity services were not significantly linked to long-term hospitalization. The inhabitants of small urban units had

  9. FPGA implementation of low complexity LDPC iterative decoder

    Science.gov (United States)

    Verma, Shivani; Sharma, Sanjay

    2016-07-01

    Low-density parity-check (LDPC) codes, proposed by Gallager, emerged as a class of codes which can yield very good performance on the additive white Gaussian noise channel as well as on the binary symmetric channel. LDPC codes have gained lots of importance due to their capacity achieving property and excellent performance in the noisy channel. Belief propagation (BP) algorithm and its approximations, most notably min-sum, are popular iterative decoding algorithms used for LDPC and turbo codes. The trade-off between the hardware complexity and the decoding throughput is a critical factor in the implementation of the practical decoder. This article presents introduction to LDPC codes and its various decoding algorithms followed by realisation of LDPC decoder by using simplified message passing algorithm and partially parallel decoder architecture. Simplified message passing algorithm has been proposed for trade-off between low decoding complexity and decoder performance. It greatly reduces the routing and check node complexity of the decoder. Partially parallel decoder architecture possesses high speed and reduced complexity. The improved design of the decoder possesses a maximum symbol throughput of 92.95 Mbps and a maximum of 18 decoding iterations. The article presents implementation of 9216 bits, rate-1/2, (3, 6) LDPC decoder on Xilinx XC3D3400A device from Spartan-3A DSP family.

  10. Word Decoding Development during Phonics Instruction in Children at Risk for Dyslexia.

    Science.gov (United States)

    Schaars, Moniek M H; Segers, Eliane; Verhoeven, Ludo

    2017-05-01

    In the present study, we examined the early word decoding development of 73 children at genetic risk of dyslexia and 73 matched controls. We conducted monthly curriculum-embedded word decoding measures during the first 5 months of phonics-based reading instruction followed by standardized word decoding measures halfway and by the end of first grade. In kindergarten, vocabulary, phonological awareness, lexical retrieval, and verbal and visual short-term memory were assessed. The results showed that the children at risk were less skilled in phonemic awareness in kindergarten. During the first 5 months of reading instruction, children at risk were less efficient in word decoding and the discrepancy increased over the months. In subsequent months, the discrepancy prevailed for simple words but increased for more complex words. Phonemic awareness and lexical retrieval predicted the reading development in children at risk and controls to the same extent. It is concluded that children at risk are behind their typical peers in word decoding development starting from the very beginning. Furthermore, it is concluded that the disadvantage increased during phonics instruction and that the same predictors underlie the development of word decoding in the two groups of children. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.

  11. Asynchronous L1-gain control of uncertain switched positive linear systems with dwell time.

    Science.gov (United States)

    Li, Yang; Zhang, Hongbin

    2018-04-01

    In this paper, dwell time (DT) stability, L 1 -gain performance analysis and asynchronous L 1 -gain controller design problems of uncertain switched positive linear systems (SPLSs) are investigated. Via a time-scheduled multiple linear co-positive Lyapunov function (TSMLCLF) approach, convex sufficient conditions of DT stability and L 1 -gain performance of SPLSs with interval and polytopic uncertainties are presented. Furthermore, by utilizing the feature that the TSMLCLF keeps decreasing even if the controller is running asynchronously with the system, the asynchronous L 1 -gain controller design problem of SPLSs with interval and polytopic uncertainties is investigated. Convex sufficient conditions of the existence of time-varying asynchronous state-feedback controller which can ensure the closed-loop system's positivity, stability and L 1 -gain performance are established, and the controller gain matrices can be calculated instantaneously online. The obtained L 1 -gain in the paper is standard. All the results are presented in terms of linear programming. A practical example is provided to show the effectiveness of the results. Copyright © 2018 ISA. Published by Elsevier Ltd. All rights reserved.

  12. Asynchronous machines. Direct torque control; Machines asynchrones. Commande par controle direct de couple

    Energy Technology Data Exchange (ETDEWEB)

    Fornel, B. de [Institut National Polytechnique, 31 - Toulouse (France)

    2006-05-15

    The asynchronous machine, with its low cost and robustness, is today the most widely used motor to make speed variators. However, its main drawback is that the same current generates both the magnetic flux and the torque, and thus any torque variation creates a flux variation. Such a coupling gives to the asynchronous machine a nonlinear behaviour which makes its control much more complex. The direct self control (DSC) method has been developed to improve the low efficiency of the scalar control method and for the specific railway drive application. The direct torque control (DTC) method is derived from the DSC method but corresponds to other type of applications. The DSC and DTC algorithms for asynchronous motors are presented in this article: 1 - direct control of the stator flux (DSC): principle, flux control, torque control, switching frequency of the inverter, speed estimation; 2 - direct torque control (DTC): principle, electromagnetic torque derivative, signals shape and switching frequency, some results, DTC speed variator without speed sensor, DTC application to multi-machine multi-converter systems; 3 - conclusion. (J.S.)

  13. Identity Presentation: The Construction of Identity in Asynchronous Discussion

    Directory of Open Access Journals (Sweden)

    Brian Morgan

    2008-08-01

    Full Text Available This study examines the use of e-mail as a tool for long term discussion between teachers and grade six students. E-mail messages between grade six students and teachers were collected over the course of one academic year. Methods of conversation analysis within a framework of social practice are used to examine the data. While identity is more readily constructed and more fully developed in contexts which allow for physical embodiment such as face-to-face discussion, this analysis found that identity can be constructed in a context that does not provide for the physical embodiment of identity: Identity was constructed using the social, cultural, and technological tools provided and supported by e-mail to develop social practices germane to the e-mail discussion. This study has implications for further understanding the relation between identity, goals, constraints and affordances, and the collaborative creation of social practices in asynchronous computer mediated communication. URN: urn:nbn:de:0114-fqs0803185

  14. Asynchronous vs didactic education: it's too early to throw in the towel on tradition.

    Science.gov (United States)

    Jordan, Jaime; Jalali, Azadeh; Clarke, Samuel; Dyne, Pamela; Spector, Tahlia; Coates, Wendy

    2013-08-08

    Asynchronous, computer based instruction is cost effective, allows self-directed pacing and review, and addresses preferences of millennial learners. Current research suggests there is no significant difference in learning compared to traditional classroom instruction. Data are limited for novice learners in emergency medicine. The objective of this study was to compare asynchronous, computer-based instruction with traditional didactics for senior medical students during a week-long intensive course in acute care. We hypothesized both modalities would be equivalent. This was a prospective observational quasi-experimental study of 4th year medical students who were novice learners with minimal prior exposure to curricular elements. We assessed baseline knowledge with an objective pre-test. The curriculum was delivered in either traditional lecture format (shock, acute abdomen, dyspnea, field trauma) or via asynchronous, computer-based modules (chest pain, EKG interpretation, pain management, trauma). An interactive review covering all topics was followed by a post-test. Knowledge retention was measured after 10 weeks. Pre and post-test items were written by a panel of medical educators and validated with a reference group of learners. Mean scores were analyzed using dependent t-test and attitudes were assessed by a 5-point Likert scale. 44 of 48 students completed the protocol. Students initially acquired more knowledge from didactic education as demonstrated by mean gain scores (didactic: 28.39% ± 18.06; asynchronous 9.93% ± 23.22). Mean difference between didactic and asynchronous = 18.45% with 95% CI [10.40 to 26.50]; p = 0.0001. Retention testing demonstrated similar knowledge attrition: mean gain scores -14.94% (didactic); -17.61% (asynchronous), which was not significantly different: 2.68% ± 20.85, 95% CI [-3.66 to 9.02], p = 0.399. The attitudinal survey revealed that 60.4% of students believed the asynchronous modules were educational and 95

  15. Bounded-Angle Iterative Decoding of LDPC Codes

    Science.gov (United States)

    Dolinar, Samuel; Andrews, Kenneth; Pollara, Fabrizio; Divsalar, Dariush

    2009-01-01

    Bounded-angle iterative decoding is a modified version of conventional iterative decoding, conceived as a means of reducing undetected-error rates for short low-density parity-check (LDPC) codes. For a given code, bounded-angle iterative decoding can be implemented by means of a simple modification of the decoder algorithm, without redesigning the code. Bounded-angle iterative decoding is based on a representation of received words and code words as vectors in an n-dimensional Euclidean space (where n is an integer).

  16. Long-term potentiation and long-term depression: a clinical perspective

    Directory of Open Access Journals (Sweden)

    Timothy V.P. Bliss

    2011-01-01

    Full Text Available Long-term potentiation and long-term depression are enduring changes in synaptic strength, induced by specific patterns of synaptic activity, that have received much attention as cellular models of information storage in the central nervous system. Work in a number of brain regions, from the spinal cord to the cerebral cortex, and in many animal species, ranging from invertebrates to humans, has demonstrated a reliable capacity for chemical synapses to undergo lasting changes in efficacy in response to a variety of induction protocols. In addition to their physiological relevance, long-term potentiation and depression may have important clinical applications. A growing insight into the molecular mechanisms underlying these processes, and technological advances in non-invasive manipulation of brain activity, now puts us at the threshold of harnessing long-term potentiation and depression and other forms of synaptic, cellular and circuit plasticity to manipulate synaptic strength in the human nervous system. Drugs may be used to erase or treat pathological synaptic states and non-invasive stimulation devices may be used to artificially induce synaptic plasticity to ameliorate conditions arising from disrupted synaptic drive. These approaches hold promise for the treatment of a variety of neurological conditions, including neuropathic pain, epilepsy, depression, amblyopia, tinnitus and stroke.

  17. Soft-decision decoding of RS codes

    DEFF Research Database (Denmark)

    Justesen, Jørn

    2005-01-01

    By introducing a few simplifying assumptions we derive a simple condition for successful decoding using the Koetter-Vardy algorithm for soft-decision decoding of RS codes. We show that the algorithm has a significant advantage over hard decision decoding when the code rate is low, when two or more...

  18. Current Trends in High-Level Synthesis of Asynchronous Circuits

    DEFF Research Database (Denmark)

    Sparsø, Jens

    2009-01-01

    This paper is a survey paper presenting what the author sees as two major and promising trends in the current research in CAD-tools and design-methods for asynchronous circuits. One branch of research builds on top of existing asynchronous CAD-tools that perform syntax directed translation, e...... a conventional synchronous circuit as the starting point, and then adds some form of handshake-based flow-control. One approach keeps the global clock and implements discrete-time asynchronous operation. Another approach substitutes the clocked registers by asynchronous handshake-registers, thus creating truly...

  19. Long-Term Symbolic Learning

    National Research Council Canada - National Science Library

    Kennedy, William G; Trafton, J. G

    2007-01-01

    What are the characteristics of long-term learning? We investigated the characteristics of long-term, symbolic learning using the Soar and ACT-R cognitive architectures running cognitive models of two simple tasks...

  20. An Optimized Three-Level Design of Decoder Based on Nanoscale Quantum-Dot Cellular Automata

    Science.gov (United States)

    Seyedi, Saeid; Navimipour, Nima Jafari

    2018-03-01

    Quantum-dot Cellular Automata (QCA) has been potentially considered as a supersede to Complementary Metal-Oxide-Semiconductor (CMOS) because of its inherent advantages. Many QCA-based logic circuits with smaller feature size, improved operating frequency, and lower power consumption than CMOS have been offered. This technology works based on electron relations inside quantum-dots. Due to the importance of designing an optimized decoder in any digital circuit, in this paper, we design, implement and simulate a new 2-to-4 decoder based on QCA with low delay, area, and complexity. The logic functionality of the 2-to-4 decoder is verified using the QCADesigner tool. The results have shown that the proposed QCA-based decoder has high performance in terms of a number of cells, covered area, and time delay. Due to the lower clock pulse frequency, the proposed 2-to-4 decoder is helpful for building QCA-based sequential digital circuits with high performance.

  1. Pediatric polytrauma : Short-term and long-term outcomes

    NARCIS (Netherlands)

    vanderSluis, CK; Kingma, J; Eisma, WH; tenDuis, HJ

    Objective: To assess the short-term and long-term outcomes of pediatric polytrauma patients and to analyze the extent to which short-term outcomes can predict long-term outcomes. Materials and Methods: Ail pediatric polytrauma patients (Injury Severity Score of greater than or equal to 16, less than

  2. The Treeterbi and Parallel Treeterbi algorithms: efficient, optimal decoding for ordinary, generalized and pair HMMs

    DEFF Research Database (Denmark)

    Keibler, Evan; Arumugam, Manimozhiyan; Brent, Michael R

    2007-01-01

    MOTIVATION: Hidden Markov models (HMMs) and generalized HMMs been successfully applied to many problems, but the standard Viterbi algorithm for computing the most probable interpretation of an input sequence (known as decoding) requires memory proportional to the length of the sequence, which can...... be prohibitive. Existing approaches to reducing memory usage either sacrifice optimality or trade increased running time for reduced memory. RESULTS: We developed two novel decoding algorithms, Treeterbi and Parallel Treeterbi, and implemented them in the TWINSCAN/N-SCAN gene-prediction system. The worst case...... asymptotic space and time are the same as for standard Viterbi, but in practice, Treeterbi optimally decodes arbitrarily long sequences with generalized HMMs in bounded memory without increasing running time. Parallel Treeterbi uses the same ideas to split optimal decoding across processors, dividing latency...

  3. Analysis of error floor of LDPC codes under LP decoding over the BSC

    Energy Technology Data Exchange (ETDEWEB)

    Chertkov, Michael [Los Alamos National Laboratory; Chilappagari, Shashi [UNIV OF AZ; Vasic, Bane [UNIV OF AZ; Stepanov, Mikhail [UNIV OF AZ

    2009-01-01

    We consider linear programming (LP) decoding of a fixed low-density parity-check (LDPC) code over the binary symmetric channel (BSC). The LP decoder fails when it outputs a pseudo-codeword which is not a codeword. We propose an efficient algorithm termed the instanton search algorithm (ISA) which, given a random input, generates a set of flips called the BSC-instanton and prove that: (a) the LP decoder fails for any set of flips with support vector including an instanton; (b) for any input, the algorithm outputs an instanton in the number of steps upper-bounded by twice the number of flips in the input. We obtain the number of unique instantons of different sizes by running the ISA sufficient number of times. We then use the instanton statistics to predict the performance of the LP decoding over the BSC in the error floor region. We also propose an efficient semi-analytical method to predict the performance of LP decoding over a large range of transition probabilities of the BSC.

  4. A Scalable Multicore Architecture With Heterogeneous Memory Structures for Dynamic Neuromorphic Asynchronous Processors (DYNAPs).

    Science.gov (United States)

    Moradi, Saber; Qiao, Ning; Stefanini, Fabio; Indiveri, Giacomo

    2018-02-01

    Neuromorphic computing systems comprise networks of neurons that use asynchronous events for both computation and communication. This type of representation offers several advantages in terms of bandwidth and power consumption in neuromorphic electronic systems. However, managing the traffic of asynchronous events in large scale systems is a daunting task, both in terms of circuit complexity and memory requirements. Here, we present a novel routing methodology that employs both hierarchical and mesh routing strategies and combines heterogeneous memory structures for minimizing both memory requirements and latency, while maximizing programming flexibility to support a wide range of event-based neural network architectures, through parameter configuration. We validated the proposed scheme in a prototype multicore neuromorphic processor chip that employs hybrid analog/digital circuits for emulating synapse and neuron dynamics together with asynchronous digital circuits for managing the address-event traffic. We present a theoretical analysis of the proposed connectivity scheme, describe the methods and circuits used to implement such scheme, and characterize the prototype chip. Finally, we demonstrate the use of the neuromorphic processor with a convolutional neural network for the real-time classification of visual symbols being flashed to a dynamic vision sensor (DVS) at high speed.

  5. Video coding for decoding power-constrained embedded devices

    Science.gov (United States)

    Lu, Ligang; Sheinin, Vadim

    2004-01-01

    Low power dissipation and fast processing time are crucial requirements for embedded multimedia devices. This paper presents a technique in video coding to decrease the power consumption at a standard video decoder. Coupled with a small dedicated video internal memory cache on a decoder, the technique can substantially decrease the amount of data traffic to the external memory at the decoder. A decrease in data traffic to the external memory at decoder will result in multiple benefits: faster real-time processing and power savings. The encoder, given prior knowledge of the decoder"s dedicated video internal memory cache management scheme, regulates its choice of motion compensated predictors to reduce the decoder"s external memory accesses. This technique can be used in any standard or proprietary encoder scheme to generate a compliant output bit stream decodable by standard CPU-based and dedicated hardware-based decoders for power savings with the best quality-power cost trade off. Our simulation results show that with a relatively small amount of dedicated video internal memory cache, the technique may decrease the traffic between CPU and external memory over 50%.

  6. Asynchronous communication in real space process algebra

    NARCIS (Netherlands)

    Baeten, J.C.M.; Bergstra, J.A.

    1991-01-01

    A version of classical real space process algebra is given in which messages travel with constant speed through a three-dimensional medium. It follows that communication is asynchronous and has a broadcasting character. A state operator is used to describe asynchronous message transfer and a

  7. Asynchronous communication in real space process algebra

    NARCIS (Netherlands)

    Bergstra, J.A.; Baeten, J.C.M.

    1992-01-01

    A version of classical real space process algebra is given in which messages travel with constant speed through a three-dimensional medium. It follows that communication is asynchronous and has a broadcasting character. A state operator is used to describe asynchronous message transfer and a

  8. A novel parallel pipeline structure of VP9 decoder

    Science.gov (United States)

    Qin, Huabiao; Chen, Wu; Yi, Sijun; Tan, Yunfei; Yi, Huan

    2018-04-01

    To improve the efficiency of VP9 decoder, a novel parallel pipeline structure of VP9 decoder is presented in this paper. According to the decoding workflow, VP9 decoder can be divided into sub-modules which include entropy decoding, inverse quantization, inverse transform, intra prediction, inter prediction, deblocking and pixel adaptive compensation. By analyzing the computing time of each module, hotspot modules are located and the causes of low efficiency of VP9 decoder can be found. Then, a novel pipeline decoder structure is designed by using mixed parallel decoding methods of data division and function division. The experimental results show that this structure can greatly improve the decoding efficiency of VP9.

  9. Asynchronous and corrected-asynchronous numerical solutions of parabolic PDES on MIMD multiprocessors

    Science.gov (United States)

    Amitai, Dganit; Averbuch, Amir; Itzikowitz, Samuel; Turkel, Eli

    1991-01-01

    A major problem in achieving significant speed-up on parallel machines is the overhead involved with synchronizing the concurrent process. Removing the synchronization constraint has the potential of speeding up the computation. The authors present asynchronous (AS) and corrected-asynchronous (CA) finite difference schemes for the multi-dimensional heat equation. Although the discussion concentrates on the Euler scheme for the solution of the heat equation, it has the potential for being extended to other schemes and other parabolic partial differential equations (PDEs). These schemes are analyzed and implemented on the shared memory multi-user Sequent Balance machine. Numerical results for one and two dimensional problems are presented. It is shown experimentally that the synchronization penalty can be about 50 percent of run time: in most cases, the asynchronous scheme runs twice as fast as the parallel synchronous scheme. In general, the efficiency of the parallel schemes increases with processor load, with the time level, and with the problem dimension. The efficiency of the AS may reach 90 percent and over, but it provides accurate results only for steady-state values. The CA, on the other hand, is less efficient, but provides more accurate results for intermediate (non steady-state) values.

  10. NP-hardness of decoding quantum error-correction codes

    Science.gov (United States)

    Hsieh, Min-Hsiu; Le Gall, François

    2011-05-01

    Although the theory of quantum error correction is intimately related to classical coding theory and, in particular, one can construct quantum error-correction codes (QECCs) from classical codes with the dual-containing property, this does not necessarily imply that the computational complexity of decoding QECCs is the same as their classical counterparts. Instead, decoding QECCs can be very much different from decoding classical codes due to the degeneracy property. Intuitively, one expects degeneracy would simplify the decoding since two different errors might not and need not be distinguished in order to correct them. However, we show that general quantum decoding problem is NP-hard regardless of the quantum codes being degenerate or nondegenerate. This finding implies that no considerably fast decoding algorithm exists for the general quantum decoding problems and suggests the existence of a quantum cryptosystem based on the hardness of decoding QECCs.

  11. NP-hardness of decoding quantum error-correction codes

    International Nuclear Information System (INIS)

    Hsieh, Min-Hsiu; Le Gall, Francois

    2011-01-01

    Although the theory of quantum error correction is intimately related to classical coding theory and, in particular, one can construct quantum error-correction codes (QECCs) from classical codes with the dual-containing property, this does not necessarily imply that the computational complexity of decoding QECCs is the same as their classical counterparts. Instead, decoding QECCs can be very much different from decoding classical codes due to the degeneracy property. Intuitively, one expects degeneracy would simplify the decoding since two different errors might not and need not be distinguished in order to correct them. However, we show that general quantum decoding problem is NP-hard regardless of the quantum codes being degenerate or nondegenerate. This finding implies that no considerably fast decoding algorithm exists for the general quantum decoding problems and suggests the existence of a quantum cryptosystem based on the hardness of decoding QECCs.

  12. Evaluation framework for K-best sphere decoders

    KAUST Repository

    Shen, Chungan; Eltawil, Ahmed M.; Salama, Khaled N.

    2010-01-01

    or receive antennas. Tree-searching type decoder structures such as Sphere decoder and K-best decoder present an interesting trade-off between complexity and performance. Many algorithmic developments and VLSI implementations have been reported in literature

  13. Time-varying block codes for synchronisation errors: maximum a posteriori decoder and practical issues

    Directory of Open Access Journals (Sweden)

    Johann A. Briffa

    2014-06-01

    Full Text Available In this study, the authors consider time-varying block (TVB codes, which generalise a number of previous synchronisation error-correcting codes. They also consider various practical issues related to maximum a posteriori (MAP decoding of these codes. Specifically, they give an expression for the expected distribution of drift between transmitter and receiver because of synchronisation errors. They determine an appropriate choice for state space limits based on the drift probability distribution. In turn, they obtain an expression for the decoder complexity under given channel conditions in terms of the state space limits used. For a given state space, they also give a number of optimisations that reduce the algorithm complexity with no further loss of decoder performance. They also show how the MAP decoder can be used in the absence of known frame boundaries, and demonstrate that an appropriate choice of decoder parameters allows the decoder to approach the performance when frame boundaries are known, at the expense of some increase in complexity. Finally, they express some existing constructions as TVB codes, comparing performance with published results and showing that improved performance is possible by taking advantage of the flexibility of TVB codes.

  14. Asynchronous vs didactic education: it’s too early to throw in the towel on tradition

    Science.gov (United States)

    2013-01-01

    Background Asynchronous, computer based instruction is cost effective, allows self-directed pacing and review, and addresses preferences of millennial learners. Current research suggests there is no significant difference in learning compared to traditional classroom instruction. Data are limited for novice learners in emergency medicine. The objective of this study was to compare asynchronous, computer-based instruction with traditional didactics for senior medical students during a week-long intensive course in acute care. We hypothesized both modalities would be equivalent. Methods This was a prospective observational quasi-experimental study of 4th year medical students who were novice learners with minimal prior exposure to curricular elements. We assessed baseline knowledge with an objective pre-test. The curriculum was delivered in either traditional lecture format (shock, acute abdomen, dyspnea, field trauma) or via asynchronous, computer-based modules (chest pain, EKG interpretation, pain management, trauma). An interactive review covering all topics was followed by a post-test. Knowledge retention was measured after 10 weeks. Pre and post-test items were written by a panel of medical educators and validated with a reference group of learners. Mean scores were analyzed using dependent t-test and attitudes were assessed by a 5-point Likert scale. Results 44 of 48 students completed the protocol. Students initially acquired more knowledge from didactic education as demonstrated by mean gain scores (didactic: 28.39% ± 18.06; asynchronous 9.93% ± 23.22). Mean difference between didactic and asynchronous = 18.45% with 95% CI [10.40 to 26.50]; p = 0.0001. Retention testing demonstrated similar knowledge attrition: mean gain scores −14.94% (didactic); -17.61% (asynchronous), which was not significantly different: 2.68% ± 20.85, 95% CI [−3.66 to 9.02], p = 0.399. The attitudinal survey revealed that 60.4% of students believed the asynchronous

  15. Neural Decoder for Topological Codes

    Science.gov (United States)

    Torlai, Giacomo; Melko, Roger G.

    2017-07-01

    We present an algorithm for error correction in topological codes that exploits modern machine learning techniques. Our decoder is constructed from a stochastic neural network called a Boltzmann machine, of the type extensively used in deep learning. We provide a general prescription for the training of the network and a decoding strategy that is applicable to a wide variety of stabilizer codes with very little specialization. We demonstrate the neural decoder numerically on the well-known two-dimensional toric code with phase-flip errors.

  16. Sum rates of asynchronous GFDMA and SC-FDMA for 5G uplink

    Directory of Open Access Journals (Sweden)

    Woojin Park

    2015-12-01

    Full Text Available The fifth generation (5G of mobile communication envisions ultralow latency less than 1 ms for radio interface. To this end, frameless asynchronous multiple access may be needed to allow users to transmit instantly without waiting for the next frame start. In this paper, generalized frequency division multiple-access (GFDMA, one of the promising multiple-access candidates for 5G mobile, is compared with the conventional single-carrier FDMA (SC-FDMA in terms of the uplink sum rate when both techniques are adapted for the asynchronous scenario. In particular, a waveform windowing technique is applied to both schemes to mitigate the inter-user interference due to non-zero out-of-band emission.

  17. LDPC Decoding on GPU for Mobile Device

    Directory of Open Access Journals (Sweden)

    Yiqin Lu

    2016-01-01

    Full Text Available A flexible software LDPC decoder that exploits data parallelism for simultaneous multicode words decoding on the mobile device is proposed in this paper, supported by multithreading on OpenCL based graphics processing units. By dividing the check matrix into several parts to make full use of both the local memory and private memory on GPU and properly modify the code capacity each time, our implementation on a mobile phone shows throughputs above 100 Mbps and delay is less than 1.6 millisecond in decoding, which make high-speed communication like video calling possible. To realize efficient software LDPC decoding on the mobile device, the LDPC decoding feature on communication baseband chip should be replaced to save the cost and make it easier to upgrade decoder to be compatible with a variety of channel access schemes.

  18. On the average complexity of sphere decoding in lattice space-time coded multiple-input multiple-output channel

    KAUST Repository

    Abediseid, Walid

    2012-12-21

    The exact average complexity analysis of the basic sphere decoder for general space-time codes applied to multiple-input multiple-output (MIMO) wireless channel is known to be difficult. In this work, we shed the light on the computational complexity of sphere decoding for the quasi- static, lattice space-time (LAST) coded MIMO channel. Specifically, we drive an upper bound of the tail distribution of the decoder\\'s computational complexity. We show that when the computational complexity exceeds a certain limit, this upper bound becomes dominated by the outage probability achieved by LAST coding and sphere decoding schemes. We then calculate the minimum average computational complexity that is required by the decoder to achieve near optimal performance in terms of the system parameters. Our results indicate that there exists a cut-off rate (multiplexing gain) for which the average complexity remains bounded. Copyright © 2012 John Wiley & Sons, Ltd.

  19. The serial message-passing schedule for LDPC decoding algorithms

    Science.gov (United States)

    Liu, Mingshan; Liu, Shanshan; Zhou, Yuan; Jiang, Xue

    2015-12-01

    The conventional message-passing schedule for LDPC decoding algorithms is the so-called flooding schedule. It has the disadvantage that the updated messages cannot be used until next iteration, thus reducing the convergence speed . In this case, the Layered Decoding algorithm (LBP) based on serial message-passing schedule is proposed. In this paper the decoding principle of LBP algorithm is briefly introduced, and then proposed its two improved algorithms, the grouped serial decoding algorithm (Grouped LBP) and the semi-serial decoding algorithm .They can improve LBP algorithm's decoding speed while maintaining a good decoding performance.

  20. Near-Term Actions to Address Long-Term Climate Risk

    Science.gov (United States)

    Lempert, R. J.

    2014-12-01

    Addressing climate change requires effective long-term policy making, which occurs when reflecting on potential events decades or more in the future causes policy makers to choose near-term actions different than those they would otherwise pursue. Contrary to some expectations, policy makers do sometimes make such long-term decisions, but not as commonly and successfully as climate change may require. In recent years however, the new capabilities of analytic decision support tools, combined with improved understanding of cognitive and organizational behaviors, has significantly improved the methods available for organizations to manage longer-term climate risks. In particular, these tools allow decision makers to understand what near-term actions consistently contribute to achieving both short- and long-term societal goals, even in the face of deep uncertainty regarding the long-term future. This talk will describe applications of these approaches for infrastructure, water, and flood risk management planning, as well as studies of how near-term choices about policy architectures can affect long-term greenhouse gas emission reduction pathways.

  1. Improved Power Decoding of One-Point Hermitian Codes

    DEFF Research Database (Denmark)

    Puchinger, Sven; Bouw, Irene; Rosenkilde, Johan Sebastian Heesemann

    2017-01-01

    We propose a new partial decoding algorithm for one-point Hermitian codes that can decode up to the same number of errors as the Guruswami–Sudan decoder. Simulations suggest that it has a similar failure probability as the latter one. The algorithm is based on a recent generalization of the power...... decoding algorithm for Reed–Solomon codes and does not require an expensive root-finding step. In addition, it promises improvements for decoding interleaved Hermitian codes....

  2. Decoding communities in networks.

    Science.gov (United States)

    Radicchi, Filippo

    2018-02-01

    According to a recent information-theoretical proposal, the problem of defining and identifying communities in networks can be interpreted as a classical communication task over a noisy channel: memberships of nodes are information bits erased by the channel, edges and nonedges in the network are parity bits introduced by the encoder but degraded through the channel, and a community identification algorithm is a decoder. The interpretation is perfectly equivalent to the one at the basis of well-known statistical inference algorithms for community detection. The only difference in the interpretation is that a noisy channel replaces a stochastic network model. However, the different perspective gives the opportunity to take advantage of the rich set of tools of coding theory to generate novel insights on the problem of community detection. In this paper, we illustrate two main applications of standard coding-theoretical methods to community detection. First, we leverage a state-of-the-art decoding technique to generate a family of quasioptimal community detection algorithms. Second and more important, we show that the Shannon's noisy-channel coding theorem can be invoked to establish a lower bound, here named as decodability bound, for the maximum amount of noise tolerable by an ideal decoder to achieve perfect detection of communities. When computed for well-established synthetic benchmarks, the decodability bound explains accurately the performance achieved by the best community detection algorithms existing on the market, telling us that only little room for their improvement is still potentially left.

  3. Decoding communities in networks

    Science.gov (United States)

    Radicchi, Filippo

    2018-02-01

    According to a recent information-theoretical proposal, the problem of defining and identifying communities in networks can be interpreted as a classical communication task over a noisy channel: memberships of nodes are information bits erased by the channel, edges and nonedges in the network are parity bits introduced by the encoder but degraded through the channel, and a community identification algorithm is a decoder. The interpretation is perfectly equivalent to the one at the basis of well-known statistical inference algorithms for community detection. The only difference in the interpretation is that a noisy channel replaces a stochastic network model. However, the different perspective gives the opportunity to take advantage of the rich set of tools of coding theory to generate novel insights on the problem of community detection. In this paper, we illustrate two main applications of standard coding-theoretical methods to community detection. First, we leverage a state-of-the-art decoding technique to generate a family of quasioptimal community detection algorithms. Second and more important, we show that the Shannon's noisy-channel coding theorem can be invoked to establish a lower bound, here named as decodability bound, for the maximum amount of noise tolerable by an ideal decoder to achieve perfect detection of communities. When computed for well-established synthetic benchmarks, the decodability bound explains accurately the performance achieved by the best community detection algorithms existing on the market, telling us that only little room for their improvement is still potentially left.

  4. Long Term Financing of Infrastructure

    OpenAIRE

    Sinha, Sidharth

    2014-01-01

    Infrastructure projects, given their long life, require long term financing. The main sources of long term financings are insurance and pension funds who seek long term investments with low credit risk. However, in India household financial savings are mainly invested in bank deposits. Insurance and pension funds account for only a small percentage of household financial savings. In addition most infrastructure projects do not qualify for investment by insurance and pension funds because of t...

  5. Decoding divergent series in nonparaxial optics.

    Science.gov (United States)

    Borghi, Riccardo; Gori, Franco; Guattari, Giorgio; Santarsiero, Massimo

    2011-03-15

    A theoretical analysis aimed at investigating the divergent character of perturbative series involved in the study of free-space nonparaxial propagation of vectorial optical beams is proposed. Our analysis predicts a factorial divergence for such series and provides a theoretical framework within which the results of recently published numerical experiments concerning nonparaxial propagation of vectorial Gaussian beams find a meaningful interpretation in terms of the decoding operated on such series by the Weniger transformation.

  6. A modular control architecture for real-time synchronous and asynchronous systems

    International Nuclear Information System (INIS)

    Butler, P.L.; Jones, J.P.

    1993-01-01

    This paper describes a control architecture for real-time control of complex robotic systems. The Modular Integrated Control Architecture (MICA), which is actually two complementary control systems, recognizes and exploits the differences between asynchronous and synchronous control. The asynchronous control system simulates shared memory on a heterogeneous network. For control information, a portable event-scheme is used. This scheme provides consistent interprocess coordination among multiple tasks on a number of distributed systems. The machines in the network can vary with respect to their native operating systems and the intemal representation of numbers they use. The synchronous control system is needed for tight real-time control of complex electromechanical systems such as robot manipulators, and the system uses multiple processors at a specified rate. Both the synchronous and asynchronous portions of MICA have been developed to be extremely modular. MICA presents a simple programming model to code developers and also considers the needs of system integrators and maintainers. MICA has been used successfully in a complex robotics project involving a mobile 7-degree-of-freedom manipulator in a heterogeneous network with a body of software totaling over 100,000 lines of code. MICA has also been used in another robotics system, controlling a commercial long-reach manipulator

  7. Unsupervised learning of facial emotion decoding skills.

    Science.gov (United States)

    Huelle, Jan O; Sack, Benjamin; Broer, Katja; Komlewa, Irina; Anders, Silke

    2014-01-01

    Research on the mechanisms underlying human facial emotion recognition has long focussed on genetically determined neural algorithms and often neglected the question of how these algorithms might be tuned by social learning. Here we show that facial emotion decoding skills can be significantly and sustainably improved by practice without an external teaching signal. Participants saw video clips of dynamic facial expressions of five different women and were asked to decide which of four possible emotions (anger, disgust, fear, and sadness) was shown in each clip. Although no external information about the correctness of the participant's response or the sender's true affective state was provided, participants showed a significant increase of facial emotion recognition accuracy both within and across two training sessions two days to several weeks apart. We discuss several similarities and differences between the unsupervised improvement of facial decoding skills observed in the current study, unsupervised perceptual learning of simple stimuli described in previous studies and practice effects often observed in cognitive tasks.

  8. Long term stability of power systems

    Energy Technology Data Exchange (ETDEWEB)

    Kundur, P; Gao, B [Powertech Labs. Inc., Surrey, BC (Canada)

    1994-12-31

    Power system long term stability is still a developing subject. In this paper we provide our perspectives and experiences related to long term stability. The paper begins with the description of the nature of the long term stability problem, followed by the discussion of issues related to the modeling and solution techniques of tools for long term stability analysis. Cases studies are presented to illustrate the voltage stability aspect and plant dynamics aspect of long term stability. (author) 20 refs., 11 figs.

  9. Jointly Decoded Raptor Codes: Analysis and Design for the BIAWGN Channel

    Directory of Open Access Journals (Sweden)

    Venkiah Auguste

    2009-01-01

    Full Text Available Abstract We are interested in the analysis and optimization of Raptor codes under a joint decoding framework, that is, when the precode and the fountain code exchange soft information iteratively. We develop an analytical asymptotic convergence analysis of the joint decoder, derive an optimization method for the design of efficient output degree distributions, and show that the new optimized distributions outperform the existing ones, both at long and moderate lengths. We also show that jointly decoded Raptor codes are robust to channel variation: they perform reasonably well over a wide range of channel capacities. This robustness property was already known for the erasure channel but not for the Gaussian channel. Finally, we discuss some finite length code design issues. Contrary to what is commonly believed, we show by simulations that using a relatively low rate for the precode , we can improve greatly the error floor performance of the Raptor code.

  10. Interior point decoding for linear vector channels

    International Nuclear Information System (INIS)

    Wadayama, T

    2008-01-01

    In this paper, a novel decoding algorithm for low-density parity-check (LDPC) codes based on convex optimization is presented. The decoding algorithm, called interior point decoding, is designed for linear vector channels. The linear vector channels include many practically important channels such as inter-symbol interference channels and partial response channels. It is shown that the maximum likelihood decoding (MLD) rule for a linear vector channel can be relaxed to a convex optimization problem, which is called a relaxed MLD problem

  11. Interior point decoding for linear vector channels

    Energy Technology Data Exchange (ETDEWEB)

    Wadayama, T [Nagoya Institute of Technology, Gokiso, Showa-ku, Nagoya, Aichi, 466-8555 (Japan)], E-mail: wadayama@nitech.ac.jp

    2008-01-15

    In this paper, a novel decoding algorithm for low-density parity-check (LDPC) codes based on convex optimization is presented. The decoding algorithm, called interior point decoding, is designed for linear vector channels. The linear vector channels include many practically important channels such as inter-symbol interference channels and partial response channels. It is shown that the maximum likelihood decoding (MLD) rule for a linear vector channel can be relaxed to a convex optimization problem, which is called a relaxed MLD problem.

  12. Developing asynchronous online interprofessional education.

    Science.gov (United States)

    Sanborn, Heidi

    2016-09-01

    For many health programmes, developing interprofessional education (IPE) has been a challenge. Evidence on the best method for design and implementation of IPE has been slow to emerge, with little research on how to best incorporate IPE in the asynchronous online learning environment. This leaves online programmes with no clear guidance when embarking upon an initiative to integrate IPE into the curriculum. One tool that can be effective at guiding the incorporation of IPE across all learning platforms is the Interprofessional Education Collaborative (IPEC) competencies. A project was designed to integrate the nationally defined IPEC competencies throughout an asynchronous, online baccalaureate nursing completion programme. A programme-wide review led to targeted revision of course and unit-level objectives, learning experiences, and assessments based on the IPEC framework. As a result of this effort, the programme curriculum now provides interprofessional learning activities across all courses. This report provides a method for using the IPEC competencies to incorporate IPE within various asynchronous learning assessments, assuring students learn about, with, and from other professions.

  13. Asynchronous Parallelization of a CFD Solver

    OpenAIRE

    Abdi, Daniel S.; Bitsuamlak, Girma T.

    2015-01-01

    The article of record as published may be found at http://dx.doi.org/10.1155/2015/295393 A Navier-Stokes equations solver is parallelized to run on a cluster of computers using the domain decomposition method. Two approaches of communication and computation are investigated, namely, synchronous and asynchronous methods. Asynchronous communication between subdomains is not commonly used inCFDcodes; however, it has a potential to alleviate scaling bottlenecks incurred due to process...

  14. Symbol synchronization for the TDRSS decoder

    Science.gov (United States)

    Costello, D. J., Jr.

    1983-01-01

    Each 8 bits out of the Viterbi decoder correspond to one symbol of the R/S code. Synchronization must be maintained here so that each 8-bit symbol delivered to the R/S decoder corresponds to an 8-bit symbol from the R/S encoder. Lack of synchronization, would cause an error in almost every R/S symbol since even a - 1-bit sync slip shifts every bit in each 8-bit symbol by one position, therby confusing the mapping betweeen 8-bit sequences and symbols. The error correcting capability of the R/S code would be exceeded. Possible ways to correcting this condition include: (1) designing the R/S decoder to recognize the overload and shifting the output sequence of the inner decoder to establish a different sync state; (2) using the characteristics of the inner decoder to establish symbol synchronization for the outer code, with or without a deinterleaver and an interleaver; and (3) modifying the encoder to alternate periodically between two sets of generators.

  15. A class of Sudan-decodable codes

    DEFF Research Database (Denmark)

    Nielsen, Rasmus Refslund

    2000-01-01

    In this article, Sudan's algorithm is modified into an efficient method to list-decode a class of codes which can be seen as a generalization of Reed-Solomon codes. The algorithm is specialized into a very efficient method for unique decoding. The code construction can be generalized based...... on algebraic-geometry codes and the decoding algorithms are generalized accordingly. Comparisons with Reed-Solomon and Hermitian codes are made....

  16. Adaptive Iterative Soft-Input Soft-Output Parallel Decision-Feedback Detectors for Asynchronous Coded DS-CDMA Systems

    Directory of Open Access Journals (Sweden)

    Zhang Wei

    2005-01-01

    Full Text Available The optimum and many suboptimum iterative soft-input soft-output (SISO multiuser detectors require a priori information about the multiuser system, such as the users' transmitted signature waveforms, relative delays, as well as the channel impulse response. In this paper, we employ adaptive algorithms in the SISO multiuser detector in order to avoid the need for this a priori information. First, we derive the optimum SISO parallel decision-feedback detector for asynchronous coded DS-CDMA systems. Then, we propose two adaptive versions of this SISO detector, which are based on the normalized least mean square (NLMS and recursive least squares (RLS algorithms. Our SISO adaptive detectors effectively exploit the a priori information of coded symbols, whose soft inputs are obtained from a bank of single-user decoders. Furthermore, we consider how to select practical finite feedforward and feedback filter lengths to obtain a good tradeoff between the performance and computational complexity of the receiver.

  17. Multi-stage decoding for multi-level block modulation codes

    Science.gov (United States)

    Lin, Shu

    1991-01-01

    In this paper, we investigate various types of multi-stage decoding for multi-level block modulation codes, in which the decoding of a component code at each stage can be either soft-decision or hard-decision, maximum likelihood or bounded-distance. Error performance of codes is analyzed for a memoryless additive channel based on various types of multi-stage decoding, and upper bounds on the probability of an incorrect decoding are derived. Based on our study and computation results, we find that, if component codes of a multi-level modulation code and types of decoding at various stages are chosen properly, high spectral efficiency and large coding gain can be achieved with reduced decoding complexity. In particular, we find that the difference in performance between the suboptimum multi-stage soft-decision maximum likelihood decoding of a modulation code and the single-stage optimum decoding of the overall code is very small: only a fraction of dB loss in SNR at the probability of an incorrect decoding for a block of 10(exp -6). Multi-stage decoding of multi-level modulation codes really offers a way to achieve the best of three worlds, bandwidth efficiency, coding gain, and decoding complexity.

  18. On Rational Interpolation-Based List-Decoding and List-Decoding Binary Goppa Codes

    DEFF Research Database (Denmark)

    Beelen, Peter; Høholdt, Tom; Nielsen, Johan Sebastian Rosenkilde

    2013-01-01

    We derive the Wu list-decoding algorithm for generalized Reed–Solomon (GRS) codes by using Gröbner bases over modules and the Euclidean algorithm as the initial algorithm instead of the Berlekamp–Massey algorithm. We present a novel method for constructing the interpolation polynomial fast. We gi...... and a duality in the choice of parameters needed for decoding, both in the case of GRS codes and in the case of Goppa codes....

  19. Coder and decoder of fractal signals of comb-type structure

    Directory of Open Access Journals (Sweden)

    Politanskyi R. L.

    2014-08-01

    Full Text Available The article presents a coder and decoder of fractal signals of comb-type structure (FSCS based on microcontrollers (MC. The coder and decoder consist of identical control modules, while their managed modules have different schematic constructions. The control module performs forming or recognition of signals, and also carries out the function of information exchange with a computer. The basic element of the control module is a PIC18F2550 microcontroller from MicroChip. The coder of the system forms fractal signals of a given order according to the information bits coming from the computer. Samples of the calculated values of the amplitudes of elementary rectangular pulses that constitute the structure of fractal pulses are stored in the memory of the microcontroller as a table. Minimum bit capacity of the DAC necessary for the generation of FSCS of fourth order is four bits. The operation algorithm, "wired" into the controller of the program, provides for encoding of the transmitted information by two-bit symbols. Recognition of the start of transmission of each byte in communication channel is performed by the transmission of the timing signal. In a decoder the microcontroller carries out reception and decoding of the received fractal signals which are then transmitted to the computer. The developed algorithm of the program for the microcontroller of the decoder is carried out by determination of order of fractal impulse after the value of sum of amplitudes of elementary impulses, constituents fractal signal. The programs for coder and decoder are written in "C". In the most critical places of the program influencing on the fast-acting of chart “assembler” insertions are done. The blocks of the coder and decoder were connected with a coaxial 10 meters long cable with an impendance of 75 Ohm. The signals generated by the developed coder of FSCS, were studied using a digital oscillograph. On the basis of the obtained spectrums, it is possible

  20. Asynchronous SAR ADC

    NARCIS (Netherlands)

    2011-01-01

    An asynchronous analog to digital convertor for converting an analog input signal into a digital output is presented. According to an embodiment, the analog to digital convertor comprises a clock input operable to receive an external clock signal having a clock period, a comparator operable to

  1. Decoding Finger Flexion From Band-specific ECoG Signals in Humans

    Directory of Open Access Journals (Sweden)

    Nanying eLiang

    2012-06-01

    Full Text Available This article presents the method that won the BCI competition IV addressed to the pre- diction of the finger flexion from ECoG signals. ECoG-based BCIs have recently drawn the attention from the community. Indeed, ECoG can provide a higher spatial resolution, a higher signal quality and is more suitable for long-term use than classical EEG recordings. These characteristics allow to decode precise brain activities and to realize efficient ECoG-based neu- roprostheses. Signal processing is a very important task in BCIs research for translating brain signals into commands. Here, we present a linear regression method based on the amplitude modulation of band-specific ECoG including a short term memory for individual finger flexion prediction. The effectiveness of the method was proven by achieving the highest value of corre- lation coefficient between the predicted and recorded finger flexion values on data set 4 during the BCI competition IV.

  2. Joint Decoding of Concatenated VLEC and STTC System

    Directory of Open Access Journals (Sweden)

    Chen Huijun

    2008-01-01

    Full Text Available Abstract We consider the decoding of wireless communication systems with both source coding in the application layer and channel coding in the physical layer for high-performance transmission over fading channels. Variable length error correcting codes (VLECs and space time trellis codes (STTCs are used to provide bandwidth efficient data compression as well as coding and diversity gains. At the receiver, an iterative joint source and space time decoding scheme are developed to utilize redundancy in both STTC and VLEC to improve overall decoding performance. Issues such as the inseparable systematic information in the symbol level, the asymmetric trellis structure of VLEC, and information exchange between bit and symbol domains have been considered in the maximum a posteriori probability (MAP decoding algorithm. Simulation results indicate that the developed joint decoding scheme achieves a significant decoding gain over the separate decoding in fading channels, whether or not the channel information is perfectly known at the receiver. Furthermore, how rate allocation between STTC and VLEC affects the performance of the joint source and space-time decoder is investigated. Different systems with a fixed overall information rate are studied. It is shown that for a system with more redundancy dedicated to the source code and a higher order modulation of STTC, the joint decoding yields better performance, though with increased complexity.

  3. Joint Decoding of Concatenated VLEC and STTC System

    Directory of Open Access Journals (Sweden)

    Huijun Chen

    2008-07-01

    Full Text Available We consider the decoding of wireless communication systems with both source coding in the application layer and channel coding in the physical layer for high-performance transmission over fading channels. Variable length error correcting codes (VLECs and space time trellis codes (STTCs are used to provide bandwidth efficient data compression as well as coding and diversity gains. At the receiver, an iterative joint source and space time decoding scheme are developed to utilize redundancy in both STTC and VLEC to improve overall decoding performance. Issues such as the inseparable systematic information in the symbol level, the asymmetric trellis structure of VLEC, and information exchange between bit and symbol domains have been considered in the maximum a posteriori probability (MAP decoding algorithm. Simulation results indicate that the developed joint decoding scheme achieves a significant decoding gain over the separate decoding in fading channels, whether or not the channel information is perfectly known at the receiver. Furthermore, how rate allocation between STTC and VLEC affects the performance of the joint source and space-time decoder is investigated. Different systems with a fixed overall information rate are studied. It is shown that for a system with more redundancy dedicated to the source code and a higher order modulation of STTC, the joint decoding yields better performance, though with increased complexity.

  4. FPGA Realization of Memory 10 Viterbi Decoder

    DEFF Research Database (Denmark)

    Paaske, Erik; Bach, Thomas Bo; Andersen, Jakob Dahl

    1997-01-01

    sequence mode when feedback from the Reed-Solomon decoder is available. The Viterbi decoder is realized using two Altera FLEX 10K50 FPGA's. The overall operating speed is 30 kbit/s, and since up to three iterations are performed for each frame and only one decoder is used, the operating speed...

  5. Complete ML Decoding orf the (73,45) PG Code

    DEFF Research Database (Denmark)

    Justesen, Jørn; Høholdt, Tom; Hjaltason, Johan

    2005-01-01

    Our recent proof of the completeness of decoding by list bit flipping is reviewed. The proof is based on an enumeration of all cosets of low weight in terms of their minimum weight and syndrome weight. By using a geometric description of the error patterns we characterize all remaining cosets....

  6. High Speed Frame Synchronization and Viterbi Decoding

    DEFF Research Database (Denmark)

    Paaske, Erik; Justesen, Jørn; Larsen, Knud J.

    1996-01-01

    The purpose of Phase 1 of the study is to describe the system structure and algorithms in sufficient detail to allow drawing the high level architecture of units containing frame synchronization and Viterbi decoding. The systems we consider are high data rate space communication systems. Also...... components. Node synchronization performed within a Viterbi decoder is discussed, and algorithms for frame synchronization are described and analyzed. We present a list of system configurations that we find potentially useful. Further, the high level architecture of units that contain frame synchronization...... and various other functions needed in a complete system is presented. Two such units are described, one for placement before the Viterbi decoder and another for placement after the decoder. The high level architectures of three possible implementations of Viterbi decoders are described: The first...

  7. High Speed Frame Synchronization and Viterbi Decoding

    DEFF Research Database (Denmark)

    Paaske, Erik; Justesen, Jørn; Larsen, Knud J.

    1998-01-01

    The study has been divided into two phases. The purpose of Phase 1 of the study was to describe the system structure and algorithms in sufficient detail to allow drawing the high level architecture of units containing frame synchronization and Viterbi decoding. After selection of which specific...... potentially useful.Algorithms for frame synchronization are described and analyzed. Further, the high level architecture of units that contain frame synchronization and various other functions needed in a complete system is presented. Two such units are described, one for placement before the Viterbi decoder...... towards a realization in an FPGA.Node synchronization performed within a Viterbi decoder is discussed, and the high level architectures of three possible implementations of Viterbi decoders are described: The first implementation uses a number of commercially available decoders while the the two others...

  8. A Scalable Architecture of a Structured LDPC Decoder

    Science.gov (United States)

    Lee, Jason Kwok-San; Lee, Benjamin; Thorpe, Jeremy; Andrews, Kenneth; Dolinar, Sam; Hamkins, Jon

    2004-01-01

    We present a scalable decoding architecture for a certain class of structured LDPC codes. The codes are designed using a small (n,r) protograph that is replicated Z times to produce a decoding graph for a (Z x n, Z x r) code. Using this architecture, we have implemented a decoder for a (4096,2048) LDPC code on a Xilinx Virtex-II 2000 FPGA, and achieved decoding speeds of 31 Mbps with 10 fixed iterations. The implemented message-passing algorithm uses an optimized 3-bit non-uniform quantizer that operates with 0.2dB implementation loss relative to a floating point decoder.

  9. Simulating fail-stop in asynchronous distributed systems

    Science.gov (United States)

    Sabel, Laura; Marzullo, Keith

    1994-01-01

    The fail-stop failure model appears frequently in the distributed systems literature. However, in an asynchronous distributed system, the fail-stop model cannot be implemented. In particular, it is impossible to reliably detect crash failures in an asynchronous system. In this paper, we show that it is possible to specify and implement a failure model that is indistinguishable from the fail-stop model from the point of view of any process within an asynchronous system. We give necessary conditions for a failure model to be indistinguishable from the fail-stop model, and derive lower bounds on the amount of process replication needed to implement such a failure model. We present a simple one-round protocol for implementing one such failure model, which we call simulated fail-stop.

  10. Asynchronous decentralized method for interconnected electricity markets

    International Nuclear Information System (INIS)

    Huang, Anni; Joo, Sung-Kwan; Song, Kyung-Bin; Kim, Jin-Ho; Lee, Kisung

    2008-01-01

    This paper presents an asynchronous decentralized method to solve the optimization problem of interconnected electricity markets. The proposed method decomposes the optimization problem of combined electricity markets into individual optimization problems. The impact of neighboring markets' information is included in the objective function of the individual market optimization problem by the standard Lagrangian relaxation method. Most decentralized optimization methods use synchronous models of communication to exchange updated market information among markets during the iterative process. In this paper, however, the solutions of the individual optimization problems are coordinated through an asynchronous communication model until they converge to the global optimal solution of combined markets. Numerical examples are presented to demonstrate the advantages of the proposed asynchronous method over the existing synchronous methods. (author)

  11. Fast decoder for local quantum codes using Groebner basis

    Science.gov (United States)

    Haah, Jeongwan

    2013-03-01

    Based on arXiv:1204.1063. A local translation-invariant quantum code has a description in terms of Laurent polynomials. As an application of this observation, we present a fast decoding algorithm for translation-invariant local quantum codes in any spatial dimensions using the straightforward division algorithm for multivariate polynomials. The running time is O (n log n) on average, or O (n2 log n) on worst cases, where n is the number of physical qubits. The algorithm improves a subroutine of the renormalization-group decoder by Bravyi and Haah (arXiv:1112.3252) in the translation-invariant case. This work is supported in part by the Insitute for Quantum Information and Matter, an NSF Physics Frontier Center, and the Korea Foundation for Advanced Studies.

  12. Evaluation framework for K-best sphere decoders

    KAUST Repository

    Shen, Chungan

    2010-08-01

    While Maximum-Likelihood (ML) is the optimum decoding scheme for most communication scenarios, practical implementation difficulties limit its use, especially for Multiple Input Multiple Output (MIMO) systems with a large number of transmit or receive antennas. Tree-searching type decoder structures such as Sphere decoder and K-best decoder present an interesting trade-off between complexity and performance. Many algorithmic developments and VLSI implementations have been reported in literature with widely varying performance to area and power metrics. In this semi-tutorial paper we present a holistic view of different Sphere decoding techniques and K-best decoding techniques, identifying the key algorithmic and implementation trade-offs. We establish a consistent benchmark framework to investigate and compare the delay cost, power cost, and power-delay-product cost incurred by each method. Finally, using the framework, we propose and analyze a novel architecture and compare that to other published approaches. Our goal is to explicitly elucidate the overall advantages and disadvantages of each proposed algorithms in one coherent framework. © 2010 World Scientific Publishing Company.

  13. Long-term urethral catheterisation.

    Science.gov (United States)

    Turner, Bruce; Dickens, Nicola

    This article discusses long-term urethral catheterisation, focusing on the relevant anatomy and physiology, indications for the procedure, catheter selection and catheter care. It is important that nurses have a good working knowledge of long-term catheterisation as the need for this intervention will increase with the rise in chronic health conditions and the ageing population.

  14. Optimization of parameters of special asynchronous electric drives

    Science.gov (United States)

    Karandey, V. Yu; Popov, B. K.; Popova, O. B.; Afanasyev, V. L.

    2018-03-01

    The article considers the solution of the problem of parameters optimization of special asynchronous electric drives. The solution of the problem will allow one to project and create special asynchronous electric drives for various industries. The created types of electric drives will have optimum mass-dimensional and power parameters. It will allow one to realize and fulfill the set characteristics of management of technological processes with optimum level of expenses of electric energy, time of completing the process or other set parameters. The received decision allows one not only to solve a certain optimizing problem, but also to construct dependences between the optimized parameters of special asynchronous electric drives, for example, with the change of power, current in a winding of the stator or rotor, induction in a gap or steel of magnetic conductors and other parameters. On the constructed dependences, it is possible to choose necessary optimum values of parameters of special asynchronous electric drives and their components without carrying out repeated calculations.

  15. Peer Moderation of Asynchronous Online Discussions: An Exploratory Study of Peer E-Moderating Behaviour

    Science.gov (United States)

    Ghadirian, Hajar; Ayub, Ahmad Fauzi Mohd

    2017-01-01

    This study explored patterns of e-moderating behaviour students performed when they were assigned as peer moderators of asynchronous online discussions in a reciprocal manner. Eighty-four students from an undergraduate blended course were observed during a 7-week-long online discussions. Using quantitative content analysis peer moderators'…

  16. Orientation decoding: Sense in spirals?

    Science.gov (United States)

    Clifford, Colin W G; Mannion, Damien J

    2015-04-15

    The orientation of a visual stimulus can be successfully decoded from the multivariate pattern of fMRI activity in human visual cortex. Whether this capacity requires coarse-scale orientation biases is controversial. We and others have advocated the use of spiral stimuli to eliminate a potential coarse-scale bias-the radial bias toward local orientations that are collinear with the centre of gaze-and hence narrow down the potential coarse-scale biases that could contribute to orientation decoding. The usefulness of this strategy is challenged by the computational simulations of Carlson (2014), who reported the ability to successfully decode spirals of opposite sense (opening clockwise or counter-clockwise) from the pooled output of purportedly unbiased orientation filters. Here, we elaborate the mathematical relationship between spirals of opposite sense to confirm that they cannot be discriminated on the basis of the pooled output of unbiased or radially biased orientation filters. We then demonstrate that Carlson's (2014) reported decoding ability is consistent with the presence of inadvertent biases in the set of orientation filters; biases introduced by their digital implementation and unrelated to the brain's processing of orientation. These analyses demonstrate that spirals must be processed with an orientation bias other than the radial bias for successful decoding of spiral sense. Copyright © 2014 Elsevier Inc. All rights reserved.

  17. Multiparty Asynchronous Session Types

    DEFF Research Database (Denmark)

    Honda, Kohei; Yoshida, Nobuko; Carbone, Marco

    2016-01-01

    . This work extends the foregoing theories of binary session types to multiparty, asynchronous sessions, which often arise in practical communication-centered applications. Presented as a typed calculus for mobile processes, the theory introduces a new notion of types in which interactions involving multiple......Communication is a central elements in software development. As a potential typed foundation for structured communication-centered programming, session types have been studied over the past decade for a wide range of process calculi and programming languages, focusing on binary (two-party) sessions...... peers are directly abstracted as a global scenario. Global types retain the friendly type syntax of binary session types while specifying dependencies and capturing complex causal chains of multiparty asynchronous interactions. A global type plays the role of a shared agreement among communication peers...

  18. Grasp movement decoding from premotor and parietal cortex.

    Science.gov (United States)

    Townsend, Benjamin R; Subasi, Erk; Scherberger, Hansjörg

    2011-10-05

    Despite recent advances in harnessing cortical motor-related activity to control computer cursors and robotic devices, the ability to decode and execute different grasping patterns remains a major obstacle. Here we demonstrate a simple Bayesian decoder for real-time classification of grip type and wrist orientation in macaque monkeys that uses higher-order planning signals from anterior intraparietal cortex (AIP) and ventral premotor cortex (area F5). Real-time decoding was based on multiunit signals, which had similar tuning properties to cells in previous single-unit recording studies. Maximum decoding accuracy for two grasp types (power and precision grip) and five wrist orientations was 63% (chance level, 10%). Analysis of decoder performance showed that grip type decoding was highly accurate (90.6%), with most errors occurring during orientation classification. In a subsequent off-line analysis, we found small but significant performance improvements (mean, 6.25 percentage points) when using an optimized spike-sorting method (superparamagnetic clustering). Furthermore, we observed significant differences in the contributions of F5 and AIP for grasp decoding, with F5 being better suited for classification of the grip type and AIP contributing more toward decoding of object orientation. However, optimum decoding performance was maximal when using neural activity simultaneously from both areas. Overall, these results highlight quantitative differences in the functional representation of grasp movements in AIP and F5 and represent a first step toward using these signals for developing functional neural interfaces for hand grasping.

  19. O2-GIDNC: Beyond instantly decodable network coding

    KAUST Repository

    Aboutorab, Neda

    2013-06-01

    In this paper, we are concerned with extending the graph representation of generalized instantly decodable network coding (GIDNC) to a more general opportunistic network coding (ONC) scenario, referred to as order-2 GIDNC (O2-GIDNC). In the O2-GIDNC scheme, receivers can store non-instantly decodable packets (NIDPs) comprising two of their missing packets, and use them in a systematic way for later decodings. Once this graph representation is found, it can be used to extend the GIDNC graph-based analyses to the proposed O2-GIDNC scheme with a limited increase in complexity. In the proposed O2-GIDNC scheme, the information of the stored NIDPs at the receivers and the decoding opportunities they create can be exploited to improve the broadcast completion time and decoding delay compared to traditional GIDNC scheme. The completion time and decoding delay minimizing algorithms that can operate on the new O2-GIDNC graph are further described. The simulation results show that our proposed O2-GIDNC improves the completion time and decoding delay performance of the traditional GIDNC. © 2013 IEEE.

  20. On decoding of multi-level MPSK modulation codes

    Science.gov (United States)

    Lin, Shu; Gupta, Alok Kumar

    1990-01-01

    The decoding problem of multi-level block modulation codes is investigated. The hardware design of soft-decision Viterbi decoder for some short length 8-PSK block modulation codes is presented. An effective way to reduce the hardware complexity of the decoder by reducing the branch metric and path metric, using a non-uniform floating-point to integer mapping scheme, is proposed and discussed. The simulation results of the design are presented. The multi-stage decoding (MSD) of multi-level modulation codes is also investigated. The cases of soft-decision and hard-decision MSD are considered and their performance are evaluated for several codes of different lengths and different minimum squared Euclidean distances. It is shown that the soft-decision MSD reduces the decoding complexity drastically and it is suboptimum. The hard-decision MSD further simplifies the decoding while still maintaining a reasonable coding gain over the uncoded system, if the component codes are chosen properly. Finally, some basic 3-level 8-PSK modulation codes using BCH codes as component codes are constructed and their coding gains are found for hard decision multistage decoding.

  1. Multi-stage decoding of multi-level modulation codes

    Science.gov (United States)

    Lin, Shu; Kasami, Tadao; Costello, Daniel J., Jr.

    1991-01-01

    Various types of multi-stage decoding for multi-level modulation codes are investigated. It is shown that if the component codes of a multi-level modulation code and types of decoding at various stages are chosen properly, high spectral efficiency and large coding gain can be achieved with reduced decoding complexity. Particularly, it is shown that the difference in performance between the suboptimum multi-stage soft-decision maximum likelihood decoding of a modulation code and the single-stage optimum soft-decision decoding of the code is very small, only a fraction of dB loss in signal to noise ratio at a bit error rate (BER) of 10(exp -6).

  2. The Treeterbi and Parallel Treeterbi algorithms: efficient, optimal decoding for ordinary, generalized and pair HMMs.

    Science.gov (United States)

    Keibler, Evan; Arumugam, Manimozhiyan; Brent, Michael R

    2007-03-01

    Hidden Markov models (HMMs) and generalized HMMs been successfully applied to many problems, but the standard Viterbi algorithm for computing the most probable interpretation of an input sequence (known as decoding) requires memory proportional to the length of the sequence, which can be prohibitive. Existing approaches to reducing memory usage either sacrifice optimality or trade increased running time for reduced memory. We developed two novel decoding algorithms, Treeterbi and Parallel Treeterbi, and implemented them in the TWINSCAN/N-SCAN gene-prediction system. The worst case asymptotic space and time are the same as for standard Viterbi, but in practice, Treeterbi optimally decodes arbitrarily long sequences with generalized HMMs in bounded memory without increasing running time. Parallel Treeterbi uses the same ideas to split optimal decoding across processors, dividing latency to completion by approximately the number of available processors with constant average overhead per processor. Using these algorithms, we were able to optimally decode all human chromosomes with N-SCAN, which increased its accuracy relative to heuristic solutions. We also implemented Treeterbi for Pairagon, our pair HMM based cDNA-to-genome aligner. The TWINSCAN/N-SCAN/PAIRAGON open source software package is available from http://genes.cse.wustl.edu.

  3. Binary Systematic Network Coding for Progressive Packet Decoding

    OpenAIRE

    Jones, Andrew L.; Chatzigeorgiou, Ioannis; Tassi, Andrea

    2015-01-01

    We consider binary systematic network codes and investigate their capability of decoding a source message either in full or in part. We carry out a probability analysis, derive closed-form expressions for the decoding probability and show that systematic network coding outperforms conventional net- work coding. We also develop an algorithm based on Gaussian elimination that allows progressive decoding of source packets. Simulation results show that the proposed decoding algorithm can achieve ...

  4. Power decoding Reed-Solomon codes up to the Johnson radius

    DEFF Research Database (Denmark)

    Rosenkilde, Johan Sebastian Heesemann

    2018-01-01

    Power decoding, or "decoding using virtual interleaving" is a technique for decoding Reed-Solomon codes up to the Sudan radius. Since the method's inception, it has been an open question if it is possible to use this approach to decode up to the Johnson radius - the decoding radius of the Guruswami...

  5. Application of Beyond Bound Decoding for High Speed Optical Communications

    DEFF Research Database (Denmark)

    Li, Bomin; Larsen, Knud J.; Vegas Olmos, Juan José

    2013-01-01

    This paper studies the application of beyond bound decoding method for high speed optical communications. This hard-decision decoding method outperforms traditional minimum distance decoding method, with a total net coding gain of 10.36 dB.......This paper studies the application of beyond bound decoding method for high speed optical communications. This hard-decision decoding method outperforms traditional minimum distance decoding method, with a total net coding gain of 10.36 dB....

  6. Polar Coding with CRC-Aided List Decoding

    Science.gov (United States)

    2015-08-01

    TECHNICAL REPORT 2087 August 2015 Polar Coding with CRC-Aided List Decoding David Wasserman Approved...list decoding . RESULTS Our simulation results show that polar coding can produce results very similar to the FEC used in the Digital Video...standard. RECOMMENDATIONS In any application for which the DVB-S2 FEC is considered, polar coding with CRC-aided list decod - ing with N = 65536

  7. Interpolation algorithm for asynchronous ADC-data

    Directory of Open Access Journals (Sweden)

    S. Bramburger

    2017-09-01

    Full Text Available This paper presents a modified interpolation algorithm for signals with variable data rate from asynchronous ADCs. The Adaptive weights Conjugate gradient Toeplitz matrix (ACT algorithm is extended to operate with a continuous data stream. An additional preprocessing of data with constant and linear sections and a weighted overlap of step-by-step into spectral domain transformed signals improve the reconstruction of the asycnhronous ADC signal. The interpolation method can be used if asynchronous ADC data is fed into synchronous digital signal processing.

  8. Long-term associative learning predicts verbal short-term memory performance.

    Science.gov (United States)

    Jones, Gary; Macken, Bill

    2018-02-01

    Studies using tests such as digit span and nonword repetition have implicated short-term memory across a range of developmental domains. Such tests ostensibly assess specialized processes for the short-term manipulation and maintenance of information that are often argued to enable long-term learning. However, there is considerable evidence for an influence of long-term linguistic learning on performance in short-term memory tasks that brings into question the role of a specialized short-term memory system separate from long-term knowledge. Using natural language corpora, we show experimentally and computationally that performance on three widely used measures of short-term memory (digit span, nonword repetition, and sentence recall) can be predicted from simple associative learning operating on the linguistic environment to which a typical child may have been exposed. The findings support the broad view that short-term verbal memory performance reflects the application of long-term language knowledge to the experimental setting.

  9. Decoding of concatenated codes with interleaved outer codes

    DEFF Research Database (Denmark)

    Justesen, Jørn; Høholdt, Tom; Thommesen, Christian

    2004-01-01

    Recently Bleichenbacher et al. proposed a decoding algorithm for interleaved (N, K) Reed-Solomon codes, which allows close to N-K errors to be corrected in many cases. We discuss the application of this decoding algorithm to concatenated codes.......Recently Bleichenbacher et al. proposed a decoding algorithm for interleaved (N, K) Reed-Solomon codes, which allows close to N-K errors to be corrected in many cases. We discuss the application of this decoding algorithm to concatenated codes....

  10. Contributions of Phonological Awareness, Phonological Short-Term Memory, and Rapid Automated Naming, toward Decoding Ability in Students with Mild Intellectual Disability

    Science.gov (United States)

    Soltani, Amanallah; Roslan, Samsilah

    2013-01-01

    Reading decoding ability is a fundamental skill to acquire word-specific orthographic information necessary for skilled reading. Decoding ability and its underlying phonological processing skills have been heavily investigated typically among developing students. However, the issue has rarely been noticed among students with intellectual…

  11. Synchronous and Asynchronous ATM Multiplexor Properties Comparsion

    OpenAIRE

    Jan Zabka

    2006-01-01

    The article is aimed to ATM multiplexor computer model utilisation. Based on simulation runs we try to review aspects of use a synchronous and asynchronous ATM multiplexors. ATM multiplexor is the input queuing model with three inputs. Synchronous multiplexor works without an input priority. Multiplexor inputs are served periodically. Asynchronous multiplexor model supports several queuing and priority mechanisms. CLR and CTD are basic performance parameters. Input cell flows are genera...

  12. Asynchronous communication in real space process algebra

    OpenAIRE

    Baeten, JCM Jos; Bergstra, JA Jan

    1990-01-01

    A version of classical real space process algebra is given in which messages travel with constant speed through a three-dimensional medium. It follows that communication is asynchronous and has a broadcasting character. A state operator is used to describe asynchronous message transfer and a priority mechanism allows to express the broadcasting mechanism. As an application, a protocol is specified in which the receiver moves with respect to the sender.

  13. Iterative channel decoding of FEC-based multiple-description codes.

    Science.gov (United States)

    Chang, Seok-Ho; Cosman, Pamela C; Milstein, Laurence B

    2012-03-01

    Multiple description coding has been receiving attention as a robust transmission framework for multimedia services. This paper studies the iterative decoding of FEC-based multiple description codes. The proposed decoding algorithms take advantage of the error detection capability of Reed-Solomon (RS) erasure codes. The information of correctly decoded RS codewords is exploited to enhance the error correction capability of the Viterbi algorithm at the next iteration of decoding. In the proposed algorithm, an intradescription interleaver is synergistically combined with the iterative decoder. The interleaver does not affect the performance of noniterative decoding but greatly enhances the performance when the system is iteratively decoded. We also address the optimal allocation of RS parity symbols for unequal error protection. For the optimal allocation in iterative decoding, we derive mathematical equations from which the probability distributions of description erasures can be generated in a simple way. The performance of the algorithm is evaluated over an orthogonal frequency-division multiplexing system. The results show that the performance of the multiple description codes is significantly enhanced.

  14. Asynchronous Learning Sources in a High-Tech Organization

    Science.gov (United States)

    Bouhnik, Dan; Giat, Yahel; Sanderovitch, Yafit

    2009-01-01

    Purpose: The purpose of this study is to characterize learning from asynchronous sources among research and development (R&D) personnel. It aims to examine four aspects of asynchronous source learning: employee preferences regarding self-learning; extent of source usage; employee satisfaction with these sources and the effect of the sources on the…

  15. Real-time minimal-bit-error probability decoding of convolutional codes

    Science.gov (United States)

    Lee, L.-N.

    1974-01-01

    A recursive procedure is derived for decoding of rate R = 1/n binary convolutional codes which minimizes the probability of the individual decoding decisions for each information bit, subject to the constraint that the decoding delay be limited to Delta branches. This new decoding algorithm is similar to, but somewhat more complex than, the Viterbi decoding algorithm. A real-time, i.e., fixed decoding delay, version of the Viterbi algorithm is also developed and used for comparison to the new algorithm on simulated channels. It is shown that the new algorithm offers advantages over Viterbi decoding in soft-decision applications, such as in the inner coding system for concatenated coding.

  16. Real-time minimal bit error probability decoding of convolutional codes

    Science.gov (United States)

    Lee, L. N.

    1973-01-01

    A recursive procedure is derived for decoding of rate R=1/n binary convolutional codes which minimizes the probability of the individual decoding decisions for each information bit subject to the constraint that the decoding delay be limited to Delta branches. This new decoding algorithm is similar to, but somewhat more complex than, the Viterbi decoding algorithm. A real-time, i.e. fixed decoding delay, version of the Viterbi algorithm is also developed and used for comparison to the new algorithm on simulated channels. It is shown that the new algorithm offers advantages over Viterbi decoding in soft-decision applications such as in the inner coding system for concatenated coding.

  17. TCDQ-TCT retraction and losses during asynchronous beam dump

    CERN Document Server

    Bracco, Chiara; Quaranta, Elena; CERN. Geneva. ATS Department

    2016-01-01

    The protection provided by the TCDQs in case of asynchronous beam dump depends strongly on their correct setup. They have to respect the strict hierarchy of the full collimation system and shield the tertiary collimators in the experimental regions. This MD aimed at performing asynchronous beam dump tests with different configurations, in order to assess the minimum allowed retraction between TCTs and TCDQs and, as a consequence, on the The protection provided by the TCDQs in case of asynchronous beam dump depends strongly on their correct setup. They have to respect the strict hierarchy of the full collimation system and shield the tertiary collimators in the experimental regions. This MD aimed at performing asynchronous beam dump tests with different configurations, in order to assess the minimum allowed retraction between TCTs and TCDQs and, as a consequence, on the β* reach.

  18. Unsupervised learning of facial emotion decoding skills

    Directory of Open Access Journals (Sweden)

    Jan Oliver Huelle

    2014-02-01

    Full Text Available Research on the mechanisms underlying human facial emotion recognition has long focussed on genetically determined neural algorithms and often neglected the question of how these algorithms might be tuned by social learning. Here we show that facial emotion decoding skills can be significantly and sustainably improved by practise without an external teaching signal. Participants saw video clips of dynamic facial expressions of five different women and were asked to decide which of four possible emotions (anger, disgust, fear and sadness was shown in each clip. Although no external information about the correctness of the participant’s response or the sender’s true affective state was provided, participants showed a significant increase of facial emotion recognition accuracy both within and across two training sessions two days to several weeks apart. We discuss several similarities and differences between the unsupervised improvement of facial decoding skills observed in the current study, unsupervised perceptual learning of simple stimuli described in previous studies and practise effects often observed in cognitive tasks.

  19. Basic Algorithms for the Asynchronous Reconfigurable Mesh

    Directory of Open Access Journals (Sweden)

    Yosi Ben-Asher

    2002-01-01

    Full Text Available Many constant time algorithms for various problems have been developed for the reconfigurable mesh (RM in the past decade. All these algorithms are designed to work with synchronous execution, with no regard for the fact that large size RMs will probably be asynchronous. A similar observation about the PRAM model motivated many researchers to develop algorithms and complexity measures for the asynchronous PRAM (APRAM. In this work, we show how to define the asynchronous reconfigurable mesh (ARM and how to measure the complexity of asynchronous algorithms executed on it. We show that connecting all processors in a row of an n×n ARM (the analog of barrier synchronization in the APRAM model can be solved with complexity Θ(nlog⁡n. Intuitively, this is average work time for solving such a problem. Next, we describe general a technique for simulating T -step synchronous RM algorithms on the ARM with complexity of Θ(T⋅n2log⁡n. Finally, we consider the simulation of the classical synchronous algorithm for counting the number of non-zero bits in an n bits vector using (k

  20. Direct migration motion estimation and mode decision to decoder for a low-complexity decoder Wyner-Ziv video coding

    Science.gov (United States)

    Lei, Ted Chih-Wei; Tseng, Fan-Shuo

    2017-07-01

    This paper addresses the problem of high-computational complexity decoding in traditional Wyner-Ziv video coding (WZVC). The key focus is the migration of two traditionally high-computationally complex encoder algorithms, namely motion estimation and mode decision. In order to reduce the computational burden in this process, the proposed architecture adopts the partial boundary matching algorithm and four flexible types of block mode decision at the decoder. This approach does away with the need for motion estimation and mode decision at the encoder. The experimental results show that the proposed padding block-based WZVC not only decreases decoder complexity to approximately one hundredth that of the state-of-the-art DISCOVER decoding but also outperforms DISCOVER codec by up to 3 to 4 dB.

  1. Study of bifurcation behavior of two-dimensional turbo product code decoders

    International Nuclear Information System (INIS)

    He Yejun; Lau, Francis C.M.; Tse, Chi K.

    2008-01-01

    Turbo codes, low-density parity-check (LDPC) codes and turbo product codes (TPCs) are high performance error-correction codes which employ iterative algorithms for decoding. Under different conditions, the behaviors of the decoders are different. While the nonlinear dynamical behaviors of turbo code decoders and LDPC decoders have been reported in the literature, the dynamical behavior of TPC decoders is relatively unexplored. In this paper, we investigate the behavior of the iterative algorithm of a two-dimensional TPC decoder when the input signal-to-noise ratio (SNR) varies. The quantity to be measured is the mean square value of the posterior probabilities of the information bits. Unlike turbo decoders or LDPC decoders, TPC decoders do not produce a clear 'waterfall region'. This is mainly because the TPC decoding algorithm does not converge to 'indecisive' fixed points even at very low SNR values

  2. Architecture for time or transform domain decoding of reed-solomon codes

    Science.gov (United States)

    Shao, Howard M. (Inventor); Truong, Trieu-Kie (Inventor); Hsu, In-Shek (Inventor); Deutsch, Leslie J. (Inventor)

    1989-01-01

    Two pipeline (255,233) RS decoders, one a time domain decoder and the other a transform domain decoder, use the same first part to develop an errata locator polynomial .tau.(x), and an errata evaluator polynominal A(x). Both the time domain decoder and transform domain decoder have a modified GCD that uses an input multiplexer and an output demultiplexer to reduce the number of GCD cells required. The time domain decoder uses a Chien search and polynomial evaluator on the GCD outputs .tau.(x) and A(x), for the final decoding steps, while the transform domain decoder uses a transform error pattern algorithm operating on .tau.(x) and the initial syndrome computation S(x), followed by an inverse transform algorithm in sequence for the final decoding steps prior to adding the received RS coded message to produce a decoded output message.

  3. Decoding subjective mental states from fMRI activity patterns

    International Nuclear Information System (INIS)

    Tamaki, Masako; Kamitani, Yukiyasu

    2011-01-01

    In recent years, functional magnetic resonance imaging (fMRI) decoding has emerged as a powerful tool to read out detailed stimulus features from multi-voxel brain activity patterns. Moreover, the method has been extended to perform a primitive form of 'mind-reading,' by applying a decoder 'objectively' trained using stimulus features to more 'subjective' conditions. In this paper, we first introduce basic procedures for fMRI decoding based on machine learning techniques. Second, we discuss the source of information used for decoding, in particular, the possibility of extracting information from subvoxel neural structures. We next introduce two experimental designs for decoding subjective mental states: the 'objective-to-subjective design' and the 'subjective-to-subjective design.' Then, we illustrate recent studies on the decoding of a variety of mental states, such as, attention, awareness, decision making, memory, and mental imagery. Finally, we discuss the challenges and new directions of fMRI decoding. (author)

  4. Study of bifurcation behavior of two-dimensional turbo product code decoders

    Energy Technology Data Exchange (ETDEWEB)

    He Yejun [Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hunghom, Hong Kong (China); Lau, Francis C.M. [Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hunghom, Hong Kong (China)], E-mail: encmlau@polyu.edu.hk; Tse, Chi K. [Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hunghom, Hong Kong (China)

    2008-04-15

    Turbo codes, low-density parity-check (LDPC) codes and turbo product codes (TPCs) are high performance error-correction codes which employ iterative algorithms for decoding. Under different conditions, the behaviors of the decoders are different. While the nonlinear dynamical behaviors of turbo code decoders and LDPC decoders have been reported in the literature, the dynamical behavior of TPC decoders is relatively unexplored. In this paper, we investigate the behavior of the iterative algorithm of a two-dimensional TPC decoder when the input signal-to-noise ratio (SNR) varies. The quantity to be measured is the mean square value of the posterior probabilities of the information bits. Unlike turbo decoders or LDPC decoders, TPC decoders do not produce a clear 'waterfall region'. This is mainly because the TPC decoding algorithm does not converge to 'indecisive' fixed points even at very low SNR values.

  5. Interpolation decoding method with variable parameters for fractal image compression

    International Nuclear Information System (INIS)

    He Chuanjiang; Li Gaoping; Shen Xiaona

    2007-01-01

    The interpolation fractal decoding method, which is introduced by [He C, Yang SX, Huang X. Progressive decoding method for fractal image compression. IEE Proc Vis Image Signal Process 2004;3:207-13], involves generating progressively the decoded image by means of an interpolation iterative procedure with a constant parameter. It is well-known that the majority of image details are added at the first steps of iterations in the conventional fractal decoding; hence the constant parameter for the interpolation decoding method must be set as a smaller value in order to achieve a better progressive decoding. However, it needs to take an extremely large number of iterations to converge. It is thus reasonable for some applications to slow down the iterative process at the first stages of decoding and then to accelerate it afterwards (e.g., at some iteration as we need). To achieve the goal, this paper proposed an interpolation decoding scheme with variable (iteration-dependent) parameters and proved the convergence of the decoding process mathematically. Experimental results demonstrate that the proposed scheme has really achieved the above-mentioned goal

  6. Sub-quadratic decoding of one-point hermitian codes

    DEFF Research Database (Denmark)

    Nielsen, Johan Sebastian Rosenkilde; Beelen, Peter

    2015-01-01

    We present the first two sub-quadratic complexity decoding algorithms for one-point Hermitian codes. The first is based on a fast realization of the Guruswami-Sudan algorithm using state-of-the-art algorithms from computer algebra for polynomial-ring matrix minimization. The second is a power...... decoding algorithm: an extension of classical key equation decoding which gives a probabilistic decoding algorithm up to the Sudan radius. We show how the resulting key equations can be solved by the matrix minimization algorithms from computer algebra, yielding similar asymptotic complexities....

  7. DECODING OF ACADEMIC CONTENT BY THE 1st GRADE STUDENTS

    Directory of Open Access Journals (Sweden)

    Kamil Błaszczyński

    2017-07-01

    Full Text Available In the paper a comparative study conducted on the 1st grade students of sociology and pedagogy is discussed. The study was focused on the language skills of students. The most important skills tested were the abilities to decode academic content. The study shows that the students have very poor language skills in decoding the academic content on every level of its complexity. They also have noticeable problems with the definition of basic academic terms. The significance of the obtained results are high because of the innovative topic and character of the study, which was the first such study conducted on students of a Polish university. Results are also valuable for academic teachers who are interested in such problems as effective communication with students.

  8. Pharmacists' perception of synchronous versus asynchronous distance learning for continuing education programs.

    Science.gov (United States)

    Buxton, Eric C

    2014-02-12

    To evaluate and compare pharmacists' satisfaction with the content and learning environment of a continuing education program series offered as either synchronous or asynchronous webinars. An 8-lecture series of online presentations on the topic of new drug therapies was offered to pharmacists in synchronous and asynchronous webinar formats. Participants completed a 50-question online survey at the end of the program series to evaluate their perceptions of the distance learning experience. Eighty-two participants completed the survey instrument (41 participants from the live webinar series and 41 participants from the asynchronous webinar series.) Responses indicated that while both groups were satisfied with the program content, the asynchronous group showed greater satisfaction with many aspects of the learning environment. The synchronous and asynchronous webinar participants responded positively regarding the quality of the programming and the method of delivery, but asynchronous participants rated their experience more positively overall.

  9. Designing Asynchronous Circuits for Low Power: An IFIR Filter

    DEFF Research Database (Denmark)

    Nielsen, Lars Skovby; Sparsø, Jens

    1999-01-01

    This paper addresses the design of asynchronous circuits for low power through an example: a filter bank for a digital hearing aid. The asynchronous design re-implements an existing synchronous circuit which is used in a commercial product. For comparison, both designs have been fabricated...

  10. Realistic modelling of the effects of asynchronous motion at the base of bridge piers

    International Nuclear Information System (INIS)

    Romanelli, F.; Panza, G.F.; Vaccari, F.

    2002-11-01

    Frequently long-span bridges provide deep valley crossings, which require special consideration due to the possibility of local amplification of the ground motion as a consequence of topographical irregularities and local soil conditions. This does in fact cause locally enhanced seismic input with the possibility for the bridge piers to respond asynchronously. This introduces special design requirements so that possible out-of-phase ground displacements and the associated large relative displacements of adjacent piers can be accommodated without excessive damage. Assessment of the local variability of the ground motion due to local lateral heterogeneities and to attenuation properties is thus crucial toward the realistic definition of the asynchronous motion at the base of the bridge piers. We illustrate the work done in the framework of a large international cooperation to assess the importance of non-synchronous seismic excitation of long structures. To accomplish this task we compute complete synthetic accelerograms using as input a set of parameters that describes, to the best of our knowledge, the geological structure and seismotectonic setting of the investigated area. (author)

  11. FPGA BASED ASYNCHRONOUS PIPELINED MB-OFDM UWB TRANSMITTER BACKEND MODULES

    Directory of Open Access Journals (Sweden)

    M. Santhi

    2010-03-01

    Full Text Available In this paper, a novel scheme is proposed which comprises the advantages of asynchronous pipelining techniques and the advantages of FPGAs for implementing a 200Mbps MB-OFDM UWB transmitter digital backend modules. In asynchronous pipelined system, registers are used as in synchronous system. But they are controlled by handshaking signals. Since FPGAs are rich in registers, design and implementation of asynchronous pipelined MBOFDM UWB transmitter on FPGA using four-phase bundled-data protocol is considered in this paper. Novel ideas have also been proposed for designing asynchronous OFDM using Modified Radix-24 SDF and asynchronous interleaver using two RAM banks. Implementation has been performed on ALTERA STRATIX II EP2S60F1020C4 FPGA and it is operating at a speed of 350MHz. It is assured that the proposed MB-OFDM UWB system can be made to work on STRATIX III device with the operating frequency of 528MHz in compliance to the ECMA-368 standard. The proposed scheme is also applicable for FPGA from other vendors and ASIC.

  12. Integrating Asynchronous Digital Design Into the Computer Engineering Curriculum

    Science.gov (United States)

    Smith, S. C.; Al-Assadi, W. K.; Di, J.

    2010-01-01

    As demand increases for circuits with higher performance, higher complexity, and decreased feature size, asynchronous (clockless) paradigms will become more widely used in the semiconductor industry, as evidenced by the International Technology Roadmap for Semiconductors' (ITRS) prediction of a likely shift from synchronous to asynchronous design…

  13. Exploring Asynchronous and Synchronous Tool Use in Online Courses

    Science.gov (United States)

    Oztok, Murat; Zingaro, Daniel; Brett, Clare; Hewitt, Jim

    2013-01-01

    While the independent contributions of synchronous and asynchronous interaction in online learning are clear, comparatively less is known about the pedagogical consequences of using both modes in the same environment. In this study, we examine relationships between students' use of asynchronous discussion forums and synchronous private messages…

  14. An Overview of the Asynchronous Digital Systems – Part 3

    Directory of Open Access Journals (Sweden)

    Mihai Timis

    2008-01-01

    Full Text Available Implementation methods for the digital asynchronous systems use different predefined models like self timed circuits, speed independent circuits, delay insensitive circuits, handshake protocol implementation in asynchronous systems,C Muller circuits.

  15. An Overview of the Asynchronous Digital Systems – Part 2

    Directory of Open Access Journals (Sweden)

    Mihai Timis

    2008-01-01

    Full Text Available Implementation methods for the digital asynchronous systems use different predefined models like self timed circuits, speed independent circuits, delay insensitive circuits, handshake protocol implementation in asynchronous systems,C Muller circuits.

  16. Reforming Long-Term Care Funding in Alberta.

    Science.gov (United States)

    Crump, R Trafford; Repin, Nadya; Sutherland, Jason M

    2015-01-01

    Like many provinces across Canada, Alberta is facing growing demand for long-term care. Issues with the mixed funding model used to pay long-term care providers had Alberta Health Services concerned that it was not efficiently meeting the demand for long-term care. Consequently, in 2010, Alberta Health Services introduced the patient/care-based funding (PCBF) model. PCBF is similar to activity-based funding in that it directly ties the complexity and care needs of long-term care residents to the payment received by long-term care providers. This review describes PCBF and discusses some of its strengths and weaknesses. In doing so, this review is intended to inform other provinces faced with similar long-term care challenges and contemplating their own funding reforms.

  17. Router Designs for an Asynchronous Time-Division-Multiplexed Network-on-Chip

    DEFF Research Database (Denmark)

    Kasapaki, Evangelia; Sparsø, Jens; Sørensen, Rasmus Bo

    2013-01-01

    In this paper we explore the design of an asynchronous router for a time-division-multiplexed (TDM) network-on-chip (NOC) that is being developed for a multi-processor platform for hard real-time systems. TDM inherently requires a common time reference, and existing TDM-based NOC designs are either....... This adds hardware complexity and increases area and power consumption. We propose to use asynchronous routers in order to achieve a simpler, more robust and globally-asynchronous NOC, and this represents an unexplored point in the design space. The paper presents a range of alternative router designs. All...... routers have been synthesized for a 65nm CMOS technology, and the paper reports post-layout figures for area, speed and energy and compares the asynchronous designs with an existing mesochronous clocked router. The results show that an asynchronous router is 2 times smaller, marginally slower...

  18. Differentiating closed-loop cortical intention from rest: building an asynchronous electrocorticographic BCI

    Science.gov (United States)

    Williams, Jordan J.; Rouse, Adam G.; Thongpang, Sanitta; Williams, Justin C.; Moran, Daniel W.

    2013-08-01

    Objective. Recent experiments have shown that electrocorticography (ECoG) can provide robust control signals for a brain-computer interface (BCI). Strategies that attempt to adapt a BCI control algorithm by learning from past trials often assume that the subject is attending to each training trial. Likewise, automatic disabling of movement control would be desirable during resting periods when random brain fluctuations might cause unintended movements of a device. To this end, our goal was to identify ECoG differences that arise between periods of active BCI use and rest. Approach. We examined spectral differences in multi-channel, epidural micro-ECoG signals recorded from non-human primates when rest periods were interleaved between blocks of an active BCI control task. Main Results. Post-hoc analyses demonstrated that these states can be decoded accurately on both a trial-by-trial and real-time basis, and this discriminability remains robust over a period of weeks. In addition, high gamma frequencies showed greater modulation with desired movement direction, while lower frequency components demonstrated greater amplitude differences between task and rest periods, suggesting possible specialized BCI roles for these frequencies. Significance. The results presented here provide valuable insight into the neurophysiology of BCI control as well as important considerations toward the design of an asynchronous BCI system.

  19. Neuroprosthetic Decoder Training as Imitation Learning.

    Science.gov (United States)

    Merel, Josh; Carlson, David; Paninski, Liam; Cunningham, John P

    2016-05-01

    Neuroprosthetic brain-computer interfaces function via an algorithm which decodes neural activity of the user into movements of an end effector, such as a cursor or robotic arm. In practice, the decoder is often learned by updating its parameters while the user performs a task. When the user's intention is not directly observable, recent methods have demonstrated value in training the decoder against a surrogate for the user's intended movement. Here we show that training a decoder in this way is a novel variant of an imitation learning problem, where an oracle or expert is employed for supervised training in lieu of direct observations, which are not available. Specifically, we describe how a generic imitation learning meta-algorithm, dataset aggregation (DAgger), can be adapted to train a generic brain-computer interface. By deriving existing learning algorithms for brain-computer interfaces in this framework, we provide a novel analysis of regret (an important metric of learning efficacy) for brain-computer interfaces. This analysis allows us to characterize the space of algorithmic variants and bounds on their regret rates. Existing approaches for decoder learning have been performed in the cursor control setting, but the available design principles for these decoders are such that it has been impossible to scale them to naturalistic settings. Leveraging our findings, we then offer an algorithm that combines imitation learning with optimal control, which should allow for training of arbitrary effectors for which optimal control can generate goal-oriented control. We demonstrate this novel and general BCI algorithm with simulated neuroprosthetic control of a 26 degree-of-freedom model of an arm, a sophisticated and realistic end effector.

  20. Neuroprosthetic Decoder Training as Imitation Learning.

    Directory of Open Access Journals (Sweden)

    Josh Merel

    2016-05-01

    Full Text Available Neuroprosthetic brain-computer interfaces function via an algorithm which decodes neural activity of the user into movements of an end effector, such as a cursor or robotic arm. In practice, the decoder is often learned by updating its parameters while the user performs a task. When the user's intention is not directly observable, recent methods have demonstrated value in training the decoder against a surrogate for the user's intended movement. Here we show that training a decoder in this way is a novel variant of an imitation learning problem, where an oracle or expert is employed for supervised training in lieu of direct observations, which are not available. Specifically, we describe how a generic imitation learning meta-algorithm, dataset aggregation (DAgger, can be adapted to train a generic brain-computer interface. By deriving existing learning algorithms for brain-computer interfaces in this framework, we provide a novel analysis of regret (an important metric of learning efficacy for brain-computer interfaces. This analysis allows us to characterize the space of algorithmic variants and bounds on their regret rates. Existing approaches for decoder learning have been performed in the cursor control setting, but the available design principles for these decoders are such that it has been impossible to scale them to naturalistic settings. Leveraging our findings, we then offer an algorithm that combines imitation learning with optimal control, which should allow for training of arbitrary effectors for which optimal control can generate goal-oriented control. We demonstrate this novel and general BCI algorithm with simulated neuroprosthetic control of a 26 degree-of-freedom model of an arm, a sophisticated and realistic end effector.

  1. Iterative List Decoding of Concatenated Source-Channel Codes

    Directory of Open Access Journals (Sweden)

    Hedayat Ahmadreza

    2005-01-01

    Full Text Available Whenever variable-length entropy codes are used in the presence of a noisy channel, any channel errors will propagate and cause significant harm. Despite using channel codes, some residual errors always remain, whose effect will get magnified by error propagation. Mitigating this undesirable effect is of great practical interest. One approach is to use the residual redundancy of variable length codes for joint source-channel decoding. In this paper, we improve the performance of residual redundancy source-channel decoding via an iterative list decoder made possible by a nonbinary outer CRC code. We show that the list decoding of VLC's is beneficial for entropy codes that contain redundancy. Such codes are used in state-of-the-art video coders, for example. The proposed list decoder improves the overall performance significantly in AWGN and fully interleaved Rayleigh fading channels.

  2. Evaluation of discrete modeling efficiency of asynchronous electric machines

    OpenAIRE

    Byczkowska-Lipińska, Liliana; Stakhiv, Petro; Hoholyuk, Oksana; Vasylchyshyn, Ivanna

    2011-01-01

    In the paper the problem of effective mathematical macromodels in the form of state variables intended for asynchronous motor transient analysis is considered. Their comparing with traditional mathematical models of asynchronous motors including models built into MATLAB/Simulink software was carried out and analysis of their efficiency was conducted.

  3. Two Studies Examining Argumentation in Asynchronous Computer Mediated Communication

    Science.gov (United States)

    Joiner, Richard; Jones, Sarah; Doherty, John

    2008-01-01

    Asynchronous computer mediated communication (CMC) would seem to be an ideal medium for supporting development in student argumentation. This paper investigates this assumption through two studies. The first study compared asynchronous CMC with face-to-face discussions. The transactional and strategic level of the argumentation (i.e. measures of…

  4. MODELING AND INVESTIGATION OF ASYNCHRONOUS TWO-MACHINE SYSTEM MODES

    Directory of Open Access Journals (Sweden)

    V. S. Safaryan

    2014-01-01

    Full Text Available The paper considers stationary and transient processes of an asynchronous two-machine system. A mathematical model for investigation of stationary and transient modes, static characteristics and research results of dynamic process pertaining to starting-up the asynchronous two-machine system has been given in paper.

  5. On Lattice Sequential Decoding for The Unconstrained AWGN Channel

    KAUST Repository

    Abediseid, Walid; Alouini, Mohamed-Slim

    2013-01-01

    channel has been studied only under the use of the minimum Euclidean distance decoder that is commonly referred to as the \\textit{lattice decoder}. Lattice decoders based on solutions to the NP-hard closest vector problem are very complex to implement

  6. Industrial Foundations as Long-Term Owners

    DEFF Research Database (Denmark)

    Thomsen, Steen; Poulsen, Thomas; Børsting, Christa Winther

    Short-termism has become a serious concern for corporate governance, and this has inspired a search for institutional arrangements to promote long-term decision-making. In this paper, we call attention to long-term ownership by industrial foundations, which is common in Northern Europe but little...... known in the rest of the world. We use a unique Danish data set to document that industrial foundations are long-term owners that practice long-term governance. We show that foundation ownership is highly stable compared to other ownership structures. Foundation-owned companies replace managers less...... frequently. They have conservative capital structures with low financial leverage. They score higher on an index of long-termism in finance, investment, and employment. They survive longer. Overall, our paper supports the hypothesis that corporate time horizons are influenced by ownership structures...

  7. Long-term associative learning predicts verbal short-term memory performance

    OpenAIRE

    Jones, Gary; Macken, Bill

    2017-01-01

    Studies using tests such as digit span and nonword repetition have implicated short-term memory across a range of developmental domains. Such tests ostensibly assess specialized processes for the short-term manipulation and maintenance of information that are often argued to enable long-term learning. However, there is considerable evidence for an influence of long-term linguistic learning on performance in short-term memory tasks that brings into question the role of a specialized short-term...

  8. PsychVACS: a system for asynchronous telepsychiatry.

    Science.gov (United States)

    Odor, Alberto; Yellowlees, Peter; Hilty, Donald; Parish, Michelle Burke; Nafiz, Najia; Iosif, Ana-Maria

    2011-05-01

    To describe the technical development of an asynchronous telepsychiatry application, the Psychiatric Video Archiving and Communication System. A client-server application was developed in Visual Basic.Net with Microsoft(®) SQL database as the backend. It includes the capability of storing video-recorded psychiatric interviews and manages the workflow of the system with automated messaging. Psychiatric Video Archiving and Communication System has been used to conduct the first ever series of asynchronous telepsychiatry consultations worldwide. A review of the software application and the process as part of this project has led to a number of improvements that are being implemented in the next version, which is being written in Java. This is the first description of the use of video recorded data in an asynchronous telemedicine application. Primary care providers and consulting psychiatrists have found it easy to work with and a valuable resource to increase the availability of psychiatric consultation in remote rural locations.

  9. Asynchronous Knowledge Sharing and Conversation Interaction Impact on Grade in an Online Business Course

    Science.gov (United States)

    Strang, Kenneth David

    2011-01-01

    Student knowledge sharing and conversation theory interactions were coded from asynchronous discussion forums to measure the effect of learning-oriented utterances on academic performance. The sample was 3 terms of an online business course (in an accredited MBA program) at a U.S.-based university. Correlation, stepwise regression, and multiple…

  10. Singer product apertures—A coded aperture system with a fast decoding algorithm

    International Nuclear Information System (INIS)

    Byard, Kevin; Shutler, Paul M.E.

    2017-01-01

    A new type of coded aperture configuration that enables fast decoding of the coded aperture shadowgram data is presented. Based on the products of incidence vectors generated from the Singer difference sets, we call these Singer product apertures. For a range of aperture dimensions, we compare experimentally the performance of three decoding methods: standard decoding, induction decoding and direct vector decoding. In all cases the induction and direct vector methods are several orders of magnitude faster than the standard method, with direct vector decoding being significantly faster than induction decoding. For apertures of the same dimensions the increase in speed offered by direct vector decoding over induction decoding is better for lower throughput apertures.

  11. Singer product apertures—A coded aperture system with a fast decoding algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Byard, Kevin, E-mail: kevin.byard@aut.ac.nz [School of Economics, Faculty of Business, Economics and Law, Auckland University of Technology, Auckland 1142 (New Zealand); Shutler, Paul M.E. [National Institute of Education, Nanyang Technological University, 1 Nanyang Walk, Singapore 637616 (Singapore)

    2017-06-01

    A new type of coded aperture configuration that enables fast decoding of the coded aperture shadowgram data is presented. Based on the products of incidence vectors generated from the Singer difference sets, we call these Singer product apertures. For a range of aperture dimensions, we compare experimentally the performance of three decoding methods: standard decoding, induction decoding and direct vector decoding. In all cases the induction and direct vector methods are several orders of magnitude faster than the standard method, with direct vector decoding being significantly faster than induction decoding. For apertures of the same dimensions the increase in speed offered by direct vector decoding over induction decoding is better for lower throughput apertures.

  12. Performance comparison of 0/π- and ± π/2-phase-shifted superstructured Fiber Bragg grating en/decoder.

    Science.gov (United States)

    Dai, Bo; Gao, Zhensen; Wang, Xu; Kataoka, Nobuyuki; Wada, Naoya

    2011-06-20

    We compare the performances of the 0/π-phase-shifted SSFBG (0/π-SSFBG) and the ± π/2-phase-shifted SSFBG (± π/2-SSFBG) en/decoders in the three aspects: the security, coding and system performances. In terms of the security performance, we evaluate the security performance by the investigation on the encoded waveform of both encoders. We also propose and demonstrate the code extraction technique for the ± π/2-SSFBG encoder when input pulse has large pulse width. Then, we analyze the coding performance of these two kinds of en/decoders by the calculation of autocorrelation and cross-correlation with sets of 31-chip, 63-chip and 127-chip Gold codes. Furthermore, we propose and demonstrate the hybrid use of both en/decoders. To demonstrate the performance of both en/decoders and the hybrid use in the different systems, we employ four 31-chip 640 Gchip/s 0/π-SSFBG and ± π/2-SSFBG en/decoders in the 4-user 10 Gbps/user on-off keying and differential phase-shift keying OCDMA systems.

  13. ASCERTAINMENT OF THE EQUIVALENT CIRCUIT PARAMETERS OF THE ASYNCHRONOUS MACHINE

    Directory of Open Access Journals (Sweden)

    V. S. Safaryan

    2015-01-01

    Full Text Available The article considers experimental and analytical determination of the asynchronous machine equivalent-circuit parameters with application of the reference data. Transient processes investigation of the asynchronous machines necessitates the equivalent circuit parameters (resistance impedance, inductances and coefficient of the stator-rotor contours mutual inductance that help form the transitory-process mathematical simulation model. The reference books do not provide those parameters; they instead give the rated ones (active power, voltage, slide, coefficient of performance and capacity coefficient as well as the ratio of starting and nominal currents and torques. The noted studies on the asynchronous machine equivalent-circuits parametrization fail to solve the problems ad finem or solve them with admissions. The paper presents experimental and analytical determinations of the asynchronous machine equivalent-circuit parameters: the experimental one based on the results of two measurements and the analytical one where the problem boils down to solving a system of nonlineal algebraic equations. The authors investigate the equivalent asynchronous machine input-resistance properties and adduce the dependence curvatures of the input-resistances on the slide. They present a symbolic model for analytical parameterization of the asynchronous machine equivalent-circuit that represents a system of nonlineal equations and requires one of the rotor-parameters arbitrary assignment. The article demonstrates that for the asynchronous machine equivalent-circuit experimental parameterization the measures are to be conducted of the stator-circuit voltage, current and active power with two different slides and arbitrary assignment of one of the rotor parameters. The paper substantiates the fact that additional measurement does not discard the rotor-parameter choice arbitrariness. The authors establish that in motoring mode there is a critical slide by which the

  14. Effect of asynchronous updating on the stability of cellular automata

    International Nuclear Information System (INIS)

    Baetens, J.M.; Van der Weeën, P.; De Baets, B.

    2012-01-01

    Highlights: ► An upper bound on the Lyapunov exponent of asynchronously updated CA is established. ► The employed update method has repercussions on the stability of CAs. ► A decision on the employed update method should be taken with care. ► Substantial discrepancies arise between synchronously and asynchronously updated CA. ► Discrepancies between different asynchronous update schemes are less pronounced. - Abstract: Although cellular automata (CAs) were conceptualized as utter discrete mathematical models in which the states of all their spatial entities are updated simultaneously at every consecutive time step, i.e. synchronously, various CA-based models that rely on so-called asynchronous update methods have been constructed in order to overcome the limitations that are tied up with the classical way of evolving CAs. So far, only a few researchers have addressed the consequences of this way of updating on the evolved spatio-temporal patterns, and the reachable stationary states. In this paper, we exploit Lyapunov exponents to determine to what extent the stability of the rules within a family of totalistic CAs is affected by the underlying update method. For that purpose, we derive an upper bound on the maximum Lyapunov exponent of asynchronously iterated CAs, and show its validity, after which we present a comparative study between the Lyapunov exponents obtained for five different update methods, namely one synchronous method and four well-established asynchronous methods. It is found that the stability of CAs is seriously affected if one of the latter methods is employed, whereas the discrepancies arising between the different asynchronous methods are far less pronounced and, finally, we discuss the repercussions of our findings on the development of CA-based models.

  15. Fast decoding algorithms for coded aperture systems

    International Nuclear Information System (INIS)

    Byard, Kevin

    2014-01-01

    Fast decoding algorithms are described for a number of established coded aperture systems. The fast decoding algorithms for all these systems offer significant reductions in the number of calculations required when reconstructing images formed by a coded aperture system and hence require less computation time to produce the images. The algorithms may therefore be of use in applications that require fast image reconstruction, such as near real-time nuclear medicine and location of hazardous radioactive spillage. Experimental tests confirm the efficacy of the fast decoding techniques

  16. Detection of Failure in Asynchronous Motor Using Soft Computing Method

    Science.gov (United States)

    Vinoth Kumar, K.; Sony, Kevin; Achenkunju John, Alan; Kuriakose, Anto; John, Ano P.

    2018-04-01

    This paper investigates the stator short winding failure of asynchronous motor also their effects on motor current spectrums. A fuzzy logic approach i.e., model based technique possibly will help to detect the asynchronous motor failure. Actually, fuzzy logic similar to humanoid intelligent methods besides expected linguistic empowering inferences through vague statistics. The dynamic model is technologically advanced for asynchronous motor by means of fuzzy logic classifier towards investigate the stator inter turn failure in addition open phase failure. A hardware implementation was carried out with LabVIEW for the online-monitoring of faults.

  17. Modified Dynamic Decode-and-Forward Relaying Protocol for Type II Relay in LTE-Advanced and Beyond

    KAUST Repository

    Nam, Sung Sik; Alouini, Mohamed-Slim; Choi, Seyeong

    2016-01-01

    In this paper, we propose a modified dynamic decode-and-forward (MoDDF) relaying protocol to meet the critical requirements for user equipment (UE) relays in next-generation cellular systems (e.g., LTE-Advanced and beyond). The proposed MoDDF realizes the fast jump-in relaying and the sequential decoding with an application of random codeset to encoding and re-encoding process at the source and the multiple UE relays, respectively. A subframe-by-subframe decoding based on the accumulated (or buffered) messages is employed to achieve energy, information, or mixed combining. Finally, possible early termination of decoding at the end user can lead to the higher spectral efficiency and more energy saving by reducing the frequency of redundant subframe transmission and decoding. These attractive features eliminate the need of directly exchanging control messages between multiple UE relays and the end user, which is an important prerequisite for the practical UE relay deployment. Copyright: © 2016 Nam et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

  18. Modified Dynamic Decode-and-Forward Relaying Protocol for Type II Relay in LTE-Advanced and Beyond

    KAUST Repository

    Nam, Sung Sik

    2016-11-29

    In this paper, we propose a modified dynamic decode-and-forward (MoDDF) relaying protocol to meet the critical requirements for user equipment (UE) relays in next-generation cellular systems (e.g., LTE-Advanced and beyond). The proposed MoDDF realizes the fast jump-in relaying and the sequential decoding with an application of random codeset to encoding and re-encoding process at the source and the multiple UE relays, respectively. A subframe-by-subframe decoding based on the accumulated (or buffered) messages is employed to achieve energy, information, or mixed combining. Finally, possible early termination of decoding at the end user can lead to the higher spectral efficiency and more energy saving by reducing the frequency of redundant subframe transmission and decoding. These attractive features eliminate the need of directly exchanging control messages between multiple UE relays and the end user, which is an important prerequisite for the practical UE relay deployment. Copyright: © 2016 Nam et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

  19. Long-term collections

    CERN Multimedia

    Collectes à long terme

    2007-01-01

    The Committee of the Long Term Collections (CLT) asks for your attention for the following message from a young Peruvian scientist, following the earthquake which devastated part of her country a month ago.

  20. LDPC-based iterative joint source-channel decoding for JPEG2000.

    Science.gov (United States)

    Pu, Lingling; Wu, Zhenyu; Bilgin, Ali; Marcellin, Michael W; Vasic, Bane

    2007-02-01

    A framework is proposed for iterative joint source-channel decoding of JPEG2000 codestreams. At the encoder, JPEG2000 is used to perform source coding with certain error-resilience (ER) modes, and LDPC codes are used to perform channel coding. During decoding, the source decoder uses the ER modes to identify corrupt sections of the codestream and provides this information to the channel decoder. Decoding is carried out jointly in an iterative fashion. Experimental results indicate that the proposed method requires fewer iterations and improves overall system performance.

  1. Deep Learning Methods for Improved Decoding of Linear Codes

    Science.gov (United States)

    Nachmani, Eliya; Marciano, Elad; Lugosch, Loren; Gross, Warren J.; Burshtein, David; Be'ery, Yair

    2018-02-01

    The problem of low complexity, close to optimal, channel decoding of linear codes with short to moderate block length is considered. It is shown that deep learning methods can be used to improve a standard belief propagation decoder, despite the large example space. Similar improvements are obtained for the min-sum algorithm. It is also shown that tying the parameters of the decoders across iterations, so as to form a recurrent neural network architecture, can be implemented with comparable results. The advantage is that significantly less parameters are required. We also introduce a recurrent neural decoder architecture based on the method of successive relaxation. Improvements over standard belief propagation are also observed on sparser Tanner graph representations of the codes. Furthermore, we demonstrate that the neural belief propagation decoder can be used to improve the performance, or alternatively reduce the computational complexity, of a close to optimal decoder of short BCH codes.

  2. Bayesian population decoding of spiking neurons.

    Science.gov (United States)

    Gerwinn, Sebastian; Macke, Jakob; Bethge, Matthias

    2009-01-01

    The timing of action potentials in spiking neurons depends on the temporal dynamics of their inputs and contains information about temporal fluctuations in the stimulus. Leaky integrate-and-fire neurons constitute a popular class of encoding models, in which spike times depend directly on the temporal structure of the inputs. However, optimal decoding rules for these models have only been studied explicitly in the noiseless case. Here, we study decoding rules for probabilistic inference of a continuous stimulus from the spike times of a population of leaky integrate-and-fire neurons with threshold noise. We derive three algorithms for approximating the posterior distribution over stimuli as a function of the observed spike trains. In addition to a reconstruction of the stimulus we thus obtain an estimate of the uncertainty as well. Furthermore, we derive a 'spike-by-spike' online decoding scheme that recursively updates the posterior with the arrival of each new spike. We use these decoding rules to reconstruct time-varying stimuli represented by a Gaussian process from spike trains of single neurons as well as neural populations.

  3. Bayesian population decoding of spiking neurons

    Directory of Open Access Journals (Sweden)

    Sebastian Gerwinn

    2009-10-01

    Full Text Available The timing of action potentials in spiking neurons depends on the temporal dynamics of their inputs and contains information about temporal fluctuations in the stimulus. Leaky integrate-and-fire neurons constitute a popular class of encoding models, in which spike times depend directly on the temporal structure of the inputs. However, optimal decoding rules for these models have only been studied explicitly in the noiseless case. Here, we study decoding rules for probabilistic inference of a continuous stimulus from the spike times of a population of leaky integrate-and-fire neurons with threshold noise. We derive three algorithms for approximating the posterior distribution over stimuli as a function of the observed spike trains. In addition to a reconstruction of the stimulus we thus obtain an estimate of the uncertainty as well. Furthermore, we derive a `spike-by-spike' online decoding scheme that recursively updates the posterior with the arrival of each new spike. We use these decoding rules to reconstruct time-varying stimuli represented by a Gaussian process from spike trains of single neurons as well as neural populations.

  4. Three phase full wave dc motor decoder

    Science.gov (United States)

    Studer, P. A. (Inventor)

    1977-01-01

    A three phase decoder for dc motors is disclosed which employs an extremely simple six transistor circuit to derive six properly phased output signals for fullwave operation of dc motors. Six decoding transistors are coupled at their base-emitter junctions across a resistor network arranged in a delta configuration. Each point of the delta configuration is coupled to one of three position sensors which sense the rotational position of the motor. A second embodiment of the invention is disclosed in which photo-optical isolators are used in place of the decoding transistors.

  5. Decoding and Encoding Facial Expressions in Preschool-Age Children.

    Science.gov (United States)

    Zuckerman, Miron; Przewuzman, Sylvia J.

    1979-01-01

    Preschool-age children drew, decoded, and encoded facial expressions depicting five different emotions. Accuracy of drawing, decoding and encoding each of the five emotions was consistent across the three tasks; decoding ability was correlated with drawing ability among female subjects, but neither of these abilities was correlated with encoding…

  6. Belief propagation decoding of quantum channels by passing quantum messages

    International Nuclear Information System (INIS)

    Renes, Joseph M

    2017-01-01

    The belief propagation (BP) algorithm is a powerful tool in a wide range of disciplines from statistical physics to machine learning to computational biology, and is ubiquitous in decoding classical error-correcting codes. The algorithm works by passing messages between nodes of the factor graph associated with the code and enables efficient decoding of the channel, in some cases even up to the Shannon capacity. Here we construct the first BP algorithm which passes quantum messages on the factor graph and is capable of decoding the classical–quantum channel with pure state outputs. This gives explicit decoding circuits whose number of gates is quadratic in the code length. We also show that this decoder can be modified to work with polar codes for the pure state channel and as part of a decoder for transmitting quantum information over the amplitude damping channel. These represent the first explicit capacity-achieving decoders for non-Pauli channels. (fast track communication)

  7. Belief propagation decoding of quantum channels by passing quantum messages

    Science.gov (United States)

    Renes, Joseph M.

    2017-07-01

    The belief propagation (BP) algorithm is a powerful tool in a wide range of disciplines from statistical physics to machine learning to computational biology, and is ubiquitous in decoding classical error-correcting codes. The algorithm works by passing messages between nodes of the factor graph associated with the code and enables efficient decoding of the channel, in some cases even up to the Shannon capacity. Here we construct the first BP algorithm which passes quantum messages on the factor graph and is capable of decoding the classical-quantum channel with pure state outputs. This gives explicit decoding circuits whose number of gates is quadratic in the code length. We also show that this decoder can be modified to work with polar codes for the pure state channel and as part of a decoder for transmitting quantum information over the amplitude damping channel. These represent the first explicit capacity-achieving decoders for non-Pauli channels.

  8. Compensation for PKMζ in long-term potentiation and spatial long-term memory in mutant mice.

    Science.gov (United States)

    Tsokas, Panayiotis; Hsieh, Changchi; Yao, Yudong; Lesburguères, Edith; Wallace, Emma Jane Claire; Tcherepanov, Andrew; Jothianandan, Desingarao; Hartley, Benjamin Rush; Pan, Ling; Rivard, Bruno; Farese, Robert V; Sajan, Mini P; Bergold, Peter John; Hernández, Alejandro Iván; Cottrell, James E; Shouval, Harel Z; Fenton, André Antonio; Sacktor, Todd Charlton

    2016-05-17

    PKMζ is a persistently active PKC isoform proposed to maintain late-LTP and long-term memory. But late-LTP and memory are maintained without PKMζ in PKMζ-null mice. Two hypotheses can account for these findings. First, PKMζ is unimportant for LTP or memory. Second, PKMζ is essential for late-LTP and long-term memory in wild-type mice, and PKMζ-null mice recruit compensatory mechanisms. We find that whereas PKMζ persistently increases in LTP maintenance in wild-type mice, PKCι/λ, a gene-product closely related to PKMζ, persistently increases in LTP maintenance in PKMζ-null mice. Using a pharmacogenetic approach, we find PKMζ-antisense in hippocampus blocks late-LTP and spatial long-term memory in wild-type mice, but not in PKMζ-null mice without the target mRNA. Conversely, a PKCι/λ-antagonist disrupts late-LTP and spatial memory in PKMζ-null mice but not in wild-type mice. Thus, whereas PKMζ is essential for wild-type LTP and long-term memory, persistent PKCι/λ activation compensates for PKMζ loss in PKMζ-null mice.

  9. Asynchronous Operators of Sequential Logic Venjunction & Sequention

    CERN Document Server

    Vasyukevich, Vadim

    2011-01-01

    This book is dedicated to new mathematical instruments assigned for logical modeling of the memory of digital devices. The case in point is logic-dynamical operation named venjunction and venjunctive function as well as sequention and sequentional function. Venjunction and sequention operate within the framework of sequential logic. In a form of the corresponding equations, they organically fit analytical expressions of Boolean algebra. Thus, a sort of symbiosis is formed using elements of asynchronous sequential logic on the one hand and combinational logic on the other hand. So, asynchronous

  10. Design of 10Gbps optical encoder/decoder structure for FE-OCDMA system using SOA and opto-VLSI processors.

    Science.gov (United States)

    Aljada, Muhsen; Hwang, Seow; Alameh, Kamal

    2008-01-21

    In this paper we propose and experimentally demonstrate a reconfigurable 10Gbps frequency-encoded (1D) encoder/decoder structure for optical code division multiple access (OCDMA). The encoder is constructed using a single semiconductor optical amplifier (SOA) and 1D reflective Opto-VLSI processor. The SOA generates broadband amplified spontaneous emission that is dynamically sliced using digital phase holograms loaded onto the Opto-VLSI processor to generate 1D codewords. The selected wavelengths are injected back into the same SOA for amplifications. The decoder is constructed using single Opto-VLSI processor only. The encoded signal can successfully be retrieved at the decoder side only when the digital phase holograms of the encoder and the decoder are matched. The system performance is measured in terms of the auto-correlation and cross-correlation functions as well as the eye diagram.

  11. Sexuality and Physical Intimacy in Long Term Care: Sexuality, long term care, capacity assessment

    OpenAIRE

    Lichtenberg, Peter A.

    2014-01-01

    Sexuality and sexual needs in older adults remains a neglected area of clinical intervention, particularly so in long term care settings. Because older adults in medical rehabilitation and long term care beds present with significant frailties, and often significant neurocognitive disorders it makes it difficult for occupational therapists and other staff to evaluate the capacity of an older adult resident to participate in sexual relationships. The current paper reviews the current literatur...

  12. Differences in health status between long-term and short-term benzodiazepine users.

    NARCIS (Netherlands)

    Zandstra, S.M.; Furer, J.W.; Lisdonk, E.H. van de; Bor, J.H.J.; Zitman, F.G.; Weel, C. van

    2002-01-01

    BACKGROUND: Despite generally accepted advice to keep treatment short, benzodiazepines are often prescibed for more than six months. Prevention of long-term benzodiazepine use could be facilitated by the utilisation of risk indicators for long-term use. However, the characteristics of long-term

  13. Scientific Understanding from Long Term Observations: Insights from the Long Term Ecological Research (LTER) Program

    Science.gov (United States)

    Gosz, J.

    2001-12-01

    The network dedicated to Long Term Ecological Research (LTER) in the United States has grown to 24 sites since it was formed in 1980. Long-term research and monitoring are performed on parameters thatare basic to all ecosystems and are required to understand patterns, processes, and relationship to change. Collectively, the sites in the LTER Network provide opportunities to contrast marine, coastal, and continental regions, the full range of climatic gradients existing in North America, and aquatic and terrestrial habitats in a range of ecosystem types. The combination of common core areas and long-term research and monitoring in many habitats have allowed unprecedented abilities to understand and compare complex temporal and spatial dynamics associated with issues like climate change, effects of pollution, biodiversity and landuse. For example, McMurdo Dry Valley in the Antarctic has demonstrated an increase in glacier mass since 1993 which coincides with a period of cooler than normal summers and more than average snowfall. In contrast, the Bonanza Creek and Toolik Lake sites in Alaska have recorded a warming period unprecedented in the past 200 years. Nitrogen deposition effects have been identified through long-term watershed studies on biogeochemical cycles, especially at Coweeta Hydrological Lab, Harvard Forest, and the Hubbard Brook Experimental Forest. In aquatic systems, such as the Northern Temperate Lakes site, long-term data revealed time lags in effects of invaders and disturbance on lake communities. Biological recovery from an effect such as lake acidification was shown to lag behind chemical recovery. The long-term changes documented over 2 decades have been instrumental in influencing management practices in many of the LTER areas. In Puerto Rico, the Luquillo LTER demonstrated that dams obstruct migrations of fish and freshwater shrimp and water abstraction at low flows can completely obliterate downstream migration of juveniles and damage

  14. [A novel proposal explaining sleep disturbance of children in Japan--asynchronization].

    Science.gov (United States)

    Kohyama, Jun

    2008-07-01

    It has been reported that more than half of the children in Japan suffer from daytime sleepiness. In contrast, about one quarter of junior high-school students in Japan complain of insomnia. According to the International Classification of Sleep Disorders (Second edition), these children could be diagnosed as having behaviorally-induced insufficient sleep syndrome due to inadequate sleeping habits. Getting on adequate amount of sleep should solve such problems;however, such a therapeutic approach often fails. Although social factors are involved in these sleep disturbances, I feel that a novel notion - asynchronization - leads to an understanding of the pathophysiology of disturbances in these children. Further, it could contribute to resolve their problems. The essence of asynchronization is a disturbance of various aspects (e.g., cycle, amplitude, phase, and interrelationship) of the biological rhythms that normally exhibits circadian oscillation. The main cause of asynchronization is hypothesized to be the combination of light exposure during night and the lack of light exposure in the morning. Asynchronization results in the disturbance of variable systems. Thus, symptoms of asynchronization include disturbances of the autonomic nervous system (sleepiness, insomnia, disturbance of hormonal excretion, gastrointestinal problems, etc.) and higher brain function (disorientation, loss of sociality, loss of will or motivation, impaired alertness and performance, etc.). Neurological (attention deficit, aggression, impulsiveness, hyperactivity, etc.), psychiatric (depressive disorders, personality disorders, anxiety disorders, etc.) and somatic (tiredness, fatigue, etc.) disturbances could also be symptoms of asynchronization. At the initial phase of asynchronization, disturbances are functional and can be resolved relatively easily, such as by the establishment of a regular sleep-wakefulness cycle;however, without adequate intervention the disturbances could gradually

  15. Decoding Algorithms for Random Linear Network Codes

    DEFF Research Database (Denmark)

    Heide, Janus; Pedersen, Morten Videbæk; Fitzek, Frank

    2011-01-01

    We consider the problem of efficient decoding of a random linear code over a finite field. In particular we are interested in the case where the code is random, relatively sparse, and use the binary finite field as an example. The goal is to decode the data using fewer operations to potentially...... achieve a high coding throughput, and reduce energy consumption.We use an on-the-fly version of the Gauss-Jordan algorithm as a baseline, and provide several simple improvements to reduce the number of operations needed to perform decoding. Our tests show that the improvements can reduce the number...

  16. Very long-term sequelae of craniopharyngioma.

    Science.gov (United States)

    Wijnen, Mark; van den Heuvel-Eibrink, Marry M; Janssen, Joseph A M J L; Catsman-Berrevoets, Coriene E; Michiels, Erna M C; van Veelen-Vincent, Marie-Lise C; Dallenga, Alof H G; van den Berge, J Herbert; van Rij, Carolien M; van der Lely, Aart-Jan; Neggers, Sebastian J C M M

    2017-06-01

    Studies investigating long-term health conditions in patients with craniopharyngioma are limited by short follow-up durations and generally do not compare long-term health effects according to initial craniopharyngioma treatment approach. In addition, studies comparing long-term health conditions between patients with childhood- and adult-onset craniopharyngioma report conflicting results. The objective of this study was to analyse a full spectrum of long-term health effects in patients with craniopharyngioma according to initial treatment approach and age group at craniopharyngioma presentation. Cross-sectional study based on retrospective data. We studied a single-centre cohort of 128 patients with craniopharyngioma treated from 1980 onwards (63 patients with childhood-onset disease). Median follow-up since craniopharyngioma presentation was 13 years (interquartile range: 5-23 years). Initial craniopharyngioma treatment approaches included gross total resection ( n  = 25), subtotal resection without radiotherapy ( n  = 44), subtotal resection with radiotherapy ( n  = 25), cyst aspiration without radiotherapy ( n  = 8), and 90 Yttrium brachytherapy ( n  = 21). Pituitary hormone deficiencies (98%), visual disturbances (75%) and obesity (56%) were the most common long-term health conditions observed. Different initial craniopharyngioma treatment approaches resulted in similar long-term health effects. Patients with childhood-onset craniopharyngioma experienced significantly more growth hormone deficiency, diabetes insipidus, panhypopituitarism, morbid obesity, epilepsy and psychiatric conditions compared with patients with adult-onset disease. Recurrence-/progression-free survival was significantly lower after initial craniopharyngioma treatment with cyst aspiration compared with other therapeutic approaches. Survival was similar between patients with childhood- and adult-onset craniopharyngioma. Long-term health conditions were comparable after

  17. Long-term prisoner in prison isolation

    Directory of Open Access Journals (Sweden)

    Karolina Grudzińska

    2013-06-01

    Full Text Available Long-term prisoner belongs to a particular category of people who are imprisoned in prisons. On the one hand in this group are often heavily demoralized people who committed the most serious crimes, on the other hand it is a group of prisoners, who should be well thought out and programmed the impact of rehabilitation. The situation of man trapped for years poses in a complicated situation not only the prisoners, but also the entire prison staff. They have to take care of the fact that the prison isolation did not cause the state in which convicts form itself in learned helplessness and lack of skills for self-planning and decision-making. In addition, planning the rehabilitation impact of long-term prisoners should not be forgotten that these prisoners in the short or the long term will return to the libertarian environment therefore, should prevent any negative effects of long-term imprisonment. This article presents the main issues related to the execution of imprisonment against long-term prisoners. It is an attempt to systematize the knowledge of this category of people living in prison isolation.

  18. Localized radio frequency communication using asynchronous transfer mode protocol

    Science.gov (United States)

    Witzke, Edward L.; Robertson, Perry J.; Pierson, Lyndon G.

    2007-08-14

    A localized wireless communication system for communication between a plurality of circuit boards, and between electronic components on the circuit boards. Transceivers are located on each circuit board and electronic component. The transceivers communicate with one another over spread spectrum radio frequencies. An asynchronous transfer mode protocol controls communication flow with asynchronous transfer mode switches located on the circuit boards.

  19. Electrotechnics - AC motors. Asynchronous and brush-less motors - Lecture and exercises with solutions; Electrotechnique - Moteurs a courant alternatif. Moteurs asynchrones et brushless - Cours et problemes resolus

    Energy Technology Data Exchange (ETDEWEB)

    Jacob, D.

    2005-07-01

    This book proposes a presentation of AC electric motors essentially based on physics and technology. Its originality consists in avoiding to use mathematical formulations (like Park's transformation). The modeling retained, which only uses magnetic momentum, magnetic fields and reluctance concepts, leads simply and naturally to the vectorial control principle. The book develops some lecture elements which includes some topics rarely considered like the dimensioning of an asynchronous motor or of a single-phase brush-less motor. Experimental results illustrate the physical phenomena described and many original problems are resolved and commented at the end of each chapter. Content: signals and systems in electrotechnics, torque and rotating magnetic fields generation, asynchronous machine in permanent regime, speed variation of the asynchronous motor, special asynchronous motors, synchronous machine in permanent regime, brush-less motor, note about step motors, note about inverters, index. (J.S.)

  20. Behavioral Synthesis of Asynchronous Circuits Using Syntax Directed Translation as Backend

    DEFF Research Database (Denmark)

    Nielsen, Sune Fallgaard; Sparsø, Jens; Madsen, Jan

    2009-01-01

    The current state-of-the art in high-level synthesis of asynchronous circuits is syntax directed translation, which performs a one-to-one mapping of a HDL-description into a corresponding circuit. This paper presents a method for behavioral synthesis of asynchronous circuits which builds on top...... description language Balsa [1]. This ”conventional” template architecture allows us to adapt traditional synchronous synthesis techniques for resource sharing, scheduling, binding etc, to the domain of asynchronous circuits. A prototype tool has been implemented on top of the Balsa framework, and the method...... is illustrated through the implementation of a set of example circuits. The main contributions of the paper are: the fundamental idea, the template architecture and its implementation using asynchronous handshake components, and the implementation of a prototype tool....

  1. Dynamic modeling of an asynchronous squirrel-cage machine; Modelisation dynamique d'une machine asynchrone a cage

    Energy Technology Data Exchange (ETDEWEB)

    Guerette, D.

    2009-07-01

    This document presented a detailed mathematical explanation and validation of the steps leading to the development of an asynchronous squirrel-cage machine. The MatLab/Simulink software was used to model a wind turbine at variable high speeds. The asynchronous squirrel-cage machine is an electromechanical system coupled to a magnetic circuit. The resulting electromagnetic circuit can be represented as a set of resistances, leakage inductances and mutual inductances. Different models were used for a comparison study, including the Munteanu, Boldea, Wind Turbine Blockset, and SimPowerSystem. MatLab/Simulink modeling results were in good agreement with the results from other comparable models. Simulation results were in good agreement with analytical calculations. 6 refs, 2 tabs, 9 figs.

  2. Competitive short-term and long-term memory processes in spatial habituation.

    Science.gov (United States)

    Sanderson, David J; Bannerman, David M

    2011-04-01

    Exposure to a spatial location leads to habituation of exploration such that, in a novelty preference test, rodents subsequently prefer exploring a novel location to the familiar location. According to Wagner's (1981) theory of memory, short-term and long-term habituation are caused by separate and sometimes opponent processes. In the present study, this dual-process account of memory was tested. Mice received a series of exposure training trials to a location before receiving a novelty preference test. The novelty preference was greater when tested after a short, rather than a long, interval. In contrast, the novelty preference was weaker when exposure training trials were separated by a short, rather than a long interval. Furthermore, it was found that long-term habituation was determined by the independent effects of the amount of exposure training and the number of exposure training trials when factors such as the intertrial interval and the cumulative intertrial interval were controlled. A final experiment demonstrated that a long-term reduction of exploration could be caused by a negative priming effect due to associations formed during exploration. These results provide evidence against a single-process account of habituation and suggest that spatial habituation is determined by both short-term, recency-based memory and long-term, incrementally strengthened memory.

  3. Low-power priority Address-Encoder and Reset-Decoder data-driven readout for Monolithic Active Pixel Sensors for tracker system

    International Nuclear Information System (INIS)

    Yang, P.; Aglieri, G.; Cavicchioli, C.; Chalmet, P.L.; Chanlek, N.; Collu, A.; Gao, C.; Hillemanns, H.; Junique, A.; Kofarago, M.; Keil, M.; Kugathasan, T.; Kim, D.; Kim, J.; Lattuca, A.; Marin Tobon, C.A.; Marras, D.; Mager, M.; Martinengo, P.; Mazza, G.

    2015-01-01

    Active Pixel Sensors used in High Energy Particle Physics require low power consumption to reduce the detector material budget, low integration time to reduce the possibilities of pile-up and fast readout to improve the detector data capability. To satisfy these requirements, a novel Address-Encoder and Reset-Decoder (AERD) asynchronous circuit for a fast readout of a pixel matrix has been developed. The AERD data-driven readout architecture operates the address encoding and reset decoding based on an arbitration tree, and allows us to readout only the hit pixels. Compared to the traditional readout structure of the rolling shutter scheme in Monolithic Active Pixel Sensors (MAPS), AERD can achieve a low readout time and a low power consumption especially for low hit occupancies. The readout is controlled at the chip periphery with a signal synchronous with the clock, allows a good digital and analogue signal separation in the matrix and a reduction of the power consumption. The AERD circuit has been implemented in the TowerJazz 180 nm CMOS Imaging Sensor (CIS) process with full complementary CMOS logic in the pixel. It works at 10 MHz with a matrix height of 15 mm. The energy consumed to read out one pixel is around 72 pJ. A scheme to boost the readout speed to 40 MHz is also discussed. The sensor chip equipped with AERD has been produced and characterised. Test results including electrical beam measurement are presented

  4. Low-power priority Address-Encoder and Reset-Decoder data-driven readout for Monolithic Active Pixel Sensors for tracker system

    Energy Technology Data Exchange (ETDEWEB)

    Yang, P., E-mail: yangping0710@126.com [Central China Normal University, Wuhan (China); Aglieri, G.; Cavicchioli, C. [CERN, 1210 Geneva 23 (Switzerland); Chalmet, P.L. [MIND, Archamps (France); Chanlek, N. [Suranaree University of Technology, Nakhon Ratchasima (Thailand); Collu, A. [University of Cagliari, Cagliari (Italy); INFN (Italy); Gao, C. [Central China Normal University, Wuhan (China); Hillemanns, H.; Junique, A. [CERN, 1210 Geneva 23 (Switzerland); Kofarago, M. [CERN, 1210 Geneva 23 (Switzerland); University of Utrecht, Utrecht (Netherlands); Keil, M.; Kugathasan, T. [CERN, 1210 Geneva 23 (Switzerland); Kim, D. [Dongguk and Yonsei University, Seoul (Korea, Republic of); Kim, J. [Pusan National University, Busan (Korea, Republic of); Lattuca, A. [University of Torino, Torino (Italy); INFN (Italy); Marin Tobon, C.A. [CERN, 1210 Geneva 23 (Switzerland); Marras, D. [University of Cagliari, Cagliari (Italy); INFN (Italy); Mager, M.; Martinengo, P. [CERN, 1210 Geneva 23 (Switzerland); Mazza, G. [University of Torino, Torino (Italy); INFN (Italy); and others

    2015-06-11

    Active Pixel Sensors used in High Energy Particle Physics require low power consumption to reduce the detector material budget, low integration time to reduce the possibilities of pile-up and fast readout to improve the detector data capability. To satisfy these requirements, a novel Address-Encoder and Reset-Decoder (AERD) asynchronous circuit for a fast readout of a pixel matrix has been developed. The AERD data-driven readout architecture operates the address encoding and reset decoding based on an arbitration tree, and allows us to readout only the hit pixels. Compared to the traditional readout structure of the rolling shutter scheme in Monolithic Active Pixel Sensors (MAPS), AERD can achieve a low readout time and a low power consumption especially for low hit occupancies. The readout is controlled at the chip periphery with a signal synchronous with the clock, allows a good digital and analogue signal separation in the matrix and a reduction of the power consumption. The AERD circuit has been implemented in the TowerJazz 180 nm CMOS Imaging Sensor (CIS) process with full complementary CMOS logic in the pixel. It works at 10 MHz with a matrix height of 15 mm. The energy consumed to read out one pixel is around 72 pJ. A scheme to boost the readout speed to 40 MHz is also discussed. The sensor chip equipped with AERD has been produced and characterised. Test results including electrical beam measurement are presented.

  5. Low-power priority Address-Encoder and Reset-Decoder data-driven readout for Monolithic Active Pixel Sensors for tracker system

    Science.gov (United States)

    Yang, P.; Aglieri, G.; Cavicchioli, C.; Chalmet, P. L.; Chanlek, N.; Collu, A.; Gao, C.; Hillemanns, H.; Junique, A.; Kofarago, M.; Keil, M.; Kugathasan, T.; Kim, D.; Kim, J.; Lattuca, A.; Marin Tobon, C. A.; Marras, D.; Mager, M.; Martinengo, P.; Mazza, G.; Mugnier, H.; Musa, L.; Puggioni, C.; Rousset, J.; Reidt, F.; Riedler, P.; Snoeys, W.; Siddhanta, S.; Usai, G.; van Hoorne, J. W.; Yi, J.

    2015-06-01

    Active Pixel Sensors used in High Energy Particle Physics require low power consumption to reduce the detector material budget, low integration time to reduce the possibilities of pile-up and fast readout to improve the detector data capability. To satisfy these requirements, a novel Address-Encoder and Reset-Decoder (AERD) asynchronous circuit for a fast readout of a pixel matrix has been developed. The AERD data-driven readout architecture operates the address encoding and reset decoding based on an arbitration tree, and allows us to readout only the hit pixels. Compared to the traditional readout structure of the rolling shutter scheme in Monolithic Active Pixel Sensors (MAPS), AERD can achieve a low readout time and a low power consumption especially for low hit occupancies. The readout is controlled at the chip periphery with a signal synchronous with the clock, allows a good digital and analogue signal separation in the matrix and a reduction of the power consumption. The AERD circuit has been implemented in the TowerJazz 180 nm CMOS Imaging Sensor (CIS) process with full complementary CMOS logic in the pixel. It works at 10 MHz with a matrix height of 15 mm. The energy consumed to read out one pixel is around 72 pJ. A scheme to boost the readout speed to 40 MHz is also discussed. The sensor chip equipped with AERD has been produced and characterised. Test results including electrical beam measurement are presented.

  6. Oppositional Decoding as an Act of Resistance.

    Science.gov (United States)

    Steiner, Linda

    1988-01-01

    Argues that contributors to the "No Comment" feature of "Ms." magazine are engaging in oppositional decoding and speculates on why this is a satisfying group process. Also notes such decoding presents another challenge to the idea that mass media has the same effect on all audiences. (SD)

  7. Asynchronous and Synchronous Online Discussion: Real and Perceived Achievement Differences

    Science.gov (United States)

    Johnson, Genevieve Marie; Buck, George H.

    2007-01-01

    Students in an introductory educational psychology course used two WebCT communication tools (synchronous chat and asynchronous discussion) to discuss four case studies. In response to the item, "I learned the case studies best when using," 39 students selected synchronous chat and 51 students selected asynchronous discussion. Students who…

  8. Long term liquidity analysis of the firm

    Directory of Open Access Journals (Sweden)

    Jaroslav Gonos

    2009-09-01

    Full Text Available Liquidity control is a very difficult and important function. If the business is not liquid in the long term, it is under threatof bankruptcy, and on the other hand surplus of the cash in hand threaten its future efficiency, because the cash in hand is a sourceof only limited profitability. Long term liquidity is related to the ability of the short term and long term liabilities payment. Articleis trying to point out to the monitoring and analyzing of the long term liquidity in the concrete business, in this case the printing industrycompany. Hereby at the end of the article mentioned monitored and analyzed liquidity is evaluated in the five years time period.

  9. DESIGN METHODOLOGY OF SELF-EXCITED ASYNCHRONOUS GENERATOR

    Directory of Open Access Journals (Sweden)

    Berzan V.

    2012-04-01

    Full Text Available The paper sets out the methodology of designing an asynchronous generator with capacitive self-excitation. It is known that its design is possible on the basis of serial synchronous motor with squirrel cage rotor. With this approach, the design reworked only the stator winding of electrical machines, making it cost-effectively implement the creation of the generator. Therefore, the methodology for the design, optimization calculations, the development scheme and the stator winding excitation system gain, not only of practical interest, and may also be useful for specialists in the field of electrical machines in the design of asynchronous generators.

  10. Handbook of asynchronous machines with variable speed

    CERN Document Server

    Razik, Hubert

    2013-01-01

    This handbook deals with the asynchronous machine in its close environment. It was born from a reflection on this electromagnetic converter whose integration in industrial environments takes a wide part. Previously this type of motor operated at fixed speed, from now on it has been integrated more and more in processes at variable speed. For this reason it seemed useful, or necessary, to write a handbook on the various aspects from the motor in itself, via the control and while finishing by the diagnosis aspect. Indeed, an asynchronous motor is used nowadays in industry where variation speed a

  11. Reliable self-replicating machines in asynchronous cellular automata.

    Science.gov (United States)

    Lee, Jia; Adachi, Susumu; Peper, Ferdinand

    2007-01-01

    We propose a self-replicating machine that is embedded in a two-dimensional asynchronous cellular automaton with von Neumann neighborhood. The machine dynamically encodes its shape into description signals, and despite the randomness of cell updating, it is able to successfully construct copies of itself according to the description signals. Self-replication on asynchronously updated cellular automata may find application in nanocomputers, where reconfigurability is an essential property, since it allows avoidance of defective parts and simplifies programming of such computers.

  12. Generalized Sudan's List Decoding for Order Domain Codes

    DEFF Research Database (Denmark)

    Geil, Hans Olav; Matsumoto, Ryutaroh

    2007-01-01

    We generalize Sudan's list decoding algorithm without multiplicity to evaluation codes coming from arbitrary order domains. The number of correctable errors by the proposed method is larger than the original list decoding without multiplicity....

  13. Low-power Implementation of an Encryption/Decryption System with Asynchronous Techniques

    Directory of Open Access Journals (Sweden)

    Nikos Sklavos

    2002-01-01

    Full Text Available An asynchronous VLSI implementation of the International Data Encryption Algorithm (IDEA is presented in this paper. In order to evaluate the asynchronous design a synchronous version of the algorithm was also designed. VHDL hardware description language was used in order to describe the algorithm. By using Synopsys commercial available tools the VHDL code was synthesized. After placing and routing both designs were fabricated with 0.6 μm CMOS technology. With a system clock of up to 8 MHz and a power supply of 5 V the two chips were tested and evaluated comparing with the software implementation of the IDEA algorithm. This new approach proves efficiently the lowest power consumption of the asynchronous implementation compared to the existing synchronous. Therefore, the asynchronous chip performs efficiently in Wireless Encryption Protocols and high speed networks.

  14. Advanced surveillance technologies for used fuel long-term storage and transportation - 59032

    International Nuclear Information System (INIS)

    Tsai, Hanchung; Liu, Yung Y.; Nutt, Mark; Shuler, James

    2012-01-01

    Utilities worldwide are using dry-cask storage systems to handle the ever-increasing number of discharged fuel assemblies from nuclear power plants. In the United States and possibly elsewhere, this trend will continue until an acceptable disposal path is established. The recent Fukushima nuclear power plant accident, specifically the events with the storage pools, may accelerate the drive to relocate more of the used fuel assemblies from pools into dry casks. Many of the newer cask systems incorporate dual-purpose (storage and transport) or multiple-purpose (storage, transport, and disposal) canister technologies. With the prospect looming for very long term storage - possibly over multiple decades - and deferred transport, condition- and performance-based aging management of cask structures and components is now a necessity that requires immediate attention. From the standpoint of consequences, one of the greatest concerns is the rupture of a substantial number of fuel rods that would affect fuel retrievability. Used fuel cladding may become susceptible to rupture due to radial-hydride-induced embrittlement caused by water-side corrosion during the reactor operation and subsequent drying/transfer process, through early stage of storage in a dry cask, especially for high burnup fuels. Radio frequency identification (RFID) is an automated data capture and remote-sensing technology ideally suited for monitoring sensitive assets on a long-term, continuous basis. One such system, called ARG-US, has been developed by Argonne National Laboratory for the U.S. Department of Energy's Packaging Certification Program for tracking and monitoring drums containing sensitive nuclear and radioactive materials. The ARG-US RFID system is versatile and can be readily adapted for dry-cask monitoring applications. The current built-in sensor suite consists of seal, temperature, humidity, shock, and radiation sensors. With the universal asynchronous receiver/transmitter interface in

  15. Long-term contracts vs. short-term trade of natural gas - a European perspective

    International Nuclear Information System (INIS)

    Neuhoff, Karsten; Hirschhausen, Christian von

    2005-01-01

    This paper analyses the economics of long-term gas contracts under changing institutional conditions, mainly gas sector liberalisation. The paper is motivated by the increasingly tense debate in continental Europe, UK and the US on the security of long-term gas supply. We discuss the main issues regarding long-term contracts, i.e. the changing role of the flexibility clause, the effect of abandoning the destination clause, and the strategic behaviour of producers between long-term sales and spot-sales. The literature suggests consumers and producers benefit from risk hedging through long-term contracts. Furthermore long-term contracts may reduce exercise of market power. Our analysis adds an additional benefit if the long-run demand elasticity is significantly lower than the short-run elasticity, both strategic producers and consumers benefit from lower prices and larger market volume. Some policy implications of the findings are also discussed. (Author)

  16. The uranium industry: long-term planning for short-term competition

    International Nuclear Information System (INIS)

    Vottero, X.; Georges Capus, G.

    2001-01-01

    Long term planning for short term competition Today, uranium producers face new challenges in terms of both production (new regulatory, environmental and social constraints) and market conditions (new sources of uranium supply, very low prices and tough competition). In such a context, long-term planning is not just a prerequisite to survive in the nuclear fuel cycle industry. In fact, it also contributes to sustaining nuclear electricity generation facing fierce competition from other energy sources in increasingly deregulated markets. Firstly, the risk of investing in new mining projects in western countries is growing because, on the one hand, of very erratic market conditions and, on the other hand, of increasingly lengthy, complex and unpredictable regulatory conditions. Secondly, the supply of other sources of uranium (uranium derived from nuclear weapons, uranium produced in CIS countries, ...) involve other risks, mainly related to politics and commercial restrictions. Consequently, competitive uranium supply requires not only technical competence but also financial strength and good marketing capabilities in order to anticipate long-term market trends, in terms of both demand and supply. It also requires taking into account new parameters such as politics, environment, regulations, etc. Today, a supplier dedicated to the sustainable production of nuclear electricity must manage a broad range of long-term risks inherent to the procurement of uranium. Taking into account all these parameters in a context of short-term, fast-changing market is a great challenge for the future generation. World Uranium Civilian Supply and Demand. (authors)

  17. Long-term biodosimetry Redux

    International Nuclear Information System (INIS)

    Simon, Steven L.; Bouville, Andre

    2016-01-01

    This paper revisits and reiterates the needs, purposes and requirements of bio-dosimetric assays for long-term dose and health risk assessments. While the most crucial need for bio-dosimetric assays is to guide medical response for radiation accidents, the value of such techniques for improving our understanding of radiation health risk by supporting epidemiological (long-term health risk) studies is significant. As new cohorts of exposed persons are identified and new health risk studies are undertaken with the hopes that studying the exposed will result in a deeper understanding of radiation risk, the value of reliable dose reconstruction is underscored. The ultimate application of biodosimetry in long-term health risk studies would be to completely replace model-based dose reconstruction-a complex suite of methods for retrospectively estimating dose that is commonly fraught with large uncertainties due to the absence of important exposure-related information, as well as imperfect models. While biodosimetry could potentially supplant model-based doses, there are numerous limitations of presently available techniques that constrain their widespread application in health risk research, including limited ability to assess doses received far in the past, high cost, great inter-individual variability, invasiveness, higher than preferred detection limits and the inability to assess internal dose (for the most part). These limitations prevent the extensive application of biodosimetry to large cohorts and should be considered a challenge to researchers to develop new and more flexible techniques that meet the demands of long-term health risk research. Events in recent years, e.g. the Fukushima reactor accident and the increased threat of nuclear terrorism, underscore that any event that results in significant radiation exposures of a group of people will also produce a much larger population, exposed at lower levels, but that likewise needs (or demands) an exposure

  18. Long term radioactive waste management

    International Nuclear Information System (INIS)

    Lavie, J.M.

    1984-01-01

    In France, waste management, a sensitive issue in term of public opinion, is developing quickly, and due to twenty years of experience, is now reaching maturity. With the launching of the French nuclear programme, the use of radioactive sources in radiotherapy and industry, waste management has become an industrial activity. Waste management is an integrated system dealing with the wastes from their production to the long term disposal, including their identification, sortage, treatment, packaging, collection and transport. This system aims at guaranteing the protection of present and future populations with an available technology. In regard to their long term management, and the design of disposals, radioactive wastes are divided in three categories. This classification takes into account the different radioisotopes contained, their half life and their total activity. Presently short-lived wastes are stored in the shallowland disposal of the ''Centre de la Manche''. Set up within the French Atomic Energy Commission (CEA), the National Agency for waste management (ANDRA) is responsible within the framework of legislative and regulatory provisions for long term waste management in France [fr

  19. Modeling long-term dynamics of electricity markets

    International Nuclear Information System (INIS)

    Olsina, Fernando; Garces, Francisco; Haubrich, H.-J.

    2006-01-01

    In the last decade, many countries have restructured their electricity industries by introducing competition in their power generation sectors. Although some restructuring has been regarded as successful, the short experience accumulated with liberalized power markets does not allow making any founded assertion about their long-term behavior. Long-term prices and long-term supply reliability are now center of interest. This concerns firms considering investments in generation capacity and regulatory authorities interested in assuring the long-term supply adequacy and the stability of power markets. In order to gain significant insight into the long-term behavior of liberalized power markets, in this paper, a simulation model based on system dynamics is proposed and the underlying mathematical formulations extensively discussed. Unlike classical market models based on the assumption that market outcomes replicate the results of a centrally made optimization, the approach presented here focuses on replicating the system structure of power markets and the logic of relationships among system components in order to derive its dynamical response. The simulations suggest that there might be serious problems to adjust early enough the generation capacity necessary to maintain stable reserve margins, and consequently, stable long-term price levels. Because of feedback loops embedded in the structure of power markets and the existence of some time lags, the long-term market development might exhibit a quite volatile behavior. By varying some exogenous inputs, a sensitivity analysis is carried out to assess the influence of these factors on the long-run market dynamics

  20. Short-term memory and long-term memory are still different.

    Science.gov (United States)

    Norris, Dennis

    2017-09-01

    A commonly expressed view is that short-term memory (STM) is nothing more than activated long-term memory. If true, this would overturn a central tenet of cognitive psychology-the idea that there are functionally and neurobiologically distinct short- and long-term stores. Here I present an updated case for a separation between short- and long-term stores, focusing on the computational demands placed on any STM system. STM must support memory for previously unencountered information, the storage of multiple tokens of the same type, and variable binding. None of these can be achieved simply by activating long-term memory. For example, even a simple sequence of digits such as "1, 3, 1" where there are 2 tokens of the digit "1" cannot be stored in the correct order simply by activating the representations of the digits "1" and "3" in LTM. I also review recent neuroimaging data that has been presented as evidence that STM is activated LTM and show that these data are exactly what one would expect to see based on a conventional 2-store view. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  1. Turbo decoder architecture for beyond-4G applications

    CERN Document Server

    Wong, Cheng-Chi

    2013-01-01

    This book describes the most recent techniques for turbo decoder implementation, especially for 4G and beyond 4G applications. The authors reveal techniques for the design of high-throughput decoders for future telecommunication systems, enabling designers to reduce hardware cost and shorten processing time. Coverage includes an explanation of VLSI implementation of the turbo decoder, from basic functional units to advanced parallel architecture. The authors discuss both hardware architecture techniques and experimental results, showing the variations in area/throughput/performance with respec

  2. Hybrid EEG-fNIRS Asynchronous Brain-Computer Interface for Multiple Motor Tasks.

    Directory of Open Access Journals (Sweden)

    Alessio Paolo Buccino

    Full Text Available Non-invasive Brain-Computer Interfaces (BCI have demonstrated great promise for neuroprosthetics and assistive devices. Here we aim to investigate methods to combine Electroencephalography (EEG and functional Near-Infrared Spectroscopy (fNIRS in an asynchronous Sensory Motor rhythm (SMR-based BCI. We attempted to classify 4 different executed movements, namely, Right-Arm-Left-Arm-Right-Hand-Left-Hand tasks. Previous studies demonstrated the benefit of EEG-fNIRS combination. However, since normally fNIRS hemodynamic response shows a long delay, we investigated new features, involving slope indicators, in order to immediately detect changes in the signals. Moreover, Common Spatial Patterns (CSPs have been applied to both EEG and fNIRS signals. 15 healthy subjects took part in the experiments and since 25 trials per class were available, CSPs have been regularized with information from the entire population of participants and optimized using genetic algorithms. The different features have been compared in terms of performance and the dynamic accuracy over trials shows that the introduced methods diminish the fNIRS delay in the detection of changes.

  3. A best-first tree-searching approach for ML decoding in MIMO system

    KAUST Repository

    Shen, Chung-An

    2012-07-28

    In MIMO communication systems maximum-likelihood (ML) decoding can be formulated as a tree-searching problem. This paper presents a tree-searching approach that combines the features of classical depth-first and breadth-first approaches to achieve close to ML performance while minimizing the number of visited nodes. A detailed outline of the algorithm is given, including the required storage. The effects of storage size on BER performance and complexity in terms of search space are also studied. Our result demonstrates that with a proper choice of storage size the proposed method visits 40% fewer nodes than a sphere decoding algorithm at signal to noise ratio (SNR) = 20dB and by an order of magnitude at 0 dB SNR.

  4. Mapping visual stimuli to perceptual decisions via sparse decoding of mesoscopic neural activity.

    Science.gov (United States)

    Sajda, Paul

    2010-01-01

    In this talk I will describe our work investigating sparse decoding of neural activity, given a realistic mapping of the visual scene to neuronal spike trains generated by a model of primary visual cortex (V1). We use a linear decoder which imposes sparsity via an L1 norm. The decoder can be viewed as a decoding neuron (linear summation followed by a sigmoidal nonlinearity) in which there are relatively few non-zero synaptic weights. We find: (1) the best decoding performance is for a representation that is sparse in both space and time, (2) decoding of a temporal code results in better performance than a rate code and is also a better fit to the psychophysical data, (3) the number of neurons required for decoding increases monotonically as signal-to-noise in the stimulus decreases, with as little as 1% of the neurons required for decoding at the highest signal-to-noise levels, and (4) sparse decoding results in a more accurate decoding of the stimulus and is a better fit to psychophysical performance than a distributed decoding, for example one imposed by an L2 norm. We conclude that sparse coding is well-justified from a decoding perspective in that it results in a minimum number of neurons and maximum accuracy when sparse representations can be decoded from the neural dynamics.

  5. Discrete decoding based ultrafast multidimensional nuclear magnetic resonance spectroscopy

    International Nuclear Information System (INIS)

    Wei, Zhiliang; Lin, Liangjie; Ye, Qimiao; Li, Jing; Cai, Shuhui; Chen, Zhong

    2015-01-01

    The three-dimensional (3D) nuclear magnetic resonance (NMR) spectroscopy constitutes an important and powerful tool in analyzing chemical and biological systems. However, the abundant 3D information arrives at the expense of long acquisition times lasting hours or even days. Therefore, there has been a continuous interest in developing techniques to accelerate recordings of 3D NMR spectra, among which the ultrafast spatiotemporal encoding technique supplies impressive acquisition speed by compressing a multidimensional spectrum in a single scan. However, it tends to suffer from tradeoffs among spectral widths in different dimensions, which deteriorates in cases of NMR spectroscopy with more dimensions. In this study, the discrete decoding is proposed to liberate the ultrafast technique from tradeoffs among spectral widths in different dimensions by focusing decoding on signal-bearing sites. For verifying its feasibility and effectiveness, we utilized the method to generate two different types of 3D spectra. The proposed method is also applicable to cases with more than three dimensions, which, based on the experimental results, may widen applications of the ultrafast technique

  6. Discrete decoding based ultrafast multidimensional nuclear magnetic resonance spectroscopy

    Science.gov (United States)

    Wei, Zhiliang; Lin, Liangjie; Ye, Qimiao; Li, Jing; Cai, Shuhui; Chen, Zhong

    2015-07-01

    The three-dimensional (3D) nuclear magnetic resonance (NMR) spectroscopy constitutes an important and powerful tool in analyzing chemical and biological systems. However, the abundant 3D information arrives at the expense of long acquisition times lasting hours or even days. Therefore, there has been a continuous interest in developing techniques to accelerate recordings of 3D NMR spectra, among which the ultrafast spatiotemporal encoding technique supplies impressive acquisition speed by compressing a multidimensional spectrum in a single scan. However, it tends to suffer from tradeoffs among spectral widths in different dimensions, which deteriorates in cases of NMR spectroscopy with more dimensions. In this study, the discrete decoding is proposed to liberate the ultrafast technique from tradeoffs among spectral widths in different dimensions by focusing decoding on signal-bearing sites. For verifying its feasibility and effectiveness, we utilized the method to generate two different types of 3D spectra. The proposed method is also applicable to cases with more than three dimensions, which, based on the experimental results, may widen applications of the ultrafast technique.

  7. The Differential Contributions of Auditory-Verbal and Visuospatial Working Memory on Decoding Skills in Children Who Are Poor Decoders

    Science.gov (United States)

    Squires, Katie Ellen

    2013-01-01

    This study investigated the differential contribution of auditory-verbal and visuospatial working memory (WM) on decoding skills in second- and fifth-grade children identified with poor decoding. Thirty-two second-grade students and 22 fifth-grade students completed measures that assessed simple and complex auditory-verbal and visuospatial memory,…

  8. Long-Term Collections

    CERN Multimedia

    Comité des collectes à long terme

    2011-01-01

    It is the time of the year when our fireman colleagues go around the laboratory for their traditional calendars sale. A part of the money of the sales will be donated in favour of the long-term collections. We hope that you will welcome them warmly.

  9. Application of intelligent soft start in asynchronous motor

    Science.gov (United States)

    Du, Xue; Ye, Ying; Wang, Yuelong; Peng, Lei; Zhang, Suying

    2018-05-01

    The starting way of three phase asynchronous motor has full voltage start and step-down start. Direct starting brings large current impact, causing excessive local temperature to the power grid and larger starting torque will also impact the motor equipment and affect the service life of the motor. Aim at the problem of large current and torque caused by start-up, an intelligent soft starter is proposed. Through the application of intelligent soft start on asynchronous motor, highlights its application advantage in motor control.

  10. Decoding using back-project algorithm from coded image in ICF

    International Nuclear Information System (INIS)

    Jiang shaoen; Liu Zhongli; Zheng Zhijian; Tang Daoyuan

    1999-01-01

    The principle of the coded imaging and its decoding in inertial confinement fusion is described simply. The authors take ring aperture microscope for example and use back-project (BP) algorithm to decode the coded image. The decoding program has been performed for numerical simulation. Simulations of two models are made, and the results show that the accuracy of BP algorithm is high and effect of reconstruction is good. Thus, it indicates that BP algorithm is applicable to decoding for coded image in ICF experiments

  11. A real-time MPEG software decoder using a portable message-passing library

    Energy Technology Data Exchange (ETDEWEB)

    Kwong, Man Kam; Tang, P.T. Peter; Lin, Biquan

    1995-12-31

    We present a real-time MPEG software decoder that uses message-passing libraries such as MPL, p4 and MPI. The parallel MPEG decoder currently runs on the IBM SP system but can be easil ported to other parallel machines. This paper discusses our parallel MPEG decoding algorithm as well as the parallel programming environment under which it uses. Several technical issues are discussed, including balancing of decoding speed, memory limitation, 1/0 capacities, and optimization of MPEG decoding components. This project shows that a real-time portable software MPEG decoder is feasible in a general-purpose parallel machine.

  12. Efficient universal computing architectures for decoding neural activity.

    Directory of Open Access Journals (Sweden)

    Benjamin I Rapoport

    Full Text Available The ability to decode neural activity into meaningful control signals for prosthetic devices is critical to the development of clinically useful brain- machine interfaces (BMIs. Such systems require input from tens to hundreds of brain-implanted recording electrodes in order to deliver robust and accurate performance; in serving that primary function they should also minimize power dissipation in order to avoid damaging neural tissue; and they should transmit data wirelessly in order to minimize the risk of infection associated with chronic, transcutaneous implants. Electronic architectures for brain- machine interfaces must therefore minimize size and power consumption, while maximizing the ability to compress data to be transmitted over limited-bandwidth wireless channels. Here we present a system of extremely low computational complexity, designed for real-time decoding of neural signals, and suited for highly scalable implantable systems. Our programmable architecture is an explicit implementation of a universal computing machine emulating the dynamics of a network of integrate-and-fire neurons; it requires no arithmetic operations except for counting, and decodes neural signals using only computationally inexpensive logic operations. The simplicity of this architecture does not compromise its ability to compress raw neural data by factors greater than [Formula: see text]. We describe a set of decoding algorithms based on this computational architecture, one designed to operate within an implanted system, minimizing its power consumption and data transmission bandwidth; and a complementary set of algorithms for learning, programming the decoder, and postprocessing the decoded output, designed to operate in an external, nonimplanted unit. The implementation of the implantable portion is estimated to require fewer than 5000 operations per second. A proof-of-concept, 32-channel field-programmable gate array (FPGA implementation of this portion

  13. Robust pattern decoding in shape-coded structured light

    Science.gov (United States)

    Tang, Suming; Zhang, Xu; Song, Zhan; Song, Lifang; Zeng, Hai

    2017-09-01

    Decoding is a challenging and complex problem in a coded structured light system. In this paper, a robust pattern decoding method is proposed for the shape-coded structured light in which the pattern is designed as grid shape with embedded geometrical shapes. In our decoding method, advancements are made at three steps. First, a multi-template feature detection algorithm is introduced to detect the feature point which is the intersection of each two orthogonal grid-lines. Second, pattern element identification is modelled as a supervised classification problem and the deep neural network technique is applied for the accurate classification of pattern elements. Before that, a training dataset is established, which contains a mass of pattern elements with various blurring and distortions. Third, an error correction mechanism based on epipolar constraint, coplanarity constraint and topological constraint is presented to reduce the false matches. In the experiments, several complex objects including human hand are chosen to test the accuracy and robustness of the proposed method. The experimental results show that our decoding method not only has high decoding accuracy, but also owns strong robustness to surface color and complex textures.

  14. Long-Term Memory Performance in Adult ADHD.

    Science.gov (United States)

    Skodzik, Timo; Holling, Heinz; Pedersen, Anya

    2017-02-01

    Memory problems are a frequently reported symptom in adult ADHD, and it is well-documented that adults with ADHD perform poorly on long-term memory tests. However, the cause of this effect is still controversial. The present meta-analysis examined underlying mechanisms that may lead to long-term memory impairments in adult ADHD. We performed separate meta-analyses of measures of memory acquisition and long-term memory using both verbal and visual memory tests. In addition, the influence of potential moderator variables was examined. Adults with ADHD performed significantly worse than controls on verbal but not on visual long-term memory and memory acquisition subtests. The long-term memory deficit was strongly statistically related to the memory acquisition deficit. In contrast, no retrieval problems were observable. Our results suggest that memory deficits in adult ADHD reflect a learning deficit induced at the stage of encoding. Implications for clinical and research settings are presented.

  15. Performance Analysis of CRC Codes for Systematic and Nonsystematic Polar Codes with List Decoding

    Directory of Open Access Journals (Sweden)

    Takumi Murata

    2018-01-01

    Full Text Available Successive cancellation list (SCL decoding of polar codes is an effective approach that can significantly outperform the original successive cancellation (SC decoding, provided that proper cyclic redundancy-check (CRC codes are employed at the stage of candidate selection. Previous studies on CRC-assisted polar codes mostly focus on improvement of the decoding algorithms as well as their implementation, and little attention has been paid to the CRC code structure itself. For the CRC-concatenated polar codes with CRC code as their outer code, the use of longer CRC code leads to reduction of information rate, whereas the use of shorter CRC code may reduce the error detection probability, thus degrading the frame error rate (FER performance. Therefore, CRC codes of proper length should be employed in order to optimize the FER performance for a given signal-to-noise ratio (SNR per information bit. In this paper, we investigate the effect of CRC codes on the FER performance of polar codes with list decoding in terms of the CRC code length as well as its generator polynomials. Both the original nonsystematic and systematic polar codes are considered, and we also demonstrate that different behaviors of CRC codes should be observed depending on whether the inner polar code is systematic or not.

  16. Optimal and efficient decoding of concatenated quantum block codes

    International Nuclear Information System (INIS)

    Poulin, David

    2006-01-01

    We consider the problem of optimally decoding a quantum error correction code--that is, to find the optimal recovery procedure given the outcomes of partial ''check'' measurements on the system. In general, this problem is NP hard. However, we demonstrate that for concatenated block codes, the optimal decoding can be efficiently computed using a message-passing algorithm. We compare the performance of the message-passing algorithm to that of the widespread blockwise hard decoding technique. Our Monte Carlo results using the five-qubit and Steane's code on a depolarizing channel demonstrate significant advantages of the message-passing algorithms in two respects: (i) Optimal decoding increases by as much as 94% the error threshold below which the error correction procedure can be used to reliably send information over a noisy channel; and (ii) for noise levels below these thresholds, the probability of error after optimal decoding is suppressed at a significantly higher rate, leading to a substantial reduction of the error correction overhead

  17. Distributed Consensus of Stochastic Delayed Multi-agent Systems Under Asynchronous Switching.

    Science.gov (United States)

    Wu, Xiaotai; Tang, Yang; Cao, Jinde; Zhang, Wenbing

    2016-08-01

    In this paper, the distributed exponential consensus of stochastic delayed multi-agent systems with nonlinear dynamics is investigated under asynchronous switching. The asynchronous switching considered here is to account for the time of identifying the active modes of multi-agent systems. After receipt of confirmation of mode's switching, the matched controller can be applied, which means that the switching time of the matched controller in each node usually lags behind that of system switching. In order to handle the coexistence of switched signals and stochastic disturbances, a comparison principle of stochastic switched delayed systems is first proved. By means of this extended comparison principle, several easy to verified conditions for the existence of an asynchronously switched distributed controller are derived such that stochastic delayed multi-agent systems with asynchronous switching and nonlinear dynamics can achieve global exponential consensus. Two examples are given to illustrate the effectiveness of the proposed method.

  18. Error Recovery Properties and Soft Decoding of Quasi-Arithmetic Codes

    Directory of Open Access Journals (Sweden)

    Christine Guillemot

    2007-08-01

    Full Text Available This paper first introduces a new set of aggregated state models for soft-input decoding of quasi arithmetic (QA codes with a termination constraint. The decoding complexity with these models is linear with the sequence length. The aggregation parameter controls the tradeoff between decoding performance and complexity. It is shown that close-to-optimal decoding performance can be obtained with low values of the aggregation parameter, that is, with a complexity which is significantly reduced with respect to optimal QA bit/symbol models. The choice of the aggregation parameter depends on the synchronization recovery properties of the QA codes. This paper thus describes a method to estimate the probability mass function (PMF of the gain/loss of symbols following a single bit error (i.e., of the difference between the number of encoded and decoded symbols. The entropy of the gain/loss turns out to be the average amount of information conveyed by a length constraint on both the optimal and aggregated state models. This quantity allows us to choose the value of the aggregation parameter that will lead to close-to-optimal decoding performance. It is shown that the optimum position for the length constraint is not the last time instant of the decoding process. This observation leads to the introduction of a new technique for robust decoding of QA codes with redundancy which turns out to outperform techniques based on the concept of forbidden symbol.

  19. A Long-term Plan for Kalk

    DEFF Research Database (Denmark)

    2017-01-01

    In this case, the author demonstrates together with the owner-manager of KALK A/S, Mr Rasmus Jorgensen, how to use the Family Business Map to frame a constructive discussion about long-term planning. The Family Business Map is a tool for long-term planning in family firms developed by Professor...

  20. Virtual Models of Long-Term Care

    Science.gov (United States)

    Phenice, Lillian A.; Griffore, Robert J.

    2012-01-01

    Nursing homes, assisted living facilities and home-care organizations, use web sites to describe their services to potential consumers. This virtual ethnographic study developed models representing how potential consumers may understand this information using data from web sites of 69 long-term-care providers. The content of long-term-care web…

  1. Sleep facilitates long-term face adaptation

    OpenAIRE

    Ditye, Thomas; Javadi, Amir Homayoun; Carbon, Claus-Christian; Walsh, Vincent

    2013-01-01

    Adaptation is an automatic neural mechanism supporting the optimization of visual processing on the basis of previous experiences. While the short-term effects of adaptation on behaviour and physiology have been studied extensively, perceptual long-term changes associated with adaptation are still poorly understood. Here, we show that the integration of adaptation-dependent long-term shifts in neural function is facilitated by sleep. Perceptual shifts induced by adaptation to a distorted imag...

  2. The Determination of the Asynchronous Traction Motor Characteristics of Locomotive

    Directory of Open Access Journals (Sweden)

    Pavel Grigorievich Kolpakhchyan

    2017-01-01

    Full Text Available The article deals with the problem of the locomotive asynchronous traction motor control with the AC diesel-electric transmission. The limitations of the torque of the traction motor when powered by the inverter are determined. The recommendations to improve the use of asynchronous traction motor of locomotives with the AC diesel-electric transmission are given.

  3. OFDM with Index Modulation for Asynchronous mMTC Networks.

    Science.gov (United States)

    Doğan, Seda; Tusha, Armed; Arslan, Hüseyin

    2018-04-21

    One of the critical missions for next-generation wireless communication systems is to fulfill the high demand for massive Machine-Type Communications (mMTC). In mMTC systems, a sporadic transmission is performed between machine users and base station (BS). Lack of coordination between the users and BS in time destroys orthogonality between the subcarriers, and causes inter-carrier interference (ICI). Therefore, providing services to asynchronous massive machine users is a major challenge for Orthogonal Frequency Division Multiplexing (OFDM). In this study, OFDM with index modulation (OFDM-IM) is proposed as an eligible solution to alleviate ICI caused by asynchronous transmission in uncoordinated mMTC networks. In OFDM-IM, data transmission is performed not only by modulated subcarriers but also by the indices of active subcarriers. Unlike classical OFDM, fractional subcarrier activation leads to less ICI in OFDM-IM technology. A novel subcarrier mapping scheme (SMS) named as Inner Subcarrier Activation is proposed to further alleviate adjacent user interference in asynchronous OFDM-IM-based systems. ISA reduces inter-user interference since it gives more activation priority to inner subcarriers compared with the existing SMS-s. The superiority of the proposed SMS is shown through both theoretical analysis and computer-based simulations in comparison to existing mapping schemes for asynchronous systems.

  4. Prevalence of Long-Term Opioid Use in Long-Stay Nursing Home Residents.

    Science.gov (United States)

    Hunnicutt, Jacob N; Chrysanthopoulou, Stavroula A; Ulbricht, Christine M; Hume, Anne L; Tjia, Jennifer; Lapane, Kate L

    2018-01-01

    Overall and long-term opioid use among older adults have increased since 1999. Less is known about opioid use in older adults in nursing homes (NHs). Cross-sectional. U.S. NHs (N = 13,522). Long-stay NH resident Medicare beneficiaries with a Minimum Data Set 3.0 (MDS) assessment between April 1, 2012, and June 30, 2012, and 120 days of follow-up (N = 315,949). We used Medicare Part D claims to measure length of opioid use in the 120 days from the index assessment (short-term: ≤30 days, medium-term: >30-89 days, long-term: ≥90 days), adjuvants (e.g., anticonvulsants), and other pain medications (e.g., corticosteroids). MDS assessments in the follow-up period were used to measure nonpharmacological pain management use. Modified Poisson models were used to estimate adjusted prevalence ratios (aPR) and 95% confidence intervals (CI) for age, gender, race and ethnicity, cognitive and physical impairment, and long-term opioid use. Of all long-stay residents, 32.4% were prescribed any opioid, and 15.5% were prescribed opioids long-term. Opioid users (versus nonusers) were more commonly prescribed pain adjuvants (32.9% vs 14.9%), other pain medications (25.5% vs 11.0%), and nonpharmacological pain management (24.5% vs 9.3%). Long-term opioid use was higher in women (aPR = 1.21, 95% CI = 1.18-1.23) and lower in racial and ethnic minorities (non-Hispanic blacks vs whites: APR = 0.93, 95% CI = 0.90-0.94) and those with severe cognitive impairment (vs no or mild impairment, aPR = 0.82, 95% CI = 0.79-0.83). One in seven NH residents was prescribed opioids long-term. Recent guidelines on opioid prescribing for pain recommend reducing long-term opioid use, but this is challenging in NHs because residents may not benefit from nonpharmacological and nonopioid interventions. Studies to address concerns about opioid safety and effectiveness (e.g., on pain and functional status) in NHs are needed. © 2017, Copyright the Authors Journal compilation © 2017, The American Geriatrics

  5. Asynchronous control for networked systems

    CERN Document Server

    Rubio, Francisco; Bencomo, Sebastián

    2015-01-01

    This book sheds light on networked control systems; it describes different techniques for asynchronous control, moving away from the periodic actions of classical control, replacing them with state-based decisions and reducing the frequency with which communication between subsystems is required. The text focuses specially on event-based control. Split into two parts, Asynchronous Control for Networked Systems begins by addressing the problems of single-loop networked control systems, laying out various solutions which include two alternative model-based control schemes (anticipatory and predictive) and the use of H2/H∞ robust control to deal with network delays and packet losses. Results on self-triggering and send-on-delta sampling are presented to reduce the need for feedback in the loop. In Part II, the authors present solutions for distributed estimation and control. They deal first with reliable networks and then extend their results to scenarios in which delays and packet losses may occur. The novel ...

  6. Word Processing in Dyslexics: An Automatic Decoding Deficit?

    Science.gov (United States)

    Yap, Regina; Van Der Leu, Aryan

    1993-01-01

    Compares dyslexic children with normal readers on measures of phonological decoding and automatic word processing. Finds that dyslexics have a deficit in automatic phonological decoding skills. Discusses results within the framework of the phonological deficit and the automatization deficit hypotheses. (RS)

  7. On the decoding process in ternary error-correcting output codes.

    Science.gov (United States)

    Escalera, Sergio; Pujol, Oriol; Radeva, Petia

    2010-01-01

    A common way to model multiclass classification problems is to design a set of binary classifiers and to combine them. Error-Correcting Output Codes (ECOC) represent a successful framework to deal with these type of problems. Recent works in the ECOC framework showed significant performance improvements by means of new problem-dependent designs based on the ternary ECOC framework. The ternary framework contains a larger set of binary problems because of the use of a "do not care" symbol that allows us to ignore some classes by a given classifier. However, there are no proper studies that analyze the effect of the new symbol at the decoding step. In this paper, we present a taxonomy that embeds all binary and ternary ECOC decoding strategies into four groups. We show that the zero symbol introduces two kinds of biases that require redefinition of the decoding design. A new type of decoding measure is proposed, and two novel decoding strategies are defined. We evaluate the state-of-the-art coding and decoding strategies over a set of UCI Machine Learning Repository data sets and into a real traffic sign categorization problem. The experimental results show that, following the new decoding strategies, the performance of the ECOC design is significantly improved.

  8. Modeling and Analysis of Asynchronous Systems Using SAL and Hybrid SAL

    Science.gov (United States)

    Tiwari, Ashish; Dutertre, Bruno

    2013-01-01

    We present formal models and results of formal analysis of two different asynchronous systems. We first examine a mid-value select module that merges the signals coming from three different sensors that are each asynchronously sampling the same input signal. We then consider the phase locking protocol proposed by Daly, Hopkins, and McKenna. This protocol is designed to keep a set of non-faulty (asynchronous) clocks phase locked even in the presence of Byzantine-faulty clocks on the network. All models and verifications have been developed using the SAL model checking tools and the Hybrid SAL abstractor.

  9. LDPC Codes--Structural Analysis and Decoding Techniques

    Science.gov (United States)

    Zhang, Xiaojie

    2012-01-01

    Low-density parity-check (LDPC) codes have been the focus of much research over the past decade thanks to their near Shannon limit performance and to their efficient message-passing (MP) decoding algorithms. However, the error floor phenomenon observed in MP decoding, which manifests itself as an abrupt change in the slope of the error-rate curve,…

  10. Supporting collaborative discussions on asynchronous time: a technological perspective

    OpenAIRE

    Caballé, Santi

    2011-01-01

    The aim of this paper is to report on an experience of using an innovative on-line learning tool to support real, collaborative learning through discussion in asynchronous time. While asynchronous interaction gives rise to unique opportunities that support active, collaborative learning, unique problems also arise, such as frustration, caused by waiting for other peoples' reactions and feedback and the consequent loss of motivation, which has a negative impact on learning outcomes. In order t...

  11. EPOS for Coordination of Asynchronous Sensor Webs

    Data.gov (United States)

    National Aeronautics and Space Administration — Develop, integrate, and deploy software-based tools to coordinate asynchronous, distributed missions and optimize observation planning spanning simultaneous...

  12. Long-Term Stewardship Baseline Report and Transition Guidance

    Energy Technology Data Exchange (ETDEWEB)

    Kristofferson, Keith

    2001-11-01

    Long-term stewardship consists of those actions necessary to maintain and demonstrate continued protection of human health and the environment after facility cleanup is complete. As the Department of Energy’s (DOE) lead laboratory for environmental management programs, the Idaho National Engineering and Environmental Laboratory (INEEL) administers DOE’s long-term stewardship science and technology efforts. The INEEL provides DOE with technical, and scientific expertise needed to oversee its long-term environmental management obligations complexwide. Long-term stewardship is administered and overseen by the Environmental Management Office of Science and Technology. The INEEL Long-Term Stewardship Program is currently developing the management structures and plans to complete INEEL-specific, long-term stewardship obligations. This guidance document (1) assists in ensuring that the program leads transition planning for the INEEL with respect to facility and site areas and (2) describes the classes and types of criteria and data required to initiate transition for areas and sites where the facility mission has ended and cleanup is complete. Additionally, this document summarizes current information on INEEL facilities, structures, and release sites likely to enter long-term stewardship at the completion of DOE’s cleanup mission. This document is not intended to function as a discrete checklist or local procedure to determine readiness to transition. It is an overarching document meant as guidance in implementing specific transition procedures. Several documents formed the foundation upon which this guidance was developed. Principal among these documents was the Long-Term Stewardship Draft Technical Baseline; A Report to Congress on Long-Term Stewardship, Volumes I and II; Infrastructure Long-Range Plan; Comprehensive Facility Land Use Plan; INEEL End-State Plan; and INEEL Institutional Plan.

  13. Construction and decoding of matrix-product codes from nested codes

    DEFF Research Database (Denmark)

    Hernando, Fernando; Lally, Kristine; Ruano, Diego

    2009-01-01

    We consider matrix-product codes [C1 ... Cs] · A, where C1, ..., Cs  are nested linear codes and matrix A has full rank. We compute their minimum distance and provide a decoding algorithm when A is a non-singular by columns matrix. The decoding algorithm decodes up to half of the minimum distance....

  14. Design of FBG En/decoders in Coherent 2-D Time-polarization OCDMA Systems

    Science.gov (United States)

    Hou, Fen-fei; Yang, Ming

    2012-12-01

    A novel fiber Bragg grating (FBG)-based en/decoder for the two-dimensional (2-D) time-spreading and polarization multiplexer optical coding is proposed. Compared with other 2-D en/decoders, the proposed en/decoding for an optical code-division multiple-access (OCDMA) system uses a single phase-encoded FBG and coherent en/decoding. Furthermore, combined with reconstruction-equivalent-chirp technology, such en/decoders can be realized with a conventional simple fabrication setup. Experimental results of such en/decoders and the corresponding system test at a data rate of 5 Gbit/s demonstrate that this kind of 2-D FBG-based en/decoders could improve the performances of OCDMA systems.

  15. Low Complexity List Decoding for Polar Codes with Multiple CRC Codes

    Directory of Open Access Journals (Sweden)

    Jong-Hwan Kim

    2017-04-01

    Full Text Available Polar codes are the first family of error correcting codes that provably achieve the capacity of symmetric binary-input discrete memoryless channels with low complexity. Since the development of polar codes, there have been many studies to improve their finite-length performance. As a result, polar codes are now adopted as a channel code for the control channel of 5G new radio of the 3rd generation partnership project. However, the decoder implementation is one of the big practical problems and low complexity decoding has been studied. This paper addresses a low complexity successive cancellation list decoding for polar codes utilizing multiple cyclic redundancy check (CRC codes. While some research uses multiple CRC codes to reduce memory and time complexity, we consider the operational complexity of decoding, and reduce it by optimizing CRC positions in combination with a modified decoding operation. Resultingly, the proposed scheme obtains not only complexity reduction from early stopping of decoding, but also additional reduction from the reduced number of decoding paths.

  16. Decoding bipedal locomotion from the rat sensorimotor cortex

    NARCIS (Netherlands)

    Rigosa, J.; Panarese, A.; Dominici, N.; Friedli, L.; van den Brand, R.; Carpaneto, J.; DiGiovanna, J.; Courtine, G.; Micera, S.

    2015-01-01

    Objective. Decoding forelimb movements from the firing activity of cortical neurons has been interfaced with robotic and prosthetic systems to replace lost upper limb functions in humans. Despite the potential of this approach to improve locomotion and facilitate gait rehabilitation, decoding lower

  17. An FPGA Implementation of (3,6-Regular Low-Density Parity-Check Code Decoder

    Directory of Open Access Journals (Sweden)

    Tong Zhang

    2003-05-01

    Full Text Available Because of their excellent error-correcting performance, low-density parity-check (LDPC codes have recently attracted a lot of attention. In this paper, we are interested in the practical LDPC code decoder hardware implementations. The direct fully parallel decoder implementation usually incurs too high hardware complexity for many real applications, thus partly parallel decoder design approaches that can achieve appropriate trade-offs between hardware complexity and decoding throughput are highly desirable. Applying a joint code and decoder design methodology, we develop a high-speed (3,k-regular LDPC code partly parallel decoder architecture based on which we implement a 9216-bit, rate-1/2(3,6-regular LDPC code decoder on Xilinx FPGA device. This partly parallel decoder supports a maximum symbol throughput of 54 Mbps and achieves BER 10−6 at 2 dB over AWGN channel while performing maximum 18 decoding iterations.

  18. Multithreaded Asynchronous Graph Traversal for In-Memory and Semi-External Memory

    KAUST Repository

    Pearce, Roger

    2010-11-01

    Processing large graphs is becoming increasingly important for many domains such as social networks, bioinformatics, etc. Unfortunately, many algorithms and implementations do not scale with increasing graph sizes. As a result, researchers have attempted to meet the growing data demands using parallel and external memory techniques. We present a novel asynchronous approach to compute Breadth-First-Search (BFS), Single-Source-Shortest-Paths, and Connected Components for large graphs in shared memory. Our highly parallel asynchronous approach hides data latency due to both poor locality and delays in the underlying graph data storage. We present an experimental study applying our technique to both In-Memory and Semi-External Memory graphs utilizing multi-core processors and solid-state memory devices. Our experiments using synthetic and real-world datasets show that our asynchronous approach is able to overcome data latencies and provide significant speedup over alternative approaches. For example, on billion vertex graphs our asynchronous BFS scales up to 14x on 16-cores. © 2010 IEEE.

  19. Long-Term Prognosis of Plantar Fasciitis

    DEFF Research Database (Denmark)

    Hansen, Liselotte; Krogh, Thøger Persson; Ellingsen, Torkell

    2018-01-01

    , exercise-induced symptoms, bilateral heel pain, fascia thickness, and presence of a heel spur) could predict long-term outcomes, (3) to assess the long-term ultrasound (US) development in the fascia, and (4) to assess whether US-guided corticosteroid injections induce atrophy of the heel fat pad. Study....... The risk was significantly greater for women (P heel...... regardless of symptoms and had no impact on prognosis, and neither did the presence of a heel spur. Only 24% of asymptomatic patients had a normal fascia on US at long-term follow-up. A US-guided corticosteroid injection did not cause atrophy of the heel fat pad. Our observational study did not allow us...

  20. Long-term dependence in exchange rates

    Directory of Open Access Journals (Sweden)

    A. Karytinos

    2000-01-01

    Full Text Available The extent to which exchange rates of four major currencies against the Greek Drachma exhibit long-term dependence is investigated using a R/S analysis testing framework. We show that both classic R/S analysis and the modified R/S statistic if enhanced by bootstrapping techniques can be proven very reliable tools to this end. Our findings support persistence and long-term dependence with non-periodic cycles for the Deutsche Mark and the French Franc series. In addition a noisy chaos explanation is favored over fractional Brownian motion. On the contrary, the US Dollar and British Pound were found to exhibit a much more random behavior and lack of any long-term structure.

  1. Performance-complexity tradeoff in sequential decoding for the unconstrained AWGN channel

    KAUST Repository

    Abediseid, Walid; Alouini, Mohamed-Slim

    2013-01-01

    channel has been studied only under the use of the minimum Euclidean distance decoder that is commonly referred to as the lattice decoder. Lattice decoders based on solutions to the NP-hard closest vector problem are very complex to implement

  2. Short-term versus long-term contracting for uranium enrichment services

    International Nuclear Information System (INIS)

    Rudy, G.P.

    1990-01-01

    The US Department of Energy (US DOE) is the world's largest and most experienced supplier of uranium enrichment services. Through the late 1970s and early 1980s, emerging market forces transformed what was once a monopoly into a highly competitive industry. In the early 1980's the DOE lost market share. But as we enter the 1990s, new market forces have emerged. The US DOE believes a responsible balance between long-term and short-term contracting will be the key to success and the key to assuring the long-term health and reliability of the nuclear fuel industry. The US DOE intends to be in this nuclear business for a long time and will continue to offer reliable and responsive services second to none

  3. Long-Term Dynamics of Autonomous Fractional Differential Equations

    Science.gov (United States)

    Liu, Tao; Xu, Wei; Xu, Yong; Han, Qun

    This paper aims to investigate long-term dynamic behaviors of autonomous fractional differential equations with effective numerical method. The long-term dynamic behaviors predict where systems are heading after long-term evolution. We make some modification and transplant cell mapping methods to autonomous fractional differential equations. The mapping time duration of cell mapping is enlarged to deal with the long memory effect. Three illustrative examples, i.e. fractional Lotka-Volterra equation, fractional van der Pol oscillator and fractional Duffing equation, are studied with our revised generalized cell mapping method. We obtain long-term dynamics, such as attractors, basins of attraction, and saddles. Compared with some existing stability and numerical results, the validity of our method is verified. Furthermore, we find that the fractional order has its effect on the long-term dynamics of autonomous fractional differential equations.

  4. Neural network decoder for quantum error correcting codes

    Science.gov (United States)

    Krastanov, Stefan; Jiang, Liang

    Artificial neural networks form a family of extremely powerful - albeit still poorly understood - tools used in anything from image and sound recognition through text generation to, in our case, decoding. We present a straightforward Recurrent Neural Network architecture capable of deducing the correcting procedure for a quantum error-correcting code from a set of repeated stabilizer measurements. We discuss the fault-tolerance of our scheme and the cost of training the neural network for a system of a realistic size. Such decoders are especially interesting when applied to codes, like the quantum LDPC codes, that lack known efficient decoding schemes.

  5. Distributed coding/decoding complexity in video sensor networks.

    Science.gov (United States)

    Cordeiro, Paulo J; Assunção, Pedro

    2012-01-01

    Video Sensor Networks (VSNs) are recent communication infrastructures used to capture and transmit dense visual information from an application context. In such large scale environments which include video coding, transmission and display/storage, there are several open problems to overcome in practical implementations. This paper addresses the most relevant challenges posed by VSNs, namely stringent bandwidth usage and processing time/power constraints. In particular, the paper proposes a novel VSN architecture where large sets of visual sensors with embedded processors are used for compression and transmission of coded streams to gateways, which in turn transrate the incoming streams and adapt them to the variable complexity requirements of both the sensor encoders and end-user decoder terminals. Such gateways provide real-time transcoding functionalities for bandwidth adaptation and coding/decoding complexity distribution by transferring the most complex video encoding/decoding tasks to the transcoding gateway at the expense of a limited increase in bit rate. Then, a method to reduce the decoding complexity, suitable for system-on-chip implementation, is proposed to operate at the transcoding gateway whenever decoders with constrained resources are targeted. The results show that the proposed method achieves good performance and its inclusion into the VSN infrastructure provides an additional level of complexity control functionality.

  6. Best linear decoding of random mask images

    International Nuclear Information System (INIS)

    Woods, J.W.; Ekstrom, M.P.; Palmieri, T.M.; Twogood, R.E.

    1975-01-01

    In 1968 Dicke proposed coded imaging of x and γ rays via random pinholes. Since then, many authors have agreed with him that this technique can offer significant image improvement. A best linear decoding of the coded image is presented, and its superiority over the conventional matched filter decoding is shown. Experimental results in the visible light region are presented. (U.S.)

  7. Observations on Polar Coding with CRC-Aided List Decoding

    Science.gov (United States)

    2016-09-01

    TECHNICAL REPORT 3041 September 2016 Observations on Polar Coding with CRC-Aided List Decoding David Wasserman Approved for public release. SSC...described in [2, 3]. In FY15 and FY16 we used cyclic redundancy check (CRC)-aided polar list decoding [4]. Section 2 describes the basics of polar coding ...and gives details of the encoders and decoders we used. In the course of our research, we performed simulations of polar codes in hundreds of cases

  8. The Womanly World of Long Term Care: The Plight of the Long Term Care Worker. Gray Paper.

    Science.gov (United States)

    Older Women's League, Washington, DC.

    Long-term care workers (those who are paid to provide custodial care for long-term patients in nursing homes or at home) must care for a growing number of increasingly disabled or dependent persons. They are working for agencies and institutions under growing pressure to increase productivity. They face new training and competency requirements,…

  9. A Block-Asynchronous Relaxation Method for Graphics Processing Units

    OpenAIRE

    Anzt, H.; Dongarra, J.; Heuveline, Vincent; Tomov, S.

    2011-01-01

    In this paper, we analyze the potential of asynchronous relaxation methods on Graphics Processing Units (GPUs). For this purpose, we developed a set of asynchronous iteration algorithms in CUDA and compared them with a parallel implementation of synchronous relaxation methods on CPU-based systems. For a set of test matrices taken from the University of Florida Matrix Collection we monitor the convergence behavior, the average iteration time and the total time-to-solution time. Analyzing the r...

  10. Novel Asynchronous Wrapper and Its Application to GALS Systems

    Institute of Scientific and Technical Information of China (English)

    Zhuang Shengxian; Peng Anjin; Lars Wanhammar

    2006-01-01

    An asynchronous wrapper with novel handshake circuits for data communication in globally asynchronous locally synchronous (GALS) systems is proposed. The handshake circuits include two communication ports and a local clock generator. Two approaches for the implementation of communication ports are presented, one with pure standard cells and the others with Müller-C elements. The detailed design methodology for GALS systems is given and the circuits are validated with VHDL and circuits simulation in standard CMOS technology.

  11. Optimization of MPEG decoding

    DEFF Research Database (Denmark)

    Martins, Bo; Forchhammer, Søren

    1999-01-01

    MPEG-2 video decoding is examined. A unified approach to quality improvement, chrominance upsampling, de-interlacing and superresolution is presented. The information over several frames is combined as part of the processing....

  12. Emotion Decoding and Incidental Processing Fluency as Antecedents of Attitude Certainty.

    Science.gov (United States)

    Petrocelli, John V; Whitmire, Melanie B

    2017-07-01

    Previous research demonstrates that attitude certainty influences the degree to which an attitude changes in response to persuasive appeals. In the current research, decoding emotions from facial expressions and incidental processing fluency, during attitude formation, are examined as antecedents of both attitude certainty and attitude change. In Experiment 1, participants who decoded anger or happiness during attitude formation expressed their greater attitude certainty, and showed more resistance to persuasion than participants who decoded sadness. By manipulating the emotion decoded, the diagnosticity of processing fluency experienced during emotion decoding, and the gaze direction of the social targets, Experiment 2 suggests that the link between emotion decoding and attitude certainty results from incidental processing fluency. Experiment 3 demonstrated that fluency in processing irrelevant stimuli influences attitude certainty, which in turn influences resistance to persuasion. Implications for appraisal-based accounts of attitude formation and attitude change are discussed.

  13. Long term wet spent nuclear fuel storage

    International Nuclear Information System (INIS)

    1987-04-01

    The meeting showed that there is continuing confidence in the use of wet storage for spent nuclear fuel and that long-term wet storage of fuel clad in zirconium alloys can be readily achieved. The importance of maintaining good water chemistry has been identified. The long-term wet storage behaviour of sensitized stainless steel clad fuel involves, as yet, some uncertainties. However, great reliance will be placed on long-term wet storage of spent fuel into the future. The following topics were treated to some extent: Oxidation of the external surface of fuel clad, rod consolidation, radiation protection, optimum methods of treating spent fuel storage water, physical radiation effects, and the behaviour of spent fuel assemblies of long-term wet storage conditions. A number of papers on national experience are included

  14. Long-Term Collections

    CERN Multimedia

    Staff Association

    2016-01-01

    45 years helping in developing countries! CERN personnel have been helping the least fortunate people on the planet since 1971. How? With the Long-Term Collections! Dear Colleagues, The Staff Association’s Long-Term Collections (LTC) Committee is delighted to share this important milestone in the life of our Laboratory with you. Indeed, whilst the name of CERN is known worldwide for scientific discoveries, it also shines in the many humanitarian projects which have been supported by the LTC since 1971. Several schools and clinics, far and wide, carry its logo... Over the past 45 years, 74 projects have been supported (9 of which are still ongoing). This all came from a group of colleagues who wanted to share a little of what life offered them here at CERN, in this haven of mutual understanding, peace and security, with those who were less fortunate elsewhere. Thus, the LTC were born... Since then, we have worked as a team to maintain the dream of these visionaries, with the help of regular donat...

  15. Long-Term Collection

    CERN Multimedia

    Staff Association

    2016-01-01

    Dear Colleagues, As previously announced in Echo (No. 254), your delegates took action to draw attention to the projects of the Long-Term Collections (LTC), the humanitarian body of the CERN Staff Association. On Tuesday, 11 October, at noon, small Z-Cards were widely distributed at the entrances of CERN restaurants and we thank you all for your interest. We hope to have achieved an important part of our goal, which was to inform you, convince you and find new supporters among you. We will find out in the next few days! An exhibition of the LTC was also set up in the Main Building for the entire week. The Staff Association wants to celebrate the occasion of the Long-Term Collection’s 45th anniversary at CERN because, ever since 1971, CERN personnel have showed great support in helping the least fortunate people on the planet in a variety of ways according to their needs. On a regular basis, joint fundraising appeals are made with the Directorate to help the victims of natural disasters around th...

  16. Blending Online Asynchronous and Synchronous Learning

    Directory of Open Access Journals (Sweden)

    Lisa C. Yamagata-Lynch

    2014-04-01

    Full Text Available In this article I will share a qualitative self-study about a 15-week blended 100% online graduate level course facilitated through synchronous meetings on Blackboard Collaborate and asynchronous discussions on Blackboard. I taught the course at the University of Tennessee (UT during the spring 2012 semester and the course topic was online learning environments. The primary research question of this study was: How can the designer/instructor optimize learning experiences for students who are studying about online learning environments in a blended online course relying on both synchronous and asynchronous technologies? I relied on student reflections of course activities during the beginning, middle, and the end of the semester as the primary data source to obtain their insights regarding course experiences. Through the experiences involved in designing and teaching the course and engaging in this study I found that there is room in the instructional technology research community to address strategies for facilitating online synchronous learning that complement asynchronous learning. Synchronous online whole class meetings and well-structured small group meetings can help students feel a stronger sense of connection to their peers and instructor and stay engaged with course activities. In order to provide meaningful learning spaces in synchronous learning environments, the instructor/designer needs to balance the tension between embracing the flexibility that the online space affords to users and designing deliberate structures that will help them take advantage of the flexible space.

  17. A Synchronous-Asynchronous Particle Swarm Optimisation Algorithm

    Science.gov (United States)

    Ab Aziz, Nor Azlina; Mubin, Marizan; Mohamad, Mohd Saberi; Ab Aziz, Kamarulzaman

    2014-01-01

    In the original particle swarm optimisation (PSO) algorithm, the particles' velocities and positions are updated after the whole swarm performance is evaluated. This algorithm is also known as synchronous PSO (S-PSO). The strength of this update method is in the exploitation of the information. Asynchronous update PSO (A-PSO) has been proposed as an alternative to S-PSO. A particle in A-PSO updates its velocity and position as soon as its own performance has been evaluated. Hence, particles are updated using partial information, leading to stronger exploration. In this paper, we attempt to improve PSO by merging both update methods to utilise the strengths of both methods. The proposed synchronous-asynchronous PSO (SA-PSO) algorithm divides the particles into smaller groups. The best member of a group and the swarm's best are chosen to lead the search. Members within a group are updated synchronously, while the groups themselves are asynchronously updated. Five well-known unimodal functions, four multimodal functions, and a real world optimisation problem are used to study the performance of SA-PSO, which is compared with the performances of S-PSO and A-PSO. The results are statistically analysed and show that the proposed SA-PSO has performed consistently well. PMID:25121109

  18. Changing incentives for long-term gas contracts

    International Nuclear Information System (INIS)

    Bohi, D.R.

    1992-01-01

    There is much concern about the absence of long-term gas contracts with fixed price and quantity conditions, which until recent years was the standard way of doing business in the gas industry. These types of contracts performed a valuable service in the development of the gas industry, and there comparative absence today is sometimes thought to be one reason for the current malaise in the industry. One hears the argument that there must be some kind of 'market failure' that prevents buyers and sellers from entering into these long term arrangements, and recent changes in state and federal regulations are often cited as the cause of the problem. The purpose of the author's remarks is to argue that what is taken as a breakdown in the market may be simply a reaction to a decline in economic incentives to enter into long-term contracts with rigid price and quantity terms. This is, in other words, simply one more aspect of change in the gas business that Frank Heintz referred to in his opening remarks this morning. The author starts by giving a brief description of the motives for engaging in long-term contracts, and then describes how incentives to use long-term contracts have declined for both gas buyers and gas sellers. He concludes that the decline in the use of long-term contracts is not cause for regulatory concern, but a result of the continuing transformation of the gas business to one that more closely resembles other commodity markets

  19. Energy in 2010 - 2020. Long term challenges; Energie 2010-2020. Les defis du long terme

    Energy Technology Data Exchange (ETDEWEB)

    Dessus, Benjamin [ed.] [Centre National de la Recherche Scientifique (CNRS), 75 - Paris (France)

    2000-02-02

    This report presents the results of a workshop intending to anticipate the long term challenges, to guide better the short term power options, to understand the available political, economical and technical assumptions for the prospective world situation, to give some strategic hints on the necessary transition. Indeed, the difficult issue which the workshop tried to tackle was how should we prepare to reveal the energetic challenge of the development of the eight to ten billion inhabitants of our Planet in the next century without jeopardizing its existence. The energetic problems, a hardcore of the international preoccupation of both growth and environment, as it was recently evidenced by the climatic conference in Kyoto, have ever been the object of a particular attention on the part of General Commissariat of Plan. Thus, the commission 'Energy in 2010 - 2020' has been instituted in April 1996 in order to update the works done in 1990 - 1991 by the commission 'Energy 2010'. Soon it occurred to this new commission the task of illuminating its works by a long term (2050 - 2100) world prospective analysis of the challenges and problems linked to energy, growth and environment. In conclusion, this document tried to find answers to questions like: - which are the risks the energy consumption augmentation entail? - can we control them by appropriate urbanism and transport policies or technological innovation?. Four options for immediate action are suggested: - the energy efficiency should become a priority objective of policies; -coping with the long term challenges requires acting at present; - building the transition between governmental leadership and market; - taking profit of all the possible synergies between short and long term planning.

  20. Opening remarks for a panel discussion on short-term vs long-term procurement

    International Nuclear Information System (INIS)

    Courtenay, R.H.

    1990-01-01

    Long-term contracting in the late 1970's and early 1980's is blamed for some of the inequities that plague the uranium industry today. Utilities are obliged to pay prices far above prevailing levels and relatively low cost producers are forced to shut down while watching less efficient suppliers stay in business thanks to their long term supply agreements. Furthermore, it is argued that long-term contracts have contributed to supply instability by forcing the buildup of surplus inventories and by supporting excess incremental production by suppliers who have a baseload of long-term contracts. The depressed prices resulting from this oversupply are in turn jeopardizing future resource development and damaging supply reliability. In summary, the author's argument is that supply reliability will be greatly enhanced by the assurance of adequate primary supply from traditional sources such as Canada. This will not happen without long-term contracts. This conclusion may not be expected coming from a representative of Canada's largest uranium producer. But the final comment is less self serving. Many of the critics of long term contracts apparently expect a continuing and plentiful supply of East Bloc uranium to the spot market. A further question is to what extent East Bloc suppliers will eventually require long-term contracts in order to maintain production facilitates in economies that are no longer centrally planned, and where there is open competition for capital. Ultimately, reliability of supply from the non-traditional suppliers may also depend on long-term contracts

  1. Sequential decoders for large MIMO systems

    KAUST Repository

    Ali, Konpal S.

    2014-05-01

    Due to their ability to provide high data rates, multiple-input multiple-output (MIMO) systems have become increasingly popular. Decoding of these systems with acceptable error performance is computationally very demanding. In this paper, we employ the Sequential Decoder using the Fano Algorithm for large MIMO systems. A parameter called the bias is varied to attain different performance-complexity trade-offs. Low values of the bias result in excellent performance but at the expense of high complexity and vice versa for higher bias values. Numerical results are done that show moderate bias values result in a decent performance-complexity trade-off. We also attempt to bound the error by bounding the bias, using the minimum distance of a lattice. The variations in complexity with SNR have an interesting trend that shows room for considerable improvement. Our work is compared against linear decoders (LDs) aided with Element-based Lattice Reduction (ELR) and Complex Lenstra-Lenstra-Lovasz (CLLL) reduction. © 2014 IFIP.

  2. Adaptive decoding of MPEG-4 sprites for memory-constrained embedded systems

    NARCIS (Netherlands)

    Pastrnak, M.; Farin, D.S.; With, de P.H.N.; Cardinal, J.; Cerf, N.; Delgrnage, O.

    2005-01-01

    Background sprite decoding is an essential part of object-based video coding.The composition and rendering of a final scene involves the placing of individual video objects in a predefined way superimposed on the decoded background image. The MPEG-4 standard includes the decoding algorithm for

  3. Low Power LDPC Code Decoder Architecture Based on Intermediate Message Compression Technique

    Science.gov (United States)

    Shimizu, Kazunori; Togawa, Nozomu; Ikenaga, Takeshi; Goto, Satoshi

    Reducing the power dissipation for LDPC code decoder is a major challenging task to apply it to the practical digital communication systems. In this paper, we propose a low power LDPC code decoder architecture based on an intermediate message-compression technique which features as follows: (i) An intermediate message compression technique enables the decoder to reduce the required memory capacity and write power dissipation. (ii) A clock gated shift register based intermediate message memory architecture enables the decoder to decompress the compressed messages in a single clock cycle while reducing the read power dissipation. The combination of the above two techniques enables the decoder to reduce the power dissipation while keeping the decoding throughput. The simulation results show that the proposed architecture improves the power efficiency up to 52% and 18% compared to that of the decoder based on the overlapped schedule and the rapid convergence schedule without the proposed techniques respectively.

  4. Construction and decoding of a class of algebraic geometry codes

    DEFF Research Database (Denmark)

    Justesen, Jørn; Larsen, Knud J.; Jensen, Helge Elbrønd

    1989-01-01

    A class of codes derived from algebraic plane curves is constructed. The concepts and results from algebraic geometry that were used are explained in detail; no further knowledge of algebraic geometry is needed. Parameters, generator and parity-check matrices are given. The main result is a decod...... is a decoding algorithm which turns out to be a generalization of the Peterson algorithm for decoding BCH decoder codes......A class of codes derived from algebraic plane curves is constructed. The concepts and results from algebraic geometry that were used are explained in detail; no further knowledge of algebraic geometry is needed. Parameters, generator and parity-check matrices are given. The main result...

  5. Multiformat decoder for a DSP-based IP set-top box

    Science.gov (United States)

    Pescador, F.; Garrido, M. J.; Sanz, C.; Juárez, E.; Samper, D.; Antoniello, R.

    2007-05-01

    Internet Protocol Set-Top Boxes (IP STBs) based on single-processor architectures have been recently introduced in the market. In this paper, the implementation of an MPEG-4 SP/ASP video decoder for a multi-format IP STB based on a TMS320DM641 DSP is presented. An initial decoder for PC platform was fully tested and ported to the DSP. Using this code an optimization process was started achieving a 90% speedup. This process allows real-time MPEG-4 SP/ASP decoding. The MPEG-4 decoder has been integrated in an IP STB and tested in a real environment using DVD movies and TV channels with excellent results.

  6. STACK DECODING OF LINEAR BLOCK CODES FOR DISCRETE MEMORYLESS CHANNEL USING TREE DIAGRAM

    Directory of Open Access Journals (Sweden)

    H. Prashantha Kumar

    2012-03-01

    Full Text Available The boundaries between block and convolutional codes have become diffused after recent advances in the understanding of the trellis structure of block codes and the tail-biting structure of some convolutional codes. Therefore, decoding algorithms traditionally proposed for decoding convolutional codes have been applied for decoding certain classes of block codes. This paper presents the decoding of block codes using tree structure. Many good block codes are presently known. Several of them have been used in applications ranging from deep space communication to error control in storage systems. But the primary difficulty with applying Viterbi or BCJR algorithms to decode of block codes is that, even though they are optimum decoding methods, the promised bit error rates are not achieved in practice at data rates close to capacity. This is because the decoding effort is fixed and grows with block length, and thus only short block length codes can be used. Therefore, an important practical question is whether a suboptimal realizable soft decision decoding method can be found for block codes. A noteworthy result which provides a partial answer to this question is described in the following sections. This result of near optimum decoding will be used as motivation for the investigation of different soft decision decoding methods for linear block codes which can lead to the development of efficient decoding algorithms. The code tree can be treated as an expanded version of the trellis, where every path is totally distinct from every other path. We have derived the tree structure for (8, 4 and (16, 11 extended Hamming codes and have succeeded in implementing the soft decision stack algorithm to decode them. For the discrete memoryless channel, gains in excess of 1.5dB at a bit error rate of 10-5 with respect to conventional hard decision decoding are demonstrated for these codes.

  7. Population Decoding of Motor Cortical Activity using a Generalized Linear Model with Hidden States

    Science.gov (United States)

    Lawhern, Vernon; Wu, Wei; Hatsopoulos, Nicholas G.; Paninski, Liam

    2010-01-01

    Generalized linear models (GLMs) have been developed for modeling and decoding population neuronal spiking activity in the motor cortex. These models provide reasonable characterizations between neural activity and motor behavior. However, they lack a description of movement-related terms which are not observed directly in these experiments, such as muscular activation, the subject's level of attention, and other internal or external states. Here we propose to include a multi-dimensional hidden state to address these states in a GLM framework where the spike count at each time is described as a function of the hand state (position, velocity, and acceleration), truncated spike history, and the hidden state. The model can be identified by an Expectation-Maximization algorithm. We tested this new method in two datasets where spikes were simultaneously recorded using a multi-electrode array in the primary motor cortex of two monkeys. It was found that this method significantly improves the model-fitting over the classical GLM, for hidden dimensions varying from 1 to 4. This method also provides more accurate decoding of hand state (lowering the Mean Square Error by up to 29% in some cases), while retaining real-time computational efficiency. These improvements on representation and decoding over the classical GLM model suggest that this new approach could contribute as a useful tool to motor cortical decoding and prosthetic applications. PMID:20359500

  8. The Design of Finite State Machine for Asynchronous Replication Protocol

    Science.gov (United States)

    Wang, Yanlong; Li, Zhanhuai; Lin, Wei; Hei, Minglei; Hao, Jianhua

    Data replication is a key way to design a disaster tolerance system and to achieve reliability and availability. It is difficult for a replication protocol to deal with the diverse and complex environment. This means that data is less well replicated than it ought to be. To reduce data loss and to optimize replication protocols, we (1) present a finite state machine, (2) run it to manage an asynchronous replication protocol and (3) report a simple evaluation of the asynchronous replication protocol based on our state machine. It's proved that our state machine is applicable to guarantee the asynchronous replication protocol running in the proper state to the largest extent in the event of various possible events. It also can helpful to build up replication-based disaster tolerance systems to ensure the business continuity.

  9. Visual perception as retrospective Bayesian decoding from high- to low-level features.

    Science.gov (United States)

    Ding, Stephanie; Cueva, Christopher J; Tsodyks, Misha; Qian, Ning

    2017-10-24

    When a stimulus is presented, its encoding is known to progress from low- to high-level features. How these features are decoded to produce perception is less clear, and most models assume that decoding follows the same low- to high-level hierarchy of encoding. There are also theories arguing for global precedence, reversed hierarchy, or bidirectional processing, but they are descriptive without quantitative comparison with human perception. Moreover, observers often inspect different parts of a scene sequentially to form overall perception, suggesting that perceptual decoding requires working memory, yet few models consider how working-memory properties may affect decoding hierarchy. We probed decoding hierarchy by comparing absolute judgments of single orientations and relative/ordinal judgments between two sequentially presented orientations. We found that lower-level, absolute judgments failed to account for higher-level, relative/ordinal judgments. However, when ordinal judgment was used to retrospectively decode memory representations of absolute orientations, striking aspects of absolute judgments, including the correlation and forward/backward aftereffects between two reported orientations in a trial, were explained. We propose that the brain prioritizes decoding of higher-level features because they are more behaviorally relevant, and more invariant and categorical, and thus easier to specify and maintain in noisy working memory, and that more reliable higher-level decoding constrains less reliable lower-level decoding. Published under the PNAS license.

  10. Long-term risks of kidney living donation

    DEFF Research Database (Denmark)

    Maggiore, Umberto; Budde, Klemens; Heemann, Uwe

    2017-01-01

    Two recent matched cohort studies from the USA and Norway published in 2014 have raised some concerns related to the long-term safety of kidney living donation. Further studies on the long-term risks of living donation have since been published. In this position paper, Developing Education Science...... and Care for Renal Transplantation in European States (DESCARTES) board members critically review the literature in an effort to summarize the current knowledge concerning long-term risks of kidney living donation to help physicians for decision-making purposes and for providing information...... to the prospective live donors. Long-term risk of end-stage renal disease (ESRD) can be partially foreseen by trying to identify donors at risk of developing ‘de novo’ kidney diseases during life post-donation and by predicting lifetime ESRD risk. However, lifetime risk may be difficult to assess in young donors...

  11. Long-term effects of radiation

    International Nuclear Information System (INIS)

    Smith, J.; Smith, T.

    1981-01-01

    It is pointed out that sources of long-term damage from radiation are two-fold. People who have been exposed to doses of radiation from initial early fallout but have recovered from the acute effects may still suffer long-term damage from their exposure. Those who have not been exposed to early fallout may be exposed to delayed fallout, the hazards from which are almost exclusively from ingesting strontium, caesium and carbon isotopes present in food; the damage caused is relatively unimportant compared with that caused by the brief doses from initial radiation and early fallout. A brief discussion is presented of the distribution of delayed long-lived isotope fallout, and an outline is sketched of late biological effects, such as malignant disease, cataracts, retarded development, infertility and genetic effects. (U.K.)

  12. An Improved Unscented Kalman Filter Based Decoder for Cortical Brain-Machine Interfaces.

    Science.gov (United States)

    Li, Simin; Li, Jie; Li, Zheng

    2016-01-01

    Brain-machine interfaces (BMIs) seek to connect brains with machines or computers directly, for application in areas such as prosthesis control. For this application, the accuracy of the decoding of movement intentions is crucial. We aim to improve accuracy by designing a better encoding model of primary motor cortical activity during hand movements and combining this with decoder engineering refinements, resulting in a new unscented Kalman filter based decoder, UKF2, which improves upon our previous unscented Kalman filter decoder, UKF1. The new encoding model includes novel acceleration magnitude, position-velocity interaction, and target-cursor-distance features (the decoder does not require target position as input, it is decoded). We add a novel probabilistic velocity threshold to better determine the user's intent to move. We combine these improvements with several other refinements suggested by others in the field. Data from two Rhesus monkeys indicate that the UKF2 generates offline reconstructions of hand movements (mean CC 0.851) significantly more accurately than the UKF1 (0.833) and the popular position-velocity Kalman filter (0.812). The encoding model of the UKF2 could predict the instantaneous firing rate of neurons (mean CC 0.210), given kinematic variables and past spiking, better than the encoding models of these two decoders (UKF1: 0.138, p-v Kalman: 0.098). In closed-loop experiments where each monkey controlled a computer cursor with each decoder in turn, the UKF2 facilitated faster task completion (mean 1.56 s vs. 2.05 s) and higher Fitts's Law bit rate (mean 0.738 bit/s vs. 0.584 bit/s) than the UKF1. These results suggest that the modeling and decoder engineering refinements of the UKF2 improve decoding performance. We believe they can be used to enhance other decoders as well.

  13. EXIT Chart Analysis of Binary Message-Passing Decoders

    DEFF Research Database (Denmark)

    Lechner, Gottfried; Pedersen, Troels; Kramer, Gerhard

    2007-01-01

    Binary message-passing decoders for LDPC codes are analyzed using EXIT charts. For the analysis, the variable node decoder performs all computations in the L-value domain. For the special case of a hard decision channel, this leads to the well know Gallager B algorithm, while the analysis can...... be extended to channels with larger output alphabets. By increasing the output alphabet from hard decisions to four symbols, a gain of more than 1.0 dB is achieved using optimized codes. For this code optimization, the mixing property of EXIT functions has to be modified to the case of binary message......-passing decoders....

  14. Locating and decoding barcodes in fuzzy images captured by smart phones

    Science.gov (United States)

    Deng, Wupeng; Hu, Jiwei; Liu, Quan; Lou, Ping

    2017-07-01

    With the development of barcodes for commercial use, people's requirements for detecting barcodes by smart phone become increasingly pressing. The low quality of barcode image captured by mobile phone always affects the decoding and recognition rates. This paper focuses on locating and decoding EAN-13 barcodes in fuzzy images. We present a more accurate locating algorithm based on segment length and high fault-tolerant rate algorithm for decoding barcodes. Unlike existing approaches, location algorithm is based on the edge segment length of EAN -13 barcodes, while our decoding algorithm allows the appearance of fuzzy region in barcode image. Experimental results are performed on damaged, contaminated and scratched digital images, and provide a quite promising result for EAN -13 barcode location and decoding.

  15. Decoding Hermitian Codes with Sudan's Algorithm

    DEFF Research Database (Denmark)

    Høholdt, Tom; Nielsen, Rasmus Refslund

    1999-01-01

    We present an efficient implementation of Sudan's algorithm for list decoding Hermitian codes beyond half the minimum distance. The main ingredients are an explicit method to calculate so-called increasing zero bases, an efficient interpolation algorithm for finding the Q-polynomial, and a reduct......We present an efficient implementation of Sudan's algorithm for list decoding Hermitian codes beyond half the minimum distance. The main ingredients are an explicit method to calculate so-called increasing zero bases, an efficient interpolation algorithm for finding the Q...

  16. Decoding algorithm for vortex communications receiver

    Science.gov (United States)

    Kupferman, Judy; Arnon, Shlomi

    2018-01-01

    Vortex light beams can provide a tremendous alphabet for encoding information. We derive a symbol decoding algorithm for a direct detection matrix detector vortex beam receiver using Laguerre Gauss (LG) modes, and develop a mathematical model of symbol error rate (SER) for this receiver. We compare SER as a function of signal to noise ratio (SNR) for our algorithm and for the Pearson correlation algorithm. To our knowledge, this is the first comprehensive treatment of a decoding algorithm of a matrix detector for an LG receiver.

  17. Fast and Flexible Successive-Cancellation List Decoders for Polar Codes

    Science.gov (United States)

    Hashemi, Seyyed Ali; Condo, Carlo; Gross, Warren J.

    2017-11-01

    Polar codes have gained significant amount of attention during the past few years and have been selected as a coding scheme for the next generation of mobile broadband standard. Among decoding schemes, successive-cancellation list (SCL) decoding provides a reasonable trade-off between the error-correction performance and hardware implementation complexity when used to decode polar codes, at the cost of limited throughput. The simplified SCL (SSCL) and its extension SSCL-SPC increase the speed of decoding by removing redundant calculations when encountering particular information and frozen bit patterns (rate one and single parity check codes), while keeping the error-correction performance unaltered. In this paper, we improve SSCL and SSCL-SPC by proving that the list size imposes a specific number of bit estimations required to decode rate one and single parity check codes. Thus, the number of estimations can be limited while guaranteeing exactly the same error-correction performance as if all bits of the code were estimated. We call the new decoding algorithms Fast-SSCL and Fast-SSCL-SPC. Moreover, we show that the number of bit estimations in a practical application can be tuned to achieve desirable speed, while keeping the error-correction performance almost unchanged. Hardware architectures implementing both algorithms are then described and implemented: it is shown that our design can achieve 1.86 Gb/s throughput, higher than the best state-of-the-art decoders.

  18. Hard decoding algorithm for optimizing thresholds under general Markovian noise

    Science.gov (United States)

    Chamberland, Christopher; Wallman, Joel; Beale, Stefanie; Laflamme, Raymond

    2017-04-01

    Quantum error correction is instrumental in protecting quantum systems from noise in quantum computing and communication settings. Pauli channels can be efficiently simulated and threshold values for Pauli error rates under a variety of error-correcting codes have been obtained. However, realistic quantum systems can undergo noise processes that differ significantly from Pauli noise. In this paper, we present an efficient hard decoding algorithm for optimizing thresholds and lowering failure rates of an error-correcting code under general completely positive and trace-preserving (i.e., Markovian) noise. We use our hard decoding algorithm to study the performance of several error-correcting codes under various non-Pauli noise models by computing threshold values and failure rates for these codes. We compare the performance of our hard decoding algorithm to decoders optimized for depolarizing noise and show improvements in thresholds and reductions in failure rates by several orders of magnitude. Our hard decoding algorithm can also be adapted to take advantage of a code's non-Pauli transversal gates to further suppress noise. For example, we show that using the transversal gates of the 5-qubit code allows arbitrary rotations around certain axes to be perfectly corrected. Furthermore, we show that Pauli twirling can increase or decrease the threshold depending upon the code properties. Lastly, we show that even if the physical noise model differs slightly from the hypothesized noise model used to determine an optimized decoder, failure rates can still be reduced by applying our hard decoding algorithm.

  19. Network evolution induced by asynchronous stimuli through spike-timing-dependent plasticity.

    Directory of Open Access Journals (Sweden)

    Wu-Jie Yuan

    Full Text Available In sensory neural system, external asynchronous stimuli play an important role in perceptual learning, associative memory and map development. However, the organization of structure and dynamics of neural networks induced by external asynchronous stimuli are not well understood. Spike-timing-dependent plasticity (STDP is a typical synaptic plasticity that has been extensively found in the sensory systems and that has received much theoretical attention. This synaptic plasticity is highly sensitive to correlations between pre- and postsynaptic firings. Thus, STDP is expected to play an important role in response to external asynchronous stimuli, which can induce segregative pre- and postsynaptic firings. In this paper, we study the impact of external asynchronous stimuli on the organization of structure and dynamics of neural networks through STDP. We construct a two-dimensional spatial neural network model with local connectivity and sparseness, and use external currents to stimulate alternately on different spatial layers. The adopted external currents imposed alternately on spatial layers can be here regarded as external asynchronous stimuli. Through extensive numerical simulations, we focus on the effects of stimulus number and inter-stimulus timing on synaptic connecting weights and the property of propagation dynamics in the resulting network structure. Interestingly, the resulting feedforward structure induced by stimulus-dependent asynchronous firings and its propagation dynamics reflect both the underlying property of STDP. The results imply a possible important role of STDP in generating feedforward structure and collective propagation activity required for experience-dependent map plasticity in developing in vivo sensory pathways and cortices. The relevance of the results to cue-triggered recall of learned temporal sequences, an important cognitive function, is briefly discussed as well. Furthermore, this finding suggests a potential

  20. Low Complexity Approach for High Throughput Belief-Propagation based Decoding of LDPC Codes

    Directory of Open Access Journals (Sweden)

    BOT, A.

    2013-11-01

    Full Text Available The paper proposes a low complexity belief propagation (BP based decoding algorithm for LDPC codes. In spite of the iterative nature of the decoding process, the proposed algorithm provides both reduced complexity and increased BER performances as compared with the classic min-sum (MS algorithm, generally used for hardware implementations. Linear approximations of check-nodes update function are used in order to reduce the complexity of the BP algorithm. Considering this decoding approach, an FPGA based hardware architecture is proposed for implementing the decoding algorithm, aiming to increase the decoder throughput. FPGA technology was chosen for the LDPC decoder implementation, due to its parallel computation and reconfiguration capabilities. The obtained results show improvements regarding decoding throughput and BER performances compared with state-of-the-art approaches.

  1. Long-term home care scheduling

    DEFF Research Database (Denmark)

    Gamst, Mette; Jensen, Thomas Sejr

    In several countries, home care is provided for certain citizens living at home. The long-term home care scheduling problem is to generate work plans spanning several days such that a high quality of service is maintained and the overall cost is kept as low as possible. A solution to the problem...... provides detailed information on visits and visit times for each employee on each of the covered days. We propose a branch-and-price algorithm for the long-term home care scheduling problem. The pricing problem generates one-day plans for an employee, and the master problem merges the plans with respect...

  2. MED5/355: Using Web-technology for Asynchronous Telemedicine Consulting

    OpenAIRE

    Reviakin, Y; Sukhanov, A

    1999-01-01

    Introduction Common telemedicine consultations can be divided in two classes: real-time telemedicine consultations and asynchronous telemedicine consultations. The advantage of real-time consultations is obvious - this is a natural discussion between physicians, which may be realised on the basis of desktop videoconferences. But the problems are also obvious: the necessity of additional hardware and the elevated demands for channel bandwidth. Because of the latter, the use of asynchronous tel...

  3. Analysing long term discursive processes

    DEFF Research Database (Denmark)

    Horsbøl, Anders

    which extend beyond the single interaction, for instance negotiations or planning processes, seems to have played a less important role, with studies such as Iedema 2001 and Wodak 2000 as exceptions. These long term processes, however, are central to the constitution and workings of organizations......What do timescales - the notion that processes take place or can be viewed within a shorter or longer temporal range (Lemke 2005) - mean for the analysis of discourse? What are the methodological consequences of analyzing discourse at different timescales? It may be argued that discourse analysis...... in general has favored either the analysis of short term processes such as interviews, discussions, and lessons, or the analysis of non-processual entities such as (multimodal) texts, arguments, discursive repertoires, and discourses (in a Foucaultian sense). In contrast, analysis of long term processes...

  4. Nuclear Energy, Long Term Requirements

    International Nuclear Information System (INIS)

    Knapp, V.

    2006-01-01

    There are serious warnings about depletion of oil and gas and even more serious warnings about dangers of climate change caused by emission of carbon dioxide. Should developed countries be called to replace CO2 emitting energy sources as soon as possible, and the time available may not be longer then few decades, can nuclear energy answer the call and what are the requirements? Assuming optimistic contribution of renewable energy sources, can nuclear energy expand to several times present level in order to replace large part of fossil fuels use? Paper considers intermediate and long-term requirements. Future of nuclear power depends on satisfactory answers on several questions. First group of questions are those important for near and intermediate future. They deal with economics and safety of nuclear power stations in the first place. On the same time scale a generally accepted concept for radioactive waste disposal is also required. All these issues are in the focus of present research and development. Safer and more economical reactors are targets of international efforts in Generation IV and INPRO projects, but aiming further ahead these innovative projects are also addressing issues such as waste reduction and proliferation resistance. However, even assuming successful technical development of these projects, and there is no reason to doubt it, long term and large-scale nuclear power use is thereby not yet secured. If nuclear power is to play an essential role in the long-term future energy production and in reduction of CO2 emission, than several additional questions must be replied. These questions will deal with long-term nuclear fuel sufficiency, with necessary contribution of nuclear power in sectors of transport and industrial processes and with nuclear proliferation safety. This last issue is more political then technical, thus sometimes neglected by nuclear engineers, yet it will have essential role for the long-term prospects of nuclear power. The

  5. What are the differences between long-term, short-term, and working memory?

    OpenAIRE

    Cowan, Nelson

    2008-01-01

    In the recent literature there has been considerable confusion about the three types of memory: long-term, short-term, and working memory. This chapter strives to reduce that confusion and makes up-to-date assessments of these types of memory. Long- and short-term memory could differ in two fundamental ways, with only short-term memory demonstrating (1) temporal decay and (2) chunk capacity limits. Both properties of short-term memory are still controversial but the current literature is rath...

  6. High-speed architecture for the decoding of trellis-coded modulation

    Science.gov (United States)

    Osborne, William P.

    1992-01-01

    Since 1971, when the Viterbi Algorithm was introduced as the optimal method of decoding convolutional codes, improvements in circuit technology, especially VLSI, have steadily increased its speed and practicality. Trellis-Coded Modulation (TCM) combines convolutional coding with higher level modulation (non-binary source alphabet) to provide forward error correction and spectral efficiency. For binary codes, the current stare-of-the-art is a 64-state Viterbi decoder on a single CMOS chip, operating at a data rate of 25 Mbps. Recently, there has been an interest in increasing the speed of the Viterbi Algorithm by improving the decoder architecture, or by reducing the algorithm itself. Designs employing new architectural techniques are now in existence, however these techniques are currently applied to simpler binary codes, not to TCM. The purpose of this report is to discuss TCM architectural considerations in general, and to present the design, at the logic gate level, or a specific TCM decoder which applies these considerations to achieve high-speed decoding.

  7. Long-term hearing preservation in vestibular schwannoma

    DEFF Research Database (Denmark)

    Stangerup, Sven-Eric; Thomsen, Jens; Tos, Mirko

    2010-01-01

    The aim of the present study was to evaluate the long-term hearing during "wait and scan" management of vestibular schwannomas.......The aim of the present study was to evaluate the long-term hearing during "wait and scan" management of vestibular schwannomas....

  8. Asynchronous stream processing with S-Net

    NARCIS (Netherlands)

    Grelck, C.; Scholz, S.-B.; Shafarenko, A.

    2010-01-01

    We present the rationale and design of S-Net, a coordination language for asynchronous stream processing. The language achieves a near-complete separation between the application code, written in any conventional programming language, and the coordination/communication code written in S-Net. Our

  9. Using Television Sitcoms to Facilitate Asynchronous Discussions in the Online Communication Course

    Science.gov (United States)

    Tolman, Elizabeth; Asbury, Bryan

    2012-01-01

    Asynchronous discussions are a useful instructional resource in the online communication course. In discussion groups students have the opportunity to actively participate and interact with students and the instructor. Asynchronous communication allows for flexibility because "participants can interact with significant amounts of time between…

  10. Long-Term Patency of Lymphovenous Anastomoses: A Systematic Review.

    Science.gov (United States)

    Tourani, Saam S; Taylor, G Ian; Ashton, Mark W

    2016-08-01

    With advancements in technology and microsurgical techniques, lymphovenous anastomosis has become a popular reconstructive procedure in the treatment of chronic lymphedema. However, the long-term patency of these anastomoses is not clear in the literature. A systematic review of the MEDLINE and EMBASE databases was performed to assess the reported long-term patency of lymphovenous anastomoses. A total of eight studies satisfied the inclusion criteria. Pooled data from four similar experiments in normal dogs showed an average long-term (≥5 months) patency of 52 percent. The only experiment in dogs with chronic lymphedema failed to show any long-term patency. The creation of peripheral lymphovenous anastomoses with a moderate long-term patency rate has become technically possible. However, the long-term results in chronic lymphedema are limited.

  11. Long-Term Orientation in Trade

    NARCIS (Netherlands)

    Hofstede, G.J.; Jonker, C.M.; Verwaart, D.

    2008-01-01

    Trust does not work in the same way across cultures. This paper presents an agent model of behavior in trade across Hofstedes cultural dimension of long-term vs. short-term orientation. The situation is based on a gaming simulation, the Trust and Tracing game. The paper investigates the

  12. [Cost-effectiveness of Synchronous vs. Asynchronous Telepsychiatry in Prison Inmates With Depression].

    Science.gov (United States)

    Barrera-Valencia, Camilo; Benito-Devia, Alexis Vladimir; Vélez-Álvarez, Consuelo; Figueroa-Barrera, Mario; Franco-Idárraga, Sandra Milena

    Telepsychiatry is defined as the use of information and communication technology (ICT) in providing remote psychiatric services. Telepsychiatry is applied using two types of communication: synchronous (real time) and asynchronous (store and forward). To determine the cost-effectiveness of a synchronous and an asynchronous telepsychiatric model in prison inmate patients with symptoms of depression. A cost-effectiveness study was performed on a population consisting of 157 patients from the Establecimiento Penitenciario y Carcelario de Mediana Seguridad de Manizales, Colombia. The sample was determined by applying Zung self-administered surveys for depression (1965) and the Hamilton Depression Rating Scale (HDRS), the latter being the tool used for the comparison. Initial Hamilton score, arrival time, duration of system downtime, and clinical effectiveness variables had normal distributions (P>.05). There were significant differences (P<.001) between care costs for the different models, showing that the mean cost of the asynchronous model is less than synchronous model, and making the asynchronous model more cost-effective. The asynchronous model is the most cost-effective model of telepsychiatry care for patients with depression admitted to a detention centre, according to the results of clinical effectiveness, cost measurement, and patient satisfaction. Copyright © 2016 Asociación Colombiana de Psiquiatría. Publicado por Elsevier España. All rights reserved.

  13. Leveraging Energy Harvesting and Wake-Up Receivers for Long-Term Wireless Sensor Networks.

    Science.gov (United States)

    Ait Aoudia, Fayçal; Gautier, Matthieu; Magno, Michele; Berder, Olivier; Benini, Luca

    2018-05-15

    Wireless sensor nodes are traditionally powered by individual batteries, and a significant effort has been devoted to maximizing the lifetime of these devices. However, as the batteries can only store a finite amount of energy, the network is still doomed to die, and changing the batteries is not always possible. A promising solution is to enable each node to harvest energy directly in its environment, using individual energy harvesters. Moreover, novel ultra-low power wake-up receivers, which allow continuous listening of the channel with negligible power consumption, are emerging. These devices enable asynchronous communication, further reducing the power consumption related to communication, which is typically one the most energy-consuming tasks in wireless sensor networks. Energy harvesting and wake-up receivers can be combined to significantly increase the energy efficiency of sensor networks. In this paper, we propose an energy manager for energy harvesting wireless sensor nodes and an asynchronous medium access control protocol, which exploits ultra-low power wake-up receivers. The two components are designed to work together and especially to fit the stringent constraints of wireless sensor nodes. The proposed approach has been implemented on a real hardware platform and tested in the field. Experimental results demonstrate the benefits of the proposed approach in terms of energy efficiency, power consumption and throughput, which can be up to more than two-times higher compared to traditional schemes.

  14. Long-term effects of childbirth in MS

    NARCIS (Netherlands)

    D'hooghe, M.B.; Nagels, G.; Uitdehaag, B.M.J.

    2010-01-01

    Background: The uncertainty about long-term effects of childbirth presents MS patients with dilemmas. Methods: Based on clinical data of 330 female MS patients, the long-term effects of childbirth were analysed, using a cross-sectional study design. Four groups of patients were distinguished: (1)

  15. Electrophysiological difference between mental state decoding and mental state reasoning.

    Science.gov (United States)

    Cao, Bihua; Li, Yiyuan; Li, Fuhong; Li, Hong

    2012-06-29

    Previous studies have explored the neural mechanism of Theory of Mind (ToM), but the neural correlates of its two components, mental state decoding and mental state reasoning, remain unclear. In the present study, participants were presented with various photographs, showing an actor looking at 1 of 2 objects, either with a happy or an unhappy expression. They were asked to either decode the emotion of the actor (mental state decoding task), predict which object would be chosen by the actor (mental state reasoning task), or judge at which object the actor was gazing (physical task), while scalp potentials were recorded. Results showed that (1) the reasoning task elicited an earlier N2 peak than the decoding task did over the prefrontal scalp sites; and (2) during the late positive component (240-440 ms), the reasoning task elicited a more positive deflection than the other two tasks did at the prefrontal scalp sites. In addition, neither the decoding task nor the reasoning task has no left/right hemisphere difference. These findings imply that mental state reasoning differs from mental state decoding early (210 ms) after stimulus onset, and that the prefrontal lobe is the neural basis of mental state reasoning. Copyright © 2012 Elsevier B.V. All rights reserved.

  16. Partially blind instantly decodable network codes for lossy feedback environment

    KAUST Repository

    Sorour, Sameh

    2014-09-01

    In this paper, we study the multicast completion and decoding delay minimization problems for instantly decodable network coding (IDNC) in the case of lossy feedback. When feedback loss events occur, the sender falls into uncertainties about packet reception at the different receivers, which forces it to perform partially blind selections of packet combinations in subsequent transmissions. To determine efficient selection policies that reduce the completion and decoding delays of IDNC in such an environment, we first extend the perfect feedback formulation in our previous works to the lossy feedback environment, by incorporating the uncertainties resulting from unheard feedback events in these formulations. For the completion delay problem, we use this formulation to identify the maximum likelihood state of the network in events of unheard feedback and employ it to design a partially blind graph update extension to the multicast IDNC algorithm in our earlier work. For the decoding delay problem, we derive an expression for the expected decoding delay increment for any arbitrary transmission. This expression is then used to find the optimal policy that reduces the decoding delay in such lossy feedback environment. Results show that our proposed solutions both outperform previously proposed approaches and achieve tolerable degradation even at relatively high feedback loss rates.

  17. Complexity Analysis of Reed-Solomon Decoding over GF without Using Syndromes

    Directory of Open Access Journals (Sweden)

    Chen Ning

    2008-01-01

    Full Text Available Abstract There has been renewed interest in decoding Reed-Solomon (RS codes without using syndromes recently. In this paper, we investigate the complexity of syndromeless decoding, and compare it to that of syndrome-based decoding. Aiming to provide guidelines to practical applications, our complexity analysis focuses on RS codes over characteristic-2 fields, for which some multiplicative FFT techniques are not applicable. Due to moderate block lengths of RS codes in practice, our analysis is complete, without big notation. In addition to fast implementation using additive FFT techniques, we also consider direct implementation, which is still relevant for RS codes with moderate lengths. For high-rate RS codes, when compared to syndrome-based decoding algorithms, not only syndromeless decoding algorithms require more field operations regardless of implementation, but also decoder architectures based on their direct implementations have higher hardware costs and lower throughput. We also derive tighter bounds on the complexities of fast polynomial multiplications based on Cantor's approach and the fast extended Euclidean algorithm.

  18. On the Convergence of Asynchronous Parallel Pattern Search

    International Nuclear Information System (INIS)

    Tamara Gilbson Kolda

    2002-01-01

    In this paper the authors prove global convergence for asynchronous parallel pattern search. In standard pattern search, decisions regarding the update of the iterate and the step-length control parameter are synchronized implicitly across all search directions. They lose this feature in asynchronous parallel pattern search since the search along each direction proceeds semi-autonomously. By bounding the value of the step-length control parameter after any step that produces decrease along a single search direction, they can prove that all the processes share a common accumulation point and that such a point is a stationary point of the standard nonlinear unconstrained optimization problem

  19. Hippocampal Focal Knockout of CBP Affects Specific Histone Modifications, Long-Term Potentiation, and Long-Term Memory

    Science.gov (United States)

    Barrett, Ruth M; Malvaez, Melissa; Kramar, Eniko; Matheos, Dina P; Arrizon, Abraham; Cabrera, Sara M; Lynch, Gary; Greene, Robert W; Wood, Marcelo A

    2011-01-01

    To identify the role of the histone acetyltransferase (HAT) CREB-binding protein (CBP) in neurons of the CA1 region of the hippocampus during memory formation, we examine the effects of a focal homozygous knockout of CBP on histone modifications, gene expression, synaptic plasticity, and long-term memory. We show that CBP is critical for the in vivo acetylation of lysines on histones H2B, H3, and H4. CBP's homolog p300 was unable to compensate for the loss of CBP. Neurons lacking CBP maintained phosphorylation of the transcription factor CREB, yet failed to activate CREB:CBP-mediated gene expression. Loss of CBP in dorsal CA1 of the hippocampus resulted in selective impairments to long-term potentiation and long-term memory for contextual fear and object recognition. Together, these results suggest a necessary role for specific chromatin modifications, selectively mediated by CBP in the consolidation of memories. PMID:21508930

  20. A quantum algorithm for Viterbi decoding of classical convolutional codes

    OpenAIRE

    Grice, Jon R.; Meyer, David A.

    2014-01-01

    We present a quantum Viterbi algorithm (QVA) with better than classical performance under certain conditions. In this paper the proposed algorithm is applied to decoding classical convolutional codes, for instance; large constraint length $Q$ and short decode frames $N$. Other applications of the classical Viterbi algorithm where $Q$ is large (e.g. speech processing) could experience significant speedup with the QVA. The QVA exploits the fact that the decoding trellis is similar to the butter...

  1. Design of a VLSI Decoder for Partially Structured LDPC Codes

    Directory of Open Access Journals (Sweden)

    Fabrizio Vacca

    2008-01-01

    of their parity matrix can be partitioned into two disjoint sets, namely, the structured and the random ones. For the proposed class of codes a constructive design method is provided. To assess the value of this method the constructed codes performance are presented. From these results, a novel decoding method called split decoding is introduced. Finally, to prove the effectiveness of the proposed approach a whole VLSI decoder is designed and characterized.

  2. [Psychosocial issues of long-term cancer survivors].

    Science.gov (United States)

    Weis, J; Faller, H

    2012-04-01

    Although cancer incidence rates are increasing, recent statistical studies suggest that cancer patients are showing higher cure rates as well as improved overall survival rates for most cancer locations. These advances are explained by improved strategies in early diagnoses as well as improved cancer therapies. Therefore, the number of long-term cancer survivors has also increased, but only few studies, especially within the last years, have focused on psychosocial issues of this subgroup. Some studies show that overall quality of life of long-term cancer survivors is quite high and comparable to that of the normal population. Nevertheless, a substantial percentage of former patients shows reduced quality of life and suffers from various sequelae of cancer and its treatment. This review focuses on the most common psychosocial issue of long-term survivors such as reduced psychological wellbeing, neuropsychological deficits and cancer-related fatigue syndrome. Finally, recommendations for problem-oriented interventions as well as improvement of psychosocial care of long-term survivors are given.

  3. Are long-term bisphosphonate users a reality?

    DEFF Research Database (Denmark)

    Abrahamsen, B

    2012-01-01

    The prevalence of long-term bisphosphonate use may be low due to low refill compliance and gaps in treatment. An analysis of the prescription history of 58,674 bisphosphonate users in Denmark found that only 2.8 % had received ten dose years of treatment or above. INTRODUCTION: This study aims...... to describe the demographics of present bisphosphonate (BP) users, to determine the prevalence of long-term BP use, and to establish if long-term use (a 10-year history of osteoporosis treatment) translated to ten dose years of bisphosphonate prescriptions filled, given the propensity for treatment gaps...... more than ten dose years of a BP. For any osteoporosis drug, 3.0 % had received ten dose years or more, while 23.2 % had received between 5 and 10 years of treatment. CONCLUSION: Long-term users with ten dose years or more of a BP are rare due to periods of low compliance and gaps, with a discrepancy...

  4. A security/safety survey of long term care facilities.

    Science.gov (United States)

    Acorn, Jonathan R

    2010-01-01

    What are the major security/safety problems of long term care facilities? What steps are being taken by some facilities to mitigate such problems? Answers to these questions can be found in a survey of IAHSS members involved in long term care security conducted for the IAHSS Long Term Care Security Task Force. The survey, the author points out, focuses primarily on long term care facilities operated by hospitals and health systems. However, he believes, it does accurately reflect the security problems most long term facilities face, and presents valuable information on security systems and practices which should be also considered by independent and chain operated facilities.

  5. Multiuser Random Coding Techniques for Mismatched Decoding

    OpenAIRE

    Scarlett, Jonathan; Martinez, Alfonso; Guillén i Fàbregas, Albert

    2016-01-01

    This paper studies multiuser random coding techniques for channel coding with a given (possibly suboptimal) decoding rule. For the mismatched discrete memoryless multiple-access channel, an error exponent is obtained that is tight with respect to the ensemble average, and positive within the interior of Lapidoth's achievable rate region. This exponent proves the ensemble tightness of the exponent of Liu and Hughes in the case of maximum-likelihood decoding. An equivalent dual form of Lapidoth...

  6. ZONES OF STEADY CAPACITOR EXCITATION IN A MODE OF GENERATION OF TYPICAL ASYNCHRONOUS MACHINES

    Directory of Open Access Journals (Sweden)

    Postoronca Sv.

    2009-12-01

    Full Text Available In work some features of a mode of capacitor excitation of industrial asynchronous electric motors, and also generators made on their base which can be used in wind installations of low power are considered. Borders of zones of steady capacitor excitation of asynchronous electric motors in rated power of 0,25-22,0 kW and generators made on their base, and also character of influence of own losses and active capacity of loading of the equivalent circuit of the asynchronous machine resulted in parameters have been determined. Some recommendations after maintenance of stability of capacitor excitation of asynchronous machines for work in a mode of generation of electric energy are given.

  7. The Structure and Content of Long-Term and Short-Term Mate Preferences

    Directory of Open Access Journals (Sweden)

    Peter K. Jonason

    2013-12-01

    Full Text Available This study addresses two limitations in the mate preferences literature. First, research all-too-often relies on single-item assessments of mate preferences precluding more advanced statistical techniques like factor analysis. Second, when factor analysis could be done, it exclusively has done for long-term mate preferences, at the exclusion of short-term mate preferences. In this study (N = 401, we subjected 20 items designed to measure short- and long-term mate preferences to both principle components (n = 200 and confirmatory factor analysis (n = 201. In the long-term context, we replicated previous findings that there are three different categories of preferences: physical attractiveness, interpersonal warmth, and social status. In the short-term context, physical attractiveness occupied two parts of the structure, social status dropped out, and interpersonal warmth remained. Across short- and long-term contexts, there were slight changes in what defined the shared dimensions (i.e., physical attractiveness and interpersonal warmth, suggesting prior work that applies the same inventory to each context might be flawed. We also replicated sex differences and similarities in mate preferences and correlates with sociosexuality and mate value. We adopt an evolutionary paradigm to understand our results.

  8. Long-Term Memory and Learning

    Science.gov (United States)

    Crossland, John

    2011-01-01

    The English National Curriculum Programmes of Study emphasise the importance of knowledge, understanding and skills, and teachers are well versed in structuring learning in those terms. Research outcomes into how long-term memory is stored and retrieved provide support for structuring learning in this way. Four further messages are added to the…

  9. A least squares approach for efficient and reliable short-term versus long-term optimization

    DEFF Research Database (Denmark)

    Christiansen, Lasse Hjuler; Capolei, Andrea; Jørgensen, John Bagterp

    2017-01-01

    The uncertainties related to long-term forecasts of oil prices impose significant financial risk on ventures of oil production. To minimize risk, oil companies are inclined to maximize profit over short-term horizons ranging from months to a few years. In contrast, conventional production...... optimization maximizes long-term profits over horizons that span more than a decade. To address this challenge, the oil literature has introduced short-term versus long-term optimization. Ideally, this problem is solved by a posteriori multi-objective optimization methods that generate an approximation...... the balance between the objectives, leaving an unfulfilled potential to increase profits. To promote efficient and reliable short-term versus long-term optimization, this paper introduces a natural way to characterize desirable Pareto points and proposes a novel least squares (LS) method. Unlike hierarchical...

  10. Sacrococcygeal teratoma: Clinical characteristics and long-term ...

    African Journals Online (AJOL)

    Background/Purpose : The excision of sacrococcygeal teratoma (SCT) may be associated with significant long-term morbidity for the child. We reviewed our experience with SCT in a tertiary health care facility in a developing country with particular interest on the long-term sequelae. Methods : Between January 1990 and ...

  11. Gigabit Ethernet signal transmission using asynchronous optical code division multiple access.

    Science.gov (United States)

    Ma, Philip Y; Fok, Mable P; Shastri, Bhavin J; Wu, Ben; Prucnal, Paul R

    2015-12-15

    We propose and experimentally demonstrate a novel architecture for interfacing and transmitting a Gigabit Ethernet (GbE) signal using asynchronous incoherent optical code division multiple access (OCDMA). This is the first such asynchronous incoherent OCDMA system carrying GbE data being demonstrated to be working among multi-users where each user is operating with an independent clock/data rate and is granted random access to the network. Three major components, the GbE interface, the OCDMA transmitter, and the OCDMA receiver are discussed in detail. The performance of the system is studied and characterized through measuring eye diagrams, bit-error rate and packet loss rate in real-time file transfer. Our Letter also addresses the near-far problem and realizes asynchronous transmission and detection of signal.

  12. Kernel Temporal Differences for Neural Decoding

    Science.gov (United States)

    Bae, Jihye; Sanchez Giraldo, Luis G.; Pohlmeyer, Eric A.; Francis, Joseph T.; Sanchez, Justin C.; Príncipe, José C.

    2015-01-01

    We study the feasibility and capability of the kernel temporal difference (KTD)(λ) algorithm for neural decoding. KTD(λ) is an online, kernel-based learning algorithm, which has been introduced to estimate value functions in reinforcement learning. This algorithm combines kernel-based representations with the temporal difference approach to learning. One of our key observations is that by using strictly positive definite kernels, algorithm's convergence can be guaranteed for policy evaluation. The algorithm's nonlinear functional approximation capabilities are shown in both simulations of policy evaluation and neural decoding problems (policy improvement). KTD can handle high-dimensional neural states containing spatial-temporal information at a reasonable computational complexity allowing real-time applications. When the algorithm seeks a proper mapping between a monkey's neural states and desired positions of a computer cursor or a robot arm, in both open-loop and closed-loop experiments, it can effectively learn the neural state to action mapping. Finally, a visualization of the coadaptation process between the decoder and the subject shows the algorithm's capabilities in reinforcement learning brain machine interfaces. PMID:25866504

  13. Managerial Long-Term Responsibility in Family-Controlled Firms

    Directory of Open Access Journals (Sweden)

    Dietmar Sternad

    2013-01-01

    Full Text Available Evidence suggests that long-term orientation (LTO as a dominantstrategic logic contributes to the sustainable performance offamily-controlled firms (FCFS. Combining a review of the literatureon lto with stewardship theory and upper echelons theoryreasoning, this article presents a typology of managerial responsibilityand introduces the concept of long-term responsibility as amanagerial characteristic constituting a major driving force behindcreating lto. The antecedents of long-term responsibilityunder family firm-specific conditions (stemming from the familysystem, the governance system, and family-firm managers’ personalcharacteristics are also identified and presented in an integratedmodel. The paper contributes to a more comprehensiveunderstanding of intertemporal choice in fcfs and explains whythey tend to be more long-term oriented than other types of firms.

  14. Row Reduction Applied to Decoding of Rank Metric and Subspace Codes

    DEFF Research Database (Denmark)

    Puchinger, Sven; Nielsen, Johan Sebastian Rosenkilde; Li, Wenhui

    2017-01-01

    We show that decoding of ℓ-Interleaved Gabidulin codes, as well as list-ℓ decoding of Mahdavifar–Vardy (MV) codes can be performed by row reducing skew polynomial matrices. Inspired by row reduction of F[x] matrices, we develop a general and flexible approach of transforming matrices over skew...... polynomial rings into a certain reduced form. We apply this to solve generalised shift register problems over skew polynomial rings which occur in decoding ℓ-Interleaved Gabidulin codes. We obtain an algorithm with complexity O(ℓμ2) where μ measures the size of the input problem and is proportional...... to the code length n in the case of decoding. Further, we show how to perform the interpolation step of list-ℓ-decoding MV codes in complexity O(ℓn2), where n is the number of interpolation constraints....

  15. Collectes à long terme

    CERN Multimedia

    Collectes à long terme

    2014-01-01

    En cette fin d’année 2014 qui approche à grands pas, le Comité des Collectes à Long Terme remercie chaleureusement ses fidèles donatrices et donateurs réguliers pour leurs contributions à nos actions en faveur des plus démunis de notre planète. C’est très important, pour notre Comité, de pouvoir compter sur l’appui assidu que vous nous apportez. Depuis plus de 40 ans maintenant, le modèle des CLT est basé principalement sur des actions à long terme (soit une aide pendant 4-5 ans par projet, mais plus parfois selon les circonstances), et sa planification demande une grande régularité de ses soutiens financiers. Grand MERCI à vous ! D’autres dons nous parviennent au cours de l’année, et ils sont aussi les bienvenus. En particulier, nous tenons à remercier...

  16. Computing by Temporal Order: Asynchronous Cellular Automata

    Directory of Open Access Journals (Sweden)

    Michael Vielhaber

    2012-08-01

    Full Text Available Our concern is the behaviour of the elementary cellular automata with state set 0,1 over the cell set Z/nZ (one-dimensional finite wrap-around case, under all possible update rules (asynchronicity. Over the torus Z/nZ (n<= 11,we will see that the ECA with Wolfram rule 57 maps any v in F_2^n to any w in F_2^n, varying the update rule. We furthermore show that all even (element of the alternating group bijective functions on the set F_2^n = 0,...,2^n-1, can be computed by ECA57, by iterating it a sufficient number of times with varying update rules, at least for n <= 10. We characterize the non-bijective functions computable by asynchronous rules.

  17. Computational Aspects of Asynchronous CA

    OpenAIRE

    Chandesris, Jérôme; Dennunzio, Alberto; Formenti, Enrico; Manzoni, Luca

    2011-01-01

    This work studies some aspects of the computational power of fully asynchronous cellular automata (ACA). We deal with some notions of simulation between ACA and Turing Machines. In particular, we characterize the updating sequences specifying which are "universal", i.e., allowing a (specific family of) ACA to simulate any TM on any input. We also consider the computational cost of such simulations.

  18. CLIMATE CHANGE: LONG-TERM TRENDS AND SHORT-TERM OSCILLATIONS

    Institute of Scientific and Technical Information of China (English)

    GAO Xin-quan; ZHANG Xin; QIAN Wei-hong

    2006-01-01

    Identifying the Northern Hemisphere (NH) temperature reconstruction and instrumental data for the past 1000 years shows that climate change in the last millennium includes long-term trends and various oscillations. Two long-term trends and the quasi-70-year oscillation were detected in the global temperature series for the last 140 years and the NH millennium series. One important feature was emphasized that temperature decreases slowly but it increases rapidly based on the analysis of different series. Benefits can be obtained of climate change from understanding various long-term trends and oscillations. Millennial temperature proxies from the natural climate system and time series of nonlinear model system are used in understanding the natural climate change and recognizing potential benefits by using the method of wavelet transform analysis. The results from numerical modeling show that major oscillations contained in numerical solutions on the interdecadal timescale are consistent with that of natural proxies. It seems that these oscillations in the climate change are not directly linked with the solar radiation as an external forcing. This investigation may conclude that the climate variability at the interdecadal timescale strongly depends on the internal nonlinear effects in the climate system.

  19. Impact of long-term and short-term therapies on seminal parameters

    Directory of Open Access Journals (Sweden)

    Jlenia Elia

    2013-04-01

    Full Text Available Aim: The aim of this work was: i to evaluate the prevalence of male partners of subfertile couples being treated with long/short term therapies for non andrological diseases; ii to study their seminal profile for the possible effects of their treatments on spermatogenesis and/or epididymal maturation. Methods: The study group was made up of 723 subjects, aged between 25 and 47 years. Semen analysis was performed according to World Health Organization (WHO guidelines (1999. The Superimposed Image Analysis System (SIAS, which is based on the computerized superimposition of spermatozoa images, was used to assess sperm motility parameters. Results: The prevalence of subjects taking pharmacological treatments was 22.7% (164/723. The prevalence was 3.7% (27/723 for the Short-Term Group and 18.9% (137/723 for the Long-Term Group. The subjects of each group were also subdivided into subgroups according to the treatments being received. Regarding the seminal profile, we did not observe a significant difference between the Long-Term, Short-Term or the Control Group. However, regarding the subgroups, we found a significant decrease in sperm number and progressive motility percentage in the subjects receiving treatment with antihypertensive drugs compared with the other subgroups and the Control Group. Conclusions: In the management of infertile couples, the potential negative impact on seminal parameters of any drugs being taken as Long-Term Therapy should be considered. The pathogenic mechanism needs to be clarified.

  20. Coding/decoding two-dimensional images with orbital angular momentum of light.

    Science.gov (United States)

    Chu, Jiaqi; Li, Xuefeng; Smithwick, Quinn; Chu, Daping

    2016-04-01

    We investigate encoding and decoding of two-dimensional information using the orbital angular momentum (OAM) of light. Spiral phase plates and phase-only spatial light modulators are used in encoding and decoding of OAM states, respectively. We show that off-axis points and spatial variables encoded with a given OAM state can be recovered through decoding with the corresponding complimentary OAM state.

  1. A cost of long-term memory in Drosophila

    OpenAIRE

    Mery, Frederic; Kawecki, Tadeusz J.

    2005-01-01

    Two distinct forms of consolidated associative memory are known in Drosophila: long-term memory and so-called anesthesia-resistant memory. Long-term memory is more stable, but unlike anesthesia-resistant memory, its formation requires protein synthesis. We show that flies induced to form long-term memory become more susceptible to extreme stress (such as desiccation). In contrast, induction of anesthesia-resistant memory had no detectable effect on desiccation resistance. This finding may hel...

  2. The design of an asynchronous Tiny RISC TM/TR4101 microprocessor core

    DEFF Research Database (Denmark)

    Christensen, Kåre Tais; Jensen, P.; Korger, P.

    1998-01-01

    This paper presents the design of an asynchronous version of the TR4101 embedded microprocessor core developed by LSI Logic Inc. The asynchronous processor, called ARISC, was designed using the same CAD tools and the same standard cell library that was used to implement the TR4101. The paper repo...

  3. A study of the long-term effect of malar fat repositioning in face lift surgery: short-term success but long-term failure.

    Science.gov (United States)

    Hamra, Sam T

    2002-09-01

    In 1990, the author reported on a series of 403 cases of deep plane face lifts, the first published technique describing the repositioning of the cheek fat, known as malar fat, in face lift surgery. This study examines the long-term results of 20 of the original series in an attempt to determine what areas of the rejuvenated face (specifically, the malar fat) showed long-term improvement. The results were judged by comparing the preoperative and long-term postoperative views in a half-and-half same-side hemiface photograph. The anatomy of the jawline (superficial musculoaponeurotic system [SMAS]), the nasolabial fold (malar fat), and the periorbital diameter were evaluated. The results confirmed that repositioning of the SMAS remained for longer than improvement in the nasolabial fold and that the vertical diameter of the periorbit did not change at all. The early results of malar fat repositioning shown at 1 to 2 years were successful, but the long-term results showed failure of the early improvement, manifested by recurrence of the nasolabial folds. There was, however, continuation of the improved results of the forehead lift and SMAS maneuvers of the original procedure. The conclusion is that only a direct excision will produce a permanent correction of the aging nasolabial fold.

  4. ESVD: An Integrated Energy Scalable Framework for Low-Power Video Decoding Systems

    Directory of Open Access Journals (Sweden)

    Wen Ji

    2010-01-01

    Full Text Available Video applications using mobile wireless devices are a challenging task due to the limited capacity of batteries. The higher complex functionality of video decoding needs high resource requirements. Thus, power efficient control has become more critical design with devices integrating complex video processing techniques. Previous works on power efficient control in video decoding systems often aim at the low complexity design and not explicitly consider the scalable impact of subfunctions in decoding process, and seldom consider the relationship with the features of compressed video date. This paper is dedicated to developing an energy-scalable video decoding (ESVD strategy for energy-limited mobile terminals. First, ESVE can dynamically adapt the variable energy resources due to the device aware technique. Second, ESVD combines the decoder control with decoded data, through classifying the data into different partition profiles according to its characteristics. Third, it introduces utility theoretical analysis during the resource allocation process, so as to maximize the resource utilization. Finally, it adapts the energy resource as different energy budget and generates the scalable video decoding output under energy-limited systems. Experimental results demonstrate the efficiency of the proposed approach.

  5. Population decoding of motor cortical activity using a generalized linear model with hidden states.

    Science.gov (United States)

    Lawhern, Vernon; Wu, Wei; Hatsopoulos, Nicholas; Paninski, Liam

    2010-06-15

    Generalized linear models (GLMs) have been developed for modeling and decoding population neuronal spiking activity in the motor cortex. These models provide reasonable characterizations between neural activity and motor behavior. However, they lack a description of movement-related terms which are not observed directly in these experiments, such as muscular activation, the subject's level of attention, and other internal or external states. Here we propose to include a multi-dimensional hidden state to address these states in a GLM framework where the spike count at each time is described as a function of the hand state (position, velocity, and acceleration), truncated spike history, and the hidden state. The model can be identified by an Expectation-Maximization algorithm. We tested this new method in two datasets where spikes were simultaneously recorded using a multi-electrode array in the primary motor cortex of two monkeys. It was found that this method significantly improves the model-fitting over the classical GLM, for hidden dimensions varying from 1 to 4. This method also provides more accurate decoding of hand state (reducing the mean square error by up to 29% in some cases), while retaining real-time computational efficiency. These improvements on representation and decoding over the classical GLM model suggest that this new approach could contribute as a useful tool to motor cortical decoding and prosthetic applications. Copyright (c) 2010 Elsevier B.V. All rights reserved.

  6. Developmental Dyslexia and Explicit Long-Term Memory

    Science.gov (United States)

    Menghini, Deny; Carlesimo, Giovanni Augusto; Marotta, Luigi; Finzi, Alessandra; Vicari, Stefano

    2010-01-01

    The reduced verbal long-term memory capacities often reported in dyslexics are generally interpreted as a consequence of their deficit in phonological coding. The present study was aimed at evaluating whether the learning deficit exhibited by dyslexics was restricted only to the verbal component of the long-term memory abilities or also involved…

  7. Delay reduction in lossy intermittent feedback for generalized instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.

    2013-10-01

    In this paper, we study the effect of lossy intermittent feedback loss events on the multicast decoding delay performance of generalized instantly decodable network coding. These feedback loss events create uncertainty at the sender about the reception statues of different receivers and thus uncertainty to accurately determine subsequent instantly decodable coded packets. To solve this problem, we first identify the different possibilities of uncertain packets at the sender and their probabilities. We then derive the expression of the mean decoding delay. We formulate the Generalized Instantly Decodable Network Coding (G-IDNC) minimum decoding delay problem as a maximum weight clique problem. Since finding the optimal solution is NP-hard, we design a variant of the algorithm employed in [1]. Our algorithm is compared to the two blind graph update proposed in [2] through extensive simulations. Results show that our algorithm outperforms the blind approaches in all the situations and achieves a tolerable degradation, against the perfect feedback, for large feedback loss period. © 2013 IEEE.

  8. Delay reduction in lossy intermittent feedback for generalized instantly decodable network coding

    KAUST Repository

    Douik, Ahmed S.; Sorour, Sameh; Alouini, Mohamed-Slim; Ai-Naffouri, Tareq Y.

    2013-01-01

    In this paper, we study the effect of lossy intermittent feedback loss events on the multicast decoding delay performance of generalized instantly decodable network coding. These feedback loss events create uncertainty at the sender about the reception statues of different receivers and thus uncertainty to accurately determine subsequent instantly decodable coded packets. To solve this problem, we first identify the different possibilities of uncertain packets at the sender and their probabilities. We then derive the expression of the mean decoding delay. We formulate the Generalized Instantly Decodable Network Coding (G-IDNC) minimum decoding delay problem as a maximum weight clique problem. Since finding the optimal solution is NP-hard, we design a variant of the algorithm employed in [1]. Our algorithm is compared to the two blind graph update proposed in [2] through extensive simulations. Results show that our algorithm outperforms the blind approaches in all the situations and achieves a tolerable degradation, against the perfect feedback, for large feedback loss period. © 2013 IEEE.

  9. Proximity-Based Asynchronous Messaging Platform for Location-Based Internet of Things Service

    Directory of Open Access Journals (Sweden)

    Hyeong gon Jo

    2016-07-01

    Full Text Available The Internet of Things (IoT opens up tremendous opportunities to provide location-based applications. However, despite the services around a user being physically adjacent, common IoT platforms use a centralized structure, like a cloud-computing architecture, which transfers large amounts of data to a central server. This raises problems, such as traffic concentration, long service latency, and high communication cost. In this paper, we propose a physical distance-based asynchronous messaging platform that specializes in processing personalized data and location-based messages. The proposed system disperses traffic using a location-based message-delivery protocol, and has high stability.

  10. Utilizing Cross-Layer Information to Improve Performance in JPEG2000 Decoding

    Directory of Open Access Journals (Sweden)

    Hannes Persson

    2007-01-01

    Full Text Available We focus on wireless multimedia communication and investigate how cross-layer information can be used to improve performance at the application layer, using JPEG2000 as an example. The cross-layer information is in the form of soft information from the physical layer. The soft information, which is supplied by a soft decision demodulator, yields reliability measures for the received bits and is fed into two soft input iterative JPEG2000 image decoders. When errors are detected with the error detecting mechanisms in JPEG2000, the decoders utilize the soft information to point out likely transmission errors. Hence, the decoders can correct errors and increase the image quality without making time-consuming retransmissions. We believe that the proposed decoding method utilizing soft information is suitable for a general IP-based network and that it keeps the principles of a layered structure of the protocol stack intact. Further, experimental results with images transmitted over a simulated wireless channel show that a simple decoding algorithm that utilizes soft information can give high gains in image quality compared to the standard hard-decision decoding.

  11. Advancing an ethical framework for long-term care.

    Science.gov (United States)

    Carter, Mary Whelan

    2002-02-01

    This article represents an effort to formulate an ethical framework for long-term care with the explicit purpose of providing a catalyst to promote further discourse and expand consideration of what an ethic of long-term care might entail. Grounding the discussion, an introduction to traditional ethical philosophy is presented, focusing mainly on the fundamentals of deontological and teleological ethical theories. Attention then shifts to a review of the more frequently cited principles found in the long-term care ethics literature, followed by a critique of the current reliance upon principlism to resolve ethical dilemmas in long-term care. In response to this criticism, an agent-driven ethical framework stressing dignity and respect for personhood, drawn from the philosophy of Immanuel Kant, is advanced.

  12. Long-term memory, sleep, and the spacing effect.

    Science.gov (United States)

    Bell, Matthew C; Kawadri, Nader; Simone, Patricia M; Wiseheart, Melody

    2014-01-01

    Many studies have shown that memory is enhanced when study sessions are spaced apart rather than massed. This spacing effect has been shown to have a lasting benefit to long-term memory when the study phase session follows the encoding session by 24 hours. Using a spacing paradigm we examined the impact of sleep and spacing gaps on long-term declarative memory for Swahili-English word pairs by including four spacing delay gaps (massed, 12 hours same-day, 12 hours overnight, and 24 hours). Results showed that a 12-hour spacing gap that includes sleep promotes long-term memory retention similar to the 24-hour gap. The findings support the importance of sleep to the long-term benefit of the spacing effect.

  13. Efficient Dual Domain Decoding of Linear Block Codes Using Genetic Algorithms

    Directory of Open Access Journals (Sweden)

    Ahmed Azouaoui

    2012-01-01

    Full Text Available A computationally efficient algorithm for decoding block codes is developed using a genetic algorithm (GA. The proposed algorithm uses the dual code in contrast to the existing genetic decoders in the literature that use the code itself. Hence, this new approach reduces the complexity of decoding the codes of high rates. We simulated our algorithm in various transmission channels. The performance of this algorithm is investigated and compared with competitor decoding algorithms including Maini and Shakeel ones. The results show that the proposed algorithm gives large gains over the Chase-2 decoding algorithm and reach the performance of the OSD-3 for some quadratic residue (QR codes. Further, we define a new crossover operator that exploits the domain specific information and compare it with uniform and two point crossover. The complexity of this algorithm is also discussed and compared to other algorithms.

  14. Long-term course of opioid addiction.

    Science.gov (United States)

    Hser, Yih-Ing; Evans, Elizabeth; Grella, Christine; Ling, Walter; Anglin, Douglas

    2015-01-01

    Opioid addiction is associated with excess mortality, morbidities, and other adverse conditions. Guided by a life-course framework, we review the literature on the long-term course of opioid addiction in terms of use trajectories, transitions, and turning points, as well as other factors that facilitate recovery from addiction. Most long-term follow-up studies are based on heroin addicts recruited from treatment settings (mostly methadone maintenance treatment), many of whom are referred by the criminal justice system. Cumulative evidence indicates that opioid addiction is a chronic disorder with frequent relapses. Longer treatment retention is associated with a greater likelihood of abstinence, whereas incarceration is negatively related to subsequent abstinence. Over the long term, the mortality rate of opioid addicts (overdose being the most common cause) is about 6 to 20 times greater than that of the general population; among those who remain alive, the prevalence of stable abstinence from opioid use is low (less than 30% after 10-30 years of observation), and many continue to use alcohol and other drugs after ceasing to use opioids. Histories of sexual or physical abuse and comorbid mental disorders are associated with the persistence of opioid use, whereas family and social support, as well as employment, facilitates recovery. Maintaining opioid abstinence for at least five years substantially increases the likelihood of future stable abstinence. Recent advances in pharmacological treatment options (buprenorphine and naltrexone) include depot formulations offering longer duration of medication; their impact on the long-term course of opioid addiction remains to be assessed.

  15. Synchronization of Hierarchical Time-Varying Neural Networks Based on Asynchronous and Intermittent Sampled-Data Control.

    Science.gov (United States)

    Xiong, Wenjun; Patel, Ragini; Cao, Jinde; Zheng, Wei Xing

    In this brief, our purpose is to apply asynchronous and intermittent sampled-data control methods to achieve the synchronization of hierarchical time-varying neural networks. The asynchronous and intermittent sampled-data controllers are proposed for two reasons: 1) the controllers may not transmit the control information simultaneously and 2) the controllers cannot always exist at any time . The synchronization is then discussed for a kind of hierarchical time-varying neural networks based on the asynchronous and intermittent sampled-data controllers. Finally, the simulation results are given to illustrate the usefulness of the developed criteria.In this brief, our purpose is to apply asynchronous and intermittent sampled-data control methods to achieve the synchronization of hierarchical time-varying neural networks. The asynchronous and intermittent sampled-data controllers are proposed for two reasons: 1) the controllers may not transmit the control information simultaneously and 2) the controllers cannot always exist at any time . The synchronization is then discussed for a kind of hierarchical time-varying neural networks based on the asynchronous and intermittent sampled-data controllers. Finally, the simulation results are given to illustrate the usefulness of the developed criteria.

  16. Essays on long-term mortality and interest rate risk

    NARCIS (Netherlands)

    de Kort, J.P.

    2017-01-01

    This dissertation comprises a study of long-term risks which play a major role in actuarial science. In Part I we analyse long-term mortality risk and its impact on consumption and investment decisions of economic agents, while Part II focuses on the mathematical modelling of long-term interest

  17. ERDA's long-term waste management goals and programs

    International Nuclear Information System (INIS)

    Perge, A.F.; Trice, V.G. Jr.; Walton, R.D. Jr.

    1976-01-01

    This paper presents an overview of the ERDA's major program for the long-term waste management of radioactive waste and provides a perspective for symposium participants with regard to the interrelationship of specific components of the program that are discussed in detail in other ERDA-sponsored papers. Needs, goals, and plans are reviewed for ERDA's management of the commercially generated wastes which are expected to be delivered to ERDA in accordance with Federal regulations. At present, ERDA responsibilities include long-term management of commercial-level wastes. Possible future regulations may give ERDA responsibility for the long-term management of commercial low-level solid wastes contaminated with transuranic nuclides. Primary planning goals and programs for the development of terminal storage facilities and waste processing technology to produce acceptable waste forms for long-term management are reviewed for each of the waste types identified above. The status of development programs for the long-term management of airborne radionuclides, which may be required at some time in the future, is also reviewed. (author)

  18. Optimal coding-decoding for systems controlled via a communication channel

    Science.gov (United States)

    Yi-wei, Feng; Guo, Ge

    2013-12-01

    In this article, we study the problem of controlling plants over a signal-to-noise ratio (SNR) constrained communication channel. Different from previous research, this article emphasises the importance of the actual channel model and coder/decoder in the study of network performance. Our major objectives include coder/decoder design for an additive white Gaussian noise (AWGN) channel with both standard network configuration and Youla parameter network architecture. We find that the optimal coder and decoder can be realised for different network configuration. The results are useful in determining the minimum channel capacity needed in order to stabilise plants over communication channels. The coder/decoder obtained can be used to analyse the effect of uncertainty on the channel capacity. An illustrative example is provided to show the effectiveness of the results.

  19. A BMI-based occupational therapy assist suit: asynchronous control by SSVEP.

    Science.gov (United States)

    Sakurada, Takeshi; Kawase, Toshihiro; Takano, Kouji; Komatsu, Tomoaki; Kansaku, Kenji

    2013-01-01

    A brain-machine interface (BMI) is an interface technology that uses neurophysiological signals from the brain to control external machines. Recent invasive BMI technologies have succeeded in the asynchronous control of robot arms for a useful series of actions, such as reaching and grasping. In this study, we developed non-invasive BMI technologies aiming to make such useful movements using the subject's own hands by preparing a BMI-based occupational therapy assist suit (BOTAS). We prepared a pre-recorded series of useful actions-a grasping-a-ball movement and a carrying-the-ball movement-and added asynchronous control using steady-state visual evoked potential (SSVEP) signals. A SSVEP signal was used to trigger the grasping-a-ball movement and another SSVEP signal was used to trigger the carrying-the-ball movement. A support vector machine was used to classify EEG signals recorded from the visual cortex (Oz) in real time. Untrained, able-bodied participants (n = 12) operated the system successfully. Classification accuracy and time required for SSVEP detection were ~88% and 3 s, respectively. We further recruited three patients with upper cervical spinal cord injuries (SCIs); they also succeeded in operating the system without training. These data suggest that our BOTAS system is potentially useful in terms of rehabilitation of patients with upper limb disabilities.

  20. Long-Term Clock Behavior of GPS IIR Satellites

    National Research Council Canada - National Science Library

    Epstein, Marvin; Dass, Todd; Rajan, John; Gilmour, Paul

    2007-01-01

    .... Rubidium clocks, as opposed to cesium clocks, have significant long-term drift. The current literature describes an initial model of drift aging for rubidium atomic clocks followed by a long-term characteristic...

  1. Towards behavioral synthesis of asynchronous circuits - an implementation template targeting syntax directed compilation

    DEFF Research Database (Denmark)

    Nielsen, Sune Fallgaard; Sparsø, Jens; Madsen, Jan

    2004-01-01

    This paper presents a method for behavioral synthesis of asynchronous circuits. Our approach aims at providing a synthesis flow which is very similar to what is found in existing synchronous design tools. We adapt the synchronous behavioral synthesis abstraction into the asynchronous handshake...

  2. Long-term properties of bituminized waste products

    International Nuclear Information System (INIS)

    Snellman, M.; Valkiainen, M.

    1985-10-01

    This study is a survey of the factors of importance for the long term behaviour of bituminized ion exchange resins. Phenomena occurring in bituminized waste products affected by the treatment, storage and disposal are identified. Test methods have been developed for characterizing product properties, which are important for the long-term behaviour. The long-term properties of bituminized ion-exchange resins are studied in a repository environment with access of water equilibrated with concrete. In these circumstances the most important properties are related to the interactions of bituminized waste with the surrounding barriers. The most important phenomena are water uptake due to rehydration of the resins and subsequent swelling of the product

  3. Long-term properties of bituminized waste products

    International Nuclear Information System (INIS)

    Snellman, M.; Valkiainen, M.

    1985-10-01

    This study is a survey of the factors of importance for long term behaviour of bituminized ion exchange resins. Phenomena occurring in bituminized waste products affected by the treatment, storage and disposal are identified. Test methods have been developed for characterizing product properties, which are important for the long-term behaviour. The long-term properties of bituminized ion-exchange resins are studied in a repository environment with access of water equilibrated with concrete. In these circumstances the most important properties are ralated to the interactions of bituminized waste with the surrounding barriers. The most important phenomena are water uptake due to rehydration of the resins and subsequent swelling of the product. (author)

  4. Space ventures and society long-term perspectives

    Science.gov (United States)

    Brown, W. M.

    1985-01-01

    A futuristic evaluation of mankind's potential long term future in space is presented. Progress in space will not be inhibited by shortages of the Earth's physical resources, since long term economic growth will be focused on ways to constrain industrial productivity by changing social values, management styles, or government competence. Future technological progress is likely to accelerate with an emphasis on international cooperation, making possible such large joint projects as lunar colonies or space stations on Mars. The long term future in space looks exceedingly bright even in relatively pessimistic scenarios. The principal driving forces will be technological progress, commercial and public-oriented satellites, space industrialization, space travel, and eventually space colonization.

  5. Achievable Information Rates for Coded Modulation With Hard Decision Decoding for Coherent Fiber-Optic Systems

    Science.gov (United States)

    Sheikh, Alireza; Amat, Alexandre Graell i.; Liva, Gianluigi

    2017-12-01

    We analyze the achievable information rates (AIRs) for coded modulation schemes with QAM constellations with both bit-wise and symbol-wise decoders, corresponding to the case where a binary code is used in combination with a higher-order modulation using the bit-interleaved coded modulation (BICM) paradigm and to the case where a nonbinary code over a field matched to the constellation size is used, respectively. In particular, we consider hard decision decoding, which is the preferable option for fiber-optic communication systems where decoding complexity is a concern. Recently, Liga \\emph{et al.} analyzed the AIRs for bit-wise and symbol-wise decoders considering what the authors called \\emph{hard decision decoder} which, however, exploits \\emph{soft information} of the transition probabilities of discrete-input discrete-output channel resulting from the hard detection. As such, the complexity of the decoder is essentially the same as the complexity of a soft decision decoder. In this paper, we analyze instead the AIRs for the standard hard decision decoder, commonly used in practice, where the decoding is based on the Hamming distance metric. We show that if standard hard decision decoding is used, bit-wise decoders yield significantly higher AIRs than symbol-wise decoders. As a result, contrary to the conclusion by Liga \\emph{et al.}, binary decoders together with the BICM paradigm are preferable for spectrally-efficient fiber-optic systems. We also design binary and nonbinary staircase codes and show that, in agreement with the AIRs, binary codes yield better performance.

  6. Short-term and long-term sick-leave in Sweden

    DEFF Research Database (Denmark)

    Blank, N; Diderichsen, Finn

    1995-01-01

    The primary aim of the study was to analyse similarities and differences between repeated spells of short-term sick-leave (more than 3 spells of less than 7 days' duration in a 12-month period) and long-term absence through sickness (at least 1 spell of more than 59 days' duration in a 12-month p...

  7. Embedded Vehicle Speed Estimation System Using an Asynchronous Temporal Contrast Vision Sensor

    Directory of Open Access Journals (Sweden)

    D. Bauer

    2007-01-01

    Full Text Available This article presents an embedded multilane traffic data acquisition system based on an asynchronous temporal contrast vision sensor, and algorithms for vehicle speed estimation developed to make efficient use of the asynchronous high-precision timing information delivered by this sensor. The vision sensor features high temporal resolution with a latency of less than 100 μs, wide dynamic range of 120 dB of illumination, and zero-redundancy, asynchronous data output. For data collection, processing and interfacing, a low-cost digital signal processor is used. The speed of the detected vehicles is calculated from the vision sensor's asynchronous temporal contrast event data. We present three different algorithms for velocity estimation and evaluate their accuracy by means of calibrated reference measurements. The error of the speed estimation of all algorithms is near zero mean and has a standard deviation better than 3% for both traffic flow directions. The results and the accuracy limitations as well as the combined use of the algorithms in the system are discussed.

  8. Pediatric emergency medicine asynchronous e-learning: a multicenter randomized controlled Solomon four-group study.

    Science.gov (United States)

    Chang, Todd P; Pham, Phung K; Sobolewski, Brad; Doughty, Cara B; Jamal, Nazreen; Kwan, Karen Y; Little, Kim; Brenkert, Timothy E; Mathison, David J

    2014-08-01

    Asynchronous e-learning allows for targeted teaching, particularly advantageous when bedside and didactic education is insufficient. An asynchronous e-learning curriculum has not been studied across multiple centers in the context of a clinical rotation. We hypothesize that an asynchronous e-learning curriculum during the pediatric emergency medicine (EM) rotation improves medical knowledge among residents and students across multiple participating centers. Trainees on pediatric EM rotations at four large pediatric centers from 2012 to 2013 were randomized in a Solomon four-group design. The experimental arms received an asynchronous e-learning curriculum consisting of nine Web-based, interactive, peer-reviewed Flash/HTML5 modules. Postrotation testing and in-training examination (ITE) scores quantified improvements in knowledge. A 2 × 2 analysis of covariance (ANCOVA) tested interaction and main effects, and Pearson's correlation tested associations between module usage, scores, and ITE scores. A total of 256 of 458 participants completed all study elements; 104 had access to asynchronous e-learning modules, and 152 were controls who used the current education standards. No pretest sensitization was found (p = 0.75). Use of asynchronous e-learning modules was associated with an improvement in posttest scores (p effect (partial η(2) = 0.19). Posttest scores correlated with ITE scores (r(2) = 0.14, p e-learning is an effective educational tool to improve knowledge in a clinical rotation. Web-based asynchronous e-learning is a promising modality to standardize education among multiple institutions with common curricula, particularly in clinical rotations where scheduling difficulties, seasonality, and variable experiences limit in-hospital learning. © 2014 by the Society for Academic Emergency Medicine.

  9. Increasing Student Engagement Using Asynchronous Learning

    Science.gov (United States)

    Northey, Gavin; Bucic, Tania; Chylinski, Mathew; Govind, Rahul

    2015-01-01

    Student engagement is an ongoing concern for educators because of its positive association with deep learning and educational outcomes. This article tests the use of a social networking site (Facebook) as a tool to facilitate asynchronous learning opportunities that complement face-to-face interactions and thereby enable a stronger learning…

  10. Asynchronous Gossip for Averaging and Spectral Ranking

    Science.gov (United States)

    Borkar, Vivek S.; Makhijani, Rahul; Sundaresan, Rajesh

    2014-08-01

    We consider two variants of the classical gossip algorithm. The first variant is a version of asynchronous stochastic approximation. We highlight a fundamental difficulty associated with the classical asynchronous gossip scheme, viz., that it may not converge to a desired average, and suggest an alternative scheme based on reinforcement learning that has guaranteed convergence to the desired average. We then discuss a potential application to a wireless network setting with simultaneous link activation constraints. The second variant is a gossip algorithm for distributed computation of the Perron-Frobenius eigenvector of a nonnegative matrix. While the first variant draws upon a reinforcement learning algorithm for an average cost controlled Markov decision problem, the second variant draws upon a reinforcement learning algorithm for risk-sensitive control. We then discuss potential applications of the second variant to ranking schemes, reputation networks, and principal component analysis.

  11. Decoding error-correcting codes with Gröbner bases

    NARCIS (Netherlands)

    Bulygin, S.; Pellikaan, G.R.; Veldhuis, R.; Cronie, H.; Hoeksema, H.

    2007-01-01

    The decoding of arbitrary linear block codes is accomplished by solving a system of quadratic equations by means of Buchberger’s algorithm for finding a Gröbner basis. This generalizes the algorithm of Berlekamp-Massey for decoding Reed Solomon, Goppa and cyclic codes up to half the true minimum

  12. Experimental Researches on Long-Term Strength of Granite Gneiss

    Directory of Open Access Journals (Sweden)

    Lin Liu

    2015-01-01

    Full Text Available It is important to confirm the long-term strength of rock materials for the purpose of evaluating the long-term stability of rock engineering. In this study, a series of triaxial creep tests were conducted on granite gneiss under different pore pressures. Based on the test data, we proposed two new quantitative methods, tangent method and intersection method, to confirm the long-term strength of rock. Meanwhile, the isochronous stress-strain curve method was adopted to make sure of the accuracy and operability of the two new methods. It is concluded that the new methods are suitable for the study of the long-term strength of rock. The effect of pore pressure on the long-term strength of rock in triaxial creep tests is also discussed.

  13. Violation of the equivalence principle for stressed bodies in asynchronous relativity

    Energy Technology Data Exchange (ETDEWEB)

    Andrade Martins, R. de (Centro de Logica, Epistemologia e Historia da Ciencia, Campinas (Brazil))

    1983-12-11

    In the recently developed asynchronous formulation of the relativistic theory of extended bodies, the inertial mass of a body does not explicitly depend on its pressure or stress. The detailed analysis of the weight of a box filled with a gas and placed in a weak gravitational field shows that this feature of asynchronous relativity implies a breakdown of the equivalence between inertial and passive gravitational mass for stressed systems.

  14. Decoding Individual Finger Movements from One Hand Using Human EEG Signals

    Science.gov (United States)

    Gonzalez, Jania; Ding, Lei

    2014-01-01

    Brain computer interface (BCI) is an assistive technology, which decodes neurophysiological signals generated by the human brain and translates them into control signals to control external devices, e.g., wheelchairs. One problem challenging noninvasive BCI technologies is the limited control dimensions from decoding movements of, mainly, large body parts, e.g., upper and lower limbs. It has been reported that complicated dexterous functions, i.e., finger movements, can be decoded in electrocorticography (ECoG) signals, while it remains unclear whether noninvasive electroencephalography (EEG) signals also have sufficient information to decode the same type of movements. Phenomena of broadband power increase and low-frequency-band power decrease were observed in EEG in the present study, when EEG power spectra were decomposed by a principal component analysis (PCA). These movement-related spectral structures and their changes caused by finger movements in EEG are consistent with observations in previous ECoG study, as well as the results from ECoG data in the present study. The average decoding accuracy of 77.11% over all subjects was obtained in classifying each pair of fingers from one hand using movement-related spectral changes as features to be decoded using a support vector machine (SVM) classifier. The average decoding accuracy in three epilepsy patients using ECoG data was 91.28% with the similarly obtained features and same classifier. Both decoding accuracies of EEG and ECoG are significantly higher than the empirical guessing level (51.26%) in all subjects (pEEG as in ECoG, and demonstrates the feasibility of discriminating finger movements from one hand using EEG. These findings are promising to facilitate the development of BCIs with rich control signals using noninvasive technologies. PMID:24416360

  15. Long-term skeletal findings in Menkes disease

    International Nuclear Information System (INIS)

    Amador, Eva; Domene, Ruth; Fuentes, Cristian; Carreno, Juan-Carlos; Enriquez, Goya

    2010-01-01

    Skeletal findings in infants with Menkes disease, the most characteristic of which are metaphyseal spurs, long-bone fractures and wormian bones, have been widely reported. However, the changes in skeletal features over time are not well known. The long-term findings differ completely from those initially observed and consist of undertubulation and metaphyseal flaring, similar to the findings seen in some types of bone dysplasia. The initial and long-term radiological features in an 8-year-old boy with Menkes disease are illustrated. (orig.)

  16. The interaction of short-term and long-term memory in phonetic category formation

    Science.gov (United States)

    Harnsberger, James D.

    2002-05-01

    This study examined the role that short-term memory capacity plays in the relationship between novel stimuli (e.g., non-native speech sounds, native nonsense words) and phonetic categories in long-term memory. Thirty native speakers of American English were administered five tests: categorial AXB discrimination using nasal consonants from Malayalam; categorial identification, also using Malayalam nasals, which measured the influence of phonetic categories in long-term memory; digit span; nonword span, a short-term memory measure mediated by phonetic categories in long-term memory; and paired-associate word learning (word-word and word-nonword pairs). The results showed that almost all measures were significantly correlated with one another. The strongest predictor for the discrimination and word-nonword learning results was nonword (r=+0.62) and digit span (r=+0.51), respectively. When the identification test results were partialed out, only nonword span significantly correlated with discrimination. The results show a strong influence of short-term memory capacity on the encoding of phonetic detail within phonetic categories and suggest that long-term memory representations regulate the capacity of short-term memory to preserve information for subsequent encoding. The results of this study will also be discussed with regards to resolving the tension between episodic and abstract models of phonetic category structure.

  17. Application of source biasing technique for energy efficient DECODER circuit design: memory array application

    Science.gov (United States)

    Gupta, Neha; Parihar, Priyanka; Neema, Vaibhav

    2018-04-01

    Researchers have proposed many circuit techniques to reduce leakage power dissipation in memory cells. If we want to reduce the overall power in the memory system, we have to work on the input circuitry of memory architecture i.e. row and column decoder. In this research work, low leakage power with a high speed row and column decoder for memory array application is designed and four new techniques are proposed. In this work, the comparison of cluster DECODER, body bias DECODER, source bias DECODER, and source coupling DECODER are designed and analyzed for memory array application. Simulation is performed for the comparative analysis of different DECODER design parameters at 180 nm GPDK technology file using the CADENCE tool. Simulation results show that the proposed source bias DECODER circuit technique decreases the leakage current by 99.92% and static energy by 99.92% at a supply voltage of 1.2 V. The proposed circuit also improves dynamic power dissipation by 5.69%, dynamic PDP/EDP 65.03% and delay 57.25% at 1.2 V supply voltage.

  18. Systolic array processing of the sequential decoding algorithm

    Science.gov (United States)

    Chang, C. Y.; Yao, K.

    1989-01-01

    A systolic array processing technique is applied to implementing the stack algorithm form of the sequential decoding algorithm. It is shown that sorting, a key function in the stack algorithm, can be efficiently realized by a special type of systolic arrays known as systolic priority queues. Compared to the stack-bucket algorithm, this approach is shown to have the advantages that the decoding always moves along the optimal path, that it has a fast and constant decoding speed and that its simple and regular hardware architecture is suitable for VLSI implementation. Three types of systolic priority queues are discussed: random access scheme, shift register scheme and ripple register scheme. The property of the entries stored in the systolic priority queue is also investigated. The results are applicable to many other basic sorting type problems.

  19. [Participation as Target of Social Medicine and Nursing Care: - Legal Definition of Long-Term Care Dependency - Strategies to Prevent Long-Term Care Dependency].

    Science.gov (United States)

    Nüchtern, Elisabeth; Gansweid, Barbara; Gerber, Hans; von Mittelstaedt, Gert

    2017-01-01

    Objective: By the "Second Bill to Strengthen Long-Term Care", a new concept of long-term care dependency will be introduced, valid from 2017. Long-term care dependency according to Social Code XI will be defined covering more aspects than today. Therefore, the working group "Nursing Care" of the division "Social Medicine in Practice and Rehabilitation" in the German Society for Social Medicine and Prevention presents their results after working on the social medicine perspective of the definition and prevention of long-term care dependency. Methods: Both the definition and strategies to prevent long-term care dependency are systematically taken into consideration from the point of view of social medicine on the basis of the International Classification of Functioning, Disability and Health (ICF), as long-term care dependency means a defined condition of disability. Results: Both the current and the new concept of long-term care dependency focus activity limitations. The perspective of social medicine considers the interactions of health condition, its effects on daily activities and personal as well as environmental factors. From this point of view approaches for social benefits concerning prevention and rehabilitation can be identified systematically so as to work against the development and progression of long-term care dependency. The reference to the ICF can facilitate the communication between different professions. The new "graduation" of long-term care dependency would allow an international "translation" referring to the ICF. Conclusion: Experts from the field of social medicine as well as those of nursing care, care-givers and nursing researchers have in common the objective that persons in need of nursing care can participate in as many aspects of life of importance to them in an autonomous and self-determined way. The point of view of social medicine on long-term care dependency is fundamental for all occupational groups that are involved and for their

  20. A new image for long-term care.

    Science.gov (United States)

    Wager, Richard; Creelman, William

    2004-04-01

    To counter widely held negative images of long-term care, managers in the industry should implement quality-improvement initiatives that include six key strategies: Manage the expectations of residents and their families. Address customers' concerns early. Build long-term customer satisfaction. Allocate resources to achieve exceptional outcomes in key areas. Respond to adverse events with compassion. Reinforce the facility's credibility.

  1. Techniques and Architectures for Hazard-Free Semi-Parallel Decoding of LDPC Codes

    Directory of Open Access Journals (Sweden)

    Rovini Massimo

    2009-01-01

    Full Text Available The layered decoding algorithm has recently been proposed as an efficient means for the decoding of low-density parity-check (LDPC codes, thanks to the remarkable improvement in the convergence speed (2x of the decoding process. However, pipelined semi-parallel decoders suffer from violations or "hazards" between consecutive updates, which not only violate the layered principle but also enforce the loops in the code, thus spoiling the error correction performance. This paper describes three different techniques to properly reschedule the decoding updates, based on the careful insertion of "idle" cycles, to prevent the hazards of the pipeline mechanism. Also, different semi-parallel architectures of a layered LDPC decoder suitable for use with such techniques are analyzed. Then, taking the LDPC codes for the wireless local area network (IEEE 802.11n as a case study, a detailed analysis of the performance attained with the proposed techniques and architectures is reported, and results of the logic synthesis on a 65 nm low-power CMOS technology are shown.

  2. On the theoretical gap between synchronous and asynchronous MPC protocols

    DEFF Research Database (Denmark)

    Beerliová-Trubíniová, Zuzana; Hirt, Martin; Nielsen, Jesper Buus

    2010-01-01

    that in the cryptographic setting (with setup), the sole reason for it is the distribution of inputs: given an oracle for input distribution, cryptographically-secure asynchronous MPC is possible with the very same condition as synchronous MPC, namely t ..., we show that such an input-distribution oracle can be reduced to an oracle that allows each party to synchronously broadcast one single message. This means that when one single round of synchronous broadcast is available, then asynchronous MPC is possible at the same condition as synchronous MPC...

  3. Online decoding of object-based attention using real-time fMRI.

    Science.gov (United States)

    Niazi, Adnan M; van den Broek, Philip L C; Klanke, Stefan; Barth, Markus; Poel, Mannes; Desain, Peter; van Gerven, Marcel A J

    2014-01-01

    Visual attention is used to selectively filter relevant information depending on current task demands and goals. Visual attention is called object-based attention when it is directed to coherent forms or objects in the visual field. This study used real-time functional magnetic resonance imaging for moment-to-moment decoding of attention to spatially overlapped objects belonging to two different object categories. First, a whole-brain classifier was trained on pictures of faces and places. Subjects then saw transparently overlapped pictures of a face and a place, and attended to only one of them while ignoring the other. The category of the attended object, face or place, was decoded on a scan-by-scan basis using the previously trained decoder. The decoder performed at 77.6% accuracy indicating that despite competing bottom-up sensory input, object-based visual attention biased neural patterns towards that of the attended object. Furthermore, a comparison between different classification approaches indicated that the representation of faces and places is distributed rather than focal. This implies that real-time decoding of object-based attention requires a multivariate decoding approach that can detect these distributed patterns of cortical activity. © 2013 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  4. Long-term multipactor discharge in multicarrier systems

    International Nuclear Information System (INIS)

    Anza, S.; Vicente, C.; Gimeno, B.; Boria, V. E.; Armendariz, J.

    2007-01-01

    A new mechanism of long-term multipactor in multicarrier systems is studied employing both analytical and numerical methods. In particular, the investigation is focused on the impact that a realistic secondary emission yield at low energies produces on the development of long term multipactor. A novel analytical model for this interperiod charge accumulation is presented using the traditional multipactor theory for parallel plates, and approximating the multicarrier signal as a single-carrier signal modulated by a pulsed signal envelope. The analytical predictions are verified by numerical simulations for a typical rectangular waveguide. The analytical and numerical results demonstrate that the susceptibility of the system to develop a long-term multipactor discharge increases with higher values of low-energy secondary emission yield

  5. Multiple LDPC decoding for distributed source coding and video coding

    DEFF Research Database (Denmark)

    Forchhammer, Søren; Luong, Huynh Van; Huang, Xin

    2011-01-01

    Distributed source coding (DSC) is a coding paradigm for systems which fully or partly exploit the source statistics at the decoder to reduce the computational burden at the encoder. Distributed video coding (DVC) is one example. This paper considers the use of Low Density Parity Check Accumulate...... (LDPCA) codes in a DSC scheme with feed-back. To improve the LDPC coding performance in the context of DSC and DVC, while retaining short encoder blocks, this paper proposes multiple parallel LDPC decoding. The proposed scheme passes soft information between decoders to enhance performance. Experimental...

  6. The age-dependent epigenetic and physiological changes in an Arabidopsis T87 cell suspension culture during long-term cultivation

    Energy Technology Data Exchange (ETDEWEB)

    Kwiatkowska, Aleksandra, E-mail: A.Kwiatkows@gmail.com [Department of Botany, University of Rzeszow, Kolbuszowa (Poland); Zebrowski, Jacek [Department of Plant Physiology, University of Rzeszow, Kolbuszowa (Poland); Oklejewicz, Bernadetta [Department of Genetics, University of Rzeszow, Kolbuszowa (Poland); Czarnik, Justyna [Department of Botany, University of Rzeszow, Kolbuszowa (Poland); Halibart-Puzio, Joanna [Department of Plant Physiology, University of Rzeszow, Kolbuszowa (Poland); Wnuk, Maciej [Department of Genetics, University of Rzeszow, Kolbuszowa (Poland)

    2014-05-02

    Highlights: • A decrease in proliferation rate during long-term cultivation of Arabidopsis cells. • Age-dependent increase in senescence-associated gene expression in Arabidopsis cells. • Age-related increase in DNA methylation, H3K9me2, and H3K27me3 in Arabidopsis cells. • High potential of photosynthetic efficiency of long-term cultured Arabidopsis cells. - Abstract: Plant cell suspension cultures represent good model systems applicable for both basic research and biotechnological purposes. Nevertheless, it is widely known that a prolonged in vitro cultivation of plant cells is associated with genetic and epigenetic instabilities, which may limit the usefulness of plant lines. In this study, the age-dependent epigenetic and physiological changes in an asynchronous Arabidopsis T87 cell culture were examined. A prolonged cultivation period was found to be correlated with a decrease in the proliferation rate and a simultaneous increase in the expression of senescence-associated genes, indicating that the aging process started at the late growth phase of the culture. In addition, increases in the heterochromatin-specific epigenetic markers, i.e., global DNA methylation, H3K9 dimethylation, and H3K27 trimethylation, were observed, suggesting the onset of chromatin condensation, a hallmark of the early stages of plant senescence. Although the number of live cells decreased with an increase in the age of the culture, the remaining viable cells retained a high potential to efficiently perform photosynthesis and did not exhibit any symptoms of photosystem II damage.

  7. The age-dependent epigenetic and physiological changes in an Arabidopsis T87 cell suspension culture during long-term cultivation

    International Nuclear Information System (INIS)

    Kwiatkowska, Aleksandra; Zebrowski, Jacek; Oklejewicz, Bernadetta; Czarnik, Justyna; Halibart-Puzio, Joanna; Wnuk, Maciej

    2014-01-01

    Highlights: • A decrease in proliferation rate during long-term cultivation of Arabidopsis cells. • Age-dependent increase in senescence-associated gene expression in Arabidopsis cells. • Age-related increase in DNA methylation, H3K9me2, and H3K27me3 in Arabidopsis cells. • High potential of photosynthetic efficiency of long-term cultured Arabidopsis cells. - Abstract: Plant cell suspension cultures represent good model systems applicable for both basic research and biotechnological purposes. Nevertheless, it is widely known that a prolonged in vitro cultivation of plant cells is associated with genetic and epigenetic instabilities, which may limit the usefulness of plant lines. In this study, the age-dependent epigenetic and physiological changes in an asynchronous Arabidopsis T87 cell culture were examined. A prolonged cultivation period was found to be correlated with a decrease in the proliferation rate and a simultaneous increase in the expression of senescence-associated genes, indicating that the aging process started at the late growth phase of the culture. In addition, increases in the heterochromatin-specific epigenetic markers, i.e., global DNA methylation, H3K9 dimethylation, and H3K27 trimethylation, were observed, suggesting the onset of chromatin condensation, a hallmark of the early stages of plant senescence. Although the number of live cells decreased with an increase in the age of the culture, the remaining viable cells retained a high potential to efficiently perform photosynthesis and did not exhibit any symptoms of photosystem II damage

  8. Partially blind instantly decodable network codes for lossy feedback environment

    KAUST Repository

    Sorour, Sameh; Douik, Ahmed S.; Valaee, Shahrokh; Al-Naffouri, Tareq Y.; Alouini, Mohamed-Slim

    2014-01-01

    an expression for the expected decoding delay increment for any arbitrary transmission. This expression is then used to find the optimal policy that reduces the decoding delay in such lossy feedback environment. Results show that our proposed solutions both

  9. Long-term characteristics of nuclear emulsion

    International Nuclear Information System (INIS)

    Naganawa, N; Kuwabara, K

    2010-01-01

    Long-term characteristics of the nuclear emulsion so called 'OPERA film' used in the neutrino oscillation experiment, OPERA, has been studied for 8 years since its production or refreshing after it. In the results, it turned out to be excellent in sensitivity, amount of random noise, and refreshing characteristics. The retention capacity of latent image of tracks was also studied. The result will open the way to the recycling of 7,000,000 emulsion films which will remain not developed after 5 years of OPERA's run, and other long-term experiments with emulsion.

  10. Long-term characteristics of nuclear emulsion

    Science.gov (United States)

    Naganawa, N.; Kuwabara, K.

    2010-02-01

    Long-term characteristics of the nuclear emulsion so called ``OPERA film'' used in the neutrino oscillation experiment, OPERA, has been studied for 8 years since its production or refreshing after it. In the results, it turned out to be excellent in sensitivity, amount of random noise, and refreshing characteristics. The retention capacity of latent image of tracks was also studied. The result will open the way to the recycling of 7,000,000 emulsion films which will remain not developed after 5 years of OPERA's run, and other long-term experiments with emulsion.

  11. Asynchronous Multiparty Computation

    DEFF Research Database (Denmark)

    Damgård, Ivan Bjerre; Geisler, Martin; Krøigaard, Mikkel

    2009-01-01

    guarantees termination if the adversary allows a preprocessing phase to terminate, in which no information is released. The communication complexity of this protocol is the same as that of a passively secure solution up to a constant factor. It is secure against an adaptive and active adversary corrupting...... less than n/3 players. We also present a software framework for implementation of asynchronous protocols called VIFF (Virtual Ideal Functionality Framework), which allows automatic parallelization of primitive operations such as secure multiplications, without having to resort to complicated...... multithreading. Benchmarking of a VIFF implementation of our protocol confirms that it is applicable to practical non-trivial secure computations....

  12. In search of the elusive long-term price

    International Nuclear Information System (INIS)

    Connor, M.J.; Combs, J.

    1989-01-01

    The Uranium Institute, WNFM, and past USCEA sessions described and compared existing price reporting systems. The McGraw-Hill conference led to a rather heated discussion as to the propriety of spot prices having the influence they do on amounts paid in long-term contracts. The Ux representative proposed a future's market as a way that producers could hedge against some of the uncertainty of volatile spot market. In discussing the search for the elusive long-term price, there are two interrelated issues. The first is obvious-the search for a starting or initializing price that is representative of recently-signed or pending long-term contracts. The second is less obvious, but perhaps more important-the search for a successful mechanism for determining later delivery values in long-term contracts. This paper addresses the question of pricing mechanisms first

  13. Argo: A Time-Elastic Time-Division-Multiplexed NOC using Asynchronous Routers

    DEFF Research Database (Denmark)

    Kasapaki, Evangelia; Sparsø, Jens

    2014-01-01

    are either synchronous or mesochronous. We use asynchronous routers to achieve a simpler, smaller, and more robust, self-timed design. Our design exploits the fact that pipelined asynchronous circuits also behave as ripple FIFOs. Thus, it avoids the need for explicit synchronization FIFOs between the routers......In this paper we explore the use of asynchronous routers in a time-division-multiplexed (TDM) network-on-chip (NOC), Argo, that is being developed for a multi-processor platform for hard real-time systems. TDM inherently requires a common time reference, and existing TDM-based NOC designs...... delays derived from a 65nm CMOS implementation, a worstcase analysis shows that a typical design can tolerate a skew of 1-5 cycles (depending on FIFO depths and NI clock frequency). Simulation results of a 2 x 2 NOC confirm this....

  14. Very-long-term and short-term chromatic adaptation: are their influences cumulative?

    Science.gov (United States)

    Belmore, Suzanne C; Shevell, Steven K

    2011-02-09

    Very-long-term (VLT) chromatic adaptation results from exposure to an altered chromatic environment for days or weeks. Color shifts from VLT adaptation are observed hours or days after leaving the altered environment. Short-term chromatic adaptation, on the other hand, results from exposure for a few minutes or less, with color shifts measured within seconds or a few minutes after the adapting light is extinguished; recovery to the pre-adapted state is complete in less than an hour. Here, both types of adaptation were combined. All adaptation was to reddish-appearing long-wavelength light. Shifts in unique yellow were measured following adaptation. Previous studies demonstrate shifts in unique yellow due to VLT chromatic adaptation, but shifts from short-term chromatic adaptation to comparable adapting light can be far greater than from VLT adaptation. The question considered here is whether the color shifts from VLT adaptation are cumulative with large shifts from short-term adaptation or, alternatively, does simultaneous short-term adaptation eliminate color shifts caused by VLT adaptation. The results show the color shifts from VLT and short-term adaptation together are cumulative, which indicates that both short-term and very-long-term chromatic adaptation affect color perception during natural viewing. Copyright © 2010 Elsevier Ltd. All rights reserved.

  15. Marketing in the long-term care continuum.

    Science.gov (United States)

    Laurence, J Nathan; Kash, Bita A

    2010-04-01

    Today, long-term care facilities are composed of independent, assisted living, and skilled nursing facilities along with many variations of those themes in between. The clientele for these various types of facilities differ because of the level of care the facility provides as well as the amenities long-term care consumers are looking for. However, there many similarities and common approaches to how reaching the target audience through effective marketing activities. Knowing who the target audience is, how to reach them, and how to communicate with them will serve any facility well in this competitive market. Developing marketing strategies for long-term care settings is as important as understanding what elements of care can be marketed individually as a niche market. Determining the market base for a facility is equally crucial since the target populations differ among the three types of facilities. By reviewing current marketing articles and applying marketing practices, we have crafted some general principles for which each facility type can learn from. Finally, we will discuss the types of marketing and how they related to the spectrum of long-term care facilities.

  16. VLSI Architectures for Sliding-Window-Based Space-Time Turbo Trellis Code Decoders

    Directory of Open Access Journals (Sweden)

    Georgios Passas

    2012-01-01

    Full Text Available The VLSI implementation of SISO-MAP decoders used for traditional iterative turbo coding has been investigated in the literature. In this paper, a complete architectural model of a space-time turbo code receiver that includes elementary decoders is presented. These architectures are based on newly proposed building blocks such as a recursive add-compare-select-offset (ACSO unit, A-, B-, Γ-, and LLR output calculation modules. Measurements of complexity and decoding delay of several sliding-window-technique-based MAP decoder architectures and a proposed parameter set lead to defining equations and comparison between those architectures.

  17. An Evaluation of Parallel Synchronous and Conservative Asynchronous Logic-Level Simulations

    Directory of Open Access Journals (Sweden)

    Ausif Mahmood

    1996-01-01

    a circuit remain fixed during the entire simulation. We remove this limitation and, by extending the analyses to multi-input, multi-output circuits with an arbitrary number of input events, show that the conservative asynchronous simulation extracts more parallelism and executes faster than synchronous simulation in general. Our conclusions are supported by a comparison of the idealized execution times of synchronous and conservative asynchronous algorithms on ISCAS combinational and sequential benchmark circuits.

  18. Effects of synchronous versus asynchronous mode of propulsion on wheelchair basketball sprinting.

    Science.gov (United States)

    Faupin, Arnaud; Borel, Benoit; Meyer, Christophe; Gorce, Philippe; Watelain, Eric

    2013-11-01

    This study aimed to first investigate synchronous (SYN) versus asynchronous (ASY) mode of propulsion and, second, investigate the wheel camber effects on sprinting performance as well as temporal parameters. Seven wheelchair basketball players performed four maximal eight-second sprints on a wheelchair ergometer. They repeated the test according to two modes of propulsion (SYN and ASY) and two wheel cambers (9° and 15°). The mean maximal velocity and push power output was greater in the synchronous mode compared to the asynchronous mode for both camber angles. However, the fluctuation in the velocity profile is inferior for ASY versus SYN mode for both camber angles. Greater push time/cycle time (Pt/Ct) and arm frequency (AF) for synchronous mode versus asynchronous mode and inversely, lesser Ct and rest time (Rt) values for the synchronous mode, for which greater velocity were observed. SYN mode leads to better performance than ASY mode in terms of maximal propulsion velocity. However, ASY propulsion allows greater continuity of the hand-rim force application, reducing fluctuations in the velocity profile. The camber angle had no effect on ASY and SYN mean maximal velocity and push power output. The study of wheelchair propulsion strategies is important for better understanding physiological and biomechanical impacts of wheelchair propulsion for individuals with disabilities. From a kinematical point of view, this study highlights synchronous mode of propulsion to be more efficient, with regards to mean maximal velocity reaching during maximal sprinting exercises. Even if this study focuses on well-trained wheelchair athletes, results from this study could complement the knowledge on the physiological and biomechanical adaptations to wheelchair propulsion and therefore, might be interesting for wheelchair modifications for purposes of rehabilitation.

  19. Psychophysiological effects of synchronous versus asynchronous music during cycling.

    Science.gov (United States)

    Lim, Harry B T; Karageorghis, Costas I; Romer, Lee M; Bishop, Daniel T

    2014-02-01

    Synchronizing movement to a musical beat may reduce the metabolic cost of exercise, but findings to date have been equivocal. Our aim was to examine the degree to which the synchronous application of music moderates the metabolic demands of a cycle ergometer task. Twenty-three recreationally active men made two laboratory visits. During the first visit, participants completed a maximal incremental ramp test on a cycle ergometer. At the second visit, they completed four randomized 6-min cycling bouts at 90% of ventilatory threshold (control, metronome, synchronous music, and asynchronous music). Main outcome variables were oxygen uptake, HR, ratings of dyspnea and limb discomfort, affective valence, and arousal. No significant differences were evident for oxygen uptake. HR was lower under the metronome condition (122 ± 15 bpm) compared to asynchronous music (124 ± 17 bpm) and control (125 ± 16 bpm). Limb discomfort was lower while listening to the metronome (2.5 ± 1.2) and synchronous music (2.3 ± 1.1) compared to control (3.0 ± 1.5). Both music conditions, synchronous (1.9 ± 1.2) and asynchronous (2.1 ± 1.3), elicited more positive affective valence compared to metronome (1.2 ± 1.4) and control (1.2 ± 1.2), while arousal was higher with synchronous music (3.4 ± 0.9) compared to metronome (2.8 ± 1.0) and control (2.8 ± 0.9). Synchronizing movement to a rhythmic stimulus does not reduce metabolic cost but may lower limb discomfort. Moreover, synchronous music has a stronger effect on limb discomfort and arousal when compared to asynchronous music.

  20. Designing indicators of long-term energy supply security

    International Nuclear Information System (INIS)

    Jansen, J.C.; Van Arkel, W.G.; Boots, M.G.

    2004-01-01

    To our knowledge, so far amazingly little research work has been undertaken to construct meaningful indicators of long-run energy supply security for a particular nation or region. Currently, in addressing energy supply security, policy makers tend to emphasise short-term supply disruptions. In contrast, this pre-study accords with the broader Sustainability Outlook in considering the long-term perspective. This report starts with taking stock, in a concise way, of the official EU energy outlook and issues related to the opportunities to administer changes in the energy mix at the level of major energy use categories. Then a brief survey of relevant literature is made on long-term strategies to ensure survival of systems - be it biological, social, etc. - in an environment largely characterised by high uncertainty and a lot of unchartered territory. We found the work of Andrew Stirling very inspiring in this context. Based on his work and considering the limitations of the present research activity, we retained the Shannon index as the best 'simple' indicator of diversity. In the core of the report, the Shannon index is elaborated into four indicators of long-term energy supply security. Stepwise, additional aspects of long-term energy supply security are introduced. These aspects are: Diversification of energy sources in energy supply; Diversification of imports with respect to imported energy sources; Long-term political stability in regions of origin; The resource base in regions of origin, including the home region/country itself. After small adjustments to allow for data availability, these indicators were applied to the reference year 2030 of four long-term scenarios with data of base year 1995 and projections for underlying variables provided by the Netherlands Environmental Assessment Agency (MNP). Preliminary interpretation of the results suggests the usefulness of the indicators presented in this report. A second activity undertaken in this report was

  1. Sleep Quality, Short-Term and Long-Term CPAP Adherence

    Science.gov (United States)

    Somiah, Manya; Taxin, Zachary; Keating, Joseph; Mooney, Anne M.; Norman, Robert G.; Rapoport, David M.; Ayappa, Indu

    2012-01-01

    Study Objectives: Adherence to CPAP therapy is low in patients with obstructive sleep apnea/hypopnea syndrome (OSAHS). The purpose of the present study was to evaluate the utility of measures of sleep architecture and sleep continuity on the CPAP titration study as predictors of both short- and long-term CPAP adherence. Methods: 93 patients with OSAHS (RDI 42.8 ± 34.3/h) underwent in-laboratory diagnostic polysomnography, CPAP titration, and follow-up polysomnography (NPSG) on CPAP. Adherence to CPAP was objectively monitored. Short-term (ST) CPAP adherence was averaged over 14 days immediately following the titration study. Long-term (LT) CPAP adherence was obtained in 56/93 patients after approximately 2 months of CPAP use. Patients were grouped into CPAP adherence groups for ST ( 4 h) and LT adherence ( 4 h). Sleep architecture, sleep disordered breathing (SDB) indices, and daytime outcome variables from the diagnostic and titration NPSGs were compared between CPAP adherence groups. Results: There was a significant relationship between ST and LT CPAP adherence (r = 0.81, p CPAP adherence groups had significantly lower %N2 and greater %REM on the titration NPSG. A model combining change in sleep efficiency and change in sleep continuity between the diagnostic and titration NPSGs predicted 17% of the variance in LT adherence (p = 0.006). Conclusions: These findings demonstrate that characteristics of sleep architecture, even on the titration NPSG, may predict some of the variance in CPAP adherence. Better sleep quality on the titration night was related to better CPAP adherence, suggesting that interventions to improve sleep on/prior to the CPAP titration study might be used as a therapeutic intervention to improve CPAP adherence. Citation: Somiah M; Taxin Z; Keating J; Mooney AM; Norman RG; Rapoport DM; Ayappa I. Sleep quality, short-term and long-term CPAP adherence. J Clin Sleep Med 2012;8(5):489-500. PMID:23066359

  2. On Modeling Large-Scale Multi-Agent Systems with Parallel, Sequential and Genuinely Asynchronous Cellular Automata

    International Nuclear Information System (INIS)

    Tosic, P.T.

    2011-01-01

    We study certain types of Cellular Automata (CA) viewed as an abstraction of large-scale Multi-Agent Systems (MAS). We argue that the classical CA model needs to be modified in several important respects, in order to become a relevant and sufficiently general model for the large-scale MAS, and so that thus generalized model can capture many important MAS properties at the level of agent ensembles and their long-term collective behavior patterns. We specifically focus on the issue of inter-agent communication in CA, and propose sequential cellular automata (SCA) as the first step, and genuinely Asynchronous Cellular Automata (ACA) as the ultimate deterministic CA-based abstract models for large-scale MAS made of simple reactive agents. We first formulate deterministic and nondeterministic versions of sequential CA, and then summarize some interesting configuration space properties (i.e., possible behaviors) of a restricted class of sequential CA. In particular, we compare and contrast those properties of sequential CA with the corresponding properties of the classical (that is, parallel and perfectly synchronous) CA with the same restricted class of update rules. We analytically demonstrate failure of the studied sequential CA models to simulate all possible behaviors of perfectly synchronous parallel CA, even for a very restricted class of non-linear totalistic node update rules. The lesson learned is that the interleaving semantics of concurrency, when applied to sequential CA, is not refined enough to adequately capture the perfect synchrony of parallel CA updates. Last but not least, we outline what would be an appropriate CA-like abstraction for large-scale distributed computing insofar as the inter-agent communication model is concerned, and in that context we propose genuinely asynchronous CA. (author)

  3. Emphasis on the Impact of Asynchronous Media

    African Journals Online (AJOL)

    ICTs and their utilization is one of the most pertinent issues in the education industry today. ... The paper pointed out specific impact of asynchronous ICT media in ... The paper finally noted that the struggle to be part of the digital world is ...

  4. Decoding English Alphabet Letters Using EEG Phase Information

    Directory of Open Access Journals (Sweden)

    YiYan Wang

    2018-02-01

    Full Text Available Increasing evidence indicates that the phase pattern and power of the low frequency oscillations of brain electroencephalograms (EEG contain significant information during the human cognition of sensory signals such as auditory and visual stimuli. Here, we investigate whether and how the letters of the alphabet can be directly decoded from EEG phase and power data. In addition, we investigate how different band oscillations contribute to the classification and determine the critical time periods. An English letter recognition task was assigned, and statistical analyses were conducted to decode the EEG signal corresponding to each letter visualized on a computer screen. We applied support vector machine (SVM with gradient descent method to learn the potential features for classification. It was observed that the EEG phase signals have a higher decoding accuracy than the oscillation power information. Low-frequency theta and alpha oscillations have phase information with higher accuracy than do other bands. The decoding performance was best when the analysis period began from 180 to 380 ms after stimulus presentation, especially in the lateral occipital and posterior temporal scalp regions (PO7 and PO8. These results may provide a new approach for brain-computer interface techniques (BCI and may deepen our understanding of EEG oscillations in cognition.

  5. Multi-Trial Guruswami–Sudan Decoding for Generalised Reed–Solomon Codes

    DEFF Research Database (Denmark)

    Nielsen, Johan Sebastian Rosenkilde; Zeh, Alexander

    2013-01-01

    An iterated refinement procedure for the Guruswami–Sudan list decoding algorithm for Generalised Reed–Solomon codes based on Alekhnovich’s module minimisation is proposed. The method is parametrisable and allows variants of the usual list decoding approach. In particular, finding the list...

  6. An Asynchronous Cellular Automata-Based Adaptive Illumination Facility

    Science.gov (United States)

    Bandini, Stefania; Bonomi, Andrea; Vizzari, Giuseppe; Acconci, Vito

    The term Ambient Intelligence refers to electronic environments that are sensitive and responsive to the presence of people; in the described scenario the environment itself is endowed with a set of sensors (to perceive humans or other physical entities such as dogs, bicycles, etc.), interacting with a set of actuators (lights) that choose their actions (i.e. state of illumination) in an attempt improve the overall experience of these users. The model for the interaction and action of sensors and actuators is an asynchronous Cellular Automata (CA) with memory, supporting a self-organization of the system as a response to the presence and movements of people inside it. The paper will introduce the model, as well as an ad hoc user interface for the specification of the relevant parameters of the CA transition rule that determines the overall system behaviour.

  7. On the relationship between short- and long-term memory

    DEFF Research Database (Denmark)

    Sørensen, Thomas Alrik

    James (1890) divided memory into separate stores; primary and secondary – or short-term and long-term memory. The interaction between the two stores often assumes that information initially is represented in volatile short-term store before entering and consolidating in the more durable long-term......, accepted). Counter to popular beliefs this suggest that long-term memory precedes short-term memory and not vice versa....... memory system (e.g. Atkinson & Shiffrin, 1968). Short-term memory seems to provide a surprising processing bottleneck where only a very limited amount of information can be represented at any given moment (Miller, 1956; Cowan, 2001). A number of studies have investigated the nature of this processing...

  8. Multiple-Symbol Decision-Feedback Space-Time Differential Decoding in Fading Channels

    Directory of Open Access Journals (Sweden)

    Wang Xiaodong

    2002-01-01

    Full Text Available Space-time differential coding (STDC is an effective technique for exploiting transmitter diversity while it does not require the channel state information at the receiver. However, like conventional differential modulation schemes, it exhibits an error floor in fading channels. In this paper, we develop an STDC decoding technique based on multiple-symbol detection and decision-feedback, which makes use of the second-order statistic of the fading processes and has a very low computational complexity. This decoding method can significantly lower the error floor of the conventional STDC decoding algorithm, especially in fast fading channels. The application of the proposed multiple-symbol decision-feedback STDC decoding technique in orthogonal frequency-division multiplexing (OFDM system is also discussed.

  9. Decoding Interleaved Gabidulin Codes using Alekhnovich's Algorithm

    DEFF Research Database (Denmark)

    Puchinger, Sven; Müelich, Sven; Mödinger, David

    2017-01-01

    We prove that Alekhnovich's algorithm can be used for row reduction of skew polynomial matrices. This yields an O(ℓ3n(ω+1)/2log⁡(n)) decoding algorithm for ℓ-Interleaved Gabidulin codes of length n, where ω is the matrix multiplication exponent.......We prove that Alekhnovich's algorithm can be used for row reduction of skew polynomial matrices. This yields an O(ℓ3n(ω+1)/2log⁡(n)) decoding algorithm for ℓ-Interleaved Gabidulin codes of length n, where ω is the matrix multiplication exponent....

  10. Decoding LDPC Convolutional Codes on Markov Channels

    Directory of Open Access Journals (Sweden)

    Kashyap Manohar

    2008-01-01

    Full Text Available Abstract This paper describes a pipelined iterative technique for joint decoding and channel state estimation of LDPC convolutional codes over Markov channels. Example designs are presented for the Gilbert-Elliott discrete channel model. We also compare the performance and complexity of our algorithm against joint decoding and state estimation of conventional LDPC block codes. Complexity analysis reveals that our pipelined algorithm reduces the number of operations per time step compared to LDPC block codes, at the expense of increased memory and latency. This tradeoff is favorable for low-power applications.

  11. Decoding LDPC Convolutional Codes on Markov Channels

    Directory of Open Access Journals (Sweden)

    Chris Winstead

    2008-04-01

    Full Text Available This paper describes a pipelined iterative technique for joint decoding and channel state estimation of LDPC convolutional codes over Markov channels. Example designs are presented for the Gilbert-Elliott discrete channel model. We also compare the performance and complexity of our algorithm against joint decoding and state estimation of conventional LDPC block codes. Complexity analysis reveals that our pipelined algorithm reduces the number of operations per time step compared to LDPC block codes, at the expense of increased memory and latency. This tradeoff is favorable for low-power applications.

  12. Heating calculation features at self-start of large asynchronous motor

    Science.gov (United States)

    Shevchenko, A. A.; Temlyakova, Z. S.; Grechkin, V. V.; Vilberger, M. E.

    2017-10-01

    The article proposes a method for optimizing the incremental heating calculation in the active volume of a large asynchronous motor for certain kinds of load characteristics. The incremental heating calculation is conditioned by the need to determine the aging level of the insulation and to predict a decrease in the electric machine service life. The method for optimizing the incremental heating calculation of asynchronous motor active volume is based on the automation of calculating the heating when simulating the self-starting process of the motor after eliminating an AC drop.

  13. Design issues in the semantics and scheduling of asynchronous tasks.

    Energy Technology Data Exchange (ETDEWEB)

    Olivier, Stephen L.

    2013-07-01

    The asynchronous task model serves as a useful vehicle for shared memory parallel programming, particularly on multicore and manycore processors. As adoption of model among programmers has increased, support has emerged for the integration of task parallel language constructs into mainstream programming languages, e.g., C and C++. This paper examines some of the design decisions in Cilk and OpenMP concerning semantics and scheduling of asynchronous tasks with the aim of informing the efforts of committees considering language integration, as well as developers of new task parallel languages and libraries.

  14. Toward a universal decoder of linguistic meaning from brain activation.

    Science.gov (United States)

    Pereira, Francisco; Lou, Bin; Pritchett, Brianna; Ritter, Samuel; Gershman, Samuel J; Kanwisher, Nancy; Botvinick, Matthew; Fedorenko, Evelina

    2018-03-06

    Prior work decoding linguistic meaning from imaging data has been largely limited to concrete nouns, using similar stimuli for training and testing, from a relatively small number of semantic categories. Here we present a new approach for building a brain decoding system in which words and sentences are represented as vectors in a semantic space constructed from massive text corpora. By efficiently sampling this space to select training stimuli shown to subjects, we maximize the ability to generalize to new meanings from limited imaging data. To validate this approach, we train the system on imaging data of individual concepts, and show it can decode semantic vector representations from imaging data of sentences about a wide variety of both concrete and abstract topics from two separate datasets. These decoded representations are sufficiently detailed to distinguish even semantically similar sentences, and to capture the similarity structure of meaning relationships between sentences.

  15. A Parallel Decoding Algorithm for Short Polar Codes Based on Error Checking and Correcting

    Science.gov (United States)

    Pan, Xiaofei; Pan, Kegang; Ye, Zhan; Gong, Chao

    2014-01-01

    We propose a parallel decoding algorithm based on error checking and correcting to improve the performance of the short polar codes. In order to enhance the error-correcting capacity of the decoding algorithm, we first derive the error-checking equations generated on the basis of the frozen nodes, and then we introduce the method to check the errors in the input nodes of the decoder by the solutions of these equations. In order to further correct those checked errors, we adopt the method of modifying the probability messages of the error nodes with constant values according to the maximization principle. Due to the existence of multiple solutions of the error-checking equations, we formulate a CRC-aided optimization problem of finding the optimal solution with three different target functions, so as to improve the accuracy of error checking. Besides, in order to increase the throughput of decoding, we use a parallel method based on the decoding tree to calculate probability messages of all the nodes in the decoder. Numerical results show that the proposed decoding algorithm achieves better performance than that of some existing decoding algorithms with the same code length. PMID:25540813

  16. Analysis and Design of Binary Message-Passing Decoders

    DEFF Research Database (Denmark)

    Lechner, Gottfried; Pedersen, Troels; Kramer, Gerhard

    2012-01-01

    Binary message-passing decoders for low-density parity-check (LDPC) codes are studied by using extrinsic information transfer (EXIT) charts. The channel delivers hard or soft decisions and the variable node decoder performs all computations in the L-value domain. A hard decision channel results...... message-passing decoders. Finally, it is shown that errors on cycles consisting only of degree two and three variable nodes cannot be corrected and a necessary and sufficient condition for the existence of a cycle-free subgraph is derived....... in the well-know Gallager B algorithm, and increasing the output alphabet from hard decisions to two bits yields a gain of more than 1.0 dB in the required signal to noise ratio when using optimized codes. The code optimization requires adapting the mixing property of EXIT functions to the case of binary...

  17. UNIVERSAL REGULAR AUTONOMOUS ASYNCHRONOUS SYSTEMS: ω-LIMIT SETS, INVARIANCE AND BASINS OF ATTRACTION

    Directory of Open Access Journals (Sweden)

    Serban Vlad

    2011-07-01

    Full Text Available The asynchronous systems are the non-deterministic real timebinarymodels of the asynchronous circuits from electrical engineering.Autonomy means that the circuits and their models have no input.Regularity means analogies with the dynamical systems, thus such systems may be considered to be real time dynamical systems with a’vector field’, Universality refers to the case when the state space of the system is the greatest possible in the sense of theinclusion. The purpose of this paper is that of defining, by analogy with the dynamical systems theory, the omega-limit sets, the invariance and the basins of attraction of the universal regular autonomous asynchronous systems.

  18. PSA modeling of long-term accident sequences

    International Nuclear Information System (INIS)

    Georgescu, Gabriel; Corenwinder, Francois; Lanore, Jeanne-Marie

    2014-01-01

    In the context of the extension of PSA scope to include external hazards, in France, both operator (EDF) and IRSN work for the improvement of methods to better take into account in the PSA the accident sequences induced by initiators which affect a whole site containing several nuclear units (reactors, fuel pools,...). These methodological improvements represent an essential prerequisite for the development of external hazards PSA. However, it has to be noted that in French PSA, even before Fukushima, long term accident sequences were taken into account: many insight were therefore used, as complementary information, to enhance the safety level of the plants. IRSN proposed an external events PSA development program. One of the first steps of the program is the development of methods to model in the PSA the long term accident sequences, based on the experience gained. At short term IRSN intends to enhance the modeling of the 'long term' accident sequences induced by the loss of the heat sink or/and the loss of external power supply. The experience gained by IRSN and EDF from the development of several probabilistic studies treating long term accident sequences shows that the simple extension of the mission time of the mitigation systems from 24 hours to longer times is not sufficient to realistically quantify the risk and to obtain a correct ranking of the risk contributions and that treatment of recoveries is also necessary. IRSN intends to develop a generic study which can be used as a general methodology for the assessment of the long term accident sequences, mainly generated by external hazards and their combinations. This first attempt to develop this generic study allowed identifying some aspects, which may be hazard (or combinations of hazards) or related to initial boundary conditions, which should be taken into account for further developments. (authors)

  19. Long-term care financing: lessons from France.

    Science.gov (United States)

    Doty, Pamela; Nadash, Pamela; Racco, Nathalie

    2015-06-01

    POLICY POINTS: France's model of third-party coverage for long-term services and supports (LTSS) combines a steeply income-adjusted universal public program for people 60 or older with voluntary supplemental private insurance. French and US policies differ: the former pay cash; premiums are lower; and take-up rates are higher, in part because employer sponsorship, with and without subsidization, is more common-but also because coverage targets higher levels of need and pays a smaller proportion of costs. Such inexpensive, bare-bones private coverage, especially if marketed as a supplement to a limited public benefit, would be more affordable to those Americans currently most at risk of "spending down" to Medicaid. An aging population leads to a growing demand for long-term services and supports (LTSS). In 2002, France introduced universal, income-adjusted, public long-term care coverage for adults 60 and older, whereas the United States funds means-tested benefits only. Both countries have private long-term care insurance (LTCI) markets: American policies create alternatives to out-of-pocket spending and protect purchasers from relying on Medicaid. Sales, however, have stagnated, and the market's viability is uncertain. In France, private LTCI supplements public coverage, and sales are growing, although its potential to alleviate the long-term care financing problem is unclear. We explore whether France's very different approach to structuring public and private financing for long-term care could inform the United States' long-term care financing reform efforts. We consulted insurance experts and conducted a detailed review of public reports, academic studies, and other documents to understand the public and private LTCI systems in France, their advantages and disadvantages, and the factors affecting their development. France provides universal public coverage for paid assistance with functional dependency for people 60 and older. Benefits are steeply income

  20. What Legislators Need to Know about Long-Term Care Insurance.

    Science.gov (United States)

    Landes, David

    This booklet discusses the potential importance to states of long-term care insurance, describes general policy characteristics, and summarizes state actions to both regulate and promote long-term care insurance. It is intended as a resource for legislators and others involved in long-term care financing and public policy formulation. Long-term…

  1. Private long-term care insurance and state tax incentives.

    Science.gov (United States)

    Stevenson, David G; Frank, Richard G; Tau, Jocelyn

    2009-01-01

    To increase the role of private insurance in financing long-term care, tax incentives for long-term care insurance have been implemented at both the federal and state levels. To date, there has been surprisingly little study of these initiatives. Using a panel of national data, we find that market take-up for long-term care insurance increased over the last decade, but state tax incentives were responsible for only a small portion of this growth. Ultimately, the modest ability of state tax incentives to lower premiums implies that they should be viewed as a small piece of the long-term care financing puzzle.

  2. The Long-Term Conditions Questionnaire: conceptual framework and item development.

    Science.gov (United States)

    Peters, Michele; Potter, Caroline M; Kelly, Laura; Hunter, Cheryl; Gibbons, Elizabeth; Jenkinson, Crispin; Coulter, Angela; Forder, Julien; Towers, Ann-Marie; A'Court, Christine; Fitzpatrick, Ray

    2016-01-01

    To identify the main issues of importance when living with long-term conditions to refine a conceptual framework for informing the item development of a patient-reported outcome measure for long-term conditions. Semi-structured qualitative interviews (n=48) were conducted with people living with at least one long-term condition. Participants were recruited through primary care. The interviews were transcribed verbatim and analyzed by thematic analysis. The analysis served to refine the conceptual framework, based on reviews of the literature and stakeholder consultations, for developing candidate items for a new measure for long-term conditions. Three main organizing concepts were identified: impact of long-term conditions, experience of services and support, and self-care. The findings helped to refine a conceptual framework, leading to the development of 23 items that represent issues of importance in long-term conditions. The 23 candidate items formed the first draft of the measure, currently named the Long-Term Conditions Questionnaire. The aim of this study was to refine the conceptual framework and develop items for a patient-reported outcome measure for long-term conditions, including single and multiple morbidities and physical and mental health conditions. Qualitative interviews identified the key themes for assessing outcomes in long-term conditions, and these underpinned the development of the initial draft of the measure. These initial items will undergo cognitive testing to refine the items prior to further validation in a survey.

  3. Characteristics of asthma attack with long-term management for bronchial asthma.

    Science.gov (United States)

    Kawahara, Noriko; Hasegawa, Shunji; Hashimoto, Kunio; Matsubara, Tomoyo; Ichiyama, Takashi; Furukawa, Susumu

    2009-10-01

    There have been no reports on the evaluation of the usefulness of long-term asthma management based on the Japanese Pediatric Guideline for the Treatment and Management of Bronchial Asthma 2005 (JPGL 2005). The purpose of the present study was to retrospectively investigate the records of 350 patients admitted to Yamaguchi University Hospital who had asthma attacks from January 2006 to June 2008. There were 149 patients who were treated for more than 3 months in accordance with the guideline (long-term management group) and 201 who were not (non-long-term management group). The patients were divided into three age groups: 100 infants, 159 toddlers, and 91 schoolchildren. The onset age of asthma in the long-term management group was earlier than that in the non-long-term management group in toddlers and schoolchildren. The white blood cell counts and C-reactive protein levels were higher in the non-long-term management group in schoolchildren, suggesting the complication of some infections. The severity of asthma in the long-term management group was greater than that in the non-long-term management group among all three age groups. There were no significant differences, however, in the severity of asthma attack at admission between the long-term and non-long-term management groups in the three age groups. Patients who had severe asthma tended to be treated with long-term management, which suggests that long-term asthma management according to JPGL 2005 may reduce the severity of asthma attack at that admission, because the severity of asthma in patients undergoing long-term management correlates with the severity of asthma attack.

  4. An overview of turbo decoding on fading channels

    OpenAIRE

    ATILGAN, Doğan

    2009-01-01

    A review of turbo coding and decoding has been presented in the literature [1]. In that paper, turbo coding and decoding on AWGN (Additive White Gaussian Noise) channels has been elaborated. In wireless communications, a phenomennon called multipath fading is frequently encountered. Therefore, investigation of efficient techniques to tackle with the destructive effects of fading is essential. Turbo coding has been proven as an efficient channel coding technique for AWGN channels. Some of the ...

  5. Design and Analysis of Adaptive Message Coding on LDPC Decoder with Faulty Storage

    Directory of Open Access Journals (Sweden)

    Guangjun Ge

    2018-01-01

    Full Text Available Unreliable message storage severely degrades the performance of LDPC decoders. This paper discusses the impacts of message errors on LDPC decoders and schemes improving the robustness. Firstly, we develop a discrete density evolution analysis for faulty LDPC decoders, which indicates that protecting the sign bits of messages is effective enough for finite-precision LDPC decoders. Secondly, we analyze the effects of quantization precision loss for static sign bit protection and propose an embedded dynamic coding scheme by adaptively employing the least significant bits (LSBs to protect the sign bits. Thirdly, we give a construction of Hamming product code for the adaptive coding and present low complexity decoding algorithms. Theoretic analysis indicates that the proposed scheme outperforms traditional triple modular redundancy (TMR scheme in decoding both threshold and residual errors, while Monte Carlo simulations show that the performance loss is less than 0.2 dB when the storage error probability varies from 10-3 to 10-4.

  6. A novel asynchronous access method with binary interfaces

    Directory of Open Access Journals (Sweden)

    Torres-Solis Jorge

    2008-10-01

    Full Text Available Abstract Background Traditionally synchronous access strategies require users to comply with one or more time constraints in order to communicate intent with a binary human-machine interface (e.g., mechanical, gestural or neural switches. Asynchronous access methods are preferable, but have not been used with binary interfaces in the control of devices that require more than two commands to be successfully operated. Methods We present the mathematical development and evaluation of a novel asynchronous access method that may be used to translate sporadic activations of binary interfaces into distinct outcomes for the control of devices requiring an arbitrary number of commands to be controlled. With this method, users are required to activate their interfaces only when the device under control behaves erroneously. Then, a recursive algorithm, incorporating contextual assumptions relevant to all possible outcomes, is used to obtain an informed estimate of user intention. We evaluate this method by simulating a control task requiring a series of target commands to be tracked by a model user. Results When compared to a random selection, the proposed asynchronous access method offers a significant reduction in the number of interface activations required from the user. Conclusion This novel access method offers a variety of advantages over traditionally synchronous access strategies and may be adapted to a wide variety of contexts, with primary relevance to applications involving direct object manipulation.

  7. NUMERICAL WITHOUT ITERATION METHOD OF MODELING OF ELECTROMECHANICAL PROCESSES IN ASYNCHRONOUS ENGINES

    Directory of Open Access Journals (Sweden)

    D. G. Patalakh

    2018-02-01

    Full Text Available Purpose. Development of calculation of electromagnetic and electromechanic transients is in asynchronous engines without iterations. Methodology. Numeral methods of integration of usual differential equations, programming. Findings. As the system of equations, describing the dynamics of asynchronous engine, contents the products of rotor and stator currents and product of rotation frequency of rotor and currents, so this system is nonlinear one. The numeral solution of nonlinear differential equations supposes an iteration process on every step of integration. Time-continuing and badly converging iteration process may be the reason of calculation slowing. The improvement of numeral method by the way of an iteration process removing is offered. As result the modeling time is reduced. The improved numeral method is applied for integration of differential equations, describing the dynamics of asynchronous engine. Originality. The improvement of numeral method allowing to execute numeral integrations of differential equations containing product of functions is offered, that allows to avoid an iteration process on every step of integration and shorten modeling time. Practical value. On the basis of the offered methodology the universal program of modeling of electromechanics processes in asynchronous engines could be developed as taking advantage on fast-acting.

  8. Vectorization of Reed Solomon decoding and mapping on the EVP

    NARCIS (Netherlands)

    Kumar, A.; Berkel, van C.H.

    2008-01-01

    Reed Solomon (RS) codes are used in a variety of (wireless) communication systems. Although commonly implemented in dedicated hardware, this paper explores the mapping of high-throughput RS decoding on vector DSPs. The four modules of such a decoder, viz. Syndrome Computation, Key Equation Solver,

  9. Quantification of long term emission potential from landfills

    NARCIS (Netherlands)

    Heimovaara, T.J.

    2011-01-01

    Novel approaches for the after-care of Municipal Solid Waste (MSW) landfills are based on technological measures to reduce the long term emission potential in a short time period. Biological degradation in landfills is a means to significantly reduce the long term emission potential. Leachate

  10. Dynamic Performances of Asynchronous Machines | Ubeku ...

    African Journals Online (AJOL)

    The per-phase parameters of a 1.5 hp, 380 V, 50 Hz, 4 poles, 3 phase asynchronous machine used in the simulation were computed with reading obtained from a dc, no-load and blocked rotor tests carried out on the machine in the laboratory. The results obtained from the computer simulations confirmed the capabilities ...

  11. Long-term care financing through Federal tax incentives.

    Science.gov (United States)

    Moran, D W; Weingart, J M

    1988-12-01

    Congress and the Administration are currently exploring various methods of promoting access to long-term care. In this article, an inventory of recent legislative proposals for using the Federal tax code to expand access to long-term care services is provided. Proposals are arrayed along a functional typology that includes tax mechanisms to encourage accumulation of funds, promote purchase of long-term care insurance, or induce the diversion of funds accumulated for another purpose (such as individual retirement accounts). The proposals are evaluated against the public policy objective of encouraging risk pooling to minimize social cost.

  12. Evaluation of long term leaching of borosilicate glasses

    International Nuclear Information System (INIS)

    Lanza, F.; Parnisari, E.

    1978-01-01

    For the evaluation of long term hazard of glass, data on long term glass leaching are needed. Moreover for long term leaching a model of homogeneous dissolution seems reasonable and ask for confirmation. Tests were performed at 30 0 , 80 0 , 100 0 , using an apparatus of the Soxhlet type, to 3.600 hours. Results were obtained as a weight loss and analysed following a relation with time composed by a parabolic and a linear part. Analysis of the surface layer using energy dispersion X ray spectrometry were performed. A critical analysis of the results and of the apparatus is presented

  13. Long-Term Stewardship Program Science and Technology Requirements

    Energy Technology Data Exchange (ETDEWEB)

    Joan McDonald

    2002-09-01

    Many of the United States’ hazardous and radioactively contaminated waste sites will not be sufficiently remediated to allow unrestricted land use because funding and technology limitations preclude cleanup to pristine conditions. This means that after cleanup is completed, the Department of Energy will have long-term stewardship responsibilities to monitor and safeguard more than 100 sites that still contain residual contamination. Long-term stewardship encompasses all physical and institutional controls, institutions, information, and other mechanisms required to protect human health and the environment from the hazards remaining. The Department of Energy Long-Term Stewardship National Program is in the early stages of development, so considerable planning is still required to identify all the specific roles and responsibilities, policies, and activities needed over the next few years to support the program’s mission. The Idaho National Engineering and Environmental Laboratory was tasked with leading the development of Science and Technology within the Long-Term Stewardship National Program. As part of that role, a task was undertaken to identify the existing science and technology related requirements, identify gaps and conflicts that exist, and make recommendations to the Department of Energy for future requirements related to science and technology requirements for long-term stewardship. This work is summarized in this document.

  14. Long-term psychosocial consequences of surgical congenital malformations.

    Science.gov (United States)

    Diseth, Trond H; Emblem, Ragnhild

    2017-10-01

    Surgical congenital malformations often represent years of treatment, large number of hospital stays, treatment procedures, and long-term functional sequels affecting patients' psychosocial functioning. Both functional defects and psychosocial difficulties that occur commonly in childhood may pass through adolescence on to adulthood. This overview presents reports published over the past 3 decades to elucidate the long-term psychosocial consequences of surgical congenital malformations. Literature searches conducted on PubMed database revealed that less than 1% of all the records of surgical congenital malformations described long-term psychosocial consequences, but with diverse findings. This inconsistency may be due to methodological differences or deficiencies; especially in study design, patient sampling, and methods. Most of the studies revealed that the functional deficits may have great impact on patients' mental health, psychosocial functioning, and QoL; both short- and long-term negative consequences. Factors other than functional problems, e.g., repeated anesthesia, multiple hospitalization, traumatic treatment procedures, and parental dysfunctioning, may also predict long-term mental health and psychosocial functioning. Through multidisciplinary approach, pediatric surgeons should also be aware of deficits in emotional and psychosocial functioning. To achieve overall optimal psychosocial functioning, the challenge is to find a compromise between physically optimal treatment procedures and procedures that are not psychologically detrimental. Copyright © 2017. Published by Elsevier Inc.

  15. Short-term and long-term deflection of reinforced hollow core ...

    African Journals Online (AJOL)

    This paper presents a study on different methods of analysis that are currently used by design codes to predict the short-term and long-term deflection of reinforced concrete slab systems and compares the predicted deflections with measured deflections. The experimental work to measure deflections involved the testing of ...

  16. THE ROLE OF OFFLINE METALANGUAGE TALK IN ASYNCHRONOUS COMPUTER-MEDIATED COMMUNICATION

    Directory of Open Access Journals (Sweden)

    Keiko Kitade

    2008-02-01

    Full Text Available In order to demonstrate how learners utilize the text-based asynchronous attributes of the Bulletin Board System, this study explored Japanese-as-a-second-language learners' metalanguage episodes (Swain & Lapkin, 1995, 1998 in offline verbal peer speech and online asynchronous discussions with their Japanese key pals. The findings suggest the crucial role of offline collaborative dialogue, the interactional modes in which the episodes occur, and the unique discourse structure of metalanguage episodes concerning online and offline interactions. A high score on the posttest also suggests the high retention of linguistic knowledge constructed through offline peer dialogue. In the offline mode, the learners were able to collaboratively construct knowledge with peers in the stipulated time, while simultaneously focusing on task content in the online interaction. The retrospective interviews and questionnaires reveal the factors that could affect the benefits of the asynchronous computer-mediated communication medium for language learning.

  17. Instantly Decodable Network Coding: From Centralized to Device-to-Device Communications

    KAUST Repository

    Douik, Ahmed S.

    2015-05-01

    From its introduction to its quindecennial, network coding have built a strong reputation in enhancing packet recovery process and achieving maximum information flow in both wires and wireless networks. Traditional studies focused on optimizing the throughput of the network by proposing complex schemes that achieve optimal delay. With the shift toward distributed computing at mobile devices, throughput and complexity become both critical factors that affect the efficiency of a coding scheme. Instantly decodable network coding imposed itself as a new paradigm in network coding that trades off this two aspects. This paper presents a survey of instantly decodable network coding schemes that are proposed in the literature. The various schemes are identified, categorized and evaluated. Two categories can be distinguished namely the conventional centralized schemes and the distributed or cooperative schemes. For each scheme, the comparison is carried out in terms of reliability, performance, complexity and packet selection methodology. Although the performance is generally inversely proportional to the computation complexity, numerous successful schemes from both the performance and complexity viewpoint are identified.

  18. Instantly Decodable Network Coding: From Centralized to Device-to-Device Communications

    KAUST Repository

    Douik, Ahmed S.

    2015-01-01

    From its introduction to its quindecennial, network coding have built a strong reputation in enhancing packet recovery process and achieving maximum information flow in both wires and wireless networks. Traditional studies focused on optimizing the throughput of the network by proposing complex schemes that achieve optimal delay. With the shift toward distributed computing at mobile devices, throughput and complexity become both critical factors that affect the efficiency of a coding scheme. Instantly decodable network coding imposed itself as a new paradigm in network coding that trades off this two aspects. This paper presents a survey of instantly decodable network coding schemes that are proposed in the literature. The various schemes are identified, categorized and evaluated. Two categories can be distinguished namely the conventional centralized schemes and the distributed or cooperative schemes. For each scheme, the comparison is carried out in terms of reliability, performance, complexity and packet selection methodology. Although the performance is generally inversely proportional to the computation complexity, numerous successful schemes from both the performance and complexity viewpoint are identified.

  19. Do Short-Term Managerial Objectives Lead to Under- or Over-Investment in Long-Term Projects

    OpenAIRE

    Lucian Arye Bebchuk; Lars A. Stole

    1994-01-01

    This paper studies managerial decisions about investment in long-run projects in the presence of imperfect information (the market knows less about such investments than the firm's managers) and short-term managerial objectives (the managers are concerned about the short-term stock price as well as the long-term stock price). Prior work has suggested that imperfect information and short-term managerial objectives induce managers to underinvest in long-run projects. We show that either underin...

  20. Observation of long term potentiation in papain-based memory devices

    KAUST Repository

    Bag, A.; Hota, Mrinal Kanti; Mallik, Sandipan B.; Maì ti, Chinmay Kumar

    2014-01-01

    Biological synaptic behavior in terms of long term potentiation has been observed in papain-based (plant protein) memory devices (memristors) for the first time. Improvement in long term potentiation depends on pulse amplitude and width (duration). Continuous/repetitive dc voltage sweep leads to an increase in memristor conductivity leading to a long term memory in the 'learning' processes.