WorldWideScience

Sample records for irregular ldpc code

  1. Threshold Multi Split-Row algorithm for decoding irregular LDPC codes

    Directory of Open Access Journals (Sweden)

    Chakir Aqil

    2017-12-01

    Full Text Available In this work, we propose a new threshold multi split-row algorithm in order to improve the multi split-row algorithm for LDPC irregular codes decoding. We give a complete description of our algorithm as well as its advantages for the LDPC codes. The simulation results over an additive white gaussian channel show that an improvement in code error performance between 0.4 dB and 0.6 dB compared to the multi split-row algorithm.

  2. Ensemble Weight Enumerators for Protograph LDPC Codes

    Science.gov (United States)

    Divsalar, Dariush

    2006-01-01

    Recently LDPC codes with projected graph, or protograph structures have been proposed. In this paper, finite length ensemble weight enumerators for LDPC codes with protograph structures are obtained. Asymptotic results are derived as the block size goes to infinity. In particular we are interested in obtaining ensemble average weight enumerators for protograph LDPC codes which have minimum distance that grows linearly with block size. As with irregular ensembles, linear minimum distance property is sensitive to the proportion of degree-2 variable nodes. In this paper the derived results on ensemble weight enumerators show that linear minimum distance condition on degree distribution of unstructured irregular LDPC codes is a sufficient but not a necessary condition for protograph LDPC codes.

  3. Low Complexity Encoder of High Rate Irregular QC-LDPC Codes for Partial Response Channels

    Directory of Open Access Journals (Sweden)

    IMTAWIL, V.

    2011-11-01

    Full Text Available High rate irregular QC-LDPC codes based on circulant permutation matrices, for efficient encoder implementation, are proposed in this article. The structure of the code is an approximate lower triangular matrix. In addition, we present two novel efficient encoding techniques for generating redundant bits. The complexity of the encoder implementation depends on the number of parity bits of the code for the one-stage encoding and the length of the code for the two-stage encoding. The advantage of both encoding techniques is that few XOR-gates are used in the encoder implementation. Simulation results on partial response channels also show that the BER performance of the proposed code has gain over other QC-LDPC codes.

  4. A novel construction scheme of QC-LDPC codes based on the RU algorithm for optical transmission systems

    Science.gov (United States)

    Yuan, Jian-guo; Liang, Meng-qi; Wang, Yong; Lin, Jin-zhao; Pang, Yu

    2016-03-01

    A novel lower-complexity construction scheme of quasi-cyclic low-density parity-check (QC-LDPC) codes for optical transmission systems is proposed based on the structure of the parity-check matrix for the Richardson-Urbanke (RU) algorithm. Furthermore, a novel irregular QC-LDPC(4 288, 4 020) code with high code-rate of 0.937 is constructed by this novel construction scheme. The simulation analyses show that the net coding gain ( NCG) of the novel irregular QC-LDPC(4 288,4 020) code is respectively 2.08 dB, 1.25 dB and 0.29 dB more than those of the classic RS(255, 239) code, the LDPC(32 640, 30 592) code and the irregular QC-LDPC(3 843, 3 603) code at the bit error rate ( BER) of 10-6. The irregular QC-LDPC(4 288, 4 020) code has the lower encoding/decoding complexity compared with the LDPC(32 640, 30 592) code and the irregular QC-LDPC(3 843, 3 603) code. The proposed novel QC-LDPC(4 288, 4 020) code can be more suitable for the increasing development requirements of high-speed optical transmission systems.

  5. Enhancement of Unequal Error Protection Properties of LDPC Codes

    Directory of Open Access Journals (Sweden)

    Poulliat Charly

    2007-01-01

    Full Text Available It has been widely recognized in the literature that irregular low-density parity-check (LDPC codes exhibit naturally an unequal error protection (UEP behavior. In this paper, we propose a general method to emphasize and control the UEP properties of LDPC codes. The method is based on a hierarchical optimization of the bit node irregularity profile for each sensitivity class within the codeword by maximizing the average bit node degree while guaranteeing a minimum degree as high as possible. We show that this optimization strategy is efficient, since the codes that we optimize show better UEP capabilities than the codes optimized for the additive white Gaussian noise channel.

  6. Weight Distribution for Non-binary Cluster LDPC Code Ensemble

    Science.gov (United States)

    Nozaki, Takayuki; Maehara, Masaki; Kasai, Kenta; Sakaniwa, Kohichi

    In this paper, we derive the average weight distributions for the irregular non-binary cluster low-density parity-check (LDPC) code ensembles. Moreover, we give the exponential growth rate of the average weight distribution in the limit of large code length. We show that there exist $(2,d_c)$-regular non-binary cluster LDPC code ensembles whose normalized typical minimum distances are strictly positive.

  7. Fast QC-LDPC code for free space optical communication

    Science.gov (United States)

    Wang, Jin; Zhang, Qi; Udeh, Chinonso Paschal; Wu, Rangzhong

    2017-02-01

    Free Space Optical (FSO) Communication systems use the atmosphere as a propagation medium. Hence the atmospheric turbulence effects lead to multiplicative noise related with signal intensity. In order to suppress the signal fading induced by multiplicative noise, we propose a fast Quasi-Cyclic (QC) Low-Density Parity-Check (LDPC) code for FSO Communication systems. As a linear block code based on sparse matrix, the performances of QC-LDPC is extremely near to the Shannon limit. Currently, the studies on LDPC code in FSO Communications is mainly focused on Gauss-channel and Rayleigh-channel, respectively. In this study, the LDPC code design over atmospheric turbulence channel which is nether Gauss-channel nor Rayleigh-channel is closer to the practical situation. Based on the characteristics of atmospheric channel, which is modeled as logarithmic-normal distribution and K-distribution, we designed a special QC-LDPC code, and deduced the log-likelihood ratio (LLR). An irregular QC-LDPC code for fast coding, of which the rates are variable, is proposed in this paper. The proposed code achieves excellent performance of LDPC codes and can present the characteristics of high efficiency in low rate, stable in high rate and less number of iteration. The result of belief propagation (BP) decoding shows that the bit error rate (BER) obviously reduced as the Signal-to-Noise Ratio (SNR) increased. Therefore, the LDPC channel coding technology can effectively improve the performance of FSO. At the same time, the BER, after decoding reduces with the increase of SNR arbitrarily, and not having error limitation platform phenomenon with error rate slowing down.

  8. Optimized Irregular Low-Density Parity-Check Codes for Multicarrier Modulations over Frequency-Selective Channels

    Directory of Open Access Journals (Sweden)

    Valérian Mannoni

    2004-09-01

    Full Text Available This paper deals with optimized channel coding for OFDM transmissions (COFDM over frequency-selective channels using irregular low-density parity-check (LDPC codes. Firstly, we introduce a new characterization of the LDPC code irregularity called “irregularity profile.” Then, using this parameterization, we derive a new criterion based on the minimization of the transmission bit error probability to design an irregular LDPC code suited to the frequency selectivity of the channel. The optimization of this criterion is done using the Gaussian approximation technique. Simulations illustrate the good performance of our approach for different transmission channels.

  9. LDPC concatenated space-time block coded system in multipath fading environment: Analysis and evaluation

    Directory of Open Access Journals (Sweden)

    Surbhi Sharma

    2011-06-01

    Full Text Available Irregular low-density parity-check (LDPC codes have been found to show exceptionally good performance for single antenna systems over a wide class of channels. In this paper, the performance of LDPC codes with multiple antenna systems is investigated in flat Rayleigh and Rician fading channels for different modulation schemes. The focus of attention is mainly on the concatenation of irregular LDPC codes with complex orthogonal space-time codes. Iterative decoding is carried out with a density evolution method that sets a threshold above which the code performs well. For the proposed concatenated system, the simulation results show that the QAM technique achieves a higher coding gain of 8.8 dB and 3.2 dB over the QPSK technique in Rician (LOS and Rayleigh (NLOS faded environments respectively.

  10. Interleaved Product LDPC Codes

    OpenAIRE

    Baldi, Marco; Cancellieri, Giovanni; Chiaraluce, Franco

    2011-01-01

    Product LDPC codes take advantage of LDPC decoding algorithms and the high minimum distance of product codes. We propose to add suitable interleavers to improve the waterfall performance of LDPC decoding. Interleaving also reduces the number of low weight codewords, that gives a further advantage in the error floor region.

  11. A novel construction method of QC-LDPC codes based on CRT for optical communications

    Science.gov (United States)

    Yuan, Jian-guo; Liang, Meng-qi; Wang, Yong; Lin, Jin-zhao; Pang, Yu

    2016-05-01

    A novel construction method of quasi-cyclic low-density parity-check (QC-LDPC) codes is proposed based on Chinese remainder theory (CRT). The method can not only increase the code length without reducing the girth, but also greatly enhance the code rate, so it is easy to construct a high-rate code. The simulation results show that at the bit error rate ( BER) of 10-7, the net coding gain ( NCG) of the regular QC-LDPC(4 851, 4 546) code is respectively 2.06 dB, 1.36 dB, 0.53 dB and 0.31 dB more than those of the classic RS(255, 239) code in ITU-T G.975, the LDPC(32 640, 30 592) code in ITU-T G.975.1, the QC-LDPC(3 664, 3 436) code constructed by the improved combining construction method based on CRT and the irregular QC-LDPC(3 843, 3 603) code constructed by the construction method based on the Galois field ( GF( q)) multiplicative group. Furthermore, all these five codes have the same code rate of 0.937. Therefore, the regular QC-LDPC(4 851, 4 546) code constructed by the proposed construction method has excellent error-correction performance, and can be more suitable for optical transmission systems.

  12. Discussion on LDPC Codes and Uplink Coding

    Science.gov (United States)

    Andrews, Ken; Divsalar, Dariush; Dolinar, Sam; Moision, Bruce; Hamkins, Jon; Pollara, Fabrizio

    2007-01-01

    This slide presentation reviews the progress that the workgroup on Low-Density Parity-Check (LDPC) for space link coding. The workgroup is tasked with developing and recommending new error correcting codes for near-Earth, Lunar, and deep space applications. Included in the presentation is a summary of the technical progress of the workgroup. Charts that show the LDPC decoder sensitivity to symbol scaling errors are reviewed, as well as a chart showing the performance of several frame synchronizer algorithms compared to that of some good codes and LDPC decoder tests at ESTL. Also reviewed is a study on Coding, Modulation, and Link Protocol (CMLP), and the recommended codes. A design for the Pseudo-Randomizer with LDPC Decoder and CRC is also reviewed. A chart that summarizes the three proposed coding systems is also presented.

  13. QC-LDPC code-based cryptography

    CERN Document Server

    Baldi, Marco

    2014-01-01

    This book describes the fundamentals of cryptographic primitives based on quasi-cyclic low-density parity-check (QC-LDPC) codes, with a special focus on the use of these codes in public-key cryptosystems derived from the McEliece and Niederreiter schemes. In the first part of the book, the main characteristics of QC-LDPC codes are reviewed, and several techniques for their design are presented, while tools for assessing the error correction performance of these codes are also described. Some families of QC-LDPC codes that are best suited for use in cryptography are also presented. The second part of the book focuses on the McEliece and Niederreiter cryptosystems, both in their original forms and in some subsequent variants. The applicability of QC-LDPC codes in these frameworks is investigated by means of theoretical analyses and numerical tools, in order to assess their benefits and drawbacks in terms of system efficiency and security. Several examples of QC-LDPC code-based public key cryptosystems are prese...

  14. Constructing LDPC Codes from Loop-Free Encoding Modules

    Science.gov (United States)

    Divsalar, Dariush; Dolinar, Samuel; Jones, Christopher; Thorpe, Jeremy; Andrews, Kenneth

    2009-01-01

    channel capacity limits can be achieved for the codes of the type in question having low maximum variable node degrees. The decoding thresholds in these examples are lower than those of the best-known unstructured irregular LDPC codes constrained to have the same maximum node degrees. Furthermore, the present method enables the construction of codes of any desired rate with thresholds that stay uniformly close to their respective channel capacity thresholds.

  15. Analysis of Non-binary Hybrid LDPC Codes

    OpenAIRE

    Sassatelli, Lucile; Declercq, David

    2008-01-01

    In this paper, we analyse asymptotically a new class of LDPC codes called Non-binary Hybrid LDPC codes, which has been recently introduced. We use density evolution techniques to derive a stability condition for hybrid LDPC codes, and prove their threshold behavior. We study this stability condition to conclude on asymptotic advantages of hybrid LDPC codes compared to their non-hybrid counterparts.

  16. LDPC Codes with Minimum Distance Proportional to Block Size

    Science.gov (United States)

    Divsalar, Dariush; Jones, Christopher; Dolinar, Samuel; Thorpe, Jeremy

    2009-01-01

    Low-density parity-check (LDPC) codes characterized by minimum Hamming distances proportional to block sizes have been demonstrated. Like the codes mentioned in the immediately preceding article, the present codes are error-correcting codes suitable for use in a variety of wireless data-communication systems that include noisy channels. The previously mentioned codes have low decoding thresholds and reasonably low error floors. However, the minimum Hamming distances of those codes do not grow linearly with code-block sizes. Codes that have this minimum-distance property exhibit very low error floors. Examples of such codes include regular LDPC codes with variable degrees of at least 3. Unfortunately, the decoding thresholds of regular LDPC codes are high. Hence, there is a need for LDPC codes characterized by both low decoding thresholds and, in order to obtain acceptably low error floors, minimum Hamming distances that are proportional to code-block sizes. The present codes were developed to satisfy this need. The minimum Hamming distances of the present codes have been shown, through consideration of ensemble-average weight enumerators, to be proportional to code block sizes. As in the cases of irregular ensembles, the properties of these codes are sensitive to the proportion of degree-2 variable nodes. A code having too few such nodes tends to have an iterative decoding threshold that is far from the capacity threshold. A code having too many such nodes tends not to exhibit a minimum distance that is proportional to block size. Results of computational simulations have shown that the decoding thresholds of codes of the present type are lower than those of regular LDPC codes. Included in the simulations were a few examples from a family of codes characterized by rates ranging from low to high and by thresholds that adhere closely to their respective channel capacity thresholds; the simulation results from these examples showed that the codes in question have low

  17. Protograph LDPC Codes with Node Degrees at Least 3

    Science.gov (United States)

    Divsalar, Dariush; Jones, Christopher

    2006-01-01

    In this paper we present protograph codes with a small number of degree-3 nodes and one high degree node. The iterative decoding threshold for proposed rate 1/2 codes are lower, by about 0.2 dB, than the best known irregular LDPC codes with degree at least 3. The main motivation is to gain linear minimum distance to achieve low error floor. Also to construct rate-compatible protograph-based LDPC codes for fixed block length that simultaneously achieves low iterative decoding threshold and linear minimum distance. We start with a rate 1/2 protograph LDPC code with degree-3 nodes and one high degree node. Higher rate codes are obtained by connecting check nodes with degree-2 non-transmitted nodes. This is equivalent to constraint combining in the protograph. The condition where all constraints are combined corresponds to the highest rate code. This constraint must be connected to nodes of degree at least three for the graph to have linear minimum distance. Thus having node degree at least 3 for rate 1/2 guarantees linear minimum distance property to be preserved for higher rates. Through examples we show that the iterative decoding threshold as low as 0.544 dB can be achieved for small protographs with node degrees at least three. A family of low- to high-rate codes with minimum distance linearly increasing in block size and with capacity-approaching performance thresholds is presented. FPGA simulation results for a few example codes show that the proposed codes perform as predicted.

  18. Protograph based LDPC codes with minimum distance linearly growing with block size

    Science.gov (United States)

    Divsalar, Dariush; Jones, Christopher; Dolinar, Sam; Thorpe, Jeremy

    2005-01-01

    We propose several LDPC code constructions that simultaneously achieve good threshold and error floor performance. Minimum distance is shown to grow linearly with block size (similar to regular codes of variable degree at least 3) by considering ensemble average weight enumerators. Our constructions are based on projected graph, or protograph, structures that support high-speed decoder implementations. As with irregular ensembles, our constructions are sensitive to the proportion of degree-2 variable nodes. A code with too few such nodes tends to have an iterative decoding threshold that is far from the capacity threshold. A code with too many such nodes tends to not exhibit a minimum distance that grows linearly in block length. In this paper we also show that precoding can be used to lower the threshold of regular LDPC codes. The decoding thresholds of the proposed codes, which have linearly increasing minimum distance in block size, outperform that of regular LDPC codes. Furthermore, a family of low to high rate codes, with thresholds that adhere closely to their respective channel capacity thresholds, is presented. Simulation results for a few example codes show that the proposed codes have low error floors as well as good threshold SNFt performance.

  19. Joint design of QC-LDPC codes for coded cooperation system with joint iterative decoding

    Science.gov (United States)

    Zhang, Shunwai; Yang, Fengfan; Tang, Lei; Ejaz, Saqib; Luo, Lin; Maharaj, B. T.

    2016-03-01

    In this paper, we investigate joint design of quasi-cyclic low-density-parity-check (QC-LDPC) codes for coded cooperation system with joint iterative decoding in the destination. First, QC-LDPC codes based on the base matrix and exponent matrix are introduced, and then we describe two types of girth-4 cycles in QC-LDPC codes employed by the source and relay. In the equivalent parity-check matrix corresponding to the jointly designed QC-LDPC codes employed by the source and relay, all girth-4 cycles including both type I and type II are cancelled. Theoretical analysis and numerical simulations show that the jointly designed QC-LDPC coded cooperation well combines cooperation gain and channel coding gain, and outperforms the coded non-cooperation under the same conditions. Furthermore, the bit error rate performance of the coded cooperation employing jointly designed QC-LDPC codes is better than those of random LDPC codes and separately designed QC-LDPC codes over AWGN channels.

  20. LDPC coded OFDM over the atmospheric turbulence channel.

    Science.gov (United States)

    Djordjevic, Ivan B; Vasic, Bane; Neifeld, Mark A

    2007-05-14

    Low-density parity-check (LDPC) coded optical orthogonal frequency division multiplexing (OFDM) is shown to significantly outperform LDPC coded on-off keying (OOK) over the atmospheric turbulence channel in terms of both coding gain and spectral efficiency. In the regime of strong turbulence at a bit-error rate of 10(-5), the coding gain improvement of the LDPC coded single-side band unclipped-OFDM system with 64 sub-carriers is larger than the coding gain of the LDPC coded OOK system by 20.2 dB for quadrature-phase-shift keying (QPSK) and by 23.4 dB for binary-phase-shift keying (BPSK).

  1. Non-binary Hybrid LDPC Codes: Structure, Decoding and Optimization

    OpenAIRE

    Sassatelli, Lucile; Declercq, David

    2007-01-01

    In this paper, we propose to study and optimize a very general class of LDPC codes whose variable nodes belong to finite sets with different orders. We named this class of codes Hybrid LDPC codes. Although efficient optimization techniques exist for binary LDPC codes and more recently for non-binary LDPC codes, they both exhibit drawbacks due to different reasons. Our goal is to capitalize on the advantages of both families by building codes with binary (or small finite set order) and non-bin...

  2. Non-Binary Protograph-Based LDPC Codes: Analysis,Enumerators and Designs

    OpenAIRE

    Sun, Yizeng

    2013-01-01

    Non-binary LDPC codes can outperform binary LDPC codes using sum-product algorithm with higher computation complexity. Non-binary LDPC codes based on protographs have the advantage of simple hardware architecture. In the first part of this thesis, we will use EXIT chart analysis to compute the thresholds of different protographs over GF(q). Based on threshold computation, some non-binary protograph-based LDPC codes are designed and their frame error rates are compared with binary LDPC codes. ...

  3. Optimisation des codes LDPC irréguliers et algorithmes de décodage des codes LDPC q-aires

    OpenAIRE

    Cances , Jean-Pierre

    2013-01-01

    Cette note technique rappelle les principes d'optimisation pour obtenir les profils de codes LDPC irréguliers performants et rappelle les principes des algorithmes de décodage utilizes pour les codes LDPC q-aires à grande efficacité spectrale.

  4. Decoding LDPC Convolutional Codes on Markov Channels

    Directory of Open Access Journals (Sweden)

    Kashyap Manohar

    2008-01-01

    Full Text Available Abstract This paper describes a pipelined iterative technique for joint decoding and channel state estimation of LDPC convolutional codes over Markov channels. Example designs are presented for the Gilbert-Elliott discrete channel model. We also compare the performance and complexity of our algorithm against joint decoding and state estimation of conventional LDPC block codes. Complexity analysis reveals that our pipelined algorithm reduces the number of operations per time step compared to LDPC block codes, at the expense of increased memory and latency. This tradeoff is favorable for low-power applications.

  5. Decoding LDPC Convolutional Codes on Markov Channels

    Directory of Open Access Journals (Sweden)

    Chris Winstead

    2008-04-01

    Full Text Available This paper describes a pipelined iterative technique for joint decoding and channel state estimation of LDPC convolutional codes over Markov channels. Example designs are presented for the Gilbert-Elliott discrete channel model. We also compare the performance and complexity of our algorithm against joint decoding and state estimation of conventional LDPC block codes. Complexity analysis reveals that our pipelined algorithm reduces the number of operations per time step compared to LDPC block codes, at the expense of increased memory and latency. This tradeoff is favorable for low-power applications.

  6. Multilevel LDPC Codes Design for Multimedia Communication CDMA System

    Directory of Open Access Journals (Sweden)

    Hou Jia

    2004-01-01

    Full Text Available We design multilevel coding (MLC with a semi-bit interleaved coded modulation (BICM scheme based on low density parity check (LDPC codes. Different from the traditional designs, we joined the MLC and BICM together by using the Gray mapping, which is suitable to transmit the data over several equivalent channels with different code rates. To perform well at signal-to-noise ratio (SNR to be very close to the capacity of the additive white Gaussian noise (AWGN channel, random regular LDPC code and a simple semialgebra LDPC (SA-LDPC code are discussed in MLC with parallel independent decoding (PID. The numerical results demonstrate that the proposed scheme could achieve both power and bandwidth efficiency.

  7. Construction of type-II QC-LDPC codes with fast encoding based on perfect cyclic difference sets

    Science.gov (United States)

    Li, Ling-xiang; Li, Hai-bing; Li, Ji-bi; Jiang, Hua

    2017-09-01

    In view of the problems that the encoding complexity of quasi-cyclic low-density parity-check (QC-LDPC) codes is high and the minimum distance is not large enough which leads to the degradation of the error-correction performance, the new irregular type-II QC-LDPC codes based on perfect cyclic difference sets (CDSs) are constructed. The parity check matrices of these type-II QC-LDPC codes consist of the zero matrices with weight of 0, the circulant permutation matrices (CPMs) with weight of 1 and the circulant matrices with weight of 2 (W2CMs). The introduction of W2CMs in parity check matrices makes it possible to achieve the larger minimum distance which can improve the error- correction performance of the codes. The Tanner graphs of these codes have no girth-4, thus they have the excellent decoding convergence characteristics. In addition, because the parity check matrices have the quasi-dual diagonal structure, the fast encoding algorithm can reduce the encoding complexity effectively. Simulation results show that the new type-II QC-LDPC codes can achieve a more excellent error-correction performance and have no error floor phenomenon over the additive white Gaussian noise (AWGN) channel with sum-product algorithm (SPA) iterative decoding.

  8. Protograph LDPC Codes for the Erasure Channel

    Science.gov (United States)

    Pollara, Fabrizio; Dolinar, Samuel J.; Divsalar, Dariush

    2006-01-01

    This viewgraph presentation reviews the use of protograph Low Density Parity Check (LDPC) codes for erasure channels. A protograph is a Tanner graph with a relatively small number of nodes. A "copy-and-permute" operation can be applied to the protograph to obtain larger derived graphs of various sizes. For very high code rates and short block sizes, a low asymptotic threshold criterion is not the best approach to designing LDPC codes. Simple protographs with much regularity and low maximum node degrees appear to be the best choices Quantized-rateless protograph LDPC codes can be built by careful design of the protograph such that multiple puncturing patterns will still permit message passing decoding to proceed

  9. On Analyzing LDPC Codes over Multiantenna MC-CDMA System

    Directory of Open Access Journals (Sweden)

    S. Suresh Kumar

    2014-01-01

    Full Text Available Multiantenna multicarrier code-division multiple access (MC-CDMA technique has been attracting much attention for designing future broadband wireless systems. In addition, low-density parity-check (LDPC code, a promising near-optimal error correction code, is also being widely considered in next generation communication systems. In this paper, we propose a simple method to construct a regular quasicyclic low-density parity-check (QC-LDPC code to improve the transmission performance over the precoded MC-CDMA system with limited feedback. Simulation results show that the coding gain of the proposed QC-LDPC codes is larger than that of the Reed-Solomon codes, and the performance of the multiantenna MC-CDMA system can be greatly improved by these QC-LDPC codes when the data rate is high.

  10. Encoding of QC-LDPC Codes of Rank Deficient Parity Matrix

    Directory of Open Access Journals (Sweden)

    Mohammed Kasim Mohammed Al-Haddad

    2016-05-01

    Full Text Available the encoding of long low density parity check (LDPC codes presents a challenge compared to its decoding. The Quasi Cyclic (QC LDPC codes offer the advantage for reducing the complexity for both encoding and decoding due to its QC structure. Most QC-LDPC codes have rank deficient parity matrix and this introduces extra complexity over the codes with full rank parity matrix. In this paper an encoding scheme of QC-LDPC codes is presented that is suitable for codes with full rank parity matrix and rank deficient parity matrx. The extra effort required by the codes with rank deficient parity matrix over the codes of full rank parity matrix is investigated.

  11. Multiple LDPC decoding for distributed source coding and video coding

    DEFF Research Database (Denmark)

    Forchhammer, Søren; Luong, Huynh Van; Huang, Xin

    2011-01-01

    Distributed source coding (DSC) is a coding paradigm for systems which fully or partly exploit the source statistics at the decoder to reduce the computational burden at the encoder. Distributed video coding (DVC) is one example. This paper considers the use of Low Density Parity Check Accumulate...... (LDPCA) codes in a DSC scheme with feed-back. To improve the LDPC coding performance in the context of DSC and DVC, while retaining short encoder blocks, this paper proposes multiple parallel LDPC decoding. The proposed scheme passes soft information between decoders to enhance performance. Experimental...

  12. Differentially Encoded LDPC Codes—Part II: General Case and Code Optimization

    Directory of Open Access Journals (Sweden)

    Li (Tiffany Jing

    2008-01-01

    Full Text Available This two-part series of papers studies the theory and practice of differentially encoded low-density parity-check (DE-LDPC codes, especially in the context of noncoherent detection. Part I showed that a special class of DE-LDPC codes, product accumulate codes, perform very well with both coherent and noncoherent detections. The analysis here reveals that a conventional LDPC code, however, is not fitful for differential coding and does not, in general, deliver a desirable performance when detected noncoherently. Through extrinsic information transfer (EXIT analysis and a modified "convergence-constraint" density evolution (DE method developed here, we provide a characterization of the type of LDPC degree profiles that work in harmony with differential detection (or a recursive inner code in general, and demonstrate how to optimize these LDPC codes. The convergence-constraint method provides a useful extension to the conventional "threshold-constraint" method, and can match an outer LDPC code to any given inner code with the imperfectness of the inner decoder taken into consideration.

  13. Performance Analysis for Cooperative Communication System with QC-LDPC Codes Constructed with Integer Sequences

    Directory of Open Access Journals (Sweden)

    Yan Zhang

    2015-01-01

    Full Text Available This paper presents four different integer sequences to construct quasi-cyclic low-density parity-check (QC-LDPC codes with mathematical theory. The paper introduces the procedure of the coding principle and coding. Four different integer sequences constructing QC-LDPC code are compared with LDPC codes by using PEG algorithm, array codes, and the Mackey codes, respectively. Then, the integer sequence QC-LDPC codes are used in coded cooperative communication. Simulation results show that the integer sequence constructed QC-LDPC codes are effective, and overall performance is better than that of other types of LDPC codes in the coded cooperative communication. The performance of Dayan integer sequence constructed QC-LDPC is the most excellent performance.

  14. Design of ACM system based on non-greedy punctured LDPC codes

    Science.gov (United States)

    Lu, Zijun; Jiang, Zihong; Zhou, Lin; He, Yucheng

    2017-08-01

    In this paper, an adaptive coded modulation (ACM) scheme based on rate-compatible LDPC (RC-LDPC) codes was designed. The RC-LDPC codes were constructed by a non-greedy puncturing method which showed good performance in high code rate region. Moreover, the incremental redundancy scheme of LDPC-based ACM system over AWGN channel was proposed. By this scheme, code rates vary from 2/3 to 5/6 and the complication of the ACM system is lowered. Simulations show that more and more obvious coding gain can be obtained by the proposed ACM system with higher throughput.

  15. Spatially coupled LDPC coding in cooperative wireless networks

    NARCIS (Netherlands)

    Jayakody, D.N.K.; Skachek, V.; Chen, B.

    2016-01-01

    This paper proposes a novel technique of spatially coupled low-density parity-check (SC-LDPC) code-based soft forwarding relaying scheme for a two-way relay system. We introduce an array-based optimized SC-LDPC codes in relay channels. A more precise model is proposed to characterize the residual

  16. Construction of Quasi-Cyclic LDPC Codes Based on Fundamental Theorem of Arithmetic

    Directory of Open Access Journals (Sweden)

    Hai Zhu

    2018-01-01

    Full Text Available Quasi-cyclic (QC LDPC codes play an important role in 5G communications and have been chosen as the standard codes for 5G enhanced mobile broadband (eMBB data channel. In this paper, we study the construction of QC LDPC codes based on an arbitrary given expansion factor (or lifting degree. First, we analyze the cycle structure of QC LDPC codes and give the necessary and sufficient condition for the existence of short cycles. Based on the fundamental theorem of arithmetic in number theory, we divide the integer factorization into three cases and present three classes of QC LDPC codes accordingly. Furthermore, a general construction method of QC LDPC codes with girth of at least 6 is proposed. Numerical results show that the constructed QC LDPC codes perform well over the AWGN channel when decoded with the iterative algorithms.

  17. DNA Barcoding through Quaternary LDPC Codes.

    Science.gov (United States)

    Tapia, Elizabeth; Spetale, Flavio; Krsticevic, Flavia; Angelone, Laura; Bulacio, Pilar

    2015-01-01

    For many parallel applications of Next-Generation Sequencing (NGS) technologies short barcodes able to accurately multiplex a large number of samples are demanded. To address these competitive requirements, the use of error-correcting codes is advised. Current barcoding systems are mostly built from short random error-correcting codes, a feature that strongly limits their multiplexing accuracy and experimental scalability. To overcome these problems on sequencing systems impaired by mismatch errors, the alternative use of binary BCH and pseudo-quaternary Hamming codes has been proposed. However, these codes either fail to provide a fine-scale with regard to size of barcodes (BCH) or have intrinsic poor error correcting abilities (Hamming). Here, the design of barcodes from shortened binary BCH codes and quaternary Low Density Parity Check (LDPC) codes is introduced. Simulation results show that although accurate barcoding systems of high multiplexing capacity can be obtained with any of these codes, using quaternary LDPC codes may be particularly advantageous due to the lower rates of read losses and undetected sample misidentification errors. Even at mismatch error rates of 10(-2) per base, 24-nt LDPC barcodes can be used to multiplex roughly 2000 samples with a sample misidentification error rate in the order of 10(-9) at the expense of a rate of read losses just in the order of 10(-6).

  18. DNA Barcoding through Quaternary LDPC Codes.

    Directory of Open Access Journals (Sweden)

    Elizabeth Tapia

    Full Text Available For many parallel applications of Next-Generation Sequencing (NGS technologies short barcodes able to accurately multiplex a large number of samples are demanded. To address these competitive requirements, the use of error-correcting codes is advised. Current barcoding systems are mostly built from short random error-correcting codes, a feature that strongly limits their multiplexing accuracy and experimental scalability. To overcome these problems on sequencing systems impaired by mismatch errors, the alternative use of binary BCH and pseudo-quaternary Hamming codes has been proposed. However, these codes either fail to provide a fine-scale with regard to size of barcodes (BCH or have intrinsic poor error correcting abilities (Hamming. Here, the design of barcodes from shortened binary BCH codes and quaternary Low Density Parity Check (LDPC codes is introduced. Simulation results show that although accurate barcoding systems of high multiplexing capacity can be obtained with any of these codes, using quaternary LDPC codes may be particularly advantageous due to the lower rates of read losses and undetected sample misidentification errors. Even at mismatch error rates of 10(-2 per base, 24-nt LDPC barcodes can be used to multiplex roughly 2000 samples with a sample misidentification error rate in the order of 10(-9 at the expense of a rate of read losses just in the order of 10(-6.

  19. Codeword Structure Analysis for LDPC Convolutional Codes

    Directory of Open Access Journals (Sweden)

    Hua Zhou

    2015-12-01

    Full Text Available The codewords of a low-density parity-check (LDPC convolutional code (LDPC-CC are characterised into structured and non-structured. The number of the structured codewords is dominated by the size of the polynomial syndrome former matrix H T ( D , while the number of the non-structured ones depends on the particular monomials or polynomials in H T ( D . By evaluating the relationship of the codewords between the mother code and its super codes, the low weight non-structured codewords in the super codes can be eliminated by appropriately choosing the monomials or polynomials in H T ( D , resulting in improved distance spectrum of the mother code.

  20. Protograph LDPC Codes Over Burst Erasure Channels

    Science.gov (United States)

    Divsalar, Dariush; Dolinar, Sam; Jones, Christopher

    2006-01-01

    In this paper we design high rate protograph based LDPC codes suitable for binary erasure channels. To simplify the encoder and decoder implementation for high data rate transmission, the structure of codes are based on protographs and circulants. These LDPC codes can improve data link and network layer protocols in support of communication networks. Two classes of codes were designed. One class is designed for large block sizes with an iterative decoding threshold that approaches capacity of binary erasure channels. The other class is designed for short block sizes based on maximizing minimum stopping set size. For high code rates and short blocks the second class outperforms the first class.

  1. Error floor behavior study of LDPC codes for concatenated codes design

    Science.gov (United States)

    Chen, Weigang; Yin, Liuguo; Lu, Jianhua

    2007-11-01

    Error floor behavior of low-density parity-check (LDPC) codes using quantized decoding algorithms is statistically studied with experimental results on a hardware evaluation platform. The results present the distribution of the residual errors after decoding failure and reveal that the number of residual error bits in a codeword is usually very small using quantized sum-product (SP) algorithm. Therefore, LDPC code may serve as the inner code in a concatenated coding system with a high code rate outer code and thus an ultra low error floor can be achieved. This conclusion is also verified by the experimental results.

  2. Short-Block Protograph-Based LDPC Codes

    Science.gov (United States)

    Divsalar, Dariush; Dolinar, Samuel; Jones, Christopher

    2010-01-01

    Short-block low-density parity-check (LDPC) codes of a special type are intended to be especially well suited for potential applications that include transmission of command and control data, cellular telephony, data communications in wireless local area networks, and satellite data communications. [In general, LDPC codes belong to a class of error-correcting codes suitable for use in a variety of wireless data-communication systems that include noisy channels.] The codes of the present special type exhibit low error floors, low bit and frame error rates, and low latency (in comparison with related prior codes). These codes also achieve low maximum rate of undetected errors over all signal-to-noise ratios, without requiring the use of cyclic redundancy checks, which would significantly increase the overhead for short blocks. These codes have protograph representations; this is advantageous in that, for reasons that exceed the scope of this article, the applicability of protograph representations makes it possible to design highspeed iterative decoders that utilize belief- propagation algorithms.

  3. Differentially Encoded LDPC Codes—Part II: General Case and Code Optimization

    Directory of Open Access Journals (Sweden)

    Jing Li (Tiffany

    2008-04-01

    Full Text Available This two-part series of papers studies the theory and practice of differentially encoded low-density parity-check (DE-LDPC codes, especially in the context of noncoherent detection. Part I showed that a special class of DE-LDPC codes, product accumulate codes, perform very well with both coherent and noncoherent detections. The analysis here reveals that a conventional LDPC code, however, is not fitful for differential coding and does not, in general, deliver a desirable performance when detected noncoherently. Through extrinsic information transfer (EXIT analysis and a modified “convergence-constraint” density evolution (DE method developed here, we provide a characterization of the type of LDPC degree profiles that work in harmony with differential detection (or a recursive inner code in general, and demonstrate how to optimize these LDPC codes. The convergence-constraint method provides a useful extension to the conventional “threshold-constraint” method, and can match an outer LDPC code to any given inner code with the imperfectness of the inner decoder taken into consideration.

  4. LDPC-PPM Coding Scheme for Optical Communication

    Science.gov (United States)

    Barsoum, Maged; Moision, Bruce; Divsalar, Dariush; Fitz, Michael

    2009-01-01

    In a proposed coding-and-modulation/demodulation-and-decoding scheme for a free-space optical communication system, an error-correcting code of the low-density parity-check (LDPC) type would be concatenated with a modulation code that consists of a mapping of bits to pulse-position-modulation (PPM) symbols. Hence, the scheme is denoted LDPC-PPM. This scheme could be considered a competitor of a related prior scheme in which an outer convolutional error-correcting code is concatenated with an interleaving operation, a bit-accumulation operation, and a PPM inner code. Both the prior and present schemes can be characterized as serially concatenated pulse-position modulation (SCPPM) coding schemes. Figure 1 represents a free-space optical communication system based on either the present LDPC-PPM scheme or the prior SCPPM scheme. At the transmitting terminal, the original data (u) are processed by an encoder into blocks of bits (a), and the encoded data are mapped to PPM of an optical signal (c). For the purpose of design and analysis, the optical channel in which the PPM signal propagates is modeled as a Poisson point process. At the receiving terminal, the arriving optical signal (y) is demodulated to obtain an estimate (a^) of the coded data, which is then processed by a decoder to obtain an estimate (u^) of the original data.

  5. LDPC-coded orbital angular momentum (OAM) modulation for free-space optical communication.

    Science.gov (United States)

    Djordjevic, Ivan B; Arabaci, Murat

    2010-11-22

    An orbital angular momentum (OAM) based LDPC-coded modulation scheme suitable for use in FSO communication is proposed. We demonstrate that the proposed scheme can operate under strong atmospheric turbulence regime and enable 100 Gb/s optical transmission while employing 10 Gb/s components. Both binary and nonbinary LDPC-coded OAM modulations are studied. In addition to providing better BER performance, the nonbinary LDPC-coded modulation reduces overall decoder complexity and latency. The nonbinary LDPC-coded OAM modulation provides a net coding gain of 9.3 dB at the BER of 10(-8). The maximum-ratio combining scheme outperforms the corresponding equal-gain combining scheme by almost 2.5 dB.

  6. A good performance watermarking LDPC code used in high-speed optical fiber communication system

    Science.gov (United States)

    Zhang, Wenbo; Li, Chao; Zhang, Xiaoguang; Xi, Lixia; Tang, Xianfeng; He, Wenxue

    2015-07-01

    A watermarking LDPC code, which is a strategy designed to improve the performance of the traditional LDPC code, was introduced. By inserting some pre-defined watermarking bits into original LDPC code, we can obtain a more correct estimation about the noise level in the fiber channel. Then we use them to modify the probability distribution function (PDF) used in the initial process of belief propagation (BP) decoding algorithm. This algorithm was tested in a 128 Gb/s PDM-DQPSK optical communication system and results showed that the watermarking LDPC code had a better tolerances to polarization mode dispersion (PMD) and nonlinearity than that of traditional LDPC code. Also, by losing about 2.4% of redundancy for watermarking bits, the decoding efficiency of the watermarking LDPC code is about twice of the traditional one.

  7. Construction of Short-Length High-Rates LDPC Codes Using Difference Families

    Directory of Open Access Journals (Sweden)

    Deny Hamdani

    2010-10-01

    Full Text Available Low-density parity-check (LDPC code is linear-block error-correcting code defined by sparse parity-check matrix. It is decoded using the massage-passing algorithm, and in many cases, capable of outperforming turbo code. This paper presents a class of low-density parity-check (LDPC codes showing good performance with low encoding complexity. The code is constructed using difference families from  combinatorial design. The resulting code, which is designed to have short code length and high code rate, can be encoded with low complexity due to its quasi-cyclic structure, and performs well when it is iteratively decoded with the sum-product algorithm. These properties of LDPC code are quite suitable for applications in future wireless local area network.

  8. Performance analysis of WS-EWC coded optical CDMA networks with/without LDPC codes

    Science.gov (United States)

    Huang, Chun-Ming; Huang, Jen-Fa; Yang, Chao-Chin

    2010-10-01

    One extended Welch-Costas (EWC) code family for the wavelength-division-multiplexing/spectral-amplitude coding (WDM/SAC; WS) optical code-division multiple-access (OCDMA) networks is proposed. This system has a superior performance as compared to the previous modified quadratic congruence (MQC) coded OCDMA networks. However, since the performance of such a network is unsatisfactory when the data bit rate is higher, one class of quasi-cyclic low-density parity-check (QC-LDPC) code is adopted to improve that. Simulation results show that the performance of the high-speed WS-EWC coded OCDMA network can be greatly improved by using the LDPC codes.

  9. Construction of Protograph LDPC Codes with Linear Minimum Distance

    Science.gov (United States)

    Divsalar, Dariush; Dolinar, Sam; Jones, Christopher

    2006-01-01

    A construction method for protograph-based LDPC codes that simultaneously achieve low iterative decoding threshold and linear minimum distance is proposed. We start with a high-rate protograph LDPC code with variable node degrees of at least 3. Lower rate codes are obtained by splitting check nodes and connecting them by degree-2 nodes. This guarantees the linear minimum distance property for the lower-rate codes. Excluding checks connected to degree-1 nodes, we show that the number of degree-2 nodes should be at most one less than the number of checks for the protograph LDPC code to have linear minimum distance. Iterative decoding thresholds are obtained by using the reciprocal channel approximation. Thresholds are lowered by using either precoding or at least one very high-degree node in the base protograph. A family of high- to low-rate codes with minimum distance linearly increasing in block size and with capacity-approaching performance thresholds is presented. FPGA simulation results for a few example codes show that the proposed codes perform as predicted.

  10. Experimental demonstration of nonbinary LDPC convolutional codes for DP-64QAM/256QAM

    NARCIS (Netherlands)

    Koike-Akino, T.; Sugihara, K.; Millar, D.S.; Pajovic, M.; Matsumoto, W.; Alvarado, A.; Maher, R.; Lavery, D.; Paskov, M.; Kojima, K.; Parsons, K.; Thomsen, B.C.; Savory, S.J.; Bayvel, P.

    2016-01-01

    We show the great potential of nonbinary LDPC convolutional codes (NB-LDPC-CC) with low-latency windowed decoding. It is experimentally demonstrated that NB-LDPC-CC can offer a performance improvement of up to 5 dB compared with binary coding.

  11. The application of LDPC code in MIMO-OFDM system

    Science.gov (United States)

    Liu, Ruian; Zeng, Beibei; Chen, Tingting; Liu, Nan; Yin, Ninghao

    2018-03-01

    The combination of MIMO and OFDM technology has become one of the key technologies of the fourth generation mobile communication., which can overcome the frequency selective fading of wireless channel, increase the system capacity and improve the frequency utilization. Error correcting coding introduced into the system can further improve its performance. LDPC (low density parity check) code is a kind of error correcting code which can improve system reliability and anti-interference ability, and the decoding is simple and easy to operate. This paper mainly discusses the application of LDPC code in MIMO-OFDM system.

  12. Power Allocation Optimization: Linear Precoding Adapted to NB-LDPC Coded MIMO Transmission

    Directory of Open Access Journals (Sweden)

    Tarek Chehade

    2015-01-01

    Full Text Available In multiple-input multiple-output (MIMO transmission systems, the channel state information (CSI at the transmitter can be used to add linear precoding to the transmitted signals in order to improve the performance and the reliability of the transmission system. This paper investigates how to properly join precoded closed-loop MIMO systems and nonbinary low density parity check (NB-LDPC. The q elements in the Galois field, GF(q, are directly mapped to q transmit symbol vectors. This allows NB-LDPC codes to perfectly fit with a MIMO precoding scheme, unlike binary LDPC codes. The new transmission model is detailed and studied for several linear precoders and various designed LDPC codes. We show that NB-LDPC codes are particularly well suited to be jointly used with precoding schemes based on the maximization of the minimum Euclidean distance (max-dmin criterion. These results are theoretically supported by extrinsic information transfer (EXIT analysis and are confirmed by numerical simulations.

  13. Construction of Short-length High-rates Ldpc Codes Using Difference Families

    OpenAIRE

    Deny Hamdani; Ery Safrianti

    2007-01-01

    Low-density parity-check (LDPC) code is linear-block error-correcting code defined by sparse parity-check matrix. It isdecoded using the massage-passing algorithm, and in many cases, capable of outperforming turbo code. This paperpresents a class of low-density parity-check (LDPC) codes showing good performance with low encoding complexity.The code is constructed using difference families from combinatorial design. The resulting code, which is designed tohave short code length and high code r...

  14. Improved Design of Unequal Error Protection LDPC Codes

    Directory of Open Access Journals (Sweden)

    Sandberg Sara

    2010-01-01

    Full Text Available We propose an improved method for designing unequal error protection (UEP low-density parity-check (LDPC codes. The method is based on density evolution. The degree distribution with the best UEP properties is found, under the constraint that the threshold should not exceed the threshold of a non-UEP code plus some threshold offset. For different codeword lengths and different construction algorithms, we search for good threshold offsets for the UEP code design. The choice of the threshold offset is based on the average a posteriori variable node mutual information. Simulations reveal the counter intuitive result that the short-to-medium length codes designed with a suitable threshold offset all outperform the corresponding non-UEP codes in terms of average bit-error rate. The proposed codes are also compared to other UEP-LDPC codes found in the literature.

  15. Evaluation of four-dimensional nonbinary LDPC-coded modulation for next-generation long-haul optical transport networks.

    Science.gov (United States)

    Zhang, Yequn; Arabaci, Murat; Djordjevic, Ivan B

    2012-04-09

    Leveraging the advanced coherent optical communication technologies, this paper explores the feasibility of using four-dimensional (4D) nonbinary LDPC-coded modulation (4D-NB-LDPC-CM) schemes for long-haul transmission in future optical transport networks. In contrast to our previous works on 4D-NB-LDPC-CM which considered amplified spontaneous emission (ASE) noise as the dominant impairment, this paper undertakes transmission in a more realistic optical fiber transmission environment, taking into account impairments due to dispersion effects, nonlinear phase noise, Kerr nonlinearities, and stimulated Raman scattering in addition to ASE noise. We first reveal the advantages of using 4D modulation formats in LDPC-coded modulation instead of conventional two-dimensional (2D) modulation formats used with polarization-division multiplexing (PDM). Then we demonstrate that 4D LDPC-coded modulation schemes with nonbinary LDPC component codes significantly outperform not only their conventional PDM-2D counterparts but also the corresponding 4D bit-interleaved LDPC-coded modulation (4D-BI-LDPC-CM) schemes, which employ binary LDPC codes as component codes. We also show that the transmission reach improvement offered by the 4D-NB-LDPC-CM over 4D-BI-LDPC-CM increases as the underlying constellation size and hence the spectral efficiency of transmission increases. Our results suggest that 4D-NB-LDPC-CM can be an excellent candidate for long-haul transmission in next-generation optical networks.

  16. Performance Analysis of Faulty Gallager-B Decoding of QC-LDPC Codes with Applications

    Directory of Open Access Journals (Sweden)

    O. Al Rasheed

    2014-06-01

    Full Text Available In this paper we evaluate the performance of Gallager-B algorithm, used for decoding low-density parity-check (LDPC codes, under unreliable message computation. Our analysis is restricted to LDPC codes constructed from circular matrices (QC-LDPC codes. Using Monte Carlo simulation we investigate the effects of different code parameters on coding system performance, under a binary symmetric communication channel and independent transient faults model. One possible application of the presented analysis in designing memory architecture with unreliable components is considered.

  17. A novel QC-LDPC code based on the finite field multiplicative group for optical communications

    Science.gov (United States)

    Yuan, Jian-guo; Xu, Liang; Tong, Qing-zhen

    2013-09-01

    A novel construction method of quasi-cyclic low-density parity-check (QC-LDPC) code is proposed based on the finite field multiplicative group, which has easier construction, more flexible code-length code-rate adjustment and lower encoding/decoding complexity. Moreover, a regular QC-LDPC(5334,4962) code is constructed. The simulation results show that the constructed QC-LDPC(5334,4962) code can gain better error correction performance under the condition of the additive white Gaussian noise (AWGN) channel with iterative decoding sum-product algorithm (SPA). At the bit error rate (BER) of 10-6, the net coding gain (NCG) of the constructed QC-LDPC(5334,4962) code is 1.8 dB, 0.9 dB and 0.2 dB more than that of the classic RS(255,239) code in ITU-T G.975, the LDPC(32640,30592) code in ITU-T G.975.1 and the SCG-LDPC(3969,3720) code constructed by the random method, respectively. So it is more suitable for optical communication systems.

  18. ANÁLISIS Y ESTUDIO DEL DESEMPEÑO DEL CÓDIGO CHEQUEO DE PARIDAD DE BAJA DENSIDAD IRREGULAR EN UN CANAL DE LÍNEA ELÉCTRICA DE BAJA TENSIÓN ANALYSIS AND STUDY OF THE PERFORMANCE OF THE IRREGULAR LOW DENSITY PARITY CHECK CODE IN LOW VOLTAGE POWERLINE

    Directory of Open Access Journals (Sweden)

    Washington Fernández R

    2009-04-01

    Full Text Available En este trabajo se analiza y estudia el código chequeo de paridad de baja densidad irregular (LDPC, en un canal de línea eléctrica de baja tensión. Se analizan los modelos de ruido existentes para las líneas eléctricas de baja tensión. Se evalúa el desempeño del código LDPC irregular, en un canal de línea eléctrica de baja tensión para diferentes velocidades de transmisión (3, 10, 15 y 30 Mbps, considerando como parámetro de desempeño el BER versus SNR.In this work the irregular Low Density Parity Check code (LDPC is analyzed and studied, is used in a low voltage powerline. Noise models are analyzed for these low voltage lines. The performance of the LDPC code is evaluated in a low voltage line channel for rates of 3, 5, 15 and 30 Mbps, considering of BER vs SNR a parameter.

  19. New Technique for Improving Performance of LDPC Codes in the Presence of Trapping Sets

    Directory of Open Access Journals (Sweden)

    Mohamed Adnan Landolsi

    2008-06-01

    Full Text Available Trapping sets are considered the primary factor for degrading the performance of low-density parity-check (LDPC codes in the error-floor region. The effect of trapping sets on the performance of an LDPC code becomes worse as the code size decreases. One approach to tackle this problem is to minimize trapping sets during LDPC code design. However, while trapping sets can be reduced, their complete elimination is infeasible due to the presence of cycles in the underlying LDPC code bipartite graph. In this work, we introduce a new technique based on trapping sets neutralization to minimize the negative effect of trapping sets under belief propagation (BP decoding. Simulation results for random, progressive edge growth (PEG and MacKay LDPC codes demonstrate the effectiveness of the proposed technique. The hardware cost of the proposed technique is also shown to be minimal.

  20. Implementation of Layered Decoding Architecture for LDPC Code using Layered Min-Sum Algorithm

    OpenAIRE

    Sandeep Kakde; Atish Khobragade; Shrikant Ambatkar; Pranay Nandanwar

    2017-01-01

    For binary field and long code lengths, Low Density Parity Check (LDPC) code approaches Shannon limit performance. LDPC codes provide remarkable error correction performance and therefore enlarge the design space for communication systems.In this paper, we have compare different digital modulation techniques and found that BPSK modulation technique is better than other modulation techniques in terms of BER. It also gives error performance of LDPC decoder over AWGN channel using Min-Sum algori...

  1. High Girth Column-Weight-Two LDPC Codes Based on Distance Graphs

    Directory of Open Access Journals (Sweden)

    Gabofetswe Malema

    2007-01-01

    Full Text Available LDPC codes of column weight of two are constructed from minimal distance graphs or cages. Distance graphs are used to represent LDPC code matrices such that graph vertices that represent rows and edges are columns. The conversion of a distance graph into matrix form produces an adjacency matrix with column weight of two and girth double that of the graph. The number of 1's in each row (row weight is equal to the degree of the corresponding vertex. By constructing graphs with different vertex degrees, we can vary the rate of corresponding LDPC code matrices. Cage graphs are used as examples of distance graphs to design codes with different girths and rates. Performance of obtained codes depends on girth and structure of the corresponding distance graphs.

  2. Cooperative optimization and their application in LDPC codes

    Science.gov (United States)

    Chen, Ke; Rong, Jian; Zhong, Xiaochun

    2008-10-01

    Cooperative optimization is a new way for finding global optima of complicated functions of many variables. The proposed algorithm is a class of message passing algorithms and has solid theory foundations. It can achieve good coding gains over the sum-product algorithm for LDPC codes. For (6561, 4096) LDPC codes, the proposed algorithm can achieve 2.0 dB gains over the sum-product algorithm at BER of 4×10-7. The decoding complexity of the proposed algorithm is lower than the sum-product algorithm can do; furthermore, the former can achieve much lower error floor than the latter can do after the Eb / No is higher than 1.8 dB.

  3. Rate-Compatible Protograph LDPC Codes

    Science.gov (United States)

    Nguyen, Thuy V. (Inventor); Nosratinia, Aria (Inventor); Divsalar, Dariush (Inventor)

    2014-01-01

    Digital communication coding methods resulting in rate-compatible low density parity-check (LDPC) codes built from protographs. Described digital coding methods start with a desired code rate and a selection of the numbers of variable nodes and check nodes to be used in the protograph. Constraints are set to satisfy a linear minimum distance growth property for the protograph. All possible edges in the graph are searched for the minimum iterative decoding threshold and the protograph with the lowest iterative decoding threshold is selected. Protographs designed in this manner are used in decode and forward relay channels.

  4. CONSTRUCTION OF REGULAR LDPC LIKE CODES BASED ON FULL RANK CODES AND THEIR ITERATIVE DECODING USING A PARITY CHECK TREE

    Directory of Open Access Journals (Sweden)

    H. Prashantha Kumar

    2011-09-01

    Full Text Available Low density parity check (LDPC codes are capacity-approaching codes, which means that practical constructions exist that allow the noise threshold to be set very close to the theoretical Shannon limit for a memory less channel. LDPC codes are finding increasing use in applications like LTE-Networks, digital television, high density data storage systems, deep space communication systems etc. Several algebraic and combinatorial methods are available for constructing LDPC codes. In this paper we discuss a novel low complexity algebraic method for constructing regular LDPC like codes derived from full rank codes. We demonstrate that by employing these codes over AWGN channels, coding gains in excess of 2dB over un-coded systems can be realized when soft iterative decoding using a parity check tree is employed.

  5. Evaluation of large girth LDPC codes for PMD compensation by turbo equalization.

    Science.gov (United States)

    Minkov, Lyubomir L; Djordjevic, Ivan B; Xu, Lei; Wang, Ting; Kueppers, Franko

    2008-08-18

    Large-girth quasi-cyclic LDPC codes have been experimentally evaluated for use in PMD compensation by turbo equalization for a 10 Gb/s NRZ optical transmission system, and observing one sample per bit. Net effective coding gain improvement for girth-10, rate 0.906 code of length 11936 over maximum a posteriori probability (MAP) detector for differential group delay of 125 ps is 6.25 dB at BER of 10(-6). Girth-10 LDPC code of rate 0.8 outperforms the girth-10 code of rate 0.906 by 2.75 dB, and provides the net effective coding gain improvement of 9 dB at the same BER. It is experimentally determined that girth-10 LDPC codes of length around 15000 approach channel capacity limit within 1.25 dB.

  6. Mutiple LDPC Decoding using Bitplane Correlation for Transform Domain Wyner-Ziv Video Coding

    DEFF Research Database (Denmark)

    Luong, Huynh Van; Huang, Xin; Forchhammer, Søren

    2011-01-01

    Distributed video coding (DVC) is an emerging video coding paradigm for systems which fully or partly exploit the source statistics at the decoder to reduce the computational burden at the encoder. This paper considers a Low Density Parity Check (LDPC) based Transform Domain Wyner-Ziv (TDWZ) video...... codec. To improve the LDPC coding performance in the context of TDWZ, this paper proposes a Wyner-Ziv video codec using bitplane correlation through multiple parallel LDPC decoding. The proposed scheme utilizes inter bitplane correlation to enhance the bitplane decoding performance. Experimental results...

  7. Construction method of QC-LDPC codes based on multiplicative group of finite field in optical communication

    Science.gov (United States)

    Huang, Sheng; Ao, Xiang; Li, Yuan-yuan; Zhang, Rui

    2016-09-01

    In order to meet the needs of high-speed development of optical communication system, a construction method of quasi-cyclic low-density parity-check (QC-LDPC) codes based on multiplicative group of finite field is proposed. The Tanner graph of parity check matrix of the code constructed by this method has no cycle of length 4, and it can make sure that the obtained code can get a good distance property. Simulation results show that when the bit error rate ( BER) is 10-6, in the same simulation environment, the net coding gain ( NCG) of the proposed QC-LDPC(3 780, 3 540) code with the code rate of 93.7% in this paper is improved by 2.18 dB and 1.6 dB respectively compared with those of the RS(255, 239) code in ITU-T G.975 and the LDPC(3 2640, 3 0592) code in ITU-T G.975.1. In addition, the NCG of the proposed QC-LDPC(3 780, 3 540) code is respectively 0.2 dB and 0.4 dB higher compared with those of the SG-QC-LDPC(3 780, 3 540) code based on the two different subgroups in finite field and the AS-QC-LDPC(3 780, 3 540) code based on the two arbitrary sets of a finite field. Thus, the proposed QC-LDPC(3 780, 3 540) code in this paper can be well applied in optical communication systems.

  8. Performance analysis of LDPC codes on OOK terahertz wireless channels

    International Nuclear Information System (INIS)

    Liu Chun; Wang Chang; Cao Jun-Cheng

    2016-01-01

    Atmospheric absorption, scattering, and scintillation are the major causes to deteriorate the transmission quality of terahertz (THz) wireless communications. An error control coding scheme based on low density parity check (LDPC) codes with soft decision decoding algorithm is proposed to improve the bit-error-rate (BER) performance of an on-off keying (OOK) modulated THz signal through atmospheric channel. The THz wave propagation characteristics and channel model in atmosphere is set up. Numerical simulations validate the great performance of LDPC codes against the atmospheric fading and demonstrate the huge potential in future ultra-high speed beyond Gbps THz communications. (paper)

  9. Simultaneous chromatic dispersion and PMD compensation by using coded-OFDM and girth-10 LDPC codes.

    Science.gov (United States)

    Djordjevic, Ivan B; Xu, Lei; Wang, Ting

    2008-07-07

    Low-density parity-check (LDPC)-coded orthogonal frequency division multiplexing (OFDM) is studied as an efficient coded modulation scheme suitable for simultaneous chromatic dispersion and polarization mode dispersion (PMD) compensation. We show that, for aggregate rate of 10 Gb/s, accumulated dispersion over 6500 km of SMF and differential group delay of 100 ps can be simultaneously compensated with penalty within 1.5 dB (with respect to the back-to-back configuration) when training sequence based channel estimation and girth-10 LDPC codes of rate 0.8 are employed.

  10. Using LDPC Code Constraints to Aid Recovery of Symbol Timing

    Science.gov (United States)

    Jones, Christopher; Villasnor, John; Lee, Dong-U; Vales, Esteban

    2008-01-01

    A method of utilizing information available in the constraints imposed by a low-density parity-check (LDPC) code has been proposed as a means of aiding the recovery of symbol timing in the reception of a binary-phase-shift-keying (BPSK) signal representing such a code in the presence of noise, timing error, and/or Doppler shift between the transmitter and the receiver. This method and the receiver architecture in which it would be implemented belong to a class of timing-recovery methods and corresponding receiver architectures characterized as pilotless in that they do not require transmission and reception of pilot signals. Acquisition and tracking of a signal of the type described above have traditionally been performed upstream of, and independently of, decoding and have typically involved utilization of a phase-locked loop (PLL). However, the LDPC decoding process, which is iterative, provides information that can be fed back to the timing-recovery receiver circuits to improve performance significantly over that attainable in the absence of such feedback. Prior methods of coupling LDPC decoding with timing recovery had focused on the use of output code words produced as the iterations progress. In contrast, in the present method, one exploits the information available from the metrics computed for the constraint nodes of an LDPC code during the decoding process. In addition, the method involves the use of a waveform model that captures, better than do the waveform models of the prior methods, distortions introduced by receiver timing errors and transmitter/ receiver motions. An LDPC code is commonly represented by use of a bipartite graph containing two sets of nodes. In the graph corresponding to an (n,k) code, the n variable nodes correspond to the code word symbols and the n-k constraint nodes represent the constraints that the code places on the variable nodes in order for them to form a valid code word. The decoding procedure involves iterative computation

  11. Iterative decoding of SOVA and LDPC product code for bit-patterned media recoding

    Science.gov (United States)

    Jeong, Seongkwon; Lee, Jaejin

    2018-05-01

    The demand for high-density storage systems has increased due to the exponential growth of data. Bit-patterned media recording (BPMR) is one of the promising technologies to achieve the density of 1Tbit/in2 and higher. To increase the areal density in BPMR, the spacing between islands needs to be reduced, yet this aggravates inter-symbol interference and inter-track interference and degrades the bit error rate performance. In this paper, we propose a decision feedback scheme using low-density parity check (LDPC) product code for BPMR. This scheme can improve the decoding performance using an iterative approach with extrinsic information and log-likelihood ratio value between iterative soft output Viterbi algorithm and LDPC product code. Simulation results show that the proposed LDPC product code can offer 1.8dB and 2.3dB gains over the one LDPC code at the density of 2.5 and 3 Tb/in2, respectively, when bit error rate is 10-6.

  12. Weighted-Bit-Flipping-Based Sequential Scheduling Decoding Algorithms for LDPC Codes

    Directory of Open Access Journals (Sweden)

    Qing Zhu

    2013-01-01

    Full Text Available Low-density parity-check (LDPC codes can be applied in a lot of different scenarios such as video broadcasting and satellite communications. LDPC codes are commonly decoded by an iterative algorithm called belief propagation (BP over the corresponding Tanner graph. The original BP updates all the variable-nodes simultaneously, followed by all the check-nodes simultaneously as well. We propose a sequential scheduling algorithm based on weighted bit-flipping (WBF algorithm for the sake of improving the convergence speed. Notoriously, WBF is a low-complexity and simple algorithm. We combine it with BP to obtain advantages of these two algorithms. Flipping function used in WBF is borrowed to determine the priority of scheduling. Simulation results show that it can provide a good tradeoff between FER performance and computation complexity for short-length LDPC codes.

  13. Experimental research and comparison of LDPC and RS channel coding in ultraviolet communication systems.

    Science.gov (United States)

    Wu, Menglong; Han, Dahai; Zhang, Xiang; Zhang, Feng; Zhang, Min; Yue, Guangxin

    2014-03-10

    We have implemented a modified Low-Density Parity-Check (LDPC) codec algorithm in ultraviolet (UV) communication system. Simulations are conducted with measured parameters to evaluate the LDPC-based UV system performance. Moreover, LDPC (960, 480) and RS (18, 10) are implemented and experimented via a non-line-of-sight (NLOS) UV test bed. The experimental results are in agreement with the simulation and suggest that based on the given power and 10(-3)bit error rate (BER), in comparison with an uncoded system, average communication distance increases 32% with RS code, while 78% with LDPC code.

  14. Bounded-Angle Iterative Decoding of LDPC Codes

    Science.gov (United States)

    Dolinar, Samuel; Andrews, Kenneth; Pollara, Fabrizio; Divsalar, Dariush

    2009-01-01

    Bounded-angle iterative decoding is a modified version of conventional iterative decoding, conceived as a means of reducing undetected-error rates for short low-density parity-check (LDPC) codes. For a given code, bounded-angle iterative decoding can be implemented by means of a simple modification of the decoder algorithm, without redesigning the code. Bounded-angle iterative decoding is based on a representation of received words and code words as vectors in an n-dimensional Euclidean space (where n is an integer).

  15. Experimental study of non-binary LDPC coding for long-haul coherent optical QPSK transmissions.

    Science.gov (United States)

    Zhang, Shaoliang; Arabaci, Murat; Yaman, Fatih; Djordjevic, Ivan B; Xu, Lei; Wang, Ting; Inada, Yoshihisa; Ogata, Takaaki; Aoki, Yasuhiro

    2011-09-26

    The performance of rate-0.8 4-ary LDPC code has been studied in a 50 GHz-spaced 40 Gb/s DWDM system with PDM-QPSK modulation. The net effective coding gain of 10 dB is obtained at BER of 10(-6). With the aid of time-interleaving polarization multiplexing and MAP detection, 10,560 km transmission over legacy dispersion managed fiber is achieved without any countable errors. The proposed nonbinary quasi-cyclic LDPC code achieves an uncoded BER threshold at 4×10(-2). Potential issues like phase ambiguity and coding length are also discussed when implementing LDPC in current coherent optical systems. © 2011 Optical Society of America

  16. Product code optimization for determinate state LDPC decoding in robust image transmission.

    Science.gov (United States)

    Thomos, Nikolaos; Boulgouris, Nikolaos V; Strintzis, Michael G

    2006-08-01

    We propose a novel scheme for error-resilient image transmission. The proposed scheme employs a product coder consisting of low-density parity check (LDPC) codes and Reed-Solomon codes in order to deal effectively with bit errors. The efficiency of the proposed scheme is based on the exploitation of determinate symbols in Tanner graph decoding of LDPC codes and a novel product code optimization technique based on error estimation. Experimental evaluation demonstrates the superiority of the proposed system in comparison to recent state-of-the-art techniques for image transmission.

  17. Implementation of Layered Decoding Architecture for LDPC Code using Layered Min-Sum Algorithm

    Directory of Open Access Journals (Sweden)

    Sandeep Kakde

    2017-12-01

    Full Text Available For binary field and long code lengths, Low Density Parity Check (LDPC code approaches Shannon limit performance. LDPC codes provide remarkable error correction performance and therefore enlarge the design space for communication systems.In this paper, we have compare different digital modulation techniques and found that BPSK modulation technique is better than other modulation techniques in terms of BER. It also gives error performance of LDPC decoder over AWGN channel using Min-Sum algorithm. VLSI Architecture is proposed which uses the value re-use property of min-sum algorithm and gives high throughput. The proposed work has been implemented and tested on Xilinx Virtex 5 FPGA. The MATLAB result of LDPC decoder for low bit error rate (BER gives bit error rate in the range of 10-1 to 10-3.5 at SNR=1 to 2 for 20 no of iterations. So it gives good bit error rate performance. The latency of the parallel design of LDPC decoder has also reduced. It has accomplished 141.22 MHz maximum frequency and throughput of 2.02 Gbps while consuming less area of the design.

  18. Design and Analysis of Adaptive Message Coding on LDPC Decoder with Faulty Storage

    Directory of Open Access Journals (Sweden)

    Guangjun Ge

    2018-01-01

    Full Text Available Unreliable message storage severely degrades the performance of LDPC decoders. This paper discusses the impacts of message errors on LDPC decoders and schemes improving the robustness. Firstly, we develop a discrete density evolution analysis for faulty LDPC decoders, which indicates that protecting the sign bits of messages is effective enough for finite-precision LDPC decoders. Secondly, we analyze the effects of quantization precision loss for static sign bit protection and propose an embedded dynamic coding scheme by adaptively employing the least significant bits (LSBs to protect the sign bits. Thirdly, we give a construction of Hamming product code for the adaptive coding and present low complexity decoding algorithms. Theoretic analysis indicates that the proposed scheme outperforms traditional triple modular redundancy (TMR scheme in decoding both threshold and residual errors, while Monte Carlo simulations show that the performance loss is less than 0.2 dB when the storage error probability varies from 10-3 to 10-4.

  19. Ultra high speed optical transmission using subcarrier-multiplexed four-dimensional LDPC-coded modulation.

    Science.gov (United States)

    Batshon, Hussam G; Djordjevic, Ivan; Schmidt, Ted

    2010-09-13

    We propose a subcarrier-multiplexed four-dimensional LDPC bit-interleaved coded modulation scheme that is capable of achieving beyond 480 Gb/s single-channel transmission rate over optical channels. Subcarrier-multiplexed four-dimensional LDPC coded modulation scheme outperforms the corresponding dual polarization schemes by up to 4.6 dB in OSNR at BER 10(-8).

  20. Structured LDPC Codes over Integer Residue Rings

    Directory of Open Access Journals (Sweden)

    Mo Elisa

    2008-01-01

    Full Text Available Abstract This paper presents a new class of low-density parity-check (LDPC codes over represented by regular, structured Tanner graphs. These graphs are constructed using Latin squares defined over a multiplicative group of a Galois ring, rather than a finite field. Our approach yields codes for a wide range of code rates and more importantly, codes whose minimum pseudocodeword weights equal their minimum Hamming distances. Simulation studies show that these structured codes, when transmitted using matched signal sets over an additive-white-Gaussian-noise channel, can outperform their random counterparts of similar length and rate.

  1. Construction of LDPC codes over GF(q) with modified progressive edge growth

    Institute of Scientific and Technical Information of China (English)

    CHEN Xin; MEN Ai-dong; YANG Bo; QUAN Zi-yi

    2009-01-01

    A parity check matrix construction method for constructing a low-density parity-check (LDPC) codes over GF(q) (q>2) based on the modified progressive edge growth (PEG) algorithm is introduced. First, the nonzero locations of the parity check matrix are selected using the PEG algorithm. Then the nonzero elements are defined by avoiding the definition of subcode. A proof is given to show the good minimum distance property of constructed GF(q)-LDPC codes. Simulations are also presented to illustrate the good error performance of the designed codes.

  2. Polarization-multiplexed rate-adaptive non-binary-quasi-cyclic-LDPC-coded multilevel modulation with coherent detection for optical transport networks.

    Science.gov (United States)

    Arabaci, Murat; Djordjevic, Ivan B; Saunders, Ross; Marcoccia, Roberto M

    2010-02-01

    In order to achieve high-speed transmission over optical transport networks (OTNs) and maximize its throughput, we propose using a rate-adaptive polarization-multiplexed coded multilevel modulation with coherent detection based on component non-binary quasi-cyclic (QC) LDPC codes. Compared to prior-art bit-interleaved LDPC-coded modulation (BI-LDPC-CM) scheme, the proposed non-binary LDPC-coded modulation (NB-LDPC-CM) scheme not only reduces latency due to symbol- instead of bit-level processing but also provides either impressive reduction in computational complexity or striking improvements in coding gain depending on the constellation size. As the paper presents, compared to its prior-art binary counterpart, the proposed NB-LDPC-CM scheme addresses the needs of future OTNs, which are achieving the target BER performance and providing maximum possible throughput both over the entire lifetime of the OTN, better.

  3. Differentially Encoded LDPC Codes—Part I: Special Case of Product Accumulate Codes

    Directory of Open Access Journals (Sweden)

    (Tiffany JingLi

    2008-01-01

    Full Text Available Part I of a two-part series investigates product accumulate codes, a special class of differentially-encoded low density parity check (DE-LDPC codes with high performance and low complexity, on flat Rayleigh fading channels. In the coherent detection case, Divsalar's simple bounds and iterative thresholds using density evolution are computed to quantify the code performance at finite and infinite lengths, respectively. In the noncoherent detection case, a simple iterative differential detection and decoding (IDDD receiver is proposed and shown to be robust for different Doppler shifts. Extrinsic information transfer (EXIT charts reveal that, with pilot symbol assisted differential detection, the widespread practice of inserting pilot symbols to terminate the trellis actually incurs a loss in capacity, and a more efficient way is to separate pilots from the trellis. Through analysis and simulations, it is shown that PA codes perform very well with both coherent and noncoherent detections. The more general case of DE-LDPC codes, where the LDPC part may take arbitrary degree profiles, is studied in Part II Li 2008.

  4. Advanced GF(32) nonbinary LDPC coded modulation with non-uniform 9-QAM outperforming star 8-QAM.

    Science.gov (United States)

    Liu, Tao; Lin, Changyu; Djordjevic, Ivan B

    2016-06-27

    In this paper, we first describe a 9-symbol non-uniform signaling scheme based on Huffman code, in which different symbols are transmitted with different probabilities. By using the Huffman procedure, prefix code is designed to approach the optimal performance. Then, we introduce an algorithm to determine the optimal signal constellation sets for our proposed non-uniform scheme with the criterion of maximizing constellation figure of merit (CFM). The proposed nonuniform polarization multiplexed signaling 9-QAM scheme has the same spectral efficiency as the conventional 8-QAM. Additionally, we propose a specially designed GF(32) nonbinary quasi-cyclic LDPC code for the coded modulation system based on the 9-QAM non-uniform scheme. Further, we study the efficiency of our proposed non-uniform 9-QAM, combined with nonbinary LDPC coding, and demonstrate by Monte Carlo simulation that the proposed GF(23) nonbinary LDPC coded 9-QAM scheme outperforms nonbinary LDPC coded uniform 8-QAM by at least 0.8dB.

  5. LDPC code decoding adapted to the precoded partial response magnetic recording channels

    International Nuclear Information System (INIS)

    Lee, Jun; Kim, Kyuyong; Lee, Jaejin; Yang, Gijoo

    2004-01-01

    We propose a signal processing technique using LDPC (low-density parity-check) code instead of PRML (partial response maximum likelihood) system for the longitudinal magnetic recording channel. The scheme is designed by the precoder admitting level detection at the receiver-end and modifying the likelihood function for LDPC code decoding. The scheme can be collaborated with other decoder for turbo-like systems. The proposed algorithm can contribute to improve the performance of the conventional turbo-like systems

  6. LDPC code decoding adapted to the precoded partial response magnetic recording channels

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Jun E-mail: leejun28@sait.samsung.co.kr; Kim, Kyuyong; Lee, Jaejin; Yang, Gijoo

    2004-05-01

    We propose a signal processing technique using LDPC (low-density parity-check) code instead of PRML (partial response maximum likelihood) system for the longitudinal magnetic recording channel. The scheme is designed by the precoder admitting level detection at the receiver-end and modifying the likelihood function for LDPC code decoding. The scheme can be collaborated with other decoder for turbo-like systems. The proposed algorithm can contribute to improve the performance of the conventional turbo-like systems.

  7. Structured LDPC Codes over Integer Residue Rings

    Directory of Open Access Journals (Sweden)

    Marc A. Armand

    2008-07-01

    Full Text Available This paper presents a new class of low-density parity-check (LDPC codes over ℤ2a represented by regular, structured Tanner graphs. These graphs are constructed using Latin squares defined over a multiplicative group of a Galois ring, rather than a finite field. Our approach yields codes for a wide range of code rates and more importantly, codes whose minimum pseudocodeword weights equal their minimum Hamming distances. Simulation studies show that these structured codes, when transmitted using matched signal sets over an additive-white-Gaussian-noise channel, can outperform their random counterparts of similar length and rate.

  8. Characterization and Optimization of LDPC Codes for the 2-User Gaussian Multiple Access Channel

    Directory of Open Access Journals (Sweden)

    Declercq David

    2007-01-01

    Full Text Available We address the problem of designing good LDPC codes for the Gaussian multiple access channel (MAC. The framework we choose is to design multiuser LDPC codes with joint belief propagation decoding on the joint graph of the 2-user case. Our main result compared to existing work is to express analytically EXIT functions of the multiuser decoder with two different approximations of the density evolution. This allows us to propose a very simple linear programming optimization for the complicated problem of LDPC code design with joint multiuser decoding. The stability condition for our case is derived and used in the optimization constraints. The codes that we obtain for the 2-user case are quite good for various rates, especially if we consider the very simple optimization procedure.

  9. Rate-compatible protograph LDPC code families with linear minimum distance

    Science.gov (United States)

    Divsalar, Dariush (Inventor); Dolinar, Jr., Samuel J. (Inventor); Jones, Christopher R. (Inventor)

    2012-01-01

    Digital communication coding methods are shown, which generate certain types of low-density parity-check (LDPC) codes built from protographs. A first method creates protographs having the linear minimum distance property and comprising at least one variable node with degree less than 3. A second method creates families of protographs of different rates, all structurally identical for all rates except for a rate-dependent designation of certain variable nodes as transmitted or non-transmitted. A third method creates families of protographs of different rates, all structurally identical for all rates except for a rate-dependent designation of the status of certain variable nodes as non-transmitted or set to zero. LDPC codes built from the protographs created by these methods can simultaneously have low error floors and low iterative decoding thresholds.

  10. LDPC Codes--Structural Analysis and Decoding Techniques

    Science.gov (United States)

    Zhang, Xiaojie

    2012-01-01

    Low-density parity-check (LDPC) codes have been the focus of much research over the past decade thanks to their near Shannon limit performance and to their efficient message-passing (MP) decoding algorithms. However, the error floor phenomenon observed in MP decoding, which manifests itself as an abrupt change in the slope of the error-rate curve,…

  11. Peeling Decoding of LDPC Codes with Applications in Compressed Sensing

    Directory of Open Access Journals (Sweden)

    Weijun Zeng

    2016-01-01

    Full Text Available We present a new approach for the analysis of iterative peeling decoding recovery algorithms in the context of Low-Density Parity-Check (LDPC codes and compressed sensing. The iterative recovery algorithm is particularly interesting for its low measurement cost and low computational complexity. The asymptotic analysis can track the evolution of the fraction of unrecovered signal elements in each iteration, which is similar to the well-known density evolution analysis in the context of LDPC decoding algorithm. Our analysis shows that there exists a threshold on the density factor; if under this threshold, the recovery algorithm is successful; otherwise it will fail. Simulation results are also provided for verifying the agreement between the proposed asymptotic analysis and recovery algorithm. Compared with existing works of peeling decoding algorithm, focusing on the failure probability of the recovery algorithm, our proposed approach gives accurate evolution of performance with different parameters of measurement matrices and is easy to implement. We also show that the peeling decoding algorithm performs better than other schemes based on LDPC codes.

  12. Unitals and ovals of symmetric block designs in LDPC and space-time coding

    Science.gov (United States)

    Andriamanalimanana, Bruno R.

    2004-08-01

    An approach to the design of LDPC (low density parity check) error-correction and space-time modulation codes involves starting with known mathematical and combinatorial structures, and deriving code properties from structure properties. This paper reports on an investigation of unital and oval configurations within generic symmetric combinatorial designs, not just classical projective planes, as the underlying structure for classes of space-time LDPC outer codes. Of particular interest are the encoding and iterative (sum-product) decoding gains that these codes may provide. Various small-length cases have been numerically implemented in Java and Matlab for a number of channel models.

  13. Pilotless Frame Synchronization Using LDPC Code Constraints

    Science.gov (United States)

    Jones, Christopher; Vissasenor, John

    2009-01-01

    A method of pilotless frame synchronization has been devised for low- density parity-check (LDPC) codes. In pilotless frame synchronization , there are no pilot symbols; instead, the offset is estimated by ex ploiting selected aspects of the structure of the code. The advantag e of pilotless frame synchronization is that the bandwidth of the sig nal is reduced by an amount associated with elimination of the pilot symbols. The disadvantage is an increase in the amount of receiver data processing needed for frame synchronization.

  14. Construction and Iterative Decoding of LDPC Codes Over Rings for Phase-Noisy Channels

    Directory of Open Access Journals (Sweden)

    Karuppasami Sridhar

    2008-01-01

    Full Text Available Abstract This paper presents the construction and iterative decoding of low-density parity-check (LDPC codes for channels affected by phase noise. The LDPC code is based on integer rings and designed to converge under phase-noisy channels. We assume that phase variations are small over short blocks of adjacent symbols. A part of the constructed code is inherently built with this knowledge and hence able to withstand a phase rotation of radians, where " " is the number of phase symmetries in the signal set, that occur at different observation intervals. Another part of the code estimates the phase ambiguity present in every observation interval. The code makes use of simple blind or turbo phase estimators to provide phase estimates over every observation interval. We propose an iterative decoding schedule to apply the sum-product algorithm (SPA on the factor graph of the code for its convergence. To illustrate the new method, we present the performance results of an LDPC code constructed over with quadrature phase shift keying (QPSK modulated signals transmitted over a static channel, but affected by phase noise, which is modeled by the Wiener (random-walk process. The results show that the code can withstand phase noise of standard deviation per symbol with small loss.

  15. Memory-efficient decoding of LDPC codes

    Science.gov (United States)

    Kwok-San Lee, Jason; Thorpe, Jeremy; Hawkins, Jon

    2005-01-01

    We present a low-complexity quantization scheme for the implementation of regular (3,6) LDPC codes. The quantization parameters are optimized to maximize the mutual information between the source and the quantized messages. Using this non-uniform quantized belief propagation algorithm, we have simulated that an optimized 3-bit quantizer operates with 0.2dB implementation loss relative to a floating point decoder, and an optimized 4-bit quantizer operates less than 0.1dB quantization loss.

  16. Techniques and Architectures for Hazard-Free Semi-Parallel Decoding of LDPC Codes

    Directory of Open Access Journals (Sweden)

    Rovini Massimo

    2009-01-01

    Full Text Available The layered decoding algorithm has recently been proposed as an efficient means for the decoding of low-density parity-check (LDPC codes, thanks to the remarkable improvement in the convergence speed (2x of the decoding process. However, pipelined semi-parallel decoders suffer from violations or "hazards" between consecutive updates, which not only violate the layered principle but also enforce the loops in the code, thus spoiling the error correction performance. This paper describes three different techniques to properly reschedule the decoding updates, based on the careful insertion of "idle" cycles, to prevent the hazards of the pipeline mechanism. Also, different semi-parallel architectures of a layered LDPC decoder suitable for use with such techniques are analyzed. Then, taking the LDPC codes for the wireless local area network (IEEE 802.11n as a case study, a detailed analysis of the performance attained with the proposed techniques and architectures is reported, and results of the logic synthesis on a 65 nm low-power CMOS technology are shown.

  17. Bilayer expurgated LDPC codes with uncoded relaying

    Directory of Open Access Journals (Sweden)

    Md. Noor-A-Rahim

    2017-08-01

    Full Text Available Bilayer low-density parity-check (LDPC codes are an effective coding technique for decode-and-forward relaying, where the relay forwards extra parity bits to help the destination to decode the source bits correctly. In the existing bilayer coding scheme, these parity bits are protected by an error correcting code and assumed reliably available at the receiver. We propose an uncoded relaying scheme, where the extra parity bits are forwarded to the destination without any protection. Through density evolution analysis and simulation results, we show that our proposed scheme achieves better performance in terms of bit erasure probability than the existing relaying scheme. In addition, our proposed scheme results in lower complexity at the relay.

  18. A Golay complementary TS-based symbol synchronization scheme in variable rate LDPC-coded MB-OFDM UWBoF system

    Science.gov (United States)

    He, Jing; Wen, Xuejie; Chen, Ming; Chen, Lin

    2015-09-01

    In this paper, a Golay complementary training sequence (TS)-based symbol synchronization scheme is proposed and experimentally demonstrated in multiband orthogonal frequency division multiplexing (MB-OFDM) ultra-wideband over fiber (UWBoF) system with a variable rate low-density parity-check (LDPC) code. Meanwhile, the coding gain and spectral efficiency in the variable rate LDPC-coded MB-OFDM UWBoF system are investigated. By utilizing the non-periodic auto-correlation property of the Golay complementary pair, the start point of LDPC-coded MB-OFDM UWB signal can be estimated accurately. After 100 km standard single-mode fiber (SSMF) transmission, at the bit error rate of 1×10-3, the experimental results show that the short block length 64QAM-LDPC coding provides a coding gain of 4.5 dB, 3.8 dB and 2.9 dB for a code rate of 62.5%, 75% and 87.5%, respectively.

  19. Improving a Power Line Communications Standard with LDPC Codes

    Directory of Open Access Journals (Sweden)

    Hsu Christine

    2007-01-01

    Full Text Available We investigate a power line communications (PLC scheme that could be used to enhance the HomePlug 1.0 standard, specifically its ROBO mode which provides modest throughput for the worst case PLC channel. The scheme is based on using a low-density parity-check (LDPC code, in lieu of the concatenated Reed-Solomon and convolutional codes in ROBO mode. The PLC channel is modeled with multipath fading and Middleton's class A noise. Clipping is introduced to mitigate the effect of impulsive noise. A simple and effective method is devised to estimate the variance of the clipped noise for LDPC decoding. Simulation results show that the proposed scheme outperforms the HomePlug 1.0 ROBO mode and has lower computational complexity. The proposed scheme also dispenses with the repetition of information bits in ROBO mode to gain time diversity, resulting in 4-fold increase in physical layer throughput.

  20. Min-Max decoding for non binary LDPC codes

    OpenAIRE

    Savin, Valentin

    2008-01-01

    Iterative decoding of non-binary LDPC codes is currently performed using either the Sum-Product or the Min-Sum algorithms or slightly different versions of them. In this paper, several low-complexity quasi-optimal iterative algorithms are proposed for decoding non-binary codes. The Min-Max algorithm is one of them and it has the benefit of two possible LLR domain implementations: a standard implementation, whose complexity scales as the square of the Galois field's cardinality and a reduced c...

  1. A PEG Construction of LDPC Codes Based on the Betweenness Centrality Metric

    Directory of Open Access Journals (Sweden)

    BHURTAH-SEEWOOSUNGKUR, I.

    2016-05-01

    Full Text Available Progressive Edge Growth (PEG constructions are usually based on optimizing the distance metric by using various methods. In this work however, the distance metric is replaced by a different one, namely the betweenness centrality metric, which was shown to enhance routing performance in wireless mesh networks. A new type of PEG construction for Low-Density Parity-Check (LDPC codes is introduced based on the betweenness centrality metric borrowed from social networks terminology given that the bipartite graph describing the LDPC is analogous to a network of nodes. The algorithm is very efficient in filling edges on the bipartite graph by adding its connections in an edge-by-edge manner. The smallest graph size the new code could construct surpasses those obtained from a modified PEG algorithm - the RandPEG algorithm. To the best of the authors' knowledge, this paper produces the best regular LDPC column-weight two graphs. In addition, the technique proves to be competitive in terms of error-correcting performance. When compared to MacKay, PEG and other recent modified-PEG codes, the algorithm gives better performance over high SNR due to its particular edge and local graph properties.

  2. Experimental demonstration of the transmission performance for LDPC-coded multiband OFDM ultra-wideband over fiber system

    Science.gov (United States)

    He, Jing; Wen, Xuejie; Chen, Ming; Chen, Lin; Su, Jinshu

    2015-01-01

    To improve the transmission performance of multiband orthogonal frequency division multiplexing (MB-OFDM) ultra-wideband (UWB) over optical fiber, a pre-coding scheme based on low-density parity-check (LDPC) is adopted and experimentally demonstrated in the intensity-modulation and direct-detection MB-OFDM UWB over fiber system. Meanwhile, a symbol synchronization and pilot-aided channel estimation scheme is implemented on the receiver of the MB-OFDM UWB over fiber system. The experimental results show that the LDPC pre-coding scheme can work effectively in the MB-OFDM UWB over fiber system. After 70 km standard single-mode fiber (SSMF) transmission, at the bit error rate of 1 × 10-3, the receiver sensitivities are improved about 4 dB when the LDPC code rate is 75%.

  3. Error Correction using Quantum Quasi-Cyclic Low-Density Parity-Check(LDPC) Codes

    Science.gov (United States)

    Jing, Lin; Brun, Todd; Quantum Research Team

    Quasi-cyclic LDPC codes can approach the Shannon capacity and have efficient decoders. Manabu Hagiwara et al., 2007 presented a method to calculate parity check matrices with high girth. Two distinct, orthogonal matrices Hc and Hd are used. Using submatrices obtained from Hc and Hd by deleting rows, we can alter the code rate. The submatrix of Hc is used to correct Pauli X errors, and the submatrix of Hd to correct Pauli Z errors. We simulated this system for depolarizing noise on USC's High Performance Computing Cluster, and obtained the block error rate (BER) as a function of the error weight and code rate. From the rates of uncorrectable errors under different error weights we can extrapolate the BER to any small error probability. Our results show that this code family can perform reasonably well even at high code rates, thus considerably reducing the overhead compared to concatenated and surface codes. This makes these codes promising as storage blocks in fault-tolerant quantum computation. Error Correction using Quantum Quasi-Cyclic Low-Density Parity-Check(LDPC) Codes.

  4. Transmission over UWB channels with OFDM system using LDPC coding

    Science.gov (United States)

    Dziwoki, Grzegorz; Kucharczyk, Marcin; Sulek, Wojciech

    2009-06-01

    Hostile wireless environment requires use of sophisticated signal processing methods. The paper concerns on Ultra Wideband (UWB) transmission over Personal Area Networks (PAN) including MB-OFDM specification of physical layer. In presented work the transmission system with OFDM modulation was connected with LDPC encoder/decoder. Additionally the frame and bit error rate (FER and BER) of the system was decreased using results from the LDPC decoder in a kind of turbo equalization algorithm for better channel estimation. Computational block using evolutionary strategy, from genetic algorithms family, was also used in presented system. It was placed after SPA (Sum-Product Algorithm) decoder and is conditionally turned on in the decoding process. The result is increased effectiveness of the whole system, especially lower FER. The system was tested with two types of LDPC codes, depending on type of parity check matrices: randomly generated and constructed deterministically, optimized for practical decoder architecture implemented in the FPGA device.

  5. Construction and Iterative Decoding of LDPC Codes Over Rings for Phase-Noisy Channels

    Directory of Open Access Journals (Sweden)

    William G. Cowley

    2008-04-01

    Full Text Available This paper presents the construction and iterative decoding of low-density parity-check (LDPC codes for channels affected by phase noise. The LDPC code is based on integer rings and designed to converge under phase-noisy channels. We assume that phase variations are small over short blocks of adjacent symbols. A part of the constructed code is inherently built with this knowledge and hence able to withstand a phase rotation of 2π/M radians, where “M” is the number of phase symmetries in the signal set, that occur at different observation intervals. Another part of the code estimates the phase ambiguity present in every observation interval. The code makes use of simple blind or turbo phase estimators to provide phase estimates over every observation interval. We propose an iterative decoding schedule to apply the sum-product algorithm (SPA on the factor graph of the code for its convergence. To illustrate the new method, we present the performance results of an LDPC code constructed over ℤ4 with quadrature phase shift keying (QPSK modulated signals transmitted over a static channel, but affected by phase noise, which is modeled by the Wiener (random-walk process. The results show that the code can withstand phase noise of 2∘ standard deviation per symbol with small loss.

  6. UEP Concepts in Modulation and Coding

    Directory of Open Access Journals (Sweden)

    Werner Henkel

    2010-01-01

    Full Text Available First unequal error protection (UEP proposals date back to the 1960's (Masnick and Wolf; 1967, but now with the introduction of scalable video, UEP develops to a key concept for the transport of multimedia data. The paper presents an overview of some new approaches realizing UEP properties in physical transport, especially multicarrier modulation, or with LDPC and Turbo codes. For multicarrier modulation, UEP bit-loading together with hierarchical modulation is described allowing for an arbitrary number of classes, arbitrary SNR margins between the classes, and arbitrary number of bits per class. In Turbo coding, pruning, as a counterpart of puncturing is presented for flexible bit-rate adaptations, including tables with optimized pruning patterns. Bit- and/or check-irregular LDPC codes may be designed to provide UEP to its code bits. However, irregular degree distributions alone do not ensure UEP, and other necessary properties of the parity-check matrix for providing UEP are also pointed out. Pruning is also the means for constructing variable-rate LDPC codes for UEP, especially controlling the check-node profile.

  7. Analysis and Construction of Full-Diversity Joint Network-LDPC Codes for Cooperative Communications

    Directory of Open Access Journals (Sweden)

    Capirone Daniele

    2010-01-01

    Full Text Available Transmit diversity is necessary in harsh environments to reduce the required transmit power for achieving a given error performance at a certain transmission rate. In networks, cooperative communication is a well-known technique to yield transmit diversity and network coding can increase the spectral efficiency. These two techniques can be combined to achieve a double diversity order for a maximum coding rate on the Multiple-Access Relay Channel (MARC, where two sources share a common relay in their transmission to the destination. However, codes have to be carefully designed to obtain the intrinsic diversity offered by the MARC. This paper presents the principles to design a family of full-diversity LDPC codes with maximum rate. Simulation of the word error rate performance of the new proposed family of LDPC codes for the MARC confirms the full diversity.

  8. Reply to "Comments on Techniques and Architectures for Hazard-Free Semi-Parallel Decoding of LDPC Codes"

    Directory of Open Access Journals (Sweden)

    Rovini Massimo

    2009-01-01

    Full Text Available This is a reply to the comments by Gunnam et al. "Comments on 'Techniques and architectures for hazard-free semi-parallel decoding of LDPC codes'", EURASIP Journal on Embedded Systems, vol. 2009, Article ID 704174 on our recent work "Techniques and architectures for hazard-free semi-parallel decoding of LDPC codes", EURASIP Journal on Embedded Systems, vol. 2009, Article ID 723465.

  9. PMD compensation in fiber-optic communication systems with direct detection using LDPC-coded OFDM.

    Science.gov (United States)

    Djordjevic, Ivan B

    2007-04-02

    The possibility of polarization-mode dispersion (PMD) compensation in fiber-optic communication systems with direct detection using a simple channel estimation technique and low-density parity-check (LDPC)-coded orthogonal frequency division multiplexing (OFDM) is demonstrated. It is shown that even for differential group delay (DGD) of 4/BW (BW is the OFDM signal bandwidth), the degradation due to the first-order PMD can be completely compensated for. Two classes of LDPC codes designed based on two different combinatorial objects (difference systems and product of combinatorial designs) suitable for use in PMD compensation are introduced.

  10. Performance Analysis of Iterative Decoding Algorithms for PEG LDPC Codes in Nakagami Fading Channels

    Directory of Open Access Journals (Sweden)

    O. Al Rasheed

    2013-11-01

    Full Text Available In this paper we give a comparative analysis of decoding algorithms of Low Density Parity Check (LDPC codes in a channel with the Nakagami distribution of the fading envelope. We consider the Progressive Edge-Growth (PEG method and Improved PEG method for the parity check matrix construction, which can be used to avoid short girths, small trapping sets and a high level of error floor. A comparative analysis of several classes of LDPC codes in various propagation conditions and decoded using different decoding algorithms is also presented.

  11. High-throughput GPU-based LDPC decoding

    Science.gov (United States)

    Chang, Yang-Lang; Chang, Cheng-Chun; Huang, Min-Yu; Huang, Bormin

    2010-08-01

    Low-density parity-check (LDPC) code is a linear block code known to approach the Shannon limit via the iterative sum-product algorithm. LDPC codes have been adopted in most current communication systems such as DVB-S2, WiMAX, WI-FI and 10GBASE-T. LDPC for the needs of reliable and flexible communication links for a wide variety of communication standards and configurations have inspired the demand for high-performance and flexibility computing. Accordingly, finding a fast and reconfigurable developing platform for designing the high-throughput LDPC decoder has become important especially for rapidly changing communication standards and configurations. In this paper, a new graphic-processing-unit (GPU) LDPC decoding platform with the asynchronous data transfer is proposed to realize this practical implementation. Experimental results showed that the proposed GPU-based decoder achieved 271x speedup compared to its CPU-based counterpart. It can serve as a high-throughput LDPC decoder.

  12. A novel construction method of QC-LDPC codes based on the subgroup of the finite field multiplicative group for optical transmission systems

    Science.gov (United States)

    Yuan, Jian-guo; Zhou, Guang-xiang; Gao, Wen-chun; Wang, Yong; Lin, Jin-zhao; Pang, Yu

    2016-01-01

    According to the requirements of the increasing development for optical transmission systems, a novel construction method of quasi-cyclic low-density parity-check (QC-LDPC) codes based on the subgroup of the finite field multiplicative group is proposed. Furthermore, this construction method can effectively avoid the girth-4 phenomena and has the advantages such as simpler construction, easier implementation, lower encoding/decoding complexity, better girth properties and more flexible adjustment for the code length and code rate. The simulation results show that the error correction performance of the QC-LDPC(3 780,3 540) code with the code rate of 93.7% constructed by this proposed method is excellent, its net coding gain is respectively 0.3 dB, 0.55 dB, 1.4 dB and 1.98 dB higher than those of the QC-LDPC(5 334,4 962) code constructed by the method based on the inverse element characteristics in the finite field multiplicative group, the SCG-LDPC(3 969,3 720) code constructed by the systematically constructed Gallager (SCG) random construction method, the LDPC(32 640,30 592) code in ITU-T G.975.1 and the classic RS(255,239) code which is widely used in optical transmission systems in ITU-T G.975 at the bit error rate ( BER) of 10-7. Therefore, the constructed QC-LDPC(3 780,3 540) code is more suitable for optical transmission systems.

  13. Multiple component codes based generalized LDPC codes for high-speed optical transport.

    Science.gov (United States)

    Djordjevic, Ivan B; Wang, Ting

    2014-07-14

    A class of generalized low-density parity-check (GLDPC) codes suitable for optical communications is proposed, which consists of multiple local codes. It is shown that Hamming, BCH, and Reed-Muller codes can be used as local codes, and that the maximum a posteriori probability (MAP) decoding of these local codes by Ashikhmin-Lytsin algorithm is feasible in terms of complexity and performance. We demonstrate that record coding gains can be obtained from properly designed GLDPC codes, derived from multiple component codes. We then show that several recently proposed classes of LDPC codes such as convolutional and spatially-coupled codes can be described using the concept of GLDPC coding, which indicates that the GLDPC coding can be used as a unified platform for advanced FEC enabling ultra-high speed optical transport. The proposed class of GLDPC codes is also suitable for code-rate adaption, to adjust the error correction strength depending on the optical channel conditions.

  14. Performance analysis of LDPC codes on OOK terahertz wireless channels

    Science.gov (United States)

    Chun, Liu; Chang, Wang; Jun-Cheng, Cao

    2016-02-01

    Atmospheric absorption, scattering, and scintillation are the major causes to deteriorate the transmission quality of terahertz (THz) wireless communications. An error control coding scheme based on low density parity check (LDPC) codes with soft decision decoding algorithm is proposed to improve the bit-error-rate (BER) performance of an on-off keying (OOK) modulated THz signal through atmospheric channel. The THz wave propagation characteristics and channel model in atmosphere is set up. Numerical simulations validate the great performance of LDPC codes against the atmospheric fading and demonstrate the huge potential in future ultra-high speed beyond Gbps THz communications. Project supported by the National Key Basic Research Program of China (Grant No. 2014CB339803), the National High Technology Research and Development Program of China (Grant No. 2011AA010205), the National Natural Science Foundation of China (Grant Nos. 61131006, 61321492, and 61204135), the Major National Development Project of Scientific Instrument and Equipment (Grant No. 2011YQ150021), the National Science and Technology Major Project (Grant No. 2011ZX02707), the International Collaboration and Innovation Program on High Mobility Materials Engineering of the Chinese Academy of Sciences, and the Shanghai Municipal Commission of Science and Technology (Grant No. 14530711300).

  15. Design LDPC Codes without Cycles of Length 4 and 6

    Directory of Open Access Journals (Sweden)

    Kiseon Kim

    2008-04-01

    Full Text Available We present an approach for constructing LDPC codes without cycles of length 4 and 6. Firstly, we design 3 submatrices with different shifting functions given by the proposed schemes, then combine them into the matrix specified by the proposed approach, and, finally, expand the matrix into a desired parity-check matrix using identity matrices and cyclic shift matrices of the identity matrices. The simulation result in AWGN channel verifies that the BER of the proposed code is close to those of Mackay's random codes and Tanner's QC codes, and the good BER performance of the proposed can remain at high code rates.

  16. LDPC product coding scheme with extrinsic information for bit patterned media recoding

    Science.gov (United States)

    Jeong, Seongkwon; Lee, Jaejin

    2017-05-01

    Since the density limit of the current perpendicular magnetic storage system will soon be reached, bit patterned media recording (BPMR) is a promising candidate for the next generation storage system to achieve an areal density beyond 1 Tb/in2. Each recording bit is stored in a fabricated magnetic island and the space between the magnetic islands is nonmagnetic in BPMR. To approach recording densities of 1 Tb/in2, the spacing of the magnetic islands must be less than 25 nm. Consequently, severe inter-symbol interference (ISI) and inter-track interference (ITI) occur. ITI and ISI degrade the performance of BPMR. In this paper, we propose a low-density parity check (LDPC) product coding scheme that exploits extrinsic information for BPMR. This scheme shows an improved bit error rate performance compared to that in which one LDPC code is used.

  17. Rate-Compatible LDPC Codes with Linear Minimum Distance

    Science.gov (United States)

    Divsalar, Dariush; Jones, Christopher; Dolinar, Samuel

    2009-01-01

    A recently developed method of constructing protograph-based low-density parity-check (LDPC) codes provides for low iterative decoding thresholds and minimum distances proportional to block sizes, and can be used for various code rates. A code constructed by this method can have either fixed input block size or fixed output block size and, in either case, provides rate compatibility. The method comprises two submethods: one for fixed input block size and one for fixed output block size. The first mentioned submethod is useful for applications in which there are requirements for rate-compatible codes that have fixed input block sizes. These are codes in which only the numbers of parity bits are allowed to vary. The fixed-output-blocksize submethod is useful for applications in which framing constraints are imposed on the physical layers of affected communication systems. An example of such a system is one that conforms to one of many new wireless-communication standards that involve the use of orthogonal frequency-division modulation

  18. LDPC product coding scheme with extrinsic information for bit patterned media recoding

    Directory of Open Access Journals (Sweden)

    Seongkwon Jeong

    2017-05-01

    Full Text Available Since the density limit of the current perpendicular magnetic storage system will soon be reached, bit patterned media recording (BPMR is a promising candidate for the next generation storage system to achieve an areal density beyond 1 Tb/in2. Each recording bit is stored in a fabricated magnetic island and the space between the magnetic islands is nonmagnetic in BPMR. To approach recording densities of 1 Tb/in2, the spacing of the magnetic islands must be less than 25 nm. Consequently, severe inter-symbol interference (ISI and inter-track interference (ITI occur. ITI and ISI degrade the performance of BPMR. In this paper, we propose a low-density parity check (LDPC product coding scheme that exploits extrinsic information for BPMR. This scheme shows an improved bit error rate performance compared to that in which one LDPC code is used.

  19. LDPC-coded MIMO optical communication over the atmospheric turbulence channel using Q-ary pulse-position modulation.

    Science.gov (United States)

    Djordjevic, Ivan B

    2007-08-06

    We describe a coded power-efficient transmission scheme based on repetition MIMO principle suitable for communication over the atmospheric turbulence channel, and determine its channel capacity. The proposed scheme employs the Q-ary pulse-position modulation. We further study how to approach the channel capacity limits using low-density parity-check (LDPC) codes. Component LDPC codes are designed using the concept of pairwise-balanced designs. Contrary to the several recent publications, bit-error rates and channel capacities are reported assuming non-ideal photodetection. The atmospheric turbulence channel is modeled using the Gamma-Gamma distribution function due to Al-Habash et al. Excellent bit-error rate performance improvement, over uncoded case, is found.

  20. Low Power LDPC Code Decoder Architecture Based on Intermediate Message Compression Technique

    Science.gov (United States)

    Shimizu, Kazunori; Togawa, Nozomu; Ikenaga, Takeshi; Goto, Satoshi

    Reducing the power dissipation for LDPC code decoder is a major challenging task to apply it to the practical digital communication systems. In this paper, we propose a low power LDPC code decoder architecture based on an intermediate message-compression technique which features as follows: (i) An intermediate message compression technique enables the decoder to reduce the required memory capacity and write power dissipation. (ii) A clock gated shift register based intermediate message memory architecture enables the decoder to decompress the compressed messages in a single clock cycle while reducing the read power dissipation. The combination of the above two techniques enables the decoder to reduce the power dissipation while keeping the decoding throughput. The simulation results show that the proposed architecture improves the power efficiency up to 52% and 18% compared to that of the decoder based on the overlapped schedule and the rapid convergence schedule without the proposed techniques respectively.

  1. Opportunistic Adaptive Transmission for Network Coding Using Nonbinary LDPC Codes

    Directory of Open Access Journals (Sweden)

    Cocco Giuseppe

    2010-01-01

    Full Text Available Network coding allows to exploit spatial diversity naturally present in mobile wireless networks and can be seen as an example of cooperative communication at the link layer and above. Such promising technique needs to rely on a suitable physical layer in order to achieve its best performance. In this paper, we present an opportunistic packet scheduling method based on physical layer considerations. We extend channel adaptation proposed for the broadcast phase of asymmetric two-way bidirectional relaying to a generic number of sinks and apply it to a network context. The method consists of adapting the information rate for each receiving node according to its channel status and independently of the other nodes. In this way, a higher network throughput can be achieved at the expense of a slightly higher complexity at the transmitter. This configuration allows to perform rate adaptation while fully preserving the benefits of channel and network coding. We carry out an information theoretical analysis of such approach and of that typically used in network coding. Numerical results based on nonbinary LDPC codes confirm the effectiveness of our approach with respect to previously proposed opportunistic scheduling techniques.

  2. FPGA implementation of low complexity LDPC iterative decoder

    Science.gov (United States)

    Verma, Shivani; Sharma, Sanjay

    2016-07-01

    Low-density parity-check (LDPC) codes, proposed by Gallager, emerged as a class of codes which can yield very good performance on the additive white Gaussian noise channel as well as on the binary symmetric channel. LDPC codes have gained lots of importance due to their capacity achieving property and excellent performance in the noisy channel. Belief propagation (BP) algorithm and its approximations, most notably min-sum, are popular iterative decoding algorithms used for LDPC and turbo codes. The trade-off between the hardware complexity and the decoding throughput is a critical factor in the implementation of the practical decoder. This article presents introduction to LDPC codes and its various decoding algorithms followed by realisation of LDPC decoder by using simplified message passing algorithm and partially parallel decoder architecture. Simplified message passing algorithm has been proposed for trade-off between low decoding complexity and decoder performance. It greatly reduces the routing and check node complexity of the decoder. Partially parallel decoder architecture possesses high speed and reduced complexity. The improved design of the decoder possesses a maximum symbol throughput of 92.95 Mbps and a maximum of 18 decoding iterations. The article presents implementation of 9216 bits, rate-1/2, (3, 6) LDPC decoder on Xilinx XC3D3400A device from Spartan-3A DSP family.

  3. On the performance of 1-level LDPC lattices

    OpenAIRE

    Sadeghi, Mohammad-Reza; Sakzad, Amin

    2013-01-01

    The low-density parity-check (LDPC) lattices perform very well in high dimensions under generalized min-sum iterative decoding algorithm. In this work we focus on 1-level LDPC lattices. We show that these lattices are the same as lattices constructed based on Construction A and low-density lattice-code (LDLC) lattices. In spite of having slightly lower coding gain, 1-level regular LDPC lattices have remarkable performances. The lower complexity nature of the decoding algorithm for these type ...

  4. Optical LDPC decoders for beyond 100 Gbits/s optical transmission.

    Science.gov (United States)

    Djordjevic, Ivan B; Xu, Lei; Wang, Ting

    2009-05-01

    We present an optical low-density parity-check (LDPC) decoder suitable for implementation above 100 Gbits/s, which provides large coding gains when based on large-girth LDPC codes. We show that a basic building block, the probabilities multiplier circuit, can be implemented using a Mach-Zehnder interferometer, and we propose corresponding probabilistic-domain sum-product algorithm (SPA). We perform simulations of a fully parallel implementation employing girth-10 LDPC codes and proposed SPA. The girth-10 LDPC(24015,19212) code of the rate of 0.8 outperforms the BCH(128,113)xBCH(256,239) turbo-product code of the rate of 0.82 by 0.91 dB (for binary phase-shift keying at 100 Gbits/s and a bit error rate of 10(-9)), and provides a net effective coding gain of 10.09 dB.

  5. Hardwarearchitektur für einen universellen LDPC Decoder

    Directory of Open Access Journals (Sweden)

    C. Beuschel

    2009-05-01

    Full Text Available Im vorliegenden Beitrag wird eine universelle Decoderarchitektur für einen Low-Density Parity-Check (LDPC Code Decoder vorgestellt. Anders als bei den in der Literatur häufig beschriebenen Architekturen für strukturierte Codes ist die hier vorgestellte Architektur frei programmierbar, so dass jeder beliebige LDPC Code durch eine Änderung der Initialisierung des Speichers für die Prüfmatrix mit derselben Hardware decodiert werden kann. Die größte Herausforderung beim Entwurf von teilparallelen LDPC Decoder Architekturen liegt im konfliktfreien Datenaustausch zwischen mehreren parallelen Speichern und Berechnungseinheiten, wozu ein Mapping und Scheduling Algorithmus benötigt wird. Der hier vorgestellte Algorithmus stützt sich auf Graphentheorie und findet für jeden beliebigen LDPC Code eine für die Architektur optimale Lösung. Damit sind keine Wartezyklen notwendig und die Parallelität der Architektur wird zu jedem Zeitpunkt voll ausgenutzt.

  6. FPGA implementation of high-performance QC-LDPC decoder for optical communications

    Science.gov (United States)

    Zou, Ding; Djordjevic, Ivan B.

    2015-01-01

    Forward error correction is as one of the key technologies enabling the next-generation high-speed fiber optical communications. Quasi-cyclic (QC) low-density parity-check (LDPC) codes have been considered as one of the promising candidates due to their large coding gain performance and low implementation complexity. In this paper, we present our designed QC-LDPC code with girth 10 and 25% overhead based on pairwise balanced design. By FPGAbased emulation, we demonstrate that the 5-bit soft-decision LDPC decoder can achieve 11.8dB net coding gain with no error floor at BER of 10-15 avoiding using any outer code or post-processing method. We believe that the proposed single QC-LDPC code is a promising solution for 400Gb/s optical communication systems and beyond.

  7. Um estudo sobre a construção, desempenho e implementação em VHDL de códigos LDPC binários, irregulares e estruturados para aplicação em comunicações ópticas

    OpenAIRE

    Antônio Unias de Lucena

    2015-01-01

    Resumo: O emprego de códigos LDPC em comunicações ópticas vem recebendo especial atenção nos últimos anos devido à sua elevada capacidade de correção de erros, fato que possibilita enlaces mais longos e com maior capacidade de transmissão. A presente dissertação apresenta um estudo de códigos LDPC binários, irregulares e estruturados (IE-LDPC), bem como, uma comparação do desempenho de dois algoritmos de decodificação comumente utilizados na decodificação de códigos LDPC: o algoritmo soma-pro...

  8. Kódování a efektivita LDPC kódů

    OpenAIRE

    Kozlík, Andrew

    2011-01-01

    Low-density parity-check (LDPC) codes are linear error correcting codes which are capable of performing near channel capacity. Furthermore, they admit efficient decoding algorithms that provide near optimum performance. Their main disadvantage is that most LDPC codes have relatively complex encoders. In this thesis, we begin by giving a detailed discussion of the sum-product decoding algorithm, we then study the performance of LDPC codes on the binary erasure channel under sum-product decodin...

  9. Low Complexity Approach for High Throughput Belief-Propagation based Decoding of LDPC Codes

    Directory of Open Access Journals (Sweden)

    BOT, A.

    2013-11-01

    Full Text Available The paper proposes a low complexity belief propagation (BP based decoding algorithm for LDPC codes. In spite of the iterative nature of the decoding process, the proposed algorithm provides both reduced complexity and increased BER performances as compared with the classic min-sum (MS algorithm, generally used for hardware implementations. Linear approximations of check-nodes update function are used in order to reduce the complexity of the BP algorithm. Considering this decoding approach, an FPGA based hardware architecture is proposed for implementing the decoding algorithm, aiming to increase the decoder throughput. FPGA technology was chosen for the LDPC decoder implementation, due to its parallel computation and reconfiguration capabilities. The obtained results show improvements regarding decoding throughput and BER performances compared with state-of-the-art approaches.

  10. Construction of Rate-Compatible LDPC Codes Utilizing Information Shortening and Parity Puncturing

    Directory of Open Access Journals (Sweden)

    Jones Christopher R

    2005-01-01

    Full Text Available This paper proposes a method for constructing rate-compatible low-density parity-check (LDPC codes. The construction considers the problem of optimizing a family of rate-compatible degree distributions as well as the placement of bipartite graph edges. A hybrid approach that combines information shortening and parity puncturing is proposed. Local graph conditioning techniques for the suppression of error floors are also included in the construction methodology.

  11. A Scalable Architecture of a Structured LDPC Decoder

    Science.gov (United States)

    Lee, Jason Kwok-San; Lee, Benjamin; Thorpe, Jeremy; Andrews, Kenneth; Dolinar, Sam; Hamkins, Jon

    2004-01-01

    We present a scalable decoding architecture for a certain class of structured LDPC codes. The codes are designed using a small (n,r) protograph that is replicated Z times to produce a decoding graph for a (Z x n, Z x r) code. Using this architecture, we have implemented a decoder for a (4096,2048) LDPC code on a Xilinx Virtex-II 2000 FPGA, and achieved decoding speeds of 31 Mbps with 10 fixed iterations. The implemented message-passing algorithm uses an optimized 3-bit non-uniform quantizer that operates with 0.2dB implementation loss relative to a floating point decoder.

  12. BER EVALUATION OF LDPC CODES WITH GMSK IN NAKAGAMI FADING CHANNEL

    Directory of Open Access Journals (Sweden)

    Surbhi Sharma

    2010-06-01

    Full Text Available LDPC codes (Low Density Parity Check Codes have already proved its efficacy while showing its performance near to the Shannon limit. Channel coding schemes are spectrally inefficient as using an unfiltered binary data stream to modulate an RF carrier that will produce an RF spectrum of considerable bandwidth. Techniques have been developed to improve this bandwidth inefficiency or spectral efficiency, and ease detection. GMSK or Gaussian-filtered Minimum Shift Keying uses a Gaussian Filter of an appropriate bandwidth so as to make system spectrally efficient. A Nakagami model provides a better explanation to less and more severe conditions than the Rayleigh and Rician model and provide a better fit to the mobile communication channel data. In this paper we have demonstrated the performance of Low Density Parity Check codes with GMSK modulation (BT product=0.25 technique in Nakagami fading channel. In results it is shown that average bit error rate decreases as the ‘m’ parameter increases (Less fading.

  13. Statistical mechanics analysis of LDPC coding in MIMO Gaussian channels

    Energy Technology Data Exchange (ETDEWEB)

    Alamino, Roberto C; Saad, David [Neural Computing Research Group, Aston University, Birmingham B4 7ET (United Kingdom)

    2007-10-12

    Using analytical methods of statistical mechanics, we analyse the typical behaviour of a multiple-input multiple-output (MIMO) Gaussian channel with binary inputs under low-density parity-check (LDPC) network coding and joint decoding. The saddle point equations for the replica symmetric solution are found in particular realizations of this channel, including a small and large number of transmitters and receivers. In particular, we examine the cases of a single transmitter, a single receiver and symmetric and asymmetric interference. Both dynamical and thermodynamical transitions from the ferromagnetic solution of perfect decoding to a non-ferromagnetic solution are identified for the cases considered, marking the practical and theoretical limits of the system under the current coding scheme. Numerical results are provided, showing the typical level of improvement/deterioration achieved with respect to the single transmitter/receiver result, for the various cases.

  14. Statistical mechanics analysis of LDPC coding in MIMO Gaussian channels

    International Nuclear Information System (INIS)

    Alamino, Roberto C; Saad, David

    2007-01-01

    Using analytical methods of statistical mechanics, we analyse the typical behaviour of a multiple-input multiple-output (MIMO) Gaussian channel with binary inputs under low-density parity-check (LDPC) network coding and joint decoding. The saddle point equations for the replica symmetric solution are found in particular realizations of this channel, including a small and large number of transmitters and receivers. In particular, we examine the cases of a single transmitter, a single receiver and symmetric and asymmetric interference. Both dynamical and thermodynamical transitions from the ferromagnetic solution of perfect decoding to a non-ferromagnetic solution are identified for the cases considered, marking the practical and theoretical limits of the system under the current coding scheme. Numerical results are provided, showing the typical level of improvement/deterioration achieved with respect to the single transmitter/receiver result, for the various cases

  15. Study regarding the density evolution of messages and the characteristic functions associated of a LDPC code

    Science.gov (United States)

    Drăghici, S.; Proştean, O.; Răduca, E.; Haţiegan, C.; Hălălae, I.; Pădureanu, I.; Nedeloni, M.; (Barboni Haţiegan, L.

    2017-01-01

    In this paper a method with which a set of characteristic functions are associated to a LDPC code is shown and also functions that represent the evolution density of messages that go along the edges of a Tanner graph. Graphic representations of the density evolution are shown respectively the study and simulation of likelihood threshold that render asymptotic boundaries between which there are decodable codes were made using MathCad V14 software.

  16. Percolation bounds for decoding thresholds with correlated erasures in quantum LDPC codes

    Science.gov (United States)

    Hamilton, Kathleen; Pryadko, Leonid

    Correlations between errors can dramatically affect decoding thresholds, in some cases eliminating the threshold altogether. We analyze the existence of a threshold for quantum low-density parity-check (LDPC) codes in the case of correlated erasures. When erasures are positively correlated, the corresponding multi-variate Bernoulli distribution can be modeled in terms of cluster errors, where qubits in clusters of various size can be marked all at once. In a code family with distance scaling as a power law of the code length, erasures can be always corrected below percolation on a qubit adjacency graph associated with the code. We bound this correlated percolation transition by weighted (uncorrelated) percolation on a specially constructed cluster connectivity graph, and apply our recent results to construct several bounds for the latter. This research was supported in part by the NSF Grant PHY-1416578 and by the ARO Grant W911NF-14-1-0272.

  17. Optimal Codes for the Burst Erasure Channel

    Science.gov (United States)

    Hamkins, Jon

    2010-01-01

    Deep space communications over noisy channels lead to certain packets that are not decodable. These packets leave gaps, or bursts of erasures, in the data stream. Burst erasure correcting codes overcome this problem. These are forward erasure correcting codes that allow one to recover the missing gaps of data. Much of the recent work on this topic concentrated on Low-Density Parity-Check (LDPC) codes. These are more complicated to encode and decode than Single Parity Check (SPC) codes or Reed-Solomon (RS) codes, and so far have not been able to achieve the theoretical limit for burst erasure protection. A block interleaved maximum distance separable (MDS) code (e.g., an SPC or RS code) offers near-optimal burst erasure protection, in the sense that no other scheme of equal total transmission length and code rate could improve the guaranteed correctible burst erasure length by more than one symbol. The optimality does not depend on the length of the code, i.e., a short MDS code block interleaved to a given length would perform as well as a longer MDS code interleaved to the same overall length. As a result, this approach offers lower decoding complexity with better burst erasure protection compared to other recent designs for the burst erasure channel (e.g., LDPC codes). A limitation of the design is its lack of robustness to channels that have impairments other than burst erasures (e.g., additive white Gaussian noise), making its application best suited for correcting data erasures in layers above the physical layer. The efficiency of a burst erasure code is the length of its burst erasure correction capability divided by the theoretical upper limit on this length. The inefficiency is one minus the efficiency. The illustration compares the inefficiency of interleaved RS codes to Quasi-Cyclic (QC) LDPC codes, Euclidean Geometry (EG) LDPC codes, extended Irregular Repeat Accumulate (eIRA) codes, array codes, and random LDPC codes previously proposed for burst erasure

  18. A Low-Complexity Euclidean Orthogonal LDPC Architecture for Low Power Applications.

    Science.gov (United States)

    Revathy, M; Saravanan, R

    2015-01-01

    Low-density parity-check (LDPC) codes have been implemented in latest digital video broadcasting, broadband wireless access (WiMax), and fourth generation of wireless standards. In this paper, we have proposed a high efficient low-density parity-check code (LDPC) decoder architecture for low power applications. This study also considers the design and analysis of check node and variable node units and Euclidean orthogonal generator in LDPC decoder architecture. The Euclidean orthogonal generator is used to reduce the error rate of the proposed LDPC architecture, which can be incorporated between check and variable node architecture. This proposed decoder design is synthesized on Xilinx 9.2i platform and simulated using Modelsim, which is targeted to 45 nm devices. Synthesis report proves that the proposed architecture greatly reduces the power consumption and hardware utilizations on comparing with different conventional architectures.

  19. A Low-Complexity Euclidean Orthogonal LDPC Architecture for Low Power Applications

    Directory of Open Access Journals (Sweden)

    M. Revathy

    2015-01-01

    Full Text Available Low-density parity-check (LDPC codes have been implemented in latest digital video broadcasting, broadband wireless access (WiMax, and fourth generation of wireless standards. In this paper, we have proposed a high efficient low-density parity-check code (LDPC decoder architecture for low power applications. This study also considers the design and analysis of check node and variable node units and Euclidean orthogonal generator in LDPC decoder architecture. The Euclidean orthogonal generator is used to reduce the error rate of the proposed LDPC architecture, which can be incorporated between check and variable node architecture. This proposed decoder design is synthesized on Xilinx 9.2i platform and simulated using Modelsim, which is targeted to 45 nm devices. Synthesis report proves that the proposed architecture greatly reduces the power consumption and hardware utilizations on comparing with different conventional architectures.

  20. On the reduced-complexity of LDPC decoders for ultra-high-speed optical transmission.

    Science.gov (United States)

    Djordjevic, Ivan B; Xu, Lei; Wang, Ting

    2010-10-25

    We propose two reduced-complexity (RC) LDPC decoders, which can be used in combination with large-girth LDPC codes to enable ultra-high-speed serial optical transmission. We show that optimally attenuated RC min-sum sum algorithm performs only 0.46 dB (at BER of 10(-9)) worse than conventional sum-product algorithm, while having lower storage memory requirements and much lower latency. We further study the use of RC LDPC decoding algorithms in multilevel coded modulation with coherent detection and show that with RC decoding algorithms we can achieve the net coding gain larger than 11 dB at BERs below 10(-9).

  1. Analysis of Minimal LDPC Decoder System on a Chip Implementation

    Directory of Open Access Journals (Sweden)

    T. Palenik

    2015-09-01

    Full Text Available This paper presents a practical method of potential replacement of several different Quasi-Cyclic Low-Density Parity-Check (QC-LDPC codes with one, with the intention of saving as much memory as required to implement the LDPC encoder and decoder in a memory-constrained System on a Chip (SoC. The presented method requires only a very small modification of the existing encoder and decoder, making it suitable for utilization in a Software Defined Radio (SDR platform. Besides the analysis of the effects of necessary variable-node value fixation during the Belief Propagation (BP decoding algorithm, practical standard-defined code parameters are scrutinized in order to evaluate the feasibility of the proposed LDPC setup simplification. Finally, the error performance of the modified system structure is evaluated and compared with the original system structure by means of simulation.

  2. LDPC Decoding on GPU for Mobile Device

    Directory of Open Access Journals (Sweden)

    Yiqin Lu

    2016-01-01

    Full Text Available A flexible software LDPC decoder that exploits data parallelism for simultaneous multicode words decoding on the mobile device is proposed in this paper, supported by multithreading on OpenCL based graphics processing units. By dividing the check matrix into several parts to make full use of both the local memory and private memory on GPU and properly modify the code capacity each time, our implementation on a mobile phone shows throughputs above 100 Mbps and delay is less than 1.6 millisecond in decoding, which make high-speed communication like video calling possible. To realize efficient software LDPC decoding on the mobile device, the LDPC decoding feature on communication baseband chip should be replaced to save the cost and make it easier to upgrade decoder to be compatible with a variety of channel access schemes.

  3. Instanton-based techniques for analysis and reduction of error floors of LDPC codes

    International Nuclear Information System (INIS)

    Chertkov, Michael; Chilappagari, Shashi K.; Stepanov, Mikhail G.; Vasic, Bane

    2008-01-01

    We describe a family of instanton-based optimization methods developed recently for the analysis of the error floors of low-density parity-check (LDPC) codes. Instantons are the most probable configurations of the channel noise which result in decoding failures. We show that the general idea and the respective optimization technique are applicable broadly to a variety of channels, discrete or continuous, and variety of sub-optimal decoders. Specifically, we consider: iterative belief propagation (BP) decoders, Gallager type decoders, and linear programming (LP) decoders performing over the additive white Gaussian noise channel (AWGNC) and the binary symmetric channel (BSC). The instanton analysis suggests that the underlying topological structures of the most probable instanton of the same code but different channels and decoders are related to each other. Armed with this understanding of the graphical structure of the instanton and its relation to the decoding failures, we suggest a method to construct codes whose Tanner graphs are free of these structures, and thus have less significant error floors.

  4. Instanton-based techniques for analysis and reduction of error floor of LDPC codes

    Energy Technology Data Exchange (ETDEWEB)

    Chertkov, Michael [Los Alamos National Laboratory; Chilappagari, Shashi K [Los Alamos National Laboratory; Stepanov, Mikhail G [Los Alamos National Laboratory; Vasic, Bane [SENIOR MEMBER, IEEE

    2008-01-01

    We describe a family of instanton-based optimization methods developed recently for the analysis of the error floors of low-density parity-check (LDPC) codes. Instantons are the most probable configurations of the channel noise which result in decoding failures. We show that the general idea and the respective optimization technique are applicable broadly to a variety of channels, discrete or continuous, and variety of sub-optimal decoders. Specifically, we consider: iterative belief propagation (BP) decoders, Gallager type decoders, and linear programming (LP) decoders performing over the additive white Gaussian noise channel (AWGNC) and the binary symmetric channel (BSC). The instanton analysis suggests that the underlying topological structures of the most probable instanton of the same code but different channels and decoders are related to each other. Armed with this understanding of the graphical structure of the instanton and its relation to the decoding failures, we suggest a method to construct codes whose Tanner graphs are free of these structures, and thus have less significant error floors.

  5. Adaptive transmission based on multi-relay selection and rate-compatible LDPC codes

    Science.gov (United States)

    Su, Hualing; He, Yucheng; Zhou, Lin

    2017-08-01

    In order to adapt to the dynamical changeable channel condition and improve the transmissive reliability of the system, a cooperation system of rate-compatible low density parity check (RC-LDPC) codes combining with multi-relay selection protocol is proposed. In traditional relay selection protocol, only the channel state information (CSI) of source-relay and the CSI of relay-destination has been considered. The multi-relay selection protocol proposed by this paper takes the CSI between relays into extra account in order to obtain more chances of collabration. Additionally, the idea of hybrid automatic request retransmission (HARQ) and rate-compatible are introduced. Simulation results show that the transmissive reliability of the system can be significantly improved by the proposed protocol.

  6. 428-Gb/s single-channel coherent optical OFDM transmission over 960-km SSMF with constellation expansion and LDPC coding.

    Science.gov (United States)

    Yang, Qi; Al Amin, Abdullah; Chen, Xi; Ma, Yiran; Chen, Simin; Shieh, William

    2010-08-02

    High-order modulation formats and advanced error correcting codes (ECC) are two promising techniques for improving the performance of ultrahigh-speed optical transport networks. In this paper, we present record receiver sensitivity for 107 Gb/s CO-OFDM transmission via constellation expansion to 16-QAM and rate-1/2 LDPC coding. We also show the single-channel transmission of a 428-Gb/s CO-OFDM signal over 960-km standard-single-mode-fiber (SSMF) without Raman amplification.

  7. LDPC Code Design for Nonuniform Power-Line Channels

    Directory of Open Access Journals (Sweden)

    Sanaei Ali

    2007-01-01

    Full Text Available We investigate low-density parity-check code design for discrete multitone channels over power lines. Discrete multitone channels are well modeled as nonuniform channels, that is, different bits experience various channel parameters. We propose a coding system for discrete multitone channels that allows for using a single code over a nonuniform channel. The number of code parameters for the proposed system is much greater than the number of code parameters in conventional channel. Therefore, search-based optimization methods are impractical. We first formulate the problem of optimizing the rate of an irregular low-density parity-check code, with guaranteed convergence over a general nonuniform channel, as an iterative linear programming which is significantly more efficient than search-based methods. Then we use this technique for a typical power-line channel. The methodology of this paper is directly applicable to all decoding algorithms for which a density evolution analysis is possible.

  8. Recursive construction of (J,L (J,L QC LDPC codes with girth 6

    Directory of Open Access Journals (Sweden)

    Mohammad Gholami

    2016-06-01

    Full Text Available ‎In this paper‎, ‎a recursive algorithm is presented to generate some exponent matrices which correspond to Tanner graphs with girth at least 6‎. ‎For a J×L J×L exponent matrix E E‎, ‎the lower bound Q(E Q(E is obtained explicitly such that (J,L (J,L QC LDPC codes with girth at least 6 exist for any circulant permutation matrix (CPM size m≥Q(E m≥Q(E‎. ‎The results show that the exponent matrices constructed with our recursive algorithm have smaller lower-bound than the ones proposed recently with girth 6‎

  9. Near-Capacity Coding for Discrete Multitone Systems with Impulse Noise

    Directory of Open Access Journals (Sweden)

    Kschischang Frank R

    2006-01-01

    Full Text Available We consider the design of near-capacity-achieving error-correcting codes for a discrete multitone (DMT system in the presence of both additive white Gaussian noise and impulse noise. Impulse noise is one of the main channel impairments for digital subscriber lines (DSL. One way to combat impulse noise is to detect the presence of the impulses and to declare an erasure when an impulse occurs. In this paper, we propose a coding system based on low-density parity-check (LDPC codes and bit-interleaved coded modulation that is capable of taking advantage of the knowledge of erasures. We show that by carefully choosing the degree distribution of an irregular LDPC code, both the additive noise and the erasures can be handled by a single code, thus eliminating the need for an outer code. Such a system can perform close to the capacity of the channel and for the same redundancy is significantly more immune to the impulse noise than existing methods based on an outer Reed-Solomon (RS code. The proposed method has a lower implementation complexity than the concatenated coding approach.

  10. MIMO-OFDM System's Performance Using LDPC Codes for a Mobile Robot

    Science.gov (United States)

    Daoud, Omar; Alani, Omar

    This work deals with the performance of a Sniffer Mobile Robot (SNFRbot)-based spatial multiplexed wireless Orthogonal Frequency Division Multiplexing (OFDM) transmission technology. The use of Multi-Input Multi-Output (MIMO)-OFDM technology increases the wireless transmission rate without increasing transmission power or bandwidth. A generic multilayer architecture of the SNFRbot is proposed with low power and low cost. Some experimental results are presented and show the efficiency of sniffing deadly gazes, sensing high temperatures and sending live videos of the monitored situation. Moreover, simulation results show the achieved performance by tackling the Peak-to-Average Power Ratio (PAPR) problem of the used technology using Low Density Parity Check (LDPC) codes; and the effect of combating the PAPR on the bit error rate (BER) and the signal to noise ratio (SNR) over a Doppler spread channel.

  11. High performance reconciliation for continuous-variable quantum key distribution with LDPC code

    Science.gov (United States)

    Lin, Dakai; Huang, Duan; Huang, Peng; Peng, Jinye; Zeng, Guihua

    2015-03-01

    Reconciliation is a significant procedure in a continuous-variable quantum key distribution (CV-QKD) system. It is employed to extract secure secret key from the resulted string through quantum channel between two users. However, the efficiency and the speed of previous reconciliation algorithms are low. These problems limit the secure communication distance and the secure key rate of CV-QKD systems. In this paper, we proposed a high-speed reconciliation algorithm through employing a well-structured decoding scheme based on low density parity-check (LDPC) code. The complexity of the proposed algorithm is reduced obviously. By using a graphics processing unit (GPU) device, our method may reach a reconciliation speed of 25 Mb/s for a CV-QKD system, which is currently the highest level and paves the way to high-speed CV-QKD.

  12. LDPC and SHA based iris recognition for image authentication

    Directory of Open Access Journals (Sweden)

    K. Seetharaman

    2012-11-01

    Full Text Available We introduce a novel way to authenticate an image using Low Density Parity Check (LDPC and Secure Hash Algorithm (SHA based iris recognition method with reversible watermarking scheme, which is based on Integer Wavelet Transform (IWT and threshold embedding technique. The parity checks and parity matrix of LDPC encoding and cancellable biometrics i.e., hash string of unique iris code from SHA-512 are embedded into an image for authentication purpose using reversible watermarking scheme based on IWT and threshold embedding technique. Simply by reversing the embedding process, the original image, parity checks, parity matrix and SHA-512 hash are extracted back from watermarked-image. For authentication, the new hash string produced by employing SHA-512 on error corrected iris code from live person is compared with hash string extracted from watermarked-image. The LDPC code reduces the hamming distance for genuine comparisons by a larger amount than for the impostor comparisons. This results in better separation between genuine and impostor users which improves the authentication performance. Security of this scheme is very high due to the security complexity of SHA-512, which is 2256 under birthday attack. Experimental results show that this approach can assure more accurate authentication with a low false rejection or false acceptance rate and outperforms the prior arts in terms of PSNR.

  13. Resource Efficient LDPC Decoders for Multimedia Communication

    OpenAIRE

    Chandrasetty, Vikram Arkalgud; Aziz, Syed Mahfuzul

    2013-01-01

    Achieving high image quality is an important aspect in an increasing number of wireless multimedia applications. These applications require resource efficient error correction hardware to detect and correct errors introduced by the communication channel. This paper presents an innovative flexible architecture for error correction using Low-Density Parity-Check (LDPC) codes. The proposed partially-parallel decoder architecture utilizes a novel code construction technique based on multi-level H...

  14. A modified non-binary LDPC scheme based on watermark symbols in high speed optical transmission systems

    Science.gov (United States)

    Wang, Liming; Qiao, Yaojun; Yu, Qian; Zhang, Wenbo

    2016-04-01

    We introduce a watermark non-binary low-density parity check code (NB-LDPC) scheme, which can estimate the time-varying noise variance by using prior information of watermark symbols, to improve the performance of NB-LDPC codes. And compared with the prior-art counterpart, the watermark scheme can bring about 0.25 dB improvement in net coding gain (NCG) at bit error rate (BER) of 1e-6 and 36.8-81% reduction of the iteration numbers. Obviously, the proposed scheme shows great potential in terms of error correction performance and decoding efficiency.

  15. Efficacy analysis of LDPC coded APSK modulated differential space-time-frequency coded for wireless body area network using MB-pulsed OFDM UWB technology.

    Science.gov (United States)

    Manimegalai, C T; Gauni, Sabitha; Kalimuthu, K

    2017-12-04

    Wireless body area network (WBAN) is a breakthrough technology in healthcare areas such as hospital and telemedicine. The human body has a complex mixture of different tissues. It is expected that the nature of propagation of electromagnetic signals is distinct in each of these tissues. This forms the base for the WBAN, which is different from other environments. In this paper, the knowledge of Ultra Wide Band (UWB) channel is explored in the WBAN (IEEE 802.15.6) system. The measurements of parameters in frequency range from 3.1-10.6 GHz are taken. The proposed system, transmits data up to 480 Mbps by using LDPC coded APSK Modulated Differential Space-Time-Frequency Coded MB-OFDM to increase the throughput and power efficiency.

  16. 45 Gb/s low complexity optical front-end for soft-decision LDPC decoders.

    Science.gov (United States)

    Sakib, Meer Nazmus; Moayedi, Monireh; Gross, Warren J; Liboiron-Ladouceur, Odile

    2012-07-30

    In this paper a low complexity and energy efficient 45 Gb/s soft-decision optical front-end to be used with soft-decision low-density parity-check (LDPC) decoders is demonstrated. The results show that the optical front-end exhibits a net coding gain of 7.06 and 9.62 dB for post forward error correction bit error rate of 10(-7) and 10(-12) for long block length LDPC(32768,26803) code. The performance over a hard decision front-end is 1.9 dB for this code. It is shown that the soft-decision circuit can also be used as a 2-bit flash type analog-to-digital converter (ADC), in conjunction with equalization schemes. At bit rate of 15 Gb/s using RS(255,239), LDPC(672,336), (672, 504), (672, 588), and (1440, 1344) used with a 6-tap finite impulse response (FIR) equalizer will result in optical power savings of 3, 5, 7, 9.5 and 10.5 dB, respectively. The 2-bit flash ADC consumes only 2.71 W at 32 GSamples/s. At 45 GSamples/s the power consumption is estimated to be 4.95 W.

  17. An Area-Efficient Reconfigurable LDPC Decoder with Conflict Resolution

    Science.gov (United States)

    Zhou, Changsheng; Huang, Yuebin; Huang, Shuangqu; Chen, Yun; Zeng, Xiaoyang

    Based on Turbo-Decoding Message-Passing (TDMP) and Normalized Min-Sum (NMS) algorithm, an area efficient LDPC decoder that supports both structured and unstructured LDPC codes is proposed in this paper. We introduce a solution to solve the memory access conflict problem caused by TDMP algorithm. We also arrange the main timing schedule carefully to handle the operations of our solution while avoiding much additional hardware consumption. To reduce the memory bits needed, the extrinsic message storing strategy is also optimized. Besides the extrinsic message recover and the accumulate operation are merged together. To verify our architecture, a LDPC decoder that supports both China Multimedia Mobile Broadcasting (CMMB) and Digital Terrestrial/ Television Multimedia Broadcasting (DTMB) standards is developed using SMIC 0.13µm standard CMOS process. The core area is 4.75mm2 and the maximum operating clock frequency is 200MHz. The estimated power consumption is 48.4mW at 25MHz for CMMB and 130.9mW at 50MHz for DTMB with 5 iterations and 1.2V supply.

  18. Coded Cooperation for Multiway Relaying in Wireless Sensor Networks.

    Science.gov (United States)

    Si, Zhongwei; Ma, Junyang; Thobaben, Ragnar

    2015-06-29

    Wireless sensor networks have been considered as an enabling technology for constructing smart cities. One important feature of wireless sensor networks is that the sensor nodes collaborate in some manner for communications. In this manuscript, we focus on the model of multiway relaying with full data exchange where each user wants to transmit and receive data to and from all other users in the network. We derive the capacity region for this specific model and propose a coding strategy through coset encoding. To obtain good performance with practical codes, we choose spatially-coupled LDPC (SC-LDPC) codes for the coded cooperation. In particular, for the message broadcasting from the relay, we construct multi-edge-type (MET) SC-LDPC codes by repeatedly applying coset encoding. Due to the capacity-achieving property of the SC-LDPC codes, we prove that the capacity region can theoretically be achieved by the proposed MET SC-LDPC codes. Numerical results with finite node degrees are provided, which show that the achievable rates approach the boundary of the capacity region in both binary erasure channels and additive white Gaussian channels.

  19. Spatially coupled low-density parity-check error correction for holographic data storage

    Science.gov (United States)

    Ishii, Norihiko; Katano, Yutaro; Muroi, Tetsuhiko; Kinoshita, Nobuhiro

    2017-09-01

    The spatially coupled low-density parity-check (SC-LDPC) was considered for holographic data storage. The superiority of SC-LDPC was studied by simulation. The simulations show that the performance of SC-LDPC depends on the lifting number, and when the lifting number is over 100, SC-LDPC shows better error correctability compared with irregular LDPC. SC-LDPC is applied to the 5:9 modulation code, which is one of the differential codes. The error-free point is near 2.8 dB and over 10-1 can be corrected in simulation. From these simulation results, this error correction code can be applied to actual holographic data storage test equipment. Results showed that 8 × 10-2 can be corrected, furthermore it works effectively and shows good error correctability.

  20. Entanglement-assisted quantum quasicyclic low-density parity-check codes

    Science.gov (United States)

    Hsieh, Min-Hsiu; Brun, Todd A.; Devetak, Igor

    2009-03-01

    We investigate the construction of quantum low-density parity-check (LDPC) codes from classical quasicyclic (QC) LDPC codes with girth greater than or equal to 6. We have shown that the classical codes in the generalized Calderbank-Skor-Steane construction do not need to satisfy the dual-containing property as long as preshared entanglement is available to both sender and receiver. We can use this to avoid the many four cycles which typically arise in dual-containing LDPC codes. The advantage of such quantum codes comes from the use of efficient decoding algorithms such as sum-product algorithm (SPA). It is well known that in the SPA, cycles of length 4 make successive decoding iterations highly correlated and hence limit the decoding performance. We show the principle of constructing quantum QC-LDPC codes which require only small amounts of initial shared entanglement.

  1. Coded Cooperation for Multiway Relaying in Wireless Sensor Networks

    Directory of Open Access Journals (Sweden)

    Zhongwei Si

    2015-06-01

    Full Text Available Wireless sensor networks have been considered as an enabling technology for constructing smart cities. One important feature of wireless sensor networks is that the sensor nodes collaborate in some manner for communications. In this manuscript, we focus on the model of multiway relaying with full data exchange where each user wants to transmit and receive data to and from all other users in the network. We derive the capacity region for this specific model and propose a coding strategy through coset encoding. To obtain good performance with practical codes, we choose spatially-coupled LDPC (SC-LDPC codes for the coded cooperation. In particular, for the message broadcasting from the relay, we construct multi-edge-type (MET SC-LDPC codes by repeatedly applying coset encoding. Due to the capacity-achieving property of the SC-LDPC codes, we prove that the capacity region can theoretically be achieved by the proposed MET SC-LDPC codes. Numerical results with finite node degrees are provided, which show that the achievable rates approach the boundary of the capacity region in both binary erasure channels and additive white Gaussian channels.

  2. On the reduced-complexity of LDPC decoders for beyond 400 Gb/s serial optical transmission

    Science.gov (United States)

    Djordjevic, Ivan B.; Xu, Lei; Wang, Ting

    2010-12-01

    Two reduced-complexity (RC) LDPC decoders are proposed, which can be used in combination with large-girth LDPC codes to enable beyond 400 Gb/s serial optical transmission. We show that optimally attenuated RC min-sum sum algorithm performs only 0.45 dB worse than conventional sum-product algorithm, while having lower storage memory requirements and much lower latency. We further evaluate the proposed algorithms for use in beyond 400 Gb/s serial optical transmission in combination with PolMUX 32-IPQ-based signal constellation and show that low BERs can be achieved for medium optical SNRs, while achieving the net coding gain above 11.4 dB.

  3. Cooperative MIMO Communication at Wireless Sensor Network: An Error Correcting Code Approach

    Science.gov (United States)

    Islam, Mohammad Rakibul; Han, Young Shin

    2011-01-01

    Cooperative communication in wireless sensor network (WSN) explores the energy efficient wireless communication schemes between multiple sensors and data gathering node (DGN) by exploiting multiple input multiple output (MIMO) and multiple input single output (MISO) configurations. In this paper, an energy efficient cooperative MIMO (C-MIMO) technique is proposed where low density parity check (LDPC) code is used as an error correcting code. The rate of LDPC code is varied by varying the length of message and parity bits. Simulation results show that the cooperative communication scheme outperforms SISO scheme in the presence of LDPC code. LDPC codes with different code rates are compared using bit error rate (BER) analysis. BER is also analyzed under different Nakagami fading scenario. Energy efficiencies are compared for different targeted probability of bit error pb. It is observed that C-MIMO performs more efficiently when the targeted pb is smaller. Also the lower encoding rate for LDPC code offers better error characteristics. PMID:22163732

  4. Cooperative MIMO communication at wireless sensor network: an error correcting code approach.

    Science.gov (United States)

    Islam, Mohammad Rakibul; Han, Young Shin

    2011-01-01

    Cooperative communication in wireless sensor network (WSN) explores the energy efficient wireless communication schemes between multiple sensors and data gathering node (DGN) by exploiting multiple input multiple output (MIMO) and multiple input single output (MISO) configurations. In this paper, an energy efficient cooperative MIMO (C-MIMO) technique is proposed where low density parity check (LDPC) code is used as an error correcting code. The rate of LDPC code is varied by varying the length of message and parity bits. Simulation results show that the cooperative communication scheme outperforms SISO scheme in the presence of LDPC code. LDPC codes with different code rates are compared using bit error rate (BER) analysis. BER is also analyzed under different Nakagami fading scenario. Energy efficiencies are compared for different targeted probability of bit error p(b). It is observed that C-MIMO performs more efficiently when the targeted p(b) is smaller. Also the lower encoding rate for LDPC code offers better error characteristics.

  5. Optimized Min-Sum Decoding Algorithm for Low Density Parity Check Codes

    OpenAIRE

    Mohammad Rakibul Islam; Dewan Siam Shafiullah; Muhammad Mostafa Amir Faisal; Imran Rahman

    2011-01-01

    Low Density Parity Check (LDPC) code approaches Shannon–limit performance for binary field and long code lengths. However, performance of binary LDPC code is degraded when the code word length is small. An optimized min-sum algorithm for LDPC code is proposed in this paper. In this algorithm unlike other decoding methods, an optimization factor has been introduced in both check node and bit node of the Min-sum algorithm. The optimization factor is obtained before decoding program, and the sam...

  6. Design and performance investigation of LDPC-coded upstream transmission systems in IM/DD OFDM-PONs

    Science.gov (United States)

    Gong, Xiaoxue; Guo, Lei; Wu, Jingjing; Ning, Zhaolong

    2016-12-01

    In Intensity-Modulation Direct-Detection (IM/DD) Orthogonal Frequency Division Multiplexing Passive Optical Networks (OFDM-PONs), aside from Subcarrier-to-Subcarrier Intermixing Interferences (SSII) induced by square-law detection, the same laser frequency for data sending from Optical Network Units (ONUs) results in ONU-to-ONU Beating Interferences (OOBI) at the receiver. To mitigate those interferences, we design a Low-Density Parity Check (LDPC)-coded and spectrum-efficient upstream transmission system. A theoretical channel model is also derived, in order to analyze the detrimental factors influencing system performances. Simulation results demonstrate that the receiver sensitivity is improved 3.4 dB and 2.5 dB under QPSK and 8QAM, respectively, after 100 km Standard Single-Mode Fiber (SSMF) transmission. Furthermore, the spectrum efficiency can be improved by about 50%.

  7. Blind Estimation of the Phase and Carrier Frequency Offsets for LDPC-Coded Systems

    Directory of Open Access Journals (Sweden)

    Houcke Sebastien

    2010-01-01

    Full Text Available Abstract We consider in this paper the problem of phase offset and Carrier Frequency Offset (CFO estimation for Low-Density Parity-Check (LDPC coded systems. We propose new blind estimation techniques based on the calculation and minimization of functions of the Log-Likelihood Ratios (LLR of the syndrome elements obtained according to the parity check matrix of the error-correcting code. In the first part of this paper, we consider phase offset estimation for a Binary Phase Shift Keying (BPSK modulation and propose a novel estimation technique. Simulation results show that the proposed method is very effective and outperforms many existing algorithms. Then, we modify the estimation criterion so that it can work for higher-order modulations. One interesting feature of the proposed algorithm when applied to high-order modulations is that the phase offset of the channel can be blindly estimated without any ambiguity. In the second part of the paper, we consider the problem of CFO estimation and propose estimation techniques that are based on the same concept as the ones presented for the phase offset estimation. The Mean Squared Error (MSE and Bit Error Rate (BER curves show the efficiency of the proposed estimation techniques.

  8. Polynomial theory of error correcting codes

    CERN Document Server

    Cancellieri, Giovanni

    2015-01-01

    The book offers an original view on channel coding, based on a unitary approach to block and convolutional codes for error correction. It presents both new concepts and new families of codes. For example, lengthened and modified lengthened cyclic codes are introduced as a bridge towards time-invariant convolutional codes and their extension to time-varying versions. The novel families of codes include turbo codes and low-density parity check (LDPC) codes, the features of which are justified from the structural properties of the component codes. Design procedures for regular LDPC codes are proposed, supported by the presented theory. Quasi-cyclic LDPC codes, in block or convolutional form, represent one of the most original contributions of the book. The use of more than 100 examples allows the reader gradually to gain an understanding of the theory, and the provision of a list of more than 150 definitions, indexed at the end of the book, permits rapid location of sought information.

  9. High-efficiency Gaussian key reconciliation in continuous variable quantum key distribution

    Science.gov (United States)

    Bai, ZengLiang; Wang, XuYang; Yang, ShenShen; Li, YongMin

    2016-01-01

    Efficient reconciliation is a crucial step in continuous variable quantum key distribution. The progressive-edge-growth (PEG) algorithm is an efficient method to construct relatively short block length low-density parity-check (LDPC) codes. The qua-sicyclic construction method can extend short block length codes and further eliminate the shortest cycle. In this paper, by combining the PEG algorithm and qua-si-cyclic construction method, we design long block length irregular LDPC codes with high error-correcting capacity. Based on these LDPC codes, we achieve high-efficiency Gaussian key reconciliation with slice recon-ciliation based on multilevel coding/multistage decoding with an efficiency of 93.7%.

  10. Coded Cooperation for Multiway Relaying in Wireless Sensor Networks †

    Science.gov (United States)

    Si, Zhongwei; Ma, Junyang; Thobaben, Ragnar

    2015-01-01

    Wireless sensor networks have been considered as an enabling technology for constructing smart cities. One important feature of wireless sensor networks is that the sensor nodes collaborate in some manner for communications. In this manuscript, we focus on the model of multiway relaying with full data exchange where each user wants to transmit and receive data to and from all other users in the network. We derive the capacity region for this specific model and propose a coding strategy through coset encoding. To obtain good performance with practical codes, we choose spatially-coupled LDPC (SC-LDPC) codes for the coded cooperation. In particular, for the message broadcasting from the relay, we construct multi-edge-type (MET) SC-LDPC codes by repeatedly applying coset encoding. Due to the capacity-achieving property of the SC-LDPC codes, we prove that the capacity region can theoretically be achieved by the proposed MET SC-LDPC codes. Numerical results with finite node degrees are provided, which show that the achievable rates approach the boundary of the capacity region in both binary erasure channels and additive white Gaussian channels. PMID:26131675

  11. Parallel Subspace Subcodes of Reed-Solomon Codes for Magnetic Recording Channels

    Science.gov (United States)

    Wang, Han

    2010-01-01

    Read channel architectures based on a single low-density parity-check (LDPC) code are being considered for the next generation of hard disk drives. However, LDPC-only solutions suffer from the error floor problem, which may compromise reliability, if not handled properly. Concatenated architectures using an LDPC code plus a Reed-Solomon (RS) code…

  12. LDPC-based iterative joint source-channel decoding for JPEG2000.

    Science.gov (United States)

    Pu, Lingling; Wu, Zhenyu; Bilgin, Ali; Marcellin, Michael W; Vasic, Bane

    2007-02-01

    A framework is proposed for iterative joint source-channel decoding of JPEG2000 codestreams. At the encoder, JPEG2000 is used to perform source coding with certain error-resilience (ER) modes, and LDPC codes are used to perform channel coding. During decoding, the source decoder uses the ER modes to identify corrupt sections of the codestream and provides this information to the channel decoder. Decoding is carried out jointly in an iterative fashion. Experimental results indicate that the proposed method requires fewer iterations and improves overall system performance.

  13. Entanglement-assisted quantum low-density parity-check codes

    International Nuclear Information System (INIS)

    Fujiwara, Yuichiro; Clark, David; Tonchev, Vladimir D.; Vandendriessche, Peter; De Boeck, Maarten

    2010-01-01

    This article develops a general method for constructing entanglement-assisted quantum low-density parity-check (LDPC) codes, which is based on combinatorial design theory. Explicit constructions are given for entanglement-assisted quantum error-correcting codes with many desirable properties. These properties include the requirement of only one initial entanglement bit, high error-correction performance, high rates, and low decoding complexity. The proposed method produces several infinite families of codes with a wide variety of parameters and entanglement requirements. Our framework encompasses the previously known entanglement-assisted quantum LDPC codes having the best error-correction performance and many other codes with better block error rates in simulations over the depolarizing channel. We also determine important parameters of several well-known classes of quantum and classical LDPC codes for previously unsettled cases.

  14. Analysis of error floor of LDPC codes under LP decoding over the BSC

    Energy Technology Data Exchange (ETDEWEB)

    Chertkov, Michael [Los Alamos National Laboratory; Chilappagari, Shashi [UNIV OF AZ; Vasic, Bane [UNIV OF AZ; Stepanov, Mikhail [UNIV OF AZ

    2009-01-01

    We consider linear programming (LP) decoding of a fixed low-density parity-check (LDPC) code over the binary symmetric channel (BSC). The LP decoder fails when it outputs a pseudo-codeword which is not a codeword. We propose an efficient algorithm termed the instanton search algorithm (ISA) which, given a random input, generates a set of flips called the BSC-instanton and prove that: (a) the LP decoder fails for any set of flips with support vector including an instanton; (b) for any input, the algorithm outputs an instanton in the number of steps upper-bounded by twice the number of flips in the input. We obtain the number of unique instantons of different sizes by running the ISA sufficient number of times. We then use the instanton statistics to predict the performance of the LP decoding over the BSC in the error floor region. We also propose an efficient semi-analytical method to predict the performance of LP decoding over a large range of transition probabilities of the BSC.

  15. Crosstalk eliminating and low-density parity-check codes for photochromic dual-wavelength storage

    Science.gov (United States)

    Wang, Meicong; Xiong, Jianping; Jian, Jiqi; Jia, Huibo

    2005-01-01

    Multi-wavelength storage is an approach to increase the memory density with the problem of crosstalk to be deal with. We apply Low Density Parity Check (LDPC) codes as error-correcting codes in photochromic dual-wavelength optical storage based on the investigation of LDPC codes in optical data storage. A proper method is applied to reduce the crosstalk and simulation results show that this operation is useful to improve Bit Error Rate (BER) performance. At the same time we can conclude that LDPC codes outperform RS codes in crosstalk channel.

  16. An LDPC decoder architecture for wireless sensor network applications.

    Science.gov (United States)

    Biroli, Andrea Dario Giancarlo; Martina, Maurizio; Masera, Guido

    2012-01-01

    The pervasive use of wireless sensors in a growing spectrum of human activities reinforces the need for devices with low energy dissipation. In this work, coded communication between a couple of wireless sensor devices is considered as a method to reduce the dissipated energy per transmitted bit with respect to uncoded communication. Different Low Density Parity Check (LDPC) codes are considered to this purpose and post layout results are shown for a low-area low-energy decoder, which offers percentage energy savings with respect to the uncoded solution in the range of 40%-80%, depending on considered environment, distance and bit error rate.

  17. Advanced error-prediction LDPC with temperature compensation for highly reliable SSDs

    Science.gov (United States)

    Tokutomi, Tsukasa; Tanakamaru, Shuhei; Iwasaki, Tomoko Ogura; Takeuchi, Ken

    2015-09-01

    To improve the reliability of NAND Flash memory based solid-state drives (SSDs), error-prediction LDPC (EP-LDPC) has been proposed for multi-level-cell (MLC) NAND Flash memory (Tanakamaru et al., 2012, 2013), which is effective for long retention times. However, EP-LDPC is not as effective for triple-level cell (TLC) NAND Flash memory, because TLC NAND Flash has higher error rates and is more sensitive to program-disturb error. Therefore, advanced error-prediction LDPC (AEP-LDPC) has been proposed for TLC NAND Flash memory (Tokutomi et al., 2014). AEP-LDPC can correct errors more accurately by precisely describing the error phenomena. In this paper, the effects of AEP-LDPC are investigated in a 2×nm TLC NAND Flash memory with temperature characterization. Compared with LDPC-with-BER-only, the SSD's data-retention time is increased by 3.4× and 9.5× at room-temperature (RT) and 85 °C, respectively. Similarly, the acceptable BER is increased by 1.8× and 2.3×, respectively. Moreover, AEP-LDPC can correct errors with pre-determined tables made at higher temperatures to shorten the measurement time before shipping. Furthermore, it is found that one table can cover behavior over a range of temperatures in AEP-LDPC. As a result, the total table size can be reduced to 777 kBytes, which makes this approach more practical.

  18. Structured Low-Density Parity-Check Codes with Bandwidth Efficient Modulation

    Science.gov (United States)

    Cheng, Michael K.; Divsalar, Dariush; Duy, Stephanie

    2009-01-01

    In this work, we study the performance of structured Low-Density Parity-Check (LDPC) Codes together with bandwidth efficient modulations. We consider protograph-based LDPC codes that facilitate high-speed hardware implementations and have minimum distances that grow linearly with block sizes. We cover various higher- order modulations such as 8-PSK, 16-APSK, and 16-QAM. During demodulation, a demapper transforms the received in-phase and quadrature samples into reliability information that feeds the binary LDPC decoder. We will compare various low-complexity demappers and provide simulation results for assorted coded-modulation combinations on the additive white Gaussian noise and independent Rayleigh fading channels.

  19. Joint Carrier-Phase Synchronization and LDPC Decoding

    Science.gov (United States)

    Simon, Marvin; Valles, Esteban

    2009-01-01

    A method has been proposed to increase the degree of synchronization of a radio receiver with the phase of a suppressed carrier signal modulated with a binary- phase-shift-keying (BPSK) or quaternary- phase-shift-keying (QPSK) signal representing a low-density parity-check (LDPC) code. This method is an extended version of the method described in Using LDPC Code Constraints to Aid Recovery of Symbol Timing (NPO-43112), NASA Tech Briefs, Vol. 32, No. 10 (October 2008), page 54. Both methods and the receiver architectures in which they would be implemented belong to a class of timing- recovery methods and corresponding receiver architectures characterized as pilotless in that they do not require transmission and reception of pilot signals. The proposed method calls for the use of what is known in the art as soft decision feedback to remove the modulation from a replica of the incoming signal prior to feeding this replica to a phase-locked loop (PLL) or other carrier-tracking stage in the receiver. Soft decision feedback refers to suitably processed versions of intermediate results of iterative computations involved in the LDPC decoding process. Unlike a related prior method in which hard decision feedback (the final sequence of decoded symbols) is used to remove the modulation, the proposed method does not require estimation of the decoder error probability. In a basic digital implementation of the proposed method, the incoming signal (having carrier phase theta theta (sub c) plus noise would first be converted to inphase (I) and quadrature (Q) baseband signals by mixing it with I and Q signals at the carrier frequency [wc/(2 pi)] generated by a local oscillator. The resulting demodulated signals would be processed through one-symbol-period integrate and- dump filters, the outputs of which would be sampled and held, then multiplied by a soft-decision version of the baseband modulated signal. The resulting I and Q products consist of terms proportional to the cosine

  20. Quasi Cyclic Low Density Parity Check Code for High SNR Data Transfer

    Directory of Open Access Journals (Sweden)

    M. R. Islam

    2010-06-01

    Full Text Available An improved Quasi Cyclic Low Density Parity Check code (QC-LDPC is proposed to reduce the complexity of the Low Density Parity Check code (LDPC while obtaining the similar performance. The proposed QC-LDPC presents an improved construction at high SNR with circulant sub-matrices. The proposed construction yields a performance gain of about 1 dB at a 0.0003 bit error rate (BER and it is tested on 4 different decoding algorithms. Proposed QC-LDPC is compared with the existing QC-LDPC and the simulation results show that the proposed approach outperforms the existing one at high SNR. Simulations are also performed varying the number of horizontal sub matrices and the results show that the parity check matrix with smaller horizontal concatenation shows better performance.

  1. An FPGA Implementation of (3,6-Regular Low-Density Parity-Check Code Decoder

    Directory of Open Access Journals (Sweden)

    Tong Zhang

    2003-05-01

    Full Text Available Because of their excellent error-correcting performance, low-density parity-check (LDPC codes have recently attracted a lot of attention. In this paper, we are interested in the practical LDPC code decoder hardware implementations. The direct fully parallel decoder implementation usually incurs too high hardware complexity for many real applications, thus partly parallel decoder design approaches that can achieve appropriate trade-offs between hardware complexity and decoding throughput are highly desirable. Applying a joint code and decoder design methodology, we develop a high-speed (3,k-regular LDPC code partly parallel decoder architecture based on which we implement a 9216-bit, rate-1/2(3,6-regular LDPC code decoder on Xilinx FPGA device. This partly parallel decoder supports a maximum symbol throughput of 54 Mbps and achieves BER 10−6 at 2 dB over AWGN channel while performing maximum 18 decoding iterations.

  2. Efficient decoding of random errors for quantum expander codes

    OpenAIRE

    Fawzi , Omar; Grospellier , Antoine; Leverrier , Anthony

    2017-01-01

    We show that quantum expander codes, a constant-rate family of quantum LDPC codes, with the quasi-linear time decoding algorithm of Leverrier, Tillich and Z\\'emor can correct a constant fraction of random errors with very high probability. This is the first construction of a constant-rate quantum LDPC code with an efficient decoding algorithm that can correct a linear number of random errors with a negligible failure probability. Finding codes with these properties is also motivated by Gottes...

  3. An LDPC Decoder Architecture for Wireless Sensor Network Applications

    Science.gov (United States)

    Giancarlo Biroli, Andrea Dario; Martina, Maurizio; Masera, Guido

    2012-01-01

    The pervasive use of wireless sensors in a growing spectrum of human activities reinforces the need for devices with low energy dissipation. In this work, coded communication between a couple of wireless sensor devices is considered as a method to reduce the dissipated energy per transmitted bit with respect to uncoded communication. Different Low Density Parity Check (LDPC) codes are considered to this purpose and post layout results are shown for a low-area low-energy decoder, which offers percentage energy savings with respect to the uncoded solution in the range of 40%–80%, depending on considered environment, distance and bit error rate. PMID:22438724

  4. A Low-Complexity Joint Detection-Decoding Algorithm for Nonbinary LDPC-Coded Modulation Systems

    OpenAIRE

    Wang, Xuepeng; Bai, Baoming; Ma, Xiao

    2010-01-01

    In this paper, we present a low-complexity joint detection-decoding algorithm for nonbinary LDPC codedmodulation systems. The algorithm combines hard-decision decoding using the message-passing strategy with the signal detector in an iterative manner. It requires low computational complexity, offers good system performance and has a fast rate of decoding convergence. Compared to the q-ary sum-product algorithm (QSPA), it provides an attractive candidate for practical applications of q-ary LDP...

  5. Study of bifurcation behavior of two-dimensional turbo product code decoders

    International Nuclear Information System (INIS)

    He Yejun; Lau, Francis C.M.; Tse, Chi K.

    2008-01-01

    Turbo codes, low-density parity-check (LDPC) codes and turbo product codes (TPCs) are high performance error-correction codes which employ iterative algorithms for decoding. Under different conditions, the behaviors of the decoders are different. While the nonlinear dynamical behaviors of turbo code decoders and LDPC decoders have been reported in the literature, the dynamical behavior of TPC decoders is relatively unexplored. In this paper, we investigate the behavior of the iterative algorithm of a two-dimensional TPC decoder when the input signal-to-noise ratio (SNR) varies. The quantity to be measured is the mean square value of the posterior probabilities of the information bits. Unlike turbo decoders or LDPC decoders, TPC decoders do not produce a clear 'waterfall region'. This is mainly because the TPC decoding algorithm does not converge to 'indecisive' fixed points even at very low SNR values

  6. Channel coding techniques for wireless communications

    CERN Document Server

    Deergha Rao, K

    2015-01-01

    The book discusses modern channel coding techniques for wireless communications such as turbo codes, low-density parity check (LDPC) codes, space–time (ST) coding, RS (or Reed–Solomon) codes and convolutional codes. Many illustrative examples are included in each chapter for easy understanding of the coding techniques. The text is integrated with MATLAB-based programs to enhance the understanding of the subject’s underlying theories. It includes current topics of increasing importance such as turbo codes, LDPC codes, Luby transform (LT) codes, Raptor codes, and ST coding in detail, in addition to the traditional codes such as cyclic codes, BCH (or Bose–Chaudhuri–Hocquenghem) and RS codes and convolutional codes. Multiple-input and multiple-output (MIMO) communications is a multiple antenna technology, which is an effective method for high-speed or high-reliability wireless communications. PC-based MATLAB m-files for the illustrative examples are provided on the book page on Springer.com for free dow...

  7. Quantum quasi-cyclic low-density parity-check error-correcting codes

    International Nuclear Information System (INIS)

    Yuan, Li; Gui-Hua, Zeng; Lee, Moon Ho

    2009-01-01

    In this paper, we propose the approach of employing circulant permutation matrices to construct quantum quasicyclic (QC) low-density parity-check (LDPC) codes. Using the proposed approach one may construct some new quantum codes with various lengths and rates of no cycles-length 4 in their Tanner graphs. In addition, these constructed codes have the advantages of simple implementation and low-complexity encoding. Finally, the decoding approach for the proposed quantum QC LDPC is investigated. (general)

  8. Soft-Decision-Data Reshuffle to Mitigate Pulsed Radio Frequency Interference Impact on Low-Density-Parity-Check Code Performance

    Science.gov (United States)

    Ni, Jianjun David

    2011-01-01

    This presentation briefly discusses a research effort on mitigation techniques of pulsed radio frequency interference (RFI) on a Low-Density-Parity-Check (LDPC) code. This problem is of considerable interest in the context of providing reliable communications to the space vehicle which might suffer severe degradation due to pulsed RFI sources such as large radars. The LDPC code is one of modern forward-error-correction (FEC) codes which have the decoding performance to approach the Shannon Limit. The LDPC code studied here is the AR4JA (2048, 1024) code recommended by the Consultative Committee for Space Data Systems (CCSDS) and it has been chosen for some spacecraft design. Even though this code is designed as a powerful FEC code in the additive white Gaussian noise channel, simulation data and test results show that the performance of this LDPC decoder is severely degraded when exposed to the pulsed RFI specified in the spacecraft s transponder specifications. An analysis work (through modeling and simulation) has been conducted to evaluate the impact of the pulsed RFI and a few implemental techniques have been investigated to mitigate the pulsed RFI impact by reshuffling the soft-decision-data available at the input of the LDPC decoder. The simulation results show that the LDPC decoding performance of codeword error rate (CWER) under pulsed RFI can be improved up to four orders of magnitude through a simple soft-decision-data reshuffle scheme. This study reveals that an error floor of LDPC decoding performance appears around CWER=1E-4 when the proposed technique is applied to mitigate the pulsed RFI impact. The mechanism causing this error floor remains unknown, further investigation is necessary.

  9. Study of bifurcation behavior of two-dimensional turbo product code decoders

    Energy Technology Data Exchange (ETDEWEB)

    He Yejun [Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hunghom, Hong Kong (China); Lau, Francis C.M. [Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hunghom, Hong Kong (China)], E-mail: encmlau@polyu.edu.hk; Tse, Chi K. [Department of Electronic and Information Engineering, Hong Kong Polytechnic University, Hunghom, Hong Kong (China)

    2008-04-15

    Turbo codes, low-density parity-check (LDPC) codes and turbo product codes (TPCs) are high performance error-correction codes which employ iterative algorithms for decoding. Under different conditions, the behaviors of the decoders are different. While the nonlinear dynamical behaviors of turbo code decoders and LDPC decoders have been reported in the literature, the dynamical behavior of TPC decoders is relatively unexplored. In this paper, we investigate the behavior of the iterative algorithm of a two-dimensional TPC decoder when the input signal-to-noise ratio (SNR) varies. The quantity to be measured is the mean square value of the posterior probabilities of the information bits. Unlike turbo decoders or LDPC decoders, TPC decoders do not produce a clear 'waterfall region'. This is mainly because the TPC decoding algorithm does not converge to 'indecisive' fixed points even at very low SNR values.

  10. Channel coding for underwater acoustic single-carrier CDMA communication system

    Science.gov (United States)

    Liu, Lanjun; Zhang, Yonglei; Zhang, Pengcheng; Zhou, Lin; Niu, Jiong

    2017-01-01

    CDMA is an effective multiple access protocol for underwater acoustic networks, and channel coding can effectively reduce the bit error rate (BER) of the underwater acoustic communication system. For the requirements of underwater acoustic mobile networks based on CDMA, an underwater acoustic single-carrier CDMA communication system (UWA/SCCDMA) based on the direct-sequence spread spectrum is proposed, and its channel coding scheme is studied based on convolution, RA, Turbo and LDPC coding respectively. The implementation steps of the Viterbi algorithm of convolutional coding, BP and minimum sum algorithms of RA coding, Log-MAP and SOVA algorithms of Turbo coding, and sum-product algorithm of LDPC coding are given. An UWA/SCCDMA simulation system based on Matlab is designed. Simulation results show that the UWA/SCCDMA based on RA, Turbo and LDPC coding have good performance such that the communication BER is all less than 10-6 in the underwater acoustic channel with low signal to noise ratio (SNR) from -12 dB to -10dB, which is about 2 orders of magnitude lower than that of the convolutional coding. The system based on Turbo coding with Log-MAP algorithm has the best performance.

  11. A software reconfigurable optical multiband UWB system utilizing a bit-loading combined with adaptive LDPC code rate scheme

    Science.gov (United States)

    He, Jing; Dai, Min; Chen, Qinghui; Deng, Rui; Xiang, Changqing; Chen, Lin

    2017-07-01

    In this paper, an effective bit-loading combined with adaptive LDPC code rate algorithm is proposed and investigated in software reconfigurable multiband UWB over fiber system. To compensate the power fading and chromatic dispersion for the high frequency of multiband OFDM UWB signal transmission over standard single mode fiber (SSMF), a Mach-Zehnder modulator (MZM) with negative chirp parameter is utilized. In addition, the negative power penalty of -1 dB for 128 QAM multiband OFDM UWB signal are measured at the hard-decision forward error correction (HD-FEC) limitation of 3.8 × 10-3 after 50 km SSMF transmission. The experimental results show that, compared to the fixed coding scheme with the code rate of 75%, the signal-to-noise (SNR) is improved by 2.79 dB for 128 QAM multiband OFDM UWB system after 100 km SSMF transmission using ALCR algorithm. Moreover, by employing bit-loading combined with ALCR algorithm, the bit error rate (BER) performance of system can be further promoted effectively. The simulation results present that, at the HD-FEC limitation, the value of Q factor is improved by 3.93 dB at the SNR of 19.5 dB over 100 km SSMF transmission, compared to the fixed modulation with uncoded scheme at the same spectrum efficiency (SE).

  12. Fundamentals of convolutional coding

    CERN Document Server

    Johannesson, Rolf

    2015-01-01

    Fundamentals of Convolutional Coding, Second Edition, regarded as a bible of convolutional coding brings you a clear and comprehensive discussion of the basic principles of this field * Two new chapters on low-density parity-check (LDPC) convolutional codes and iterative coding * Viterbi, BCJR, BEAST, list, and sequential decoding of convolutional codes * Distance properties of convolutional codes * Includes a downloadable solutions manual

  13. Decoding Codes on Graphs

    Indian Academy of Sciences (India)

    Shannon limit of the channel. Among the earliest discovered codes that approach the. Shannon limit were the low density parity check (LDPC) codes. The term low density arises from the property of the parity check matrix defining the code. We will now define this matrix and the role that it plays in decoding. 2. Linear Codes.

  14. Statistical mechanics of low-density parity-check codes

    Energy Technology Data Exchange (ETDEWEB)

    Kabashima, Yoshiyuki [Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology, Yokohama 2268502 (Japan); Saad, David [Neural Computing Research Group, Aston University, Birmingham B4 7ET (United Kingdom)

    2004-02-13

    We review recent theoretical progress on the statistical mechanics of error correcting codes, focusing on low-density parity-check (LDPC) codes in general, and on Gallager and MacKay-Neal codes in particular. By exploiting the relation between LDPC codes and Ising spin systems with multi-spin interactions, one can carry out a statistical mechanics based analysis that determines the practical and theoretical limitations of various code constructions, corresponding to dynamical and thermodynamical transitions, respectively, as well as the behaviour of error-exponents averaged over the corresponding code ensemble as a function of channel noise. We also contrast the results obtained using methods of statistical mechanics with those derived in the information theory literature, and show how these methods can be generalized to include other channel types and related communication problems. (topical review)

  15. Statistical mechanics of low-density parity-check codes

    International Nuclear Information System (INIS)

    Kabashima, Yoshiyuki; Saad, David

    2004-01-01

    We review recent theoretical progress on the statistical mechanics of error correcting codes, focusing on low-density parity-check (LDPC) codes in general, and on Gallager and MacKay-Neal codes in particular. By exploiting the relation between LDPC codes and Ising spin systems with multi-spin interactions, one can carry out a statistical mechanics based analysis that determines the practical and theoretical limitations of various code constructions, corresponding to dynamical and thermodynamical transitions, respectively, as well as the behaviour of error-exponents averaged over the corresponding code ensemble as a function of channel noise. We also contrast the results obtained using methods of statistical mechanics with those derived in the information theory literature, and show how these methods can be generalized to include other channel types and related communication problems. (topical review)

  16. Trellis and turbo coding iterative and graph-based error control coding

    CERN Document Server

    Schlegel, Christian B

    2015-01-01

    This new edition has been extensively revised to reflect the progress in error control coding over the past few years. Over 60% of the material has been completely reworked, and 30% of the material is original. Convolutional, turbo, and low density parity-check (LDPC) coding and polar codes in a unified framework. Advanced research-related developments such as spatial coupling. A focus on algorithmic and implementation aspects of error control coding.

  17. Uplink Coding

    Science.gov (United States)

    Andrews, Ken; Divsalar, Dariush; Dolinar, Sam; Moision, Bruce; Hamkins, Jon; Pollara, Fabrizio

    2007-01-01

    This slide presentation reviews the objectives, meeting goals and overall NASA goals for the NASA Data Standards Working Group. The presentation includes information on the technical progress surrounding the objective, short LDPC codes, and the general results on the Pu-Pw tradeoff.

  18. LDPC kódy

    OpenAIRE

    Hrouza, Ondřej

    2012-01-01

    Práce se zabývá problematikou LDPC kódů. Jsou zde popsány metody vytváření paritní matice, kde je kladen důraz především na strukturované vytváření této matice za použití konečné geometrie: Euklidovské geometrie a projektivní geometrie. Další oblastí, které se práce věnuje je dekódování LDPC kódů. Práce porovnává čtyři dekódovací metody: Hard-Decision algoritmus, Bit-Flipping algoritmus, The Sum-Product algoritmus a Log Likelihood algoritmus, při kterých je kladen důraz především na iterativn...

  19. Statistical physics inspired energy-efficient coded-modulation for optical communications.

    Science.gov (United States)

    Djordjevic, Ivan B; Xu, Lei; Wang, Ting

    2012-04-15

    Because Shannon's entropy can be obtained by Stirling's approximation of thermodynamics entropy, the statistical physics energy minimization methods are directly applicable to the signal constellation design. We demonstrate that statistical physics inspired energy-efficient (EE) signal constellation designs, in combination with large-girth low-density parity-check (LDPC) codes, significantly outperform conventional LDPC-coded polarization-division multiplexed quadrature amplitude modulation schemes. We also describe an EE signal constellation design algorithm. Finally, we propose the discrete-time implementation of D-dimensional transceiver and corresponding EE polarization-division multiplexed system. © 2012 Optical Society of America

  20. Parity-Check Network Coding for Multiple Access Relay Channel in Wireless Sensor Cooperative Communications

    Directory of Open Access Journals (Sweden)

    Du Bing

    2010-01-01

    Full Text Available A recently developed theory suggests that network coding is a generalization of source coding and channel coding and thus yields a significant performance improvement in terms of throughput and spatial diversity. This paper proposes a cooperative design of a parity-check network coding scheme in the context of a two-source multiple access relay channel (MARC model, a common compact model in hierarchical wireless sensor networks (WSNs. The scheme uses Low-Density Parity-Check (LDPC as the surrogate to build up a layered structure which encapsulates the multiple constituent LDPC codes in the source and relay nodes. Specifically, the relay node decodes the messages from two sources, which are used to generate extra parity-check bits by a random network coding procedure to fill up the rate gap between Source-Relay and Source-Destination transmissions. Then, we derived the key algebraic relationships among multidimensional LDPC constituent codes as one of the constraints for code profile optimization. These extra check bits are sent to the destination to realize a cooperative diversity as well as to approach MARC decode-and-forward (DF capacity.

  1. Typical performance of regular low-density parity-check codes over general symmetric channels

    Energy Technology Data Exchange (ETDEWEB)

    Tanaka, Toshiyuki [Department of Electronics and Information Engineering, Tokyo Metropolitan University, 1-1 Minami-Osawa, Hachioji-shi, Tokyo 192-0397 (Japan); Saad, David [Neural Computing Research Group, Aston University, Aston Triangle, Birmingham B4 7ET (United Kingdom)

    2003-10-31

    Typical performance of low-density parity-check (LDPC) codes over a general binary-input output-symmetric memoryless channel is investigated using methods of statistical mechanics. Relationship between the free energy in statistical-mechanics approach and the mutual information used in the information-theory literature is established within a general framework; Gallager and MacKay-Neal codes are studied as specific examples of LDPC codes. It is shown that basic properties of these codes known for particular channels, including their potential to saturate Shannon's bound, hold for general symmetric channels. The binary-input additive-white-Gaussian-noise channel and the binary-input Laplace channel are considered as specific channel models.

  2. Typical performance of regular low-density parity-check codes over general symmetric channels

    International Nuclear Information System (INIS)

    Tanaka, Toshiyuki; Saad, David

    2003-01-01

    Typical performance of low-density parity-check (LDPC) codes over a general binary-input output-symmetric memoryless channel is investigated using methods of statistical mechanics. Relationship between the free energy in statistical-mechanics approach and the mutual information used in the information-theory literature is established within a general framework; Gallager and MacKay-Neal codes are studied as specific examples of LDPC codes. It is shown that basic properties of these codes known for particular channels, including their potential to saturate Shannon's bound, hold for general symmetric channels. The binary-input additive-white-Gaussian-noise channel and the binary-input Laplace channel are considered as specific channel models

  3. An experimental comparison of coded modulation strategies for 100 Gb/s transceivers

    NARCIS (Netherlands)

    Sillekens, E.; Alvarado, A.; Okonkwo, C.; Thomsen, B.C.

    2016-01-01

    Coded modulation is a key technique to increase the spectral efficiency of coherent optical communication systems. Two popular strategies for coded modulation are turbo trellis-coded modulation (TTCM) and bit-interleaved coded modulation (BICM) based on low-density parity-check (LDPC) codes.

  4. Two-terminal video coding.

    Science.gov (United States)

    Yang, Yang; Stanković, Vladimir; Xiong, Zixiang; Zhao, Wei

    2009-03-01

    Following recent works on the rate region of the quadratic Gaussian two-terminal source coding problem and limit-approaching code designs, this paper examines multiterminal source coding of two correlated, i.e., stereo, video sequences to save the sum rate over independent coding of both sequences. Two multiterminal video coding schemes are proposed. In the first scheme, the left sequence of the stereo pair is coded by H.264/AVC and used at the joint decoder to facilitate Wyner-Ziv coding of the right video sequence. The first I-frame of the right sequence is successively coded by H.264/AVC Intracoding and Wyner-Ziv coding. An efficient stereo matching algorithm based on loopy belief propagation is then adopted at the decoder to produce pixel-level disparity maps between the corresponding frames of the two decoded video sequences on the fly. Based on the disparity maps, side information for both motion vectors and motion-compensated residual frames of the right sequence are generated at the decoder before Wyner-Ziv encoding. In the second scheme, source splitting is employed on top of classic and Wyner-Ziv coding for compression of both I-frames to allow flexible rate allocation between the two sequences. Experiments with both schemes on stereo video sequences using H.264/AVC, LDPC codes for Slepian-Wolf coding of the motion vectors, and scalar quantization in conjunction with LDPC codes for Wyner-Ziv coding of the residual coefficients give a slightly lower sum rate than separate H.264/AVC coding of both sequences at the same video quality.

  5. Combining Ratio Estimation for Low Density Parity Check (LDPC) Coding

    Science.gov (United States)

    Mahmoud, Saad; Hi, Jianjun

    2012-01-01

    The Low Density Parity Check (LDPC) Code decoding algorithm make use of a scaled receive signal derived from maximizing the log-likelihood ratio of the received signal. The scaling factor (often called the combining ratio) in an AWGN channel is a ratio between signal amplitude and noise variance. Accurately estimating this ratio has shown as much as 0.6 dB decoding performance gain. This presentation briefly describes three methods for estimating the combining ratio: a Pilot-Guided estimation method, a Blind estimation method, and a Simulation-Based Look-Up table. The Pilot Guided Estimation method has shown that the maximum likelihood estimates of signal amplitude is the mean inner product of the received sequence and the known sequence, the attached synchronization marker (ASM) , and signal variance is the difference of the mean of the squared received sequence and the square of the signal amplitude. This method has the advantage of simplicity at the expense of latency since several frames worth of ASMs. The Blind estimation method s maximum likelihood estimator is the average of the product of the received signal with the hyperbolic tangent of the product combining ratio and the received signal. The root of this equation can be determined by an iterative binary search between 0 and 1 after normalizing the received sequence. This method has the benefit of requiring one frame of data to estimate the combining ratio which is good for faster changing channels compared to the previous method, however it is computationally expensive. The final method uses a look-up table based on prior simulated results to determine signal amplitude and noise variance. In this method the received mean signal strength is controlled to a constant soft decision value. The magnitude of the deviation is averaged over a predetermined number of samples. This value is referenced in a look up table to determine the combining ratio that prior simulation associated with the average magnitude of

  6. Double-Layer Low-Density Parity-Check Codes over Multiple-Input Multiple-Output Channels

    Directory of Open Access Journals (Sweden)

    Yun Mao

    2012-01-01

    Full Text Available We introduce a double-layer code based on the combination of a low-density parity-check (LDPC code with the multiple-input multiple-output (MIMO system, where the decoding can be done in both inner-iteration and outer-iteration manners. The present code, called low-density MIMO code (LDMC, has a double-layer structure, that is, one layer defines subcodes that are embedded in each transmission vector and another glues these subcodes together. It supports inner iterations inside the LDPC decoder and outeriterations between detectors and decoders, simultaneously. It can also achieve the desired design rates due to the full rank of the deployed parity-check matrix. Simulations show that the LDMC performs favorably over the MIMO systems.

  7. Joint nonbinary low-density parity-check codes and modulation diversity over fading channels

    Science.gov (United States)

    Shi, Zhiping; Li, Tiffany Jing; Zhang, Zhongpei

    2010-09-01

    A joint exploitation of coding and diversity techniques to achieve efficient, reliable wireless transmission is considered. The system comprises a powerful non-binary low-density parity-check (LDPC) code that will be soft-decoded to supply strong error protection, a quadratic amplitude modulator (QAM) that directly takes in the non-binary LDPC symbols and a modulation diversity operator that will provide power- and bandwidth-efficient diversity gain. By relaxing the rate of the modulation diversity rotation matrices to below 1, we show that a better rate allocation can be arranged between the LDPC codes and the modulation diversity, which brings significant performance gain over previous systems. To facilitate the design and evaluation of the relaxed modulation diversity rotation matrices, based on a set of criteria, three practical design methods are given and their point pairwise error rate are analyzed. With EXIT chart, we investigate the convergence between demodulator and decoder.A rate match method is presented based on EXIT analysis. Through analysis and simulations, we show that our strategies are very effective in combating random fading and strong noise on fading channels.

  8. Non-binary Entanglement-assisted Stabilizer Quantum Codes

    OpenAIRE

    Riguang, Leng; Zhi, Ma

    2011-01-01

    In this paper, we show how to construct non-binary entanglement-assisted stabilizer quantum codes by using pre-shared entanglement between the sender and receiver. We also give an algorithm to determine the circuit for non-binary entanglement-assisted stabilizer quantum codes and some illustrated examples. The codes we constructed do not require the dual-containing constraint, and many non-binary classical codes, like non-binary LDPC codes, which do not satisfy the condition, can be used to c...

  9. A new LDPC decoding scheme for PDM-8QAM BICM coherent optical communication system

    Science.gov (United States)

    Liu, Yi; Zhang, Wen-bo; Xi, Li-xia; Tang, Xian-feng; Zhang, Xiao-guang

    2015-11-01

    A new log-likelihood ratio (LLR) message estimation method is proposed for polarization-division multiplexing eight quadrature amplitude modulation (PDM-8QAM) bit-interleaved coded modulation (BICM) optical communication system. The formulation of the posterior probability is theoretically analyzed, and the way to reduce the pre-decoding bit error rate ( BER) of the low density parity check (LDPC) decoder for PDM-8QAM constellations is presented. Simulation results show that it outperforms the traditional scheme, i.e., the new post-decoding BER is decreased down to 50% of that of the traditional post-decoding algorithm.

  10. On locality of Generalized Reed-Muller codes over the broadcast erasure channel

    KAUST Repository

    Alloum, Amira; Lin, Sian Jheng; Al-Naffouri, Tareq Y.

    2016-01-01

    , and more specifically at the application layer where Rateless, LDPC, Reed Slomon codes and network coding schemes have been extensively studied, optimized and standardized in the past. Beyond reusing, extending or adapting existing application layer packet

  11. PMD compensation in multilevel coded-modulation schemes with coherent detection using BLAST algorithm and iterative polarization cancellation.

    Science.gov (United States)

    Djordjevic, Ivan B; Xu, Lei; Wang, Ting

    2008-09-15

    We present two PMD compensation schemes suitable for use in multilevel (M>or=2) block-coded modulation schemes with coherent detection. The first scheme is based on a BLAST-type polarization-interference cancellation scheme, and the second scheme is based on iterative polarization cancellation. Both schemes use the LDPC codes as channel codes. The proposed PMD compensations schemes are evaluated by employing coded-OFDM and coherent detection. When used in combination with girth-10 LDPC codes those schemes outperform polarization-time coding based OFDM by 1 dB at BER of 10(-9), and provide two times higher spectral efficiency. The proposed schemes perform comparable and are able to compensate even 1200 ps of differential group delay with negligible penalty.

  12. LDPC coding for QKD at higher photon flux levels based on spatial entanglement of twin beams in PDC

    International Nuclear Information System (INIS)

    Daneshgaran, Fred; Mondin, Marina; Bari, Inam

    2014-01-01

    Twin beams generated by Parametric Down Conversion (PDC) exhibit quantum correlations that has been effectively used as a tool for many applications including calibration of single photon detectors. By now, detection of multi-mode spatial correlations is a mature field and in principle, only depends on the transmission and detection efficiency of the devices and the channel. In [2, 4, 5], the authors utilized their know-how on almost perfect selection of modes of pairwise correlated entangled beams and the optimization of the noise reduction to below the shot-noise level, for absolute calibration of Charge Coupled Device (CCD) cameras. The same basic principle is currently being considered by the same authors for possible use in Quantum Key Distribution (QKD) [3, 1]. The main advantage in such an approach would be the ability to work with much higher photon fluxes than that of a single photon regime that is theoretically required for discrete variable QKD applications (in practice, very weak laser pulses with mean photon count below one are used).The natural setup of quantization of CCD detection area and subsequent measurement of the correlation statistic needed to detect the presence of the eavesdropper Eve, leads to a QKD channel model that is a Discrete Memoryless Channel (DMC) with a number of inputs and outputs that can be more than two (i.e., the channel is a multi-level DMC). This paper investigates the use of Low Density Parity Check (LDPC) codes for information reconciliation on the effective parallel channels associated with the multi-level DMC. The performance of such codes are shown to be close to the theoretical limits.

  13. Performance optimization of PM-16QAM transmission system enabled by real-time self-adaptive coding.

    Science.gov (United States)

    Qu, Zhen; Li, Yao; Mo, Weiyang; Yang, Mingwei; Zhu, Shengxiang; Kilper, Daniel C; Djordjevic, Ivan B

    2017-10-15

    We experimentally demonstrate self-adaptive coded 5×100  Gb/s WDM polarization multiplexed 16 quadrature amplitude modulation transmission over a 100 km fiber link, which is enabled by a real-time control plane. The real-time optical signal-to-noise ratio (OSNR) is measured using an optical performance monitoring device. The OSNR measurement is processed and fed back using control plane logic and messaging to the transmitter side for code adaptation, where the binary data are adaptively encoded with three types of low-density parity-check (LDPC) codes with code rates of 0.8, 0.75, and 0.7 of large girth. The total code-adaptation latency is measured to be 2273 ms. Compared with transmission without adaptation, average net capacity improvements of 102%, 36%, and 7.5% are obtained, respectively, by adaptive LDPC coding.

  14. Bilayer Protograph Codes for Half-Duplex Relay Channels

    Science.gov (United States)

    Divsalar, Dariush; VanNguyen, Thuy; Nosratinia, Aria

    2013-01-01

    Direct to Earth return links are limited by the size and power of lander devices. A standard alternative is provided by a two-hops return link: a proximity link (from lander to orbiter relay) and a deep-space link (from orbiter relay to Earth). Although direct to Earth return links are limited by the size and power of lander devices, using an additional link and a proposed coding for relay channels, one can obtain a more reliable signal. Although significant progress has been made in the relay coding problem, existing codes must be painstakingly optimized to match to a single set of channel conditions, many of them do not offer easy encoding, and most of them do not have structured design. A high-performing LDPC (low-density parity-check) code for the relay channel addresses simultaneously two important issues: a code structure that allows low encoding complexity, and a flexible rate-compatible code that allows matching to various channel conditions. Most of the previous high-performance LDPC codes for the relay channel are tightly optimized for a given channel quality, and are not easily adapted without extensive re-optimization for various channel conditions. This code for the relay channel combines structured design and easy encoding with rate compatibility to allow adaptation to the three links involved in the relay channel, and furthermore offers very good performance. The proposed code is constructed by synthesizing a bilayer structure with a pro to graph. In addition to the contribution to relay encoding, an improved family of protograph codes was produced for the point-to-point AWGN (additive white Gaussian noise) channel whose high-rate members enjoy thresholds that are within 0.07 dB of capacity. These LDPC relay codes address three important issues in an integrative manner: low encoding complexity, modular structure allowing for easy design, and rate compatibility so that the code can be easily matched to a variety of channel conditions without extensive

  15. A Simple Scheme for Belief Propagation Decoding of BCH and RS Codes in Multimedia Transmissions

    Directory of Open Access Journals (Sweden)

    Marco Baldi

    2008-01-01

    Full Text Available Classic linear block codes, like Bose-Chaudhuri-Hocquenghem (BCH and Reed-Solomon (RS codes, are widely used in multimedia transmissions, but their soft-decision decoding still represents an open issue. Among the several approaches proposed for this purpose, an important role is played by the iterative belief propagation principle, whose application to low-density parity-check (LDPC codes permits to approach the channel capacity. In this paper, we elaborate a new technique for decoding classic binary and nonbinary codes through the belief propagation algorithm. We focus on RS codes included in the recent CDMA2000 standard, and compare the proposed technique with the adaptive belief propagation approach, that is able to ensure very good performance but with higher complexity. Moreover, we consider the case of long BCH codes included in the DVB-S2 standard, for which we show that the usage of “pure” LDPC codes would provide better performance.

  16. A Low-Complexity and High-Performance 2D Look-Up Table for LDPC Hardware Implementation

    Science.gov (United States)

    Chen, Jung-Chieh; Yang, Po-Hui; Lain, Jenn-Kaie; Chung, Tzu-Wen

    In this paper, we propose a low-complexity, high-efficiency two-dimensional look-up table (2D LUT) for carrying out the sum-product algorithm in the decoding of low-density parity-check (LDPC) codes. Instead of employing adders for the core operation when updating check node messages, in the proposed scheme, the main term and correction factor of the core operation are successfully merged into a compact 2D LUT. Simulation results indicate that the proposed 2D LUT not only attains close-to-optimal bit error rate performance but also enjoys a low complexity advantage that is suitable for hardware implementation.

  17. A mean field theory of coded CDMA systems

    International Nuclear Information System (INIS)

    Yano, Toru; Tanaka, Toshiyuki; Saad, David

    2008-01-01

    We present a mean field theory of code-division multiple-access (CDMA) systems with error-control coding. On the basis of the relation between the free energy and mutual information, we obtain an analytical expression of the maximum spectral efficiency of the coded CDMA system, from which a mean-field description of the coded CDMA system is provided in terms of a bank of scalar Gaussian channels whose variances in general vary at different code symbol positions. Regular low-density parity-check (LDPC)-coded CDMA systems are also discussed as an example of the coded CDMA systems

  18. A mean field theory of coded CDMA systems

    Energy Technology Data Exchange (ETDEWEB)

    Yano, Toru [Graduate School of Science and Technology, Keio University, Hiyoshi, Kohoku-ku, Yokohama-shi, Kanagawa 223-8522 (Japan); Tanaka, Toshiyuki [Graduate School of Informatics, Kyoto University, Yoshida Hon-machi, Sakyo-ku, Kyoto-shi, Kyoto 606-8501 (Japan); Saad, David [Neural Computing Research Group, Aston University, Birmingham B4 7ET (United Kingdom)], E-mail: yano@thx.appi.keio.ac.jp

    2008-08-15

    We present a mean field theory of code-division multiple-access (CDMA) systems with error-control coding. On the basis of the relation between the free energy and mutual information, we obtain an analytical expression of the maximum spectral efficiency of the coded CDMA system, from which a mean-field description of the coded CDMA system is provided in terms of a bank of scalar Gaussian channels whose variances in general vary at different code symbol positions. Regular low-density parity-check (LDPC)-coded CDMA systems are also discussed as an example of the coded CDMA systems.

  19. Error-correction coding and decoding bounds, codes, decoders, analysis and applications

    CERN Document Server

    Tomlinson, Martin; Ambroze, Marcel A; Ahmed, Mohammed; Jibril, Mubarak

    2017-01-01

    This book discusses both the theory and practical applications of self-correcting data, commonly known as error-correcting codes. The applications included demonstrate the importance of these codes in a wide range of everyday technologies, from smartphones to secure communications and transactions. Written in a readily understandable style, the book presents the authors’ twenty-five years of research organized into five parts: Part I is concerned with the theoretical performance attainable by using error correcting codes to achieve communications efficiency in digital communications systems. Part II explores the construction of error-correcting codes and explains the different families of codes and how they are designed. Techniques are described for producing the very best codes. Part III addresses the analysis of low-density parity-check (LDPC) codes, primarily to calculate their stopping sets and low-weight codeword spectrum which determines the performance of these codes. Part IV deals with decoders desi...

  20. Encoders for block-circulant LDPC codes

    Science.gov (United States)

    Divsalar, Dariush (Inventor); Abbasfar, Aliazam (Inventor); Jones, Christopher R. (Inventor); Dolinar, Samuel J. (Inventor); Thorpe, Jeremy C. (Inventor); Andrews, Kenneth S. (Inventor); Yao, Kung (Inventor)

    2009-01-01

    Methods and apparatus to encode message input symbols in accordance with an accumulate-repeat-accumulate code with repetition three or four are disclosed. Block circulant matrices are used. A first method and apparatus make use of the block-circulant structure of the parity check matrix. A second method and apparatus use block-circulant generator matrices.

  1. Code-Hopping Based Transmission Scheme for Wireless Physical-Layer Security

    Directory of Open Access Journals (Sweden)

    Liuguo Yin

    2018-01-01

    Full Text Available Due to the broadcast and time-varying natures of wireless channels, traditional communication systems that provide data encryption at the application layer suffer many challenges such as error diffusion. In this paper, we propose a code-hopping based secrecy transmission scheme that uses dynamic nonsystematic low-density parity-check (LDPC codes and automatic repeat-request (ARQ mechanism to jointly encode and encrypt source messages at the physical layer. In this scheme, secret keys at the transmitter and the legitimate receiver are generated dynamically upon the source messages that have been transmitted successfully. During the transmission, each source message is jointly encoded and encrypted by a parity-check matrix, which is dynamically selected from a set of LDPC matrices based on the shared dynamic secret key. As for the eavesdropper (Eve, the uncorrectable decoding errors prevent her from generating the same secret key as the legitimate parties. Thus she cannot select the correct LDPC matrix to recover the source message. We demonstrate that our scheme can be compatible with traditional cryptosystems and enhance the security without sacrificing the error-correction performance. Numerical results show that the bit error rate (BER of Eve approaches 0.5 as the number of transmitted source messages increases and the security gap of the system is small.

  2. The Analysis and the Performance Simulation of the Capacity of Bit-interleaved Coded Modulation System

    Directory of Open Access Journals (Sweden)

    Hongwei ZHAO

    2014-09-01

    Full Text Available In this paper, the capacity of the BICM system over AWGN channels is first analyzed; the curves of BICM capacity versus SNR are also got by the Monte-Carlo simulations===?=== and compared with the curves of the CM capacity. Based on the analysis results, we simulate the error performances of BICM system with LDPC codes. Simulation results show that the capacity of BICM system with LDPC codes is enormously influenced by the mapping methods. Given a certain modulation method, the BICM system can obtain about 2-3 dB gain with Gray mapping compared with Non-Gray mapping. Meanwhile, the simulation results also demonstrate the correctness of the theory analysis.

  3. On the Performance of a Multi-Edge Type LDPC Code for Coded Modulation

    NARCIS (Netherlands)

    Cronie, H.S.

    2005-01-01

    We present a method to combine error-correction coding and spectral-efficient modulation for transmission over the Additive White Gaussian Noise (AWGN) channel. The code employs signal shaping which can provide a so-called shaping gain. The code belongs to the family of sparse graph codes for which

  4. A new Monte Carlo code for simulation of the effect of irregular surfaces on X-ray spectra

    Energy Technology Data Exchange (ETDEWEB)

    Brunetti, Antonio, E-mail: brunetti@uniss.it; Golosio, Bruno

    2014-04-01

    Generally, quantitative X-ray fluorescence (XRF) analysis estimates the content of chemical elements in a sample based on the areas of the fluorescence peaks in the energy spectrum. Besides the concentration of the elements, the peak areas depend also on the geometrical conditions. In fact, the estimate of the peak areas is simple if the sample surface is smooth and if the spectrum shows a good statistic (large-area peaks). For this reason often the sample is prepared as a pellet. However, this approach is not always feasible, for instance when cultural heritage or valuable samples must be analyzed. In this case, the sample surface cannot be smoothed. In order to address this problem, several works have been reported in the literature, based on experimental measurements on a few sets of specific samples or on Monte Carlo simulations. The results obtained with the first approach are limited by the specific class of samples analyzed, while the second approach cannot be applied to arbitrarily irregular surfaces. The present work describes a more general analysis tool based on a new fast Monte Carlo algorithm, which is virtually able to simulate any kind of surface. At the best of our knowledge, it is the first Monte Carlo code with this option. A study of the influence of surface irregularities on the measured spectrum is performed and some results reported. - Highlights: • We present a fast Monte Carlo code with the possibility to simulate any irregularly rough surfaces. • We show applications to multilayer measurements. • Real time simulations are available.

  5. LDPC decoder with a limited-precision FPGA-based floating-point multiplication coprocessor

    Science.gov (United States)

    Moberly, Raymond; O'Sullivan, Michael; Waheed, Khurram

    2007-09-01

    Implementing the sum-product algorithm, in an FPGA with an embedded processor, invites us to consider a tradeoff between computational precision and computational speed. The algorithm, known outside of the signal processing community as Pearl's belief propagation, is used for iterative soft-decision decoding of LDPC codes. We determined the feasibility of a coprocessor that will perform product computations. Our FPGA-based coprocessor (design) performs computer algebra with significantly less precision than the standard (e.g. integer, floating-point) operations of general purpose processors. Using synthesis, targeting a 3,168 LUT Xilinx FPGA, we show that key components of a decoder are feasible and that the full single-precision decoder could be constructed using a larger part. Soft-decision decoding by the iterative belief propagation algorithm is impacted both positively and negatively by a reduction in the precision of the computation. Reducing precision reduces the coding gain, but the limited-precision computation can operate faster. A proposed solution offers custom logic to perform computations with less precision, yet uses the floating-point format to interface with the software. Simulation results show the achievable coding gain. Synthesis results help theorize the the full capacity and performance of an FPGA-based coprocessor.

  6. Nonlinear demodulation and channel coding in EBPSK scheme.

    Science.gov (United States)

    Chen, Xianqing; Wu, Lenan

    2012-01-01

    The extended binary phase shift keying (EBPSK) is an efficient modulation technique, and a special impacting filter (SIF) is used in its demodulator to improve the bit error rate (BER) performance. However, the conventional threshold decision cannot achieve the optimum performance, and the SIF brings more difficulty in obtaining the posterior probability for LDPC decoding. In this paper, we concentrate not only on reducing the BER of demodulation, but also on providing accurate posterior probability estimates (PPEs). A new approach for the nonlinear demodulation based on the support vector machine (SVM) classifier is introduced. The SVM method which selects only a few sampling points from the filter output was used for getting PPEs. The simulation results show that the accurate posterior probability can be obtained with this method and the BER performance can be improved significantly by applying LDPC codes. Moreover, we analyzed the effect of getting the posterior probability with different methods and different sampling rates. We show that there are more advantages of the SVM method under bad condition and it is less sensitive to the sampling rate than other methods. Thus, SVM is an effective method for EBPSK demodulation and getting posterior probability for LDPC decoding.

  7. Neural network decoder for quantum error correcting codes

    Science.gov (United States)

    Krastanov, Stefan; Jiang, Liang

    Artificial neural networks form a family of extremely powerful - albeit still poorly understood - tools used in anything from image and sound recognition through text generation to, in our case, decoding. We present a straightforward Recurrent Neural Network architecture capable of deducing the correcting procedure for a quantum error-correcting code from a set of repeated stabilizer measurements. We discuss the fault-tolerance of our scheme and the cost of training the neural network for a system of a realistic size. Such decoders are especially interesting when applied to codes, like the quantum LDPC codes, that lack known efficient decoding schemes.

  8. Coded Modulation in C and MATLAB

    Science.gov (United States)

    Hamkins, Jon; Andrews, Kenneth S.

    2011-01-01

    This software, written separately in C and MATLAB as stand-alone packages with equivalent functionality, implements encoders and decoders for a set of nine error-correcting codes and modulators and demodulators for five modulation types. The software can be used as a single program to simulate the performance of such coded modulation. The error-correcting codes implemented are the nine accumulate repeat-4 jagged accumulate (AR4JA) low-density parity-check (LDPC) codes, which have been approved for international standardization by the Consultative Committee for Space Data Systems, and which are scheduled to fly on a series of NASA missions in the Constellation Program. The software implements the encoder and decoder functions, and contains compressed versions of generator and parity-check matrices used in these operations.

  9. Quantum Kronecker sum-product low-density parity-check codes with finite rate

    Science.gov (United States)

    Kovalev, Alexey A.; Pryadko, Leonid P.

    2013-07-01

    We introduce an ansatz for quantum codes which gives the hypergraph-product (generalized toric) codes by Tillich and Zémor and generalized bicycle codes by MacKay as limiting cases. The construction allows for both the lower and the upper bounds on the minimum distance; they scale as a square root of the block length. Many thus defined codes have a finite rate and limited-weight stabilizer generators, an analog of classical low-density parity-check (LDPC) codes. Compared to the hypergraph-product codes, hyperbicycle codes generally have a wider range of parameters; in particular, they can have a higher rate while preserving the estimated error threshold.

  10. Binary Linear-Time Erasure Decoding for Non-Binary LDPC codes

    OpenAIRE

    Savin, Valentin

    2009-01-01

    In this paper, we first introduce the extended binary representation of non-binary codes, which corresponds to a covering graph of the bipartite graph associated with the non-binary code. Then we show that non-binary codewords correspond to binary codewords of the extended representation that further satisfy some simplex-constraint: that is, bits lying over the same symbol-node of the non-binary graph must form a codeword of a simplex code. Applied to the binary erasure channel, this descript...

  11. Advanced hardware design for error correcting codes

    CERN Document Server

    Coussy, Philippe

    2015-01-01

    This book provides thorough coverage of error correcting techniques. It includes essential basic concepts and the latest advances on key topics in design, implementation, and optimization of hardware/software systems for error correction. The book’s chapters are written by internationally recognized experts in this field. Topics include evolution of error correction techniques, industrial user needs, architectures, and design approaches for the most advanced error correcting codes (Polar Codes, Non-Binary LDPC, Product Codes, etc). This book provides access to recent results, and is suitable for graduate students and researchers of mathematics, computer science, and engineering. • Examines how to optimize the architecture of hardware design for error correcting codes; • Presents error correction codes from theory to optimized architecture for the current and the next generation standards; • Provides coverage of industrial user needs advanced error correcting techniques.

  12. Photonic circuits for iterative decoding of a class of low-density parity-check codes

    International Nuclear Information System (INIS)

    Pavlichin, Dmitri S; Mabuchi, Hideo

    2014-01-01

    Photonic circuits in which stateful components are coupled via guided electromagnetic fields are natural candidates for resource-efficient implementation of iterative stochastic algorithms based on propagation of information around a graph. Conversely, such message=passing algorithms suggest novel circuit architectures for signal processing and computation that are well matched to nanophotonic device physics. Here, we construct and analyze a quantum optical model of a photonic circuit for iterative decoding of a class of low-density parity-check (LDPC) codes called expander codes. Our circuit can be understood as an open quantum system whose autonomous dynamics map straightforwardly onto the subroutines of an LDPC decoding scheme, with several attractive features: it can operate in the ultra-low power regime of photonics in which quantum fluctuations become significant, it is robust to noise and component imperfections, it achieves comparable performance to known iterative algorithms for this class of codes, and it provides an instructive example of how nanophotonic cavity quantum electrodynamic components can enable useful new information technology even if the solid-state qubits on which they are based are heavily dephased and cannot support large-scale entanglement. (paper)

  13. Information rates of probabilistically shaped coded modulation for a multi-span fiber-optic communication system with 64QAM

    Science.gov (United States)

    Fehenberger, Tobias

    2018-02-01

    This paper studies probabilistic shaping in a multi-span wavelength-division multiplexing optical fiber system with 64-ary quadrature amplitude modulation (QAM) input. In split-step fiber simulations and via an enhanced Gaussian noise model, three figures of merit are investigated, which are signal-to-noise ratio (SNR), achievable information rate (AIR) for capacity-achieving forward error correction (FEC) with bit-metric decoding, and the information rate achieved with low-density parity-check (LDPC) FEC. For the considered system parameters and different shaped input distributions, shaping is found to decrease the SNR by 0.3 dB yet simultaneously increases the AIR by up to 0.4 bit per 4D-symbol. The information rates of LDPC-coded modulation with shaped 64QAM input are improved by up to 0.74 bit per 4D-symbol, which is larger than the shaping gain when considering AIRs. This increase is attributed to the reduced coding gap of the higher-rate code that is used for decoding the nonuniform QAM input.

  14. Joint beam design and user selection over non-binary coded MIMO interference channel

    Science.gov (United States)

    Li, Haitao; Yuan, Haiying

    2013-03-01

    In this paper, we discuss the problem of sum rate improvement for coded MIMO interference system, and propose joint beam design and user selection over interference channel. Firstly, we have formulated non-binary LDPC coded MIMO interference networks model. Then, the least square beam design for MIMO interference system is derived, and the low complexity user selection is presented. Simulation results confirm that the sum rate can be improved by the joint user selection and beam design comparing with single interference aligning beamformer.

  15. Design of a VLSI Decoder for Partially Structured LDPC Codes

    Directory of Open Access Journals (Sweden)

    Fabrizio Vacca

    2008-01-01

    of their parity matrix can be partitioned into two disjoint sets, namely, the structured and the random ones. For the proposed class of codes a constructive design method is provided. To assess the value of this method the constructed codes performance are presented. From these results, a novel decoding method called split decoding is introduced. Finally, to prove the effectiveness of the proposed approach a whole VLSI decoder is designed and characterized.

  16. Analysis of an Irregular RC Multi-storeyed Building Subjected to Dynamic Loading

    Science.gov (United States)

    AkashRaut; Pachpor, Prabodh; Dautkhani, Sanket

    2018-03-01

    Many buildings in the present scenario have irregular configurations both in plan and elevation. This in future may subject to devastating earthquakes. So it is necessary to analyze the structure. The present paper is made to study three type of irregularity wiz vertical, mass and plan irregularity as per clause 7.1 of IS 1893 (part1)2002 code. The paper discusses the analysis of RC (Reinforced Concrete) Buildings with vertical irregularity. The study as a whole makes an effort to evaluate the effect of vertical irregularity on RC buildings for which comparison of three parameters namely shear force, bending moment and deflection are taken into account.

  17. Simulasi Low Density Parity Check (Ldpc) dengan Standar Dvb-t2

    OpenAIRE

    Kurniawan, Yusuf; Hafizh, Idham

    2014-01-01

    Artikel ini berisi implementasi simulasi encoding-decoding yang dilakukanpada suatu sampel data biner acak sesuai dengan standar yang digunakanpada Digital Video Broadcasting – Terrestrial 2nd Generation (DVB-T2),dengan menggunakan MATLAB. Low Density Parity Check (LDPC)digunakan dalam proses encoding-decoding sebagai fitur untuk melakukankoreksi kesalahan pada saat pengiriman data. Modulasi yang digunakandalam simulasi adalah BPSK dengan model kanal AWGN. Dalam simulasitersebut, diperbanding...

  18. Irreducible normalizer operators and thresholds for degenerate quantum codes with sublinear distances

    Science.gov (United States)

    Pryadko, Leonid P.; Dumer, Ilya; Kovalev, Alexey A.

    2015-03-01

    We construct a lower (existence) bound for the threshold of scalable quantum computation which is applicable to all stabilizer codes, including degenerate quantum codes with sublinear distance scaling. The threshold is based on enumerating irreducible operators in the normalizer of the code, i.e., those that cannot be decomposed into a product of two such operators with non-overlapping support. For quantum LDPC codes with logarithmic or power-law distances, we get threshold values which are parametrically better than the existing analytical bound based on percolation. The new bound also gives a finite threshold when applied to other families of degenerate quantum codes, e.g., the concatenated codes. This research was supported in part by the NSF Grant PHY-1416578 and by the ARO Grant W911NF-11-1-0027.

  19. On locality of Generalized Reed-Muller codes over the broadcast erasure channel

    KAUST Repository

    Alloum, Amira

    2016-07-28

    One to Many communications are expected to be among the killer applications for the currently discussed 5G standard. The usage of coding mechanisms is impacting broadcasting standard quality, as coding is involved at several levels of the stack, and more specifically at the application layer where Rateless, LDPC, Reed Slomon codes and network coding schemes have been extensively studied, optimized and standardized in the past. Beyond reusing, extending or adapting existing application layer packet coding mechanisms based on previous schemes and designed for the foregoing LTE or other broadcasting standards; our purpose is to investigate the use of Generalized Reed Muller codes and the value of their locality property in their progressive decoding for Broadcast/Multicast communication schemes with real time video delivery. Our results are meant to bring insight into the use of locally decodable codes in Broadcasting. © 2016 IEEE.

  20. The serial message-passing schedule for LDPC decoding algorithms

    Science.gov (United States)

    Liu, Mingshan; Liu, Shanshan; Zhou, Yuan; Jiang, Xue

    2015-12-01

    The conventional message-passing schedule for LDPC decoding algorithms is the so-called flooding schedule. It has the disadvantage that the updated messages cannot be used until next iteration, thus reducing the convergence speed . In this case, the Layered Decoding algorithm (LBP) based on serial message-passing schedule is proposed. In this paper the decoding principle of LBP algorithm is briefly introduced, and then proposed its two improved algorithms, the grouped serial decoding algorithm (Grouped LBP) and the semi-serial decoding algorithm .They can improve LBP algorithm's decoding speed while maintaining a good decoding performance.

  1. Improved Iterative Hard- and Soft-Reliability Based Majority-Logic Decoding Algorithms for Non-Binary Low-Density Parity-Check Codes

    Science.gov (United States)

    Xiong, Chenrong; Yan, Zhiyuan

    2014-10-01

    Non-binary low-density parity-check (LDPC) codes have some advantages over their binary counterparts, but unfortunately their decoding complexity is a significant challenge. The iterative hard- and soft-reliability based majority-logic decoding algorithms are attractive for non-binary LDPC codes, since they involve only finite field additions and multiplications as well as integer operations and hence have significantly lower complexity than other algorithms. In this paper, we propose two improvements to the majority-logic decoding algorithms. Instead of the accumulation of reliability information in the existing majority-logic decoding algorithms, our first improvement is a new reliability information update. The new update not only results in better error performance and fewer iterations on average, but also further reduces computational complexity. Since existing majority-logic decoding algorithms tend to have a high error floor for codes whose parity check matrices have low column weights, our second improvement is a re-selection scheme, which leads to much lower error floors, at the expense of more finite field operations and integer operations, by identifying periodic points, re-selecting intermediate hard decisions, and changing reliability information.

  2. Combined Source-Channel Coding of Images under Power and Bandwidth Constraints

    Directory of Open Access Journals (Sweden)

    Fossorier Marc

    2007-01-01

    Full Text Available This paper proposes a framework for combined source-channel coding for a power and bandwidth constrained noisy channel. The framework is applied to progressive image transmission using constant envelope -ary phase shift key ( -PSK signaling over an additive white Gaussian noise channel. First, the framework is developed for uncoded -PSK signaling (with . Then, it is extended to include coded -PSK modulation using trellis coded modulation (TCM. An adaptive TCM system is also presented. Simulation results show that, depending on the constellation size, coded -PSK signaling performs 3.1 to 5.2 dB better than uncoded -PSK signaling. Finally, the performance of our combined source-channel coding scheme is investigated from the channel capacity point of view. Our framework is further extended to include powerful channel codes like turbo and low-density parity-check (LDPC codes. With these powerful codes, our proposed scheme performs about one dB away from the capacity-achieving SNR value of the QPSK channel.

  3. On the Automatic Parallelization of Sparse and Irregular Fortran Programs

    Directory of Open Access Journals (Sweden)

    Yuan Lin

    1999-01-01

    Full Text Available Automatic parallelization is usually believed to be less effective at exploiting implicit parallelism in sparse/irregular programs than in their dense/regular counterparts. However, not much is really known because there have been few research reports on this topic. In this work, we have studied the possibility of using an automatic parallelizing compiler to detect the parallelism in sparse/irregular programs. The study with a collection of sparse/irregular programs led us to some common loop patterns. Based on these patterns new techniques were derived that produced good speedups when manually applied to our benchmark codes. More importantly, these parallelization methods can be implemented in a parallelizing compiler and can be applied automatically.

  4. Optimized Fast Walsh–Hadamard Transform on GPUs for non-binary LDPC decoding

    OpenAIRE

    Andrade, Joao; Falcao, Gabriel; Silva, Vitor

    2014-01-01

    The Fourier Transform Sum-Product Algorithm (FT-SPA) used in non-binary Low-Density Parity-Check (LDPC) decoding makes extensive use of the Walsh–Hadamard Transform (WHT). We have developed a massively parallel Fast Walsh–Hadamard Transform (FWHT) which exploits the Graphics Processing Unit (GPU) pipeline and memory hierarchy, thereby minimizing the level of memory bank conflicts and maximizing the number of returned instructions per clock cycle for different generations of graphics processor...

  5. Performance of Low-Density Parity-Check Coded Modulation

    Science.gov (United States)

    Hamkins, Jon

    2010-01-01

    This paper reports the simulated performance of each of the nine accumulate-repeat-4-jagged-accumulate (AR4JA) low-density parity-check (LDPC) codes [3] when used in conjunction with binary phase-shift-keying (BPSK), quadrature PSK (QPSK), 8-PSK, 16-ary amplitude PSK (16- APSK), and 32-APSK.We also report the performance under various mappings of bits to modulation symbols, 16-APSK and 32-APSK ring scalings, log-likelihood ratio (LLR) approximations, and decoder variations. One of the simple and well-performing LLR approximations can be expressed in a general equation that applies to all of the modulation types.

  6. Combined Source-Channel Coding of Images under Power and Bandwidth Constraints

    Directory of Open Access Journals (Sweden)

    Marc Fossorier

    2007-01-01

    Full Text Available This paper proposes a framework for combined source-channel coding for a power and bandwidth constrained noisy channel. The framework is applied to progressive image transmission using constant envelope M-ary phase shift key (M-PSK signaling over an additive white Gaussian noise channel. First, the framework is developed for uncoded M-PSK signaling (with M=2k. Then, it is extended to include coded M-PSK modulation using trellis coded modulation (TCM. An adaptive TCM system is also presented. Simulation results show that, depending on the constellation size, coded M-PSK signaling performs 3.1 to 5.2 dB better than uncoded M-PSK signaling. Finally, the performance of our combined source-channel coding scheme is investigated from the channel capacity point of view. Our framework is further extended to include powerful channel codes like turbo and low-density parity-check (LDPC codes. With these powerful codes, our proposed scheme performs about one dB away from the capacity-achieving SNR value of the QPSK channel.

  7. Co-operation of digital nonlinear equalizers and soft-decision LDPC FEC in nonlinear transmission.

    Science.gov (United States)

    Tanimura, Takahito; Oda, Shoichiro; Hoshida, Takeshi; Aoki, Yasuhiko; Tao, Zhenning; Rasmussen, Jens C

    2013-12-30

    We experimentally and numerically investigated the characteristics of 128 Gb/s dual polarization - quadrature phase shift keying signals received with two types of nonlinear equalizers (NLEs) followed by soft-decision (SD) low-density parity-check (LDPC) forward error correction (FEC). Successful co-operation among SD-FEC and NLEs over various nonlinear transmissions were demonstrated by optimization of parameters for NLEs.

  8. Performance of sparse graph codes on a four-dimensional CDMA System in AWGN and multipath fading

    CSIR Research Space (South Africa)

    Vlok, JD

    2007-09-01

    Full Text Available (bit) = 1× 10−5. Index Terms—Block Turbo codes (BTC), complex spread- ing sequences (CSS), channel modelling, log-likelihood ratio (LLR), low-density parity-check (LDPC) codes, multi-layered- modulation (MLM), multi-dimensional (MD), repeat...~ ~ ~ ~ 1,3 −1 1,2 −1 1,2 1,2 1,3 1,3 2,3 2,3 −1 Fig. 4. Three-dimensional block turbo decoder structure The output of SISO module m is a 3D cube ΛE,m;m = 1, 2, 3, containing the extrinsic log-likelihood ratio (LLR) of each data bit xk...

  9. Progressive transmission of images over fading channels using rate-compatible LDPC codes.

    Science.gov (United States)

    Pan, Xiang; Banihashemi, Amir H; Cuhadar, Aysegul

    2006-12-01

    In this paper, we propose a combined source/channel coding scheme for transmission of images over fading channels. The proposed scheme employs rate-compatible low-density parity-check codes along with embedded image coders such as JPEG2000 and set partitioning in hierarchical trees (SPIHT). The assignment of channel coding rates to source packets is performed by a fast trellis-based algorithm. We examine the performance of the proposed scheme over correlated and uncorrelated Rayleigh flat-fading channels with and without side information. Simulation results for the expected peak signal-to-noise ratio of reconstructed images, which are within 1 dB of the capacity upper bound over a wide range of channel signal-to-noise ratios, show considerable improvement compared to existing results under similar conditions. We also study the sensitivity of the proposed scheme in the presence of channel estimation error at the transmitter and demonstrate that under most conditions our scheme is more robust compared to existing schemes.

  10. A Novel Modified Algorithm with Reduced Complexity LDPC Code Decoder

    Directory of Open Access Journals (Sweden)

    Song Yang

    2014-10-01

    Full Text Available A novel efficient decoding algorithm reduced the sum-product algorithm (SPA Complexity with LPDC code is proposed. Base on the hyperbolic tangent rule, modified the Check node update with two horizontal process, which have similar calculation, Motivated by the finding that sun- min (MS algorithm reduce the complexity reducing the approximation error in the horizontal process, simplify the information weight small part. Compared with the exiting approximations, the proposed method is less computational complexity than SPA algorithm. Simulation results show that the author algorithm can achieve performance very close SPA.

  11. Validation of AEGIS/SCOPE2 system through actual core follow calculations with irregular operational conditions

    International Nuclear Information System (INIS)

    Tabuchi, M.; Tatsumi, M.; Ohoka, Y.; Nagano, H.; Ishizaki, K.

    2017-01-01

    This paper describes overview of AEGIS/SCOPE2 system, an advanced in-core fuel management system for pressurized water reactors, and its validation results of actual core follow calculations including irregular operational conditions. AEGIS and SCOPE2 codes adopt more detailed and accurate calculation models compared to the current core design codes while computational cost is minimized with various techniques on numerical and computational algorithms. Verification and validation of AEGIS/SCOPE2 has been intensively performed to confirm validity of the system. As a part of the validation, core follow calculations have been carried out mainly for typical operational conditions. After the Fukushima Daiichi nuclear power plant accident, however, all the nuclear reactors in Japan suffered from long suspension and irregular operational conditions. In such situations, measured data in the restart and operation of the reactors should be good examinations for validation of the codes. Therefore, core follow calculations were carried out with AEGIS/SCOPE2 for various cases including zero power reactor physics tests with irregular operational conditions. Comparisons between measured data and predictions by AEGIS/SCOPE2 revealed the validity and robustness of the system. (author)

  12. Performance and Complexity Evaluation of Iterative Receiver for Coded MIMO-OFDM Systems

    Directory of Open Access Journals (Sweden)

    Rida El Chall

    2016-01-01

    Full Text Available Multiple-input multiple-output (MIMO technology in combination with channel coding technique is a promising solution for reliable high data rate transmission in future wireless communication systems. However, these technologies pose significant challenges for the design of an iterative receiver. In this paper, an efficient receiver combining soft-input soft-output (SISO detection based on low-complexity K-Best (LC-K-Best decoder with various forward error correction codes, namely, LTE turbo decoder and LDPC decoder, is investigated. We first investigate the convergence behaviors of the iterative MIMO receivers to determine the required inner and outer iterations. Consequently, the performance of LC-K-Best based receiver is evaluated in various LTE channel environments and compared with other MIMO detection schemes. Moreover, the computational complexity of the iterative receiver with different channel coding techniques is evaluated and compared with different modulation orders and coding rates. Simulation results show that LC-K-Best based receiver achieves satisfactory performance-complexity trade-offs.

  13. Comments on “Techniques and Architectures for Hazard-Free Semi-Parallel Decoding of LDPC Codes”

    Directory of Open Access Journals (Sweden)

    Mark B. Yeary

    2009-01-01

    Full Text Available This is a comment article on the publication “Techniques and Architectures for Hazard-Free Semi-Parallel Decoding of LDPC Codes” Rovini et al. (2009. We mention that there has been similar work reported in the literature before, and the previous work has not been cited correctly, for example Gunnam et al. (2006, 2007. This brief note serves to clarify these issues.

  14. Characterizing neural activities evoked by manual acupuncture through spiking irregularity measures

    International Nuclear Information System (INIS)

    Xue Ming; Wang Jiang; Deng Bin; Wei Xi-Le; Yu Hai-Tao; Chen Ying-Yuan

    2013-01-01

    The neural system characterizes information in external stimulations by different spiking patterns. In order to examine how neural spiking patterns are related to acupuncture manipulations, experiments are designed in such a way that different types of manual acupuncture (MA) manipulations are taken at the ‘Zusanli’ point of experimental rats, and the induced electrical signals in the spinal dorsal root ganglion are detected and recorded. The interspike interval (ISI) statistical histogram is fitted by the gamma distribution, which has two parameters: one is the time-dependent firing rate and the other is a shape parameter characterizing the spiking irregularities. The shape parameter is the measure of spiking irregularities and can be used to identify the type of MA manipulations. The coefficient of variation is mostly used to measure the spike time irregularity, but it overestimates the irregularity in the case of pronounced firing rate changes. However, experiments show that each acupuncture manipulation will lead to changes in the firing rate. So we combine four relatively rate-independent measures to study the irregularity of spike trains evoked by different types of MA manipulations. Results suggest that the MA manipulations possess unique spiking statistics and characteristics and can be distinguished according to the spiking irregularity measures. These studies have offered new insights into the coding processes and information transfer of acupuncture. (interdisciplinary physics and related areas of science and technology)

  15. Irregular Applications: Architectures & Algorithms

    Energy Technology Data Exchange (ETDEWEB)

    Feo, John T.; Villa, Oreste; Tumeo, Antonino; Secchi, Simone

    2012-02-06

    Irregular applications are characterized by irregular data structures, control and communication patterns. Novel irregular high performance applications which deal with large data sets and require have recently appeared. Unfortunately, current high performance systems and software infrastructures executes irregular algorithms poorly. Only coordinated efforts by end user, area specialists and computer scientists that consider both the architecture and the software stack may be able to provide solutions to the challenges of modern irregular applications.

  16. Diseño de decodificadores de altas prestaciones para código LDPC

    OpenAIRE

    Angarita Preciado, Fabian Enrique

    2013-01-01

    En esta tesis se han investigado los algoritmos de decodificación para códigos de comprobación de paridad de baja densidad (LDPC) y las arquitecturas para la implementación hardware de éstos. El trabajo realizado se centra en los algoritmos del tipo de intercambio de mensajes para códigos estructurados los cuales se incluyen en varios estándares de comunicaciones. Inicialmente se han evaluado las prestaciones de los algoritmos existentes Sum-product, Min-Sum y las principales variantes de...

  17. Constellation labeling optimization for bit-interleaved coded APSK

    Science.gov (United States)

    Xiang, Xingyu; Mo, Zijian; Wang, Zhonghai; Pham, Khanh; Blasch, Erik; Chen, Genshe

    2016-05-01

    This paper investigates the constellation and mapping optimization for amplitude phase shift keying (APSK) modulation, which is deployed in Digital Video Broadcasting Satellite - Second Generation (DVB-S2) and Digital Video Broadcasting - Satellite services to Handhelds (DVB-SH) broadcasting standards due to its merits of power and spectral efficiency together with the robustness against nonlinear distortion. The mapping optimization is performed for 32-APSK according to combined cost functions related to Euclidean distance and mutual information. A Binary switching algorithm and its modified version are used to minimize the cost function and the estimated error between the original and received data. The optimized constellation mapping is tested by combining DVB-S2 standard Low-Density Parity-Check (LDPC) codes in both Bit-Interleaved Coded Modulation (BICM) and BICM with iterative decoding (BICM-ID) systems. The simulated results validate the proposed constellation labeling optimization scheme which yields better performance against conventional 32-APSK constellation defined in DVB-S2 standard.

  18. Deriving the Normalized Min-Sum Algorithm from Cooperative Optimization

    OpenAIRE

    Huang, Xiaofei

    2006-01-01

    The normalized min-sum algorithm can achieve near-optimal performance at decoding LDPC codes. However, it is a critical question to understand the mathematical principle underlying the algorithm. Traditionally, people thought that the normalized min-sum algorithm is a good approximation to the sum-product algorithm, the best known algorithm for decoding LDPC codes and Turbo codes. This paper offers an alternative approach to understand the normalized min-sum algorithm. The algorithm is derive...

  19. Potts glass reflection of the decoding threshold for qudit quantum error correcting codes

    Science.gov (United States)

    Jiang, Yi; Kovalev, Alexey A.; Pryadko, Leonid P.

    We map the maximum likelihood decoding threshold for qudit quantum error correcting codes to the multicritical point in generalized Potts gauge glass models, extending the map constructed previously for qubit codes. An n-qudit quantum LDPC code, where a qudit can be involved in up to m stabilizer generators, corresponds to a ℤd Potts model with n interaction terms which can couple up to m spins each. We analyze general properties of the phase diagram of the constructed model, give several bounds on the location of the transitions, bounds on the energy density of extended defects (non-local analogs of domain walls), and discuss the correlation functions which can be used to distinguish different phases in the original and the dual models. This research was supported in part by the Grants: NSF PHY-1415600 (AAK), NSF PHY-1416578 (LPP), and ARO W911NF-14-1-0272 (LPP).

  20. A novel method of the image processing on irregular triangular meshes

    Science.gov (United States)

    Vishnyakov, Sergey; Pekhterev, Vitaliy; Sokolova, Elizaveta

    2018-04-01

    The paper describes a novel method of the image processing based on irregular triangular meshes implementation. The triangular mesh is adaptive to the image content, least mean square linear approximation is proposed for the basic interpolation within the triangle. It is proposed to use triangular numbers to simplify using of the local (barycentric) coordinates for the further analysis - triangular element of the initial irregular mesh is to be represented through the set of the four equilateral triangles. This allows to use fast and simple pixels indexing in local coordinates, e.g. "for" or "while" loops for access to the pixels. Moreover, representation proposed allows to use discrete cosine transform of the simple "rectangular" symmetric form without additional pixels reordering (as it is used for shape-adaptive DCT forms). Furthermore, this approach leads to the simple form of the wavelet transform on triangular mesh. The results of the method application are presented. It is shown that advantage of the method proposed is a combination of the flexibility of the image-adaptive irregular meshes with the simple form of the pixel indexing in local triangular coordinates and the using of the common forms of the discrete transforms for triangular meshes. Method described is proposed for the image compression, pattern recognition, image quality improvement, image search and indexing. It also may be used as a part of video coding (intra-frame or inter-frame coding, motion detection).

  1. Analysis of Mid-Latitude Plasma Density Irregularities in the Presence of Finite Larmor Radius Effects

    Science.gov (United States)

    Sotnikov, V. I.; Kim, T. C.; Mishin, E. V.; Kil, H.; Kwak, Y. S.; Paraschiv, I.

    2017-12-01

    Ionospheric irregularities cause scintillations of electromagnetic signals that can severely affect navigation and transionospheric communication, in particular during space storms. At mid-latitudes the source of F-region Field Aligned Irregularities (FAI) is yet to be determined. They can be created in enhanced subauroral flow channels (SAI/SUBS), where strong gradients of electric field, density and plasma temperature are present. Another important source of FAI is connected with Medium-scale travelling ionospheric disturbances (MSTIDs). Related shear flows and plasma density troughs point to interchange and Kelvin-Helmholtz type instabilities as a possible source of plasma irregularities. A model of nonlinear development of these instabilities based on the two-fluid hydrodynamic description with inclusion of finite Larmor radius effects will be presented. This approach allows to resolve density irregularities on the meter scale. A numerical code in C language to solve the derived nonlinear equations for analysis of interchange and flow velocity shear instabilities in the ionosphere was developed. This code will be used to analyze competition between interchange and Kelvin-Helmholtz instabilities in the mid-latitude region. The high-resolution simulations with continuous density and velocity profiles will be driven by the ambient conditions corresponding to the in situ data obtained during the 2016 Daejeon (Korea) and MU (Japan) radar campaign and data collected simultaneously by the Swarm satellites passed over Korea and Japan. PA approved #: 88ABW-2017-3641

  2. PSEUDO-CODEWORD LANDSCAPE

    Energy Technology Data Exchange (ETDEWEB)

    CHERTKOV, MICHAEL [Los Alamos National Laboratory; STEPANOV, MIKHAIL [Los Alamos National Laboratory

    2007-01-10

    The authors discuss performance of Low-Density-Parity-Check (LDPC) codes decoded by Linear Programming (LP) decoding at moderate and large Signal-to-Noise-Ratios (SNR). Frame-Error-Rate (FER) dependence on SNR and the noise space landscape of the coding/decoding scheme are analyzed by a combination of the previously introduced instanton/pseudo-codeword-search method and a new 'dendro' trick. To reduce complexity of the LP decoding for a code with high-degree checks, {ge} 5, they introduce its dendro-LDPC counterpart, that is the code performing identifically to the original one under Maximum-A-Posteriori (MAP) decoding but having reduced (down to three) check connectivity degree. Analyzing number of popular LDPC codes and their dendro versions performing over the Additive-White-Gaussian-Noise (AWGN) channel, they observed two qualitatively different regimes: (i) error-floor sets early, at relatively low SNR, and (ii) FER decays with SNR increase faster at moderate SNR than at the largest SNR. They explain these regimes in terms of the pseudo-codeword spectra of the codes.

  3. Star Formation in Irregular Galaxies.

    Science.gov (United States)

    Hunter, Deidre; Wolff, Sidney

    1985-01-01

    Examines mechanisms of how stars are formed in irregular galaxies. Formation in giant irregular galaxies, formation in dwarf irregular galaxies, and comparisons with larger star-forming regions found in spiral galaxies are considered separately. (JN)

  4. Parallel iterative decoding of transform domain Wyner-Ziv video using cross bitplane correlation

    DEFF Research Database (Denmark)

    Luong, Huynh Van; Huang, Xin; Forchhammer, Søren

    2011-01-01

    decoding scheme is proposed to improve the coding efficiency of TDWZ video codecs. The proposed parallel iterative LDPC decoding scheme is able to utilize cross bitplane correlation during decoding, by iteratively refining the soft-input, updating a modeled noise distribution and thereafter enhancing......In recent years, Transform Domain Wyner-Ziv (TDWZ) video coding has been proposed as an efficient Distributed Video Coding (DVC) solution, which fully or partly exploits the source statistics at the decoder to reduce the computational burden at the encoder. In this paper, a parallel iterative LDPC...

  5. Beiträge zur 6. Diskussionssitzung der ITG-Fachgruppe "Angewandte Informationstheorie"

    OpenAIRE

    Meurer, Michael

    2005-01-01

    Langvorträge: T. Schorr, A. Dittrich, W. Sauer-Greff, R. Urbansky (Lehrstuhl für Nachrichtentechnik, TU Kaiserslautern): Iterative Equalization in Fibre Optical Systems Using High-Rate RCPR, BCH and LDPC Codes A. Doenmez, T. Hehn, J. B. Huber (Lehrstuhl für Informationsübertragung, Universität Erlangen-Nürnberg): Analytical Calculation of Thresholds for LDPC Codes transmitted over Binary Erasure Channels S. Deng, T. Weber (Institut für Nachrichtentechnik und Informationselektronik, Universitä...

  6. Photonic entanglement-assisted quantum low-density parity-check encoders and decoders.

    Science.gov (United States)

    Djordjevic, Ivan B

    2010-05-01

    I propose encoder and decoder architectures for entanglement-assisted (EA) quantum low-density parity-check (LDPC) codes suitable for all-optical implementation. I show that two basic gates needed for EA quantum error correction, namely, controlled-NOT (CNOT) and Hadamard gates can be implemented based on Mach-Zehnder interferometer. In addition, I show that EA quantum LDPC codes from balanced incomplete block designs of unitary index require only one entanglement qubit to be shared between source and destination.

  7. Identification of Sparse Audio Tampering Using Distributed Source Coding and Compressive Sensing Techniques

    Directory of Open Access Journals (Sweden)

    Valenzise G

    2009-01-01

    Full Text Available In the past few years, a large amount of techniques have been proposed to identify whether a multimedia content has been illegally tampered or not. Nevertheless, very few efforts have been devoted to identifying which kind of attack has been carried out, especially due to the large data required for this task. We propose a novel hashing scheme which exploits the paradigms of compressive sensing and distributed source coding to generate a compact hash signature, and we apply it to the case of audio content protection. The audio content provider produces a small hash signature by computing a limited number of random projections of a perceptual, time-frequency representation of the original audio stream; the audio hash is given by the syndrome bits of an LDPC code applied to the projections. At the content user side, the hash is decoded using distributed source coding tools. If the tampering is sparsifiable or compressible in some orthonormal basis or redundant dictionary, it is possible to identify the time-frequency position of the attack, with a hash size as small as 200 bits/second; the bit saving obtained by introducing distributed source coding ranges between 20% to 70%.

  8. A new Fortran 90 program to compute regular and irregular associated Legendre functions (new version announcement)

    Science.gov (United States)

    Schneider, Barry I.; Segura, Javier; Gil, Amparo; Guan, Xiaoxu; Bartschat, Klaus

    2018-04-01

    This is a revised and updated version of a modern Fortran 90 code to compute the regular Plm (x) and irregular Qlm (x) associated Legendre functions for all x ∈(- 1 , + 1) (on the cut) and | x | > 1 and integer degree (l) and order (m). The necessity to revise the code comes as a consequence of some comments of Prof. James Bremer of the UC//Davis Mathematics Department, who discovered that there were errors in the code for large integer degree and order for the normalized regular Legendre functions on the cut.

  9. Spherical reconciliation for a continuous-variable quantum key distribution

    International Nuclear Information System (INIS)

    Lu Zhao; Shi Jian-Hong; Li Feng-Guang

    2017-01-01

    Information reconciliation is a significant step for a continuous-variable quantum key distribution (CV-QKD) system. We propose a reconciliation method that allows two authorized parties to extract a consistent and secure binary key in a CV-QKD protocol, which is based on Gaussian-modulated coherent states and homodyne detection. This method named spherical reconciliation is based on spherical quantization and non-binary low-density parity-check (LDPC) codes. With the suitable signal-to-noise ratio (SNR) and code rate of non-binary LDPC codes, spherical reconciliation algorithm has a high efficiency and can extend the transmission distance of CV-QKD. (paper)

  10. Development of a calculation methodology for potential flow over irregular topographies

    International Nuclear Information System (INIS)

    Del Carmen, Alejandra F.; Ferreri, Juan C.; Boutet, Luis I.

    2003-01-01

    Full text: Computer codes for the calculation of potential flow fields over surfaces with irregular topographies have been developed. The flows past multiple simple obstacles and past the neighboring region of the Embalse Nuclear Power Station have been considered. The codes developed allow the calculation of velocities quite near the surface. It, in turn, imposed developing high accuracy techniques. The Boundary Element Method, using a linear approximation on triangular plane elements and an analytical integration methodology has been applied. A particular and quite efficient technique for the calculation of the solid angle at each node vertex was also considered. The results so obtained will be applied to predict the dispersion of passive pollutants coming from discontinuous emissions. (authors)

  11. On the optimum signal constellation design for high-speed optical transport networks.

    Science.gov (United States)

    Liu, Tao; Djordjevic, Ivan B

    2012-08-27

    In this paper, we first describe an optimum signal constellation design algorithm, which is optimum in MMSE-sense, called MMSE-OSCD, for channel capacity achieving source distribution. Secondly, we introduce a feedback channel capacity inspired optimum signal constellation design (FCC-OSCD) to further improve the performance of MMSE-OSCD, inspired by the fact that feedback channel capacity is higher than that of systems without feedback. The constellations obtained by FCC-OSCD are, however, OSNR dependent. The optimization is jointly performed together with regular quasi-cyclic low-density parity-check (LDPC) code design. Such obtained coded-modulation scheme, in combination with polarization-multiplexing, is suitable as both 400 Gb/s and multi-Tb/s optical transport enabling technology. Using large girth LDPC code, we demonstrate by Monte Carlo simulations that a 32-ary signal constellation, obtained by FCC-OSCD, outperforms previously proposed optimized 32-ary CIPQ signal constellation by 0.8 dB at BER of 10(-7). On the other hand, the LDPC-coded 16-ary FCC-OSCD outperforms 16-QAM by 1.15 dB at the same BER.

  12. Capture of irregular satellites at Jupiter

    International Nuclear Information System (INIS)

    Nesvorný, David; Vokrouhlický, David; Deienno, Rogerio

    2014-01-01

    The irregular satellites of outer planets are thought to have been captured from heliocentric orbits. The exact nature of the capture process, however, remains uncertain. We examine the possibility that irregular satellites were captured from the planetesimal disk during the early solar system instability when encounters between the outer planets occurred. Nesvorný et al. already showed that the irregular satellites of Saturn, Uranus, and Neptune were plausibly captured during planetary encounters. Here we find that the current instability models present favorable conditions for capture of irregular satellites at Jupiter as well, mainly because Jupiter undergoes a phase of close encounters with an ice giant. We show that the orbital distribution of bodies captured during planetary encounters provides a good match to the observed distribution of irregular satellites at Jupiter. The capture efficiency for each particle in the original transplanetary disk is found to be (1.3-3.6) × 10 –8 . This is roughly enough to explain the observed population of jovian irregular moons. We also confirm Nesvorný et al.'s results for the irregular satellites of Saturn, Uranus, and Neptune.

  13. Irregular Migrants and the Law

    OpenAIRE

    Kassim, Azizah; Mat Zin, Ragayah Hj.

    2013-01-01

    This paper examines Malaysia`s policy on irregular migrants and its implementation, and discusses its impact. A survey and interview covering 404 respondents was conducted between July 2010 and June 2011 to ascertain the real situations surrounding irregular migrants in Malaysia, which is one of the major host countries of international migrants from developing nations. The policy on foreign workers was formulated in the mid-1980s to deal with the large number of irregular migrants and their ...

  14. Finite Length Analysis of Irregular Repetition Slotted ALOHA in the Waterfall Region

    OpenAIRE

    Amat, Alexandre Graell i; Liva, Gianluigi

    2018-01-01

    A finite length analysis is introduced for irregular repetition slotted ALOHA (IRSA) that enables to accurately estimate its performance in the moderate-to-high packet loss probability regime, i.e., in the so-called waterfall region. The analysis is tailored to the collision channel model, which enables mapping the description of the successive interference cancellation process onto the iterative erasure decoding of low-density parity-check codes. The analysis provides accurate estimates of t...

  15. A rate-compatible family of protograph-based LDPC codes built by expurgation and lengthening

    Science.gov (United States)

    Dolinar, Sam

    2005-01-01

    We construct a protograph-based rate-compatible family of low-density parity-check codes that cover a very wide range of rates from 1/2 to 16/17, perform within about 0.5 dB of their capacity limits for all rates, and can be decoded conveniently and efficiently with a common hardware implementation.

  16. Fast algorithm for two-dimensional data table use in hydrodynamic and radiative-transfer codes

    International Nuclear Information System (INIS)

    Slattery, W.L.; Spangenberg, W.H.

    1982-01-01

    A fast algorithm for finding interpolated atomic data in irregular two-dimensional tables with differing materials is described. The algorithm is tested in a hydrodynamic/radiative transfer code and shown to be of comparable speed to interpolation in regularly spaced tables, which require no table search. The concepts presented are expected to have application in any situation with irregular vector lengths. Also, the procedures that were rejected either because they were too slow or because they involved too much assembly coding are described

  17. BER reduction for mQAM high-speed transmission in optical fiber ...

    African Journals Online (AJOL)

    SANOUSerge

    Indeed, OFDM using mQAM modulation seems to be a good candidate in high ... and Matlab software. The simulation results showed that error correcting codes and particularly LDPC codes are .... channel coding as the process consist on.

  18. Comparison of soft-input-soft-output detection methods for dual-polarized quadrature duobinary system

    Science.gov (United States)

    Chang, Chun; Huang, Benxiong; Xu, Zhengguang; Li, Bin; Zhao, Nan

    2018-02-01

    Three soft-input-soft-output (SISO) detection methods for dual-polarized quadrature duobinary (DP-QDB), including maximum-logarithmic-maximum-a-posteriori-probability-algorithm (Max-log-MAP)-based detection, soft-output-Viterbi-algorithm (SOVA)-based detection, and a proposed SISO detection, which can all be combined with SISO decoding, are presented. The three detection methods are investigated at 128 Gb/s in five-channel wavelength-division-multiplexing uncoded and low-density-parity-check (LDPC) coded DP-QDB systems by simulations. Max-log-MAP-based detection needs the returning-to-initial-states (RTIS) process despite having the best performance. When the LDPC code with a code rate of 0.83 is used, the detecting-and-decoding scheme with the SISO detection does not need RTIS and has better bit error rate (BER) performance than the scheme with SOVA-based detection. The former can reduce the optical signal-to-noise ratio (OSNR) requirement (at BER=10-5) by 2.56 dB relative to the latter. The application of the SISO iterative detection in LDPC-coded DP-QDB systems makes a good trade-off between requirements on transmission efficiency, OSNR requirement, and transmission distance, compared with the other two SISO methods.

  19. Capacity Maximizing Constellations

    Science.gov (United States)

    Barsoum, Maged; Jones, Christopher

    2010-01-01

    Some non-traditional signal constellations have been proposed for transmission of data over the Additive White Gaussian Noise (AWGN) channel using such channel-capacity-approaching codes as low-density parity-check (LDPC) or turbo codes. Computational simulations have shown performance gains of more than 1 dB over traditional constellations. These gains could be translated to bandwidth- efficient communications, variously, over longer distances, using less power, or using smaller antennas. The proposed constellations have been used in a bit-interleaved coded modulation system employing state-ofthe-art LDPC codes. In computational simulations, these constellations were shown to afford performance gains over traditional constellations as predicted by the gap between the parallel decoding capacity of the constellations and the Gaussian capacity

  20. Stability analysis by ERATO code

    International Nuclear Information System (INIS)

    Tsunematsu, Toshihide; Takeda, Tatsuoki; Matsuura, Toshihiko; Azumi, Masafumi; Kurita, Gen-ichi

    1979-12-01

    Problems in MHD stability calculations by ERATO code are described; which concern convergence property of results, equilibrium codes, and machine optimization of ERATO code. It is concluded that irregularity on a convergence curve is not due to a fault of the ERATO code itself but due to inappropriate choice of the equilibrium calculation meshes. Also described are a code to calculate an equilibrium as a quasi-inverse problem and a code to calculate an equilibrium as a result of a transport process. Optimization of the code with respect to I/O operations reduced both CPU time and I/O time considerably. With the FACOM230-75 APU/CPU multiprocessor system, the performance is about 6 times as high as with the FACOM230-75 CPU, showing the effectiveness of a vector processing computer for the kind of MHD computations. This report is a summary of the material presented at the ERATO workshop 1979(ORNL), supplemented with some details. (author)

  1. Two media method for linear attenuation coefficient determination of irregular soil samples

    International Nuclear Information System (INIS)

    Vici, Carlos Henrique Georges

    2004-01-01

    In several situations of nuclear applications, the knowledge of gamma-ray linear attenuation coefficient for irregular samples is necessary, such as in soil physics and geology. This work presents the validation of a methodology for the determination of the linear attenuation coefficient (μ) of irregular shape samples, in such a way that it is not necessary to know the thickness of the considered sample. With this methodology irregular soil samples (undeformed field samples) from Londrina region, north of Parana were studied. It was employed the two media method for the μ determination. It consists of the μ determination through the measurement of a gamma-ray beam attenuation by the sample sequentially immersed in two different media, with known and appropriately chosen attenuation coefficients. For comparison, the theoretical value of μ was calculated by the product of the mass attenuation coefficient, obtained by the WinXcom code, and the measured value of the density sample. This software employs the chemical composition of the samples and supplies a table of the mass attenuation coefficients versus the photon energy. To verify the validity of the two media method, compared with the simple gamma ray transmission method, regular pome stone samples were used. With these results for the attenuation coefficients and their respective deviations, it was possible to compare the two methods. In this way we concluded that the two media method is a good tool for the determination of the linear attenuation coefficient of irregular materials, particularly in the study of soils samples. (author)

  2. Measurements of linear attenuation coefficients of irregular shaped samples by two media method

    International Nuclear Information System (INIS)

    Singh, Sukhpal; Kumar, Ashok; Thind, Kulwant Singh; Mudahar, Gurmel S.

    2008-01-01

    The linear attenuation coefficient values of regular and irregular shaped flyash materials have been measured without knowing the thickness of a sample using a new technique namely 'two media method'. These values have also been measured with a standard gamma ray transmission method and obtained theoretically with winXCOM computer code. From the comparison it is reported that the two media method has given accurate results of attenuation coefficients of flyash materials

  3. Strategic Analysis of Irregular Warfare

    Science.gov (United States)

    2010-03-01

    the same mathematical equations used by Lanchester .10 Irregular Warfare Theory and Doctrine It is time to develop new analytical methods and models...basis on which to build, similar to what Lanchester provided almost 100 years ago. Figure 9 portrays both Lanchester’s approach and an irregular 17

  4. Irregular Migration in Jordan, 1995-2007

    OpenAIRE

    AROURI, Fathi A.

    2008-01-01

    Euro-Mediterranean Consortium for Applied Research on International Migration (CARIM) This paper tackles the question of irregular migration in Jordan through its four main aspects. The first concerns irregular labour migrants and has been approached by using figures showing the socio-economic profile of non Jordanians working in Jordan and, additionally, unemployment in Jordan. This is done by assuming close similarities between legal and irregular labour migrants. The second is an attemp...

  5. Two media method for linear attenuation coefficient determination of irregular soil samples; Metodo dos dois meios para a determinacao do coeficiente de atenuacao linear de amostras irregulares de solos

    Energy Technology Data Exchange (ETDEWEB)

    Vici, Carlos Henrique Georges

    2004-07-01

    In several situations of nuclear applications, the knowledge of gamma-ray linear attenuation coefficient for irregular samples is necessary, such as in soil physics and geology. This work presents the validation of a methodology for the determination of the linear attenuation coefficient ({mu}) of irregular shape samples, in such a way that it is not necessary to know the thickness of the considered sample. With this methodology irregular soil samples (undeformed field samples) from Londrina region, north of Parana were studied. It was employed the two media method for the {mu} determination. It consists of the {mu} determination through the measurement of a gamma-ray beam attenuation by the sample sequentially immersed in two different media, with known and appropriately chosen attenuation coefficients. For comparison, the theoretical value of {mu} was calculated by the product of the mass attenuation coefficient, obtained by the WinXcom code, and the measured value of the density sample. This software employs the chemical composition of the samples and supplies a table of the mass attenuation coefficients versus the photon energy. To verify the validity of the two media method, compared with the simple gamma ray transmission method, regular pome stone samples were used. With these results for the attenuation coefficients and their respective deviations, it was possible to compare the two methods. In this way we concluded that the two media method is a good tool for the determination of the linear attenuation coefficient of irregular materials, particularly in the study of soils samples. (author)

  6. Performance Comparisons of Improved Regular Repeat Accumulate (RA and Irregular Repeat Accumulate (IRA Turbo Decoding

    Directory of Open Access Journals (Sweden)

    Ahmed Abdulkadhim Hamad

    2017-08-01

    Full Text Available In this paper, different techniques are used to improve the turbo decoding of regular repeat accumulate (RA and irregular repeat accumulate (IRA codes. The adaptive scaling of a-posteriori information produced by Soft-output Viterbi decoder (SOVA is proposed. The encoded pilots are another scheme that applied for short length RA codes. This work also suggests a simple and a fast method to generate a random interleaver having a free 4 cycle Tanner graph. Progressive edge growth algorithm (PEG is also studied and simulated to create the Tanner graphs which have a great girth.

  7. Landsat Data Continuity Mission (LDCM) - Optimizing X-Band Usage

    Science.gov (United States)

    Garon, H. M.; Gal-Edd, J. S.; Dearth, K. W.; Sank, V. I.

    2010-01-01

    The NASA version of the low-density parity check (LDPC) 7/8-rate code, shortened to the dimensions of (8160, 7136), has been implemented as the forward error correction (FEC) schema for the Landsat Data Continuity Mission (LDCM). This is the first flight application of this code. In order to place a 440 Msps link within the 375 MHz wide X band we found it necessary to heavily bandpass filter the satellite transmitter output . Despite the significant amplitude and phase distortions that accompanied the spectral truncation, the mission required BER is maintained at LDPC code and the amplitude and phase compensation provided in the receiver. Similar results were obtained with receivers from several vendors.

  8. Fitting Irregular Shape Figures into Irregular Shape Areas for the Nesting Problem in the Leather Industry

    Directory of Open Access Journals (Sweden)

    Guevara-Palma Luis

    2015-01-01

    Full Text Available The nesting problem of irregular shapes within irregular areas has been studied from several approaches due to their application in different industries. The particular case of cutting leather involves several restrictions that add complexity to this problem, it is necessary to generate products that comply with the quality required by customers This paper presents a methodology for the accommodation of irregular shapes in an irregular area (leather considering the constraints set by the footwear industry, and the results of this methodology when applied by a computer system. The scope of the system is to develop a working prototype that operates under the guidelines of a commercial production line of a sponsor company. Preliminary results got a reduction of 70% of processing time and improvement of 5% to 7% of the area usage when compared with manual accommodation.

  9. FPGA implementation of advanced FEC schemes for intelligent aggregation networks

    Science.gov (United States)

    Zou, Ding; Djordjevic, Ivan B.

    2016-02-01

    In state-of-the-art fiber-optics communication systems the fixed forward error correction (FEC) and constellation size are employed. While it is important to closely approach the Shannon limit by using turbo product codes (TPC) and low-density parity-check (LDPC) codes with soft-decision decoding (SDD) algorithm; rate-adaptive techniques, which enable increased information rates over short links and reliable transmission over long links, are likely to become more important with ever-increasing network traffic demands. In this invited paper, we describe a rate adaptive non-binary LDPC coding technique, and demonstrate its flexibility and good performance exhibiting no error floor at BER down to 10-15 in entire code rate range, by FPGA-based emulation, making it a viable solution in the next-generation high-speed intelligent aggregation networks.

  10. The Manifestation of Stopping Sets and Absorbing Sets as Deviations on the Computation Trees of LDPC Codes

    Directory of Open Access Journals (Sweden)

    Eric Psota

    2010-01-01

    Full Text Available The error mechanisms of iterative message-passing decoders for low-density parity-check codes are studied. A tutorial review is given of the various graphical structures, including trapping sets, stopping sets, and absorbing sets that are frequently used to characterize the errors observed in simulations of iterative decoding of low-density parity-check codes. The connections between trapping sets and deviations on computation trees are explored in depth using the notion of problematic trapping sets in order to bridge the experimental and analytic approaches to these error mechanisms. A new iterative algorithm for finding low-weight problematic trapping sets is presented and shown to be capable of identifying many trapping sets that are frequently observed during iterative decoding of low-density parity-check codes on the additive white Gaussian noise channel. Finally, a new method is given for characterizing the weight of deviations that result from problematic trapping sets.

  11. Opportunistic error correction for OFDM-based DVB systems

    NARCIS (Netherlands)

    Shao, X.; Slump, Cornelis H.

    2013-01-01

    DVB-T2 (second generation terrestrial digital video broadcasting) employs LDPC (Low Density Parity Check) codes combined with BCH (Bose-Chaudhuri-Hocquengham) codes, which has a better performance in comparison to convolutional and Reed-Solomon codes used in other OFDM-based DVB systems. However,

  12. Legal aspects of the EU policy on irregular immigration

    Directory of Open Access Journals (Sweden)

    Voinikov Vadim

    2015-12-01

    Full Text Available This article addresses the issues pertaining to the adoption and development of legislation on irregular migration in the context of uncontrolled growth in the number of immigrants from North Africa and the Middle East to the EU. The article attempts at studying the EU legislation on irregular migration, classifying it, and analysing the prospects of EU migration legislation in the light of an increase in irregular immigration into the EU. The author systematises, classifies the current EU legislation on irregular immigration, and analyses the conditions, in which this legislation was developed. Using the legislation analysis method, the author proposes the following system of EU legislation on irregular immigration: rules preventing assistance to irregular immigration, rules preventing employment of irregular immigrants, rules on the return of irregular migrants and readmission, rules on border control, and rules on collaboration with third countries. The author pays special attention to analysing the current state of irregular immigration to the EU, which was dubbed the ‘greatest migration crisis in Europe’. The conclusion is that the European Union succeeded in the development of pioneering legislation on irregular immigration, which can serve as the basis for reception by other states. However, changes in the political and economic situation in the EU’s southern borderlands made the current legal mechanisms incapable of withstanding new threats. It necessitates a radical reform of the legislation on irregular immigration.

  13. Irregular menstruation according to occupational status.

    Science.gov (United States)

    Kwak, Yeunhee; Kim, Yoonjung

    2017-07-06

    This cross-sectional study explored associations of irregular menstruation with occupational characteristics, using secondary analyses of data from 4,731 women aged 19-54 years, collected from a nationally representative sample, the Korea National Health and Nutrition Examination Survey-V during 2010-2012. The associations between irregular menstruation and occupation were explored using multiple logistic regression. Compared to non-manual workers, service/sales workers had a greater odds of irregular menstruation (adjusted odds ratio [aOR]: 1.44; 95percent confidence interval [CI]: 1.04-1.99) as did manual workers and unemployed women (aOR: 1.56; 95percent CI: 1.10-2.22, aOR: 1.46; 95percent CI: 1.14-1.89, respectively). Compared to regular workers, temporary workers and unemployed women had aORs of 1.52 (95percent CI: 1.08-2.13) and 1.33 (95percent CI: 1.05-1.69), respectively. Also, when compared to full-time workers, part-time workers and unemployed women had greater odds of irregular menstruation (aOR: 1.41; 95percent CI: 1.00-2.00 and aOR: 1.29; 95percent CI: 1.03-1.63, respectively). Furthermore, compared to daytime workers, shift workers and unemployed women had greater odds irregular menstruation (aOR: 1.39; 95percent CI: 1.03-1.88 and aOR: 1.28; 95percent CI: 1.04-1.59, respectively). Women with these occupational characteristics should be screened for early diagnosis and intervention for irregular menstruation.

  14. Irregular conformal block, spectral curve and flow equations

    International Nuclear Information System (INIS)

    Choi, Sang Kwan; Rim, Chaiho; Zhang, Hong

    2016-01-01

    Irregular conformal block is motivated by the Argyres-Douglas type of N=2 super conformal gauge theory. We investigate the classical/NS limit of irregular conformal block using the spectral curve on a Riemann surface with irregular punctures, which is equivalent to the loop equation of irregular matrix model. The spectral curve is reduced to the second order (Virasoro symmetry, SU(2) for the gauge theory) and third order (W_3 symmetry, SU(3)) differential equations of a polynomial with finite degree. The conformal and W symmetry generate the flow equations in the spectral curve and determine the irregular conformal block, hence the partition function of the Argyres-Douglas theory ala AGT conjecture.

  15. 16 CFR 501.6 - Cellulose sponges, irregular dimensions.

    Science.gov (United States)

    2010-01-01

    ... 16 Commercial Practices 1 2010-01-01 2010-01-01 false Cellulose sponges, irregular dimensions. 501... REQUIREMENTS AND PROHIBITIONS UNDER PART 500 § 501.6 Cellulose sponges, irregular dimensions. Variety packages of cellulose sponges of irregular dimensions, are exempted from the requirements of § 500.25 of this...

  16. Ethical issues in irregular migration research

    NARCIS (Netherlands)

    Duvell, F.; Triandafyllidou, A.; Vollmer, B.

    2008-01-01

    This paper is concerned with the ethical issues arising for researchers engaged in the study of irregular migration. Irregular migration is by definition an elusive phenomenon as it takes place in violation of the law and at the margins of society. This very nature of the phenomenon raises important

  17. Improving soft FEC performance for higher-order modulations via optimized bit channel mappings.

    Science.gov (United States)

    Häger, Christian; Amat, Alexandre Graell I; Brännström, Fredrik; Alvarado, Alex; Agrell, Erik

    2014-06-16

    Soft forward error correction with higher-order modulations is often implemented in practice via the pragmatic bit-interleaved coded modulation paradigm, where a single binary code is mapped to a nonbinary modulation. In this paper, we study the optimization of the mapping of the coded bits to the modulation bits for a polarization-multiplexed fiber-optical system without optical inline dispersion compensation. Our focus is on protograph-based low-density parity-check (LDPC) codes which allow for an efficient hardware implementation, suitable for high-speed optical communications. The optimization is applied to the AR4JA protograph family, and further extended to protograph-based spatially coupled LDPC codes assuming a windowed decoder. Full field simulations via the split-step Fourier method are used to verify the analysis. The results show performance gains of up to 0.25 dB, which translate into a possible extension of the transmission reach by roughly up to 8%, without significantly increasing the system complexity.

  18. Locating irregularly shaped clusters of infection intensity

    Directory of Open Access Journals (Sweden)

    Niko Yiannakoulias

    2010-05-01

    Full Text Available Patterns of disease may take on irregular geographic shapes, especially when features of the physical environment influence risk. Identifying these patterns can be important for planning, and also identifying new environmental or social factors associated with high or low risk of illness. Until recently, cluster detection methods were limited in their ability to detect irregular spatial patterns, and limited to finding clusters that were roughly circular in shape. This approach has less power to detect irregularly-shaped, yet important spatial anomalies, particularly at high spatial resolutions. We employ a new method of finding irregularly-shaped spatial clusters at micro-geographical scales using both simulated and real data on Schistosoma mansoni and hookworm infection intensities. This method, which we refer to as the “greedy growth scan”, is a modification of the spatial scan method for cluster detection. Real data are based on samples of hookworm and S. mansoni from Kitengei, Makueni district, Kenya. Our analysis of simulated data shows how methods able to find irregular shapes are more likely to identify clusters along rivers than methods constrained to fixed geometries. Our analysis of infection intensity identifies two small areas within the study region in which infection intensity is elevated, possibly due to local features of the physical or social environment. Collectively, our results show that the “greedy growth scan” is a suitable method for exploratory geographical analysis of infection intensity data when irregular shapes are suspected, especially at micro-geographical scales.

  19. Detecting chaos in irregularly sampled time series.

    Science.gov (United States)

    Kulp, C W

    2013-09-01

    Recently, Wiebe and Virgin [Chaos 22, 013136 (2012)] developed an algorithm which detects chaos by analyzing a time series' power spectrum which is computed using the Discrete Fourier Transform (DFT). Their algorithm, like other time series characterization algorithms, requires that the time series be regularly sampled. Real-world data, however, are often irregularly sampled, thus, making the detection of chaotic behavior difficult or impossible with those methods. In this paper, a characterization algorithm is presented, which effectively detects chaos in irregularly sampled time series. The work presented here is a modification of Wiebe and Virgin's algorithm and uses the Lomb-Scargle Periodogram (LSP) to compute a series' power spectrum instead of the DFT. The DFT is not appropriate for irregularly sampled time series. However, the LSP is capable of computing the frequency content of irregularly sampled data. Furthermore, a new method of analyzing the power spectrum is developed, which can be useful for differentiating between chaotic and non-chaotic behavior. The new characterization algorithm is successfully applied to irregularly sampled data generated by a model as well as data consisting of observations of variable stars.

  20. Two-stage cross-talk mitigation in an orbital-angular-momentum-based free-space optical communication system.

    Science.gov (United States)

    Qu, Zhen; Djordjevic, Ivan B

    2017-08-15

    We propose and experimentally demonstrate a two-stage cross-talk mitigation method in an orbital-angular-momentum (OAM)-based free-space optical communication system, which is enabled by combining spatial offset and low-density parity-check (LDPC) coded nonuniform signaling. Different from traditional OAM multiplexing, where the OAM modes are centrally aligned for copropagation, the adjacent OAM modes (OAM states 2 and -6 and OAM states -2 and 6) in our proposed scheme are spatially offset to mitigate the mode cross talk. Different from traditional rectangular modulation formats, which transmit equidistant signal points with uniform probability, the 5-quadrature amplitude modulation (5-QAM) and 9-QAM are introduced to relieve cross-talk-induced performance degradation. The 5-QAM and 9-QAM formats are based on the Huffman coding technique, which can potentially achieve great cross-talk tolerance by combining them with corresponding nonbinary LDPC codes. We demonstrate that cross talk can be reduced by 1.6 dB and 1 dB via spatial offset for OAM states ±2 and ±6, respectively. Compared to quadrature phase shift keying and 8-QAM formats, the LDPC-coded 5-QAM and 9-QAM are able to bring 1.1 dB and 5.4 dB performance improvements in the presence of atmospheric turbulence, respectively.

  1. Fatigue damage estimation using irregularity factor. First report, irregularity factor calculations for narrow and broadband random time histories

    Science.gov (United States)

    Susuki, I.

    1981-11-01

    The results of an analysis of the irregularity factors of stationary and Gaussian random processes which are generated by filtering the output of a pure or a band-limited white noise are presented. An ideal band pass filter, a trapezoidal filter, and a Butterworth type band pass filter were examined. It was found that the values of the irregularity factors were approximately equal among these filters if only the end-slopes were the same rates. As the band width of filters increases, irregularity factors increase monotonically and approach the respective constant values depending on the end-slopes. This implies that the noise characteristics relevant to the fatigue damage such as statistical aspects of the height of the rise and fall or the distribution of the peak values are not changed for a broad band random time history. It was also found that the effect of band limitation of input white noise on irregularity factors is negligibly small.

  2. Optimal signal constellation design for ultra-high-speed optical transport in the presence of nonlinear phase noise.

    Science.gov (United States)

    Liu, Tao; Djordjevic, Ivan B

    2014-12-29

    In this paper, we first describe an optimal signal constellation design algorithm suitable for the coherent optical channels dominated by the linear phase noise. Then, we modify this algorithm to be suitable for the nonlinear phase noise dominated channels. In optimization procedure, the proposed algorithm uses the cumulative log-likelihood function instead of the Euclidian distance. Further, an LDPC coded modulation scheme is proposed to be used in combination with signal constellations obtained by proposed algorithm. Monte Carlo simulations indicate that the LDPC-coded modulation schemes employing the new constellation sets, obtained by our new signal constellation design algorithm, outperform corresponding QAM constellations significantly in terms of transmission distance and have better nonlinearity tolerance.

  3. An Efficient Downlink Scheduling Strategy Using Normal Graphs for Multiuser MIMO Wireless Systems

    Science.gov (United States)

    Chen, Jung-Chieh; Wu, Cheng-Hsuan; Lee, Yao-Nan; Wen, Chao-Kai

    Inspired by the success of the low-density parity-check (LDPC) codes in the field of error-control coding, in this paper we propose transforming the downlink multiuser multiple-input multiple-output scheduling problem into an LDPC-like problem using the normal graph. Based on the normal graph framework, soft information, which indicates the probability that each user will be scheduled to transmit packets at the access point through a specified angle-frequency sub-channel, is exchanged among the local processors to iteratively optimize the multiuser transmission schedule. Computer simulations show that the proposed algorithm can efficiently schedule simultaneous multiuser transmission which then increases the overall channel utilization and reduces the average packet delay.

  4. On irregularity strength of disjoint union of friendship graphs

    Directory of Open Access Journals (Sweden)

    Ali Ahmad

    2013-11-01

    Full Text Available We investigate the vertex total and edge total modication of the well-known irregularity strength of graphs. We have determined the exact values of the total vertex irregularity strength and the total edge irregularity strength of a disjoint union of friendship graphs.

  5. BER Reduction for mQAM High-Speed Tranmission in Optical Fiber ...

    African Journals Online (AJOL)

    The study was conducted in a software cosimulation environment with VPITransmissionMaker and Matlab software. The simulation results showed that error correcting codes and particularly LDPC codes are effective and provide satisfactory solutions to reduce the BER by fighting against optical transmission channel ...

  6. Characteristics of ionospheric irregularities causing scintillations at VHF/UHF

    International Nuclear Information System (INIS)

    Vats, H.O.; Deshpande, M.R.; Rastogi, R.G.

    1978-01-01

    Some properties of ionization irregularities using amplitude scintillation records of radio beacons from ATS-6 (phase II) at Ootacamund, India have been investigated. For the estimation of scale-size and strength of the irregularities a simple diffraction model has been used which explains only weak and moderate equatorial scintillation observations. It was found that the scale sizes of day time E-region irregularities are smaller than those in the F-region during night time in addition, irregularities are generated initially at large scale sizes which later break up into smaller scale sizes

  7. 14 CFR 135.65 - Reporting mechanical irregularities.

    Science.gov (United States)

    2010-01-01

    ... irregularities and their correction. (b) The pilot in command shall enter or have entered in the aircraft maintenance log each mechanical irregularity that comes to the pilot's attention during flight time. Before each flight, the pilot in command shall, if the pilot does not already know, determine the status of...

  8. General Lines of Disregard for the Legal Personality on Irregular Dissolution the Company

    Directory of Open Access Journals (Sweden)

    Fábio Augusto Barcelos Moreira Corrêa

    2016-12-01

    Full Text Available This article will analyze the Institute of disregard for the legal personality in situations involving irregular dissolution the limited liability company, in light of the jurisprudence of the Superior Court Tribunal. We highlight the impact that new code of Civil procedure will provide for analysis to safeguard the autonomy of assets of the legal person, as well as the guarantee of due process and of ample defense, directly impacting on business law. The analysis aims to contribute to the understanding of the Institute, and the systematic procedure. Adopting the dialectical methodology and criticism.

  9. SimCommSys: taking the errors out of error-correcting code simulations

    Directory of Open Access Journals (Sweden)

    Johann A. Briffa

    2014-06-01

    Full Text Available In this study, we present SimCommSys, a simulator of communication systems that we are releasing under an open source license. The core of the project is a set of C + + libraries defining communication system components and a distributed Monte Carlo simulator. Of principal interest is the error-control coding component, where various kinds of binary and non-binary codes are implemented, including turbo, LDPC, repeat-accumulate and Reed–Solomon. The project also contains a number of ready-to-build binaries implementing various stages of the communication system (such as the encoder and decoder, a complete simulator and a system benchmark. Finally, SimCommSys also provides a number of shell and python scripts to encapsulate routine use cases. As long as the required components are already available in SimCommSys, the user may simulate complete communication systems of their own design without any additional programming. The strict separation of development (needed only to implement new components and use (to simulate specific constructions encourages reproducibility of experimental work and reduces the likelihood of error. Following an overview of the framework, we provide some examples of how to use the framework, including the implementation of a simple codec, the specification of communication systems and their simulation.

  10. Orbital and Collisional Evolution of the Irregular Satellites

    Science.gov (United States)

    Nesvorný, David; Alvarellos, Jose L. A.; Dones, Luke; Levison, Harold F.

    2003-07-01

    The irregular moons of the Jovian planets are a puzzling part of the solar system inventory. Unlike regular satellites, the irregular moons revolve around planets at large distances in tilted and eccentric orbits. Their origin, which is intimately linked with the origin of the planets themselves, is yet to be explained. Here we report a study of the orbital and collisional evolution of the irregular satellites from times after their formation to the present epoch. The purpose of this study is to find out the features of the observed irregular moons that can be attributed to this evolution and separate them from signatures of the formation process. We numerically integrated ~60,000 test satellite orbits to map orbital locations that are stable on long time intervals. We found that the orbits highly inclined to the ecliptic are unstable due to the effect of the Kozai resonance, which radially stretches them so that satellites either escape from the Hill sphere, collide with massive inner moons, or impact the parent planet. We also found that prograde satellite orbits with large semimajor axes are unstable due to the effect of the evection resonance, which locks the orbit's apocenter to the apparent motion of the Sun around the parent planet. In such a resonance, the effect of solar tides on a resonant moon accumulates at each apocenter passage of the moon, which causes a radially outward drift of its orbital apocenter; once close to the Hill sphere, the moon escapes. By contrast, retrograde moons with large orbital semimajor axes are long-lived. We have developed an analytic model of the distant satellite orbits and used it to explain the results of our numerical experiments. In particular, we analytically studied the effect of the Kozai resonance. We numerically integrated the orbits of the 50 irregular moons (known by 2002 August 16) for 108 yr. All orbits were stable on this time interval and did not show any macroscopic variations that would indicate

  11. Decomposing Oriented Graphs into Six Locally Irregular Oriented Graphs

    DEFF Research Database (Denmark)

    Bensmail, Julien; Renault, Gabriel

    2016-01-01

    An undirected graph G is locally irregular if every two of its adjacent vertices have distinct degrees. We say that G is decomposable into k locally irregular graphs if there exists a partition E1∪E2∪⋯∪Ek of the edge set E(G) such that each Ei induces a locally irregular graph. It was recently co...

  12. Design Optimization of Irregular Cellular Structure for Additive Manufacturing

    Science.gov (United States)

    Song, Guo-Hua; Jing, Shi-Kai; Zhao, Fang-Lei; Wang, Ye-Dong; Xing, Hao; Zhou, Jing-Tao

    2017-09-01

    Irregularcellular structurehas great potential to be considered in light-weight design field. However, the research on optimizing irregular cellular structures has not yet been reporteddue to the difficulties in their modeling technology. Based on the variable density topology optimization theory, an efficient method for optimizing the topology of irregular cellular structures fabricated through additive manufacturing processes is proposed. The proposed method utilizes tangent circles to automatically generate the main outline of irregular cellular structure. The topological layoutof each cellstructure is optimized using the relative density informationobtained from the proposed modified SIMP method. A mapping relationship between cell structure and relative densityelement is builtto determine the diameter of each cell structure. The results show that the irregular cellular structure can be optimized with the proposed method. The results of simulation and experimental test are similar for irregular cellular structure, which indicate that the maximum deformation value obtained using the modified Solid Isotropic Microstructures with Penalization (SIMP) approach is lower 5.4×10-5 mm than that using the SIMP approach under the same under the same external load. The proposed research provides the instruction to design the other irregular cellular structure.

  13. Design strategies for irregularly adapting parallel applications

    International Nuclear Information System (INIS)

    Oliker, Leonid; Biswas, Rupak; Shan, Hongzhang; Sing, Jaswinder Pal

    2000-01-01

    Achieving scalable performance for dynamic irregular applications is eminently challenging. Traditional message-passing approaches have been making steady progress towards this goal; however, they suffer from complex implementation requirements. The use of a global address space greatly simplifies the programming task, but can degrade the performance of dynamically adapting computations. In this work, we examine two major classes of adaptive applications, under five competing programming methodologies and four leading parallel architectures. Results indicate that it is possible to achieve message-passing performance using shared-memory programming techniques by carefully following the same high level strategies. Adaptive applications have computational work loads and communication patterns which change unpredictably at runtime, requiring dynamic load balancing to achieve scalable performance on parallel machines. Efficient parallel implementations of such adaptive applications are therefore a challenging task. This work examines the implementation of two typical adaptive applications, Dynamic Remeshing and N-Body, across various programming paradigms and architectural platforms. We compare several critical factors of the parallel code development, including performance, programmability, scalability, algorithmic development, and portability

  14. Plasma Irregularity Production in the Polar Cap F-Region Ionosphere

    Science.gov (United States)

    Lamarche, Leslie

    Plasma in the Earth's ionosphere is highly irregular on scales ranging between a few centimeters and hundreds of kilometers. Small-scale irregularities or plasma waves can scatter radio waves resulting in a loss of signal for navigation and communication networks. The polar region is particularly susceptible to strong disturbances due to its direct connection with the Sun's magnetic field and energetic particles. In this thesis, factors that contribute to the production of decameter-scale plasma irregularities in the polar F region ionosphere are investigated. Both global and local control of irregularity production are studied, i.e. we consider global solar control through solar illumination and solar wind as well as much more local control by plasma density gradients and convection electric field. In the first experimental study, solar control of irregularity production is investigated using the Super Dual Auroral Radar Network (SuperDARN) radar at McMurdo, Antarctica. The occurrence trends for irregularities are analyzed statistically and a model is developed that describes the location of radar echoes within the radar's field-of-view. The trends are explained through variations in background plasma density with solar illumination affecting radar beam propagation. However, it is found that the irregularity occurrence during the night is higher than expected from ray tracing simulations based on a standard ionospheric density model. The high occurrence at night implies an additional source of plasma density and it is proposed that large-scale density enhancements called polar patches may be the source of this density. Additionally, occurrence maximizes around the terminator due to different competing irregularity production processes that favor a more or less sunlit ionosphere. The second study is concerned with modeling irregularity characteristics near a large-scale density gradient reversal, such as those expected near polar patches, with a particular focus on

  15. Irregular activity arises as a natural consequence of synaptic inhibition

    International Nuclear Information System (INIS)

    Terman, D.; Rubin, J. E.; Diekman, C. O.

    2013-01-01

    Irregular neuronal activity is observed in a variety of brain regions and states. This work illustrates a novel mechanism by which irregular activity naturally emerges in two-cell neuronal networks featuring coupling by synaptic inhibition. We introduce a one-dimensional map that captures the irregular activity occurring in our simulations of conductance-based differential equations and mathematically analyze the instability of fixed points corresponding to synchronous and antiphase spiking for this map. We find that the irregular solutions that arise exhibit expansion, contraction, and folding in phase space, as expected in chaotic dynamics. Our analysis shows that these features are produced from the interplay of synaptic inhibition with sodium, potassium, and leak currents in a conductance-based framework and provides precise conditions on parameters that ensure that irregular activity will occur. In particular, the temporal details of spiking dynamics must be present for a model to exhibit this irregularity mechanism and must be considered analytically to capture these effects

  16. Irregular activity arises as a natural consequence of synaptic inhibition

    Energy Technology Data Exchange (ETDEWEB)

    Terman, D., E-mail: terman@math.ohio-state.edu [Department of Mathematics, The Ohio State University, Columbus, Ohio 43210 (United States); Rubin, J. E., E-mail: jonrubin@pitt.edu [Department of Mathematics, University of Pittsburgh, Pittsburgh, Pennsylvania 15260 (United States); Diekman, C. O., E-mail: diekman@njit.edu [Department of Mathematical Sciences, New Jersey Institute of Technology, Newark, New Jersey 07102 (United States)

    2013-12-15

    Irregular neuronal activity is observed in a variety of brain regions and states. This work illustrates a novel mechanism by which irregular activity naturally emerges in two-cell neuronal networks featuring coupling by synaptic inhibition. We introduce a one-dimensional map that captures the irregular activity occurring in our simulations of conductance-based differential equations and mathematically analyze the instability of fixed points corresponding to synchronous and antiphase spiking for this map. We find that the irregular solutions that arise exhibit expansion, contraction, and folding in phase space, as expected in chaotic dynamics. Our analysis shows that these features are produced from the interplay of synaptic inhibition with sodium, potassium, and leak currents in a conductance-based framework and provides precise conditions on parameters that ensure that irregular activity will occur. In particular, the temporal details of spiking dynamics must be present for a model to exhibit this irregularity mechanism and must be considered analytically to capture these effects.

  17. Parallel Computing Strategies for Irregular Algorithms

    Science.gov (United States)

    Biswas, Rupak; Oliker, Leonid; Shan, Hongzhang; Biegel, Bryan (Technical Monitor)

    2002-01-01

    Parallel computing promises several orders of magnitude increase in our ability to solve realistic computationally-intensive problems, but relies on their efficient mapping and execution on large-scale multiprocessor architectures. Unfortunately, many important applications are irregular and dynamic in nature, making their effective parallel implementation a daunting task. Moreover, with the proliferation of parallel architectures and programming paradigms, the typical scientist is faced with a plethora of questions that must be answered in order to obtain an acceptable parallel implementation of the solution algorithm. In this paper, we consider three representative irregular applications: unstructured remeshing, sparse matrix computations, and N-body problems, and parallelize them using various popular programming paradigms on a wide spectrum of computer platforms ranging from state-of-the-art supercomputers to PC clusters. We present the underlying problems, the solution algorithms, and the parallel implementation strategies. Smart load-balancing, partitioning, and ordering techniques are used to enhance parallel performance. Overall results demonstrate the complexity of efficiently parallelizing irregular algorithms.

  18. Advances in electron dosimetry of irregular fields; Avances en dosimetria de electrones de campos irregulares

    Energy Technology Data Exchange (ETDEWEB)

    Mendez V, J. [Departamento de Radioterapia, Instituto de Enfermedades Neoplasicas, Avenida Angamos Este 2520, Lima 34 (Peru)

    1998-12-31

    In this work it is presented an advance in Electron dosimetry of irregular fields for beams emitted by linear accelerators. At present diverse methods exist which are coming to apply in the Radiotherapy centers. In this work it is proposed a method for irregular fields dosimetry. It will be allow to calculate the dose rate absorbed required for evaluating the time for the treatment of cancer patients. Utilizing the results obtained by the dosimetric system, it has been possible to prove the validity of the method describe for 12 MeV energy and for square field 7.5 x 7.5 cm{sup 2} with percentile error less than 1 % . (Author)

  19. Role of parametric decay instabilities in generating ionospheric irregularities

    International Nuclear Information System (INIS)

    Kuo, S.P.; Cheo, B.R.; Lee, M.C.

    1983-01-01

    We show that purely growing instabilities driven by the saturation spectrum of parametric decay instabilities can produce a broad spectrum of ionospheric irregularities. The threshold field Vertical BarE/sub th/Vertical Bar of the instabilities decreases with the scale lengths lambda of the ionospheric irregularities as Vertical BarE/sub th/Vertical Barproportionallambda -2 in the small-scale range ( -2 with scale lengths larger than a few kilometers. The excitation of kilometer-scale irregularities is strictly restricted by the instabilities themselves and by the spatial inhomogeneity of the medium. These results are drawn from the analyses of four-wave interaction. Ion-neutral collisions impose no net effect on the instabilities when the excited ionospheric irregularities have a field-aligned nature

  20. Joint Schemes for Physical Layer Security and Error Correction

    Science.gov (United States)

    Adamo, Oluwayomi

    2011-01-01

    The major challenges facing resource constraint wireless devices are error resilience, security and speed. Three joint schemes are presented in this research which could be broadly divided into error correction based and cipher based. The error correction based ciphers take advantage of the properties of LDPC codes and Nordstrom Robinson code. A…

  1. Synchronous spikes are necessary but not sufficient for a synchrony code in populations of spiking neurons.

    Science.gov (United States)

    Grewe, Jan; Kruscha, Alexandra; Lindner, Benjamin; Benda, Jan

    2017-03-07

    Synchronous activity in populations of neurons potentially encodes special stimulus features. Selective readout of either synchronous or asynchronous activity allows formation of two streams of information processing. Theoretical work predicts that such a synchrony code is a fundamental feature of populations of spiking neurons if they operate in specific noise and stimulus regimes. Here we experimentally test the theoretical predictions by quantifying and comparing neuronal response properties in tuberous and ampullary electroreceptor afferents of the weakly electric fish Apteronotus leptorhynchus These related systems show similar levels of synchronous activity, but only in the more irregularly firing tuberous afferents a synchrony code is established, whereas in the more regularly firing ampullary afferents it is not. The mere existence of synchronous activity is thus not sufficient for a synchrony code. Single-cell features such as the irregularity of spiking and the frequency dependence of the neuron's transfer function determine whether synchronous spikes possess a distinct meaning for the encoding of time-dependent signals.

  2. A novel chaotic encryption scheme based on arithmetic coding

    International Nuclear Information System (INIS)

    Mi Bo; Liao Xiaofeng; Chen Yong

    2008-01-01

    In this paper, under the combination of arithmetic coding and logistic map, a novel chaotic encryption scheme is presented. The plaintexts are encrypted and compressed by using an arithmetic coder whose mapping intervals are changed irregularly according to a keystream derived from chaotic map and plaintext. Performance and security of the scheme are also studied experimentally and theoretically in detail

  3. New Opportunities for Remote Sensing Ionospheric Irregularities by Fitting Scintillation Spectra

    Science.gov (United States)

    Carrano, C. S.; Rino, C. L.; Groves, K. M.

    2017-12-01

    In a recent paper, we presented a phase screen theory for the spectrum of intensity scintillations when the refractive index irregularities follow a two-component power law [Carrano and Rino, DOI: 10.1002/2015RS005903]. More recently we have investigated the inverse problem, whereby phase screen parameters are inferred from scintillation time series. This is accomplished by fitting the spectrum of intensity fluctuations with a parametrized theoretical model using Maximum Likelihood (ML) methods. The Markov-Chain Monte-Carlo technique provides a-posteriori errors and confidence intervals. The Akaike Information Criterion (AIC) provides justification for the use of one- or two-component irregularity models. We refer to this fitting as Irregularity Parameter Estimation (IPE) since it provides a statistical description of the irregularities from the scintillations they produce. In this talk, we explore some new opportunities for remote sensing ionospheric irregularities afforded by IPE. Statistical characterization of irregularities and the plasma bubbles in which they are embedded provides insight into the development of the underlying instability. In a companion paper by Rino et al., IPE is used to interpret scintillation due to simulated EPB structure. IPE can be used to reconcile multi-frequency scintillation observations and to construct high fidelity scintillation simulation tools. In space-to-ground propagation scenarios, for which an estimate of the distance to the scattering region is available a-priori, IPE enables retrieval of zonal irregularity drift. In radio occultation scenarios, the distance to the irregularities is generally unknown but IPE enables retrieval of Fresnel frequency. A geometric model for the effective scan velocity maps Fresnel frequency to Fresnel scale, yielding the distance to the irregularities. We demonstrate this approach by geolocating irregularities observed by the CORISS instrument onboard the C/NOFS satellite.

  4. State reconstruction and irregular wavefunctions for the hydrogen atom

    Science.gov (United States)

    Krähmer, D. S.; Leonhardt, U.

    1997-07-01

    Inspired by a recently proposed procedure by Leonhardt and Raymer for wavepacket reconstruction, we calculate the irregular wavefunctions for the bound states of the Coulomb potential. We select the irregular solutions which have the simplest semiclassical limit.

  5. Worst configurations (instantons) for compressed sensing over reals: a channel coding approach

    International Nuclear Information System (INIS)

    Chertkov, Michael; Chilappagari, Shashi K.; Vasic, Bane

    2010-01-01

    We consider Linear Programming (LP) solution of a Compressed Sensing (CS) problem over reals, also known as the Basis Pursuit (BasP) algorithm. The BasP allows interpretation as a channel-coding problem, and it guarantees the error-free reconstruction over reals for properly chosen measurement matrix and sufficiently sparse error vectors. In this manuscript, we examine how the BasP performs on a given measurement matrix and develop a technique to discover sparse vectors for which the BasP fails. The resulting algorithm is a generalization of our previous results on finding the most probable error-patterns, so called instantons, degrading performance of a finite size Low-Density Parity-Check (LDPC) code in the error-floor regime. The BasP fails when its output is different from the actual error-pattern. We design CS-Instanton Search Algorithm (ISA) generating a sparse vector, called CS-instanton, such that the BasP fails on the instanton, while its action on any modification of the CS-instanton decreasing a properly defined norm is successful. We also prove that, given a sufficiently dense random input for the error-vector, the CS-ISA converges to an instanton in a small finite number of steps. Performance of the CS-ISA is tested on example of a randomly generated 512 * 120 matrix, that outputs the shortest instanton (error vector) pattern of length 11.

  6. On a new process for cusp irregularity production

    Directory of Open Access Journals (Sweden)

    H. C. Carlson

    2008-09-01

    Full Text Available Two plasma instability mechanisms were thought until 2007 to dominate the formation of plasma irregularities in the F region high latitude and polar ionosphere; the gradient-drift driven instability, and the velocity-shear driven instability. The former mechanism was accepted as accounting for plasma structuring in polar cap patches, the latter for plasma structuring in polar cap sun aligned arcs. Recent work has established the need to replace this view of the past two decades with a new patch plasma structuring process (not a new mechanism, whereby shear-driven instabilities first rapidly structure the entering plasma, after which gradient drift instabilities build on these large "seed" irregularities. Correct modeling of cusp and early polar cap patch structuring will not be accomplished without allowing for this compound process. This compound process explains several previously unexplained characteristics of cusp and early polar cap patch irregularities. Here we introduce additional data, coincident in time and space, to extend that work to smaller irregularity scale sizes and relate it to the structured cusp current system.

  7. Ionospheric Irregularities at Mars Probed by MARSIS Topside Sounding

    Science.gov (United States)

    Harada, Y.; Gurnett, D. A.; Kopf, A. J.; Halekas, J. S.; Ruhunusiri, S.

    2018-01-01

    The upper ionosphere of Mars contains a variety of perturbations driven by solar wind forcing from above and upward propagating atmospheric waves from below. Here we explore the global distribution and variability of ionospheric irregularities around the exobase at Mars by analyzing topside sounding data from the Mars Advanced Radar for Subsurface and Ionosphere Sounding (MARSIS) instrument on board Mars Express. As irregular structure gives rise to off-vertical echoes with excess propagation time, the diffuseness of ionospheric echo traces can be used as a diagnostic tool for perturbed reflection surfaces. The observed properties of diffuse echoes above unmagnetized regions suggest that ionospheric irregularities with horizontal wavelengths of tens to hundreds of kilometers are particularly enhanced in the winter hemisphere and at high solar zenith angles. Given the known inverse dependence of neutral gravity wave amplitudes on the background atmospheric temperature, the ionospheric irregularities probed by MARSIS are most likely associated with plasma perturbations driven by atmospheric gravity waves. Though extreme events with unusually diffuse echoes are more frequently observed for high solar wind dynamic pressures during some time intervals, the vast majority of the diffuse echo events are unaffected by varying solar wind conditions, implying limited influence of solar wind forcing on the generation of ionospheric irregularities. Combination of remote and in situ measurements of ionospheric irregularities would offer the opportunity for a better understanding of the ionospheric dynamics at Mars.

  8. Bottomside sinusoidal irregularities in the equatorial F region

    Science.gov (United States)

    Valladares, C. E.; Hanson, W. B.; Mcclure, J. P.; Cragin, B. L.

    1983-01-01

    By using the Ogo 6 satellite, McClure and Hanson (1973) have discovered sinusoidal irregularities in the equatorial F region ion number density. In the present investigation, a description is provided of the properties of a distinct category of sinusoidal irregularities found in equatorial data from the AE-C and AE-E satellites. The observed scale sizes vary from about 300 m to 3 km in the direction perpendicular to B, overlapping with and extending the range observed by using Ogo 6. Attention is given to low and high resolution data, a comparison with Huancayo ionograms, the confinement of 'bottomside sinusoidal' (BSS) irregularities essentially to the bottomside of the F layer, spectral characteristics, and BSS, scintillation, and ionosonde observations.

  9. The Impact of Irregular Warfare on the US Army

    National Research Council Canada - National Science Library

    McDonald, III, Roger L

    2006-01-01

    Although the U.S. Army has yet to clearly define irregular warfare, it is imperative that the Army take near-term action to enhance the ability of Soldiers and units to operate effectively in an irregular warfare environment...

  10. Influence of initial stress, irregularity and heterogeneity on Love-type wave propagation in double pre-stressed irregular layers lying over a pre-stressed half-space

    Science.gov (United States)

    Singh, Abhishek Kumar; Das, Amrita; Parween, Zeenat; Chattopadhyay, Amares

    2015-10-01

    The present paper deals with the propagation of Love-type wave in an initially stressed irregular vertically heterogeneous layer lying over an initially stressed isotropic layer and an initially stressed isotropic half-space. Two different types of irregularities, viz., rectangular and parabolic, are considered at the interface of uppermost initially stressed heterogeneous layer and intermediate initially stressed isotropic layer. Dispersion equations are obtained in closed form for both cases of irregularities, distinctly. The effect of size and shape of irregularity, horizontal compressive initial stress, horizontal tensile initial stress, heterogeneity of the uppermost layer and width ratio of the layers on phase velocity of Love-type wave are the major highlights of the study. Comparative study has been made to identify the effects of different shapes of irregularity, presence of heterogeneity and initial stresses. Numerical computations have been carried out and depicted by means of graphs for the present study.

  11. Analysis on applicable error-correcting code strength of storage class memory and NAND flash in hybrid storage

    Science.gov (United States)

    Matsui, Chihiro; Kinoshita, Reika; Takeuchi, Ken

    2018-04-01

    A hybrid of storage class memory (SCM) and NAND flash is a promising technology for high performance storage. Error correction is inevitable on SCM and NAND flash because their bit error rate (BER) increases with write/erase (W/E) cycles, data retention, and program/read disturb. In addition, scaling and multi-level cell technologies increase BER. However, error-correcting code (ECC) degrades storage performance because of extra memory reading and encoding/decoding time. Therefore, applicable ECC strength of SCM and NAND flash is evaluated independently by fixing ECC strength of one memory in the hybrid storage. As a result, weak BCH ECC with small correctable bit is recommended for the hybrid storage with large SCM capacity because SCM is accessed frequently. In contrast, strong and long-latency LDPC ECC can be applied to NAND flash in the hybrid storage with large SCM capacity because large-capacity SCM improves the storage performance.

  12. Numerical simulations of type II gradient drift irregularities in the equatorial electrojet

    International Nuclear Information System (INIS)

    Ferch, R.L.; Sudan, R.N.

    1977-01-01

    Two-dimensional numerical studies of the development of type II irregularities in the equatorial electrojet have been carried out using a method similar to that of McDonald et al., (1974) except that ion inertia has been neglected. This simplification is shown to be a valid approximation whenever the electron drift velocity is small in comparison with the ion acoustic velocity and the values of the other parameters are those appropriate for the equatorial E layer. This code enables us to follow the development of quasi-steady state turbulence from appropriate initial pertubations. The two-dimensional turbulent spectrum of electron density perturbations excited is studied both for the case of devlopment from initial perturbations and for the case of a continuously pumped single driving wave

  13. Ultraviolet Communication for Medical Applications

    Science.gov (United States)

    2014-05-01

    MATLAB features were coded but not used in the test bench. First, FEC using low density parity checking ( LDPC ) codes was implemented but not used...modulation, and demodulation. MATLAB code is freely available for encryption, and a toolbox is available for the Zephyr Bioharness to capture streams... MATLAB 5 MATLAB communication system code was implemented in a basic, functional way for the first milestone test. Tested data rate was 50 kbps and

  14. Software support for irregular and loosely synchronous problems

    Science.gov (United States)

    Choudhary, A.; Fox, G.; Hiranandani, S.; Kennedy, K.; Koelbel, C.; Ranka, S.; Saltz, J.

    1992-01-01

    A large class of scientific and engineering applications may be classified as irregular and loosely synchronous from the perspective of parallel processing. We present a partial classification of such problems. This classification has motivated us to enhance FORTRAN D to provide language support for irregular, loosely synchronous problems. We present techniques for parallelization of such problems in the context of FORTRAN D.

  15. Irregular Shaped Building Design Optimization with Building Information Modelling

    Directory of Open Access Journals (Sweden)

    Lee Xia Sheng

    2016-01-01

    Full Text Available This research is to recognise the function of Building Information Modelling (BIM in design optimization for irregular shaped buildings. The study focuses on a conceptual irregular shaped “twisted” building design similar to some existing sculpture-like architectures. Form and function are the two most important aspects of new buildings, which are becoming more sophisticated as parts of equally sophisticated “systems” that we are living in. Nowadays, it is common to have irregular shaped or sculpture-like buildings which are very different when compared to regular buildings. Construction industry stakeholders are facing stiff challenges in many aspects such as buildability, cost effectiveness, delivery time and facility management when dealing with irregular shaped building projects. Building Information Modelling (BIM is being utilized to enable architects, engineers and constructors to gain improved visualization for irregular shaped buildings; this has a purpose of identifying critical issues before initiating physical construction work. In this study, three variations of design options differing in rotating angle: 30 degrees, 60 degrees and 90 degrees are created to conduct quantifiable comparisons. Discussions are focused on three major aspects including structural planning, usable building space, and structural constructability. This research concludes that Building Information Modelling is instrumental in facilitating design optimization for irregular shaped building. In the process of comparing different design variations, instead of just giving “yes or no” type of response, stakeholders can now easily visualize, evaluate and decide to achieve the right balance based on their own criteria. Therefore, construction project stakeholders are empowered with superior evaluation and decision making capability.

  16. Recent developments of the MARC/PN transport theory code including a treatment of anisotropic scatter

    International Nuclear Information System (INIS)

    Fletcher, J.K.

    1987-12-01

    The computer code MARC/PN provides a solution of the multigroup transport equation by expanding the flux in spherical harmonics. The coefficients of the series so obtained satisfy linked first order differential equations, and on eliminating terms associated with odd parity harmonics a second order system results which can be solved by established finite difference or finite element techniques. This report describes modifications incorporated in MARC/PN to allow for anisotropic scattering, and the modelling of irregular exterior boundaries in the finite element option. The latter development leads to substantial reductions in problem size, particularly for three dimensions. Also, links to an interactive graphics mesh generator (SUPERTAB) have been added. The final section of the report contains results from problems showing the effects of anisotropic scatter and the ability of the code to model irregular geometries. (author)

  17. Saturn's Irregular Moon Ymir

    Science.gov (United States)

    Denk, Tilmann; Mottola, S.

    2012-10-01

    Ymir (diameter 18 km), Saturn's second largest retrograde outer or irregular moon, has been observed six times by the Cassini narrow-angle camera (NAC) during the first 7 months in 2012. The observations span phase angles from 2° up to 102° and were taken at ranges between 15 and 18 million kilometers. From such a distance, Ymir is smaller than a pixel in the Cassini NAC. The data reveal a sidereal rotation period of 11.93 hrs, which is 1.6x longer than the previously reported value (Denk et al. 2011, EPSC/DPS #1452). Reason for this discrepancy is that the rotational light curve shows a rather uncommon 3-maxima and 3-minima shape at least in the phase angle range 50° to 100°, which was not recognizable in earlier data. The data cover several rotations from different viewing and illumination geometries and allow for a convex shape inversion with possibly a unique solution for the pole direction. The model reproduces the observed light curves to a very good accuracy without requiring albedo variegation, thereby suggesting that the lightcurve is dominated by the shape of Ymir. Among Saturn's irregular moons, the phenomenon of more than two maxima and minima at moderate to high phase angles is not unique to Ymir. At least Siarnaq and Paaliaq also show light curves with a strong deviation from a double-sine curve. Their rotation periods, however, remain unknown until more data can be taken. The light curve of Phoebe is fundamentally different to Ymir's because it is mainly shaped by local albedo differences and not by shape. Other reliable rotation periods of irregular satellites measured by Cassini include: Mundilfari 6.74 h; Kari 7.70 h; Albiorix 13.32 h; Kiviuq 21.82 h. More uncertain values are: Skathi 12 h; Bebhionn 16 h; Thrymr 27 h; Erriapus 28 h.

  18. Uniform irradiation of irregularly shaped cavities for photodynamic therapy

    NARCIS (Netherlands)

    Rem, A. I.; van Gemert, M. J.; van der Meulen, F. W.; Gijsbers, G. H.; Beek, J. F.

    1997-01-01

    It is difficult to achieve a uniform light distribution in irregularly shaped cavities. We have conducted a study on the use of hollow 'integrating' moulds for more uniform light delivery of photodynamic therapy in irregularly shaped cavities such as the oral cavity. Simple geometries such as a

  19. Spatial irregularities in Jupiter's upper ionosphere observed by Voyager radio occultations

    Science.gov (United States)

    Hinson, D. P.; Tyler, G. L.

    1982-01-01

    Radio scintillations (at 3.6 and 13 cm) produced by scattering from ionospheric irregularities during the Voyager occultations are interpreted using a weak-scattering theory. Least squares solutions for ionospheric parameters derived from the observed fluctuation spectra yield estimates of (1) the axial ratio, (2) angular orientation of the anisotropic irregularities, (3) the power law exponent of the spatial spectrum of irregularities, and (4) the magnitude of the spatial variations in electron density. It is shown that the measured angular orientation of the anisotropic irregularities indicates magnetic field direction and may provide a basis for refining Jovian magnetic field models.

  20. Irregular Dwarf Galaxy IC 1613

    Science.gov (United States)

    2005-01-01

    Ultraviolet image (left) and visual image (right) of the irregular dwarf galaxy IC 1613. Low surface brightness galaxies, such as IC 1613, are more easily detected in the ultraviolet because of the low background levels compared to visual wavelengths.

  1. Regularisation of irregular verbs in child English second language ...

    African Journals Online (AJOL)

    Data was collected from the language of English medium preschool children. The study concludes that when the Blocking Principle interferes, children resort to a novel interlanguage rule that regularises irregular verbs. This interlanguage rule applies in a similar way to all irregular verbs, thus children produce utterances ...

  2. Low frequency sound reproduction in irregular rooms using CABS (Control Acoustic Bass System)

    DEFF Research Database (Denmark)

    Celestinos, Adrian; Nielsen, Sofus Birkedal

    2011-01-01

    of an irregular room model using the FDTD (Finite Difference Time Domain) method has been presented. CABS has been simulated in the irregular room model. Measurements of CABS in a real irregular room have been performed. The performance of CABS was affected by the irregular shape of the room due to the corner...

  3. High energy model for irregular absorbing particles

    International Nuclear Information System (INIS)

    Chiappetta, Pierre.

    1979-05-01

    In the framework of a high energy formulation of relativistic quantum scattering a model is presented which describes the scattering functions and polarization of irregular absorbing particles, whose dimensions are greater than the incident wavelength. More precisely in the forward direction an amplitude parametrization of eikonal type is defined which generalizes the usual diffraction theory, and in the backward direction a reflective model is used including a shadow function. The model predictions are in good agreement with the scattering measurements off irregular compact and fluffy particles performed by Zerull, Giese and Weiss (1977)

  4. Total edge irregularity strength of (n,t)-kite graph

    Science.gov (United States)

    Winarsih, Tri; Indriati, Diari

    2018-04-01

    Let G(V, E) be a simple, connected, and undirected graph with vertex set V and edge set E. A total k-labeling is a map that carries vertices and edges of a graph G into a set of positive integer labels {1, 2, …, k}. An edge irregular total k-labeling λ :V(G)\\cup E(G)\\to \\{1,2,\\ldots,k\\} of a graph G is a labeling of vertices and edges of G in such a way that for any different edges e and f, weights wt(e) and wt(f) are distinct. The weight wt(e) of an edge e = xy is the sum of the labels of vertices x and y and the label of the edge e. The total edge irregularity strength of G, tes(G), is defined as the minimum k for which a graph G has an edge irregular total k-labeling. An (n, t)-kite graph consist of a cycle of length n with a t-edge path (the tail) attached to one vertex of a cycle. In this paper, we investigate the total edge irregularity strength of the (n, t)-kite graph, with n > 3 and t > 1. We obtain the total edge irregularity strength of the (n, t)-kite graph is tes((n, t)-kite) = \\lceil \\frac{n+t+2}{3}\\rceil .

  5. Classical limit of irregular blocks and Mathieu functions

    International Nuclear Information System (INIS)

    Piątek, Marcin; Pietrykowski, Artur R.

    2016-01-01

    The Nekrasov-Shatashvili limit of the N = 2 SU(2) pure gauge (Ω-deformed) super Yang-Mills theory encodes the information about the spectrum of the Mathieu operator. On the other hand, the Mathieu equation emerges entirely within the frame of two-dimensional conformal field theory (2d CFT) as the classical limit of the null vector decoupling equation for some degenerate irregular block. Therefore, it seems to be possible to investigate the spectrum of the Mathieu operator employing the techniques of 2d CFT. To exploit this strategy, a full correspondence between the Mathieu equation and its realization within 2d CFT has to be established. In our previous paper http://dx.doi.org/10.1007/JHEP12(2014)032, we have found that the expression of the Mathieu eigenvalue given in terms of the classical irregular block exactly coincides with the well known weak coupling expansion of this eigenvalue in the case in which the auxiliary parameter is the noninteger Floquet exponent. In the present work we verify that the formula for the corresponding eigenfunction obtained from the irregular block reproduces the so-called Mathieu exponent from which the noninteger order elliptic cosine and sine functions may be constructed. The derivation of the Mathieu equation within the formalism of 2d CFT is based on conjectures concerning the asymptotic behaviour of irregular blocks in the classical limit. A proof of these hypotheses is sketched. Finally, we speculate on how it could be possible to use the methods of 2d CFT in order to get from the irregular block the eigenvalues of the Mathieu operator in other regions of the coupling constant.

  6. Track Irregularity Time Series Analysis and Trend Forecasting

    Directory of Open Access Journals (Sweden)

    Jia Chaolong

    2012-01-01

    Full Text Available The combination of linear and nonlinear methods is widely used in the prediction of time series data. This paper analyzes track irregularity time series data by using gray incidence degree models and methods of data transformation, trying to find the connotative relationship between the time series data. In this paper, GM (1,1 is based on first-order, single variable linear differential equations; after an adaptive improvement and error correction, it is used to predict the long-term changing trend of track irregularity at a fixed measuring point; the stochastic linear AR, Kalman filtering model, and artificial neural network model are applied to predict the short-term changing trend of track irregularity at unit section. Both long-term and short-term changes prove that the model is effective and can achieve the expected accuracy.

  7. Examining U.S. Irregular Warfare Doctrine

    National Research Council Canada - National Science Library

    Kimbrough, IV, James M

    2008-01-01

    ... of insurgency and terrorism. In response to the associated strategic challenges, a growing debate occurred among military historians, strategists, and leaders about the proper principles necessary for contemporary irregular...

  8. Traffic dispersion through a series of signals with irregular split

    Science.gov (United States)

    Nagatani, Takashi

    2016-01-01

    We study the traffic behavior of a group of vehicles moving through a sequence of signals with irregular splits on a roadway. We present the stochastic model of vehicular traffic controlled by signals. The dynamic behavior of vehicular traffic is clarified by analyzing traffic pattern and travel time numerically. The group of vehicles breaks up more and more by the irregularity of signal's split. The traffic dispersion is induced by the irregular split. We show that the traffic dispersion depends highly on the cycle time and the strength of split's irregularity. Also, we study the traffic behavior through the series of signals at the green-wave strategy. The dependence of the travel time on offset time is derived for various values of cycle time. The region map of the traffic dispersion is shown in (cycle time, offset time)-space.

  9. GARCH and Irregularly Spaced Data

    NARCIS (Netherlands)

    Meddahi, N.; Renault, E.; Werker, B.J.M.

    2003-01-01

    An exact discretization of continuous time stochastic volatility processes observed at irregularly spaced times is used to give insights on how a coherent GARCH model can be specified for such data. The relation of our approach with those in the existing literature is studied.

  10. Modified hybrid subcarrier/amplitude/ phase/polarization LDPC-coded modulation for 400 Gb/s optical transmission and beyond.

    Science.gov (United States)

    Batshon, Hussam G; Djordjevic, Ivan; Xu, Lei; Wang, Ting

    2010-06-21

    In this paper, we present a modified coded hybrid subcarrier/ amplitude/phase/polarization (H-SAPP) modulation scheme as a technique capable of achieving beyond 400 Gb/s single-channel transmission over optical channels. The modified H-SAPP scheme profits from the available resources in addition to geometry to increase the bandwidth efficiency of the transmission system, and so increases the aggregate rate of the system. In this report we present the modified H-SAPP scheme and focus on an example that allows 11 bits/Symbol that can achieve 440 Gb/s transmission using components of 50 Giga Symbol/s (GS/s).

  11. Study of electromagnetic wave scattering by periodic density irregularities in plasma

    International Nuclear Information System (INIS)

    Lyle, R.; Kuo, S.P.; Huang, J.

    1995-01-01

    A quasi-particle approach is used to formulate wave propagation and scattering in a periodically structured plasma. The theory is then applied to study the effect of bottomside sinusoidal (BSS) irregularities on the propagation of beacon satellites signals through the ionosphere. In this approach, the radio wave is treated as a distribution of quasi-particles described by a Wigner distribution function governed by a transport equation. The irregularities providing the collisional effect are modeled as a two dimensional density modulation on a uniform background plasma. The present work generalizes the previous work by including the spectral bandwidth (Δk/k) effect of the spatially periodic irregularities on the transionospheric signal propagation. The collision of quasi-particles with the irregularities modifies the quasi-particle distribution and give rise to the wave scattering phenomenon. The multiple scattering process is generally considered in this deterministic analysis of radio wave scattering off the ionospheric density irregularities. The analysis shows that this two dimensional density grating effectively modulates the intensity of the beacon satellite signals. This spatial modulation of the wave intensity is converted into time modulation due to the drift of the ionospheric irregularities, which then results in the scintillation of the beacon satellite signals

  12. Crime among irregular immigrants and the influence of internal border control

    OpenAIRE

    Leerkes, Arjen; Engbersen, Godfried; Leun, Joanne

    2012-01-01

    textabstractBoth the number of crime suspects without legal status and the number of irregular or undocumented immigrants held in detention facilities increased substantially in theNetherlands between 1997 and 2003. In this period, theDutch state increasingly attempted to exclude irregular immigrants from the formal labour market and public provisions. At the same time the registered crime among irregular migrants rose. The 'marginalisation thesis' asserts that a larger number of migrants hav...

  13. Introduction to digital mobile communication

    CERN Document Server

    Akaiwa, Yoshihiko

    2015-01-01

    Introduces digital mobile communications with an emphasis on digital transmission methods This book presents mathematical analyses of signals, mobile radio channels, and digital modulation methods. The new edition covers the evolution of wireless communications technologies and systems. The major new topics are OFDM (orthogonal frequency domain multiplexing), MIMO (multi-input multi-output) systems, frequency-domain equalization, the turbo codes, LDPC (low density parity check code), ACELP (algebraic code excited linear predictive) voice coding, dynamic scheduling for wireless packet data t

  14. Measurements of electron density irregularities in the ionosphere of Jupiter by Pioneer 10

    International Nuclear Information System (INIS)

    Woo, R.; Yang, F.

    1976-01-01

    In this paper we demonstrate that when the frequency spectrum of the log amplitude fluctuations is used, the radio occultation experiment is a powerful tool for detecting, identifying, and studying ionospheric irregularities. Analysis of the Pioneer 10 radio occultation measurements reveals that the Jovian ionosphere possesses electron density irregularities which are very similar to those found in the earth's ionosphere. This is the first time such irregularities have been found in a planetary ionosphere other than that of earth. The Pioneer 10 results indicate that the spatial wave number spectrum of the electron density irregularities is close to the Kolmogorov spectrum and that the outer scale size is greater than the Fresnel size (6.15 km). This type of spectrum suggests that the irregularities are probably produced by the turbulent dissipation of irregularities larger than the outer scale size

  15. Log-Likelihood Ratio Calculation for Iterative Decoding on Rayleigh Fading Channels Using Padé Approximation

    Directory of Open Access Journals (Sweden)

    Gou Hosoya

    2013-01-01

    Full Text Available Approximate calculation of channel log-likelihood ratio (LLR for wireless channels using Padé approximation is presented. LLR is used as an input of iterative decoding for powerful error-correcting codes such as low-density parity-check (LDPC codes or turbo codes. Due to the lack of knowledge of the channel state information of a wireless fading channel, such as uncorrelated fiat Rayleigh fading channels, calculations of exact LLR for these channels are quite complicated for a practical implementation. The previous work, an LLR calculation using the Taylor approximation, quickly becomes inaccurate as the channel output leaves some derivative point. This becomes a big problem when higher order modulation scheme is employed. To overcome this problem, a new LLR approximation using Padé approximation, which expresses the original function by a rational form of two polynomials with the same total number of coefficients of the Taylor series and can accelerate the Taylor approximation, is devised. By applying the proposed approximation to the iterative decoding and the LDPC codes with some modulation schemes, we show the effectiveness of the proposed methods by simulation results and analysis based on the density evolution.

  16. Influence of initial stress, irregularity and heterogeneity on Love-type ...

    Indian Academy of Sciences (India)

    The present paper deals with the propagation of Love-type wave in an initially stressed irregular vertically heterogeneous layer lying over an initially stressed isotropic layer and an initially stressed isotropic half- space. Two different types of irregularities, viz., rectangular and parabolic, are considered at the interface.

  17. Advances in electron dosimetry of irregular fields

    International Nuclear Information System (INIS)

    Mendez V, J.

    1998-01-01

    In this work it is presented an advance in Electron dosimetry of irregular fields for beams emitted by linear accelerators. At present diverse methods exist which are coming to apply in the Radiotherapy centers. In this work it is proposed a method for irregular fields dosimetry. It will be allow to calculate the dose rate absorbed required for evaluating the time for the treatment of cancer patients. Utilizing the results obtained by the dosimetric system, it has been possible to prove the validity of the method describe for 12 MeV energy and for square field 7.5 x 7.5 cm 2 with percentile error less than 1 % . (Author)

  18. New Model for Ionospheric Irregularities at Mars

    Science.gov (United States)

    Keskinen, M. J.

    2018-03-01

    A new model for ionospheric irregularities at Mars is presented. It is shown that wind-driven currents in the dynamo region of the Martian ionosphere can be unstable to the electromagnetic gradient drift instability. This plasma instability can generate ionospheric density and magnetic field irregularities with scale sizes of approximately 15-20 km down to a few kilometers. We show that the instability-driven magnetic field fluctuation amplitudes relative to background are correlated with the ionospheric density fluctuation amplitudes relative to background. Our results can explain recent observations made by the Mars Atmosphere and Volatile EvolutioN spacecraft in the Martian ionosphere dynamo region.

  19. Edge irregular total labellings for graphs of linear size

    DEFF Research Database (Denmark)

    Brandt, Stephan; Rautenbach, D.; Miškuf, J.

    2009-01-01

    As an edge variant of the well-known irregularity strength of a graph G = (V, E) we investigate edge irregular total labellings, i.e. functions f : V ∪ E → {1, 2, ..., k} such that f (u) + f (u v) + f (v) ≠ f (u) + f (u v) + f (v) for every pair of different edges u v, u v ∈ E. The smallest possi...

  20. Spatial-Aided Low-Delay Wyner-Ziv Video Coding

    Directory of Open Access Journals (Sweden)

    Bo Wu

    2009-01-01

    Full Text Available In distributed video coding, the side information (SI quality plays an important role in Wyner-Ziv (WZ frame coding. Usually, SI is generated at the decoder by the motion-compensated interpolation (MCI from the past and future key frames under the assumption that the motion trajectory between the adjacent frames is translational with constant velocity. However, this assumption is not always true and thus, the coding efficiency for WZ coding is often unsatisfactory in video with high and/or irregular motion. This situation becomes more serious in low-delay applications since only motion-compensated extrapolation (MCE can be applied to yield SI. In this paper, a spatial-aided Wyner-Ziv video coding (WZVC in low-delay application is proposed. In SA-WZVC, at the encoder, each WZ frame is coded as performed in the existing common Wyner-Ziv video coding scheme and meanwhile, the auxiliary information is also coded with the low-complexity DPCM. At the decoder, for the WZ frame decoding, auxiliary information should be decoded firstly and then SI is generated with the help of this auxiliary information by the spatial-aided motion-compensated extrapolation (SA-MCE. Theoretical analysis proved that when a good tradeoff between the auxiliary information coding and WZ frame coding is achieved, SA-WZVC is able to achieve better rate distortion performance than the conventional MCE-based WZVC without auxiliary information. Experimental results also demonstrate that SA-WZVC can efficiently improve the coding performance of WZVC in low-delay application.

  1. Study of Track Irregularity Time Series Calibration and Variation Pattern at Unit Section

    Directory of Open Access Journals (Sweden)

    Chaolong Jia

    2014-01-01

    Full Text Available Focusing on problems existing in track irregularity time series data quality, this paper first presents abnormal data identification, data offset correction algorithm, local outlier data identification, and noise cancellation algorithms. And then proposes track irregularity time series decomposition and reconstruction through the wavelet decomposition and reconstruction approach. Finally, the patterns and features of track irregularity standard deviation data sequence in unit sections are studied, and the changing trend of track irregularity time series is discovered and described.

  2. Spectral classification of medium-scale high-latitude F region plasma density irregularities

    International Nuclear Information System (INIS)

    Singh, M.; Rodriguez, P.; Szuszczewicz, E.P.; Sachs Freeman Associates, Bowie, MD)

    1985-01-01

    The high-latitude ionosphere represents a highly structured plasma. Rodriguez and Szuszczewicz (1984) reported a wide range of plasma density irregularities (150 km to 75 m) at high latitudes near 200 km. They have shown that the small-scale irregularities (7.5 km to 75 m) populated the dayside oval more often than the other phenomenological regions. It was suggested that in the lower F region the chemical recombination is fast enough to remove small-scale irregularities before convection can transport them large distances, leaving structured particle precipitation as the dominant source term for irregularities. The present paper provides the results of spectral analyses of pulsed plasma probe data collected in situ aboard the STP/S3-4 satellite during the period March-September 1978. A quantitative description of irregularity spectra in the high-latitude lower F region plasma density is given. 22 references

  3. Characterizing spontaneous irregular behavior in coupled map lattices

    International Nuclear Information System (INIS)

    Dobyns, York; Atmanspacher, Harald

    2005-01-01

    Two-dimensional coupled map lattices display, in a specific parameter range, a stable phase (quasi-) periodic in both space and time. With small changes to the model parameters, this stable phase develops spontaneous eruptions of non-periodic behavior. Although this behavior itself appears irregular, it can be characterized in a systematic fashion. In particular, parameter-independent features of the spontaneous eruptions may allow useful empirical characterizations of other phenomena that are intrinsically hard to predict and reproduce. Specific features of the distributions of lifetimes and emergence rates of irregular states display such parameter-independent properties

  4. Characterizing spontaneous irregular behavior in coupled map lattices

    Energy Technology Data Exchange (ETDEWEB)

    Dobyns, York [PEAR, Princeton University Princeton, NJ 08544-5263 (United States); Atmanspacher, Harald [Institut fuer Grenzgebiete der Psychologie und Psychohygiene Wilhelmstrasse 3a, Freiburg 79098 (Germany)]. E-mail: haa@igpp.de

    2005-04-01

    Two-dimensional coupled map lattices display, in a specific parameter range, a stable phase (quasi-) periodic in both space and time. With small changes to the model parameters, this stable phase develops spontaneous eruptions of non-periodic behavior. Although this behavior itself appears irregular, it can be characterized in a systematic fashion. In particular, parameter-independent features of the spontaneous eruptions may allow useful empirical characterizations of other phenomena that are intrinsically hard to predict and reproduce. Specific features of the distributions of lifetimes and emergence rates of irregular states display such parameter-independent properties.

  5. Dose calculations for irregular fields using three-dimensional first-scatter integration

    International Nuclear Information System (INIS)

    Boesecke, R.; Scharfenberg, H.; Schlegel, W.; Hartmann, G.H.

    1986-01-01

    This paper describes a method of dose calculations for irregular fields which requires only the mean energy of the incident photons, the geometrical properties of the irregular field and of the therapy unit, and the attenuation coefficient of tissue. The method goes back to an approach including spatial aspects of photon scattering for inhomogeneities for the calculation of dose reduction factors as proposed by Sontag and Cunningham (1978). It is based on the separation of dose into a primary component and a scattered component. The scattered component can generally be calculated for each field by integration over dose contributions from scattering in neighbouring volume elements. The quotient of this scattering contribution in the irregular field and the scattering contribution in the equivalent open field is then the correction factor for scattering in an irregular field. A correction factor for the primary component can be calculated if the attenuation of the photons in the shielding block is properly taken into account. The correction factor is simply given by the quotient of primary photons of the irregular field and the primary photons of the open field. (author)

  6. Using forbidden ordinal patterns to detect determinism in irregularly sampled time series.

    Science.gov (United States)

    Kulp, C W; Chobot, J M; Niskala, B J; Needhammer, C J

    2016-02-01

    It is known that when symbolizing a time series into ordinal patterns using the Bandt-Pompe (BP) methodology, there will be ordinal patterns called forbidden patterns that do not occur in a deterministic series. The existence of forbidden patterns can be used to identify deterministic dynamics. In this paper, the ability to use forbidden patterns to detect determinism in irregularly sampled time series is tested on data generated from a continuous model system. The study is done in three parts. First, the effects of sampling time on the number of forbidden patterns are studied on regularly sampled time series. The next two parts focus on two types of irregular-sampling, missing data and timing jitter. It is shown that forbidden patterns can be used to detect determinism in irregularly sampled time series for low degrees of sampling irregularity (as defined in the paper). In addition, comments are made about the appropriateness of using the BP methodology to symbolize irregularly sampled time series.

  7. On the equivalence of Ising models on ‘small-world’ networks and LDPC codes on channels with memory

    International Nuclear Information System (INIS)

    Neri, Izaak; Skantzos, Nikos S

    2014-01-01

    We demonstrate the equivalence between thermodynamic observables of Ising spin-glass models on small-world lattices and the decoding properties of error-correcting low-density parity-check codes on channels with memory. In particular, the self-consistent equations for the effective field distributions in the spin-glass model within the replica symmetric ansatz are equivalent to the density evolution equations forr Gilbert–Elliott channels. This relationship allows us to present a belief-propagation decoding algorithm for finite-state Markov channels and to compute its performance at infinite block lengths from the density evolution equations. We show that loss of reliable communication corresponds to a first order phase transition from a ferromagnetic phase to a paramagnetic phase in the spin glass model. The critical noise levels derived for Gilbert–Elliott channels are in very good agreement with existing results in coding theory. Furthermore, we use our analysis to derive critical noise levels for channels with both memory and asymmetry in the noise. The resulting phase diagram shows that the combination of asymmetry and memory in the channel allows for high critical noise levels: in particular, we show that successful decoding is possible at any noise level of the bad channel when the good channel is good enough. Theoretical results at infinite block lengths using density evolution equations aree compared with average error probabilities calculated from a practical implementation of the corresponding decoding algorithms at finite block lengths. (paper)

  8. Scintillations associated with bottomside sinusoidal irregularities in the equatorial F region

    Science.gov (United States)

    Basu, S.; Basu, S.; Valladares, C. E.; Dasgupta, A.; Whitney, H. E.

    1986-01-01

    Multisatellite scintillation observations and spaced receiver drift measurements are presented for a category of equatorial F region plasma irregularities characterized by nearly sinusoidal waveforms in the ion number density. The observations were made at Huancayo, Peru, and the measurements at Ancon, Peru, associated with irregularities observed by the Atmospheric-Explorer-E satellite on a few nights in December 1979. Utilizing ray paths to various geostationary satellites, it was found that the irregularities grow and decay almost simultaneously in long-lived patches extending at least 1000 km in the east-west direction.

  9. Propagation and scattering of electromagnetic waves by the ionospheric irregularities

    International Nuclear Information System (INIS)

    Ho, A.Y.; Kuo, S.P.; Lee, M.C.

    1993-01-01

    The problem of wave propagation and scattering in the ionosphere is particularly important in the areas of communications, remote-sensing and detection. The ionosphere is often perturbed with coherently structured (quasiperiodic) density irregularities. Experimental observations suggest that these irregularities could give rise to significant ionospheric effect on wave propagation such as causing spread-F of the probing HF sounding signals and scintillation of beacon satellite signals. It was show by the latter that scintillation index S 4 ∼ 0.5 and may be as high as 0.8. In this work a quasi-particle theory is developed to study the scintillation phenomenon. A Wigner distribution function for the wave intensity in the (k,r) space is introduced and its governing equation is derived with an effective collision term giving rise to the attenuation and scattering of the wave. This kinetic equation leads to a hierarchy of moment equations in r space. This systems of equations is then truncated to the second moment which is equivalent to assuming a cold quasi-particle distribution In this analysis, the irregularities are modeled as a two dimensional density modulation on an uniform background plasma. The analysis shows that this two dimensional density grating, effectively modulates the intensity of the beacon satellite signals. This spatial modulation of the wave intensity is converted into time modulation due to the drift of the ionospheric irregularities, which then contributes to the scintillation of the beacon satellite signals. Using the proper plasma parameters and equatorial measured data of irregularities, it is shown that the scintillation index defined by S4=( 2 >- 2 )/ 2 where stands for spatial average over an irregularity wavelength is in the range of the experimentally detected values

  10. The structure of plasma-density irregularities in the interplanetary medium

    International Nuclear Information System (INIS)

    Singleton, D.G.

    1975-01-01

    The conflict in the literature as to whether the plasma-density spatial spectrum of the irregularities in the interplanetary medium is of Gaussian or power law form is discussed. Particular attention is paid to the interplanetary scintillation effects ascribed to these irregularities. It is shown that the phase-screen theory of scintillations can be invoked to devise a set of critical tests which provide a means of discriminating between the conflicting hypotheses. Differences in the predicted behaviour of the single sensor temporal spectra of the scintillations for the two irregularity forms provide the main tests of the conflicting hypotheses. However, it is also shown that the two hypotheses lead to different forms of the variation of scintillation index with the observing frequency and the solar elongation of the scintillating source. Consideration is given to the optimum conditions for observing the Fourier and Bessel temporal spectra modulation which is due to the Fresnel filtering of the spatial spectrum. Determination of irregularity shape, orientation and motion in terms of this modulation is also discussed. (author)

  11. NEOWISE: OBSERVATIONS OF THE IRREGULAR SATELLITES OF JUPITER AND SATURN

    Energy Technology Data Exchange (ETDEWEB)

    Grav, T. [Planetary Science Institute, Tucson, AZ 85719 (United States); Bauer, J. M.; Mainzer, A. K.; Masiero, J. R.; Sonnett, S.; Kramer, E. [Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA 91109 (United States); Nugent, C. R.; Cutri, R. M., E-mail: tgrav@psi.edu [Infrared Processing and Analysis Center, California Institute of Technology, Pasadena, CA 91125 (United States)

    2015-08-10

    We present thermal model fits for 11 Jovian and 3 Saturnian irregular satellites based on measurements from the WISE/NEOWISE data set. Our fits confirm spacecraft-measured diameters for the objects with in situ observations (Himalia and Phoebe) and provide diameters and albedo for 12 previously unmeasured objects, 10 Jovian and 2 Saturnian irregular satellites. The best-fit thermal model beaming parameters are comparable to what is observed for other small bodies in the outer solar system, while the visible, W1, and W2 albedos trace the taxonomic classifications previously established in the literature. Reflectance properties for the irregular satellites measured are similar to the Jovian Trojan and Hilda Populations, implying common origins.

  12. Experimental demonstration of multi-pilot aided carrier phase estimation for DP-64QAM and DP-256QAM

    NARCIS (Netherlands)

    Pajovic, M.; Millar, D.S.; Koike-Akino, T.; Maher, R.; Lavery, D.; Alvarado, A.; Paskov, M.; Kojima, K.; Parsons, K.; Thomsen, B.C.; Savory, S.J.; Bayvel, P.

    2015-01-01

    We present a statistical inference based multi-pilot aided CPE algorithm and analyze its performance via simulations. We experimentally verify LDPC coded back-to-back performance using 10 GBd DP-64QAM and DP-256QAM modulation, with transmitter and receiver linewidths of 100 kHz.

  13. Crime among irregular immigrants and the influence of internal border control

    NARCIS (Netherlands)

    Leerkes, A.S.; Engbersen, G.; Leun, van der J.P.

    2012-01-01

    Abstract Both the number of crime suspects without legal status and the number of irregular or undocumented immigrants held in detention facilities increased substantially in theNetherlands between 1997 and 2003. In this period, theDutch state increasingly attempted to exclude irregular immigrants

  14. Crime among irregular immigrants and the influence of internal border control

    NARCIS (Netherlands)

    A.S. Leerkes (Arjen); G.B.M. Engbersen (Godfried); J.P. van der Leun (Joanne)

    2012-01-01

    textabstractBoth the number of crime suspects without legal status and the number of irregular or undocumented immigrants held in detention facilities increased substantially in theNetherlands between 1997 and 2003. In this period, theDutch state increasingly attempted to exclude irregular

  15. Geostatistical regularization operators for geophysical inverse problems on irregular meshes

    Science.gov (United States)

    Jordi, C.; Doetsch, J.; Günther, T.; Schmelzbach, C.; Robertsson, J. OA

    2018-05-01

    Irregular meshes allow to include complicated subsurface structures into geophysical modelling and inverse problems. The non-uniqueness of these inverse problems requires appropriate regularization that can incorporate a priori information. However, defining regularization operators for irregular discretizations is not trivial. Different schemes for calculating smoothness operators on irregular meshes have been proposed. In contrast to classical regularization constraints that are only defined using the nearest neighbours of a cell, geostatistical operators include a larger neighbourhood around a particular cell. A correlation model defines the extent of the neighbourhood and allows to incorporate information about geological structures. We propose an approach to calculate geostatistical operators for inverse problems on irregular meshes by eigendecomposition of a covariance matrix that contains the a priori geological information. Using our approach, the calculation of the operator matrix becomes tractable for 3-D inverse problems on irregular meshes. We tested the performance of the geostatistical regularization operators and compared them against the results of anisotropic smoothing in inversions of 2-D surface synthetic electrical resistivity tomography (ERT) data as well as in the inversion of a realistic 3-D cross-well synthetic ERT scenario. The inversions of 2-D ERT and seismic traveltime field data with geostatistical regularization provide results that are in good accordance with the expected geology and thus facilitate their interpretation. In particular, for layered structures the geostatistical regularization provides geologically more plausible results compared to the anisotropic smoothness constraints.

  16. Effects of Irregular Bridge Columns and Feasibility of Seismic Regularity

    Science.gov (United States)

    Thomas, Abey E.

    2018-05-01

    Bridges with unequal column height is one of the main irregularities in bridge design particularly while negotiating steep valleys, making the bridges vulnerable to seismic action. The desirable behaviour of bridge columns towards seismic loading is that, they should perform in a regular fashion, i.e. the capacity of each column should be utilized evenly. But, this type of behaviour is often missing when the column heights are unequal along the length of the bridge, allowing short columns to bear the maximum lateral load. In the present study, the effects of unequal column height on the global seismic performance of bridges are studied using pushover analysis. Codes such as CalTrans (Engineering service center, earthquake engineering branch, 2013) and EC-8 (EN 1998-2: design of structures for earthquake resistance. Part 2: bridges, European Committee for Standardization, Brussels, 2005) suggests seismic regularity criterion for achieving regular seismic performance level at all the bridge columns. The feasibility of adopting these seismic regularity criterions along with those mentioned in literatures will be assessed for bridges designed as per the Indian Standards in the present study.

  17. Generating Performance Models for Irregular Applications

    Energy Technology Data Exchange (ETDEWEB)

    Friese, Ryan D.; Tallent, Nathan R.; Vishnu, Abhinav; Kerbyson, Darren J.; Hoisie, Adolfy

    2017-05-30

    Many applications have irregular behavior --- non-uniform input data, input-dependent solvers, irregular memory accesses, unbiased branches --- that cannot be captured using today's automated performance modeling techniques. We describe new hierarchical critical path analyses for the \\Palm model generation tool. To create a model's structure, we capture tasks along representative MPI critical paths. We create a histogram of critical tasks with parameterized task arguments and instance counts. To model each task, we identify hot instruction-level sub-paths and model each sub-path based on data flow, instruction scheduling, and data locality. We describe application models that generate accurate predictions for strong scaling when varying CPU speed, cache speed, memory speed, and architecture. We present results for the Sweep3D neutron transport benchmark; Page Rank on multiple graphs; Support Vector Machine with pruning; and PFLOTRAN's reactive flow/transport solver with domain-induced load imbalance.

  18. A Novel Strategy Using Factor Graphs and the Sum-Product Algorithm for Satellite Broadcast Scheduling Problems

    Science.gov (United States)

    Chen, Jung-Chieh

    This paper presents a low complexity algorithmic framework for finding a broadcasting schedule in a low-altitude satellite system, i. e., the satellite broadcast scheduling (SBS) problem, based on the recent modeling and computational methodology of factor graphs. Inspired by the huge success of the low density parity check (LDPC) codes in the field of error control coding, in this paper, we transform the SBS problem into an LDPC-like problem through a factor graph instead of using the conventional neural network approaches to solve the SBS problem. Based on a factor graph framework, the soft-information, describing the probability that each satellite will broadcast information to a terminal at a specific time slot, is exchanged among the local processing in the proposed framework via the sum-product algorithm to iteratively optimize the satellite broadcasting schedule. Numerical results show that the proposed approach not only can obtain optimal solution but also enjoys the low complexity suitable for integral-circuit implementation.

  19. Effects of magnetic storm phases on F layer irregularities below the auroral oval

    International Nuclear Information System (INIS)

    Aarons, J.; Gurgiolo, C.; Rodger, A.S.

    1988-01-01

    Observations of F-layer irregularity development and intensity were obtained between September and October 1981, primarily over subauroral latitudes in the area of the plasmapause. The results reveal the descent of the auroral irregularity region to include subauroral latitudes in the general area of the plasmapause during the main phases of a series of magnetic storms. Irregularities were found primarily at lower latitudes during the subauroral or plasmapause storm. A model for the subauroral irregularities in recovery phases of magnetic storms is proposed in which energy stored in the ring current is slowly released. 27 references

  20. Dependence on zenith angle of the strength of 3-meter equatorial electrojet irregularities

    International Nuclear Information System (INIS)

    Ierkic, H.M.; Fejer, B.G.; Farley, D.T.

    1980-01-01

    Radar measurements in Peru were used to deduce the zenith angle dependence of the scattering cross section of plasma irregularities generated by instabilities in the equatorial electrojet. The irregularities probed by the 50 MHz Jicamarca radar had a wavelength of 3m. The cross section for the type 2 irregularities was isotopic in the plane perpendicular to the magnetic field, while the cross section for the stronger type 1 irregularities varied with zenith angle at a rate of approximately 0.3 dB/degree; the horizontally traveling waves were more than 100 times stronger than those traveling vertically

  1. Spatial irregularities in Jupiter's upper ionosphere observed by voyager radio occultations

    Energy Technology Data Exchange (ETDEWEB)

    Hinson, D.P.; Tyler, G.L.

    1982-07-01

    Dual frequency radio occultation experiments carried out with Voyagers 1 and 2 provided data on the spatial irregularities in Jupiter's ionosphere at four different locations. Sample spectra of weak fluctuations in amplitude and phase of the 3.6-cm and 13-cm wavelength radio signals can be interpreted by using the theory for scattering from an anisotropic power law phase screen. Least squares solutions for ionospheric parameters derived from the observed fluctuation spectra yielded estimates of (1) the axial ratio, (2) angular orientation of the anisotropic irregularities, (3) the power law exponent of the spatial spectrum of irregularities, and (4) the magnitude of the spatial variations in electron density. Equipment limitations and the method of analysis constrain the observations to irregularities of approximate size 1--200 km. No evidence of the inner or outer scale of the irregularities was found. For length scales in the range given, the three-dimensional spatial spectrum obeys a power law with exponent varying from -3.0 to -3.7, and the root mean square fractional variations in electron density are 1--15%. All observed irregularities appear to be anisotropic with axial ratios between 2:1 and 10:1. Ionospheric parameters vary with altitude and latitude. We conclude that the measured angular orientation of the anisotropic irregularities indicates magnetic field direction and may provide a basis for refining Jovian magnetic field models.

  2. Measurement of Dynamic Friction Coefficient on the Irregular Free Surface

    International Nuclear Information System (INIS)

    Yeom, S. H.; Seo, K. S.; Lee, J. H.; Lee, K. H.

    2007-01-01

    A spent fuel storage cask must be estimated for a structural integrity when an earthquake occurs because it freely stands on ground surface without a restriction condition. Usually the integrity estimation for a seismic load is performed by a FEM analysis, the friction coefficient for a standing surface is an important parameter in seismic analysis when a sliding happens. When a storage cask is placed on an irregular ground surface, measuring a friction coefficient of an irregular surface is very difficult because the friction coefficient is affected by the surface condition. In this research, dynamic friction coefficients on the irregular surfaces between a concrete cylinder block and a flat concrete slab are measured with two methods by one direction actuator

  3. 30 MHz radar observations of artificial E region field-aligned plasma irregularities

    Directory of Open Access Journals (Sweden)

    D. L. Hysell

    2008-02-01

    Full Text Available Artificial E region field aligned irregularities (FAIs have been observed during heating experiments at the HAARP facility using a new 30 MHz coherent scatter radar imager deployed near Homer, Alaska. Irregularities were observed during brief experiments on three quiet days in July and August, 2007, when the daytime E region critical frequency was close to 3 MHz. Irregularities were consistently generated and detected during experiments with O-mode HF pumping on zenith with a 1-min on, 1-min off CW modulation. The scattering cross sections, rise, and fall times of the echoes were observed as well as their spectral properties. Results were found to be mainly in agreement with observations from other mid- and high-latitude sites with some discrepancies. Radar images of the irregularity-filled volume on one case exhibited clear variations in backscatter power and Doppler shift across the volume. The images furthermore show the emergence of a small irregularity-filled region to the south southwest of the main region in the approximate direction of magnetic zenith.

  4. Measurement Techniques for Clock Jitter

    Science.gov (United States)

    Lansdowne, Chatwin; Schlesinger, Adam

    2012-01-01

    NASA is in the process of modernizing its communications infrastructure to accompany the development of a Crew Exploration Vehicle (CEV) to replace the shuttle. With this effort comes the opportunity to infuse more advanced coded modulation techniques, including low-density parity-check (LDPC) codes that offer greater coding gains than the current capability. However, in order to take full advantage of these codes, the ground segment receiver synchronization loops must be able to operate at a lower signal-to-noise ratio (SNR) than supported by equipment currently in use.

  5. Irregular menses: an independent risk factor for gestational diabetes mellitus.

    Science.gov (United States)

    Haver, Mary Claire; Locksmith, Gregory J; Emmet, Emily

    2003-05-01

    Our purpose was to determine whether a history of irregular menses predicts gestational diabetes mellitus independently of traditional risk factors. We analyzed demographic characteristics, body mass index, and menstrual history of 85 pregnant women with gestational diabetes mellitus and compared them with 85 systematically selected control subjects who were matched for age, race, and delivery year. Subjects with pregestational diabetes mellitus, previous gestational diabetes mellitus, family history of diabetes mellitus, weight >200 pounds, previous macrosomic infants, or previous stillbirth were excluded. Demographic characteristics between case and control groups were similar. Mean body mass index was higher among cases (26.5 kg/m(2)) versus control subjects (24.5 kg/m(2), P =.004). Irregular cycles were more prevalent in the cases (24% vs 7%, P =.006). With the use of body mass index as a stratification factor, menstrual irregularity maintained a strong association with gestational diabetes mellitus (P =.014). A history of irregular menstrual cycles was a significant independent predictor of gestational diabetes mellitus. If selective screening is implemented for gestational diabetes mellitus, such history should be considered in the decision of whom to test.

  6. Backscatter measurements of 11-cm equatorial spread-F irregularities

    International Nuclear Information System (INIS)

    Tsunoda, R.T.

    1980-01-01

    In the equatorial F-region ionosphere, a turbulent cascade process has been found to exist that extends from irregularity spatial wavelengths longer than tens of kilometers down to wavelengths as short as 36 cm. To investigate the small-scale regime of wavelengths less than 36 cm, an equatorial radar experiment was conducted using a frequency of 1320 MHz that corresponds to an irregularity wavelength of 11 cm. The first observations of radar backscatter from 11-cm field-aligned irregularities (FAI) are described. These measurements extend the spatial wavelength regime of F-region FAI to lengths that approach both electron gyroradius and the Debye length. Agreement of these results with the theory of high-frequency drift waves suggests that these observations may be unique to the equatorial ionosphere. That is, the requirement of low electron densities for which the theroy calls may preclude the existence of 11-cm FAI elsewhere in the F-region ionosphere, except in equatorial plasma bubbles

  7. Stromal haze, myofibroblasts, and surface irregularity after PRK.

    Science.gov (United States)

    Netto, Marcelo V; Mohan, Rajiv R; Sinha, Sunilima; Sharma, Ajay; Dupps, William; Wilson, Steven E

    2006-05-01

    The aim of this study was to investigate the relationship between the level of stromal surface irregularity after photorefractive keratectomy (PRK) and myofibroblast generation along with the development of corneal haze. Variable levels of stromal surface irregularity were generated in rabbit corneas by positioning a fine mesh screen in the path of excimer laser during ablation for a variable percentage of the terminal pulses of the treatment for myopia that does not otherwise generate significant opacity. Ninety-six rabbits were divided into eight groups: [see table in text]. Slit lamp analysis and haze grading were performed in all groups. Rabbits were sacrificed at 4 hr or 4 weeks after surgery and histochemical analysis was performed on corneas for apoptosis (TUNEL assay), myofibroblast marker alpha-smooth muscle actin (SMA), and integrin alpha4 to delineate the epithelial basement membrane. Slit-lamp grading revealed severe haze formation in corneas in groups IV and VI, with significantly less haze in groups II, III, and VII and insignificant haze compared with the unwounded control in groups I and V. Analysis of SMA staining at 4 weeks after surgery, the approximate peak of haze formation in rabbits, revealed low myofibroblast formation in group I (1.2+/-0.2 cells/400x field) and group V (1.8+/-0.4), with significantly more in groups II (3.5+/-1.8), III (6.8+/-1.6), VII (7.9+/-3.8), IV (12.4+/-4.2) and VI (14.6+/-5.1). The screened groups were significantly different from each other (p PRK groups. The -9.0 diopter PRK group VI had significantly more myofibroblast generation than the -9.0 diopter PRK with PTK-smoothing group VII (p PRK and the level of stromal surface irregularity. PTK-smoothing with methylcellulose was an effective method to reduce stromal surface irregularity and decreased both haze and associated myofibroblast density. We hypothesize that stromal surface irregularity after PRK for high myopia results in defective basement membrane

  8. Criticality predicts maximum irregularity in recurrent networks of excitatory nodes.

    Directory of Open Access Journals (Sweden)

    Yahya Karimipanah

    Full Text Available A rigorous understanding of brain dynamics and function requires a conceptual bridge between multiple levels of organization, including neural spiking and network-level population activity. Mounting evidence suggests that neural networks of cerebral cortex operate at a critical regime, which is defined as a transition point between two phases of short lasting and chaotic activity. However, despite the fact that criticality brings about certain functional advantages for information processing, its supporting evidence is still far from conclusive, as it has been mostly based on power law scaling of size and durations of cascades of activity. Moreover, to what degree such hypothesis could explain some fundamental features of neural activity is still largely unknown. One of the most prevalent features of cortical activity in vivo is known to be spike irregularity of spike trains, which is measured in terms of the coefficient of variation (CV larger than one. Here, using a minimal computational model of excitatory nodes, we show that irregular spiking (CV > 1 naturally emerges in a recurrent network operating at criticality. More importantly, we show that even at the presence of other sources of spike irregularity, being at criticality maximizes the mean coefficient of variation of neurons, thereby maximizing their spike irregularity. Furthermore, we also show that such a maximized irregularity results in maximum correlation between neuronal firing rates and their corresponding spike irregularity (measured in terms of CV. On the one hand, using a model in the universality class of directed percolation, we propose new hallmarks of criticality at single-unit level, which could be applicable to any network of excitable nodes. On the other hand, given the controversy of the neural criticality hypothesis, we discuss the limitation of this approach to neural systems and to what degree they support the criticality hypothesis in real neural networks. Finally

  9. Target Tracking of a Linear Time Invariant System under Irregular Sampling

    Directory of Open Access Journals (Sweden)

    Jin Xue-Bo

    2012-11-01

    Full Text Available Due to event-triggered sampling in a system, or maybe with the aim of reducing data storage, tracking many applications will encounter irregular sampling time. By calculating the matrix exponential using an inverse Laplace transform, this paper transforms the irregular sampling tracking problem to the problem of tracking with time-varying parameters of a system. Using the common Kalman filter, the developed method is used to track a target for the simulated trajectory and video tracking. The results of simulation experiments have shown that it can obtain good estimation performance even at a very high irregular rate of measurement sampling time.

  10. Application of the pothole DAF method to vehicles traversing periodic roadway irregularities

    Science.gov (United States)

    Pesterev, A. V.; Bergman, L. A.; Tan, C. A.; Yang, B.

    2005-01-01

    This paper is a sequel to the work discussed in Pesterev et al. (Journal of Sound and Vibration, in press). In that paper, it was suggested that the technique to determine the effect of a local road surface irregularity on the dynamics of a vehicle modelled as a linear multi-degree-of-freedom system relies on the so-called pothole dynamic amplification factor (DAF), which is a complex-valued function specific to the irregularity shape. This paper discusses the companion problem of how to determine the DAF function for an irregularity represented as a superposition of simpler ones. Another purpose of this paper is to demonstrate the application of the pothole DAF functions technique to finding a priori estimates of the effect of irregularities with a repeated structure. Specifically, we solve the problem of finding the conditions under which the dynamic effect of two identical potholes located one after another is greater than that due to the single pothole. We also find the estimate for the number of periods of a periodic irregularity that are sufficient in order to consider the oscillator response as steady state. The discussions are illustrated by numerical examples.

  11. Irregular singularities in Liouville theory and Argyres-Douglas type gauge theories, I

    Energy Technology Data Exchange (ETDEWEB)

    Gaiotto, D. [Institute for Advanced Study (IAS), Princeton, NJ (United States); Teschner, J. [Deutsches Elektronen-Synchrotron (DESY), Hamburg (Germany)

    2012-03-15

    Motivated by problems arising in the study of N=2 supersymmetric gauge theories we introduce and study irregular singularities in two-dimensional conformal field theory, here Liouville theory. Irregular singularities are associated to representations of the Virasoro algebra in which a subset of the annihilation part of the algebra act diagonally. In this paper we define natural bases for the space of conformal blocks in the presence of irregular singularities, describe how to calculate their series expansions, and how such conformal blocks can be constructed by some delicate limiting procedure from ordinary conformal blocks. This leads us to a proposal for the structure functions appearing in the decomposition of physical correlation functions with irregular singularities into conformal blocks. Taken together, we get a precise prediction for the partition functions of some Argyres-Douglas type theories on S{sup 4}. (orig.)

  12. Irregular singularities in Liouville theory and Argyres-Douglas type gauge theories, I

    International Nuclear Information System (INIS)

    Gaiotto, D.; Teschner, J.

    2012-03-01

    Motivated by problems arising in the study of N=2 supersymmetric gauge theories we introduce and study irregular singularities in two-dimensional conformal field theory, here Liouville theory. Irregular singularities are associated to representations of the Virasoro algebra in which a subset of the annihilation part of the algebra act diagonally. In this paper we define natural bases for the space of conformal blocks in the presence of irregular singularities, describe how to calculate their series expansions, and how such conformal blocks can be constructed by some delicate limiting procedure from ordinary conformal blocks. This leads us to a proposal for the structure functions appearing in the decomposition of physical correlation functions with irregular singularities into conformal blocks. Taken together, we get a precise prediction for the partition functions of some Argyres-Douglas type theories on S 4 . (orig.)

  13. Research on formation of microsatellite communication with genetic algorithm.

    Science.gov (United States)

    Wu, Guoqiang; Bai, Yuguang; Sun, Zhaowei

    2013-01-01

    For the formation of three microsatellites which fly in the same orbit and perform three-dimensional solid mapping for terra, this paper proposes an optimizing design method of space circular formation order based on improved generic algorithm and provides an intersatellite direct spread spectrum communication system. The calculating equation of LEO formation flying satellite intersatellite links is guided by the special requirements of formation-flying microsatellite intersatellite links, and the transmitter power is also confirmed throughout the simulation. The method of space circular formation order optimizing design based on improved generic algorithm is given, and it can keep formation order steady for a long time under various absorb impetus. The intersatellite direct spread spectrum communication system is also provided. It can be found that, when the distance is 1 km and the data rate is 1 Mbps, the input wave matches preferably with the output wave. And LDPC code can improve the communication performance. The correct capability of (512, 256) LDPC code is better than (2, 1, 7) convolution code, distinctively. The design system can satisfy the communication requirements of microsatellites. So, the presented method provides a significant theory foundation for formation-flying and intersatellite communication.

  14. Effects of surface irregularities on intensity data from laser scanning: an experimental approach.

    Directory of Open Access Journals (Sweden)

    G. Teza

    2008-06-01

    Full Text Available The results of an experiment carried out with the aim to investigate the role of surface irregularities on the intensity data provided by a terrestrial laser scanner (TLS survey are reported here. Depending on surface roughness, the interaction between an electromagnetic wave and microscopic irregularities leads to a Lambertian-like diffusive light reflection, allowing the TLS to receive the backscattered component of the signal. The described experiment consists in a series of TLS-based acquisitions of a rotating artificial target specifically conceived in order to highlight the effects on the intensity data due to surface irregularity. This target is articulated in a flat plate and in an irregular surface, whose macro-roughness has a characteristic length with the same order of the spot size. Results point out the different behavior of the plates. The intensity of the signal backscattered by the planar element decreases if the incidence angle increases, whereas the intensity of the signal backscattered by the irregular surface is almost constant if the incidence angle varies. Since the typical surfaces acquired in a geological/geophysical survey are generally irregular, these results imply that the intensity data can be easily used in order to evaluate the reflectance of the material at the considered wavelength, e.g. for pattern recognition purposes.

  15. Radar Observations of 8.3-m scale equatorial spread F irregularities over Trivandrum

    Directory of Open Access Journals (Sweden)

    D. Tiwari

    2004-03-01

    Full Text Available In this paper, we present observations of equatorial spread F (ESF irregularities made using a newly installed 18MHz radar located at Trivandrum. We characterize the morphology and the spectral parameters of the 8.3-m ESF irregularities which are found to be remarkably different from that observed so extensively at the 3-m scale size. We also present statistical results of the irregularities in the form of percentage occurrence of the echoes and spectral parameters (SNR, Doppler velocity, Spectral width. The Doppler spectra are narrower, less structured and less variable in time as compared to those observed for 3-m scale size. We have never observed the ESF irregularity velocities to be supersonic here unlike those at Jicamarca, and the velocities are found to be within ±200ms–1. The spectral widths are found to be less than 150ms–1. Hence, the velocities and spectral width both are smaller than those reported for 3-m scale size. The velocities and spectral widths are further found to be much smaller than those of the American sector. These observations are compared with those reported elsewhere and discussed in the light of present understanding on the ESF irregularities at different wavelengths. Key words. Ionoshphere (equatorial ionosphere, plasma waves and instabilities; ionospheric irregularities

  16. Modification of transmission dose algorithm for irregularly shaped radiation field and tissue deficit

    Energy Technology Data Exchange (ETDEWEB)

    Yun, Hyong Geon; Shin, Kyo Chul [Dankook Univ., College of Medicine, Seoul (Korea, Republic of); Huh, Soon Nyung; Woo, Hong Gyun; Ha, Sung Whan [Seoul National Univ., College of Medicine, Seoul (Korea, Republic of); Lee, Hyoung Koo [The Catholic Univ., College of Medicine, Seoul (Korea, Republic of)

    2002-07-01

    Algorithm for estimation of transmission dose was modified for use in partially blocked radiation fields and in cases with tissue deficit. The beam data was measured with flat solid phantom in various conditions of beam block. And an algorithm for correction of transmission dose in cases of partially blocked radiation field was developed from the measured data. The algorithm was tested in some clinical settings with irregular shaped field. Also, another algorithm for correction of transmission dose for tissue deficit was developed by physical reasoning. This algorithm was tested in experimental settings with irregular contours mimicking breast cancer patients by using multiple sheets of solid phantoms. The algorithm for correction of beam block could accurately reflect the effect of beam block, with error within {+-}1.0%, both with square fields and irregularly shaped fields. The correction algorithm for tissue deficit could accurately reflect the effect of tissue deficit with errors within {+-}1.0% in most situations and within {+-}3.0% in experimental settings with irregular contours mimicking breast cancer treatment set-up. Developed algorithms could accurately estimate the transmission dose in most radiation treatment settings including irregularly shaped field and irregularly shaped body contour with tissue deficit in transmission dosimetry.

  17. Modification of transmission dose algorithm for irregularly shaped radiation field and tissue deficit

    International Nuclear Information System (INIS)

    Yun, Hyong Geon; Shin, Kyo Chul; Huh, Soon Nyung; Woo, Hong Gyun; Ha, Sung Whan; Lee, Hyoung Koo

    2002-01-01

    Algorithm for estimation of transmission dose was modified for use in partially blocked radiation fields and in cases with tissue deficit. The beam data was measured with flat solid phantom in various conditions of beam block. And an algorithm for correction of transmission dose in cases of partially blocked radiation field was developed from the measured data. The algorithm was tested in some clinical settings with irregular shaped field. Also, another algorithm for correction of transmission dose for tissue deficit was developed by physical reasoning. This algorithm was tested in experimental settings with irregular contours mimicking breast cancer patients by using multiple sheets of solid phantoms. The algorithm for correction of beam block could accurately reflect the effect of beam block, with error within ±1.0%, both with square fields and irregularly shaped fields. The correction algorithm for tissue deficit could accurately reflect the effect of tissue deficit with errors within ±1.0% in most situations and within ±3.0% in experimental settings with irregular contours mimicking breast cancer treatment set-up. Developed algorithms could accurately estimate the transmission dose in most radiation treatment settings including irregularly shaped field and irregularly shaped body contour with tissue deficit in transmission dosimetry

  18. Afrique Science: Revue Internationale des Sciences et Technologie ...

    African Journals Online (AJOL)

    OFDM codée pour le haut débit en fibre optique avec les codes correcteurs convolutifs, BCH, RS et LDPC · EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT DOWNLOAD FULL TEXT DOWNLOAD FULL TEXT. S R Sanou, F Zougmoré, Z Koalaga, M Kebré, 1-16 ...

  19. 1-Pb/s (32 SDM/46 WDM/768 Gb/s) C-band Dense SDM Transmission over 205.6-km of Single-mode Heterogeneous Multi-core Fiber using 96-Gbaud PDM-16QAM Channels

    DEFF Research Database (Denmark)

    Kobayashi, Takayuki; Nakamura, M.; Hamaoka, F.

    2017-01-01

    We demonstrate the first 1-Pb/s unidirectional inline-amplified transmission over 205.6-km of single-mode 32-core fiber within C-band only. 96-Gbaud LDPC-coded PDM-16QAM channels with FEC redundancy of 12.75% realize high-aggregate spectral efficiency of 217.6 b/s/Hz...

  20. PERIODIC ORBIT FAMILIES IN THE GRAVITATIONAL FIELD OF IRREGULAR-SHAPED BODIES

    Energy Technology Data Exchange (ETDEWEB)

    Jiang, Yu [State Key Laboratory of Astronautic Dynamics, Xi’an Satellite Control Center, Xi’an 710043 (China); Baoyin, Hexi, E-mail: jiangyu_xian_china@163.com [School of Aerospace Engineering, Tsinghua University, Beijing 100084 (China)

    2016-11-01

    The discovery of binary and triple asteroids in addition to the execution of space missions to minor celestial bodies in the past several years have focused increasing attention on periodic orbits around irregular-shaped celestial bodies. In the present work, we adopt a polyhedron shape model for providing an accurate representation of irregular-shaped bodies and employ the model to calculate their corresponding gravitational and effective potentials. We also investigate the characteristics of periodic orbit families and the continuation of periodic orbits. We prove a fact, which provides a conserved quantity that permits restricting the number of periodic orbits in a fixed energy curved surface about an irregular-shaped body. The collisions of Floquet multipliers are maintained during the continuation of periodic orbits around the comet 1P/Halley. Multiple bifurcations in the periodic orbit families about irregular-shaped bodies are also discussed. Three bifurcations in the periodic orbit family have been found around the asteroid 216 Kleopatra, which include two real saddle bifurcations and one period-doubling bifurcation.

  1. PERIODIC ORBIT FAMILIES IN THE GRAVITATIONAL FIELD OF IRREGULAR-SHAPED BODIES

    International Nuclear Information System (INIS)

    Jiang, Yu; Baoyin, Hexi

    2016-01-01

    The discovery of binary and triple asteroids in addition to the execution of space missions to minor celestial bodies in the past several years have focused increasing attention on periodic orbits around irregular-shaped celestial bodies. In the present work, we adopt a polyhedron shape model for providing an accurate representation of irregular-shaped bodies and employ the model to calculate their corresponding gravitational and effective potentials. We also investigate the characteristics of periodic orbit families and the continuation of periodic orbits. We prove a fact, which provides a conserved quantity that permits restricting the number of periodic orbits in a fixed energy curved surface about an irregular-shaped body. The collisions of Floquet multipliers are maintained during the continuation of periodic orbits around the comet 1P/Halley. Multiple bifurcations in the periodic orbit families about irregular-shaped bodies are also discussed. Three bifurcations in the periodic orbit family have been found around the asteroid 216 Kleopatra, which include two real saddle bifurcations and one period-doubling bifurcation.

  2. Short-term magnetic field alignment variations of equatorial ionospheric irregularities

    International Nuclear Information System (INIS)

    Johnson, A.L.

    1988-01-01

    The ionospheric irregularities that cause equatorial scintillation are elongated along the north-south magnetic field lines. During a 1981 field campaign at Ascension Island, 250-MHz receivers were spaced from 300 m to 1.6 km along the field lines, and the signals received from the Marisat satellite were cross correlated. Data collected during eight nights of fading showed a linear relationship between fading rate and cross correlation. The alignment of the antennas was adjusted to give a zero time lag between the widely spaced receivers with a measurement accuracy of 0.03 s. Since the average irregularity velocity was 125 m/s, this time accuracy translated to an angular measurement accuracy of 0.1 deg. During a 4-hour period of nightly fading, occasional differences in time of arrival were noted that corresponded to a tilt in the north-south alignment of + or - 1 deg. Data from several nights of fading were analyzed, and each night exhibited the same variance in the north-south irregularity alignment. It is postulated that the shift in the measured peak correlation may have been caused by patches of irregularities at different altitudes where the magnetic field lines have a slightly different direction. 13 references

  3. On the Total Edge Irregularity Strength of Generalized Butterfly Graph

    Science.gov (United States)

    Dwi Wahyuna, Hafidhyah; Indriati, Diari

    2018-04-01

    Let G(V, E) be a connected, simple, and undirected graph with vertex set V and edge set E. A total k-labeling is a map that carries vertices and edges of a graph G into a set of positive integer labels {1, 2, …, k}. An edge irregular total k-labeling λ: V(G) ∪ E(G) → {1, 2, …, k} of a graph G is a total k-labeling such that the weights calculated for all edges are distinct. The weight of an edge uv in G, denoted by wt(uv), is defined as the sum of the label of u, the label of v, and the label of uv. The total edge irregularity strength of G, denoted by tes(G), is the minimum value of the largest label k over all such edge irregular total k-labelings. A generalized butterfly graph, BFn , obtained by inserting vertices to every wing with assumption that sum of inserting vertices to every wing are same then it has 2n + 1 vertices and 4n ‑ 2 edges. In this paper, we investigate the total edge irregularity strength of generalized butterfly graph, BFn , for n > 2. The result is tes(B{F}n)=\\lceil \\frac{4n}{3}\\rceil .

  4. Une approche de coloriage d’arrêtes pour la conception d’architectures parallèles d’entrelaceurs matériels

    OpenAIRE

    Awais Hussein , Sani

    2012-01-01

    Nowadays, Turbo and LDPC codes are two families of codes that are extensively used in current communication standards due to their excellent error correction capabilities. However, hardware design of coders and decoders for high data rate applications is not a straightforward process. For high data rates, decoders are implemented on parallel architectures in which more than one processing elements decode the received data. To achieve high memory bandwidth, the main memory is divided into smal...

  5. Exploring Manycore Multinode Systems for Irregular Applications with FPGA Prototyping

    Energy Technology Data Exchange (ETDEWEB)

    Ceriani, Marco; Palermo, Gianluca; Secchi, Simone; Tumeo, Antonino; Villa, Oreste

    2013-04-29

    We present a prototype of a multi-core architecture implemented on FPGA, designed to enable efficient execution of irregular applications on distributed shared memory machines, while maintaining high performance on regular workloads. The architecture is composed of off-the-shelf soft-core cores, local interconnection and memory interface, integrated with custom components that optimize it for irregular applications. It relies on three key elements: a global address space, multithreading, and fine-grained synchronization. Global addresses are scrambled to reduce the formation of network hot-spots, while the latency of the transactions is covered by integrating an hardware scheduler within the custom load/store buffers to take advantage from the availability of multiple executions threads, increasing the efficiency in a transparent way to the application. We evaluated a dual node system irregular kernels showing scalability in the number of cores and threads.

  6. Seismic performance for vertical geometric irregularity frame structures

    Science.gov (United States)

    Ismail, R.; Mahmud, N. A.; Ishak, I. S.

    2018-04-01

    This research highlights the result of vertical geometric irregularity frame structures. The aid of finite element analysis software, LUSAS was used to analyse seismic performance by focusing particularly on type of irregular frame on the differences in height floors and continued in the middle of the building. Malaysia’s building structures were affected once the earthquake took place in the neighbouring country such as Indonesia (Sumatera Island). In Malaysia, concrete is widely used in building construction and limited tension resistance to prevent it. Analysing structural behavior with horizontal and vertical static load is commonly analyses by using the Plane Frame Analysis. The case study of this research is to determine the stress and displacement in the seismic response under this type of irregular frame structures. This study is based on seven-storey building of Clinical Training Centre located in Sungai Buloh, Selayang, Selangor. Since the largest earthquake occurs in Acheh, Indonesia on December 26, 2004, the data was recorded and used in conducting this research. The result of stress and displacement using IMPlus seismic analysis in LUSAS Modeller Software under the seismic response of a formwork frame system states that the building is safe to withstand the ground and in good condition under the variation of seismic performance.

  7. Simulating Seismic Wave Propagation in Viscoelastic Media with an Irregular Free Surface

    Science.gov (United States)

    Liu, Xiaobo; Chen, Jingyi; Zhao, Zhencong; Lan, Haiqiang; Liu, Fuping

    2018-05-01

    In seismic numerical simulations of wave propagation, it is very important for us to consider surface topography and attenuation, which both have large effects (e.g., wave diffractions, conversion, amplitude/phase change) on seismic imaging and inversion. An irregular free surface provides significant information for interpreting the characteristics of seismic wave propagation in areas with rugged or rapidly varying topography, and viscoelastic media are a better representation of the earth's properties than acoustic/elastic media. In this study, we develop an approach for seismic wavefield simulation in 2D viscoelastic isotropic media with an irregular free surface. Based on the boundary-conforming grid method, the 2D time-domain second-order viscoelastic isotropic equations and irregular free surface boundary conditions are transferred from a Cartesian coordinate system to a curvilinear coordinate system. Finite difference operators with second-order accuracy are applied to discretize the viscoelastic wave equations and the irregular free surface in the curvilinear coordinate system. In addition, we select the convolutional perfectly matched layer boundary condition in order to effectively suppress artificial reflections from the edges of the model. The snapshot and seismogram results from numerical tests show that our algorithm successfully simulates seismic wavefields (e.g., P-wave, Rayleigh wave and converted waves) in viscoelastic isotropic media with an irregular free surface.

  8. Radar Observations of 8.3-m scale equatorial spread F irregularities over Trivandrum

    Directory of Open Access Journals (Sweden)

    D. Tiwari

    2004-03-01

    Full Text Available In this paper, we present observations of equatorial spread F (ESF irregularities made using a newly installed 18MHz radar located at Trivandrum. We characterize the morphology and the spectral parameters of the 8.3-m ESF irregularities which are found to be remarkably different from that observed so extensively at the 3-m scale size. We also present statistical results of the irregularities in the form of percentage occurrence of the echoes and spectral parameters (SNR, Doppler velocity, Spectral width. The Doppler spectra are narrower, less structured and less variable in time as compared to those observed for 3-m scale size. We have never observed the ESF irregularity velocities to be supersonic here unlike those at Jicamarca, and the velocities are found to be within ±200ms–1. The spectral widths are found to be less than 150ms–1. Hence, the velocities and spectral width both are smaller than those reported for 3-m scale size. The velocities and spectral widths are further found to be much smaller than those of the American sector. These observations are compared with those reported elsewhere and discussed in the light of present understanding on the ESF irregularities at different wavelengths.

    Key words. Ionoshphere (equatorial ionosphere, plasma waves and instabilities; ionospheric irregularities

  9. Efficient irregular wavefront propagation algorithms on Intel® Xeon Phi™

    OpenAIRE

    Gomes, Jeremias M.; Teodoro, George; de Melo, Alba; Kong, Jun; Kurc, Tahsin; Saltz, Joel H.

    2015-01-01

    We investigate the execution of the Irregular Wavefront Propagation Pattern (IWPP), a fundamental computing structure used in several image analysis operations, on the Intel® Xeon Phi™ co-processor. An efficient implementation of IWPP on the Xeon Phi is a challenging problem because of IWPP’s irregularity and the use of atomic instructions in the original IWPP algorithm to resolve race conditions. On the Xeon Phi, the use of SIMD and vectorization instructions is critical to attain high perfo...

  10. Irregular menses predicts ovarian cancer: Prospective evidence from the Child Health and Development Studies.

    Science.gov (United States)

    Cirillo, Piera M; Wang, Erica T; Cedars, Marcelle I; Chen, Lee-May; Cohn, Barbara A

    2016-09-01

    We tested the hypothesis that irregular menstruation predicts lower risk for ovarian cancer, possibly due to less frequent ovulation. We conducted a 50-year prospective study of 15,528 mothers in the Child Health and Development Studies cohort recruited from the Kaiser Foundation Health Plan from 1959 to 1966. Irregular menstruation was classified via medical record and self-report at age 26. We identified 116 cases and 84 deaths due to ovarian cancer through 2011 via linkage to the California Cancer Registry and Vital Statistics. Contrary to expectation, women with irregular menstrual cycles had a higher risk of ovarian cancer incidence and mortality over the 50-year follow-up. Associations increased with age (p irregular menstruation and ovarian cancer-we unexpectedly found higher risk for women with irregular cycles. These women are easy to identify and many may have polycystic ovarian syndrome. Classifying high-risk phenotypes such as irregular menstruation creates opportunities to find novel early biomarkers, refine clinical screening protocols and potentially develop new risk reduction strategies. These efforts can lead to earlier detection and better survival for ovarian cancer. © 2016 UICC.

  11. Comparison of correlation analysis techniques for irregularly sampled time series

    Directory of Open Access Journals (Sweden)

    K. Rehfeld

    2011-06-01

    Full Text Available Geoscientific measurements often provide time series with irregular time sampling, requiring either data reconstruction (interpolation or sophisticated methods to handle irregular sampling. We compare the linear interpolation technique and different approaches for analyzing the correlation functions and persistence of irregularly sampled time series, as Lomb-Scargle Fourier transformation and kernel-based methods. In a thorough benchmark test we investigate the performance of these techniques.

    All methods have comparable root mean square errors (RMSEs for low skewness of the inter-observation time distribution. For high skewness, very irregular data, interpolation bias and RMSE increase strongly. We find a 40 % lower RMSE for the lag-1 autocorrelation function (ACF for the Gaussian kernel method vs. the linear interpolation scheme,in the analysis of highly irregular time series. For the cross correlation function (CCF the RMSE is then lower by 60 %. The application of the Lomb-Scargle technique gave results comparable to the kernel methods for the univariate, but poorer results in the bivariate case. Especially the high-frequency components of the signal, where classical methods show a strong bias in ACF and CCF magnitude, are preserved when using the kernel methods.

    We illustrate the performances of interpolation vs. Gaussian kernel method by applying both to paleo-data from four locations, reflecting late Holocene Asian monsoon variability as derived from speleothem δ18O measurements. Cross correlation results are similar for both methods, which we attribute to the long time scales of the common variability. The persistence time (memory is strongly overestimated when using the standard, interpolation-based, approach. Hence, the Gaussian kernel is a reliable and more robust estimator with significant advantages compared to other techniques and suitable for large scale application to paleo-data.

  12. Drug Intoxicated Irregular Fighters: Complications, Dangers, and Responses

    National Research Council Canada - National Science Library

    Kan, Paul R

    2008-01-01

    .... Drug consumption in contemporary wars has coincided with the use of child soldiers, has led to increased unpredictability among irregular fighters, provided the conditions for the breakdown of social...

  13. Artificial periodic irregularities in the auroral ionosphere

    Directory of Open Access Journals (Sweden)

    M. T. Rietveld

    1996-12-01

    Full Text Available Artificial periodic irregularities (API are produced in the ionospheric plasma by a powerful standing electromagnetic wave reflected off the F region. The resulting electron-density irregularities can scatter other high-frequency waves if the Bragg scattering condition is met. Such measurements have been performed at mid-latitudes for two decades and have been developed into a useful ionospheric diagnostic technique. We report here the first measurements from a high-latitude station, using the EISCAT heating facility near Tromsø, Norway. Both F-region and lower-altitude ionospheric echoes have been obtained, but the bulk of the data has been in the E and D regions with echoes extending down to 52-km altitude. Examples of API are shown, mainly from the D region, together with simultaneous VHF incoherent-scatter-radar (ISR data. Vertical velocities derived from the rate of phase change during the irregularity decay are shown and compared with velocities derived from the ISR. Some of the API-derived velocities in the 75–115-km height range appear consistent with vertical neutral winds as shown by their magnitudes and by evidence of gravity waves, while other data in the 50–70-km range show an unrealistically large bias. For a comparison with ISR data it has proved difficult to get good quality data sets overlapping in height and time. The initial comparisons show some agreement, but discrepancies of several metres per second do not yet allow us to conclude that the two techniques are measuring the same quantity. The irregularity decay time-constants between about 53 and 70 km are compared with the results of an advanced ion-chemistry model, and height profiles of recorded signal power are compared with model estimates in the same altitude range. The calculated amplitude shows good agreement with the data in that the maximum occurs at about the same height as that of the measured amplitude. The calculated time-constant agrees very well with the

  14. Kilometer-Spaced GNSS Array for Ionospheric Irregularity Monitoring

    Science.gov (United States)

    Su, Yang

    This dissertation presents automated, systematic data collection, processing, and analysis methods for studying the spatial-temporal properties of Global Navigation Satellite Systems (GNSS) scintillations produced by ionospheric irregularities at high latitudes using a closely spaced multi-receiver array deployed in the northern auroral zone. The main contributions include 1) automated scintillation monitoring, 2) estimation of drift and anisotropy of the irregularities, 3) error analysis of the drift estimates, and 4) multi-instrument study of the ionosphere. A radio wave propagating through the ionosphere, consisting of ionized plasma, may suffer from rapid signal amplitude and/or phase fluctuations known as scintillation. Caused by non-uniform structures in the ionosphere, intense scintillation can lead to GNSS navigation and high-frequency (HF) communication failures. With specialized GNSS receivers, scintillation can be studied to better understand the structure and dynamics of the ionospheric irregularities, which can be parameterized by altitude, drift motion, anisotropy of the shape, horizontal spatial extent and their time evolution. To study the structuring and motion of ionospheric irregularities at the sub-kilometer scale sizes that produce L-band scintillations, a closely-spaced GNSS array has been established in the auroral zone at Poker Flat Research Range, Alaska to investigate high latitude scintillation and irregularities. Routinely collecting low-rate scintillation statistics, the array database also provides 100 Hz power and phase data for each channel at L1/L2C frequency. In this work, a survey of seasonal and hourly dependence of L1 scintillation events over the course of a year is discussed. To efficiently and systematically study scintillation events, an automated low-rate scintillation detection routine is established and performed for each day by screening the phase scintillation index. The spaced-receiver technique is applied to cross

  15. Irregular ionization and scintillation of the ionosphere in equator region

    International Nuclear Information System (INIS)

    Shinno, Kenji

    1974-01-01

    The latest studies on the scintillation in satellite communication and its related irregularities of ionosphere are reviewed. They were made clear by means of spread-F, the direct measurement with scientific satellites, VHF radar observation, and radio wave propagation in equator region. The fundamental occurrence mechanism may be instability of plasma caused by the interaction of movement of neutral atmosphere and magnetic field. Comparison of the main characteristics of scintillation, namely the dependence on region, solar activity, season, local time, geomagnetic activity, movement in ionosphere, scattering source, frequency and transmission mode, was made and the correlation among spread-F, TEP and scintillation was summarized. The latest principal studies were the observations made by Intelsat and by ATS. Scintillation of Syncom-3 and Intelsat-II-F2 and spread-F by ionosphere observation were compared by Huang. It is reasonable to consider that the occurrence of scintillation is caused by the irregularities in ionosphere which are particular in equator region, because of the similar characteristics of spread-F and VHF propagation in the equator region. These three phenomena may occur in relation to the irregularities of ionosphere. Interpretation of spread-F and the abnormal propagation wave across the equator are given. The study using VHF radar and the movement of irregular ionization by the direct observation with artificial satellites are reviewd. (Iwakiri, K.)

  16. Exploring performance and energy tradeoffs for irregular applications: A case study on the Tilera many-core architecture

    Energy Technology Data Exchange (ETDEWEB)

    Panyala, Ajay; Chavarría-Miranda, Daniel; Manzano, Joseph B.; Tumeo, Antonino; Halappanavar, Mahantesh

    2017-06-01

    High performance, parallel applications with irregular data accesses are becoming a critical workload class for modern systems. In particular, the execution of such workloads on emerging many-core systems is expected to be a significant component of applications in data mining, machine learning, scientific computing and graph analytics. However, power and energy constraints limit the capabilities of individual cores, memory hierarchy and on-chip interconnect of such systems, thus leading to architectural and software trade-os that must be understood in the context of the intended application’s behavior. Irregular applications are notoriously hard to optimize given their data-dependent access patterns, lack of structured locality and complex data structures and code patterns. We have ported two irregular applications, graph community detection using the Louvain method (Grappolo) and high-performance conjugate gradient (HPCCG), to the Tilera many-core system and have conducted a detailed study of platform-independent and platform-specific optimizations that improve their performance as well as reduce their overall energy consumption. To conduct this study, we employ an auto-tuning based approach that explores the optimization design space along three dimensions - memory layout schemes, GCC compiler flag choices and OpenMP loop scheduling options. We leverage MIT’s OpenTuner auto-tuning framework to explore and recommend energy optimal choices for different combinations of parameters. We then conduct an in-depth architectural characterization to understand the memory behavior of the selected workloads. Finally, we perform a correlation study to demonstrate the interplay between the hardware behavior and application characteristics. Using auto-tuning, we demonstrate whole-node energy savings and performance improvements of up to 49:6% and 60% relative to a baseline instantiation, and up to 31% and 45:4% relative to manually optimized variants.

  17. Irregular Warfare: New Challenges for Civil-Military Relations

    National Research Council Canada - National Science Library

    Cronin, Patrick M

    2008-01-01

    .... Irregular warfare introduces new complications to what Eliot Cohen has called an unequal dialogue between civilian and military leaders in which civilian leaders hold the true power but must modulate...

  18. FINELM: a multigroup finite element diffusion code. Part I

    International Nuclear Information System (INIS)

    Davierwalla, D.M.

    1980-12-01

    The author presents a two dimensional code for multigroup diffusion using the finite element method. It was realized that the extensive connectivity which contributes significantly to the accuracy, results in a matrix which, although symmetric and positive definite, is wide band and possesses an irregular profile. Hence, it was decided to introduce sparsity techniques into the code. The introduction of the R-Z geometry lead to a great deal of changes in the code since the rotational invariance of the removal matrices in X-Y geometry did not carry over in R-Z geometry. Rectangular elements were introduced to remedy the inability of the triangles to model essentially one dimensional problems such as slab geometry. The matter is discussed briefly in the text in the section on benchmark problems. This report is restricted to the general theory of the triangular elements and to the sparsity techniques viz. incomplete disections. The latter makes the size of the problem that can be handled independent of core memory and dependent only on disc storage capacity which is virtually unlimited. (Auth.)

  19. Method of determining effects of heat-induced irregular refractive index on an optical system.

    Science.gov (United States)

    Song, Xifa; Li, Lin; Huang, Yifan

    2015-09-01

    The effects of an irregular refractive index on optical performance are examined. A method was developed to express a lens's irregular refractive index distribution. An optical system and its mountings were modeled by a thermomechanical finite element (FE) program in the predicted operating temperature range, -45°C-50°C. FE outputs were elaborated using a MATLAB optimization routine; a nonlinear least squares algorithm was adopted to determine which gradient equation best fit each lens's refractive index distribution. The obtained gradient data were imported into Zemax for sequential ray-tracing analysis. The root mean square spot diameter, modulation transfer function, and diffraction ensquared energy were computed for an optical system under an irregular refractive index and under thermoelastic deformation. These properties are greatly reduced by the irregular refractive index effect, which is one-third to five-sevenths the size of the thermoelastic deformation effect. Thus, thermal analyses of optical systems should consider not only thermoelastic deformation but also refractive index irregularities caused by inhomogeneous temperature.

  20. Significance of scatter radar studies of E and F region irregularities at high latitudes

    International Nuclear Information System (INIS)

    Greenwald, R.A.

    1983-01-01

    This chapter considers the mechanisms by which electron density irregularities may be generated in the high latitude ionosphere and the techniques through which they are observed with ground base radars. The capabilities of radars used for studying these irregularities are compared with the capabilities of radars used for incoherent scatter measurements. The use of irregularity scatter techniques for dynamic studies of larger scale structured phenomena is discussed. Topics considered include E-region irregularities, observations with auroral radars, plasma drifts associated with a westward travelling surge, and ionospheric plasma motions associated with resonant waves. It is shown why high latitude F-region irregularity studies must be made in the HF frequency band (3-30 MHz). The joint use of the European Incoherent Scatter Association (EISCAT), STARE and SAFARI facilities is examined, and it is concluded that the various techniques will enhance each other and provide a better understanding of the various processes being studied

  1. [Comparision of Different Methods of Area Measurement in Irregular Scar].

    Science.gov (United States)

    Ran, D; Li, W J; Sun, Q G; Li, J Q; Xia, Q

    2016-10-01

    To determine a measurement standard of irregular scar area by comparing the advantages and disadvantages of different measurement methods in measuring same irregular scar area. Irregular scar area was scanned by digital scanning and measured by coordinate reading method, AutoCAD pixel method, Photoshop lasso pixel method, Photoshop magic bar filled pixel method and Foxit PDF reading software, and some aspects of these methods such as measurement time, repeatability, whether could be recorded and whether could be traced were compared and analyzed. There was no significant difference in the scar areas by the measurement methods above. However, there was statistical difference in the measurement time and repeatability by one or multi performers and only Foxit PDF reading software could be traced back. The methods above can be used for measuring scar area, but each one has its advantages and disadvantages. It is necessary to develop new measurement software for forensic identification. Copyright© by the Editorial Department of Journal of Forensic Medicine

  2. Irregular flowering patterns in terrestrial orchids: theories vs. empirical data

    Directory of Open Access Journals (Sweden)

    P. Kindlmann

    2001-11-01

    Full Text Available Empirical data on many species of terrestrial orchids suggest that their between-year flowering pattern is extremely irregular and unpredictable. A long search for the reason has hitherto proved inconclusive. Here we summarise and critically review the hypotheses that were put forward as explanations of this phenomenon: irregular flowering was attributed to costs associated with sexual reproduction, to herbivory, or to the chaotic behaviour of the system represented by difference equations describing growth of the vegetative and reproductive organs. None of these seems to explain fully the events of a transition from flowering one year to sterility or absence the next year. Data on the seasonal growth of leaves and inflorescence of two terrestrial orchid species, Epipactis albensis and Dactylorhiza fuchsii and our previous results are then used here to fill gaps in what has been published until now and to test alternative explanations of the irregular flowering patterns of orchids.

  3. New prospective 4D-CT for mitigating the effects of irregular respiratory motion

    Science.gov (United States)

    Pan, Tinsu; Martin, Rachael M.; Luo, Dershan

    2017-08-01

    Artifact caused by irregular respiration is a major source of error in 4D-CT imaging. We propose a new prospective 4D-CT to mitigate this source of error without new hardware, software or off-line data-processing on the GE CT scanner. We utilize the cine CT scan in the design of the new prospective 4D-CT. The cine CT scan at each position can be stopped by the operator when an irregular respiration occurs, and resumed when the respiration becomes regular. This process can be repeated at one or multiple scan positions. After the scan, a retrospective reconstruction is initiated on the CT console to reconstruct only the images corresponding to the regular respiratory cycles. The end result is a 4D-CT free of irregular respiration. To prove feasibility, we conducted a phantom and six patient studies. The artifacts associated with the irregular respiratory cycles could be removed from both the phantom and patient studies. A new prospective 4D-CT scanning and processing technique to mitigate the impact of irregular respiration in 4D-CT has been demonstrated. This technique can save radiation dose because the repeat scans are only at the scan positions where an irregular respiration occurs. Current practice is to repeat the scans at all positions. There is no cost to apply this technique because it is applicable on the GE CT scanner without new hardware, software or off-line data-processing.

  4. Observations of inner plasmasphere irregularities with a satellite-beacon radio-interferometer array

    International Nuclear Information System (INIS)

    Jacobson, A.R.; Hoogeveen, G.; Carlos, R.C.; Wu, G.; Fejer, B.G.; Kelley, M.C.

    1996-01-01

    A radio-interferometer array illuminated by 136-MHz beacons of several geosynchronous satellites has been used to study small (≥10 13 m -2 ) transient disturbances in the total electron content along the lines of sight to the satellites. High-frequency (f>3 mHz) electron content oscillations are persistently observed, particularly during night and particularly during geomagnetically disturbed periods. The oscillations move across the array plane at speeds in the range 200 endash 2000 m/s, with propagation azimuths that are strongly peaked in lobes toward the western half-plane. Detailed analysis of this azimuth behavior, involving comparison between observations on various satellite positions, indicates compellingly that the phase oscillations originate in radio refraction due to geomagnetically aligned plasma density perturbations in the inner plasmasphere. The motion of the phase perturbations across the array plane is caused by EXB drift of the plasma medium in which the irregularities are embedded. We review the statistics of 2.5 years of around-the-clock data on the local time, magnetic disturbance, seasonal, and line-of-sight variations of these observed irregularities. We compare the irregularities close-quote inferred electrodynamic drifts to what is known about midlatitude plasma drift from incoherent scatter. Finally, we show in detail how the observation of these irregularities provides a unique and complementary monitor of inner plasmasphere irregularity incidence and zonal drift.copyright 1996 American Geophysical Union

  5. Influence of long-wavelength track irregularities on the motion of a high-speed train

    Science.gov (United States)

    Hung, C. F.; Hsu, W. L.

    2018-01-01

    Vertical track irregularities over viaducts in high-speed rail systems could be possibly caused by concrete creep if pre-stressed concrete bridges are used. For bridge spans that are almost uniformly distributed, track irregularity exhibits a near-regular wave profile that excites car bodies as a high-speed train moves over the bridge system. A long-wavelength irregularity induces low-frequency excitation that may be close to the natural frequencies of the train suspension system, thereby causing significant vibration of the car body. This paper investigates the relationship between the levels of car vibration, bridge vibration, track irregularity, and the train speed. First, this study investigates the vibration levels of a high-speed train and bridge system using 3D finite-element (FE) transient dynamic analysis, before and after adjustment of vertical track irregularities by means of installing shimming plates under rail pads. The analysis models are validated by in situ measurements and on-board measurement. Parametric studies of car body vibration and bridge vibration under three different levels of track irregularity at five train speeds and over two bridge span lengths are conducted using the FE model. Finally, a discontinuous shimming pattern is proposed to avoid vehicle suspension resonance.

  6. Synchronizing data from irregularly sampled sensors

    Science.gov (United States)

    Uluyol, Onder

    2017-07-11

    A system and method include receiving a set of sampled measurements for each of multiple sensors, wherein the sampled measurements are at irregular intervals or different rates, re-sampling the sampled measurements of each of the multiple sensors at a higher rate than one of the sensor's set of sampled measurements, and synchronizing the sampled measurements of each of the multiple sensors.

  7. The analysis of impact of irregularity in radionuclide coating of scaffold on the distribution of absorbed dose produced by grid of microsources

    Directory of Open Access Journals (Sweden)

    N. A. Nerosin

    2015-01-01

    Full Text Available The impact of irregularity in radionuclide coating of scaffold on the distribution of absorbed dose produced by grid of microsources was analyzed. On engineering software MATHCAD the program for calculation of absorbed dose produced by grid of microsources was created. To verify this algorithm the calculation model for MCNP code was established and represented the area consisted of soft biological tissue or any other tissue in which the grid of microsources was incorporated. Using the developed system the value of possible systematic irregular coating of radioactivity on the microsource’s core was analyzed. The distribution of activity along the surface of microsource was simulated to create distribution of absorbed dose rate corresponding to experimental data on radiation injury. The obtained model of microsource with irregular distribution of activity was compared to conventional microsource with core coated regularly along the entire area of the silver stem by main dosimetry characteristics. The results showed that even for extremely irregular distribution of activity the distribution of dose rate produced by microsource in the tumor area was not substantially different from dose-rate field obtained for microsource with regularly coated activity. The differences in dose rates (up to 10% in areas which were the nearest to the center of the grid were significantly lower than its decline from center to periphery of the grid. For spatial distribution of absorbed dose for specific configuration of microsource set and tracing of curves of equal level by selected cut-off the program SEEDPLAN was developed. The developed program represents precisely enough the spatial distribution of selected configuration set of microsources using results of calculation data for absorbed dose around the single microsource as basic data and may be used for optimal planning of brachytherapy with microsources. 

  8. Author Details

    African Journals Online (AJOL)

    Kebre, M. Vol 12 (2013) - Articles BER Reduction for mQAM High-Speed Tranmission in Optical Fiber Networks Using Convalutional BCH, RS and LDPC Error Correcting Codes Abstract PDF. ISSN: 1596-292X. AJOL African Journals Online. HOW TO USE AJOL... for Researchers · for Librarians · for Authors · FAQ's · More ...

  9. A Bandwidth-Optimized Multi-Core Architecture for Irregular Applications

    Energy Technology Data Exchange (ETDEWEB)

    Secchi, Simone; Tumeo, Antonino; Villa, Oreste

    2012-05-31

    This paper presents an architecture template for next-generation high performance computing systems specifically targeted to irregular applications. We start our work by considering that future generation interconnection and memory bandwidth full-system numbers are expected to grow by a factor of 10. In order to keep up with such a communication capacity, while still resorting to fine-grained multithreading as the main way to tolerate unpredictable memory access latencies of irregular applications, we show how overall performance scaling can benefit from the multi-core paradigm. At the same time, we also show how such an architecture template must be coupled with specific techniques in order to optimize bandwidth utilization and achieve the maximum scalability. We propose a technique based on memory references aggregation, together with the related hardware implementation, as one of such optimization techniques. We explore the proposed architecture template by focusing on the Cray XMT architecture and, using a dedicated simulation infrastructure, validate the performance of our template with two typical irregular applications. Our experimental results prove the benefits provided by both the multi-core approach and the bandwidth optimization reference aggregation technique.

  10. Uniform irradiation of irregularly shaped cavities for photodynamic therapy.

    Science.gov (United States)

    Rem, A I; van Gemert, M J; van der Meulen, F W; Gijsbers, G H; Beek, J F

    1997-03-01

    It is difficult to achieve a uniform light distribution in irregularly shaped cavities. We have conducted a study on the use of hollow 'integrating' moulds for more uniform light delivery of photodynamic therapy in irregularly shaped cavities such as the oral cavity. Simple geometries such as a cubical box, a sphere, a cylinder and a 'bottle-neck' geometry have been investigated experimentally and the results have been compared with computed light distributions obtained using the 'radiosity method'. A high reflection coefficient of the mould and the best uniform direct irradiance possible on the inside of the mould were found to be important determinants for achieving a uniform light distribution.

  11. Post-midnight equatorial irregularity distributions and vertical drift velocity variations during solstices

    Science.gov (United States)

    Su, S.-Y.; Liu, C. H.; Chao, C.-K.

    2018-04-01

    Longitudinal distributions of post-midnight equatorial ionospheric irregularity occurrences observed by ROCSAT-1 (1st satellite of the Republic of China) during moderate to high solar activity years in two solstices are studied with respect to the vertical drift velocity and density variations. The post-midnight irregularity distributions are found to be similar to the well-documented pre-midnight ones, but are different from some published distributions taken during solar minimum years. Even though the post-midnight ionosphere is sinking in general, longitudes of frequent positive vertical drift and high density seems to coincide with the longitudes of high irregularity occurrences. Large scatters found in the vertical drift velocity and density around the dip equator in different ROCSAT-1 orbits indicate the existence of large and frequent variations in the vertical drift velocity and density that seem to be able to provide sufficient perturbations for the Rayleigh-Taylor (RT) instability to cause the irregularity occurrences. The need of seeding agents such as gravity waves from atmospheric convective clouds to initiate the Rayleigh-Taylor instability may not be necessary.

  12. Breaking Down Anonymity: Digital surveillance on irregular migrants in Germany and the Netherlands

    NARCIS (Netherlands)

    D.W.J. Broeders (Dennis)

    2009-01-01

    textabstractThe presence of irregular migrants causes a tough problem for policy makers. Political and popular aversion against the presence of irregular migrants has mounted in most West-European societies for years, yet their presence remains. Their exact numbers are obviously unknown - only

  13. Simultaneous radar and spaced receiver VHF scintillation observations of ESF irregularities

    Directory of Open Access Journals (Sweden)

    D. Tiwari

    2006-07-01

    Full Text Available Simultaneous observations of equatorial spread F (ESF irregularities made on 10 nights during March-April 1998 and 1999, using an 18-MHz radar at Trivandrum (77° E, 8.5° N, dip 0.5° N and two spaced receivers recording scintillations on a 251-MHz signal at Tirunelveli (77.8° E, 8.7° N, dip 0.4° N, have been used to study the evolution of Equatorial Spread F (ESF irregularities. Case studies have been carried out on the day-to-day variability in ESF structure and dynamics, as observed by 18-MHz radar, and with spaced receiver measurements of average zonal drift Vo of the 251-MHz radio wave diffraction pattern on the ground, random velocity Vc, which is a measure of random changes in the characteristics of scintillation-producing irregularities, and maximum cross-correlation CI of the spaced receivers signals. Results show that in the initial phase of plasma bubble development, the greater the maximum height of ESF irregularities responsible for the radar backscatter, the greater the decorrelation is of the spaced receiver scintillation signals, indicating greater turbulence. The relationship of the maximum spectral width derived from the radar observations and CI also supports this result.

  14. [Artificial cycle therapy of acupuncture and moxibustion for irregular menstruation].

    Science.gov (United States)

    Wu, Jie; Yang, Lijie; Chen, Yajie; Li, Qing; Chen, Lin

    2015-03-01

    Through the discussion on TCM physiological characters of females in follicular, ovulatory, luteal and menstrual phases and treatment principles, the clinical application of artificial cycle therapy of acupuncture and moxibustion was introduced for irregular menstruation and the typical cases were attached. It is suggested that the menstrual cycle follows the growth-consumption rule of yin, yang, qi and blood. The corresponding treatment principles should be applied in accordance with the change rule of menstrual cycle. Hence, it is worth to adopt the artificial cycle therapy of acupuncture and moxibustion for irregular menstruation in clinical application.

  15. Properties of the Irregular Satellite System around Uranus Inferred from K2, Herschel, and Spitzer Observations

    Science.gov (United States)

    Farkas-Takács, A.; Kiss, Cs.; Pál, A.; Molnár, L.; Szabó, Gy. M.; Hanyecz, O.; Sárneczky, K.; Szabó, R.; Marton, G.; Mommert, M.; Szakáts, R.; Müller, T.; Kiss, L. L.

    2017-09-01

    In this paper, we present visible-range light curves of the irregular Uranian satellites Sycorax, Caliban, Prospero, Ferdinand, and Setebos taken with the Kepler Space Telescope over the course of the K2 mission. Thermal emission measurements obtained with the Herschel/PACS and Spitzer/MIPS instruments of Sycorax and Caliban were also analyzed and used to determine size, albedo, and surface characteristics of these bodies. We compare these properties with the rotational and surface characteristics of irregular satellites in other giant planet systems and also with those of main belt and Trojan asteroids and trans-Neptunian objects. Our results indicate that the Uranian irregular satellite system likely went through a more intense collisional evolution than the irregular satellites of Jupiter and Saturn. Surface characteristics of Uranian irregular satellites seem to resemble the Centaurs and trans-Neptunian objects more than irregular satellites around other giant planets, suggesting the existence of a compositional discontinuity in the young solar system inside the orbit of Uranus.

  16. Properties of the Irregular Satellite System around Uranus Inferred from K2 , Herschel , and Spitzer Observations

    Energy Technology Data Exchange (ETDEWEB)

    Farkas-Takács, A.; Kiss, Cs.; Pál, A.; Molnár, L.; Szabó, Gy. M.; Hanyecz, O.; Sárneczky, K.; Szabó, R.; Marton, G.; Szakáts, R.; Kiss, L. L. [Konkoly Observatory, Research Centre for Astronomy and Earth Sciences, Hungarian Academy of Sciences, Konkoly Thege Miklós út 15-17, H-1121 Budapest (Hungary); Mommert, M. [Department of Physics and Astronomy, Northern Arizona University, P.O. Box 6010, Flagstaff, AZ 86011 (United States); Müller, T., E-mail: farkas.aniko@csfk.mta.hu [Max-Plank-Institut für extraterrestrsiche Pyhsik, Garching (Germany)

    2017-09-01

    In this paper, we present visible-range light curves of the irregular Uranian satellites Sycorax, Caliban, Prospero, Ferdinand, and Setebos taken with the Kepler Space Telescope over the course of the K2 mission. Thermal emission measurements obtained with the Herschel /PACS and Spitzer /MIPS instruments of Sycorax and Caliban were also analyzed and used to determine size, albedo, and surface characteristics of these bodies. We compare these properties with the rotational and surface characteristics of irregular satellites in other giant planet systems and also with those of main belt and Trojan asteroids and trans-Neptunian objects. Our results indicate that the Uranian irregular satellite system likely went through a more intense collisional evolution than the irregular satellites of Jupiter and Saturn. Surface characteristics of Uranian irregular satellites seem to resemble the Centaurs and trans-Neptunian objects more than irregular satellites around other giant planets, suggesting the existence of a compositional discontinuity in the young solar system inside the orbit of Uranus.

  17. On the total irregularity strength of caterpillar with each internal vertex has degree three

    Science.gov (United States)

    Indriati, Diari; Rosyida, Isnaini; Widodo

    2018-04-01

    Let G be a simple, connected and undirected graph with vertex set V and edge set E. A total k-labeling f:V \\cup E\\to \\{1,2,\\ldots,k\\} is defined as totally irregular total k-labeling if the weights of any two different both vertices and edges are distinct. The weight of vertex x is defined as wt(x)=f(x)+{\\sum }xy\\in Ef(xy), while the weight of edge xy is wt(xy)=f(x)+f(xy)+f(y). A minimum k for which G has totally irregular total k-labeling is mentioned as total irregularity strength of G and denoted by ts(G). This paper contains investigation of totally irregular total k-labeling and determination of their total irregularity strengths for caterpillar graphs with each internal vertex between two stars has degree three. The results are ts({S}n,3,n)=\\lceil \\frac{2n}{2}\\rceil, ts({S}n,3,3,n)=\\lceil \\frac{2n+1}{2}\\rceil and ts({S}n,3,3,3,n)=\\lceil \\frac{2n+2}{2}\\rceil for n > 4:

  18. The effect of earthquake on architecture geometry with non-parallel system irregularity configuration

    Science.gov (United States)

    Teddy, Livian; Hardiman, Gagoek; Nuroji; Tudjono, Sri

    2017-12-01

    Indonesia is an area prone to earthquake that may cause casualties and damage to buildings. The fatalities or the injured are not largely caused by the earthquake, but by building collapse. The collapse of the building is resulted from the building behaviour against the earthquake, and it depends on many factors, such as architectural design, geometry configuration of structural elements in horizontal and vertical plans, earthquake zone, geographical location (distance to earthquake center), soil type, material quality, and construction quality. One of the geometry configurations that may lead to the collapse of the building is irregular configuration of non-parallel system. In accordance with FEMA-451B, irregular configuration in non-parallel system is defined to have existed if the vertical lateral force-retaining elements are neither parallel nor symmetric with main orthogonal axes of the earthquake-retaining axis system. Such configuration may lead to torque, diagonal translation and local damage to buildings. It does not mean that non-parallel irregular configuration should not be formed on architectural design; however the designer must know the consequence of earthquake behaviour against buildings with irregular configuration of non-parallel system. The present research has the objective to identify earthquake behaviour in architectural geometry with irregular configuration of non-parallel system. The present research was quantitative with simulation experimental method. It consisted of 5 models, where architectural data and model structure data were inputted and analyzed using the software SAP2000 in order to find out its performance, and ETAB2015 to determine the eccentricity occurred. The output of the software analysis was tabulated, graphed, compared and analyzed with relevant theories. For areas of strong earthquake zones, avoid designing buildings which wholly form irregular configuration of non-parallel system. If it is inevitable to design a

  19. Introduction to digital communication systems

    CERN Document Server

    Wesolowski, Krzysztof

    2009-01-01

    Combining theoretical knowledge and practical applications, this advanced-level textbook covers the most important aspects of contemporary digital communication systems. Introduction to Digital Communication Systems focuses on the rules of functioning digital communication system blocks, starting with the performance limits set by the information theory. Drawing on information relating to turbo codes and LDPC codes, the text presents the basic methods of error correction and detection, followed by baseband transmission methods, and single- and multi-carrier digital modulations. The basi

  20. Evaluation of irregular menses in perimenarcheal girls: a pilot study.

    Science.gov (United States)

    Browner-Elhanan, Karen J; Epstein, Jonathan; Alderman, Elizabeth M

    2003-12-01

    Acyclic vaginal bleeding in girls within three years of menarche is most commonly attributed to an immature hypothalamic-pituitary-ovarian axis. Assuming this diagnosis may preclude the practitioner from performing more definitive studies and thereby diagnosing other, treatable causes of menstrual irregularities. A retrospective chart review of 178 girls presenting to an inner-city hospital-based adolescent clinic within three years of menarche was performed. Personal and family medical and menarcheal history was assessed, and findings on physical and laboratory examination performed were evaluated. Of the 178 girls still perimenarcheal at presentation, 47 were the focus of this study. Of these, 39 had no significant findings on physical examination, while 3 had signs of functional ovarian hyperandrogenism (FOH) including obesity, hirsutism, and moderate acne with corresponding LH/FSH ratios>3, although pelvic ultrasound examination revealed normal ovaries. Four of the 39 patients with normal physical exams had LH/FSH testing done, and 1 of the 4 had an abnormal LH/FSH ratio, indicating possible FOH. Two of the 47 patients were pregnant. Other laboratory abnormalities included microcytic, hypochromic anemia in patients, and an elevated Erythrocyte Sedimentation Rate in a patient later diagnosed with a rheumatologic disorder. Those perimenarcheal girls presenting with irregular menses and findings including obesity, acne, or pallor, were likely to have treatable causes of menstrual irregularities. In one of the four girls with a normal physical examination, hormonal testing indicated possible FOH, thus suggesting that hormonal evaluation of perimenarcheal girls with menstrual irregularities may be justified, as it may reveal previously unsuspected pathology.

  1. Irregular Warfare: Impact on Future Professional Military Education

    National Research Council Canada - National Science Library

    Paschal, David G

    2006-01-01

    ... to operate effectively in an irregular warfare environment. The utility of a decisive war between nation states continues to decline and will eventually reach critical mass based upon the extreme imbalance of military power and a U.S. monopoly...

  2. Research on making reactor buildings of irregular plan and elevation forms aseismatic

    International Nuclear Information System (INIS)

    Okawa, Izuru; Yamauchi, Yasuyuki

    1997-01-01

    The necessity of pursuing the possibility of irregular form buildings as the condition of location for construction is limited, and the rational and economical arrangement of equipment and piping is considered. In order to know the effect that irregular forms exert to the aseismatic ability of buildings, it is indispensable to develop the program for precision three-dimensional elastoplastic analysis at the time of earthquakes. As the means of solving the problem, the introduction of seismic insulation structure is conceivable. The investigation of seismic insulator and its modeling and the analysis of earthquake response were carried out, and the irregular form and the effect of seismic insulation were investigated, and the results of vibration test using test specimens were summarized. The concrete items of investigation were the characteristics of input earthquake motion, the techniques of analysis, the parametric study taking the input and various characteristics of buildings in consideration, and the synthetic assessment. The vibration table experiment and the static loading experiment for the purpose of grasping the response behavior in the case of irregular form of wall type and seismic insulation type structures were carried out, and the results are reported. (K.I.)

  3. Irregular analytical errors in diagnostic testing - a novel concept.

    Science.gov (United States)

    Vogeser, Michael; Seger, Christoph

    2018-02-23

    In laboratory medicine, routine periodic analyses for internal and external quality control measurements interpreted by statistical methods are mandatory for batch clearance. Data analysis of these process-oriented measurements allows for insight into random analytical variation and systematic calibration bias over time. However, in such a setting, any individual sample is not under individual quality control. The quality control measurements act only at the batch level. Quantitative or qualitative data derived for many effects and interferences associated with an individual diagnostic sample can compromise any analyte. It is obvious that a process for a quality-control-sample-based approach of quality assurance is not sensitive to such errors. To address the potential causes and nature of such analytical interference in individual samples more systematically, we suggest the introduction of a new term called the irregular (individual) analytical error. Practically, this term can be applied in any analytical assay that is traceable to a reference measurement system. For an individual sample an irregular analytical error is defined as an inaccuracy (which is the deviation from a reference measurement procedure result) of a test result that is so high it cannot be explained by measurement uncertainty of the utilized routine assay operating within the accepted limitations of the associated process quality control measurements. The deviation can be defined as the linear combination of the process measurement uncertainty and the method bias for the reference measurement system. Such errors should be coined irregular analytical errors of the individual sample. The measurement result is compromised either by an irregular effect associated with the individual composition (matrix) of the sample or an individual single sample associated processing error in the analytical process. Currently, the availability of reference measurement procedures is still highly limited, but LC

  4. Rocket observation of electron density irregularities in the lower E region

    International Nuclear Information System (INIS)

    Watanabe, Yuzo; Nakamura, Yoshiharu; Amemiya, Hiroshi.

    1990-01-01

    Local ionospheric electron density irregularities in the scale size of 3 m to 300 m have been measured on the ascending path from 74 km to 93 km by a fix biased Langmuir probe on board the S-310-16 sounding rocket. The rocket was launched at 22:40:00 on February 1, 1986 from Kagoshima Space Center in Japan. It is found from frequency analysis of the data that the spectral index of the irregularities is 0.9 to 1.8 and the irregularity amplitude is 1 to 15 %. The altitude where the amplitude reaches its maximum is 88 km. The generation mechanism of these irregularities is explained by the neutral turbulence theory, which indicates that the spectral index is 5/3 and has been confirmed by a chemical release experiment using rockets over India to be valid up to about 110 km. From frequency analysis of the data observed during the descent in the lower E region, we have found that the rocket-wake effect becomes larger when the probe is situated near the edge of the rocket-wake, and that this is also the case even when the rocket-wake effect does not clearly appear in the DC current signal which approximately changes in proportion to the electron density, where the probe is completely situated inside the rocket-wake region. (author)

  5. PERFORMANCE EVOLUTION OF PAPR REDUCTION IN OFDM WITH AND WITHOUT LDPC TECHNIQUE

    OpenAIRE

    Punit Upmanyu*; Prof. Saurabh Gaur

    2016-01-01

    The OFDM is one of the proven multicarrier modulation techniques, which provides high spectral efficiency, low implementation complexity, less vulnerability to echoes and non-linear distortion. Apart from the above advantages presently this technique is used by almost all wireless standards and above. The one major shortcoming in the implementation of this system is the high PAPR (peak-to-average power ratio) of this system. In this paper, Irregular Low-Density-Parity Check encoder is used ef...

  6. Irregular Polyomino-Shaped Subarrays for Space-Based Active Arrays

    Directory of Open Access Journals (Sweden)

    R. J. Mailloux

    2009-01-01

    Full Text Available This paper presents new results showing the application of polyomino-based subarrays to limited field of view and wideband, wide-angle scanning. This technology can reduce the number of phase controls in arrays used for limited sector coverage or the number of time delay devices for wideband radar or communications, and so can reduce the cost of space-based active arrays. We concentrate on the wideband application. Results are presented by comparing the gain and peak sidelobe results of irregular polyomino subarray-based arrays with those of rectangular subarrays. It is shown that using irregular polyomino subarrays can result in a major decrease in sidelobes while presenting, in most cases, only a few tenths of a dB gain reduction compared to rectangular subarrays.

  7. Third-order theory for multi-directional irregular waves

    DEFF Research Database (Denmark)

    Madsen, Per A.; Fuhrman, David R.

    2012-01-01

    A new third-order solution for multi-directional irregular water waves in finite water depth is presented. The solution includes explicit expressions for the surface elevation, the amplitude dispersion and the vertical variation of the velocity potential. Expressions for the velocity potential at...

  8. Low-complexity video encoding method for wireless image transmission in capsule endoscope.

    Science.gov (United States)

    Takizawa, Kenichi; Hamaguchi, Kiyoshi

    2010-01-01

    This paper presents a low-complexity video encoding method applicable for wireless image transmission in capsule endoscopes. This encoding method is based on Wyner-Ziv theory, in which side information available at a transmitter is treated as side information at its receiver. Therefore complex processes in video encoding, such as estimation of the motion vector, are moved to the receiver side, which has a larger-capacity battery. As a result, the encoding process is only to decimate coded original data through channel coding. We provide a performance evaluation for a low-density parity check (LDPC) coding method in the AWGN channel.

  9. Some new features of electron density irregularities over SHAR during strong spread F

    Directory of Open Access Journals (Sweden)

    S. Raizada

    Full Text Available An RH-560 rocket flight was conducted from Sriharikota rocket range (SHAR (14°N, 80°E, dip latitude 5.5°N to study electron density and electric field irregularities during spread F. The rocket was launched at 2130 local time (LT and it attained an apogee of 348 km. Results of electron density fluctuations are presented here. Two extremely sharp layers of very high electron density were observed at 105 and 130 km. The electron density increase in these layers was by a factor of 50 in a vertical extent of 10 km. Large depletions in electron density were observed around 175 and 238 km. Both sharp layers as well as depletions were observed also during the descent. The presence of sharp layers and depletions during the ascent and the descent of the rocket as well as an order of magnitude less electron density, in 150-300 km region during the descent, indicate the presence of strong large-scale horizontal gradients in the electron density. Some of the valley region irregularities (165-178 km, in the intermediate scale size range, observed during this flight, show spectral peaks at 2 km and can be interpreted in terms of the image striation theory suggested by Vickrey et al. The irregularities at 176 km do not exhibit any peak at kilometer scales and appear to be of new type. The growth rate of intermediate scale size irregularities, produced through generalized Rayleigh Taylor instability, was calculated for the 200-330 km altitude, using observed values of electron density gradients and an assumed vertically downward wind of 20 ms-1. These growth rate calculations suggest that the observed irregularities could be produced by the gradient drift instability.

    Key words: Ionosphere (equatorial ionosphere; ionospheric irregularities - Radio science (ionospheric physics

  10. Some new features of electron density irregularities over SHAR during strong spread F

    Directory of Open Access Journals (Sweden)

    S. Raizada

    2000-02-01

    Full Text Available An RH-560 rocket flight was conducted from Sriharikota rocket range (SHAR (14°N, 80°E, dip latitude 5.5°N to study electron density and electric field irregularities during spread F. The rocket was launched at 2130 local time (LT and it attained an apogee of 348 km. Results of electron density fluctuations are presented here. Two extremely sharp layers of very high electron density were observed at 105 and 130 km. The electron density increase in these layers was by a factor of 50 in a vertical extent of 10 km. Large depletions in electron density were observed around 175 and 238 km. Both sharp layers as well as depletions were observed also during the descent. The presence of sharp layers and depletions during the ascent and the descent of the rocket as well as an order of magnitude less electron density, in 150-300 km region during the descent, indicate the presence of strong large-scale horizontal gradients in the electron density. Some of the valley region irregularities (165-178 km, in the intermediate scale size range, observed during this flight, show spectral peaks at 2 km and can be interpreted in terms of the image striation theory suggested by Vickrey et al. The irregularities at 176 km do not exhibit any peak at kilometer scales and appear to be of new type. The growth rate of intermediate scale size irregularities, produced through generalized Rayleigh Taylor instability, was calculated for the 200-330 km altitude, using observed values of electron density gradients and an assumed vertically downward wind of 20 ms-1. These growth rate calculations suggest that the observed irregularities could be produced by the gradient drift instability.Key words: Ionosphere (equatorial ionosphere; ionospheric irregularities - Radio science (ionospheric physics

  11. The Relationships among Cognitive Correlates and Irregular Word, Non-Word, and Word Reading

    Science.gov (United States)

    Abu-Hamour, Bashir; University, Mu'tah; Urso, Annmarie; Mather, Nancy

    2012-01-01

    This study explored four hypotheses: (a) the relationships among rapid automatized naming (RAN) and processing speed (PS) to irregular word, non-word, and word reading; (b) the predictive power of various RAN and PS measures, (c) the cognitive correlates that best predicted irregular word, non-word, and word reading, and (d) reading performance of…

  12. SUB-SAHARAN IRREGULAR MIGRANTS IN MOROCCO & THE EXCEPTIONAL REGULARISATION PROGRAMME IN 2014

    OpenAIRE

    Naama, Mbarek

    2017-01-01

    Naama, Mbarek. Sub-Saharan Irregular Migrants in Morocco and the Exceptional Regularisation Programme in 2014. Diak South Helsinki Finland. Spring 2017. 43 Pages. Language: English. Diaconia University of Applied Sciences, Degree Programme in Social Services, Option in Community Development, Bachelor of Social Services (UAS). After decades of ignoring irregular migration and refusing to acknowledge that Morocco is no longer just a transit but also a host country for increasing complex mig...

  13. Gait Characteristics Associated with Trip-Induced Falls on Level and Sloped Irregular Surfaces

    Directory of Open Access Journals (Sweden)

    Andrew Merryweather

    2011-11-01

    Full Text Available Same level falls continue to contribute to an alarming number of slip/trip/fall injuries in the mining workforce. The objective of this study was to investigate how walking on different surface types and transverse slopes influences gait parameters that may be associated with a trip event. Gait analysis was performed for ten subjects on two orientations (level and sloped on smooth, hard surface (control and irregular (gravel, larger rocks surfaces. Walking on irregular surfaces significantly increased toe clearance compared to walking on the smooth surface. There was a significant (p < 0.05 decrease in cadence (steps/min, stride length (m, and speed (m/s from control to gravel to larger rocks. Significant changes in external rotation and increased knee flexion while walking on irregular surfaces were observed. Toe and heel clearance requirements increased on irregular surfaces, which may provide an explanation for trip-induced falls; however, the gait alterations observed in the experienced workers used as subjects would likely improve stability and recovery from a trip.

  14. Geomagnetic storm effects on the occurrences of ionospheric irregularities over the African equatorial/low-latitude region

    Science.gov (United States)

    Amaechi, P. O.; Oyeyemi, E. O.; Akala, A. O.

    2018-04-01

    The study investigated the effects of intense geomagnetic storms of 2015 on the occurrences of large scale ionospheric irregularities over the African equatorial/low-latitude region. Four major/intense geomagnetic storms of 2015 were analyzed for this study. These storms occurred on 17th March 2015 (-229 nT), 22nd June 2015 (-204 nT), 7th October 2015 (-124 nT), and 20th December 2015 (-170 nT). Total Electron Content (TEC) data obtained from five African Global Navigation Satellite Systems (GNSS) stations, grouped into eastern and western sectors were used to derive the ionospheric irregularities proxy indices, e.g., rate of change of TEC (ROT), ROT index (ROTI) and ROTI daily average (ROTIAVE). These indices were characterized alongside with the disturbance storm time (Dst), the Y component of the Interplanetary Electric Field (IEFy), polar cap (PC) index and the H component of the Earth's magnetic field from ground-based magnetometers. Irregularities manifested in the form of fluctuations in TEC. Prompt penetration of electric field (PPEF) and disturbance dynamo electric field (DDEF) modulated the behaviour of irregularities during the main and recovery phases of the geomagnetic storms. The effect of electric field over both sectors depends on the local time of southward turning of IMF Bz. Consequently, westward electric field inhibited irregularities during the main phase of March and October 2015 geomagnetic storms, while for the June 2015 storm, eastward electric field triggered weak irregularities over the eastern sector. The effect of electric field on irregularities during December 2015 storm was insignificant. During the recovery phase of the storms, westward DDEF suppressed irregularities.

  15. Irregular Warfare: Special Operations Joint Professional Military Education Transformation

    National Research Council Canada - National Science Library

    Cannady, Bryan H

    2008-01-01

    ... on today's battlefront in Afghanistan and Iraq and in the Global War on Terrorism (GWOT). At the forefront of the GWOT and irregular warfare are the United States Special Operations Command (USSOCOM...

  16. The regularized monotonicity method: detecting irregular indefinite inclusions

    DEFF Research Database (Denmark)

    Garde, Henrik; Staboulis, Stratos

    2018-01-01

    inclusions, where the conductivity distribution has both more and less conductive parts relative to the background conductivity; one such method is the monotonicity method of Harrach, Seo, and Ullrich. We formulate the method for irregular indefinite inclusions, meaning that we make no regularity assumptions...

  17. Convection-diffusion lattice Boltzmann scheme for irregular lattices

    NARCIS (Netherlands)

    Sman, van der R.G.M.; Ernst, M.H.

    2000-01-01

    In this paper, a lattice Boltzmann (LB) scheme for convection diffusion on irregular lattices is presented, which is free of any interpolation or coarse graining step. The scheme is derived using the axioma that the velocity moments of the equilibrium distribution equal those of the

  18. Simultaneous radar and spaced receiver VHF scintillation observations of ESF irregularities

    Directory of Open Access Journals (Sweden)

    D. Tiwari

    2006-07-01

    Full Text Available Simultaneous observations of equatorial spread F (ESF irregularities made on 10 nights during March-April 1998 and 1999, using an 18-MHz radar at Trivandrum (77° E, 8.5° N, dip 0.5° N and two spaced receivers recording scintillations on a 251-MHz signal at Tirunelveli (77.8° E, 8.7° N, dip 0.4° N, have been used to study the evolution of Equatorial Spread F (ESF irregularities. Case studies have been carried out on the day-to-day variability in ESF structure and dynamics, as observed by 18-MHz radar, and with spaced receiver measurements of average zonal drift Vo of the 251-MHz radio wave diffraction pattern on the ground, random velocity Vc, which is a measure of random changes in the characteristics of scintillation-producing irregularities, and maximum cross-correlation CI of the spaced receivers signals. Results show that in the initial phase of plasma bubble development, the greater the maximum height of ESF irregularities responsible for the radar backscatter, the greater the decorrelation is of the spaced receiver scintillation signals, indicating greater turbulence. The relationship of the maximum spectral width derived from the radar observations and CI also supports this result.

  19. The nature and source of irregular discharges to stormwater entering Sydney estuary, Australia

    International Nuclear Information System (INIS)

    Beck, H.J.; Birch, G.F.

    2014-01-01

    Irregular discharges of polluted stormwater into drainage systems during base flow (no rainfall) result in acute ecological impacts within fluvial and estuarine environments. In this study, metal and TSS concentrations were significantly more variable during business hours of weekdays (i.e. high-business activity) than weekends/public holidays (i.e. low-business activity) within three highly-urbanised catchments of Sydney estuary (Australia), as determined by analysing multivariate dispersion (PERMDISP). Concentrations of TSS and all metals analysed (Al, Ca, Cu, Fe, Mg, Pb and Zn) were also significantly greater during high- than low-business periods within at least one of the three catchments. In no case were concentrations significantly higher during low- than high-business periods. This pattern of contamination supports the hypothesis that commercial and industrial sources are major contributors of irregular discharges of contamination to Sydney estuary. Irregular discharges and consequential ecological impacts may be effectively reduced in this environment by focussing management efforts on these activities. - Highlights: • Irregular discharges of pollution have acute impacts on aquatic ecosystems. • These discharges were thought to be made during low activity periods, such as night. • Pollution was more concentrated and erratic during high- than low-business periods. • Timing of pollution suggests commercial and industrial activities are major sources. • Discharges effectively reduced by managing commercial and industrial activities. - Previously unreported irregular, illegal discharges of polluted stormwater released to estuaries result in acute ecological impacts and are potentially related to commercial/industrial activities

  20. 16QAM transmission with 5.2 bits/s/Hz spectral efficiency over transoceanic distance.

    Science.gov (United States)

    Zhang, H; Cai, J-X; Batshon, H G; Davidson, C R; Sun, Y; Mazurczyk, M; Foursa, D G; Pilipetskii, A; Mohs, G; Bergano, Neal S

    2012-05-21

    We transmit 160 x 100 G PDM RZ 16 QAM channels with 5.2 bits/s/Hz spectral efficiency over 6,860 km. There are more than 3 billion 16 QAM symbols, i.e., 12 billion bits, processed in total. Using coded modulation and iterative decoding between a MAP decoder and an LDPC based FEC all channels are decoded with no remaining errors.

  1. Characterization of the low latitude plasma density irregularities observed using C/NOFS and SCINDA data

    Science.gov (United States)

    Andima, Geoffrey; Amabayo, Emirant B.; Jurua, Edward; Cilliers, Pierre J.

    2018-01-01

    Complex electrodynamic processes over the low latitude region often result in post sunset plasma density irregularities which degrade satellite communication and navigation. In order to forecast the density irregularities, their occurrence time, duration and location need to be quantified. Data from the Communication/Navigation Outage Forecasting System (C/NOFS) satellite was used to characterize the low latitude ion density irregularities from 2011 to 2013. This was supported by ground based data from the SCIntillation Network Decision Aid (SCINDA) receivers at Makerere (Geographic coordinate 32.6°E, 0.3°N, and dip latitude -9.3°N) and Nairobi (Geographic coordinate 36.8°E, -1.3°N, and dip latitude -10.8°N). The results show that irregularities in ion density have a daily pattern with peaks from 20:00 to 24:00 Local Time (LT). Scintillation activity at L band and VHF over East Africa peaked in 2011 and 2012 from 20:00 to 24:00 LT, though in many cases scintillation at VHF persisted longer than that at L band. A longitudinal pattern in ion density irregularity occurrence was observed with peaks over 135-180°E and 270-300°E. The likelihood of ion density irregularity occurrence decreased with increasing altitude. Analysis of C/NOFS zonal ion drift velocities showed that the largest nighttime and daytime drifts were in 270-300°E and 300-330°E longitude regions respectively. Zonal irregularity drift velocities over East Africa were for the first time estimated from L-band scintillation indices. The results show that the velocity of plasma density irregularities in 2011 and 2012 varied daily, and hourly in the range of 50-150 m s-1. The zonal drift velocity estimates from the L-band scintillation indices had good positive correlation with the zonal drift velocities derived from VHF receivers by the spaced receiver technique.

  2. Star formation histories of irregular galaxies

    International Nuclear Information System (INIS)

    Gallagher, J.S. III; Hunter, D.A.; Tutukov, A.V.

    1984-01-01

    We explore the star formation histories of a selection of irregular and spiral galaxies by using three parameters that sample the star formation rate (SFR) at different epochs: (1) the mass of a galaxy in the form of stars measures the SFR integrated over a galaxy's lifetime; (2) the blue luminosity is dominated primarily by stars formed over the past few billion years; and (3) Lyman continuum photon fluxes derived from Hα luminosities give the current ( 8 yr) SFR

  3. Bariatric Surgery Improves Hyperandrogenism, Menstrual Irregularities, and Metabolic Dysfunction Among Women with Polycystic Ovary Syndrome (PCOS).

    Science.gov (United States)

    Christ, Jacob P; Falcone, Tommaso

    2018-03-02

    To characterize the impact of bariatric surgery on reproductive and metabolic features common to polycystic ovary syndrome (PCOS) and to assess the relevance of preoperative evaluations in predicting likelihood of benefit from surgery. A retrospective chart review of records from 930 women who had undergone bariatric surgery at the Cleveland Clinic Foundation from 2009 to 2014 was completed. Cases of PCOS were identified from ICD coding and healthy women with pelvic ultrasound evaluations were identified using Healthcare Common Procedure Coding System coding. Pre- and postoperative anthropometric evaluations, menstrual cyclicity, ovarian volume (OV) as well as markers of hyperandrogenism, dyslipidemia, and dysglycemia were evaluated. Forty-four women with PCOS and 65 controls were evaluated. Both PCOS and non-PCOS had significant reductions in body mass index (BMI) and markers of dyslipidemia postoperatively (p PCOS had significant reductions in androgen levels (p irregular menses (p PCOS, independent of preoperative BMI and age, preoperative OV associated with change in hemoglobin A1c (β 95% (confidence interval) 0.202 (0.011-0.393), p = 0.04) and change in triglycerides (6.681 (1.028-12.334), p = 0.03), and preoperative free testosterone associated with change in total cholesterol (3.744 (0.906-6.583), p = 0.02) and change in non-HDL-C (3.125 (0.453-5.796), p = 0.03). Bariatric surgery improves key diagnostic features seen in women with PCOS and ovarian volume, and free testosterone may have utility in predicting likelihood of metabolic benefit from surgery.

  4. Migración y violencia en Cancún: estudio de dos asentamientos irregulares; Migração e violencia em Cancún: estudo de dois assentamentos irregulares; Migration and violence in Cancún: study of two irregular settlements

    Directory of Open Access Journals (Sweden)

    Minerva Alavez San Pedro

    2017-07-01

    Full Text Available Resumen: El presente texto ofrece una aproximación a las condiciones actuales de violencia en dos asentamientos irregulares de Cancún, que se ubica en el estado de Quintana Roo, México, como consecuencia de la migración laboral que ha experimentado esta localidad desde su creación en 1970 como Centro Turístico Integral. Para elaborar este trabajo se realizó, además de la revisión bibliográfica, un seguimiento hemerográfico de medios de comunicación locales y nacionales y un trabajo etnográfico en las colonias El Milagro y Tres Reyes de abril de 2016 a junio de 2017. De manera general se puede adelantar que las presentes manifestaciones de violencia en Cancún y el continuo surgimiento de asentamientos irregulares ponen en evidencia la incapacidad institucional del Estado para planear y atender las necesidades de seguridad y vivienda de la población. Palabras clave: Cancún, economía, migración, violencia, asentamientos.   Resumo: O presente texto ofrece uma aproximação das condiçõe atuais de violencia em dois assentamentos irregulares em Cancún, que se localiza no estado de Quintana Roo, México, como consequência da migração de trabalhadores que aconteceu nesta localidade desde sua criação em 1970 como Centro Turístico Integral. Para elaborar este trabalho se realizou, ademáis da revisão bibliográfica, um estudo hemerográfico dos meios de comunicação locais e nacionais e um trabalho etnográfico nas colônias de El Milagro e Tres Reyes de abril de 2016 a junho de 2017. De maneira geral se pode antecipar que as presentes manifestações de violencia em Cancún e o contínuo surgimento de assentamentos irregulares colocam em evidencia a capacidade institucional do Estado para planejar e atender as necesidades de segurança e vivência da população. Palavras-chaves: Cancún; economía; migração; violencia; assentamentos.   Abstract: The present text offers an approximation to the current conditions of

  5. Dynamics of long-period irregular pulsations in high latitudes during strong magnetic storms

    International Nuclear Information System (INIS)

    Kurazhkovskaya, N.A.; Klajn, B.I.

    1995-01-01

    Effects of strong magnetic storms within np type high-latitudinal long-period irregular pulsations at Mirny studied using data obtained at observatory of the magnetosphere south hemisphere. Variation of long-period irregular pulsation amplitude is shown to depend essentially on duration of storm initial phase and on the nature of solar wind heterogeneity enabling growth of strong storm. 14 refs

  6. Refraction traveltime tomography based on damped wave equation for irregular topographic model

    Science.gov (United States)

    Park, Yunhui; Pyun, Sukjoon

    2018-03-01

    Land seismic data generally have time-static issues due to irregular topography and weathered layers at shallow depths. Unless the time static is handled appropriately, interpretation of the subsurface structures can be easily distorted. Therefore, static corrections are commonly applied to land seismic data. The near-surface velocity, which is required for static corrections, can be inferred from first-arrival traveltime tomography, which must consider the irregular topography, as the land seismic data are generally obtained in irregular topography. This paper proposes a refraction traveltime tomography technique that is applicable to an irregular topographic model. This technique uses unstructured meshes to express an irregular topography, and traveltimes calculated from the frequency-domain damped wavefields using the finite element method. The diagonal elements of the approximate Hessian matrix were adopted for preconditioning, and the principle of reciprocity was introduced to efficiently calculate the Fréchet derivative. We also included regularization to resolve the ill-posed inverse problem, and used the nonlinear conjugate gradient method to solve the inverse problem. As the damped wavefields were used, there were no issues associated with artificial reflections caused by unstructured meshes. In addition, the shadow zone problem could be circumvented because this method is based on the exact wave equation, which does not require a high-frequency assumption. Furthermore, the proposed method was both robust to an initial velocity model and efficient compared to full wavefield inversions. Through synthetic and field data examples, our method was shown to successfully reconstruct shallow velocity structures. To verify our method, static corrections were roughly applied to the field data using the estimated near-surface velocity. By comparing common shot gathers and stack sections with and without static corrections, we confirmed that the proposed tomography

  7. Evaluation of statistical methods for quantifying fractal scaling in water-quality time series with irregular sampling

    Science.gov (United States)

    Zhang, Qian; Harman, Ciaran J.; Kirchner, James W.

    2018-02-01

    River water-quality time series often exhibit fractal scaling, which here refers to autocorrelation that decays as a power law over some range of scales. Fractal scaling presents challenges to the identification of deterministic trends because (1) fractal scaling has the potential to lead to false inference about the statistical significance of trends and (2) the abundance of irregularly spaced data in water-quality monitoring networks complicates efforts to quantify fractal scaling. Traditional methods for estimating fractal scaling - in the form of spectral slope (β) or other equivalent scaling parameters (e.g., Hurst exponent) - are generally inapplicable to irregularly sampled data. Here we consider two types of estimation approaches for irregularly sampled data and evaluate their performance using synthetic time series. These time series were generated such that (1) they exhibit a wide range of prescribed fractal scaling behaviors, ranging from white noise (β = 0) to Brown noise (β = 2) and (2) their sampling gap intervals mimic the sampling irregularity (as quantified by both the skewness and mean of gap-interval lengths) in real water-quality data. The results suggest that none of the existing methods fully account for the effects of sampling irregularity on β estimation. First, the results illustrate the danger of using interpolation for gap filling when examining autocorrelation, as the interpolation methods consistently underestimate or overestimate β under a wide range of prescribed β values and gap distributions. Second, the widely used Lomb-Scargle spectral method also consistently underestimates β. A previously published modified form, using only the lowest 5 % of the frequencies for spectral slope estimation, has very poor precision, although the overall bias is small. Third, a recent wavelet-based method, coupled with an aliasing filter, generally has the smallest bias and root-mean-squared error among all methods for a wide range of

  8. Justice: A Problem for Military Ethics during Irregular War

    National Research Council Canada - National Science Library

    Bauer, John W

    2008-01-01

    ... is?" or "Justice according to whom?" The relative nature of the term "justice" creates a problem for military ethics, particularly when soldiers try to determine what actions are morally acceptable when they are engaged in irregular warfare...

  9. A Generalized FDM for solving the Poisson's Equation on 3D Irregular Domains

    Directory of Open Access Journals (Sweden)

    J. Izadian

    2014-01-01

    Full Text Available In this paper a new method for solving the Poisson's equation with Dirichlet conditions on irregular domains is presented. For this purpose a generalized finite differences method is applied for numerical differentiation on irregular meshes. Three examples on cylindrical and spherical domains are considered. The numerical results are compared with analytical solution. These results show the performance and efficiency of the proposed method.

  10. Natural convection inside an irregular porous cavity

    International Nuclear Information System (INIS)

    Beltran, Jorge I. LLagostera; Trevisan, Osvair Vidal

    1990-01-01

    Natural convection flow induced by heating from below in a irregular porous cavity is investigated numerically. The influence of the modified Rayleigh number and geometric ratios on heat transfer and fluid flow is studied. Global and local Nusselt for Rayleigh numbers covering the range 0 - 1600 and for several geometric ratios. The fluid flow and the temperature field are illustrated by contour maps. (author)

  11. Swiveling Lathe Jaw Concept for Holding Irregular Pieces

    Science.gov (United States)

    David, J.

    1966-01-01

    Clamp holds irregularly shaped pieces in lathe chuck without damage and eliminates excessive time in selecting optimum mounting. Interchangeable jaws ride in standard jaw slots but swivel so that the jaw face bears evenly against the workpiece regardless of contour. The jaws can be used on both engine and turret lathes.

  12. The scholarly rebellion of the early Baker Street Irregulars

    Directory of Open Access Journals (Sweden)

    George Mills

    2017-03-01

    Full Text Available This work provides and analyzes an early institutional history of the pioneering Sherlock Holmes American fan club, the Baker Street Irregulars (BSI. Using the publications and records of these devoted Sherlockians, I track the BSI's development from a speakeasy gathering in 1934 to a national organization by the mid-1940s. This growth was built on a foundation of Victorian nostalgia and playful humor. Yet at the same time the members of the Irregulars took their fandom seriously, producing Sherlockian scholarship and creating an infrastructure of journals, conferences, and credentialing that directly mimicked the academy. They positioned themselves in contrast to prevailing scholarly practices of the period, such as New Criticism. I trace both how their fan practices developed over time and how this conflict with the academy led to many of the BSI's defining characteristics.

  13. Rocket measurements of electron density irregularities during MAC/SINE

    Science.gov (United States)

    Ulwick, J. C.

    1989-01-01

    Four Super Arcas rockets were launched at the Andoya Rocket Range, Norway, as part of the MAC/SINE campaign to measure electron density irregularities with high spatial resolution in the cold summer polar mesosphere. They were launched as part of two salvos: the turbulent/gravity wave salvo (3 rockets) and the EISCAT/SOUSY radar salvo (one rocket). In both salvos meteorological rockets, measuring temperature and winds, were also launched and the SOUSY radar, located near the launch site, measured mesospheric turbulence. Electron density irregularities and strong gradients were measured by the rocket probes in the region of most intense backscatter observed by the radar. The electron density profiles (8 to 4 on ascent and 4 on descent) show very different characteristics in the peak scattering region and show marked spatial and temporal variability. These data are intercompared and discussed.

  14. The real code of leonardo da vinci.

    Science.gov (United States)

    Ose, Leiv

    2008-02-01

    Leonardo da Vinci was born in Italy. Among the researchers and scientists, he is favourably known for his remarkable efforts in scientific work. His investigations of atherosclerosis judiciously combine three separate fields of research. In 1506, he finished his masterpiece, painting of Mona Lisa. A careful clinical examination of the famous painting reveals a yellow irregular leather-like spot at the inner end of the left upper eyelid and a soft bumpy well-defined swelling of the dorsum of the right hand beneath the index finger about 3 cm long. This is probably the first case of familial hypercholesterolemia (FH). The FH code of Leonardo da Vinci was given immense consideration by scientists like Carl Muller, who described the xanthomas tuberosum and angina pectoris. On the contrary, Akira Endo searched for microbial metabolites that would inhibit HMG-CoA reductase, the rate-limiting enzyme in the synthesis of cholesterol and finally, Michael Brown and Joseph Goldstein published a remarkable series of elegant and insightful papers in the 70s and 80s. They established that the cellular uptake of low-density lipoprotein (LDL) essentially requires the LDL receptor. this was the real Code of Leonardo da Vinci.

  15. Spiking irregularity and frequency modulate the behavioral report of single-neuron stimulation.

    Science.gov (United States)

    Doron, Guy; von Heimendahl, Moritz; Schlattmann, Peter; Houweling, Arthur R; Brecht, Michael

    2014-02-05

    The action potential activity of single cortical neurons can evoke measurable sensory effects, but it is not known how spiking parameters and neuronal subtypes affect the evoked sensations. Here, we examined the effects of spike train irregularity, spike frequency, and spike number on the detectability of single-neuron stimulation in rat somatosensory cortex. For regular-spiking, putative excitatory neurons, detectability increased with spike train irregularity and decreasing spike frequencies but was not affected by spike number. Stimulation of single, fast-spiking, putative inhibitory neurons led to a larger sensory effect compared to regular-spiking neurons, and the effect size depended only on spike irregularity. An ideal-observer analysis suggests that, under our experimental conditions, rats were using integration windows of a few hundred milliseconds or more. Our data imply that the behaving animal is sensitive to single neurons' spikes and even to their temporal patterning. Copyright © 2014 Elsevier Inc. All rights reserved.

  16. Zonal drifts of ionospheric irregularities at temperate latitude in the Indian region

    Directory of Open Access Journals (Sweden)

    S. Kumar

    1995-07-01

    Full Text Available The systematic time differences observed in the onset of postsunset VHF scintillations recorded simultaneously at Ujjain (Geogr. lat. 23.2°N, Geogr. long. 75.6°E and Bhopal (Geogr. lat. 23.2°N, Geogr. long. 77.6°E, situated at the peak of the anomaly crest in the Indian region, have been analysed to determine the zonal drifts of scintillation-producing irregularities. The method is based on the assumption that the horizontal movement of irregularities does not change while crossing the F-region cross-over points of these stations. The calculated velocities of irregularities indicate an eastward drift decreasing from about 180 m s–1 to 55 m s–1 during the course of night. In the premidnight period, the drifts are reduced under the magnetically disturbed conditions. The average east-west extension of irregularites is found to be in the range of 200–500 km.

  17. Institutional Root Cause of the Irregularity of Social Security Funds in China

    Institute of Scientific and Technical Information of China (English)

    Bingwen Zheng

    2007-01-01

    This paper analyzes institutional defects in the Chinese social security system, based on irregularities in social security funds revealed in the Audit Report by the China National Audit Office. The author divides the irregularities into five categories according to the nature of fund use. The results show that the institutional root cause of the irregularities lies in the unreasonable design and operation of the social security system, which currently faces management and institutional risks. This paper argues that simple rhetoric about strengthening regulation and supervision cannot help to reduce illegal practices, or to realize risk control. The only solution is to reform the social security system. Specifically, the Chinese Government should regulate the administrative cost of the social security system,and the behavior of its agencies, through legislation, reform the investment regime to increase rate of return of pension assets, and adjust and reshape the existing social security system, so as to elevate its pooling level.

  18. Characteristics of low latitude ionospheric E-region irregularities ...

    Indian Academy of Sciences (India)

    154°E, dip angle = 37.3°, sub-ionospheric dip = 34°) have been analyzed to study the behaviour of ionospheric E-region irregularities during the active solar and magnetic periods. The autocorrelation functions, power spectral densities, signal de-correlation times are computed to study the temporal features of ionospheric ...

  19. Randomness control of vehicular motion through a sequence of traffic signals at irregular intervals

    International Nuclear Information System (INIS)

    Nagatani, Takashi

    2010-01-01

    We study the regularization of irregular motion of a vehicle moving through the sequence of traffic signals with a disordered configuration. Each traffic signal is controlled by both cycle time and phase shift. The cycle time is the same for all signals, while the phase shift varies from signal to signal by synchronizing with intervals between a signal and the next signal. The nonlinear dynamic model of the vehicular motion is presented by the stochastic nonlinear map. The vehicle exhibits the very complex behavior with varying both cycle time and strength of irregular intervals. The irregular motion induced by the disordered configuration is regularized by adjusting the phase shift within the regularization regions.

  20. Shape of the nuclear magnetic resonance line in anisotropic superconductors with an irregular vortex lattice

    International Nuclear Information System (INIS)

    Minkin, A.V.; Tsarevskij, S.L.

    2006-01-01

    For high-temperature superconductors the shape of a NMR spectrum line is built regarding for variation of inhomogeneity of irregular vortex lattice magnetic field near superconductor surface. It is shown that the shape of a NMR line is not simply widened but noticeably varies depending on the degree of irregularity of a superconductor vortex lattice. This variation is associated with a local symmetry decrease in an irregular vortex lattice of the superconductor. Taking into account these circumstances may considerably change conclusions about the type of a vortex lattice and superconductor parameters which are commonly gained from NMR line shape analysis [ru

  1. Improving Transactional Memory Performance for Irregular Applications

    OpenAIRE

    Pedrero, Manuel; Gutiérrez, Eladio; Romero, Sergio; Plata, Óscar

    2015-01-01

    Transactional memory (TM) offers optimistic concurrency support in modern multicore archi- tectures, helping the programmers to extract parallelism in irregular applications when data dependence information is not available before runtime. In fact, recent research focus on ex- ploiting thread-level parallelism using TM approaches. However, the proposed techniques are of general use, valid for any type of application. This work presents ReduxSTM, a software TM system specially d...

  2. Morning Cortisol Levels and Perceived Stress in Irregular Shift Workers Compared with Regular Daytime Workers

    Directory of Open Access Journals (Sweden)

    Harri Lindholm

    2012-01-01

    Full Text Available The 24/7 work environment and irregular shifts may markedly enhance the psychological pressure of media work. Changes in the hypothalamic-pituitary-adrenal axis reflect adaptation to stress. We analysed the correlation between subjective stress, sleep, salivary cortisol, and melatonin hormones among Finnish media workers with regular daytime work (RDW and with irregular shift work (ISW while controlling confounders. From 874 employees with regular daytime work or with irregular shift work, 70 employees from both groups were randomly selected. The final number of employees with a complete salivary cortisol profile was 66 in the RDW group and 65 in the ISW group. Five saliva samples were gathered from each subject before and during a working day. The salivary cortisol level of the sample taken 60 minutes after awakening (T1 was compared to the salivary cortisol level taken immediately after awakening (T0, T1/T0 ratio. The ratio was higher in the ISW group than in RDW group. Irregular shift work (P<0.001, severe stress (P<0.05, and less sleep (P<0.05 were independently associated with an augmented cortisol response after awakening. A stressful work environment and irregular shift work enhance cortisol excretion after waking. In the long run, this may become detrimental to health.

  3. Deep-space and near-Earth optical communications by coded orbital angular momentum (OAM) modulation.

    Science.gov (United States)

    Djordjevic, Ivan B

    2011-07-18

    In order to achieve multi-gigabit transmission (projected for 2020) for the use in interplanetary communications, the usage of large number of time slots in pulse-position modulation (PPM), typically used in deep-space applications, is needed, which imposes stringent requirements on system design and implementation. As an alternative satisfying high-bandwidth demands of future interplanetary communications, while keeping the system cost and power consumption reasonably low, in this paper, we describe the use of orbital angular momentum (OAM) as an additional degree of freedom. The OAM is associated with azimuthal phase of the complex electric field. Because OAM eigenstates are orthogonal the can be used as basis functions for N-dimensional signaling. The OAM modulation and multiplexing can, therefore, be used, in combination with other degrees of freedom, to solve the high-bandwidth requirements of future deep-space and near-Earth optical communications. The main challenge for OAM deep-space communication represents the link between a spacecraft probe and the Earth station because in the presence of atmospheric turbulence the orthogonality between OAM states is no longer preserved. We will show that in combination with LDPC codes, the OAM-based modulation schemes can operate even under strong atmospheric turbulence regime. In addition, the spectral efficiency of proposed scheme is N2/log2N times better than that of PPM.

  4. Evaluation of statistical methods for quantifying fractal scaling in water-quality time series with irregular sampling

    Directory of Open Access Journals (Sweden)

    Q. Zhang

    2018-02-01

    Full Text Available River water-quality time series often exhibit fractal scaling, which here refers to autocorrelation that decays as a power law over some range of scales. Fractal scaling presents challenges to the identification of deterministic trends because (1 fractal scaling has the potential to lead to false inference about the statistical significance of trends and (2 the abundance of irregularly spaced data in water-quality monitoring networks complicates efforts to quantify fractal scaling. Traditional methods for estimating fractal scaling – in the form of spectral slope (β or other equivalent scaling parameters (e.g., Hurst exponent – are generally inapplicable to irregularly sampled data. Here we consider two types of estimation approaches for irregularly sampled data and evaluate their performance using synthetic time series. These time series were generated such that (1 they exhibit a wide range of prescribed fractal scaling behaviors, ranging from white noise (β  =  0 to Brown noise (β  =  2 and (2 their sampling gap intervals mimic the sampling irregularity (as quantified by both the skewness and mean of gap-interval lengths in real water-quality data. The results suggest that none of the existing methods fully account for the effects of sampling irregularity on β estimation. First, the results illustrate the danger of using interpolation for gap filling when examining autocorrelation, as the interpolation methods consistently underestimate or overestimate β under a wide range of prescribed β values and gap distributions. Second, the widely used Lomb–Scargle spectral method also consistently underestimates β. A previously published modified form, using only the lowest 5 % of the frequencies for spectral slope estimation, has very poor precision, although the overall bias is small. Third, a recent wavelet-based method, coupled with an aliasing filter, generally has the smallest bias and root-mean-squared error among

  5. Modalities of Tonometry and their Accuracy with Respect to Corneal Thickness and Irregularities

    Science.gov (United States)

    De Moraes, Carlos Gustavo V.; Prata, Tiago S.; Liebmann, Jeffrey; Ritch, Robert

    2010-01-01

    Recent advances in tonometry have led to the development of a number of devices with differing clinical applications. Their role in cases of abnormal corneal thickness and surface irregularities is particularly important, as inaccurate estimation of the true intraocular pressure (IOP) in such cases may lead to suboptimal evaluation. The purpose of the present review was to evaluate the accuracy of the most widely used devices in cases of corneal thickness and surface irregularities, based on a survey of the published literature. The analysis was based on a Medline search focusing mainly on papers that have evaluated the devices’ accuracy with respect to corneal thickness and irregularities. Nine device types (Goldmann tonometer, Tono-Pen, Perkins tonometer, Ocular Response Analyzer, non-contact tonometer, pneumatonometer, I-Care rebound tonometer, Pascal dynamic contour tonometer (DCT) and Phosphene tonometer) were described in detail. Moreover, the physical principles and major utility of each tonometer were compared. Each of the many different commercially available tonometers has specific advantages and disadvantages. New non-invasive technologies are getting closer to a precise estimation of the true IOP. However, among all tonometers, none is highly accurate when both corneal thickness and surface irregularities are present. Fifty years after its development, Goldmann tonometry remains the gold standard to which all other devices are compared.

  6. Using Radio Irregularity for Increasing Residential Energy Awareness

    Directory of Open Access Journals (Sweden)

    A. Miljković

    2012-06-01

    Full Text Available Radio irregularity phenomenon is often considered as a shortcoming of wireless networks. In this paper, the method of using radio irregularity as an efficient human presence detection sensor in smart homes is presented. The method is mainly based on monitoring variations of the received signal strength indicator (RSSI within the messages used for the communication between wireless smart power outlets. The radio signals used for the inter-outlets communication can be absorbed, diffracted or reflected by objects in their propagation paths. When a human enters the existing radio communication field, the variation of the signal strength at the receiver is even more expressed. Based on the detected changes and compared to the initial thresholds set during the initialization phase, the system detects human presence. The proposed solution increases user awareness and automates the power control in households, with the primary goal to contribute in residential energy savings. Compared to conventional sensor networks, this approach preserves the sensorial intelligence, simplicity and low installation costs, without the need for additional sensors integration.

  7. Very high latitude F-region irregularities observed by HF-radar backscatter

    International Nuclear Information System (INIS)

    Baker, K.B.; Greenwald, R.A.; Tsunoda, R.T.

    1983-01-01

    In February and March, 1982, a coherent scatter HF radar was operated from Cleary, Alaska to observe 7- to 15-m wavelength F-region plasma irregularities near the poleward edge of the auroral zone and in the polar cap. The radar operated for five days from February 25 to March 1 and produced approximately 700,000 Doppler spectra during that time. Of those nearly 700,000 spectra, approximately 10% showed backscattered power 3 dB or more above the noise level. A ray tracing technique using electron densities determined by the Chatanika incoherent scatter radar was used to predict locations where the HF waves were approximately normal to the magnetic field. If those locations were also to contain small scale electron density structure, then one would expect them to backscatter the HF waves. Several comparisons were made between predicted and observed locations of radiowave backscatter and excellent agreement was obtained. In addition, comparisons of the Doppler velocities observed by the coherent scatter HF radar and those observed by the Chatanika radar showed good agreement, suggesting that the plasma irregularities observed by the HF radar drift with the ambient plasma. In addition, average vector velocities calculated for the entire 5-day period show a flow pattern consistent with polar cap convection models. This again indicates that the irregularities drift with the plasma, as is predicted by a number of theories of F-region plasma irregularities. In the summer of 1983, the research program begun with those measurements will be continued with a steerable phased-array HF radar located at Goose Bay, Labrador, that will view the same ionospheric region as does the Sondre Stromfjord incoherent scatter radar

  8. Interagency Cooperation for Irregular Warfare at the Combatant Command

    Science.gov (United States)

    2009-01-01

    enemy’s command capability.16 Salamoni argued that the term “irregular warfare” belies an ethnocentric perspective of conflict that will limit military...duty military staffing to form the nucleus of the organization, which would receive augmentation from additional assigned reservists and interagency

  9. Simultaneous observations of ESF irregularities over Indian region using radar and GPS

    Directory of Open Access Journals (Sweden)

    S. Sripathi

    2008-10-01

    Full Text Available In this paper, we present simultaneous observations of temporal and spatial variability of total electron content (TEC and GPS amplitude scintillations on L1 frequency (1.575 GHz during the time of equatorial spread F (ESF while the MST radar (53 MHz located at Gadanki (13.5° N, 79.2° E, Dip latitude 6.3° N, a low latitude station, made simultaneous observations. In particular, the latitudinal and longitudinal extent of TEC and L-band scintillations was studied in the Indian region for different types of ESF structures observed using the MST radar during the low solar activity period of 2004 and 2005. Simultaneous radar and GPS observations during severe ESF events in the pre-midnight hour reveal that significant GPS L band scintillations, depletions in TEC, and the double derivative of the TEC index (DROTI, which is a measure of fluctuations in TEC, obtained at low latitudes coincide with the appearance of radar echoes at Gadanki. As expected, when the irregularities reach higher altitudes as seen in the radar map during pre-midnight periods, strong scintillations on an L-band signal are observed at higher latitudes. Conversely, when radar echoes are confined to only lower altitudes, weak scintillations are found and their latitudinal extent is small. During magnetically quiet periods, we have recorded plume type radar echoes during a post-midnight period that is devoid of L-band scintillations. Using spectral slopes and cross-correlation index of the VHF scintillation observations, we suggest that these irregularities could be "dead" or "fossil" bubbles which are just drifting in from west. This scenario is consistent with the observations where suppression of pre-reversal enhancement (PRE in the eastward electric field is indicated by ionosonde observations of the height of equatorial F layer and also occurrence of low spectral width in the radar observations relative to pre-midnight period. However, absence of L-band scintillations during

  10. EXIT Chart Analysis of Binary Message-Passing Decoders

    DEFF Research Database (Denmark)

    Lechner, Gottfried; Pedersen, Troels; Kramer, Gerhard

    2007-01-01

    Binary message-passing decoders for LDPC codes are analyzed using EXIT charts. For the analysis, the variable node decoder performs all computations in the L-value domain. For the special case of a hard decision channel, this leads to the well know Gallager B algorithm, while the analysis can...... be extended to channels with larger output alphabets. By increasing the output alphabet from hard decisions to four symbols, a gain of more than 1.0 dB is achieved using optimized codes. For this code optimization, the mixing property of EXIT functions has to be modified to the case of binary message......-passing decoders....

  11. Lateralized effects of orthographical irregularity and auditory memory load on the kinematics of transciption typewriting

    NARCIS (Netherlands)

    Bloemsaat, J.G.; Galen, G.P. van; Meulenbroek, R.G.J.

    2003-01-01

    This study investigated the combined effects of orthographical irregularity and auditory memory load on the kinematics of finger movements in a transcription-typewriting task. Eight right-handed touch-typists were asked to type 80 strings of ten seven-letter words. In half the trials an irregularly

  12. Magnetic storm effect on the occurrence of ionospheric irregularities at an equatorial station in the African sector

    Directory of Open Access Journals (Sweden)

    Olushola Abel Oladipo

    2014-01-01

    Full Text Available Large-scale ionospheric irregularities usually measured by GPS TEC fluctuation indices are regular occurrence at the equatorial region shortly after sunset around solar maximum. Magnetic storm can trigger or inhibit the generation of these irregularities depending on the local time the main phase of a particular storm occurs. We studied the effect of nine (9 distinct storms on the occurrence of ionospheric irregularities at Fraceville in Gabon (Lat = −1.63˚, Long = 13.55˚, dip lat. = −15.94˚, an equatorial station in the African sector. These storms occurred between November 2001 and September 2002. We used TEC fluctuation indices (i.e. ROTI and ROTIAVE estimated from 30 s interval Rinex data and also we used the storm indices (i.e. Dst, dDst/dt, and IMF BZ to predict the likely effect of each storm on the irregularities occurrence at this station. The results obtained showed that most of the storms studied inhibited ionospheric irregularities. Only one out of all the storms studied (i.e. September 4, 2002 storms with the main phase on the night of September 7-8 triggered post-midnight ionospheric irregularities. There are two of the storms during which ionospheric irregularities were observed. However, these may not be solely attributed to the storms event because the level of irregularities observed during these two storms is comparable to that observed during previous days before the storms. For this station and for the storms investigated, it seems like a little modification to the use of Aarons categories in terms of the local time the maximum negative Dst occurs could lead to a better prediction. However, it would require investigating many storms during different level of solar activities and at different latitudes to generalize this modification.

  13. Coherent and intermittent ensemble oscillations emerge from networks of irregular spiking neurons.

    Science.gov (United States)

    Hoseini, Mahmood S; Wessel, Ralf

    2016-01-01

    Local field potential (LFP) recordings from spatially distant cortical circuits reveal episodes of coherent gamma oscillations that are intermittent, and of variable peak frequency and duration. Concurrently, single neuron spiking remains largely irregular and of low rate. The underlying potential mechanisms of this emergent network activity have long been debated. Here we reproduce such intermittent ensemble oscillations in a model network, consisting of excitatory and inhibitory model neurons with the characteristics of regular-spiking (RS) pyramidal neurons, and fast-spiking (FS) and low-threshold spiking (LTS) interneurons. We find that fluctuations in the external inputs trigger reciprocally connected and irregularly spiking RS and FS neurons in episodes of ensemble oscillations, which are terminated by the recruitment of the LTS population with concurrent accumulation of inhibitory conductance in both RS and FS neurons. The model qualitatively reproduces experimentally observed phase drift, oscillation episode duration distributions, variation in the peak frequency, and the concurrent irregular single-neuron spiking at low rate. Furthermore, consistent with previous experimental studies using optogenetic manipulation, periodic activation of FS, but not RS, model neurons causes enhancement of gamma oscillations. In addition, increasing the coupling between two model networks from low to high reveals a transition from independent intermittent oscillations to coherent intermittent oscillations. In conclusion, the model network suggests biologically plausible mechanisms for the generation of episodes of coherent intermittent ensemble oscillations with irregular spiking neurons in cortical circuits. Copyright © 2016 the American Physiological Society.

  14. Young Japanese college students with dysmenorrhea have high frequency of irregular menstruation and premenstrual symptoms.

    Science.gov (United States)

    Fujiwara, Tomoko; Nakata, Rieko

    2007-01-01

    In this study, to estimate the current status of young women with menstrual disorders, the relation among dysmenorrhea, irregular menstruation and premenstrual symptoms was investigated by a questionnaire. Subjects ranging from 18 to 20 years old were recruited from 522 female students at Ashiya College in Japan. The intensity of dysmenorrhea was classified into 3 grades (score 1, not requiring analgesic; score 2, painful, requiring analgesic; score 3, painful, not relieved by analgesic). All participants were further divided into subsequent groups as having premenstrual symptoms or not and those having regular or irregular menstruation. Dysmenorrhea scores in the students with premenstrual symptoms or irregular menstruation were significantly higher than those without these symptoms (1.66+/-0.66 vs 1.41+/-0.59; 1.62+/-0.68 vs 1.49+/-0.61, respectively). There was no significant relation in the incidence between premenstrual symptoms and irregular menstruation. These findings suggest that considerable numbers of young women with dysmenorrhea are associated with premenstrual symptoms.

  15. Constructing C1 Continuous Surface on Irregular Quad Meshes

    Institute of Scientific and Technical Information of China (English)

    HE Jun; GUO Qiang

    2013-01-01

    A new method is proposed for surface construction on irregular quad meshes as extensions to uniform B-spline surfaces. Given a number of control points, which form a regular or irregular quad mesh, a weight function is constructed for each control point. The weight function is defined on a local domain and is C1 continuous. Then the whole surface is constructed by the weighted combination of all the control points. The property of the new method is that the surface is defined by piecewise C1 bi-cubic rational parametric polynomial with each quad face. It is an extension to uniform B-spline surfaces in the sense that its definition is an analogy of the B-spline surface, and it produces a uniform bi-cubic B-spline surface if the control mesh is a regular quad mesh. Examples produced by the new method are also included.

  16. Modalities of Tonometry and their Accuracy with Respect to Corneal Thickness and Irregularities

    OpenAIRE

    Carlos Gustavo V. De Moraes; Tiago S. Prata; Jeffrey Liebmann; Robert Ritch

    2008-01-01

    Recent advances in tonometry have led to the development of a number of devices with differing clinical applications. Their role in cases of abnormal corneal thickness and surface irregularities is particularly important, as inaccurate estimation of the true intraocular pressure (IOP) in such cases may lead to suboptimal evaluation. The purpose of the present review was to evaluate the accuracy of the most widely used devices in cases of corneal thickness and surface irregularities, based ...

  17. Effects of irregular-shift work and physical activity on cardiovascular risk factors in truck drivers

    Directory of Open Access Journals (Sweden)

    Elaine Cristina Marqueze

    2013-06-01

    Full Text Available OBJECTIVE: To analyze the putative effect of type of shift and its interaction with leisure-time physical activity on cardiovascular risk factors in truck drivers. METHODS: A cross-sectional study was undertaken on 57 male truck drivers working at a transportation company, of whom 31 worked irregular shifts and 26 worked on the day-shift. Participants recorded their physical activity using the International Physical Activity Questionnaire along with measurements of blood pressure, body mass index and waist-hip ratio. Participants also provided a fasting blood sample for analysis of lipid-related outcomes. Data were analyzed using a factorial model which was covariate-controlled for age, smoking, work demand, control at work and social support. RESULTS: Most of the irregular-shift and day-shift workers worked more than 8 hours per day (67.7% and 73.1%, respectively. The mean duration of experience working the irregular schedule was 15.7 years. Day-shift workers had never engaged in irregular-shift work and had been working as a truck driver for 10.8 years on average. The irregular-shift drivers had lower work demand but less control compared to day-shift drivers (p < 0.05. Moderately-active irregular-shift workers had higher systolic and diastolic arterial pressures (143.7 and 93.2 mmHg, respectively than moderately-active day-shift workers (116 and 73.3 mmHg, respectively (p < 0.05 as well as higher total cholesterol concentrations (232.1 and 145 mg/dl, respectively (p = 0.01. Irrespective of their physical activity, irregular-shift drivers had higher total cholesterol and LDL-cholesterol concentrations (211.8 and 135.7 mg/dl, respectively than day-shift workers (161.9 and 96.7 mg/dl, respectively (ANCOVA, p < 0.05. CONCLUSIONS: Truck drivers are exposed to cardiovascular risk factors due to the characteristics of the job, such as high work demand, long working hours and time in this profession, regardless of shift type or leisure-time physical

  18. Detection of increase in corneal irregularity due to pterygium using Fourier series harmonic analyses with multiple diameters.

    Science.gov (United States)

    Minami, Keiichiro; Miyata, Kazunori; Otani, Atsushi; Tokunaga, Tadatoshi; Tokuda, Shouta; Amano, Shiro

    2018-05-01

    To determine steep increase of corneal irregularity induced by advancement of pterygium. A total of 456 eyes from 456 consecutive patients with primary pterygia were examined for corneal topography and advancement of pterygium with respect to the corneal diameter. Corneal irregularity induced by the pterygium advancement was evaluated by Fourier harmonic analyses of the topographic data that were modified for a series of analysis diameters from 1 mm to 6 mm. Incidences of steep increases in the asymmetry or higher-order irregularity components (inflection points) were determined by using segmented regression analysis for each analysis diameter. The pterygium advancement ranged from 2% to 57%, with a mean of 22.0%. Both components showed steep increases from the inflection points. The inflection points in the higher-order irregularity component altered with the analysis diameter (14.0%-30.6%), while there was no alternation in the asymmetry components (35.5%-36.8%). For the former component, the values at the inflection points were obtained in a range of 0.16 to 0.25 D. The Fourier harmonic analyses for a series of analysis diameters revealed that the higher-order irregularity component increased with the pterygium advancement. The analysis results confirmed the precedence of corneal irregularity due to pterygium advancement.

  19. Experimental investigations on friction laws and dryout heat flux of particulate beds packed with multi-size spheres and irregular particles

    International Nuclear Information System (INIS)

    Li, Liangxing; Ma, Weimin

    2011-01-01

    This paper is concerned with reducing uncertainty in quantification of debris bed coolability in a hypothetical severe accident of light water reactors (LWRs). A test facility named POMECO-FL is constructed to investigate the friction laws of adiabatic single and two-phase flow in a particulate bed packed with multi-size spheres or irregular particles. The same types of particles were then loaded in the test section of the POMECO-HT facility to obtain the dryout heat flux of the volumetrically heated particulate bed. The POMECO-HT facility features a high power capacity (up to 2.1 MW/m 2 ) which enables coolability study on particulate bed with broad variations in porosity and particle diameters under both top-flooding and bottom-injection conditions. The results show that given the effective particle diameter obtained from single-phase flow through the packed bed with multi-size spheres or irregular particles, both the pressure drop and the dryout heat flux of two-phase flow through the bed can be predicted by the Reed model. The bottom injection of coolant increases the dryout heat flux significantly. Meanwhile, the elevation of the dryout position is moving upwards with increasing bottom-injection flowrate. The experimental data provides insights for interpretation of debris bed coolability, as well as high-quality data for validation of the coolability analysis models and codes. (author)

  20. The majority of irregular menstrual cycles in adolescence are ovulatory: results of a prospective study.

    Science.gov (United States)

    Peña, Alexia S; Doherty, Dorota A; Atkinson, Helen C; Hickey, Martha; Norman, Robert J; Hart, Roger

    2018-03-01

    While ovulation is most likely to occur in adolescent girls with regular menstrual cycles, there are limited data on the incidence of ovulation in girls with irregular menstrual cycles in early postmenarcheal years. The aim of the study was to evaluate the presence of ovulation in healthy postmenarcheal girls with irregular menstrual cycles. Prospective cohort study over 12 weeks including 40 healthy postmenarcheal girls recruited from the population-based cohort of adolescents from Western Australian Pregnancy Cohort (Raine) Study with irregular menstrual cycles defined by either menstrual cycles 35 days in duration or cycle length that varied from month to month by >4 days according to menstrual diaries. Ovulation defined by urinary pregnanediol-3α-glucuronide/creatinine measurements higher than three times above minimum value obtained from 12 samples (1 per week). Forty girls (37 Caucasians) with irregular menstrual cycles aged 15.1 (median (IQR) 14.9-15.4) years who were 2.3 (1.9-3.3) years postmenarche were assessed. Urinary pregnanediol-3α-glucuronide/creatinine values identified that 33 girls (82.5%) ovulated during the 3 months of observation and 7 girls had anovulatory cycles. Menstrual diaries collected for a median (IQR) of 159 (137.5-188.2) days showed median minimal and maximum menstrual cycle duration of 24 (11.5-29) and 38.5 (35-48) days, respectively. A large proportion of healthy adolescent girls with irregular menstrual cycles are still ovulating despite irregular and infrequent menses. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.