WorldWideScience

Sample records for multiple parallel allele-specific

  1. QuASAR-MPRA: accurate allele-specific analysis for massively parallel reporter assays.

    Kalita, Cynthia A; Moyerbrailean, Gregory A; Brown, Christopher; Wen, Xiaoquan; Luca, Francesca; Pique-Regi, Roger

    2018-03-01

    The majority of the human genome is composed of non-coding regions containing regulatory elements such as enhancers, which are crucial for controlling gene expression. Many variants associated with complex traits are in these regions, and may disrupt gene regulatory sequences. Consequently, it is important to not only identify true enhancers but also to test if a variant within an enhancer affects gene regulation. Recently, allele-specific analysis in high-throughput reporter assays, such as massively parallel reporter assays (MPRAs), have been used to functionally validate non-coding variants. However, we are still missing high-quality and robust data analysis tools for these datasets. We have further developed our method for allele-specific analysis QuASAR (quantitative allele-specific analysis of reads) to analyze allele-specific signals in barcoded read counts data from MPRA. Using this approach, we can take into account the uncertainty on the original plasmid proportions, over-dispersion, and sequencing errors. The provided allelic skew estimate and its standard error also simplifies meta-analysis of replicate experiments. Additionally, we show that a beta-binomial distribution better models the variability present in the allelic imbalance of these synthetic reporters and results in a test that is statistically well calibrated under the null. Applying this approach to the MPRA data, we found 602 SNPs with significant (false discovery rate 10%) allele-specific regulatory function in LCLs. We also show that we can combine MPRA with QuASAR estimates to validate existing experimental and computational annotations of regulatory variants. Our study shows that with appropriate data analysis tools, we can improve the power to detect allelic effects in high-throughput reporter assays. http://github.com/piquelab/QuASAR/tree/master/mpra. fluca@wayne.edu or rpique@wayne.edu. Supplementary data are available online at Bioinformatics. © The Author (2017). Published by

  2. Multiple Avirulence Loci and Allele-Specific Effector Recognition Control the Pm3 Race-Specific Resistance of Wheat to Powdery Mildew[OPEN

    Roffler, Stefan; Stirnweis, Daniel; Treier, Georges; Herren, Gerhard; Korol, Abraham B.; Wicker, Thomas

    2015-01-01

    In cereals, several mildew resistance genes occur as large allelic series; for example, in wheat (Triticum aestivum and Triticum turgidum), 17 functional Pm3 alleles confer agronomically important race-specific resistance to powdery mildew (Blumeria graminis). The molecular basis of race specificity has been characterized in wheat, but little is known about the corresponding avirulence genes in powdery mildew. Here, we dissected the genetics of avirulence for six Pm3 alleles and found that three major Avr loci affect avirulence, with a common locus_1 involved in all AvrPm3-Pm3 interactions. We cloned the effector gene AvrPm3a2/f2 from locus_2, which is recognized by the Pm3a and Pm3f alleles. Induction of a Pm3 allele-dependent hypersensitive response in transient assays in Nicotiana benthamiana and in wheat demonstrated specificity. Gene expression analysis of Bcg1 (encoded by locus_1) and AvrPm3 a2/f2 revealed significant differences between isolates, indicating that in addition to protein polymorphisms, expression levels play a role in avirulence. We propose a model for race specificity involving three components: an allele-specific avirulence effector, a resistance gene allele, and a pathogen-encoded suppressor of avirulence. Thus, whereas a genetically simple allelic series controls specificity in the plant host, recognition on the pathogen side is more complex, allowing flexible evolutionary responses and adaptation to resistance genes. PMID:26452600

  3. Allele specific expression and methylation in the bumblebee, Bombus terrestris

    Zoë Lonsdale

    2017-09-01

    Full Text Available The social hymenoptera are emerging as models for epigenetics. DNA methylation, the addition of a methyl group, is a common epigenetic marker. In mammals and flowering plants methylation affects allele specific expression. There is contradictory evidence for the role of methylation on allele specific expression in social insects. The aim of this paper is to investigate allele specific expression and monoallelic methylation in the bumblebee, Bombus terrestris. We found nineteen genes that were both monoallelically methylated and monoallelically expressed in a single bee. Fourteen of these genes express the hypermethylated allele, while the other five express the hypomethylated allele. We also searched for allele specific expression in twenty-nine published RNA-seq libraries. We found 555 loci with allele-specific expression. We discuss our results with reference to the functional role of methylation in gene expression in insects and in the as yet unquantified role of genetic cis effects in insect allele specific methylation and expression.

  4. QuASAR: quantitative allele-specific analysis of reads.

    Harvey, Chris T; Moyerbrailean, Gregory A; Davis, Gordon O; Wen, Xiaoquan; Luca, Francesca; Pique-Regi, Roger

    2015-04-15

    Expression quantitative trait loci (eQTL) studies have discovered thousands of genetic variants that regulate gene expression, enabling a better understanding of the functional role of non-coding sequences. However, eQTL studies are costly, requiring large sample sizes and genome-wide genotyping of each sample. In contrast, analysis of allele-specific expression (ASE) is becoming a popular approach to detect the effect of genetic variation on gene expression, even within a single individual. This is typically achieved by counting the number of RNA-seq reads matching each allele at heterozygous sites and testing the null hypothesis of a 1:1 allelic ratio. In principle, when genotype information is not readily available, it could be inferred from the RNA-seq reads directly. However, there are currently no existing methods that jointly infer genotypes and conduct ASE inference, while considering uncertainty in the genotype calls. We present QuASAR, quantitative allele-specific analysis of reads, a novel statistical learning method for jointly detecting heterozygous genotypes and inferring ASE. The proposed ASE inference step takes into consideration the uncertainty in the genotype calls, while including parameters that model base-call errors in sequencing and allelic over-dispersion. We validated our method with experimental data for which high-quality genotypes are available. Results for an additional dataset with multiple replicates at different sequencing depths demonstrate that QuASAR is a powerful tool for ASE analysis when genotypes are not available. http://github.com/piquelab/QuASAR. fluca@wayne.edu or rpique@wayne.edu Supplementary Material is available at Bioinformatics online. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  5. Keldysh formalism for multiple parallel worlds

    Ansari, M.; Nazarov, Y. V.

    2016-01-01

    We present a compact and self-contained review of the recently developed Keldysh formalism for multiple parallel worlds. The formalism has been applied to consistent quantum evaluation of the flows of informational quantities, in particular, to the evaluation of Renyi and Shannon entropy flows. We start with the formulation of the standard and extended Keldysh techniques in a single world in a form convenient for our presentation. We explain the use of Keldysh contours encompassing multiple parallel worlds. In the end, we briefly summarize the concrete results obtained with the method.

  6. Keldysh formalism for multiple parallel worlds

    Ansari, M.; Nazarov, Y. V.

    2016-03-01

    We present a compact and self-contained review of the recently developed Keldysh formalism for multiple parallel worlds. The formalism has been applied to consistent quantum evaluation of the flows of informational quantities, in particular, to the evaluation of Renyi and Shannon entropy flows. We start with the formulation of the standard and extended Keldysh techniques in a single world in a form convenient for our presentation. We explain the use of Keldysh contours encompassing multiple parallel worlds. In the end, we briefly summarize the concrete results obtained with the method.

  7. Keldysh formalism for multiple parallel worlds

    Ansari, M.; Nazarov, Y. V., E-mail: y.v.nazarov@tudelft.nl [Delft University of Technology, Kavli Institute of Nanoscience (Netherlands)

    2016-03-15

    We present a compact and self-contained review of the recently developed Keldysh formalism for multiple parallel worlds. The formalism has been applied to consistent quantum evaluation of the flows of informational quantities, in particular, to the evaluation of Renyi and Shannon entropy flows. We start with the formulation of the standard and extended Keldysh techniques in a single world in a form convenient for our presentation. We explain the use of Keldysh contours encompassing multiple parallel worlds. In the end, we briefly summarize the concrete results obtained with the method.

  8. Allele Workbench: transcriptome pipeline and interactive graphics for allele-specific expression.

    Carol A Soderlund

    Full Text Available Sequencing the transcriptome can answer various questions such as determining the transcripts expressed in a given species for a specific tissue or condition, evaluating differential expression, discovering variants, and evaluating allele-specific expression. Differential expression evaluates the expression differences between different strains, tissues, and conditions. Allele-specific expression evaluates expression differences between parental alleles. Both differential expression and allele-specific expression have been studied for heterosis (hybrid vigor, where the hybrid has improved performance over the parents for one or more traits. The Allele Workbench software was developed for a heterosis study that evaluated allele-specific expression for a mouse F1 hybrid using libraries from multiple tissues with biological replicates. This software has been made into a distributable package, which includes a pipeline, a Java interface to build the database, and a Java interface for query and display of the results. The required input is a reference genome, annotation file, and one or more RNA-Seq libraries with optional replicates. It evaluates allelic imbalance at the SNP and transcript level and flags transcripts with significant opposite directional allele-specific expression. The Java interface allows the user to view data from libraries, replicates, genes, transcripts, exons, and variants, including queries on allele imbalance for selected libraries. To determine the impact of allele-specific SNPs on protein folding, variants are annotated with their effect (e.g., missense, and the parental protein sequences may be exported for protein folding analysis. The Allele Workbench processing results in transcript files and read counts that can be used as input to the previously published Transcriptome Computational Workbench, which has a new algorithm for determining a trimmed set of gene ontology terms. The software with demo files is available

  9. Parallel magnetotransport in multiple quantum well structures

    Sheregii, E.M.; Ploch, D.; Marchewka, M.; Tomaka, G.; Kolek, A.; Stadler, A.; Mleczko, K.; Strupinski, W.; Jasik, A.; Jakiela, R.

    2004-01-01

    The results of investigations of parallel magnetotransport in AlGaAs/GaAs and InGaAs/InAlAs/InP multiple quantum wells structures (MQW's) are presented in this paper. The MQW's were obtained by metalorganic vapour phase epitaxy with different shapes of QW, numbers of QW and levels of doping. The magnetotransport measurements were performed in wide region of temperatures (0.5-300 K) and at high magnetic fields up to 30 T (B is perpendicular and current is parallel to the plane of the QW). Three types of observed effects are analyzed: quantum Hall effect and Shubnikov-de Haas oscillations at low temperatures (0.5-6 K) as well as magnetophonon resonance at higher temperatures (77-300 K)

  10. Allele-Specific DNA Methylation Detection by Pyrosequencing®

    Kristensen, Lasse Sommer; Johansen, Jens Vilstrup; Grønbæk, Kirsten

    2015-01-01

    DNA methylation is an epigenetic modification that plays important roles in healthy as well as diseased cells, by influencing the transcription of genes. In spite the fact that human somatic cells are diploid, most of the currently available methods for the study of DNA methylation do not provide......-effective protocol for allele-specific DNA methylation detection based on Pyrosequencing(®) of methylation-specific PCR (MSP) products including a single nucleotide polymorphism (SNP) within the amplicon....

  11. Conceptual design of multiple parallel switching controller

    Ugolini, D.; Yoshikawa, S.; Ozawa, K.

    1996-01-01

    This paper discusses the conceptual design and the development of a preliminary model of a multiple parallel switching (MPS) controller. The introduction of several advanced controllers has widened and improved the control capability of nonlinear dynamical systems. However, it is not possible to uniquely define a controller that always outperforms the others, and, in many situations, the controller providing the best control action depends on the operating conditions and on the intrinsic properties and behavior of the controlled dynamical system. The desire to combine the control action of several controllers with the purpose to continuously attain the best control action has motivated the development of the MPS controller. The MPS controller consists of a number of single controllers acting in parallel and of an artificial intelligence (AI) based selecting mechanism. The AI selecting mechanism analyzes the output of each controller and implements the one providing the best control performance. An inherent property of the MPS controller is the possibility to discard unreliable controllers while still being able to perform the control action. To demonstrate the feasibility and the capability of the MPS controller the simulation of the on-line operation control of a fast breeder reactor (FBR) evaporator is presented. (author)

  12. ALEA: a toolbox for allele-specific epigenomics analysis.

    Younesy, Hamid; Möller, Torsten; Heravi-Moussavi, Alireza; Cheng, Jeffrey B; Costello, Joseph F; Lorincz, Matthew C; Karimi, Mohammad M; Jones, Steven J M

    2014-04-15

    The assessment of expression and epigenomic status using sequencing based methods provides an unprecedented opportunity to identify and correlate allelic differences with epigenomic status. We present ALEA, a computational toolbox for allele-specific epigenomics analysis, which incorporates allelic variation data within existing resources, allowing for the identification of significant associations between epigenetic modifications and specific allelic variants in human and mouse cells. ALEA provides a customizable pipeline of command line tools for allele-specific analysis of next-generation sequencing data (ChIP-seq, RNA-seq, etc.) that takes the raw sequencing data and produces separate allelic tracks ready to be viewed on genome browsers. The pipeline has been validated using human and hybrid mouse ChIP-seq and RNA-seq data. The package, test data and usage instructions are available online at http://www.bcgsc.ca/platform/bioinfo/software/alea CONTACT: : mkarimi1@interchange.ubc.ca or sjones@bcgsc.ca Supplementary information: Supplementary data are available at Bioinformatics online. © The Author 2013. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  13. Allele-specific MMP-3 transcription under in vivo conditions

    Chaoyong, Zhu [Atherosclerosis Research Unit, King Gustav V Research Institute, Department of Medicine, Karolinska Institute, Stockholm (Sweden); Odeberg, Jacob [Atherosclerosis Research Unit, King Gustav V Research Institute, Department of Medicine, Karolinska Institute, Stockholm (Sweden); Department of Biotechnology, AlbaNova University Center, Royal Institute of Technology, Stockholm (Sweden); Hamsten, Anders [Atherosclerosis Research Unit, King Gustav V Research Institute, Department of Medicine, Karolinska Institute, Stockholm (Sweden); Eriksson, Per [Atherosclerosis Research Unit, King Gustav V Research Institute, Department of Medicine, Karolinska Institute, Stockholm (Sweden)

    2006-09-29

    A common matrix metalloproteinases-3 (MMP-3) -1612 5A/6A promoter polymorphism is associated with risk for cardiovascular disease, rheumatoid arthritis, and other diseases. Here we used the haplotype chromatin immunoprecipitation method to study allele-specific MMP-3 expression under in vivo conditions in heterozygous THP-1 cells. Pyrosequencing was used to analyse the ratio of 5A-allele to 6A-allele after chromatin immunoprecipitation using an antibody against phosphorylated active RNA polymerase II. There was no allele-specific difference in transcriptional activity during basal conditions, i.e., in unstimulated monocytic THP-1 cells. However, after stimulation of MMP-3 expression by monocyte differentiation or incubation with IL-1{beta}, the haplotype containing the 5A-allele was associated with higher transcriptional activity compared with the 6A-containing haplotype. Electromobility shift assay demonstrated increased binding of nuclear proteins to the 5A-allele after monocyte differentiation. In conclusion, the common MMP-3 5A/6A promoter polymorphism appears to be functional only during specific environmental conditions involving inflammation.

  14. Delimiting Allelic Imbalance of TYMS by Allele-Specific Analysis.

    Balboa-Beltrán, Emilia; Cruz, Raquel; Carracedo, Angel; Barros, Francisco

    2015-07-01

    Allelic imbalance of thymidylate synthase (TYMS) is attributed to polymorphisms in the 5'- and 3'-untranslated region (UTR). These polymorphisms have been related to the risk of suffering different cancers, for example leukemia, breast or gastric cancer, and response to different drugs, among which are methotrexate glutamates, stavudine, and specifically 5-fluorouracil (5-FU), as TYMS is its direct target. A vast literature has been published in relation to 5-FU, even suggesting the sole use of these polymorphisms to effectively manage 5-FU dosage. Estimates of the extent to which these polymorphisms influence in TYMS expression have in the past been based on functional analysis by luciferase assays and quantification of TYMS mRNA, but both these studies, as the association studies with cancer risk or with toxicity or response to 5-FU, are very contradictory. Regarding functional assays, the artificial genetic environment created in luciferase assay and the problems derived from quantitative polymerase chain reactions (qPCRs), for example the use of a reference gene, may have distorted the results. To avoid these sources of interference, we have analyzed the allelic imbalance of TYMS by allelic-specific analysis in peripheral blood mononuclear cells (PBMCs) from patients.Allelic imbalance in PBMCs, taken from 40 patients with suspected myeloproliferative haematological diseases, was determined by fluorescent fragment analysis (for the 3'-UTR polymorphism), Sanger sequencing and allelic-specific qPCR in multiplex (for the 5'-UTR polymorphisms).For neither the 3'- nor the 5'-UTR polymorphisms did the observed allelic imbalance exceed 1.5 fold. None of the TYMS polymorphisms is statistically associated with allelic imbalance.The results acquired allow us to deny the previously established assertion of an influence of 2 to 4 fold of the rs45445694 and rs2853542 polymorphisms in the expression of TYMS and narrow its allelic imbalance to 1.5 fold, in our population

  15. Comparative anatomy of chromosomal domains with imprinted and non-imprinted allele-specific DNA methylation.

    Paliwal, Anupam; Temkin, Alexis M; Kerkel, Kristi; Yale, Alexander; Yotova, Iveta; Drost, Natalia; Lax, Simon; Nhan-Chang, Chia-Ling; Powell, Charles; Borczuk, Alain; Aviv, Abraham; Wapner, Ronald; Chen, Xiaowei; Nagy, Peter L; Schork, Nicholas; Do, Catherine; Torkamani, Ali; Tycko, Benjamin

    2013-08-01

    Allele-specific DNA methylation (ASM) is well studied in imprinted domains, but this type of epigenetic asymmetry is actually found more commonly at non-imprinted loci, where the ASM is dictated not by parent-of-origin but instead by the local haplotype. We identified loci with strong ASM in human tissues from methylation-sensitive SNP array data. Two index regions (bisulfite PCR amplicons), one between the C3orf27 and RPN1 genes in chromosome band 3q21 and the other near the VTRNA2-1 vault RNA in band 5q31, proved to be new examples of imprinted DMRs (maternal alleles methylated) while a third, between STEAP3 and C2orf76 in chromosome band 2q14, showed non-imprinted haplotype-dependent ASM. Using long-read bisulfite sequencing (bis-seq) in 8 human tissues we found that in all 3 domains the ASM is restricted to single differentially methylated regions (DMRs), each less than 2kb. The ASM in the C3orf27-RPN1 intergenic region was placenta-specific and associated with allele-specific expression of a long non-coding RNA. Strikingly, the discrete DMRs in all 3 regions overlap with binding sites for the insulator protein CTCF, which we found selectively bound to the unmethylated allele of the STEAP3-C2orf76 DMR. Methylation mapping in two additional genes with non-imprinted haplotype-dependent ASM, ELK3 and CYP2A7, showed that the CYP2A7 DMR also overlaps a CTCF site. Thus, two features of imprinted domains, highly localized DMRs and allele-specific insulator occupancy by CTCF, can also be found in chromosomal domains with non-imprinted ASM. Arguing for biological importance, our analysis of published whole genome bis-seq data from hES cells revealed multiple genome-wide association study (GWAS) peaks near CTCF binding sites with ASM.

  16. Genome-wide survey of allele-specific splicing in humans

    Scheffler Konrad

    2008-06-01

    Full Text Available Abstract Background Accurate mRNA splicing depends on multiple regulatory signals encoded in the transcribed RNA sequence. Many examples of mutations within human splice regulatory regions that alter splicing qualitatively or quantitatively have been reported and allelic differences in mRNA splicing are likely to be a common and important source of phenotypic diversity at the molecular level, in addition to their contribution to genetic disease susceptibility. However, because the effect of a mutation on the efficiency of mRNA splicing is often difficult to predict, many mutations that cause disease through an effect on splicing are likely to remain undiscovered. Results We have combined a genome-wide scan for sequence polymorphisms likely to affect mRNA splicing with analysis of publicly available Expressed Sequence Tag (EST and exon array data. The genome-wide scan uses published tools and identified 30,977 SNPs located within donor and acceptor splice sites, branch points and exonic splicing enhancer elements. For 1,185 candidate splicing polymorphisms the difference in splicing between alternative alleles was corroborated by publicly available exon array data from 166 lymphoblastoid cell lines. We developed a novel probabilistic method to infer allele-specific splicing from EST data. The method uses SNPs and alternative mRNA isoforms mapped to EST sequences and models both regulated alternative splicing as well as allele-specific splicing. We have also estimated heritability of splicing and report that a greater proportion of genes show evidence of splicing heritability than show heritability of overall gene expression level. Our results provide an extensive resource that can be used to assess the possible effect on splicing of human polymorphisms in putative splice-regulatory sites. Conclusion We report a set of genes showing evidence of allele-specific splicing from an integrated analysis of genomic polymorphisms, EST data and exon array

  17. Hypergraph partitioning implementation for parallelizing matrix-vector multiplication using CUDA GPU-based parallel computing

    Murni, Bustamam, A.; Ernastuti, Handhika, T.; Kerami, D.

    2017-07-01

    Calculation of the matrix-vector multiplication in the real-world problems often involves large matrix with arbitrary size. Therefore, parallelization is needed to speed up the calculation process that usually takes a long time. Graph partitioning techniques that have been discussed in the previous studies cannot be used to complete the parallelized calculation of matrix-vector multiplication with arbitrary size. This is due to the assumption of graph partitioning techniques that can only solve the square and symmetric matrix. Hypergraph partitioning techniques will overcome the shortcomings of the graph partitioning technique. This paper addresses the efficient parallelization of matrix-vector multiplication through hypergraph partitioning techniques using CUDA GPU-based parallel computing. CUDA (compute unified device architecture) is a parallel computing platform and programming model that was created by NVIDIA and implemented by the GPU (graphics processing unit).

  18. Screening for SNPs with Allele-Specific Methylation based on Next-Generation Sequencing Data.

    Hu, Bo; Ji, Yuan; Xu, Yaomin; Ting, Angela H

    2013-05-01

    Allele-specific methylation (ASM) has long been studied but mainly documented in the context of genomic imprinting and X chromosome inactivation. Taking advantage of the next-generation sequencing technology, we conduct a high-throughput sequencing experiment with four prostate cell lines to survey the whole genome and identify single nucleotide polymorphisms (SNPs) with ASM. A Bayesian approach is proposed to model the counts of short reads for each SNP conditional on its genotypes of multiple subjects, leading to a posterior probability of ASM. We flag SNPs with high posterior probabilities of ASM by accounting for multiple comparisons based on posterior false discovery rates. Applying the Bayesian approach to the in-house prostate cell line data, we identify 269 SNPs as candidates of ASM. A simulation study is carried out to demonstrate the quantitative performance of the proposed approach.

  19. SWAMP+: multiple subsequence alignment using associative massive parallelism

    Steinfadt, Shannon Irene [Los Alamos National Laboratory; Baker, Johnnie W [KENT STATE UNIV.

    2010-10-18

    A new parallel algorithm SWAMP+ incorporates the Smith-Waterman sequence alignment on an associative parallel model known as ASC. It is a highly sensitive parallel approach that expands traditional pairwise sequence alignment. This is the first parallel algorithm to provide multiple non-overlapping, non-intersecting subsequence alignments with the accuracy of Smith-Waterman. The efficient algorithm provides multiple alignments similar to BLAST while creating a better workflow for the end users. The parallel portions of the code run in O(m+n) time using m processors. When m = n, the algorithmic analysis becomes O(n) with a coefficient of two, yielding a linear speedup. Implementation of the algorithm on the SIMD ClearSpeed CSX620 confirms this theoretical linear speedup with real timings.

  20. Allele specific expression in worker reproduction genes in the bumblebee Bombus terrestris

    Harindra E. Amarasinghe

    2015-07-01

    Full Text Available Methylation has previously been associated with allele specific expression in ants. Recently, we found methylation is important in worker reproduction in the bumblebee Bombus terrestris. Here we searched for allele specific expression in twelve genes associated with worker reproduction in bees. We found allele specific expression in Ecdysone 20 monooxygenase and IMP-L2-like. Although we were unable to confirm a genetic or epigenetic cause for this allele specific expression, the expression patterns of the two genes match those predicted for imprinted genes.

  1. Comparative anatomy of chromosomal domains with imprinted and non-imprinted allele-specific DNA methylation.

    Anupam Paliwal

    2013-08-01

    Full Text Available Allele-specific DNA methylation (ASM is well studied in imprinted domains, but this type of epigenetic asymmetry is actually found more commonly at non-imprinted loci, where the ASM is dictated not by parent-of-origin but instead by the local haplotype. We identified loci with strong ASM in human tissues from methylation-sensitive SNP array data. Two index regions (bisulfite PCR amplicons, one between the C3orf27 and RPN1 genes in chromosome band 3q21 and the other near the VTRNA2-1 vault RNA in band 5q31, proved to be new examples of imprinted DMRs (maternal alleles methylated while a third, between STEAP3 and C2orf76 in chromosome band 2q14, showed non-imprinted haplotype-dependent ASM. Using long-read bisulfite sequencing (bis-seq in 8 human tissues we found that in all 3 domains the ASM is restricted to single differentially methylated regions (DMRs, each less than 2kb. The ASM in the C3orf27-RPN1 intergenic region was placenta-specific and associated with allele-specific expression of a long non-coding RNA. Strikingly, the discrete DMRs in all 3 regions overlap with binding sites for the insulator protein CTCF, which we found selectively bound to the unmethylated allele of the STEAP3-C2orf76 DMR. Methylation mapping in two additional genes with non-imprinted haplotype-dependent ASM, ELK3 and CYP2A7, showed that the CYP2A7 DMR also overlaps a CTCF site. Thus, two features of imprinted domains, highly localized DMRs and allele-specific insulator occupancy by CTCF, can also be found in chromosomal domains with non-imprinted ASM. Arguing for biological importance, our analysis of published whole genome bis-seq data from hES cells revealed multiple genome-wide association study (GWAS peaks near CTCF binding sites with ASM.

  2. An efficient parallel algorithm for matrix-vector multiplication

    Hendrickson, B.; Leland, R.; Plimpton, S.

    1993-03-01

    The multiplication of a vector by a matrix is the kernel computation of many algorithms in scientific computation. A fast parallel algorithm for this calculation is therefore necessary if one is to make full use of the new generation of parallel supercomputers. This paper presents a high performance, parallel matrix-vector multiplication algorithm that is particularly well suited to hypercube multiprocessors. For an n x n matrix on p processors, the communication cost of this algorithm is O(n/[radical]p + log(p)), independent of the matrix sparsity pattern. The performance of the algorithm is demonstrated by employing it as the kernel in the well-known NAS conjugate gradient benchmark, where a run time of 6.09 seconds was observed. This is the best published performance on this benchmark achieved to date using a massively parallel supercomputer.

  3. Multiple Independent File Parallel I/O with HDF5

    Miller, M. C.

    2016-07-13

    The HDF5 library has supported the I/O requirements of HPC codes at Lawrence Livermore National Labs (LLNL) since the late 90’s. In particular, HDF5 used in the Multiple Independent File (MIF) parallel I/O paradigm has supported LLNL code’s scalable I/O requirements and has recently been gainfully used at scales as large as O(106) parallel tasks.

  4. Efficient multitasking: parallel versus serial processing of multiple tasks.

    Fischer, Rico; Plessow, Franziska

    2015-01-01

    In the context of performance optimizations in multitasking, a central debate has unfolded in multitasking research around whether cognitive processes related to different tasks proceed only sequentially (one at a time), or can operate in parallel (simultaneously). This review features a discussion of theoretical considerations and empirical evidence regarding parallel versus serial task processing in multitasking. In addition, we highlight how methodological differences and theoretical conceptions determine the extent to which parallel processing in multitasking can be detected, to guide their employment in future research. Parallel and serial processing of multiple tasks are not mutually exclusive. Therefore, questions focusing exclusively on either task-processing mode are too simplified. We review empirical evidence and demonstrate that shifting between more parallel and more serial task processing critically depends on the conditions under which multiple tasks are performed. We conclude that efficient multitasking is reflected by the ability of individuals to adjust multitasking performance to environmental demands by flexibly shifting between different processing strategies of multiple task-component scheduling.

  5. Allele-Specific Alternative mRNA processing (ASARP) | Informatics Technology for Cancer Research (ITCR)

    A software pipeline for prediction of allele-specific alternative RNA processing events using single RNA-seq data. The current version focuses on prediction of alternative splicing and alternative polyadenylation modulated by genetic variants.

  6. A scalable parallel algorithm for multiple objective linear programs

    Wiecek, Malgorzata M.; Zhang, Hong

    1994-01-01

    This paper presents an ADBASE-based parallel algorithm for solving multiple objective linear programs (MOLP's). Job balance, speedup and scalability are of primary interest in evaluating efficiency of the new algorithm. Implementation results on Intel iPSC/2 and Paragon multiprocessors show that the algorithm significantly speeds up the process of solving MOLP's, which is understood as generating all or some efficient extreme points and unbounded efficient edges. The algorithm gives specially good results for large and very large problems. Motivation and justification for solving such large MOLP's are also included.

  7. Simultaneous SNP identification and assessment of allele-specific bias from ChIP-seq data

    Ni Yunyun

    2012-09-01

    Full Text Available Abstract Background Single nucleotide polymorphisms (SNPs have been associated with many aspects of human development and disease, and many non-coding SNPs associated with disease risk are presumed to affect gene regulation. We have previously shown that SNPs within transcription factor binding sites can affect transcription factor binding in an allele-specific and heritable manner. However, such analysis has relied on prior whole-genome genotypes provided by large external projects such as HapMap and the 1000 Genomes Project. This requirement limits the study of allele-specific effects of SNPs in primary patient samples from diseases of interest, where complete genotypes are not readily available. Results In this study, we show that we are able to identify SNPs de novo and accurately from ChIP-seq data generated in the ENCODE Project. Our de novo identified SNPs from ChIP-seq data are highly concordant with published genotypes. Independent experimental verification of more than 100 sites estimates our false discovery rate at less than 5%. Analysis of transcription factor binding at de novo identified SNPs revealed widespread heritable allele-specific binding, confirming previous observations. SNPs identified from ChIP-seq datasets were significantly enriched for disease-associated variants, and we identified dozens of allele-specific binding events in non-coding regions that could distinguish between disease and normal haplotypes. Conclusions Our approach combines SNP discovery, genotyping and allele-specific analysis, but is selectively focused on functional regulatory elements occupied by transcription factors or epigenetic marks, and will therefore be valuable for identifying the functional regulatory consequences of non-coding SNPs in primary disease samples.

  8. Allele-Specific Chromatin Recruitment and Therapeutic Vulnerabilities of ESR1 Activating Mutations.

    Jeselsohn, Rinath; Bergholz, Johann S; Pun, Matthew; Cornwell, MacIntosh; Liu, Weihan; Nardone, Agostina; Xiao, Tengfei; Li, Wei; Qiu, Xintao; Buchwalter, Gilles; Feiglin, Ariel; Abell-Hart, Kayley; Fei, Teng; Rao, Prakash; Long, Henry; Kwiatkowski, Nicholas; Zhang, Tinghu; Gray, Nathanael; Melchers, Diane; Houtman, Rene; Liu, X Shirley; Cohen, Ofir; Wagle, Nikhil; Winer, Eric P; Zhao, Jean; Brown, Myles

    2018-02-12

    Estrogen receptor α (ER) ligand-binding domain (LBD) mutations are found in a substantial number of endocrine treatment-resistant metastatic ER-positive (ER + ) breast cancers. We investigated the chromatin recruitment, transcriptional network, and genetic vulnerabilities in breast cancer models harboring the clinically relevant ER mutations. These mutants exhibit both ligand-independent functions that mimic estradiol-bound wild-type ER as well as allele-specific neomorphic properties that promote a pro-metastatic phenotype. Analysis of the genome-wide ER binding sites identified mutant ER unique recruitment mediating the allele-specific transcriptional program. Genetic screens identified genes that are essential for the ligand-independent growth driven by the mutants. These studies provide insights into the mechanism of endocrine therapy resistance engendered by ER mutations and potential therapeutic targets. Copyright © 2018 Elsevier Inc. All rights reserved.

  9. Screening for SNPs with Allele-Specific Methylation based on Next-Generation Sequencing Data

    Hu, Bo; Ji, Yuan; Xu, Yaomin; Ting, Angela H

    2013-01-01

    Allele-specific methylation (ASM) has long been studied but mainly documented in the context of genomic imprinting and X chromosome inactivation. Taking advantage of the next-generation sequencing technology, we conduct a high-throughput sequencing experiment with four prostate cell lines to survey the whole genome and identify single nucleotide polymorphisms (SNPs) with ASM. A Bayesian approach is proposed to model the counts of short reads for each SNP conditional on its genotypes of multip...

  10. Allele-specific deletions in mouse tumors identify Fbxw7 as germline modifier of tumor susceptibility.

    Jesus Perez-Losada

    Full Text Available Genome-wide association studies (GWAS have been successful in finding associations between specific genetic variants and cancer susceptibility in human populations. These studies have identified a range of highly statistically significant associations between single nucleotide polymorphisms (SNPs and susceptibility to development of a range of human tumors. However, the effect of each SNP in isolation is very small, and all of the SNPs combined only account for a relatively minor proportion of the total genetic risk (5-10%. There is therefore a major requirement for alternative routes to the discovery of genetic risk factors for cancer. We have previously shown using mouse models that chromosomal regions harboring susceptibility genes identified by linkage analysis frequently exhibit allele-specific genetic alterations in tumors. We demonstrate here that the Fbxw7 gene, a commonly mutated gene in a wide range of mouse and human cancers, shows allele-specific deletions in mouse lymphomas and skin tumors. Lymphomas from three different F1 hybrids show 100% allele-specificity in the patterns of allelic loss. Parental alleles from 129/Sv or Spretus/Gla mice are lost in tumors from F1 hybrids with C57BL/6 animals, due to the presence of a specific non-synonymous coding sequence polymorphism at the N-terminal portion of the gene. A specific genetic test of association between this SNP and lymphoma susceptibility in interspecific backcross mice showed a significant linkage (p = 0.001, but only in animals with a functional p53 gene. These data therefore identify Fbxw7 as a p53-dependent tumor susceptibility gene. Increased p53-dependent tumor susceptibility and allele-specific losses were also seen in a mouse skin model of skin tumor development. We propose that analysis of preferential allelic imbalances in tumors may provide an efficient means of uncovering genetic variants that affect mouse and human tumor susceptibility.

  11. A small asparagine-rich protein required for S-allele-specific pollen rejection in Nicotiana.

    McClure, B; Mou, B; Canevascini, S; Bernatzky, R

    1999-11-09

    Although S-locus RNases (S-RNases) determine the specificity of pollen rejection in self-incompatible (SI) solanaceous plants, they alone are not sufficient to cause S-allele-specific pollen rejection. To identify non-S-RNase sequences that are required for pollen rejection, a Nicotiana alata cDNA library was screened by differential hybridization. One clone, designated HT, hybridized strongly to RNA from N. alata styles but not to RNA from Nicotiana plumbaginifolia, a species known to lack one or more factors necessary for S-allele-specific pollen rejection. Sequence analysis revealed a 101-residue ORF including a putative secretion signal and an asparagine-rich domain near the C terminus. RNA blot analysis showed that the HT-transcript accumulates in the stigma and style before anthesis. The timing of HT-expression lags slightly behind S(C10)-RNase in SI N. alata S(C10)S(C10) and is well correlated with the onset of S-allele-specific pollen rejection in the style. An antisense-HT construct was prepared to test for a role in pollen rejection. Transformed (N. plumbaginifolia x SI N. alata S(C10)S(C10)) hybrids with reduced levels of HT-protein continued to express S(C10)-RNase but failed to reject S(C10)-pollen. Control hybrids expressing both S(C10)-RNase and HT-protein showed a normal S-allele-specific pollen rejection response. We conclude that HT-protein is directly implicated in pollen rejection.

  12. Allele-specific physical interactions regulate the heterotic traits in hybrids of Arabidopsis thaliana ecotypes

    Babita Singh

    2017-10-01

    Full Text Available Heterosis is an important phenomenon for the breeding in agricultural crops as it influences yield related traits such as biomass yield, seed number and weight, adaptive and reproductive traits. However, the level of heterosis greatly varies for different traits and different genotypes. The present study focuses on identification of physical interactions between alleles and their role in transcriptional regulation in heterotic plants. Here, we used two Arabidopsis ecotypes; Col-0 and C24 as parent for crosses. We performed crossing between these ecotypes and screened the F1 hybrids on the basis of different SSR markers. Further, we used Hi-C to capture intra- and inter-chromosomal physical interactions between alleles on genome-wide level. Then, we identified allele-specific chromatin interactions and constructed genome-wide allele-specific contact maps at different resolutions for the entire chromosome. We also performed RNA-seq of hybrids and their parents. RNA-seq analysis identified several differentially expressed genes and non-additively expressed genes in hybrids with respect to their parents. Further, to understand the biological significance of these chromatin interactions, we annotated these interactions and correlated with the transcriptome data. Thus, our study provides alleles-specific chromatin interactions in genome-wide fashion which play a crucial role in regulation of different genes that may be important for heterosis.

  13. Parallel multiple instance learning for extremely large histopathology image analysis.

    Xu, Yan; Li, Yeshu; Shen, Zhengyang; Wu, Ziwei; Gao, Teng; Fan, Yubo; Lai, Maode; Chang, Eric I-Chao

    2017-08-03

    Histopathology images are critical for medical diagnosis, e.g., cancer and its treatment. A standard histopathology slice can be easily scanned at a high resolution of, say, 200,000×200,000 pixels. These high resolution images can make most existing imaging processing tools infeasible or less effective when operated on a single machine with limited memory, disk space and computing power. In this paper, we propose an algorithm tackling this new emerging "big data" problem utilizing parallel computing on High-Performance-Computing (HPC) clusters. Experimental results on a large-scale data set (1318 images at a scale of 10 billion pixels each) demonstrate the efficiency and effectiveness of the proposed algorithm for low-latency real-time applications. The framework proposed an effective and efficient system for extremely large histopathology image analysis. It is based on the multiple instance learning formulation for weakly-supervised learning for image classification, segmentation and clustering. When a max-margin concept is adopted for different clusters, we obtain further improvement in clustering performance.

  14. Rapid ABO genotyping by high-speed droplet allele-specific PCR using crude samples.

    Taira, Chiaki; Matsuda, Kazuyuki; Takeichi, Naoya; Furukawa, Satomi; Sugano, Mitsutoshi; Uehara, Takeshi; Okumura, Nobuo; Honda, Takayuki

    2018-01-01

    ABO genotyping has common tools for personal identification of forensic and transplantation field. We developed a new method based on a droplet allele-specific PCR (droplet-AS-PCR) that enabled rapid PCR amplification. We attempted rapid ABO genotyping using crude DNA isolated from dried blood and buccal cells. We designed allele-specific primers for three SNPs (at nucleotides 261, 526, and 803) in exons 6 and 7 of the ABO gene. We pretreated dried blood and buccal cells with proteinase K, and obtained crude DNAs without DNA purification. Droplet-AS-PCR allowed specific amplification of the SNPs at the three loci using crude DNA, with results similar to those for DNA extracted from fresh peripheral blood. The sensitivity of the methods was 5%-10%. The genotyping of extracted DNA and crude DNA were completed within 8 and 9 minutes, respectively. The genotypes determined by the droplet-AS-PCR method were always consistent with those obtained by direct sequencing. The droplet-AS-PCR method enabled rapid and specific amplification of three SNPs of the ABO gene from crude DNA treated with proteinase K. ABO genotyping by the droplet-AS-PCR has the potential to be applied to various fields including a forensic medicine and transplantation medical care. © 2017 Wiley Periodicals, Inc.

  15. Allele-specific cytokine responses at the HLA-C locus, implications for psoriasis

    Hundhausen, Christian; Bertoni, Anna; Mak, Rose K; Botti, Elisabetta; Di Meglio, Paola; Clop, Alex; Laggner, Ute; Chimenti, Sergio; Hayday, Adrian C; Barker, Jonathan N; Trembath, Richard C; Capon, Francesca; Nestle, Frank O

    2011-01-01

    Psoriasis is an inflammatory skin disorder that is inherited as a complex trait. Genetic studies have repeatedly highlighted HLA-C as the major determinant for psoriasis susceptibility, with the Cw*0602 allele conferring significant disease risk in a wide-range of populations. Despite the potential importance of HLA-C variation in psoriasis, either via an effect on peptide presentation or immuno-inhibitory activity, allele-specific expression patterns have not been investigated. Here, we used reporter assays to characterize two regulatory variants, which virtually abolished the response to TNF-α (rs2524094) and IFN-γ (rs10657191) in HLA-Cw*0602 and a cluster of related alleles. We validated these findings through the analysis of HLA-Cw*0602 expression in primary keratinocytes treated with TNF-α and IFN-γ. Finally, we showed that HLA-Cw*0602 transcripts are not increased in psoriatic skin lesions, despite highly elevated TNF-α levels. Thus, our findings demonstrate the presence of allele-specific differences in HLA-C expression and indicate that HLA-Cw*0602 is unresponsive to up-regulation by key pro-inflammatory cytokines in psoriasis. These data pave the way for functional studies into the pathogenic role of the major psoriasis susceptibility allele. PMID:22113476

  16. Allele-specific cytokine responses at the HLA-C locus: implications for psoriasis.

    Hundhausen, Christian; Bertoni, Anna; Mak, Rose K; Botti, Elisabetta; Di Meglio, Paola; Clop, Alex; Laggner, Ute; Chimenti, Sergio; Hayday, Adrian C; Barker, Jonathan N; Trembath, Richard C; Capon, Francesca; Nestle, Frank O

    2012-03-01

    Psoriasis is an inflammatory skin disorder that is inherited as a complex trait. Genetic studies have repeatedly highlighted HLA-C as the major determinant for psoriasis susceptibility, with the Cw*0602 allele conferring significant disease risk in a wide range of populations. Despite the potential importance of HLA-C variation in psoriasis, either via an effect on peptide presentation or immuno-inhibitory activity, allele-specific expression patterns have not been investigated. Here, we used reporter assays to characterize two regulatory variants, which virtually abolished the response to tumor necrosis factor (TNF)-α (rs2524094) and IFN-γ (rs10657191) in HLA-Cw*0602 and a cluster of related alleles. We validated these findings through the analysis of HLA-Cw*0602 expression in primary keratinocytes treated with TNF-α and IFN-γ. Finally, we showed that HLA-Cw*0602 transcripts are not increased in psoriatic skin lesions, despite highly elevated TNF-α levels. Thus, our findings demonstrate the presence of allele-specific differences in HLA-C expression and indicate that HLA-Cw*0602 is unresponsive to upregulation by key proinflammatory cytokines in psoriasis. These data pave the way for functional studies into the pathogenic role of the major psoriasis susceptibility allele.

  17. Parallel Beam-Beam Simulation Incorporating Multiple Bunches and Multiple Interaction Regions

    Jones, F W; Pieloni, T

    2007-01-01

    The simulation code COMBI has been developed to enable the study of coherent beam-beam effects in the full collision scenario of the LHC, with multiple bunches interacting at multiple crossing points over many turns. The program structure and input are conceived in a general way which allows arbitrary numbers and placements of bunches and interaction points (IP's), together with procedural options for head-on and parasitic collisions (in the strong-strong sense), beam transport, statistics gathering, harmonic analysis, and periodic output of simulation data. The scale of this problem, once we go beyond the simplest case of a pair of bunches interacting once per turn, quickly escalates into the parallel computing arena, and herein we will describe the construction of an MPI-based version of COMBI able to utilize arbitrary numbers of processors to support efficient calculation of multi-bunch multi-IP interactions and transport. Implementing the parallel version did not require extensive disruption of the basic ...

  18. Bamgineer: Introduction of simulated allele-specific copy number variants into exome and targeted sequence data sets.

    Samadian, Soroush; Bruce, Jeff P; Pugh, Trevor J

    2018-03-01

    Somatic copy number variations (CNVs) play a crucial role in development of many human cancers. The broad availability of next-generation sequencing data has enabled the development of algorithms to computationally infer CNV profiles from a variety of data types including exome and targeted sequence data; currently the most prevalent types of cancer genomics data. However, systemic evaluation and comparison of these tools remains challenging due to a lack of ground truth reference sets. To address this need, we have developed Bamgineer, a tool written in Python to introduce user-defined haplotype-phased allele-specific copy number events into an existing Binary Alignment Mapping (BAM) file, with a focus on targeted and exome sequencing experiments. As input, this tool requires a read alignment file (BAM format), lists of non-overlapping genome coordinates for introduction of gains and losses (bed file), and an optional file defining known haplotypes (vcf format). To improve runtime performance, Bamgineer introduces the desired CNVs in parallel using queuing and parallel processing on a local machine or on a high-performance computing cluster. As proof-of-principle, we applied Bamgineer to a single high-coverage (mean: 220X) exome sequence file from a blood sample to simulate copy number profiles of 3 exemplar tumors from each of 10 tumor types at 5 tumor cellularity levels (20-100%, 150 BAM files in total). To demonstrate feasibility beyond exome data, we introduced read alignments to a targeted 5-gene cell-free DNA sequencing library to simulate EGFR amplifications at frequencies consistent with circulating tumor DNA (10, 1, 0.1 and 0.01%) while retaining the multimodal insert size distribution of the original data. We expect Bamgineer to be of use for development and systematic benchmarking of CNV calling algorithms by users using locally-generated data for a variety of applications. The source code is freely available at http://github.com/pughlab/bamgineer.

  19. Bamgineer: Introduction of simulated allele-specific copy number variants into exome and targeted sequence data sets.

    Soroush Samadian

    2018-03-01

    Full Text Available Somatic copy number variations (CNVs play a crucial role in development of many human cancers. The broad availability of next-generation sequencing data has enabled the development of algorithms to computationally infer CNV profiles from a variety of data types including exome and targeted sequence data; currently the most prevalent types of cancer genomics data. However, systemic evaluation and comparison of these tools remains challenging due to a lack of ground truth reference sets. To address this need, we have developed Bamgineer, a tool written in Python to introduce user-defined haplotype-phased allele-specific copy number events into an existing Binary Alignment Mapping (BAM file, with a focus on targeted and exome sequencing experiments. As input, this tool requires a read alignment file (BAM format, lists of non-overlapping genome coordinates for introduction of gains and losses (bed file, and an optional file defining known haplotypes (vcf format. To improve runtime performance, Bamgineer introduces the desired CNVs in parallel using queuing and parallel processing on a local machine or on a high-performance computing cluster. As proof-of-principle, we applied Bamgineer to a single high-coverage (mean: 220X exome sequence file from a blood sample to simulate copy number profiles of 3 exemplar tumors from each of 10 tumor types at 5 tumor cellularity levels (20-100%, 150 BAM files in total. To demonstrate feasibility beyond exome data, we introduced read alignments to a targeted 5-gene cell-free DNA sequencing library to simulate EGFR amplifications at frequencies consistent with circulating tumor DNA (10, 1, 0.1 and 0.01% while retaining the multimodal insert size distribution of the original data. We expect Bamgineer to be of use for development and systematic benchmarking of CNV calling algorithms by users using locally-generated data for a variety of applications. The source code is freely available at http://github.com/pughlab/bamgineer.

  20. Kinetic characterisation of primer mismatches in allele-specific PCR: a quantitative assessment.

    Waterfall, Christy M; Eisenthal, Robert; Cobb, Benjamin D

    2002-12-20

    A novel method of estimating the kinetic parameters of Taq DNA polymerase during rapid cycle PCR is presented. A model was constructed using a simplified sigmoid function to represent substrate accumulation during PCR in combination with the general equation describing high substrate inhibition for Michaelis-Menten enzymes. The PCR progress curve was viewed as a series of independent reactions where initial rates were accurately measured for each cycle. Kinetic parameters were obtained for allele-specific PCR (AS-PCR) amplification to examine the effect of mismatches on amplification. A high degree of correlation was obtained providing evidence of substrate inhibition as a major cause of the plateau phase that occurs in the later cycles of PCR.

  1. Parallel k-means++ for Multiple Shared-Memory Architectures

    Mackey, Patrick S.; Lewis, Robert R.

    2016-09-22

    In recent years k-means++ has become a popular initialization technique for improved k-means clustering. To date, most of the work done to improve its performance has involved parallelizing algorithms that are only approximations of k-means++. In this paper we present a parallelization of the exact k-means++ algorithm, with a proof of its correctness. We develop implementations for three distinct shared-memory architectures: multicore CPU, high performance GPU, and the massively multithreaded Cray XMT platform. We demonstrate the scalability of the algorithm on each platform. In addition we present a visual approach for showing which platform performed k-means++ the fastest for varying data sizes.

  2. Detection of MPL mutations by a novel allele-specific PCR-based strategy.

    Furtado, Larissa V; Weigelin, Helmut C; Elenitoba-Johnson, Kojo S J; Betz, Bryan L

    2013-11-01

    MPL mutation testing is recommended in patients with suspected primary myelofibrosis or essential thrombocythemia who lack the JAK2 V617F mutation. MPL mutations can occur at allelic levels below 15%, which may escape detection by commonly used mutation screening methods such as Sanger sequencing. We developed a novel multiplexed allele-specific PCR assay capable of detecting most recurrent MPL exon 10 mutations associated with primary myelofibrosis and essential thrombocythemia (W515L, W515K, W515A, and S505N) down to a sensitivity of 2.5% mutant allele. Test results were reviewed from 15 reference cases and 1380 consecutive specimens referred to our laboratory for testing. Assay performance was compared to Sanger sequencing across a series of 58 specimens with MPL mutations. Positive cases consisted of 45 with W515L, 6 with S505N, 5 with W515K, 1 with W515A, and 1 with both W515L and S505N. Seven cases had mutations below 5% that were undetected by Sanger sequencing. Ten additional cases had mutation levels between 5% and 15% that were not consistently detected by sequencing. All results were easily interpreted in the allele-specific test. This assay offers a sensitive and reliable solution for MPL mutation testing. Sanger sequencing appears insufficiently sensitive for robust MPL mutation detection. Our data also suggest the relative frequency of S505N mutations may be underestimated, highlighting the necessity for inclusion of this mutation in MPL test platforms. Copyright © 2013 American Society for Investigative Pathology and the Association for Molecular Pathology. Published by Elsevier Inc. All rights reserved.

  3. Allele-Specific DNA Methylation and Its Interplay with Repressive Histone Marks at Promoter-Mutant TERT Genes

    Josh Lewis Stern

    2017-12-01

    Full Text Available A mutation in the promoter of the Telomerase Reverse Transcriptase (TERT gene is the most frequent noncoding mutation in cancer. The mutation drives unusual monoallelic expression of TERT, allowing immortalization. Here, we find that DNA methylation of the TERT CpG island (CGI is also allele-specific in multiple cancers. The expressed allele is hypomethylated, which is opposite to cancers without TERT promoter mutations. The continued presence of Polycomb repressive complex 2 (PRC2 on the inactive allele suggests that histone marks of repressed chromatin may be causally linked to high DNA methylation. Consistent with this hypothesis, TERT promoter DNA containing 5-methyl-CpG has much increased affinity for PRC2 in vitro. Thus, CpG methylation and histone marks appear to collaborate to maintain the two TERT alleles in different epigenetic states in TERT promoter mutant cancers. Finally, in several cancers, DNA methylation levels at the TERT CGI correlate with altered patient survival.

  4. Allele-Specific DNA Methylation and Its Interplay with Repressive Histone Marks at Promoter-Mutant TERT Genes.

    Stern, Josh Lewis; Paucek, Richard D; Huang, Franklin W; Ghandi, Mahmoud; Nwumeh, Ronald; Costello, James C; Cech, Thomas R

    2017-12-26

    A mutation in the promoter of the Telomerase Reverse Transcriptase (TERT) gene is the most frequent noncoding mutation in cancer. The mutation drives unusual monoallelic expression of TERT, allowing immortalization. Here, we find that DNA methylation of the TERT CpG island (CGI) is also allele-specific in multiple cancers. The expressed allele is hypomethylated, which is opposite to cancers without TERT promoter mutations. The continued presence of Polycomb repressive complex 2 (PRC2) on the inactive allele suggests that histone marks of repressed chromatin may be causally linked to high DNA methylation. Consistent with this hypothesis, TERT promoter DNA containing 5-methyl-CpG has much increased affinity for PRC2 in vitro. Thus, CpG methylation and histone marks appear to collaborate to maintain the two TERT alleles in different epigenetic states in TERT promoter mutant cancers. Finally, in several cancers, DNA methylation levels at the TERT CGI correlate with altered patient survival. Copyright © 2017 The Author(s). Published by Elsevier Inc. All rights reserved.

  5. Performance Analysis of a Threshold-Based Parallel Multiple Beam Selection Scheme for WDM FSO Systems

    Nam, Sung Sik; Alouini, Mohamed-Slim; Ko, Young-Chai

    2018-01-01

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme for a free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has occurred

  6. Allele specific LAMP- gold nanoparticle for characterization of single nucleotide polymorphisms

    Fábio Ferreira Carlos

    2017-12-01

    Full Text Available Due to their relevance as disease biomarkers and for diagnostics, screening of single nucleotide polymorphism (SNPs requires simple and straightforward strategies capable to provide results in medium throughput settings. Suitable approaches relying on isothermal amplification techniques have been evolving to substitute the cumbersome and highly specialized PCR amplification detection schemes. Nonetheless, identification of an individual’s genotype still requires sophisticated equipment and laborious methods.Here, we present a low-cost and reliable approach based on the allele specific loop-mediated isothermal amplification (AS-LAMP coupled to ssDNA functionalized gold nanoparticle (Au-nanoprobe colorimetric sequence discrimination. The Au-nanoprobe integration allows for the colorimetric detection of AS-LAMP amplification product that can be easily interpreted in less than 15 min. We targeted a clinical relevant SNP responsible for lactose intolerance (-13910C/T dbSNP rs#: 4988235 to demonstrate its proof of concept and full potential of this novel approach. Keywords: SNP, Isothermal amplification, Gold nanoparticles, Gold nanoprobes, Lactose intolerance

  7. [Molecular authentication of Jinyinhua formula granule by using allele-specific PCR].

    Jiang, Chao; Tu, Li-Chan; Yuan, Yuan; Huang, Lu-Qi; Gao, Wei; Jin, Yan

    2017-07-01

    Traditional authentication method is hard to identify herb's authenticity of traditional Chinese medicine(TCM) formula granules because they have lost all their morphological characteristics. In this study, a new allele-specific PCR method was established for identifying the authentication of Jinyinhua formula granule (made from Lonicerae Japonicae Flos) based on an SNP site in trnL-trnF fragment. Genomic DNA was successfully extracted from Lonicerae Japonicae Flos and its formula granules by using an improved spin column method and then PCR was performed with the designed primer. Approximately 110 bp specific bands was obtained only in the authentic Lonicerae Japonicae Flos and its formula granules, while no bands were found in fake mixed products. In addition, the PCR product sequence was proved from Lonicerae Japonicae Flos trnL-trnF sequence by using BLAST method. Therefore, DNA molecular authentication method could make up the limitations of character identification method and microscopic identification, and quickly identify herb's authenticity of TCM formula granules, with enormous potential for market supervision and quality control. Copyright© by the Chinese Pharmaceutical Association.

  8. Allele-specific marker generation and linkage mapping on the Xiphophorus sex chromosomes.

    Woolcock, B; Kazianis, S; Lucito, R; Walter, R B; Kallman, K D; Morizot, D C; Vielkind, J R

    2006-01-01

    There is great interest in the sex chromosomes of Xiphophorus fishes because both WY/YY and XX/XY sex-determining mechanisms function in these species, with at least one taxon possessing all three types of sex chromosomes, and because in certain interspecific hybrids melanoma arises as a consequence of inheritance of the sex-linked macromelanophore determining locus (MDL). Representational difference analysis (RDA) has been used to clone two sequences from the sex-determining region of X. maculatus, including a cholinergic receptor, nicotinic, delta polypeptide (CHRND) orthologue. Allele-specific assays for these sequences, as well as for the sex-linked XMRK1 and XMRK2 genes, were developed to distinguish W, X, and Y chromosomes derived from a X. maculatus (XX/XY) strain and a X. helleri (WY/YY) strain. Linkage mapping localized these markers to linkage group (LG) 24. No recombinants were observed between XMRK2 and MDL, confirming a role for XMRK2 in macromelanophore development. Although the master sex-determining (SD) locus certainly resides on Xiphophorus LG 24, autosomal loci are probably involved in sex determination as well, as indicated by the abnormal sex ratios in the backcross hybrids that contrast theoretical predictions based on LG 24 genotyping. Marker development and allelic discrimination on the Xiphophorus sex chromosomes should prove highly useful for studies that utilize this genus as an animal model.

  9. Allele-specific gene expression in a wild nonhuman primate population

    Tung, J.; Akinyi, M. Y.; Mutura, S.; Altmann, J.; Wray, G. A.; Alberts, S. C.

    2015-01-01

    Natural populations hold enormous potential for evolutionary genetic studies, especially when phenotypic, genetic and environmental data are all available on the same individuals. However, untangling the genotype-phenotype relationship in natural populations remains a major challenge. Here, we describe results of an investigation of one class of phenotype, allele-specific gene expression (ASGE), in the well-studied natural population of baboons of the Amboseli basin, Kenya. ASGE measurements identify cases in which one allele of a gene is overexpressed relative to the alternative allele of the same gene, within individuals, thus providing a control for background genetic and environmental effects. Here, we characterize the incidence of ASGE in the Amboseli baboon population, focusing on the genetic and environmental contributions to ASGE in a set of eleven genes involved in immunity and defence. Within this set, we identify evidence for common ASGE in four genes. We also present examples of two relationships between cis-regulatory genetic variants and the ASGE phenotype. Finally, we identify one case in which this relationship is influenced by a novel gene-environment interaction. Specifically, the dominance rank of an individual’s mother during its early life (an aspect of that individual’s social environment) influences the expression of the gene CCL5 via an interaction with cis-regulatory genetic variation. These results illustrate how environmental and ecological data can be integrated into evolutionary genetic studies of functional variation in natural populations. They also highlight the potential importance of early life environmental variation in shaping the genetic architecture of complex traits in wild mammals. PMID:21226779

  10. An informational view of accession rarity and allele specificity in germplasm banks for management and conservation.

    Reyes-Valdés, M Humberto; Burgueño, Juan; Singh, Sukhwinder; Martínez, Octavio; Sansaloni, Carolina Paola

    2018-01-01

    Germplasm banks are growing in their importance, number of accessions and amount of characterization data, with a large emphasis on molecular genetic markers. In this work, we offer an integrated view of accessions and marker data in an information theory framework. The basis of this development is the mutual information between accessions and allele frequencies for molecular marker loci, which can be decomposed in allele specificities, as well as in rarity and divergence of accessions. In this way, formulas are provided to calculate the specificity of the different marker alleles with reference to their distribution across accessions, accession rarity, defined as the weighted average of the specificity of its alleles, and divergence, defined by the Kullback-Leibler formula. Albeit being different measures, it is demonstrated that average rarity and divergence are equal for any collection. These parameters can contribute to the knowledge of the structure of a germplasm collection and to make decisions about the preservation of rare variants. The concepts herein developed served as the basis for a strategy for core subset selection called HCore, implemented in a publicly available R script. As a proof of concept, the mathematical view and tools developed in this research were applied to a large collection of Mexican wheat accessions, widely characterized by SNP markers. The most specific alleles were found to be private of a single accession, and the distribution of this parameter had its highest frequencies at low levels of specificity. Accession rarity and divergence had largely symmetrical distributions, and had a positive, albeit non-strictly linear relationship. Comparison of the HCore approach for core subset selection, with three state-of-the-art methods, showed it to be superior for average divergence and rarity, mean genetic distance and diversity. The proposed approach can be used for knowledge extraction and decision making in germplasm collections of

  11. Hierarchical approach to optimization of parallel matrix multiplication on large-scale platforms

    Hasanov, Khalid

    2014-03-04

    © 2014, Springer Science+Business Media New York. Many state-of-the-art parallel algorithms, which are widely used in scientific applications executed on high-end computing systems, were designed in the twentieth century with relatively small-scale parallelism in mind. Indeed, while in 1990s a system with few hundred cores was considered a powerful supercomputer, modern top supercomputers have millions of cores. In this paper, we present a hierarchical approach to optimization of message-passing parallel algorithms for execution on large-scale distributed-memory systems. The idea is to reduce the communication cost by introducing hierarchy and hence more parallelism in the communication scheme. We apply this approach to SUMMA, the state-of-the-art parallel algorithm for matrix–matrix multiplication, and demonstrate both theoretically and experimentally that the modified Hierarchical SUMMA significantly improves the communication cost and the overall performance on large-scale platforms.

  12. Allele-specific DNA methylation of disease susceptibility genes in Japanese patients with inflammatory bowel disease.

    Chiba, Hirofumi; Kakuta, Yoichi; Kinouchi, Yoshitaka; Kawai, Yosuke; Watanabe, Kazuhiro; Nagao, Munenori; Naito, Takeo; Onodera, Motoyuki; Moroi, Rintaro; Kuroha, Masatake; Kanazawa, Yoshitake; Kimura, Tomoya; Shiga, Hisashi; Endo, Katsuya; Negoro, Kenichi; Nagasaki, Masao; Unno, Michiaki; Shimosegawa, Tooru

    2018-01-01

    Inflammatory bowel disease (IBD) has an unknown etiology; however, accumulating evidence suggests that IBD is a multifactorial disease influenced by a combination of genetic and environmental factors. The influence of genetic variants on DNA methylation in cis and cis effects on expression have been demonstrated. We hypothesized that IBD susceptibility single-nucleotide polymorphisms (SNPs) regulate susceptibility gene expressions in cis by regulating DNA methylation around SNPs. For this, we determined cis-regulated allele-specific DNA methylation (ASM) around IBD susceptibility genes in CD4+ effector/memory T cells (Tem) in lamina propria mononuclear cells (LPMCs) in patients with IBD and examined the association between the ASM SNP genotype and neighboring susceptibility gene expressions. CD4+ effector/memory T cells (Tem) were isolated from LPMCs in 15 Japanese IBD patients (ten Crohn's disease [CD] and five ulcerative colitis [UC] patients). ASM analysis was performed by methylation-sensitive SNP array analysis. We defined ASM as a changing average relative allele score ([Formula: see text]) >0.1 after digestion by methylation-sensitive restriction enzymes. Among SNPs showing [Formula: see text] >0.1, we extracted the probes located on tag-SNPs of 200 IBD susceptibility loci and around IBD susceptibility genes as candidate ASM SNPs. To validate ASM, bisulfite-pyrosequencing was performed. Transcriptome analysis was examined in 11 IBD patients (seven CD and four UC patients). The relation between rs36221701 genotype and neighboring gene expressions were analyzed. We extracted six candidate ASM SNPs around IBD susceptibility genes. The top of [Formula: see text] (0.23) was rs1130368 located on HLA-DQB1. ASM around rs36221701 ([Formula: see text] = 0.14) located near SMAD3 was validated using bisulfite pyrosequencing. The SMAD3 expression was significantly associated with the rs36221701 genotype (p = 0.016). We confirmed the existence of cis-regulated ASM around

  13. Allele-specific DNA methylation of disease susceptibility genes in Japanese patients with inflammatory bowel disease

    Chiba, Hirofumi; Kakuta, Yoichi; Kinouchi, Yoshitaka; Kawai, Yosuke; Watanabe, Kazuhiro; Nagao, Munenori; Naito, Takeo; Onodera, Motoyuki; Moroi, Rintaro; Kuroha, Masatake; Kanazawa, Yoshitake; Kimura, Tomoya; Shiga, Hisashi; Endo, Katsuya; Negoro, Kenichi; Nagasaki, Masao; Unno, Michiaki; Shimosegawa, Tooru

    2018-01-01

    Background Inflammatory bowel disease (IBD) has an unknown etiology; however, accumulating evidence suggests that IBD is a multifactorial disease influenced by a combination of genetic and environmental factors. The influence of genetic variants on DNA methylation in cis and cis effects on expression have been demonstrated. We hypothesized that IBD susceptibility single-nucleotide polymorphisms (SNPs) regulate susceptibility gene expressions in cis by regulating DNA methylation around SNPs. For this, we determined cis-regulated allele-specific DNA methylation (ASM) around IBD susceptibility genes in CD4+ effector/memory T cells (Tem) in lamina propria mononuclear cells (LPMCs) in patients with IBD and examined the association between the ASM SNP genotype and neighboring susceptibility gene expressions. Methods CD4+ effector/memory T cells (Tem) were isolated from LPMCs in 15 Japanese IBD patients (ten Crohn's disease [CD] and five ulcerative colitis [UC] patients). ASM analysis was performed by methylation-sensitive SNP array analysis. We defined ASM as a changing average relative allele score (ΔRAS¯) >0.1 after digestion by methylation-sensitive restriction enzymes. Among SNPs showing ΔRAS¯ >0.1, we extracted the probes located on tag-SNPs of 200 IBD susceptibility loci and around IBD susceptibility genes as candidate ASM SNPs. To validate ASM, bisulfite-pyrosequencing was performed. Transcriptome analysis was examined in 11 IBD patients (seven CD and four UC patients). The relation between rs36221701 genotype and neighboring gene expressions were analyzed. Results We extracted six candidate ASM SNPs around IBD susceptibility genes. The top of ΔRAS¯ (0.23) was rs1130368 located on HLA-DQB1. ASM around rs36221701 (ΔRAS¯ = 0.14) located near SMAD3 was validated using bisulfite pyrosequencing. The SMAD3 expression was significantly associated with the rs36221701 genotype (p = 0.016). Conclusions We confirmed the existence of cis-regulated ASM around IBD

  14. Aperture and counting rate of rectangular telescopes for single and multiple parallel particles. [Spark chamber telescopes

    D' Ettorre Piazzoli, B; Mannocchi, G [Consiglio Nazionale delle Ricerche, Turin (Italy). Lab. di Cosmo-Geofisica; Melone, S [Istituto di Fisica dell' Universita, Ancona, Italy; Picchi, P; Visentin, R [Comitato Nazionale per l' Energia Nucleare, Frascati (Italy). Laboratori Nazionali di Frascati

    1976-06-01

    Expressions for the counting rate of rectangular telescopes in the case of single as well as multiple particles are given. The aperture for single particles is obtained in the form of a double integral and analytical solutions are given for some cases. The intensity for different multiplicities of parallel particles is related to the geometry of the detectors and to the features of the radiation. This allows an absolute comparison between the data recorded by different devices.

  15. Further exploration of antimicrobial ketodihydronicotinic acid derivatives by multiple parallel syntheses

    Laursen, Jane B.; Nielsen, Janne; Haack, T.

    2006-01-01

    A synthetic reexamination of a series of ketodihydronicotinic acid class antibacterial agents was undertaken in an attempt to improve their therapeutic potential. A convenient new synthesis was developed involving hetero Diels-Alder chemistry producing 74 new analogs in a multiple parallel synthe...

  16. Modelling and simulation of multiple single - phase induction motor in parallel connection

    Sujitjorn, S.

    2006-11-01

    Full Text Available A mathematical model for parallel connected n-multiple single-phase induction motors in generalized state-space form is proposed in this paper. The motor group draws electric power from one inverter. The model is developed by the dq-frame theory and was tested against four loading scenarios in which satisfactory results were obtained.

  17. Impact of pre-existing MSP142-allele specific immunity on potency of an erythrocytic Plasmodium falciparum vaccine

    Bergmann-Leitner Elke S

    2012-09-01

    Full Text Available Abstract Background MSP1 is the major surface protein on merozoites and a prime candidate for a blood stage malaria vaccine. Preclinical and seroepidemiological studies have implicated antibodies to MSP1 in protection against blood stage parasitaemia and/or reduced parasite densities, respectively. Malaria endemic areas have multiple strains of Plasmodium falciparum circulating at any given time, giving rise to complex immune responses, an issue which is generally not addressed in clinical trials conducted in non-endemic areas. A lack of understanding of the effect of pre-existing immunity to heterologous parasite strains may significantly contribute to vaccine failure in the field. The purpose of this study was to model the effect of pre-existing immunity to MSP142 on the immunogenicity of blood-stage malaria vaccines based on alternative MSP1 alleles. Methods Inbred and outbred mice were immunized with various recombinant P. falciparum MSP142 proteins that represent the two major alleles of MSP142, MAD20 (3D7 and Wellcome (K1, FVO. Humoral immune responses were analysed by ELISA and LuminexTM, and functional activity of induced MSP142-specific antibodies was assessed by growth inhibition assays. T-cell responses were characterized using ex vivo ELISpot assays. Results Analysis of the immune responses induced by various immunization regimens demonstrated a strong allele-specific response at the T cell level in both inbred and outbred mice. The success of heterologous regimens depended on the degree of homology of the N-terminal p33 portion of the MSP142, likely due to the fact that most T cell epitopes reside in this part of the molecule. Analysis of humoral immune responses revealed a marked cross-reactivity between the alleles. Functional analyses showed that some of the heterologous regimens induced antibodies with improved growth inhibitory activities. Conclusion The development of a more broadly efficacious MSP1 based vaccine may be

  18. Allele-specific expression in the germline of patients with familial pancreatic cancer: An unbiased approach to cancer gene discovery

    Tan, Aik Choon; Fan, Jian-Bing; Karikari, Collins; Bibikova, Marina; Garcia, Eliza Wickham; Zhou, Lixin; Barker, David; Serre, David; Feldmann, Georg; Hruban, Ralph H.; Klein, Alison P.; Goggins, Michael; Couch, Fergus J.; Hudson, Thomas J.; Winslow, Raimond L.

    2007-01-01

    Physiologic allele-specific expression (ASE) in germline tissues occurs during random X-chromosome inactivation1 and in genomic imprinting,2 wherein the two alleles of a gene in a heterozygous individual are not expressed equally. Recent studies have confirmed the existence of ASE in apparently non-imprinted autosomal genes;3–14 however, the extent of ASE in the human genome is unknown. We explored ASE in lymphoblastoid cell lines of 145 individuals using an oligonucleotide array based assay....

  19. Identification and Genetic Diversity of Etambutol Resistant Strains of Mycobacterium Tuberculosis by Allelic-Specific PCR and Spologiotyping

    Zahra Derakhshani Nezhad

    2012-09-01

    Full Text Available Background & Objectives: Ethambutol is one of the four main drugs in treatment of tuberculosis. The most common mutation associated with this drug resistance usually occurs in codon 306 of embB. The aim of this study was to detect ethambutol resistance using Allele-Specific PCR and Spoligotyping in various subtypes of Mycobacterium tuberculosis.   Methods : 140 sputum specimens were collected from suspected TB patients. They were digested and decontaminated using Pettrof method before culturing them on LJ medium. Drug susceptibility testing was performed on 106 culture positive specimens using proportional method. DNA was extracted from the isolated organisms and subsequently subjected to Allele-Specific PCR to detect any mutationin embB306. Spoligotyping was then used to determine the subtypes.   Results: Out of 106 cultures positive samples, 36 samples (33.9% showed resistance to ethambutol using proportional method. Allele-Specific PCR assay identified 93 as sensitive and 13 (27.6% as resistant strains. The results of PCR were in agreement with result of proportional method. The PCR method revealed that 61.5% of mutation occurred in the first and 38.5% in third nucleotides. Spoligotyping differentiated Mycobacterium tuberculosis strains into Beijing (10 9.4%, Bovis (2 1.8%, CAS (24 22.6%, EAI (1 0.9%, Haarlem (27 25.4%, LAM (5 4.7%, Manu (5 4.7%, T (27 25.4% and U( 2 1,8% families. The high frequency of mutation in embB gene was belonged to Haarlem, CAS and T subfamilies.   Conclusion: Based on results current study, mutations in the genes other than embB might have occurred in the resistant strains that gave negative result in Allele-Specific PCR assay. Therefore other mechanisms of resistance to this antibiotic should be investigated.

  20. Harmonic resonance assessment of multiple paralleled grid-connected inverters system

    Wang, Yanbo; Wang, Xiongfei; Blaabjerg, Frede

    2017-01-01

    This paper presents an eigenvalue-based impedance stability analytical method of multiple paralleled grid-connected inverter system. Different from the conventional impedance-based stability criterion, this work first built the state-space model of paralleled grid-connected inverters. On the basis...... of this, a bridge between the state-space-based modelling and impedance-based stability criterion is presented. The proposed method is able to perform stability assessment locally at the connection points of the component. Meanwhile, the eigenvalue-based sensitivity analysis is adopted to identify...

  1. Hierarchical Parallel Matrix Multiplication on Large-Scale Distributed Memory Platforms

    Quintin, Jean-Noel

    2013-10-01

    Matrix multiplication is a very important computation kernel both in its own right as a building block of many scientific applications and as a popular representative for other scientific applications. Cannon\\'s algorithm which dates back to 1969 was the first efficient algorithm for parallel matrix multiplication providing theoretically optimal communication cost. However this algorithm requires a square number of processors. In the mid-1990s, the SUMMA algorithm was introduced. SUMMA overcomes the shortcomings of Cannon\\'s algorithm as it can be used on a nonsquare number of processors as well. Since then the number of processors in HPC platforms has increased by two orders of magnitude making the contribution of communication in the overall execution time more significant. Therefore, the state of the art parallel matrix multiplication algorithms should be revisited to reduce the communication cost further. This paper introduces a new parallel matrix multiplication algorithm, Hierarchical SUMMA (HSUMMA), which is a redesign of SUMMA. Our algorithm reduces the communication cost of SUMMA by introducing a two-level virtual hierarchy into the two-dimensional arrangement of processors. Experiments on an IBM BlueGene/P demonstrate the reduction of communication cost up to 2.08 times on 2048 cores and up to 5.89 times on 16384 cores. © 2013 IEEE.

  2. Hierarchical Parallel Matrix Multiplication on Large-Scale Distributed Memory Platforms

    Quintin, Jean-Noel; Hasanov, Khalid; Lastovetsky, Alexey

    2013-01-01

    Matrix multiplication is a very important computation kernel both in its own right as a building block of many scientific applications and as a popular representative for other scientific applications. Cannon's algorithm which dates back to 1969 was the first efficient algorithm for parallel matrix multiplication providing theoretically optimal communication cost. However this algorithm requires a square number of processors. In the mid-1990s, the SUMMA algorithm was introduced. SUMMA overcomes the shortcomings of Cannon's algorithm as it can be used on a nonsquare number of processors as well. Since then the number of processors in HPC platforms has increased by two orders of magnitude making the contribution of communication in the overall execution time more significant. Therefore, the state of the art parallel matrix multiplication algorithms should be revisited to reduce the communication cost further. This paper introduces a new parallel matrix multiplication algorithm, Hierarchical SUMMA (HSUMMA), which is a redesign of SUMMA. Our algorithm reduces the communication cost of SUMMA by introducing a two-level virtual hierarchy into the two-dimensional arrangement of processors. Experiments on an IBM BlueGene/P demonstrate the reduction of communication cost up to 2.08 times on 2048 cores and up to 5.89 times on 16384 cores. © 2013 IEEE.

  3. SPEEDES - A multiple-synchronization environment for parallel discrete-event simulation

    Steinman, Jeff S.

    1992-01-01

    Synchronous Parallel Environment for Emulation and Discrete-Event Simulation (SPEEDES) is a unified parallel simulation environment. It supports multiple-synchronization protocols without requiring users to recompile their code. When a SPEEDES simulation runs on one node, all the extra parallel overhead is removed automatically at run time. When the same executable runs in parallel, the user preselects the synchronization algorithm from a list of options. SPEEDES currently runs on UNIX networks and on the California Institute of Technology/Jet Propulsion Laboratory Mark III Hypercube. SPEEDES also supports interactive simulations. Featured in the SPEEDES environment is a new parallel synchronization approach called Breathing Time Buckets. This algorithm uses some of the conservative techniques found in Time Bucket synchronization, along with the optimism that characterizes the Time Warp approach. A mathematical model derived from first principles predicts the performance of Breathing Time Buckets. Along with the Breathing Time Buckets algorithm, this paper discusses the rules for processing events in SPEEDES, describes the implementation of various other synchronization protocols supported by SPEEDES, describes some new ones for the future, discusses interactive simulations, and then gives some performance results.

  4. Design of multiple sequence alignment algorithms on parallel, distributed memory supercomputers.

    Church, Philip C; Goscinski, Andrzej; Holt, Kathryn; Inouye, Michael; Ghoting, Amol; Makarychev, Konstantin; Reumann, Matthias

    2011-01-01

    The challenge of comparing two or more genomes that have undergone recombination and substantial amounts of segmental loss and gain has recently been addressed for small numbers of genomes. However, datasets of hundreds of genomes are now common and their sizes will only increase in the future. Multiple sequence alignment of hundreds of genomes remains an intractable problem due to quadratic increases in compute time and memory footprint. To date, most alignment algorithms are designed for commodity clusters without parallelism. Hence, we propose the design of a multiple sequence alignment algorithm on massively parallel, distributed memory supercomputers to enable research into comparative genomics on large data sets. Following the methodology of the sequential progressiveMauve algorithm, we design data structures including sequences and sorted k-mer lists on the IBM Blue Gene/P supercomputer (BG/P). Preliminary results show that we can reduce the memory footprint so that we can potentially align over 250 bacterial genomes on a single BG/P compute node. We verify our results on a dataset of E.coli, Shigella and S.pneumoniae genomes. Our implementation returns results matching those of the original algorithm but in 1/2 the time and with 1/4 the memory footprint for scaffold building. In this study, we have laid the basis for multiple sequence alignment of large-scale datasets on a massively parallel, distributed memory supercomputer, thus enabling comparison of hundreds instead of a few genome sequences within reasonable time.

  5. Use of allele-specific FAIRE to determine functional regulatory polymorphism using large-scale genotyping arrays.

    Andrew J P Smith

    Full Text Available Following the widespread use of genome-wide association studies (GWAS, focus is turning towards identification of causal variants rather than simply genetic markers of diseases and traits. As a step towards a high-throughput method to identify genome-wide, non-coding, functional regulatory variants, we describe the technique of allele-specific FAIRE, utilising large-scale genotyping technology (FAIRE-gen to determine allelic effects on chromatin accessibility and regulatory potential. FAIRE-gen was explored using lymphoblastoid cells and the 50,000 SNP Illumina CVD BeadChip. The technique identified an allele-specific regulatory polymorphism within NR1H3 (coding for LXR-α, rs7120118, coinciding with a previously GWAS-identified SNP for HDL-C levels. This finding was confirmed using FAIRE-gen with the 200,000 SNP Illumina Metabochip and verified with the established method of TaqMan allelic discrimination. Examination of this SNP in two prospective Caucasian cohorts comprising 15,000 individuals confirmed the association with HDL-C levels (combined beta = 0.016; p = 0.0006, and analysis of gene expression identified an allelic association with LXR-α expression in heart tissue. Using increasingly comprehensive genotyping chips and distinct tissues for examination, FAIRE-gen has the potential to aid the identification of many causal SNPs associated with disease from GWAS.

  6. Design and Analysis of Cooperative Cable Parallel Manipulators for Multiple Mobile Cranes

    Bin Zi

    2012-11-01

    Full Text Available The design, dynamic modelling, and workspace are presented in this paper concerning cooperative cable parallel manipulators for multiple mobile cranes (CPMMCs. The CPMMCs can handle complex tasks that are more difficult or even impossible for a single mobile crane. Kinematics and dynamics of the CPMMCs are studied on the basis of geometric methodology and d'Alembert's principle, and a mathematical model of the CPMMCs is developed and presented with dynamic simulation. The constant orientation workspace analysis of the CPMMCs is carried out additionally. As an example, a cooperative cable parallel manipulator for triple mobile cranes with 6 Degrees of Freedom is investigated on the basis of the above design objectives.

  7. Prioritizing multiple therapeutic targets in parallel using automated DNA-encoded library screening

    Machutta, Carl A.; Kollmann, Christopher S.; Lind, Kenneth E.; Bai, Xiaopeng; Chan, Pan F.; Huang, Jianzhong; Ballell, Lluis; Belyanskaya, Svetlana; Besra, Gurdyal S.; Barros-Aguirre, David; Bates, Robert H.; Centrella, Paolo A.; Chang, Sandy S.; Chai, Jing; Choudhry, Anthony E.; Coffin, Aaron; Davie, Christopher P.; Deng, Hongfeng; Deng, Jianghe; Ding, Yun; Dodson, Jason W.; Fosbenner, David T.; Gao, Enoch N.; Graham, Taylor L.; Graybill, Todd L.; Ingraham, Karen; Johnson, Walter P.; King, Bryan W.; Kwiatkowski, Christopher R.; Lelièvre, Joël; Li, Yue; Liu, Xiaorong; Lu, Quinn; Lehr, Ruth; Mendoza-Losana, Alfonso; Martin, John; McCloskey, Lynn; McCormick, Patti; O'Keefe, Heather P.; O'Keeffe, Thomas; Pao, Christina; Phelps, Christopher B.; Qi, Hongwei; Rafferty, Keith; Scavello, Genaro S.; Steiginga, Matt S.; Sundersingh, Flora S.; Sweitzer, Sharon M.; Szewczuk, Lawrence M.; Taylor, Amy; Toh, May Fern; Wang, Juan; Wang, Minghui; Wilkins, Devan J.; Xia, Bing; Yao, Gang; Zhang, Jean; Zhou, Jingye; Donahue, Christine P.; Messer, Jeffrey A.; Holmes, David; Arico-Muendel, Christopher C.; Pope, Andrew J.; Gross, Jeffrey W.; Evindar, Ghotas

    2017-07-01

    The identification and prioritization of chemically tractable therapeutic targets is a significant challenge in the discovery of new medicines. We have developed a novel method that rapidly screens multiple proteins in parallel using DNA-encoded library technology (ELT). Initial efforts were focused on the efficient discovery of antibacterial leads against 119 targets from Acinetobacter baumannii and Staphylococcus aureus. The success of this effort led to the hypothesis that the relative number of ELT binders alone could be used to assess the ligandability of large sets of proteins. This concept was further explored by screening 42 targets from Mycobacterium tuberculosis. Active chemical series for six targets from our initial effort as well as three chemotypes for DHFR from M. tuberculosis are reported. The findings demonstrate that parallel ELT selections can be used to assess ligandability and highlight opportunities for successful lead and tool discovery.

  8. DIALIGN P: Fast pair-wise and multiple sequence alignment using parallel processors

    Kaufmann Michael

    2004-09-01

    Full Text Available Abstract Background Parallel computing is frequently used to speed up computationally expensive tasks in Bioinformatics. Results Herein, a parallel version of the multi-alignment program DIALIGN is introduced. We propose two ways of dividing the program into independent sub-routines that can be run on different processors: (a pair-wise sequence alignments that are used as a first step to multiple alignment account for most of the CPU time in DIALIGN. Since alignments of different sequence pairs are completely independent of each other, they can be distributed to multiple processors without any effect on the resulting output alignments. (b For alignments of large genomic sequences, we use a heuristics by splitting up sequences into sub-sequences based on a previously introduced anchored alignment procedure. For our test sequences, this combined approach reduces the program running time of DIALIGN by up to 97%. Conclusions By distributing sub-routines to multiple processors, the running time of DIALIGN can be crucially improved. With these improvements, it is possible to apply the program in large-scale genomics and proteomics projects that were previously beyond its scope.

  9. Cubic systems with invariant affine straight lines of total parallel multiplicity seven

    Alexandru Suba

    2013-12-01

    Full Text Available In this article, we study the planar cubic differential systems with invariant affine straight lines of total parallel multiplicity seven. We classify these system according to their geometric properties encoded in the configurations of invariant straight lines. We show that there are only 17 different topological phase portraits in the Poincar\\'e disc associated to this family of cubic systems up to a reversal of the sense of their orbits, and we provide representatives of every class modulo an affine change of variables and rescaling of the time variable.

  10. Identical parallel machine scheduling with nonlinear deterioration and multiple rate modifying activities

    Ömer Öztürkoğlu

    2017-07-01

    Full Text Available This study focuses on identical parallel machine scheduling of jobs with deteriorating processing times and rate-modifying activities. We consider non linearly increasing processing times of jobs based on their position assignment. Rate modifying activities are also considered to recover the increase in processing times of jobs due to deterioration. We also propose heuristics algorithms that rely on ant colony optimization and simulated annealing algorithms to solve the problem with multiple RMAs in a reasonable amount of time. Finally, we show that ant colony optimization algorithm generates close optimal solutions and superior results than simulated annealing algorithm.

  11. Parallel point-multiplication architecture using combined group operations for high-speed cryptographic applications.

    Md Selim Hossain

    Full Text Available In this paper, we propose a novel parallel architecture for fast hardware implementation of elliptic curve point multiplication (ECPM, which is the key operation of an elliptic curve cryptography processor. The point multiplication over binary fields is synthesized on both FPGA and ASIC technology by designing fast elliptic curve group operations in Jacobian projective coordinates. A novel combined point doubling and point addition (PDPA architecture is proposed for group operations to achieve high speed and low hardware requirements for ECPM. It has been implemented over the binary field which is recommended by the National Institute of Standards and Technology (NIST. The proposed ECPM supports two Koblitz and random curves for the key sizes 233 and 163 bits. For group operations, a finite-field arithmetic operation, e.g. multiplication, is designed on a polynomial basis. The delay of a 233-bit point multiplication is only 3.05 and 3.56 μs, in a Xilinx Virtex-7 FPGA, for Koblitz and random curves, respectively, and 0.81 μs in an ASIC 65-nm technology, which are the fastest hardware implementation results reported in the literature to date. In addition, a 163-bit point multiplication is also implemented in FPGA and ASIC for fair comparison which takes around 0.33 and 0.46 μs, respectively. The area-time product of the proposed point multiplication is very low compared to similar designs. The performance ([Formula: see text] and Area × Time × Energy (ATE product of the proposed design are far better than the most significant studies found in the literature.

  12. SAAS-CNV: A Joint Segmentation Approach on Aggregated and Allele Specific Signals for the Identification of Somatic Copy Number Alterations with Next-Generation Sequencing Data.

    Zhang, Zhongyang; Hao, Ke

    2015-11-01

    Cancer genomes exhibit profound somatic copy number alterations (SCNAs). Studying tumor SCNAs using massively parallel sequencing provides unprecedented resolution and meanwhile gives rise to new challenges in data analysis, complicated by tumor aneuploidy and heterogeneity as well as normal cell contamination. While the majority of read depth based methods utilize total sequencing depth alone for SCNA inference, the allele specific signals are undervalued. We proposed a joint segmentation and inference approach using both signals to meet some of the challenges. Our method consists of four major steps: 1) extracting read depth supporting reference and alternative alleles at each SNP/Indel locus and comparing the total read depth and alternative allele proportion between tumor and matched normal sample; 2) performing joint segmentation on the two signal dimensions; 3) correcting the copy number baseline from which the SCNA state is determined; 4) calling SCNA state for each segment based on both signal dimensions. The method is applicable to whole exome/genome sequencing (WES/WGS) as well as SNP array data in a tumor-control study. We applied the method to a dataset containing no SCNAs to test the specificity, created by pairing sequencing replicates of a single HapMap sample as normal/tumor pairs, as well as a large-scale WGS dataset consisting of 88 liver tumors along with adjacent normal tissues. Compared with representative methods, our method demonstrated improved accuracy, scalability to large cancer studies, capability in handling both sequencing and SNP array data, and the potential to improve the estimation of tumor ploidy and purity.

  13. Vipie: web pipeline for parallel characterization of viral populations from multiple NGS samples.

    Lin, Jake; Kramna, Lenka; Autio, Reija; Hyöty, Heikki; Nykter, Matti; Cinek, Ondrej

    2017-05-15

    Next generation sequencing (NGS) technology allows laboratories to investigate virome composition in clinical and environmental samples in a culture-independent way. There is a need for bioinformatic tools capable of parallel processing of virome sequencing data by exactly identical methods: this is especially important in studies of multifactorial diseases, or in parallel comparison of laboratory protocols. We have developed a web-based application allowing direct upload of sequences from multiple virome samples using custom parameters. The samples are then processed in parallel using an identical protocol, and can be easily reanalyzed. The pipeline performs de-novo assembly, taxonomic classification of viruses as well as sample analyses based on user-defined grouping categories. Tables of virus abundance are produced from cross-validation by remapping the sequencing reads to a union of all observed reference viruses. In addition, read sets and reports are created after processing unmapped reads against known human and bacterial ribosome references. Secured interactive results are dynamically plotted with population and diversity charts, clustered heatmaps and a sortable and searchable abundance table. The Vipie web application is a unique tool for multi-sample metagenomic analysis of viral data, producing searchable hits tables, interactive population maps, alpha diversity measures and clustered heatmaps that are grouped in applicable custom sample categories. Known references such as human genome and bacterial ribosomal genes are optionally removed from unmapped ('dark matter') reads. Secured results are accessible and shareable on modern browsers. Vipie is a freely available web-based tool whose code is open source.

  14. Practical enhancement factor model based on GM for multiple parallel reactions: Piperazine (PZ) CO2 capture

    Gaspar, Jozsef; Fosbøl, Philip Loldrup

    2017-01-01

    Reactive absorption is a key process for gas separation and purification and it is the main technology for CO2 capture. Thus, reliable and simple mathematical models for mass transfer rate calculation are essential. Models which apply to parallel interacting and non-interacting reactions, for all......, desorption and pinch conditions.In this work, we apply the GM model to multiple parallel reactions. We deduce the model for piperazine (PZ) CO2 capture and we validate it against wetted-wall column measurements using 2, 5 and 8 molal PZ for temperatures between 40 °C and 100 °C and CO2 loadings between 0.......23 and 0.41 mol CO2/2 mol PZ. We show that overall second order kinetics describes well the reaction between CO2 and PZ accounting for the carbamate and bicarbamate reactions. Here we prove the GM model for piperazine and MEA but we expect that this practical approach is applicable for various amines...

  15. Tunable multiple plasmon induced transparencies in parallel graphene sheets and its applications

    khazaee, Sara; Granpayeh, Nosrat

    2018-01-01

    Tunable plasmon induced transparency is achieved by using only two parallel graphene sheets beyond silicon diffractive grating in mid-infrared region. Excitation of the guided-wave resonance (GWR) in this structure is illustrated on the normal incident transmission spectra and plays the bright resonance mode role. Weak hybridization between two bright modes, creates plasmon induced transparency (PIT) optical response. The resonance frequency of transparency window can be tuned by different geometrical parameters. Also, variation of graphene Fermi energy can be used to achieve tunability of the resonance frequency of transparency window without reconstruction and re-fabrication of the structure. We demonstrate the existence of multiple PIT spectral responses resulting from a series of self-assembled GWRs to be used as the wavelength demultiplexer. This study can be used for design of the optical ultra-compact devices and photonic integrated circuits.

  16. Scattering by multiple parallel radially stratified infinite cylinders buried in a lossy half space.

    Lee, Siu-Chun

    2013-07-01

    The theoretical solution for scattering by an arbitrary configuration of closely spaced parallel infinite cylinders buried in a lossy half space is presented in this paper. The refractive index and permeability of the half space and cylinders are complex in general. Each cylinder is radially stratified with a distinct complex refractive index and permeability. The incident radiation is an arbitrarily polarized plane wave propagating in the plane normal to the axes of the cylinders. Analytic solutions are derived for the electric and magnetic fields and the Poynting vector of backscattered radiation emerging from the half space. Numerical examples are presented to illustrate the application of the scattering solution to calculate backscattering from a lossy half space containing multiple homogeneous and radially stratified cylinders at various depths and different angles of incidence.

  17. High-throughput analysis of candidate imprinted genes and allele-specific gene expression in the human term placenta

    Clark Taane G

    2010-04-01

    Full Text Available Abstract Background Imprinted genes show expression from one parental allele only and are important for development and behaviour. This extreme mode of allelic imbalance has been described for approximately 56 human genes. Imprinting status is often disrupted in cancer and dysmorphic syndromes. More subtle variation of gene expression, that is not parent-of-origin specific, termed 'allele-specific gene expression' (ASE is more common and may give rise to milder phenotypic differences. Using two allele-specific high-throughput technologies alongside bioinformatics predictions, normal term human placenta was screened to find new imprinted genes and to ascertain the extent of ASE in this tissue. Results Twenty-three family trios of placental cDNA, placental genomic DNA (gDNA and gDNA from both parents were tested for 130 candidate genes with the Sequenom MassArray system. Six genes were found differentially expressed but none imprinted. The Illumina ASE BeadArray platform was then used to test 1536 SNPs in 932 genes. The array was enriched for the human orthologues of 124 mouse candidate genes from bioinformatics predictions and 10 human candidate imprinted genes from EST database mining. After quality control pruning, a total of 261 informative SNPs (214 genes remained for analysis. Imprinting with maternal expression was demonstrated for the lymphocyte imprinted gene ZNF331 in human placenta. Two potential differentially methylated regions (DMRs were found in the vicinity of ZNF331. None of the bioinformatically predicted candidates tested showed imprinting except for a skewed allelic expression in a parent-specific manner observed for PHACTR2, a neighbour of the imprinted PLAGL1 gene. ASE was detected for two or more individuals in 39 candidate genes (18%. Conclusions Both Sequenom and Illumina assays were sensitive enough to study imprinting and strong allelic bias. Previous bioinformatics approaches were not predictive of new imprinted genes

  18. Mutant Allele-Specific Uncoupling of PENETRATION3 Functions Reveals Engagement of the ATP-Binding Cassette Transporter in Distinct Tryptophan Metabolic Pathways1[OPEN

    Lu, Xunli; Dittgen, Jan; Piślewska-Bednarek, Mariola; Molina, Antonio; Schneider, Bernd; Doubský, Jan; Schneeberger, Korbinian; Schulze-Lefert, Paul

    2015-01-01

    Arabidopsis (Arabidopsis thaliana) PENETRATION (PEN) genes quantitatively contribute to the execution of different forms of plant immunity upon challenge with diverse leaf pathogens. PEN3 encodes a plasma membrane-resident pleiotropic drug resistance-type ATP-binding cassette transporter and is thought to act in a pathogen-inducible and PEN2 myrosinase-dependent metabolic pathway in extracellular defense. This metabolic pathway directs the intracellular biosynthesis and activation of tryptophan-derived indole glucosinolates for subsequent PEN3-mediated efflux across the plasma membrane at pathogen contact sites. However, PEN3 also functions in abiotic stress responses to cadmium and indole-3-butyric acid (IBA)-mediated auxin homeostasis in roots, raising the possibility that PEN3 exports multiple functionally unrelated substrates. Here, we describe the isolation of a pen3 allele, designated pen3-5, that encodes a dysfunctional protein that accumulates in planta like wild-type PEN3. The specific mutation in pen3-5 uncouples PEN3 functions in IBA-stimulated root growth modulation, callose deposition induced with a conserved peptide epitope of bacterial flagellin (flg22), and pathogen-inducible salicylic acid accumulation from PEN3 activity in extracellular defense, indicating the engagement of multiple PEN3 substrates in different PEN3-dependent biological processes. We identified 4-O-β-d-glucosyl-indol-3-yl formamide (4OGlcI3F) as a pathogen-inducible, tryptophan-derived compound that overaccumulates in pen3 leaf tissue and has biosynthesis that is dependent on an intact PEN2 metabolic pathway. We propose that a precursor of 4OGlcI3F is the PEN3 substrate in extracellular pathogen defense. These precursors, the shared indole core present in IBA and 4OGlcI3F, and allele-specific uncoupling of a subset of PEN3 functions suggest that PEN3 transports distinct indole-type metabolites in distinct biological processes. PMID:26023163

  19. Development of allele-specific multiplex PCR to determine the length of poly-T in intron 8 of CFTR

    Neng Chen

    2014-07-01

    Full Text Available Cystic fibrosis transmembrane conductance regulator (CFTR gene mutation analysis has been implemented for Cystic Fibrosis (CF carrier screening, and molecular diagnosis of CF and congenital bilateral absence of the vas deferens (CBAVD. Although poly-T allele analysis in intron 8 of CFTR is required when a patient is positive for R117H, it is not recommended for routine carrier screening. Therefore, commercial kits for CFTR mutation analysis were designed either to mask the poly-T allele results, unless a patient is R117H positive, or to have the poly-T analysis as a standalone reflex test using the same commercial platform. There are other standalone assays developed to detect poly-T alleles, such as heteroduplex analysis, High Resolution Melting (HRM curve analysis, allele-specific PCR (AS-PCR and Sanger sequencing. In this report, we developed a simple and easy-to-implement multiplex AS-PCR assay using unlabeled standard length primers, which can be used as a reflex or standalone test for CFTR poly-T track analysis. Out of 115 human gDNA samples tested, results from our new AS-PCR matched to the previous known poly-T results or results from Sanger sequencing.

  20. Allele-specific Gene Silencing of Mutant mRNA Restores Cellular Function in Ullrich Congenital Muscular Dystrophy Fibroblasts

    Satoru Noguchi

    2014-01-01

    Full Text Available Ullrich congenital muscular dystrophy (UCMD is an inherited muscle disorder characterized clinically by muscle weakness, distal joint hyperlaxity, and proximal joint contractures. Sporadic and recessive mutations in the three collagen VI genes, COL6A1, COL6A2, and COL6A3, are reported to be causative. In the sporadic forms, a heterozygous point mutation causing glycine substitution in the triple helical domain has been identified in higher rate. In this study, we examined the efficacy of siRNAs, which target point mutation site, on specific knockdown toward transcripts from mutant allele and evaluated consequent cellular phenotype of UCMD fibroblasts. We evaluated the effect of siRNAs targeted to silence-specific COL6A1 alleles in UCMD fibroblasts, where simultaneous expression of both wild-type and mutant collagen VI resulted in defective collagen localization. Addition of mutant-specific siRNAs allowed normal extracellular localization of collagen VI surrounding fibroblasts, suggesting selective inhibition of mutant collagen VI. Targeting the single-nucleotide COL6A1 c.850G>A (p.G284R mutation responsible a sporadic autosomal dominant form of UCMD can potently and selectively block expression of mutant collagen VI. These results suggest that allele-specific knockdown of the mutant mRNA can potentially be considered as a therapeutic procedure in UCMD due to COL6A1 point mutations.

  1. Allelic inhibition of displacement activity: a simplified one tube allele-specific PCR for evaluation of ITPA polymorphisms.

    Galmozzi, E; Facchetti, F; Degasperi, E; Aghemo, A; Lampertico, P

    2013-02-01

    Recently, genome-wide association studies (GWAS) in patients with chronic hepatitis C virus (HCV) infection have identified two functional single nucleotide polymorphisms (SNPs) in the inosine triphosphatase (ITPA) gene, that are associated strongly and independently with hemolytic anemia in patients exposed to pegylated-interferon (Peg-IFN) plus ribavirin (RBV) combined therapy. Here has been developed a simplified allele discrimination polymerase chain reaction (PCR) assay named allelic inhibition of displacement activity (AIDA) for evaluation of ITPA polymorphisms. AIDA system relies on three unlabeled primers only, two outer common primers and one inner primer with allele-specific 3' terminus mismatch. DNA samples from 192 patients with chronic HCV infection were used to validate the AIDA system and results were compared with the gold standard TaqMan(®) SNP genotyping assay. Concordant data were obtained for all samples, granting for high specificity of the method. In conclusion, AIDA is a practical one-tube method to reproducibly and to assess accurately rs7270101 and rs1127354 ITPA SNPs. Copyright © 2012 Elsevier B.V. All rights reserved.

  2. Advantage of using allele-specific copy numbers when testing for association in regions with common copy number variants.

    Gaëlle Marenne

    Full Text Available Copy number variants (CNV can be called from SNP-arrays; however, few studies have attempted to combine both CNV and SNP calls to test for association with complex diseases. Even when SNPs are located within CNVs, two separate association analyses are necessary, to compare the distribution of bi-allelic genotypes in cases and controls (referred to as SNP-only strategy and the number of copies of a region (referred to as CNV-only strategy. However, when disease susceptibility is actually associated with allele specific copy-number states, the two strategies may not yield comparable results, raising a series of questions about the optimal analytical approach. We performed simulations of the performance of association testing under different scenarios that varied genotype frequencies and inheritance models. We show that the SNP-only strategy lacks power under most scenarios when the SNP is located within a CNV; frequently it is excluded from analysis as it does not pass quality control metrics either because of an increased rate of missing calls or a departure from fitness for Hardy-Weinberg proportion. The CNV-only strategy also lacks power because the association testing depends on the allele which copy number varies. The combined strategy performs well in most of the scenarios. Hence, we advocate the use of this combined strategy when testing for association with SNPs located within CNVs.

  3. Identification and removal of low-complexity sites in allele-specific analysis of ChIP-seq data.

    Waszak, Sebastian M; Kilpinen, Helena; Gschwind, Andreas R; Orioli, Andrea; Raghav, Sunil K; Witwicki, Robert M; Migliavacca, Eugenia; Yurovsky, Alisa; Lappalainen, Tuuli; Hernandez, Nouria; Reymond, Alexandre; Dermitzakis, Emmanouil T; Deplancke, Bart

    2014-01-15

    High-throughput sequencing technologies enable the genome-wide analysis of the impact of genetic variation on molecular phenotypes at unprecedented resolution. However, although powerful, these technologies can also introduce unexpected artifacts. We investigated the impact of library amplification bias on the identification of allele-specific (AS) molecular events from high-throughput sequencing data derived from chromatin immunoprecipitation assays (ChIP-seq). Putative AS DNA binding activity for RNA polymerase II was determined using ChIP-seq data derived from lymphoblastoid cell lines of two parent-daughter trios. We found that, at high-sequencing depth, many significant AS binding sites suffered from an amplification bias, as evidenced by a larger number of clonal reads representing one of the two alleles. To alleviate this bias, we devised an amplification bias detection strategy, which filters out sites with low read complexity and sites featuring a significant excess of clonal reads. This method will be useful for AS analyses involving ChIP-seq and other functional sequencing assays. The R package abs filter for library clonality simulations and detection of amplification-biased sites is available from http://updepla1srv1.epfl.ch/waszaks/absfilter

  4. Infrequent detection of germline allele-specific expression of TGFBR1 in lymphoblasts and tissues of colon cancer patients.

    Guda, Kishore

    2009-06-15

    Recently, germline allele-specific expression (ASE) of the gene encoding for transforming growth factor-beta type I receptor (TGFBR1) has been proposed to be a major risk factor for cancer predisposition in the colon. Germline ASE results in a lowered expression of one of the TGFBR1 alleles (>1.5-fold), and was shown to occur in approximately 20% of informative familial and sporadic colorectal cancer (CRC) cases. In the present study, using the highly quantitative pyrosequencing technique, we estimated the frequency of ASE in TGFBR1 in a cohort of affected individuals from familial clusters of advanced colon neoplasias (cancers and adenomas with high-grade dysplasia), and also from a cohort of individuals with sporadic CRCs. Cases were considered positive for the presence of ASE if demonstrating an allelic expression ratio <0.67 or >1.5. Using RNA derived from lymphoblastoid cell lines, we find that of 46 informative Caucasian advanced colon neoplasia cases with a family history, only 2 individuals display a modest ASE, with allelic ratios of 1.65 and 1.73, respectively. Given that ASE of TGFBR1, if present, would likely be more pronounced in the colon compared with other tissues, we additionally determined the allele ratios of TGFBR1 in the RNA derived from normal-appearing colonic mucosa of sporadic CRC cases. We, however, found no evidence of ASE in any of 44 informative sporadic cases analyzed. Taken together, we find that germline ASE of TGFBR1, as assayed in lymphoblastoid and colon epithelial cells of colon cancer patients, is a relatively rare event.

  5. Allele-specific expression at the androgen receptor alpha gene in a hybrid unisexual fish, the Amazon molly (Poecilia formosa.

    Fangjun Zhu

    Full Text Available The all-female Amazon molly (Poecilia formosa is the result of a hybridization of the Atlantic molly (P. mexicana and the sailfin molly (P. latipinna approximately 120,000 years ago. As a gynogenetic species, P. formosa needs to copulate with heterospecific males including males from one of its bisexual ancestral species. However, the sperm only triggers embryogenesis of the diploid eggs. The genetic information of the sperm donor typically will not contribute to the next generation of P. formosa. Hence, P. formosa possesses generally one allele from each of its ancestral species at any genetic locus. This raises the question whether both ancestral alleles are equally expressed in P. formosa. Allele-specific expression (ASE has been previously assessed in various organisms, e.g., human and fish, and ASE was found to be important in the context of phenotypic variability and disease. In this study, we utilized Real-Time PCR techniques to estimate ASE of the androgen receptor alpha (arα gene in several distinct tissues of Amazon mollies. We found an allelic bias favoring the maternal ancestor (P. mexicana allele in ovarian tissue. This allelic bias was not observed in the gill or the brain tissue. Sequencing of the promoter regions of both alleles revealed an association between an Indel in a known CpG island and differential expression. Future studies may reveal whether our observed cis-regulatory divergence is caused by an ovary-specific trans-regulatory element, preferentially activating the allele of the maternal ancestor.

  6. Hybrid sterility and evolution in Hawaiian Drosophila: differential gene and allele-specific expression analysis of backcross males.

    Brill, E; Kang, L; Michalak, K; Michalak, P; Price, D K

    2016-08-01

    The Hawaiian Drosophila are an iconic example of sequential colonization, adaptive radiation and speciation on islands. Genetic and phenotypic analysis of closely related species pairs that exhibit incomplete reproductive isolation can provide insights into the mechanisms of speciation. Drosophila silvestris from Hawai'i Island and Drosophila planitibia from Maui are two closely related allopatric Hawaiian picture-winged Drosophila that produce sterile F1 males but fertile F1 females, a pattern consistent with Haldane's rule. Backcrossing F1 hybrid females between these two species to parental species gives rise to recombinant males with three distinct sperm phenotypes despite a similar genomic background: motile sperm, no sperm (sterile), and immotile sperm. We found that these three reproductive morphologies of backcross hybrid males produce divergent gene expression profiles in testes, as measured with RNA sequencing. There were a total of 71 genes significantly differentially expressed between backcross males with no sperm compared with those backcross males with motile sperm and immotile sperm, but no significant differential gene expression between backcross males with motile sperm and backcross males with immotile sperm. All of these genes were underexpressed in males with no sperm, including a number of genes with previously known activities in adult testis. An allele-specific expression analysis showed overwhelmingly more cis-divergent than trans-divergent genes, with no significant difference in the ratio of cis- and trans-divergent genes among the sperm phenotypes. Overall, the results indicate that the regulation of gene expression involved in sperm production likely diverged relatively rapidly between these two closely related species.

  7. Development of Allele-Specific Primer PCR for a Swine TLR2 SNP and Comparison of the Frequency among Several Pig Breeds of Japan and the Czech Republic

    Muneta, Y.; Minagawa, Y.; Kusumoto, M.; Shinkai, H.; Uenishi, H.; Šplíchal, Igor

    2012-01-01

    Roč. 74, č. 5 (2012), s. 553-559 ISSN 0916-7250 R&D Projects: GA ČR GA524/09/0365 Institutional support: RVO:61388971 Keywords : allele-specific PCR * Mycoplasma hyopneumoniae * single nucleotide polymorphism Subject RIV: EC - Immunology Impact factor: 0.876, year: 2012

  8. Allele-specific gene expression patterns in primary leukemic cells reveal regulation of gene expression by CpG site methylation

    Milani, Lili; Lundmark, Anders; Nordlund, Jessica

    2008-01-01

    To identify genes that are regulated by cis-acting functional elements in acute lymphoblastic leukemia (ALL) we determined the allele-specific expression (ASE) levels of 2, 529 genes by genotyping a genome-wide panel of single nucleotide polymorphisms in RNA and DNA from bone marrow and blood...

  9. The rhizosphere microbial community in a multiple parallel mineralization system suppresses the pathogenic fungus Fusarium oxysporum

    Fujiwara, Kazuki; Iida, Yuichiro; Iwai, Takashi; Aoyama, Chihiro; Inukai, Ryuya; Ando, Akinori; Ogawa, Jun; Ohnishi, Jun; Terami, Fumihiro; Takano, Masao; Shinohara, Makoto

    2013-01-01

    The rhizosphere microbial community in a hydroponics system with multiple parallel mineralization (MPM) can potentially suppress root-borne diseases. This study focused on revealing the biological nature of the suppression against Fusarium wilt disease, which is caused by the fungus Fusarium oxysporum, and describing the factors that may influence the fungal pathogen in the MPM system. We demonstrated that the rhizosphere microbiota that developed in the MPM system could suppress Fusarium wilt disease under in vitro and greenhouse conditions. The microbiological characteristics of the MPM system were able to control the population dynamics of F. oxysporum, but did not eradicate the fungal pathogen. The roles of the microbiological agents underlying the disease suppression and the magnitude of the disease suppression in the MPM system appear to depend on the microbial density. F. oxysporum that survived in the MPM system formed chlamydospores when exposed to the rhizosphere microbiota. These results suggest that the microbiota suppresses proliferation of F. oxysporum by controlling the pathogen's morphogenesis and by developing an ecosystem that permits coexistence with F. oxysporum. PMID:24311557

  10. Performance Analysis of a Threshold-Based Parallel Multiple Beam Selection Scheme for WDM FSO Systems

    Nam, Sung Sik

    2018-04-09

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme for a free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has occurred under independent identically distributed Gamma-Gamma fading conditions. To simplify the mathematical analysis, we additionally consider Gamma turbulence conditions, which are a good approximation of Gamma-Gamma distribution. Specifically, we statistically analyze the characteristics in operation under conventional detection schemes (i.e., heterodyne detection (HD) and intensity modulation/direct detection (IM/DD) techniques) for both adaptive modulation (AM) case in addition to non-AM case (i.e., coherent/non-coherent binary modulation). Then, based on the statistically derived results, we evaluate the outage probability of a selected beam, the average spectral efficiency (ASE), the average number of selected beams (ANSB) and the average bit error rate (BER). Selected results show that we can obtain higher spectral efficiency and simultaneously reduce the potential for increasing the complexity of implementation caused by applying the selection-based beam selection scheme without considerable performance loss. Especially for the AM case, the ASE can be increased further compared to the non- AM cases. Our derived results based on the Gamma distribution as an approximation of the Gamma-Gamma distribution can be used as approximated performance measure bounds, especially, they may lead to lower bounds on the approximated considered performance measures.

  11. The rhizosphere microbial community in a multiple parallel mineralization system suppresses the pathogenic fungus Fusarium oxysporum.

    Fujiwara, Kazuki; Iida, Yuichiro; Iwai, Takashi; Aoyama, Chihiro; Inukai, Ryuya; Ando, Akinori; Ogawa, Jun; Ohnishi, Jun; Terami, Fumihiro; Takano, Masao; Shinohara, Makoto

    2013-12-01

    The rhizosphere microbial community in a hydroponics system with multiple parallel mineralization (MPM) can potentially suppress root-borne diseases. This study focused on revealing the biological nature of the suppression against Fusarium wilt disease, which is caused by the fungus Fusarium oxysporum, and describing the factors that may influence the fungal pathogen in the MPM system. We demonstrated that the rhizosphere microbiota that developed in the MPM system could suppress Fusarium wilt disease under in vitro and greenhouse conditions. The microbiological characteristics of the MPM system were able to control the population dynamics of F. oxysporum, but did not eradicate the fungal pathogen. The roles of the microbiological agents underlying the disease suppression and the magnitude of the disease suppression in the MPM system appear to depend on the microbial density. F. oxysporum that survived in the MPM system formed chlamydospores when exposed to the rhizosphere microbiota. These results suggest that the microbiota suppresses proliferation of F. oxysporum by controlling the pathogen's morphogenesis and by developing an ecosystem that permits coexistence with F. oxysporum. © 2013 The Authors. MicrobiologyOpen published by John Wiley & Sons Ltd.

  12. MSAProbs-MPI: parallel multiple sequence aligner for distributed-memory systems.

    González-Domínguez, Jorge; Liu, Yongchao; Touriño, Juan; Schmidt, Bertil

    2016-12-15

    MSAProbs is a state-of-the-art protein multiple sequence alignment tool based on hidden Markov models. It can achieve high alignment accuracy at the expense of relatively long runtimes for large-scale input datasets. In this work we present MSAProbs-MPI, a distributed-memory parallel version of the multithreaded MSAProbs tool that is able to reduce runtimes by exploiting the compute capabilities of common multicore CPU clusters. Our performance evaluation on a cluster with 32 nodes (each containing two Intel Haswell processors) shows reductions in execution time of over one order of magnitude for typical input datasets. Furthermore, MSAProbs-MPI using eight nodes is faster than the GPU-accelerated QuickProbs running on a Tesla K20. Another strong point is that MSAProbs-MPI can deal with large datasets for which MSAProbs and QuickProbs might fail due to time and memory constraints, respectively. Source code in C ++ and MPI running on Linux systems as well as a reference manual are available at http://msaprobs.sourceforge.net CONTACT: jgonzalezd@udc.esSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  13. Short term scheduling of multiple grid-parallel PEM fuel cells for microgrid applications

    El-Sharkh, M.Y.; Rahman, A.; Alam, M.S. [Dept. of Electrical and Computer Engineering, University of South Alabama, Mobile, AL 36688 (United States)

    2010-10-15

    This paper presents a short term scheduling scheme for multiple grid-parallel PEM fuel cell power plants (FCPPs) connected to supply electrical and thermal energy to a microgrid community. As in the case of regular power plants, short term scheduling of FCPP is also a cost-based optimization problem that includes the cost of operation, thermal power recovery, and the power trade with the local utility grid. Due to the ability of the microgrid community to trade power with the local grid, the power balance constraint is not applicable, other constraints like the real power operating limits of the FCPP, and minimum up and down time are therefore used. To solve the short term scheduling problem of the FCPPs, a hybrid technique based on evolutionary programming (EP) and hill climbing technique (HC) is used. The EP is used to estimate the optimal schedule and the output power from each FCPP. The HC technique is used to monitor the feasibility of the solution during the search process. The short term scheduling problem is used to estimate the schedule and the electrical and thermal power output of five FCPPs supplying a maximum power of 300 kW. (author)

  14. Category-based attentional guidance can operate in parallel for multiple target objects.

    Jenkins, Michael; Grubert, Anna; Eimer, Martin

    2018-04-30

    The question whether the control of attention during visual search is always feature-based or can also be based on the category of objects remains unresolved. Here, we employed the N2pc component as an on-line marker for target selection processes to compare the efficiency of feature-based and category-based attentional guidance. Two successive displays containing pairs of real-world objects (line drawings of kitchen or clothing items) were separated by a 10 ms SOA. In Experiment 1, target objects were defined by their category. In Experiment 2, one specific visual object served as target (exemplar-based search). On different trials, targets appeared either in one or in both displays, and participants had to report the number of targets (one or two). Target N2pc components were larger and emerged earlier during exemplar-based search than during category-based search, demonstrating the superior efficiency of feature-based attentional guidance. On trials where target objects appeared in both displays, both targets elicited N2pc components that overlapped in time, suggesting that attention was allocated in parallel to these target objects. Critically, this was the case not only in the exemplar-based task, but also when targets were defined by their category. These results demonstrate that attention can be guided by object categories, and that this type of category-based attentional control can operate concurrently for multiple target objects. Copyright © 2018 Elsevier B.V. All rights reserved.

  15. Column-Parallel Single Slope ADC with Digital Correlated Multiple Sampling for Low Noise CMOS Image Sensors

    Chen, Y.; Theuwissen, A.J.P.; Chae, Y.

    2011-01-01

    This paper presents a low noise CMOS image sensor (CIS) using 10/12 bit configurable column-parallel single slope ADCs (SS-ADCs) and digital correlated multiple sampling (CMS). The sensor used is a conventional 4T active pixel with a pinned-photodiode as photon detector. The test sensor was

  16. WASP: a Web-based Allele-Specific PCR assay designing tool for detecting SNPs and mutations

    Assawamakin Anunchai

    2007-08-01

    Full Text Available Abstract Background Allele-specific (AS Polymerase Chain Reaction is a convenient and inexpensive method for genotyping Single Nucleotide Polymorphisms (SNPs and mutations. It is applied in many recent studies including population genetics, molecular genetics and pharmacogenomics. Using known AS primer design tools to create primers leads to cumbersome process to inexperience users since information about SNP/mutation must be acquired from public databases prior to the design. Furthermore, most of these tools do not offer the mismatch enhancement to designed primers. The available web applications do not provide user-friendly graphical input interface and intuitive visualization of their primer results. Results This work presents a web-based AS primer design application called WASP. This tool can efficiently design AS primers for human SNPs as well as mutations. To assist scientists with collecting necessary information about target polymorphisms, this tool provides a local SNP database containing over 10 million SNPs of various populations from public domain databases, namely NCBI dbSNP, HapMap and JSNP respectively. This database is tightly integrated with the tool so that users can perform the design for existing SNPs without going off the site. To guarantee specificity of AS primers, the proposed system incorporates a primer specificity enhancement technique widely used in experiment protocol. In particular, WASP makes use of different destabilizing effects by introducing one deliberate 'mismatch' at the penultimate (second to last of the 3'-end base of AS primers to improve the resulting AS primers. Furthermore, WASP offers graphical user interface through scalable vector graphic (SVG draw that allow users to select SNPs and graphically visualize designed primers and their conditions. Conclusion WASP offers a tool for designing AS primers for both SNPs and mutations. By integrating the database for known SNPs (using gene ID or rs number

  17. Hierarchical approach to optimization of parallel matrix multiplication on large-scale platforms

    Hasanov, Khalid; Quintin, Jean-Noë l; Lastovetsky, Alexey

    2014-01-01

    -scale parallelism in mind. Indeed, while in 1990s a system with few hundred cores was considered a powerful supercomputer, modern top supercomputers have millions of cores. In this paper, we present a hierarchical approach to optimization of message-passing parallel

  18. A Laboratory Preparation of Aspartame Analogs Using Simultaneous Multiple Parallel Synthesis Methodology

    Qvit, Nir; Barda, Yaniv; Gilon, Chaim; Shalev, Deborah E.

    2007-01-01

    This laboratory experiment provides a unique opportunity for students to synthesize three analogues of aspartame, a commonly used artificial sweetener. The students are introduced to the powerful and useful method of parallel synthesis while synthesizing three dipeptides in parallel using solid-phase peptide synthesis (SPPS) and simultaneous…

  19. A note on the nucleation with multiple steps: Parallel and series nucleation

    Iwamatsu, Masao

    2012-01-01

    Parallel and series nucleation are the basic elements of the complex nucleation process when two saddle points exist on the free-energy landscape. It is pointed out that the nucleation rates follow formulas similar to those of parallel and series connection of resistors or conductors in an electric circuit. Necessary formulas to calculate individual nucleation rates at the saddle points and the total nucleation rate are summarized and the extension to the more complex nucleation process is su...

  20. Towards a molecular taxonomic key of the Aurantioideae subfamily using chloroplastic SNP diagnostic markers of the main clades genotyped by competitive allele-specific PCR.

    Oueslati, Amel; Ollitrault, Frederique; Baraket, Ghada; Salhi-Hannachi, Amel; Navarro, Luis; Ollitrault, Patrick

    2016-08-18

    Chloroplast DNA is a primary source of molecular variations for phylogenetic analysis of photosynthetic eukaryotes. However, the sequencing and analysis of multiple chloroplastic regions is difficult to apply to large collections or large samples of natural populations. The objective of our work was to demonstrate that a molecular taxonomic key based on easy, scalable and low-cost genotyping method should be developed from a set of Single Nucleotide Polymorphisms (SNPs) diagnostic of well-established clades. It was applied to the Aurantioideae subfamily, the largest group of the Rutaceae family that includes the cultivated citrus species. The publicly available nucleotide sequences of eight plastid genomic regions were compared for 79 accessions of the Aurantioideae subfamily to search for SNPs revealing taxonomic differentiation at the inter-tribe, inter-subtribe, inter-genus and interspecific levels. Diagnostic SNPs (DSNPs) were found for 46 of the 54 clade levels analysed. Forty DSNPs were selected to develop KASPar markers and their taxonomic value was tested by genotyping 108 accessions of the Aurantioideae subfamily. Twenty-seven markers diagnostic of 24 clades were validated and they displayed a very high rate of transferability in the Aurantioideae subfamily (only 1.2 % of missing data on average). The UPGMA from the validated markers produced a cladistic organisation that was highly coherent with the previous phylogenetic analysis based on the sequence data of the eight plasmid regions. In particular, the monophyletic origin of the "true citrus" genera plus Oxanthera was validated. However, some clarification remains necessary regarding the organisation of the other wild species of the Citreae tribe. We validated the concept that with well-established clades, DSNPs can be selected and efficiently transformed into competitive allele-specific PCR markers (KASPar method) allowing cost-effective highly efficient cladistic analysis in large collections at

  1. A genome-wide screen in human embryonic stem cells reveals novel sites of allele-specific histone modification associated with known disease loci

    Prendergast, James G D

    2012-05-19

    AbstractBackgroundChromatin structure at a given site can differ between chromosome copies in a cell, and such imbalances in chromatin structure have been shown to be important in understanding the molecular mechanisms controlling several disease loci. Human genetic variation, DNA methylation, and disease have been intensely studied, uncovering many sites of allele-specific DNA methylation (ASM). However, little is known about the genome-wide occurrence of sites of allele-specific histone modification (ASHM) and their relationship to human disease. The aim of this study was to investigate the extent and characteristics of sites of ASHM in human embryonic stem cells (hESCs).ResultsUsing a statistically rigorous protocol, we investigated the genomic distribution of ASHM in hESCs, and their relationship to sites of allele-specific expression (ASE) and DNA methylation. We found that, although they were rare, sites of ASHM were substantially enriched at loci displaying ASE. Many were also found at known imprinted regions, hence sites of ASHM are likely to be better markers of imprinted regions than sites of ASM. We also found that sites of ASHM and ASE in hESCs colocalize at risk loci for developmental syndromes mediated by deletions, providing insights into the etiology of these disorders.ConclusionThese results demonstrate the potential importance of ASHM patterns in the interpretation of disease loci, and the protocol described provides a basis for similar studies of ASHM in other cell types to further our understanding of human disease susceptibility.

  2. Systematic evaluation of the impact of ChIP-seq read designs on genome coverage, peak identification, and allele-specific binding detection.

    Zhang, Qi; Zeng, Xin; Younkin, Sam; Kawli, Trupti; Snyder, Michael P; Keleş, Sündüz

    2016-02-24

    Chromatin immunoprecipitation followed by sequencing (ChIP-seq) experiments revolutionized genome-wide profiling of transcription factors and histone modifications. Although maturing sequencing technologies allow these experiments to be carried out with short (36-50 bps), long (75-100 bps), single-end, or paired-end reads, the impact of these read parameters on the downstream data analysis are not well understood. In this paper, we evaluate the effects of different read parameters on genome sequence alignment, coverage of different classes of genomic features, peak identification, and allele-specific binding detection. We generated 101 bps paired-end ChIP-seq data for many transcription factors from human GM12878 and MCF7 cell lines. Systematic evaluations using in silico variations of these data as well as fully simulated data, revealed complex interplay between the sequencing parameters and analysis tools, and indicated clear advantages of paired-end designs in several aspects such as alignment accuracy, peak resolution, and most notably, allele-specific binding detection. Our work elucidates the effect of design on the downstream analysis and provides insights to investigators in deciding sequencing parameters in ChIP-seq experiments. We present the first systematic evaluation of the impact of ChIP-seq designs on allele-specific binding detection and highlights the power of pair-end designs in such studies.

  3. A note on the nucleation with multiple steps: parallel and series nucleation.

    Iwamatsu, Masao

    2012-01-28

    Parallel and series nucleation are the basic elements of the complex nucleation process when two saddle points exist on the free-energy landscape. It is pointed out that the nucleation rates follow formulas similar to those of parallel and series connection of resistors or conductors in an electric circuit. Necessary formulas to calculate individual nucleation rates at the saddle points and the total nucleation rate are summarized, and the extension to the more complex nucleation process is suggested. © 2012 American Institute of Physics

  4. Parallelism measurement for base plate of standard artifact with multiple tactile approaches

    Ye, Xiuling; Zhao, Yan; Wang, Yiwen; Wang, Zhong; Fu, Luhua; Liu, Changjie

    2018-01-01

    Nowadays, as workpieces become more precise and more specialized which results in more sophisticated structures and higher accuracy for the artifacts, higher requirements have been put forward for measuring accuracy and measuring methods. As an important method to obtain the size of workpieces, coordinate measuring machine (CMM) has been widely used in many industries. In order to achieve the calibration of a self-developed CMM, it is found that the parallelism of the base plate used for fixing the standard artifact is an important factor which affects the measurement accuracy in the process of studying self-made high-precision standard artifact. And aimed to measure the parallelism of the base plate, by using the existing high-precision CMM, gauge blocks, dial gauge and marble platform with the tactile approach, three methods for parallelism measurement of workpieces are employed, and comparisons are made within the measurement results. The results of experiments show that the final accuracy of all the three methods is able to reach micron level and meets the measurement requirements. Simultaneously, these three approaches are suitable for different measurement conditions which provide a basis for rapid and high-precision measurement under different equipment conditions.

  5. Stepped-wedge cluster randomised controlled trials: a generic framework including parallel and multiple-level designs.

    Hemming, Karla; Lilford, Richard; Girling, Alan J

    2015-01-30

    Stepped-wedge cluster randomised trials (SW-CRTs) are being used with increasing frequency in health service evaluation. Conventionally, these studies are cross-sectional in design with equally spaced steps, with an equal number of clusters randomised at each step and data collected at each and every step. Here we introduce several variations on this design and consider implications for power. One modification we consider is the incomplete cross-sectional SW-CRT, where the number of clusters varies at each step or where at some steps, for example, implementation or transition periods, data are not collected. We show that the parallel CRT with staggered but balanced randomisation can be considered a special case of the incomplete SW-CRT. As too can the parallel CRT with baseline measures. And we extend these designs to allow for multiple layers of clustering, for example, wards within a hospital. Building on results for complete designs, power and detectable difference are derived using a Wald test and obtaining the variance-covariance matrix of the treatment effect assuming a generalised linear mixed model. These variations are illustrated by several real examples. We recommend that whilst the impact of transition periods on power is likely to be small, where they are a feature of the design they should be incorporated. We also show examples in which the power of a SW-CRT increases as the intra-cluster correlation (ICC) increases and demonstrate that the impact of the ICC is likely to be smaller in a SW-CRT compared with a parallel CRT, especially where there are multiple levels of clustering. Finally, through this unified framework, the efficiency of the SW-CRT and the parallel CRT can be compared. © 2014 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.

  6. Generalized framework for the parallel semantic segmentation of multiple objects and posterior manipulation

    Llopart, Adrian; Ravn, Ole; Andersen, Nils Axel

    2017-01-01

    The end-to-end approach presented in this paper deals with the recognition, detection, segmentation and grasping of objects, assuming no prior knowledge of the environment nor objects. The proposed pipeline is as follows: 1) Usage of a trained Convolutional Neural Net (CNN) that recognizes up to 80...... different classes of objects in real time and generates bounding boxes around them. 2) An algorithm to derive in parallel the pointclouds of said regions of interest (ROI). 3) Eight different segmentation methods to remove background data and noise from the pointclouds and obtain a precise result...

  7. Reduced dose uncertainty in MRI-based polymer gel dosimetry using parallel RF transmission with multiple RF sources

    Sang-Young Kim; Jung-Hoon Lee; Jin-Young Jung; Do-Wan Lee; Seu-Ran Lee; Bo-Young Choe; Hyeon-Man Baek; Korea University of Science and Technology, Daejeon; Dae-Hyun Kim; Jung-Whan Min; Ji-Yeon Park

    2014-01-01

    In this work, we present the feasibility of using a parallel RF transmit with multiple RF sources imaging method (MultiTransmit imaging) in polymer gel dosimetry. Image quality and B 1 field homogeneity was statistically better in the MultiTransmit imaging method than in conventional single source RF transmission imaging method. In particular, the standard uncertainty of R 2 was lower on the MultiTransmit images than on the conventional images. Furthermore, the MultiTransmit measurement showed improved dose resolution. Improved image quality and B 1 homogeneity results in reduced dose uncertainty, thereby suggesting the feasibility of MultiTransmit MR imaging in gel dosimetry. (author)

  8. Is orthographic information from multiple parafoveal words processed in parallel: An eye-tracking study.

    Cutter, Michael G; Drieghe, Denis; Liversedge, Simon P

    2017-08-01

    In the current study we investigated whether orthographic information available from 1 upcoming parafoveal word influences the processing of another parafoveal word. Across 2 experiments we used the boundary paradigm (Rayner, 1975) to present participants with an identity preview of the 2 words after the boundary (e.g., hot pan ), a preview in which 2 letters were transposed between these words (e.g., hop tan ), or a preview in which the same 2 letters were substituted (e.g., hob fan ). We hypothesized that if these 2 words were processed in parallel in the parafovea then we may observe significant preview benefits for the condition in which the letters were transposed between words relative to the condition in which the letters were substituted. However, no such effect was observed, with participants fixating the words for the same amount of time in both conditions. This was the case both when the transposition was made between the final and first letter of the 2 words (e.g., hop tan as a preview of hot pan ; Experiment 1) and when the transposition maintained within word letter position (e.g., pit hop as a preview of hit pop ; Experiment 2). The implications of these findings are considered in relation to serial and parallel lexical processing during reading. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  9. A Fast Multiple Sampling Method for Low-Noise CMOS Image Sensors With Column-Parallel 12-bit SAR ADCs

    Min-Kyu Kim

    2015-12-01

    Full Text Available This paper presents a fast multiple sampling method for low-noise CMOS image sensor (CIS applications with column-parallel successive approximation register analog-to-digital converters (SAR ADCs. The 12-bit SAR ADC using the proposed multiple sampling method decreases the A/D conversion time by repeatedly converting a pixel output to 4-bit after the first 12-bit A/D conversion, reducing noise of the CIS by one over the square root of the number of samplings. The area of the 12-bit SAR ADC is reduced by using a 10-bit capacitor digital-to-analog converter (DAC with four scaled reference voltages. In addition, a simple up/down counter-based digital processing logic is proposed to perform complex calculations for multiple sampling and digital correlated double sampling. To verify the proposed multiple sampling method, a 256 × 128 pixel array CIS with 12-bit SAR ADCs was fabricated using 0.18 μm CMOS process. The measurement results shows that the proposed multiple sampling method reduces each A/D conversion time from 1.2 μs to 0.45 μs and random noise from 848.3 μV to 270.4 μV, achieving a dynamic range of 68.1 dB and an SNR of 39.2 dB.

  10. Enhanced specificity of TPMT*2 genotyping using unidirectional wild-type and mutant allele-specific scorpion primers in a single tube.

    Dong Chen

    Full Text Available Genotyping of thiopurine S-methyltransferase (TPMT is recommended for predicting the adverse drug response of thiopurines. In the current study, a novel version of allele-specific PCR (AS-PCR, termed competitive real-time fluorescent AS-PCR (CRAS-PCR was developed to analyze the TPMT*2 genotype in ethnic Chinese. This technique simultaneously uses wild-type and mutant allele-specific scorpion primers in a single reaction. To determine the optimal conditions for both traditional AS-PCR and CRAS-PCR, we used the Taguchi method, an engineering optimization process that balances the concentrations of all components using an orthogonal array rather than a factorial array. Instead of running up to 264 experiments with the conventional factorial method, the Taguchi method achieved the same optimization using only 16 experiments. The optimized CRAS-PCR system completely avoided non-specific amplification occurring in traditional AS-PCR and could be performed at much more relaxed reaction conditions at 1% sensitivity, similar to traditional AS-PCR. TPMT*2 genotyping of 240 clinical samples was consistent with published data. In conclusion, CRAS-PCR is a novel and robust genotyping method, and the Taguchi method is an effective tool for the optimization of molecular analysis techniques.

  11. siRNA-mediated Allele-specific Silencing of a COL6A3 Mutation in a Cellular Model of Dominant Ullrich Muscular Dystrophy

    Véronique Bolduc

    2014-01-01

    Full Text Available Congenital muscular dystrophy type Ullrich (UCMD is a severe disorder of early childhood onset for which currently there is no effective treatment. UCMD commonly is caused by dominant-negative mutations in the genes coding for collagen type VI, a major microfibrillar component of the extracellular matrix surrounding the muscle fibers. To explore RNA interference (RNAi as a potential therapy for UCMD, we designed a series of small interfering RNA (siRNA oligos that specifically target the most common mutations resulting in skipping of exon 16 in the COL6A3 gene and tested them in UCMD-derived dermal fibroblasts. Transcript analysis by semiquantitative and quantitative reverse transcriptase PCR showed that two of these siRNAs were the most allele-specific, i.e., they efficiently knocked down the expression from the mutant allele, without affecting the normal allele. In HEK293T cells, these siRNAs selectively suppressed protein expression from a reporter construct carrying the mutation, with no or minimal suppression of the wild-type (WT construct, suggesting that collagen VI protein levels are as also reduced in an allele-specific manner. Furthermore, we found that treating UCMD fibroblasts with these siRNAs considerably improved the quantity and quality of the collagen VI matrix, as assessed by confocal microscopy. Our current study establishes RNAi as a promising molecular approach for treating dominant COL6-related dystrophies.

  12. Performance analysis of a threshold-based parallel multiple beam selection scheme for WDM-based systems for Gamma-Gamma distributions

    Nam, Sung Sik; Yoon, Chang Seok; Alouini, Mohamed-Slim

    2017-01-01

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme (TPMBS) for Free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has

  13. The Pleiotropic Phenotype of Apc Mutations in the Mouse: Allele Specificity and Effects of the Genetic Background

    Halberg, Richard B.; Chen, Xiaodi; Amos-Landgraf, James M.; White, Alanna; Rasmussen, Kristin; Clipson, Linda; Pasch, Cheri; Sullivan, Ruth; Pitot, Henry C.; Dove, William F.

    2008-01-01

    Familial adenomatous polyposis (FAP) is a human cancer syndrome characterized by the development of hundreds to thousands of colonic polyps and extracolonic lesions including desmoid fibromas, osteomas, epidermoid cysts, and congenital hypertrophy of the pigmented retinal epithelium. Afflicted individuals are heterozygous for mutations in the APC gene. Detailed investigations of mice heterozygous for mutations in the ortholog Apc have shown that other genetic factors strongly influence the phenotype. Here we report qualitative and quantitative modifications of the phenotype of Apc mutants as a function of three genetic variables: Apc allele, p53 allele, and genetic background. We have found major differences between the Apc alleles Min and 1638N in multiplicity and regionality of intestinal tumors, as well as in incidence of extracolonic lesions. By contrast, Min mice homozygous for either of two different knockout alleles of p53 show similar phenotypic effects. These studies illustrate the classic principle that functional genetics is enriched by assessing penetrance and expressivity with allelic series. The mouse permits study of an allelic gene series on multiple genetic backgrounds, thereby leading to a better understanding of gene action in a range of biological processes. PMID:18723878

  14. Smoking and increased risk of multiple sclerosis: parallel trends in the sex ratio reinforce the evidence

    Palacios, Natalia; Alonso, Alvaro; Brønnum-Hansen, Henrik

    2011-01-01

    Smoking behavior in industrialized nations has changed markedly over the second half of the 20th century, with diverging patterns in male and female smoking rates. We examined whether the female/male incidence of multiple sclerosis (MS) changed concomitantly with smoking, as would be expected if ...

  15. Mixed-time parallel evolution in multiple quantum NMR experiments: sensitivity and resolution enhancement in heteronuclear NMR

    Ying Jinfa; Chill, Jordan H.; Louis, John M.; Bax, Ad

    2007-01-01

    A new strategy is demonstrated that simultaneously enhances sensitivity and resolution in three- or higher-dimensional heteronuclear multiple quantum NMR experiments. The approach, referred to as mixed-time parallel evolution (MT-PARE), utilizes evolution of chemical shifts of the spins participating in the multiple quantum coherence in parallel, thereby reducing signal losses relative to sequential evolution. The signal in a given PARE dimension, t 1 , is of a non-decaying constant-time nature for a duration that depends on the length of t 2 , and vice versa, prior to the onset of conventional exponential decay. Line shape simulations for the 1 H- 15 N PARE indicate that this strategy significantly enhances both sensitivity and resolution in the indirect 1 H dimension, and that the unusual signal decay profile results in acceptable line shapes. Incorporation of the MT-PARE approach into a 3D HMQC-NOESY experiment for measurement of H N -H N NOEs in KcsA in SDS micelles at 50 o C was found to increase the experimental sensitivity by a factor of 1.7±0.3 with a concomitant resolution increase in the indirectly detected 1 H dimension. The method is also demonstrated for a situation in which homonuclear 13 C- 13 C decoupling is required while measuring weak H3'-2'OH NOEs in an RNA oligomer

  16. Power Factor Correction Capacitors for Multiple Parallel Three-Phase ASD Systems

    Yang, Yongheng; Blaabjerg, Frede

    2017-01-01

    Today’s three-phase Adjustable Speed Drive (ASD) systems still employ Diode Rectifiers (DRs) and Silicon-Controlled Rectifiers (SCRs) as the front-end converters due to structural and control simplicity, small volume, low cost, and high reliability. However, the uncontrollable DRs and phase......-controllable SCRs bring side-effects by injecting high harmonics to the grid, which will degrade the system performance in terms of lowering the overall efficiency and overheating the system if remain uncontrolled or unattenuated. For multiple ASD systems, certain harmonics in the entire system can be mitigated...... the power factor, passive capacitors can be installed, which yet can trigger the system resonance. Hence, this paper analyzes the resonant issues in multiple ASD systems with power factor correction capacitors. Potential damping solutions are summarized. Simulations are carried out, while laboratory tests...

  17. Quantitative threefold allele-specific PCR (QuanTAS-PCR) for highly sensitive JAK2 V617F mutant allele detection

    Zapparoli, Giada V; Jorissen, Robert N; Hewitt, Chelsee A; McBean, Michelle; Westerman, David A; Dobrovic, Alexander

    2013-01-01

    The JAK2 V617F mutation is the most frequent somatic change in myeloproliferative neoplasms, making it an important tumour-specific marker for diagnostic purposes and for the detection of minimal residual disease. Sensitive quantitative assays are required for both applications, particularly for the monitoring of minimal residual disease, which requires not only high sensitivity but also very high specificity. We developed a highly sensitive probe-free quantitative mutant-allele detection method, Quantitative Threefold Allele-Specific PCR (QuanTAS-PCR), that is performed in a closed-tube system, thus eliminating the manipulation of PCR products. QuantTAS-PCR uses a threefold approach to ensure allele-specific amplification of the mutant sequence: (i) a mutant allele-specific primer, (ii) a 3′dideoxy blocker to suppress false-positive amplification from the wild-type template and (iii) a PCR specificity enhancer, also to suppress false-positive amplification from the wild-type template. Mutant alleles were quantified relative to exon 9 of JAK2. We showed that the addition of the 3′dideoxy blocker suppressed but did not eliminate false-positive amplification from the wild-type template. However, the addition of the PCR specificity enhancer near eliminated false-positive amplification from the wild-type allele. Further discrimination between true and false positives was enabled by using the quantification cycle (Cq) value of a single mutant template as a cut-off point, thus enabling robust distinction between true and false positives. As 10,000 JAK2 templates were used per replicate, the assay had a sensitivity of 1/10 -4 per replicate. Greater sensitivity could be reached by increasing the number of replicates analysed. Variation in replicates when low mutant-allele templates were present necessitated the use of a statistics-based approach to estimate the load of mutant JAK2 copies. QuanTAS-PCR showed comparable quantitative results when validated against a

  18. Detection of EGFR mutations in plasma and biopsies from non-small cell lung cancer patients by allele-specific PCR assays

    Weber, Britta; Meldgaard, Peter; Hager, Henrik

    2014-01-01

    samples with allele-specific PCR assays. METHODS: Pairs of the diagnostic biopsy and plasma obtained just prior to start of erlotinib treatment were collected from 199 patients with adenocarcinoma of non-small-cell lung cancer. DNA from both sample types was isolated and examined for the presence...... of mutations in exons 18-21 of the EGFR gene, employing the cobas(®) EGFR Tissue Test and cobas(®) EGFR Blood Test (in development, Roche Molecular Systems, Inc., CA, USA). RESULTS: Test results were obtained in all 199 (100%) plasma samples and 196/199 (98%) of the biopsies. EGFR-activating mutations were...... identified in 24/199 (12%) plasma samples and 28/196 (14%) biopsy samples, and 17/196 (9%) matched pairs contained the same mutation. Six EGFR mutations were present only in plasma samples but not in the biopsy samples. The overall concordance of the EGFR gene mutations detected in plasma and biopsy tissue...

  19. Early Parallel Activation of Semantics and Phonology in Picture Naming: Evidence from a Multiple Linear Regression MEG Study.

    Miozzo, Michele; Pulvermüller, Friedemann; Hauk, Olaf

    2015-10-01

    The time course of brain activation during word production has become an area of increasingly intense investigation in cognitive neuroscience. The predominant view has been that semantic and phonological processes are activated sequentially, at about 150 and 200-400 ms after picture onset. Although evidence from prior studies has been interpreted as supporting this view, these studies were arguably not ideally suited to detect early brain activation of semantic and phonological processes. We here used a multiple linear regression approach to magnetoencephalography (MEG) analysis of picture naming in order to investigate early effects of variables specifically related to visual, semantic, and phonological processing. This was combined with distributed minimum-norm source estimation and region-of-interest analysis. Brain activation associated with visual image complexity appeared in occipital cortex at about 100 ms after picture presentation onset. At about 150 ms, semantic variables became physiologically manifest in left frontotemporal regions. In the same latency range, we found an effect of phonological variables in the left middle temporal gyrus. Our results demonstrate that multiple linear regression analysis is sensitive to early effects of multiple psycholinguistic variables in picture naming. Crucially, our results suggest that access to phonological information might begin in parallel with semantic processing around 150 ms after picture onset. © The Author 2014. Published by Oxford University Press.

  20. Massive parallelization of a 3D finite difference electromagnetic forward solution using domain decomposition methods on multiple CUDA enabled GPUs

    Schultz, A.

    2010-12-01

    3D forward solvers lie at the core of inverse formulations used to image the variation of electrical conductivity within the Earth's interior. This property is associated with variations in temperature, composition, phase, presence of volatiles, and in specific settings, the presence of groundwater, geothermal resources, oil/gas or minerals. The high cost of 3D solutions has been a stumbling block to wider adoption of 3D methods. Parallel algorithms for modeling frequency domain 3D EM problems have not achieved wide scale adoption, with emphasis on fairly coarse grained parallelism using MPI and similar approaches. The communications bandwidth as well as the latency required to send and receive network communication packets is a limiting factor in implementing fine grained parallel strategies, inhibiting wide adoption of these algorithms. Leading Graphics Processor Unit (GPU) companies now produce GPUs with hundreds of GPU processor cores per die. The footprint, in silicon, of the GPU's restricted instruction set is much smaller than the general purpose instruction set required of a CPU. Consequently, the density of processor cores on a GPU can be much greater than on a CPU. GPUs also have local memory, registers and high speed communication with host CPUs, usually through PCIe type interconnects. The extremely low cost and high computational power of GPUs provides the EM geophysics community with an opportunity to achieve fine grained (i.e. massive) parallelization of codes on low cost hardware. The current generation of GPUs (e.g. NVidia Fermi) provides 3 billion transistors per chip die, with nearly 500 processor cores and up to 6 GB of fast (DDR5) GPU memory. This latest generation of GPU supports fast hardware double precision (64 bit) floating point operations of the type required for frequency domain EM forward solutions. Each Fermi GPU board can sustain nearly 1 TFLOP in double precision, and multiple boards can be installed in the host computer system. We

  1. The use of coded PCR primers enables high-throughput sequencing of multiple homolog amplification products by 454 parallel sequencing.

    Jonas Binladen

    2007-02-01

    Full Text Available The invention of the Genome Sequence 20 DNA Sequencing System (454 parallel sequencing platform has enabled the rapid and high-volume production of sequence data. Until now, however, individual emulsion PCR (emPCR reactions and subsequent sequencing runs have been unable to combine template DNA from multiple individuals, as homologous sequences cannot be subsequently assigned to their original sources.We use conventional PCR with 5'-nucleotide tagged primers to generate homologous DNA amplification products from multiple specimens, followed by sequencing through the high-throughput Genome Sequence 20 DNA Sequencing System (GS20, Roche/454 Life Sciences. Each DNA sequence is subsequently traced back to its individual source through 5'tag-analysis.We demonstrate that this new approach enables the assignment of virtually all the generated DNA sequences to the correct source once sequencing anomalies are accounted for (miss-assignment rate<0.4%. Therefore, the method enables accurate sequencing and assignment of homologous DNA sequences from multiple sources in single high-throughput GS20 run. We observe a bias in the distribution of the differently tagged primers that is dependent on the 5' nucleotide of the tag. In particular, primers 5' labelled with a cytosine are heavily overrepresented among the final sequences, while those 5' labelled with a thymine are strongly underrepresented. A weaker bias also exists with regards to the distribution of the sequences as sorted by the second nucleotide of the dinucleotide tags. As the results are based on a single GS20 run, the general applicability of the approach requires confirmation. However, our experiments demonstrate that 5'primer tagging is a useful method in which the sequencing power of the GS20 can be applied to PCR-based assays of multiple homologous PCR products. The new approach will be of value to a broad range of research areas, such as those of comparative genomics, complete mitochondrial

  2. Parallelizing flow-accumulation calculations on graphics processing units—From iterative DEM preprocessing algorithm to recursive multiple-flow-direction algorithm

    Qin, Cheng-Zhi; Zhan, Lijun

    2012-06-01

    As one of the important tasks in digital terrain analysis, the calculation of flow accumulations from gridded digital elevation models (DEMs) usually involves two steps in a real application: (1) using an iterative DEM preprocessing algorithm to remove the depressions and flat areas commonly contained in real DEMs, and (2) using a recursive flow-direction algorithm to calculate the flow accumulation for every cell in the DEM. Because both algorithms are computationally intensive, quick calculation of the flow accumulations from a DEM (especially for a large area) presents a practical challenge to personal computer (PC) users. In recent years, rapid increases in hardware capacity of the graphics processing units (GPUs) provided in modern PCs have made it possible to meet this challenge in a PC environment. Parallel computing on GPUs using a compute-unified-device-architecture (CUDA) programming model has been explored to speed up the execution of the single-flow-direction algorithm (SFD). However, the parallel implementation on a GPU of the multiple-flow-direction (MFD) algorithm, which generally performs better than the SFD algorithm, has not been reported. Moreover, GPU-based parallelization of the DEM preprocessing step in the flow-accumulation calculations has not been addressed. This paper proposes a parallel approach to calculate flow accumulations (including both iterative DEM preprocessing and a recursive MFD algorithm) on a CUDA-compatible GPU. For the parallelization of an MFD algorithm (MFD-md), two different parallelization strategies using a GPU are explored. The first parallelization strategy, which has been used in the existing parallel SFD algorithm on GPU, has the problem of computing redundancy. Therefore, we designed a parallelization strategy based on graph theory. The application results show that the proposed parallel approach to calculate flow accumulations on a GPU performs much faster than either sequential algorithms or other parallel GPU

  3. Dynamic modeling and hierarchical compound control of a novel 2-DOF flexible parallel manipulator with multiple actuation modes

    Liang, Dong; Song, Yimin; Sun, Tao; Jin, Xueying

    2018-03-01

    This paper addresses the problem of rigid-flexible coupling dynamic modeling and active control of a novel flexible parallel manipulator (PM) with multiple actuation modes. Firstly, based on the flexible multi-body dynamics theory, the rigid-flexible coupling dynamic model (RFDM) of system is developed by virtue of the augmented Lagrangian multipliers approach. For completeness, the mathematical models of permanent magnet synchronous motor (PMSM) and piezoelectric transducer (PZT) are further established and integrated with the RFDM of mechanical system to formulate the electromechanical coupling dynamic model (ECDM). To achieve the trajectory tracking and vibration suppression, a hierarchical compound control strategy is presented. Within this control strategy, the proportional-differential (PD) feedback controller is employed to realize the trajectory tracking of end-effector, while the strain and strain rate feedback (SSRF) controller is developed to restrain the vibration of the flexible links using PZT. Furthermore, the stability of the control algorithm is demonstrated based on the Lyapunov stability theory. Finally, two simulation case studies are performed to illustrate the effectiveness of the proposed approach. The results indicate that, under the redundant actuation mode, the hierarchical compound control strategy can guarantee the flexible PM achieves singularity-free motion and vibration attenuation within task workspace simultaneously. The systematic methodology proposed in this study can be conveniently extended for the dynamic modeling and efficient controller design of other flexible PMs, especially the emerging ones with multiple actuation modes.

  4. HAlign-II: efficient ultra-large multiple sequence alignment and phylogenetic tree reconstruction with distributed and parallel computing.

    Wan, Shixiang; Zou, Quan

    2017-01-01

    Multiple sequence alignment (MSA) plays a key role in biological sequence analyses, especially in phylogenetic tree construction. Extreme increase in next-generation sequencing results in shortage of efficient ultra-large biological sequence alignment approaches for coping with different sequence types. Distributed and parallel computing represents a crucial technique for accelerating ultra-large (e.g. files more than 1 GB) sequence analyses. Based on HAlign and Spark distributed computing system, we implement a highly cost-efficient and time-efficient HAlign-II tool to address ultra-large multiple biological sequence alignment and phylogenetic tree construction. The experiments in the DNA and protein large scale data sets, which are more than 1GB files, showed that HAlign II could save time and space. It outperformed the current software tools. HAlign-II can efficiently carry out MSA and construct phylogenetic trees with ultra-large numbers of biological sequences. HAlign-II shows extremely high memory efficiency and scales well with increases in computing resource. THAlign-II provides a user-friendly web server based on our distributed computing infrastructure. HAlign-II with open-source codes and datasets was established at http://lab.malab.cn/soft/halign.

  5. Multiple and Periodic Measurement of RBC Aggregation and ESR in Parallel Microfluidic Channels under On-Off Blood Flow Control

    Yang Jun Kang

    2018-06-01

    Full Text Available Red blood cell (RBC aggregation causes to alter hemodynamic behaviors at low flow-rate regions of post-capillary venules. Additionally, it is significantly elevated in inflammatory or pathophysiological conditions. In this study, multiple and periodic measurements of RBC aggregation and erythrocyte sedimentation rate (ESR are suggested by sucking blood from a pipette tip into parallel microfluidic channels, and quantifying image intensity, especially through single experiment. Here, a microfluidic device was prepared from a master mold using the xurography technique rather than micro-electro-mechanical-system fabrication techniques. In order to consider variations of RBC aggregation in microfluidic channels due to continuous ESR in the conical pipette tip, two indices (aggregation index (AI and erythrocyte-sedimentation-rate aggregation index (EAI are evaluated by using temporal variations of microscopic, image-based intensity. The proposed method is employed to evaluate the effect of hematocrit and dextran solution on RBC aggregation under continuous ESR in the conical pipette tip. As a result, EAI displays a significantly linear relationship with modified conventional ESR measurement obtained by quantifying time constants. In addition, EAI varies linearly within a specific concentration of dextran solution. In conclusion, the proposed method is able to measure RBC aggregation under continuous ESR in the conical pipette tip. Furthermore, the method provides multiple data of RBC aggregation and ESR through a single experiment. A future study will involve employing the proposed method to evaluate biophysical properties of blood samples collected from cardiovascular diseases.

  6. Cloud-Coffee: implementation of a parallel consistency-based multiple alignment algorithm in the T-Coffee package and its benchmarking on the Amazon Elastic-Cloud.

    Di Tommaso, Paolo; Orobitg, Miquel; Guirado, Fernando; Cores, Fernado; Espinosa, Toni; Notredame, Cedric

    2010-08-01

    We present the first parallel implementation of the T-Coffee consistency-based multiple aligner. We benchmark it on the Amazon Elastic Cloud (EC2) and show that the parallelization procedure is reasonably effective. We also conclude that for a web server with moderate usage (10K hits/month) the cloud provides a cost-effective alternative to in-house deployment. T-Coffee is a freeware open source package available from http://www.tcoffee.org/homepage.html

  7. Enhancement of allele discrimination by introduction of nucleotide mismatches into siRNA in allele-specific gene silencing by RNAi.

    Yusuke Ohnishi

    Full Text Available Allele-specific gene silencing by RNA interference (RNAi is therapeutically useful for specifically inhibiting the expression of disease-associated alleles without suppressing the expression of corresponding wild-type alleles. To realize such allele-specific RNAi (ASP-RNAi, the design and assessment of small interfering RNA (siRNA duplexes conferring ASP-RNAi is vital; however, it is also difficult. In a previous study, we developed an assay system to assess ASP-RNAi with mutant and wild-type reporter alleles encoding the Photinus and Renilla luciferase genes. In line with experiments using the system, we realized that it is necessary and important to enhance allele discrimination between mutant and corresponding wild-type alleles. Here, we describe the improvement of ASP-RNAi against mutant alleles carrying single nucleotide variations by introducing base substitutions into siRNA sequences, where original variations are present in the central position. Artificially mismatched siRNAs or short-hairpin RNAs (shRNAs against mutant alleles of the human Prion Protein (PRNP gene, which appear to be associated with susceptibility to prion diseases, were examined using this assessment system. The data indicates that introduction of a one-base mismatch into the siRNAs and shRNAs was able to enhance discrimination between the mutant and wild-type alleles. Interestingly, the introduced mismatches that conferred marked improvement in ASP-RNAi, appeared to be largely present in the guide siRNA elements, corresponding to the 'seed region' of microRNAs. Due to the essential role of the 'seed region' of microRNAs in their association with target RNAs, it is conceivable that disruption of the base-pairing interactions in the corresponding seed region, as well as the central position (involved in cleavage of target RNAs, of guide siRNA elements could influence allele discrimination. In addition, we also suggest that nucleotide mismatches at the 3'-ends of sense

  8. A and MdMYB1 allele-specific markers controlling apple (Malus x domestica Borkh.) skin color and suitability for marker-assisted selection.

    Zhang, X J; Wang, L X; Chen, X X; Liu, Y L; Meng, R; Wang, Y J; Zhao, Z Y

    2014-10-31

    Pre-selection for fruit skin color at the seedling stage would be highly advantageous, with marker-assisted selection offering a potential method for apple pre-selection. A and MdMYB1 alleles are allele-specific DNA markers that are potentially associated with apple skin color, and co-segregate with the Rf and Rni loci, respectively. Here, we assessed the potential application of these 2 alleles for marker-assisted breeding across 30 diverse cultivars and 2 apple seedling progenies. The red skin color phenotype was usually associated with the MdMYB1-1 allele and A(1) allele, respectively, while the 2 molecular markers provided approximately 91% predictability in the 'Fuji' x 'Cripps Pink' and 'Fuji' x 'Gala' progenies. The results obtained from the 30 cultivars and 2 progenies were consistent for the 2 molecular markers. Hence, the results supported that Rf and Rni could be located in a gene cluster, or even correspond to alleles of the same gene. Our results are consistent with the hypothesis that red/yellow dimorphism is controlled by a monogenic system, with the presence of the red anthocyanin pigmentation being dominant. In addition, our results supported that the practical utilization of the 2 function markers to efficiently and accurately select red-skinned apple cultivars in apple scion breeding programs.

  9. Evaluation of efficiency of nested multiplex allele-specific PCR assay for detection of multidrug resistant tuberculosis directly from sputum samples.

    Mistri, S K; Sultana, M; Kamal, S M M; Alam, M M; Irin, F; Nessa, J; Ahsan, C R; Yasmin, M

    2016-05-01

    For an effective control of tuberculosis, rapid detection of multidrug resistant tuberculosis (MDR-TB) is necessary. Therefore, we developed a modified nested multiplex allele-specific polymerase chain reaction (MAS-PCR) method that enables rapid MDR-TB detection directly from sputum samples. The efficacy of this method was evaluated using 79 sputum samples collected from suspected tuberculosis patients. The performance of nested MAS-PCR method was compared with other MDR-TB detection methods like drug susceptibility testing (DST) and DNA sequencing. As rifampicin (RIF) resistance conforms to MDR-TB in greater than 90% cases, only the presence of RIF-associated mutations in rpoB gene was determined by DNA sequencing and nested MAS-PCR to detect MDR-TB. The concordance between nested MAS-PCR and DNA sequencing results was found to be 96·3%. When compared with DST, the sensitivity and specificity of nested MAS-PCR for RIF-resistance detection were determined to be 92·9 and 100% respectively. For developing- and high-TB burden countries, molecular-based tests have been recommended by the World Health Organization for rapid detection of MDR-TB. The results of this study indicate that, nested MAS-PCR assay might be a practical and relatively cost effective molecular method for rapid detection of MDR-TB from suspected sputum samples in developing countries with resource poor settings. © 2016 The Society for Applied Microbiology.

  10. Centrifugo-pneumatic multi-liquid aliquoting - parallel aliquoting and combination of multiple liquids in centrifugal microfluidics.

    Schwemmer, F; Hutzenlaub, T; Buselmeier, D; Paust, N; von Stetten, F; Mark, D; Zengerle, R; Kosse, D

    2015-08-07

    The generation of mixtures with precisely metered volumes is essential for reproducible automation of laboratory workflows. Splitting a given liquid into well-defined metered sub-volumes, the so-called aliquoting, has been frequently demonstrated on centrifugal microfluidics. However, so far no solution exists for assays that require simultaneous aliquoting of multiple, different liquids and the subsequent pairwise combination of aliquots with full fluidic separation before combination. Here, we introduce the centrifugo-pneumatic multi-liquid aliquoting designed for parallel aliquoting and pairwise combination of multiple liquids. All pumping and aliquoting steps are based on a combination of centrifugal forces and pneumatic forces. The pneumatic forces are thereby provided intrinsically by centrifugal transport of the assay liquids into dead end chambers to compress the enclosed air. As an example, we demonstrate simultaneous aliquoting of 1.) a common assay reagent into twenty 5 μl aliquots and 2.) five different sample liquids, each into four aliquots of 5 μl. Subsequently, the reagent and sample aliquots are simultaneously transported and combined into twenty collection chambers. All coefficients of variation for metered volumes were between 0.4%-1.0% for intra-run variations and 0.5%-1.2% for inter-run variations. The aliquoting structure is compatible to common assay reagents with a wide range of liquid and material properties, demonstrated here for contact angles between 20° and 60°, densities between 789 and 1855 kg m(-3) and viscosities between 0.89 and 4.1 mPa s. The centrifugo-pneumatic multi-liquid aliquoting is implemented as a passive fluidic structure into a single fluidic layer. Fabrication is compatible to scalable fabrication technologies such as injection molding or thermoforming and does not require any additional fabrication steps such as hydrophilic or hydrophobic coatings or integration of active valves.

  11. Bandwidth scalable, coherent transmitter based on the parallel synthesis of multiple spectral slices using optical arbitrary waveform generation.

    Geisler, David J; Fontaine, Nicolas K; Scott, Ryan P; He, Tingting; Paraschis, Loukas; Gerstel, Ori; Heritage, Jonathan P; Yoo, S J B

    2011-04-25

    We demonstrate an optical transmitter based on dynamic optical arbitrary waveform generation (OAWG) which is capable of creating high-bandwidth (THz) data waveforms in any modulation format using the parallel synthesis of multiple coherent spectral slices. As an initial demonstration, the transmitter uses only 5.5 GHz of electrical bandwidth and two 10-GHz-wide spectral slices to create 100-ns duration, 20-GHz optical waveforms in various modulation formats including differential phase-shift keying (DPSK), quaternary phase-shift keying (QPSK), and eight phase-shift keying (8PSK) with only changes in software. The experimentally generated waveforms showed clear eye openings and separated constellation points when measured using a real-time digital coherent receiver. Bit-error-rate (BER) performance analysis resulted in a BER < 9.8 × 10(-6) for DPSK and QPSK waveforms. Additionally, we experimentally demonstrate three-slice, 4-ns long waveforms that highlight the bandwidth scalable nature of the optical transmitter. The various generated waveforms show that the key transmitter properties (i.e., packet length, modulation format, data rate, and modulation filter shape) are software definable, and that the optical transmitter is capable of acting as a flexible bandwidth transmitter.

  12. GENESIS 1.1: A hybrid-parallel molecular dynamics simulator with enhanced sampling algorithms on multiple computational platforms.

    Kobayashi, Chigusa; Jung, Jaewoon; Matsunaga, Yasuhiro; Mori, Takaharu; Ando, Tadashi; Tamura, Koichi; Kamiya, Motoshi; Sugita, Yuji

    2017-09-30

    GENeralized-Ensemble SImulation System (GENESIS) is a software package for molecular dynamics (MD) simulation of biological systems. It is designed to extend limitations in system size and accessible time scale by adopting highly parallelized schemes and enhanced conformational sampling algorithms. In this new version, GENESIS 1.1, new functions and advanced algorithms have been added. The all-atom and coarse-grained potential energy functions used in AMBER and GROMACS packages now become available in addition to CHARMM energy functions. The performance of MD simulations has been greatly improved by further optimization, multiple time-step integration, and hybrid (CPU + GPU) computing. The string method and replica-exchange umbrella sampling with flexible collective variable choice are used for finding the minimum free-energy pathway and obtaining free-energy profiles for conformational changes of a macromolecule. These new features increase the usefulness and power of GENESIS for modeling and simulation in biological research. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  13. Multiple-algorithm parallel fusion of infrared polarization and intensity images based on algorithmic complementarity and synergy

    Zhang, Lei; Yang, Fengbao; Ji, Linna; Lv, Sheng

    2018-01-01

    Diverse image fusion methods perform differently. Each method has advantages and disadvantages compared with others. One notion is that the advantages of different image methods can be effectively combined. A multiple-algorithm parallel fusion method based on algorithmic complementarity and synergy is proposed. First, in view of the characteristics of the different algorithms and difference-features among images, an index vector-based feature-similarity is proposed to define the degree of complementarity and synergy. This proposed index vector is a reliable evidence indicator for algorithm selection. Second, the algorithms with a high degree of complementarity and synergy are selected. Then, the different degrees of various features and infrared intensity images are used as the initial weights for the nonnegative matrix factorization (NMF). This avoids randomness of the NMF initialization parameter. Finally, the fused images of different algorithms are integrated using the NMF because of its excellent data fusing performance on independent features. Experimental results demonstrate that the visual effect and objective evaluation index of the fused images obtained using the proposed method are better than those obtained using traditional methods. The proposed method retains all the advantages that individual fusion algorithms have.

  14. T-cell libraries allow simple parallel generation of multiple peptide-specific human T-cell clones.

    Theaker, Sarah M; Rius, Cristina; Greenshields-Watson, Alexander; Lloyd, Angharad; Trimby, Andrew; Fuller, Anna; Miles, John J; Cole, David K; Peakman, Mark; Sewell, Andrew K; Dolton, Garry

    2016-03-01

    Isolation of peptide-specific T-cell clones is highly desirable for determining the role of T-cells in human disease, as well as for the development of therapies and diagnostics. However, generation of monoclonal T-cells with the required specificity is challenging and time-consuming. Here we describe a library-based strategy for the simple parallel detection and isolation of multiple peptide-specific human T-cell clones from CD8(+) or CD4(+) polyclonal T-cell populations. T-cells were first amplified by CD3/CD28 microbeads in a 96U-well library format, prior to screening for desired peptide recognition. T-cells from peptide-reactive wells were then subjected to cytokine-mediated enrichment followed by single-cell cloning, with the entire process from sample to validated clone taking as little as 6 weeks. Overall, T-cell libraries represent an efficient and relatively rapid tool for the generation of peptide-specific T-cell clones, with applications shown here in infectious disease (Epstein-Barr virus, influenza A, and Ebola virus), autoimmunity (type 1 diabetes) and cancer. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.

  15. SU94. Allele-Specific and Trauma-Related Epigenetic Changes in the FKBP5 Gene: Differences Between Psychotic Patients and Healthy Controls

    Mihaljevic, Marina; Franic, Dusica; Soldatovic, Ivan; Andric, Sanja; Mirjanic, Tijana; Novakovic, Ivana; Adzic, Miroslav; Maric, Nadja

    2017-01-01

    Abstract Background: Hypothalamic-pituitary-adrenal (HPA) axis dysregulation is a proposed etiological mechanism of psychosis. Recent studies highlighted impact of the FKBP5 gene and its functional variant rs1360780, which risk (T) allele affects the activity of HPA axis following stress exposure, on psychotic patients exposed to early trauma (1). Additionally, risk allele and trauma dependent FKBP5 demethylation in intron 7 was observed in traumatized individuals (2). Thus, the purpose of this pilot study was to investigate influence of the risk allele and trauma on FKBP5 DNA methylation levels at intron 7 in psychotic patients and to compare it with healthy individuals. Methods: The sample consisted of 24 psychosis spectrum patients and 24 controls matched by age and gender. All participants were genotyped for rs1360780 and divided into 2 groups depending on the presence of the risk allele (risk and nonrisk group). DNA methylation levels at 3 CpG sites (CpG1, CpG2, and CpG3) in intron 7 were analyzed by Sanger sequencing. Early-life adversities were measured by Childhood Trauma Questionnaire. Pearson correlation and t test were performed as appropriate. Results: Analyses revealed decreased FKBP5 methylation at targeted CpG sites and averaged methylation level (AML) at intron 7 in patients compared to controls (P = .026, P = .017, P = .027, and P = .003, respectively). Decreased AML and methylation at CpG3 were observed comparing risk and nonrisk patients’ groups (P = .018 and P = .016, respectively). Additionally, decreased methylation was found in risk patients’ group compared to risk controls’ group. No differences were found comparing nonrisk groups. Furthermore, strong negative associations between trauma and methylation at CpG3 and AML were observed only in risk controls’ group (r = −0.707, P = .007; r = −0.741, P = .004, respectively). Conclusion: Our preliminary results revealed allele-specific epigenetic changes of the FKBP

  16. Genome-wide identification and quantification of cis- and trans-regulated genes responding to Marek’s disease virus infection via analysis of allele-specific expression

    Sean eMaceachern

    2012-01-01

    Full Text Available Marek’s disease (MD is a commercially important neoplastic disease of chickens caused by Marek’s disease virus (MDV, an oncogenic alphaherpesvirus. Selecting for increased genetic resistance to MD is a control strategy that can augment vaccinal control measures. To identify high-confidence candidate MD resistance genes, we conducted a genome-wide screen for allele-specific expression (ASE amongst F1 progeny of two inbred chicken lines that differ in MD resistance. High throughput sequencing was used to profile transcriptomes from pools of uninfected and infected individuals at 4 days post-infection to identify any genes showing ASE in response to MDV infection. RNA sequencing identified 22,655 single nucleotide polymorphisms (SNPs of which 5,360 in 3,773 genes exhibited significant allelic imbalance. Illumina GoldenGate assays were subsequently used to quantify regulatory variation controlled at the gene (cis and elsewhere in the genome (trans by examining differences in expression between F1 individuals and artificial F1 RNA pools over 6 time periods in 1,536 of the most significant SNPs identified by RNA sequencing. Allelic imbalance as a result of cis-regulatory changes was confirmed in 861 of the 1,233 GoldenGate assays successfully examined. Furthermore we have identified 7 genes that display trans-regulation only in infected animals and approximately 500 SNP that show a complex interaction between cis- and trans-regulatory changes. Our results indicate ASE analyses are a powerful approach to identify regulatory variation responsible for differences in transcript abundance in genes underlying complex traits. And the genes with SNPs exhibiting ASE provide a strong foundation to further investigate the causative polymorphisms and genetic mechanisms for MD resistance. Finally, the methods used here for identifying specific genes and SNPs may have practical implications for applying marker-assisted selection to complex traits that are

  17. TumorBoost: Normalization of allele-specific tumor copy numbers from a single pair of tumor-normal genotyping microarrays

    Neuvial Pierre

    2010-05-01

    Full Text Available Abstract Background High-throughput genotyping microarrays assess both total DNA copy number and allelic composition, which makes them a tool of choice for copy number studies in cancer, including total copy number and loss of heterozygosity (LOH analyses. Even after state of the art preprocessing methods, allelic signal estimates from genotyping arrays still suffer from systematic effects that make them difficult to use effectively for such downstream analyses. Results We propose a method, TumorBoost, for normalizing allelic estimates of one tumor sample based on estimates from a single matched normal. The method applies to any paired tumor-normal estimates from any microarray-based technology, combined with any preprocessing method. We demonstrate that it increases the signal-to-noise ratio of allelic signals, making it significantly easier to detect allelic imbalances. Conclusions TumorBoost increases the power to detect somatic copy-number events (including copy-neutral LOH in the tumor from allelic signals of Affymetrix or Illumina origin. We also conclude that high-precision allelic estimates can be obtained from a single pair of tumor-normal hybridizations, if TumorBoost is combined with single-array preprocessing methods such as (allele-specific CRMA v2 for Affymetrix or BeadStudio's (proprietary XY-normalization method for Illumina. A bounded-memory implementation is available in the open-source and cross-platform R package aroma.cn, which is part of the Aroma Project (http://www.aroma-project.org/.

  18. Minority drug-resistant HIV-1 variants in treatment naïve East-African and Caucasian patients detected by allele-specific real-time PCR.

    Halime Ekici

    Full Text Available To assess the presence of two major non-nucleoside reverse transcriptase inhibitors (NNRTI drug resistance mutations (DRMs, Y181C and K103N, in minor viral quasispecies of treatment naïve HIV-1 infected East-African and Swedish patients by allele-specific polymerase chain reaction (AS-PCR.Treatment naïve adults (n=191 with three epidemiological backgrounds were included: 92 Ethiopians living in Ethiopia; 55 East-Africans who had migrated to Sweden; and 44 Caucasians living in Sweden. The pol gene was analysed by standard population sequencing and by AS-PCR for the detection of Y181C and K103N.The Y181C was detected in the minority quasispecies of six Ethiopians (6.5%, in two Caucasians (4.5%, and in one East-African (1.8%. The K103N was detected in one East- African (1.8%, by both methods. The proportion of mutants ranged from 0.25% to 17.5%. Additional DRMs were found in all three treatment naïve patient groups by population sequencing.Major NNRTI mutations can be found by AS-PCR in minor quasispecies of treatment naïve HIV-1 infected Ethiopians living in Ethiopia, in East-African and Caucasian patients living in Sweden in whom population sequencing reveal wild-type virus only. Surveys with standard sequencing are likely to underestimate transmitted drug resistance and the presence of resistant minor quasispecies in treatment naïve patients should be topic for future large scale studies.

  19. Allele-specific primer polymerase chain reaction for a single nucleotide polymorphism (C1205T) of swine Toll-like receptor 5 and comparison of the allelic frequency among several pig breeds in Japan and the Czech Republic

    Muneta, Y.; Minagawa, Y.; Kusumoto, M.; Shinkai, H.; Uenishi, H.; Šplíchal, Igor

    2012-01-01

    Roč. 56, č. 6 (2012), s. 385-391 ISSN 0385-5600 R&D Projects: GA ČR GA524/09/0365 Institutional support: RVO:61388971 Keywords : allele-specific PCR * Salmonella enterica serovar Choleraesuis * single nucleotide polymorphism Subject RIV: EC - Immunology Impact factor: 1.545, year: 2012

  20. An Overview of High-performance Parallel Big Data transfers over multiple network channels with Transport Layer Security (TLS) and TLS plus Perfect Forward Secrecy (PFS)

    Fang, Chin [SLAC National Accelerator Lab., Menlo Park, CA (United States); Corttrell, R. A. [SLAC National Accelerator Lab., Menlo Park, CA (United States)

    2015-05-06

    This Technical Note provides an overview of high-performance parallel Big Data transfers with and without encryption for data in-transit over multiple network channels. It shows that with the parallel approach, it is feasible to carry out high-performance parallel "encrypted" Big Data transfers without serious impact to throughput. But other impacts, e.g. the energy-consumption part should be investigated. It also explains our rationales of using a statistics-based approach for gaining understanding from test results and for improving the system. The presentation is of high-level nature. Nevertheless, at the end we will pose some questions and identify potentially fruitful directions for future work.

  1. Inter- and intra-individual variation in allele-specific DNA methylation and gene expression in children conceived using assisted reproductive technology.

    Nahid Turan

    2010-07-01

    Full Text Available Epidemiological studies have reported a higher incidence of rare disorders involving imprinted genes among children conceived using assisted reproductive technology (ART, suggesting that ART procedures may be disruptive to imprinted gene methylation patterns. We examined intra- and inter-individual variation in DNA methylation at the differentially methylated regions (DMRs of the IGF2/H19 and IGF2R loci in a population of children conceived in vitro or in vivo. We found substantial variation in allele-specific methylation at both loci in both groups. Aberrant methylation of the maternal IGF2/H19 DMR was more common in the in vitro group, and the overall variance was also significantly greater in the in vitro group. We estimated the number of trophoblast stem cells in each group based on approximation of the variance of the binomial distribution of IGF2/H19 methylation ratios, as well as the distribution of X chromosome inactivation scores in placenta. Both of these independent measures indicated that placentas of the in vitro group were derived from fewer stem cells than the in vivo conceived group. Both IGF2 and H19 mRNAs were significantly lower in placenta from the in vitro group. Although average birth weight was lower in the in vitro group, we found no correlation between birth weight and IGF2 or IGF2R transcript levels or the ratio of IGF2/IGF2R transcript levels. Our results show that in vitro conception is associated with aberrant methylation patterns at the IGF2/H19 locus. However, very little of the inter- or intra-individual variation in H19 or IGF2 mRNA levels can be explained by differences in maternal DMR DNA methylation, in contrast to the expectations of current transcriptional imprinting models. Extraembryonic tissues of embryos cultured in vitro appear to be derived from fewer trophoblast stem cells. It is possible that this developmental difference has an effect on placental and fetal growth.

  2. Establishment of real time allele specific locked nucleic acid quantitative PCR for detection of HBV YIDD (ATT mutation and evaluation of its application.

    Yongbin Zeng

    Full Text Available BACKGROUND: Long-term use of nucleos(tide analogues can increase risk of HBV drug-resistance mutations. The rtM204I (ATT coding for isoleucine is one of the most important resistance mutation sites. Establishing a simple, rapid, reliable and highly sensitive assay to detect the resistant mutants as early as possible is of great clinical significance. METHODS: Recombinant plasmids for HBV YMDD (tyrosine-methionine-aspartate-aspartate and YIDD (tyrosine-isoleucine-aspartate-aspartate were constructed by TA cloning. Real time allele specific locked nucleic acid quantitative PCR (RT-AS-LNA-qPCR with SYBR Green I was established by LNA-modified primers and evaluated with standard recombinant plasmids, clinical templates (the clinical wild type and mutant HBV DNA mixture and 102 serum samples from nucleos(tide analogues-experienced patients. The serum samples from a chronic hepatitis B (CHB patient firstly received LMV mono therapy and then switched to LMV + ADV combined therapy were also dynamically analyzed for 10 times. RESULTS: The linear range of the assay was between 1×10(9 copies/μl and 1 × 10(2 copies/μl. The low detection limit was 1 × 10(1 copies/μl. Sensitivity of the assay were 10(-6, 10(-4 and 10(-2 in the wild-type background of 1 × 10(9 copies/μl, 1 × 10(7 copies/μl and 1 × 10(5 copies/μl, respectively. The sensitivity of the assay in detection of clinical samples was 0.03%. The complete coincidence rate between RT-AS-LNA-qPCR and direct sequencing was 91.2% (93/102, partial coincidence rate was 8.8% (9/102, and no complete discordance was observed. The two assays showed a high concordance (Kappa = 0.676, P = 0.000. Minor variants can be detected 18 weeks earlier than the rebound of HBV DNA load and alanine aminotransferase level. CONCLUSIONS: A rapid, cost-effective, high sensitive, specific and reliable method of RT-AS-LNA-qPCR with SYBR Green I for early and absolute quantification of HBV YIDD (ATT coding for isoleucine

  3. Parallel R

    McCallum, Ethan

    2011-01-01

    It's tough to argue with R as a high-quality, cross-platform, open source statistical software product-unless you're in the business of crunching Big Data. This concise book introduces you to several strategies for using R to analyze large datasets. You'll learn the basics of Snow, Multicore, Parallel, and some Hadoop-related tools, including how to find them, how to use them, when they work well, and when they don't. With these packages, you can overcome R's single-threaded nature by spreading work across multiple CPUs, or offloading work to multiple machines to address R's memory barrier.

  4. Patterns of gene flow and selection across multiple species of Acrocephalus warblers: footprints of parallel selection on the Z chromosome

    Reifová, R.; Majerová, V.; Reif, J.; Ahola, M.; Lindholm, A.; Procházka, Petr

    2016-01-01

    Roč. 16, č. 130 (2016), s. 130 ISSN 1471-2148 Institutional support: RVO:68081766 Keywords : Adaptive radiation * Speciation * Gene flow * Parallel adaptive evolution * Z chromosome * Acrocephalus warblers Subject RIV: EG - Zoology Impact factor: 3.221, year: 2016

  5. Parallel sorting algorithms

    Akl, Selim G

    1985-01-01

    Parallel Sorting Algorithms explains how to use parallel algorithms to sort a sequence of items on a variety of parallel computers. The book reviews the sorting problem, the parallel models of computation, parallel algorithms, and the lower bounds on the parallel sorting problems. The text also presents twenty different algorithms, such as linear arrays, mesh-connected computers, cube-connected computers. Another example where algorithm can be applied is on the shared-memory SIMD (single instruction stream multiple data stream) computers in which the whole sequence to be sorted can fit in the

  6. Introduction to parallel programming

    Brawer, Steven

    1989-01-01

    Introduction to Parallel Programming focuses on the techniques, processes, methodologies, and approaches involved in parallel programming. The book first offers information on Fortran, hardware and operating system models, and processes, shared memory, and simple parallel programs. Discussions focus on processes and processors, joining processes, shared memory, time-sharing with multiple processors, hardware, loops, passing arguments in function/subroutine calls, program structure, and arithmetic expressions. The text then elaborates on basic parallel programming techniques, barriers and race

  7. A modified parallel constitutive model for elevated temperature flow behavior of Ti-6Al-4V alloy based on multiple regression

    Cai, Jun; Shi, Jiamin; Wang, Kuaishe; Wang, Wen; Wang, Qingjuan; Liu, Yingying [Xi' an Univ. of Architecture and Technology, Xi' an (China). School of Metallurgical Engineering; Li, Fuguo [Northwestern Polytechnical Univ., Xi' an (China). School of Materials Science and Engineering

    2017-07-15

    Constitutive analysis for hot working of Ti-6Al-4V alloy was carried out by using experimental stress-strain data from isothermal hot compression tests. A new kind of constitutive equation called a modified parallel constitutive model was proposed by considering the independent effects of strain, strain rate and temperature. The predicted flow stress data were compared with the experimental data. Statistical analysis was introduced to verify the validity of the developed constitutive equation. Subsequently, the accuracy of the proposed constitutive equations was evaluated by comparing with other constitutive models. The results showed that the developed modified parallel constitutive model based on multiple regression could predict flow stress of Ti-6Al-4V alloy with good correlation and generalization.

  8. The use of coded PCR primers enables high-throughput sequencing of multiple homolog amplification products by 454 parallel sequencing

    Binladen, Jonas; Gilbert, M Thomas P; Bollback, Jonathan P

    2007-01-01

    BACKGROUND: The invention of the Genome Sequence 20 DNA Sequencing System (454 parallel sequencing platform) has enabled the rapid and high-volume production of sequence data. Until now, however, individual emulsion PCR (emPCR) reactions and subsequent sequencing runs have been unable to combine...... primers that is dependent on the 5' nucleotide of the tag. In particular, primers 5' labelled with a cytosine are heavily overrepresented among the final sequences, while those 5' labelled with a thymine are strongly underrepresented. A weaker bias also exists with regards to the distribution...

  9. Detection of Multiple Parallel Transmission Outbreak of Streptococcus suis Human Infection by Use of Genome Epidemiology, China, 2005.

    Du, Pengcheng; Zheng, Han; Zhou, Jieping; Lan, Ruiting; Ye, Changyun; Jing, Huaiqi; Jin, Dong; Cui, Zhigang; Bai, Xuemei; Liang, Jianming; Liu, Jiantao; Xu, Lei; Zhang, Wen; Chen, Chen; Xu, Jianguo

    2017-02-01

    Streptococcus suis sequence type 7 emerged and caused 2 of the largest human infection outbreaks in China in 1998 and 2005. To determine the major risk factors and source of the infections, we analyzed whole genomes of 95 outbreak-associated isolates, identified 160 single nucleotide polymorphisms, and classified them into 6 clades. Molecular clock analysis revealed that clade 1 (responsible for the 1998 outbreak) emerged in October 1997. Clades 2-6 (responsible for the 2005 outbreak) emerged separately during February 2002-August 2004. A total of 41 lineages of S. suis emerged by the end of 2004 and rapidly expanded to 68 genome types through single base mutations when the outbreak occurred in June 2005. We identified 32 identical isolates and classified them into 8 groups, which were distributed in a large geographic area with no transmission link. These findings suggest that persons were infected in parallel in respective geographic sites.

  10. The Modeling and Harmonic Coupling Analysis of Multiple-Parallel Connected Inverter Using Harmonic State Space (HSS)

    Kwon, Jun Bum; Wang, Xiongfei; Bak, Claus Leth

    2015-01-01

    As the number of power electronics based systems are increasing, studies about overall stability and harmonic problems are rising. In order to analyze harmonics and stability, most research is using an analysis method, which is based on the Linear Time Invariant (LTI) approach. However, this can...... be difficult in terms of complex multi-parallel connected systems, especially in the case of renewable energy, where possibilities for intermittent operation due to the weather conditions exist. Hence, it can bring many different operating points to the power converter, and the impedance characteristics can...... can demonstrate other phenomenon, which can not be found in the conventional LTI approach. The theoretical modeling and analysis are verified by means of simulations and experiments....

  11. Airborne electromagnetic detection of shallow seafloor topographic features, including resolution of multiple sub-parallel seafloor ridges

    Vrbancich, Julian; Boyd, Graham

    2014-05-01

    The HoistEM helicopter time-domain electromagnetic (TEM) system was flown over waters in Backstairs Passage, South Australia, in 2003 to test the bathymetric accuracy and hence the ability to resolve seafloor structure in shallow and deeper waters (extending to ~40 m depth) that contain interesting seafloor topography. The topography that forms a rock peak (South Page) in the form of a mini-seamount that barely rises above the water surface was accurately delineated along its ridge from the start of its base (where the seafloor is relatively flat) in ~30 m water depth to its peak at the water surface, after an empirical correction was applied to the data to account for imperfect system calibration, consistent with earlier studies using the same HoistEM system. A much smaller submerged feature (Threshold Bank) of ~9 m peak height located in waters of 35 to 40 m depth was also accurately delineated. These observations when checked against known water depths in these two regions showed that the airborne TEM system, following empirical data correction, was effectively operating correctly. The third and most important component of the survey was flown over the Yatala Shoals region that includes a series of sub-parallel seafloor ridges (resembling large sandwaves rising up to ~20 m from the seafloor) that branch out and gradually decrease in height as the ridges spread out across the seafloor. These sub-parallel ridges provide an interesting topography because the interpreted water depths obtained from 1D inversion of TEM data highlight the limitations of the EM footprint size in resolving both the separation between the ridges (which vary up to ~300 m) and the height of individual ridges (which vary up to ~20 m), and possibly also the limitations of assuming a 1D model in areas where the topography is quasi-2D/3D.

  12. Multiple Facets of Self-Control in Arab Adolescents: Parallel Pathways to Greater Happiness and Less Physical Aggression

    Gavriel-Fried, Belle; Ronen, Tammie; Agbaria, Qutaiba; Orkibi, Hod; Hamama, Liat

    2018-01-01

    Adolescence is a period of dramatic change that necessitates using skills and strengths to reduce physical aggression and increase happiness. This study examined the multiple facets of self-control skills in achieving both goals simultaneously, in a sample of 248 Arab adolescents in Israel. We conceptualized and tested a new multi-mediator model…

  13. Lipoprotein(a) and HIV: Allele-Specific Apolipoprotein(a) Levels Predict Carotid Intima-Media Thickness in HIV-Infected Young Women in the Women's Interagency HIV Study.

    Enkhmaa, Byambaa; Anuurad, Erdembileg; Zhang, Wei; Li, Chin-Shang; Kaplan, Robert; Lazar, Jason; Merenstein, Dan; Karim, Roksana; Aouizerat, Brad; Cohen, Mardge; Butler, Kenneth; Pahwa, Savita; Ofotokun, Igho; Adimora, Adaora A; Golub, Elizabeth; Berglund, Lars

    2017-05-01

    In the general population, lipoprotein(a) [Lp(a)] has been established as an independent causal risk factor for cardiovascular disease. Lp(a) levels are to a major extent regulated by a size polymorphism in the apolipoprotein(a) [apo(a)] gene. The roles of Lp(a)/apo(a) in human immunodeficiency virus (HIV)-related elevated cardiovascular disease risk remain unclear. The associations between total plasma Lp(a) level, allele-specific apo(a) level, an Lp(a) level carried by individual apo(a) alleles, and common carotid artery intima-media thickness were assessed in 150 HIV-infected and 100 HIV-uninfected women in the WIHS (Women's Interagency HIV Study). Linear regression analyses with and without adjustments were used. The cohort was young (mean age, ≈31 years), with the majority being Blacks (≈70%). The prevalence of a small size apo(a) (≤22 Kringle repeats) or a high Lp(a) level (≥30 mg/dL) was similar by HIV status. Total plasma Lp(a) level ( P =0.029) and allele-specific apo(a) level carried by the smaller apo(a) sizes ( P =0.022) were significantly associated with carotid artery intima-media thickness in the HIV-infected women only. After accounting for confounders (age, race, smoking, body mass index, blood pressure, hepatitis C virus coinfection, menopause, plasma lipids, treatment status, CD4 + T cell count, and HIV/RNA viral load), the association remained significant for both Lp(a) ( P =0.035) and allele-specific apo(a) level carried by the smaller apo(a) sizes ( P =0.010) in the HIV-infected women. Notably, none of the other lipids/lipoproteins was associated with carotid artery intima-media thickness. Lp(a) and allele-specific apo(a) levels predict carotid artery intima-media thickness in HIV-infected young women. Further research is needed to identify underlying mechanisms of an increased Lp(a) atherogenicity in HIV infection. © 2017 American Heart Association, Inc.

  14. Photoinduced dynamics of a cyanine dye: parallel pathways of non-radiative deactivation involving multiple excited-state twisted transients.

    Upadhyayula, Srigokul; Nuñez, Vicente; Espinoza, Eli M; Larsen, Jillian M; Bao, Duoduo; Shi, Dewen; Mac, Jenny T; Anvari, Bahman; Vullev, Valentine I

    2015-04-01

    Cyanine dyes are broadly used for fluorescence imaging and other photonic applications. 3,3'-Diethylthiacyanine (THIA) is a cyanine dye composed of two identical aromatic heterocyclic moieties linked with a single methine, -CH 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 1111111111111111111111111111111111 1111111111111111111111111111111111 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 1111111111111111111111111111111111 1111111111111111111111111111111111 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 0000000000000000000000000000000000 . The torsional degrees of freedom around the methine bonds provide routes for non-radiative decay, responsible for the inherently low fluorescence quantum yields. Using transient absorption spectroscopy, we determined that upon photoexcitation, the excited state relaxes along two parallel pathways producing three excited-state transients that undergo internal conversion to the ground state. The media viscosity impedes the molecular modes of ring rotation and preferentially affects one of the pathways of non-radiative decay, exerting a dominant effect on the emission

  15. Development of an allele-specific PCR assay for simultaneous sero-typing of avian pathogenic Escherichia coli predominant O1, O2, O18 and O78 strains.

    Wang, Shaohui; Meng, Qingmei; Dai, Jianjun; Han, Xiangan; Han, Yue; Ding, Chan; Liu, Haiwen; Yu, Shengqing

    2014-01-01

    Systemic infections by avian pathogenic Escherichia coli (APEC) are economically devastating to poultry industries worldwide. E. coli strains belonging to serotypes O1, O2, O18 and O78 are preferentially associated with avian colibacillosis. The rfb gene cluster controlling O antigen synthesis is usually various among different E. coli serotypes. In present study, the rfb gene clusters of E. coli serotypes O1, O2, O18 and O78 were characterized and compared. Based on the serotype-specific genes in rfb gene cluster, an allele-specific polymerase chain reaction (PCR) assay was developed. This PCR assay was highly specific and reliable for sero-typing of APEC O1, O2, O18 and O78 strains. The sensitivity of the assay was determined as 10 pg DNA or 10 colony forming units (CFUs) bacteria for serotypes O2 and O18 strains, and 500 pg DNA or 1,000 CFUs bacteria for serotypes O1 and O78 strains. Using this PCR system, APEC isolates and the infected tissue samples were categorized successfully. Furthermore, it was able to differentiate the serotypes for the samples with multi-agglutination in the traditional serum agglutination assay. Therefore, the allele-specific PCR is more simple, rapid and accurate assay for APEC diagnosis, epidemiologic study and vaccine development.

  16. Development of an allele-specific PCR assay for simultaneous sero-typing of avian pathogenic Escherichia coli predominant O1, O2, O18 and O78 strains.

    Shaohui Wang

    Full Text Available Systemic infections by avian pathogenic Escherichia coli (APEC are economically devastating to poultry industries worldwide. E. coli strains belonging to serotypes O1, O2, O18 and O78 are preferentially associated with avian colibacillosis. The rfb gene cluster controlling O antigen synthesis is usually various among different E. coli serotypes. In present study, the rfb gene clusters of E. coli serotypes O1, O2, O18 and O78 were characterized and compared. Based on the serotype-specific genes in rfb gene cluster, an allele-specific polymerase chain reaction (PCR assay was developed. This PCR assay was highly specific and reliable for sero-typing of APEC O1, O2, O18 and O78 strains. The sensitivity of the assay was determined as 10 pg DNA or 10 colony forming units (CFUs bacteria for serotypes O2 and O18 strains, and 500 pg DNA or 1,000 CFUs bacteria for serotypes O1 and O78 strains. Using this PCR system, APEC isolates and the infected tissue samples were categorized successfully. Furthermore, it was able to differentiate the serotypes for the samples with multi-agglutination in the traditional serum agglutination assay. Therefore, the allele-specific PCR is more simple, rapid and accurate assay for APEC diagnosis, epidemiologic study and vaccine development.

  17. Can Multiple Lifestyle Behaviours Be Improved in People with Familial Hypercholesterolemia? Results of a Parallel Randomised Controlled Trial

    Broekhuizen, Karen; van Poppel, Mireille N. M.; Koppes, Lando L.; Kindt, Iris; Brug, Johannes; van Mechelen, Willem

    2012-01-01

    Objective To evaluate the efficacy of an individualised tailored lifestyle intervention on physical activity, dietary intake, smoking and compliance to statin therapy in people with Familial Hypercholesterolemia (FH). Methods Adults with FH (n = 340) were randomly assigned to a usual care control group or an intervention group. The intervention consisted of web-based tailored lifestyle advice and face-to-face counselling. Physical activity, fat, fruit and vegetable intake, smoking and compliance to statin therapy were self-reported at baseline and after 12 months. Regression analyses were conducted to examine between-group differences. Intervention reach, dose and fidelity were assessed. Results In both groups, non-significant improvements in all lifestyle behaviours were found. Post-hoc analyses showed a significant decrease in saturated fat intake among women in the intervention group (β = −1.03; CI −1.98/−0.03). In the intervention group, 95% received a log on account, of which 49% logged on and completed one module. Nearly all participants received face-to-face counselling and on average, 4.2 telephone booster calls. Intervention fidelity was low. Conclusions Individually tailored feedback is not superior to no intervention regarding changes in multiple lifestyle behaviours in people with FH. A higher received dose of computer-tailored interventions should be achieved by uplifting the website and reducing the burden of screening questionnaires. Counsellor training should be more extensive. Trial Registration Dutch Trial Register NTR1899 PMID:23251355

  18. Hydraulic Fracture Induced Seismicity During A Multi-Stage Pad Completion in Western Canada: Evidence of Activation of Multiple, Parallel Faults

    Maxwell, S.; Garrett, D.; Huang, J.; Usher, P.; Mamer, P.

    2017-12-01

    Following reports of injection induced seismicity in the Western Canadian Sedimentary Basin, regulators have imposed seismic monitoring and traffic light protocols for fracturing operations in specific areas. Here we describe a case study in one of these reservoirs, the Montney Shale in NE British Columbia, where induced seismicity was monitored with a local array during multi-stage hydraulic fracture stimulations on several wells from a single drilling pad. Seismicity primarily occurred during the injection time periods, and correlated with periods of high injection rates and wellhead pressures above fracturing pressures. Sequential hydraulic fracture stages were found to progressively activate several parallel, critically-stressed faults, as illuminated by multiple linear hypocenter patterns in the range between Mw 1 and 3. Moment tensor inversion of larger events indicated a double-couple mechanism consistent with the regional strike-slip stress state and the hypocenter lineations. The critically-stressed faults obliquely cross the well paths which were purposely drilled parallel to the minimum principal stress direction. Seismicity on specific faults started and stopped when fracture initiation points of individual injection stages were proximal to the intersection of the fault and well. The distance ranges when the seismicity occurs is consistent with expected hydraulic fracture dimensions, suggesting that the induced fault slip only occurs when a hydraulic fracture grows directly into the fault and the faults are temporarily exposed to significantly elevated fracture pressures during the injection. Some faults crossed multiple wells and the seismicity was found to restart during injection of proximal stages on adjacent wells, progressively expanding the seismogenic zone of the fault. Progressive fault slip is therefore inferred from the seismicity migrating further along the faults during successive injection stages. An accelerometer was also deployed close

  19. Performance analysis of a threshold-based parallel multiple beam selection scheme for WDM-based systems for Gamma-Gamma distributions

    Nam, Sung Sik

    2017-03-02

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme (TPMBS) for Free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has occurred for practical consideration over independent identically distributed (i.i.d.) Gamma-Gamma fading conditions. Specifically, we statistically analyze the characteristics in operation under conventional heterodyne detection (HD) scheme for both adaptive modulation (AM) case in addition to non-AM case (i.e., coherentnon-coherent binary modulation). Then, based on the statistically derived results, we evaluate the outage probability (CDF) of a selected beam, the average spectral efficiency (ASE), the average number of selected beams (ANSB), and the average bit error rate (BER). Some selected results shows that we can obtain the higher spectral efficiency and simultaneously reduce the potential increasing of the complexity of implementation caused by applying the selection based beam selection scheme without a considerable performance loss.

  20. Application of multiple parallel perfused microbioreactors: Synthesis, characterization and cytotoxicity testing of the novel rare earth complexes with indole acid as a ligand.

    Guan, Qing-Lin; Xing, Yong-Heng; Liu, Jing; Wei, Wen-Juan; Zhang, Rui; Wang, Xuan; Bai, Feng-Ying

    2013-11-01

    Three novel complexes, [La(phen)2(IAA)2]·NO3 (1), [Sm(phen)2(IAA)2]·NO3 (2) and [Sm(IBA)3(phen)]·phen·HNO3·H2O (3) (phen: 1,10-phenanthroline, IAA: indole-3-acetic acid, IBA: indole-3-butyric acid), were synthesized and characterized with spectroscopy (infrared and UV-visible), X-ray crystal diffraction and elemental analysis. Structural analysis revealed that each lanthanide atom in complexes 1-3 held a distorted tricapped trigonal prism geometry in a nine-coordinate mode. There were two types of coordination modes of the IAA ligand in complexes 1 and 2: a μ2-η(1):η(2) bridging mode linking two lanthanide atoms and a μ2-η(1):η(1) double monodentate bridging mode. There were three types of coordination modes of the IBA ligand: a μ2-η(1):η(1) double monodentate bridging mode, a μ1-η(2) bridging mode and a μ2-η(1):η(2) bridging mode linking two lanthanide atoms. Adjacent Sm atoms were linked via the μ2-bridging carboxylate groups of the IBA ligands to generate a binuclear building unit. The biological activity of the complexes was evaluated in human adipose tissue-derived stem cells (hADSCs) and Chang liver cells using a multiple parallel perfused microbioreactor. The results showed that cytotoxicity increased as the concentrations of complexes 1-3 increased. © 2013.

  1. Parallel solid-phase isothermal amplification and detection of multiple DNA targets in microliter-sized wells of a digital versatile disc

    Santiago-Felipe, Sara; Tortajada-Genaro, Luis Antonio; Puchades, Rosa; Maquieira, Ángel

    2016-01-01

    An integrated method for the parallelized detection of multiple DNA target sequences is presented by using microstructures in a digital versatile disc (DVD). Samples and reagents were managed by using both the capillary and centrifugal forces induced by disc rotation. Recombinase polymerase amplification (RPA), in a bridge solid phase format, took place in separate wells, which thereby modified their optical properties. Then the DVD drive reader recorded the modifications of the transmitted laser beam. The strategy allowed tens of genetic determinations to be made simultaneously within <2 h, with small sample volumes (3 μL), low manipulation and at low cost. The method was applied to high-throughput screening of relevant safety threats (allergens, GMOs and pathogenic bacteria) in food samples. Satisfactory results were obtained in terms of sensitivity (48.7 fg of DNA) and reproducibility (below 18 %). This scheme warrants cost-effective multiplex amplification and detection and is perceived to represent a viable tool for screening of nucleic acid targets. (author)

  2. Pharmacodynamic effects of steady-state fingolimod on antibody response in healthy volunteers: a 4-week, randomized, placebo-controlled, parallel-group, multiple-dose study.

    Boulton, Craig; Meiser, Karin; David, Olivier J; Schmouder, Robert

    2012-12-01

    Fingolimod, a first-in-class oral sphingosine 1-phosphate receptor (S1PR) modulator, is approved in many countries for relapsing-remitting multiple sclerosis, at a once-daily 0.5-mg dose. A reduction in peripheral lymphocyte count is an expected consequence of the fingolimod mechanism of S1PR modulation. The authors investigated if this pharmacodynamic effect impacts humoral and cellular immunogenicity. In this double-blind, parallel-group, 4-week study, 72 healthy volunteers were randomized to steady state, fingolimod 0.5 mg, 1.25 mg, or to placebo. The authors compared T-cell dependent and independent responses to the neoantigens, keyhole limpet hemocyanin (KLH), and pneumococcal polysaccharides vaccine (PPV-23), respectively, and additionally recall antigen response (tetanus toxoid [TT]) and delayed-type hypersensitivity (DTH) to KLH, TT, and Candida albicans. Fingolimod caused mild to moderate decreases in anti-KLH and anti-PPV-23 IgG and IgM levels versus placebo. Responder rates were identical between placebo and 0.5-mg groups for anti-KLH IgG (both > 90%) and comparable for anti-PPV-23 IgG (55% and 41%, respectively). Fingolimod did not affect anti-TT immunogenicity, and DTH response did not differ between placebo and fingolimod 0.5-mg groups. Expectedly, lymphocyte count reduced substantially in the fingolimod groups versus placebo but reversed by study end. Fingolimod was well tolerated, and the observed safety profile was consistent with previous reports.

  3. Parallel rendering

    Crockett, Thomas W.

    1995-01-01

    This article provides a broad introduction to the subject of parallel rendering, encompassing both hardware and software systems. The focus is on the underlying concepts and the issues which arise in the design of parallel rendering algorithms and systems. We examine the different types of parallelism and how they can be applied in rendering applications. Concepts from parallel computing, such as data decomposition, task granularity, scalability, and load balancing, are considered in relation to the rendering problem. We also explore concepts from computer graphics, such as coherence and projection, which have a significant impact on the structure of parallel rendering algorithms. Our survey covers a number of practical considerations as well, including the choice of architectural platform, communication and memory requirements, and the problem of image assembly and display. We illustrate the discussion with numerous examples from the parallel rendering literature, representing most of the principal rendering methods currently used in computer graphics.

  4. Parallel Computing Using Web Servers and "Servlets".

    Lo, Alfred; Bloor, Chris; Choi, Y. K.

    2000-01-01

    Describes parallel computing and presents inexpensive ways to implement a virtual parallel computer with multiple Web servers. Highlights include performance measurement of parallel systems; models for using Java and intranet technology including single server, multiple clients and multiple servers, single client; and a comparison of CGI (common…

  5. Parallel computations

    1982-01-01

    Parallel Computations focuses on parallel computation, with emphasis on algorithms used in a variety of numerical and physical applications and for many different types of parallel computers. Topics covered range from vectorization of fast Fourier transforms (FFTs) and of the incomplete Cholesky conjugate gradient (ICCG) algorithm on the Cray-1 to calculation of table lookups and piecewise functions. Single tridiagonal linear systems and vectorized computation of reactive flow are also discussed.Comprised of 13 chapters, this volume begins by classifying parallel computers and describing techn

  6. GenPlay Multi-Genome, a tool to compare and analyze multiple human genomes in a graphical interface.

    Lajugie, Julien; Fourel, Nicolas; Bouhassira, Eric E

    2015-01-01

    Parallel visualization of multiple individual human genomes is a complex endeavor that is rapidly gaining importance with the increasing number of personal, phased and cancer genomes that are being generated. It requires the display of variants such as SNPs, indels and structural variants that are unique to specific genomes and the introduction of multiple overlapping gaps in the reference sequence. Here, we describe GenPlay Multi-Genome, an application specifically written to visualize and analyze multiple human genomes in parallel. GenPlay Multi-Genome is ideally suited for the comparison of allele-specific expression and functional genomic data obtained from multiple phased genomes in a graphical interface with access to multiple-track operation. It also allows the analysis of data that have been aligned to custom genomes rather than to a standard reference and can be used as a variant calling format file browser and as a tool to compare different genome assembly, such as hg19 and hg38. GenPlay is available under the GNU public license (GPL-3) from http://genplay.einstein.yu.edu. The source code is available at https://github.com/JulienLajugie/GenPlay. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  7. The − 5 A/G single-nucleotide polymorphism in the core promoter region of MT2A and its effect on allele-specific gene expression and Cd, Zn and Cu levels in laryngeal cancer

    Starska, Katarzyna, E-mail: katarzyna.starska@umed.lodz.pl [I Department of Otolaryngology and Laryngological Oncology, Medical University of Łódź, Kopcinskiego 22, 90-153 Łódź (Poland); Krześlak, Anna; Forma, Ewa [Department of Cytobiochemistry, University of Łódź, Pomorska 142/143, 90-236 Łódź (Poland); Olszewski, Jurek [II Department of Otolaryngology and Laryngological Oncology, Medical University of Łódź, Żeromskiego 113, 90-549 Łódź (Poland); Morawiec-Sztandera, Alina [Department of Head and Neck Surgery, Medical University of Łódź, Paderewskiego 4, 93-509 Łódź (Poland); Aleksandrowicz, Paweł [Department of Otolaryngology and Laryngological Oncology, Medical University of Lublin, Jaczewskiego 8, 20-954 Lublin (Poland); Lewy-Trenda, Iwona [Department of Pathology, Medical University of Łódź, Pomorska 251, 92-213 Łódź (Poland); and others

    2014-10-15

    Metallothioneins (MTs) are low molecular weight, cysteine-rich heavy metal-binding proteins which participate in the mechanisms of Zn homeostasis, and protect against toxic metals. MTs contain metal-thiolate cluster groups and suppress metal toxicity by binding to them. The aim of this study was to determine the − 5 A/G (rs28366003) single-nucleotide polymorphism (SNP) in the core promoter region of the MT2A gene and to investigate its effect on allele-specific gene expression and Cd, Zn and Cu content in squamous cell laryngeal cancer (SCC) and non-cancerous laryngeal mucosa (NCM) as a control. The MT2A promoter region − 5 A/G SNP was determined by restriction fragment length polymorphism using 323 SCC and 116 NCM. MT2A gene analysis was performed by quantitative real-time PCR. The frequency of A allele carriage was 94.2% and 91.8% in SCC and NCM, respectively, while G allele carriage was detected in 5.8% and 8.2% of SCC and NCM samples, respectively. As a result, a significant association was identified between the − 5 A/G SNP in the MT2A gene with mRNA expression in both groups. Metal levels were analyzed by flame atomic absorption spectrometry. The significant differences were identified between A/A and both the A/G and G/G genotypes, with regard to the concentration of the contaminating metal. The Spearman rank correlation results showed that the MT2A expression and Cd, Zn, Cu levels were negatively correlated. Results obtained in this study suggest that − 5 A/G SNP in MT2A gene may have an effect on allele-specific gene expression and accumulation of metal levels in laryngeal cancer. - Highlights: • MT2A gene expression and metal content in laryngeal cancer tissues • Association between SNP (rs28366003) and expression of MT2A • Significant associations between the SNP and Cd, Zn and Cu levels • Negative correlation between MT2A gene expression and Cd, Zn and Cu levels.

  8. The − 5 A/G single-nucleotide polymorphism in the core promoter region of MT2A and its effect on allele-specific gene expression and Cd, Zn and Cu levels in laryngeal cancer

    Starska, Katarzyna; Krześlak, Anna; Forma, Ewa; Olszewski, Jurek; Morawiec-Sztandera, Alina; Aleksandrowicz, Paweł; Lewy-Trenda, Iwona

    2014-01-01

    Metallothioneins (MTs) are low molecular weight, cysteine-rich heavy metal-binding proteins which participate in the mechanisms of Zn homeostasis, and protect against toxic metals. MTs contain metal-thiolate cluster groups and suppress metal toxicity by binding to them. The aim of this study was to determine the − 5 A/G (rs28366003) single-nucleotide polymorphism (SNP) in the core promoter region of the MT2A gene and to investigate its effect on allele-specific gene expression and Cd, Zn and Cu content in squamous cell laryngeal cancer (SCC) and non-cancerous laryngeal mucosa (NCM) as a control. The MT2A promoter region − 5 A/G SNP was determined by restriction fragment length polymorphism using 323 SCC and 116 NCM. MT2A gene analysis was performed by quantitative real-time PCR. The frequency of A allele carriage was 94.2% and 91.8% in SCC and NCM, respectively, while G allele carriage was detected in 5.8% and 8.2% of SCC and NCM samples, respectively. As a result, a significant association was identified between the − 5 A/G SNP in the MT2A gene with mRNA expression in both groups. Metal levels were analyzed by flame atomic absorption spectrometry. The significant differences were identified between A/A and both the A/G and G/G genotypes, with regard to the concentration of the contaminating metal. The Spearman rank correlation results showed that the MT2A expression and Cd, Zn, Cu levels were negatively correlated. Results obtained in this study suggest that − 5 A/G SNP in MT2A gene may have an effect on allele-specific gene expression and accumulation of metal levels in laryngeal cancer. - Highlights: • MT2A gene expression and metal content in laryngeal cancer tissues • Association between SNP (rs28366003) and expression of MT2A • Significant associations between the SNP and Cd, Zn and Cu levels • Negative correlation between MT2A gene expression and Cd, Zn and Cu levels

  9. Mouse hippocampal GABAB1 but not GABAB2 subunit-containing receptor complex levels are paralleling retrieval in the multiple-T-maze

    Soheil eKeihan Falsafi

    2015-10-01

    Full Text Available GABAB receptors are heterodimeric G-protein coupled receptors known to be involved in learning and memory. Although a role for GABAB receptors in cognitive processes is evident, there is no information on hippocampal GABAB receptor complexes in a multiple T maze (MTM task, a robust paradigm for evaluation of spatial learning.Trained or untrained (yoked control C57BL/6J male mice (n=10/group were subjected to the MTM task and sacrificed 6 hours following their performance. Hippocampi were taken, membrane proteins extracted and run on blue native PAGE followed by immunoblotting with specific antibodies against GABAB1, GABAB1a and GABAB2. Immunoprecipitation with subsequent mass spectrometric identification of co-precipitates was carried out to show if GABAB1 and GABAB2 as well as other interacting proteins co-precipitate. An antibody shift assay (ASA and a proximity ligation assay (PLA were also used to see if the two GABAB subunits are present in the receptor complex.Single bands were observed on Western blots, each representing GABAB1, GABAB1a or GABAB2 at an apparent molecular weight of approximately 100 kDa. Subsequently, densitometric analysis revealed that levels of GABAB1 and GABAB1a but not GABAB2- containing receptor complexes were significantly higher in trained than untrained groups. Immunoprecipitation followed by mass spectrometric studies confirmed the presence of GABAB1, GABAB2, calcium calmodulin kinases I and II, GluA1 and GluA2 as constituents of the complex. ASA and PLA also showed the presence of the two subunits of GABAB receptor within the complex. It is shown that increased levels of GABAB1 subunit-containing complexes are paralleling performance in a land maze.

  10. Evaluation of pulsing magnetic field effects on paresthesia in multiple sclerosis patients, a randomized, double-blind, parallel-group clinical trial.

    Afshari, Daryoush; Moradian, Nasrin; Khalili, Majid; Razazian, Nazanin; Bostani, Arash; Hoseini, Jamal; Moradian, Mohamad; Ghiasian, Masoud

    2016-10-01

    Evidence is mounting that magnet therapy could alleviate the symptoms of multiple sclerosis (MS). This study was performed to test the effects of the pulsing magnetic fields on the paresthesia in MS patients. This study has been conducted as a randomized, double-blind, parallel-group clinical trial during the April 2012 to October 2013. The subjects were selected among patients referred to MS clinic of Imam Reza Hospital; affiliated to Kermanshah University of Medical Sciences, Iran. Sixty three patients with MS were included in the study and randomly were divided into two groups, 35 patients were exposed to a magnetic pulsing field of 4mT intensity and 15-Hz frequency sinusoidal wave for 20min per session 2 times per week over a period of 2 months involving 16 sessions and 28 patients was exposed to a magnetically inactive field (placebo) for 20min per session 2 times per week over a period of 2 months involving 16 sessions. The severity of paresthesia was measured by the numerical rating scale (NRS) at 30, 60days. The study primary end point was NRS change between baseline and 60days. The secondary outcome was NRS change between baseline and 30days. Patients exposing to magnetic field showed significant paresthesia improvement compared with the group of patients exposing to placebo. According to our results pulsed magnetic therapy could alleviate paresthesia in MS patients .But trials with more patients and longer duration are mandatory to describe long-term effects. Copyright © 2016 Elsevier B.V. All rights reserved.

  11. Comparison of allele-specific PCR, created restriction-site PCR, and PCR with primer-introduced restriction analysis methods used for screening complex vertebral malformation carriers in Holstein cattle

    Altınel, Ahmet

    2017-01-01

    Complex vertebral malformation (CVM) is an inherited, autosomal recessive disorder of Holstein cattle. The aim of this study was to compare sensitivity, specificity, positive and negative predictive values, accuracy, and rapidity of allele-specific polymerase chain reaction (AS-PCR), created restriction-site PCR (CRS-PCR), and PCR with primer-introduced restriction analysis (PCR-PIRA), three methods used in identification of CVM carriers in a Holstein cattle population. In order to screen for the G>T mutation in the solute carrier family 35 member A3 (SLC35A3) gene, DNA sequencing as the gold standard method was used. The prevalence of carriers and the mutant allele frequency were 3.2% and 0.016, respectively, among Holstein cattle in the Thrace region of Turkey. Among the three methods, the fastest but least accurate was AS-PCR. Although the rapidity of CRS-PCR and PCR-PIRA were nearly equal, the accuracy of PCR-PIRA was higher than that of CRS-PCR. Therefore, among the three methods, PCR-PIRA appears to be the most efficacious for screening of mutant alleles when identifying CVM carriers in a Holstein cattle population. PMID:28927256

  12. Development of an allele-specific, loop-mediated, isothermal amplification method (AS-LAMP to detect the L1014F kdr-w mutation in Anopheles gambiae s. l.

    Badolo Athanase

    2012-07-01

    Full Text Available Abstract Background Malaria control relies heavily on treated bed nets and indoor residual spraying with pyrethroid insecticides. Unfortunately, the resistance to pyrethroid insecticides, mainly due to the kdr mutation, is spreading in the main malaria vector Anopheles gambiae s.l., decreasing the insecticides’ efficacy. To manage the insecticide resistance rapidly and flexibly, simple and effective tools for the early detection of resistant mosquitoes are needed. This study aimed to develop an allele-specific, loop-mediated, isothermal amplification (AS-LAMP method to detect the West African-type kdr mutation (kdr-w; L1014F in field-collected mosquitoes. Methods DNA fragments of the wild-type and the mutated kdr gene were used to select the primers and develop the method. The primers were designed with the mutation at the 5’ end of the backward inner primer (BIP. The AS-LAMP method was compared to the AS-PCR method using the genomic DNA of 120 field-collected mosquitoes. Results The AS-LAMP method could discriminate between the wild-type homozygote, the heterozygote, and the kdr-w homozygote within 75 min. The AS-LAMP method has the advantage of being faster and at least as sensitive and specific as the AS-PCR method. Conclusions The AS-LAMP method can be used to detect the kdr mutation for quick decision-making, even in less well-equipped laboratories.

  13. Parallel algorithms

    Casanova, Henri; Robert, Yves

    2008-01-01

    ""…The authors of the present book, who have extensive credentials in both research and instruction in the area of parallelism, present a sound, principled treatment of parallel algorithms. … This book is very well written and extremely well designed from an instructional point of view. … The authors have created an instructive and fascinating text. The book will serve researchers as well as instructors who need a solid, readable text for a course on parallelism in computing. Indeed, for anyone who wants an understandable text from which to acquire a current, rigorous, and broad vi

  14. Parallel universes beguile science

    2007-01-01

    A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too. We may not be able -- as least not yet -- to prove they exist, many serious scientists say, but there are plenty of reasons to think that parallel dimensions are more than figments of eggheaded imagination.

  15. Validation of a Multiplex Allele-Specific Polymerase Chain Reaction Assay for Detection of KRAS Gene Mutations in Formalin-Fixed, Paraffin-Embedded Tissues from Colorectal Cancer Patients.

    Sirirat Seekhuntod

    Full Text Available Patients with KRAS mutations do not respond to epidermal growth factor receptor (EGFR inhibitors and fail to benefit from adjuvant chemotherapy. Mutation analysis of KRAS is needed before starting treatment with monoclonal anti-EGFR antibodies in patients with metastatic colorectal cancer (mCRC. The objective of this study is to develop a multiplex allele-specific PCR (MAS-PCR assay to detect KRAS mutations.We developed a single-tube MAS-PCR assay for the detection of seven KRAS mutations (G12D, G12A, G12R, G12C, G12S, G12V, and G13D. We performed MAS-PCR assay analysis for KRAS on DNA isolated from 270 formalin-fixed paraffin-embedded (FFPE colorectal cancer tissues. Sequences of all 270 samples were determined by pyrosequencing. Seven known point-mutation DNA samples diluted with wild-type DNA were assayed to determine the limitation of detection and reproducibility of the MAS-PCR assay.Overall, the results of MAS-PCR assay were in good concordance with pyrosequencing, and only seven discordant samples were found. The MAS-PCR assay reproducibly detected 1 to 2% mutant alleles. The most common mutations were G13D in codon 13 (49.17%, G12D (25.83% and G12V (12.50% in codon 12.The MAS-PCR assay provides a rapid, cost-effective, and reliable diagnostic tool for accurate detection of KRAS mutations in routine FFPE colorectal cancer tissues.

  16. Ultrascalable petaflop parallel supercomputer

    Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton On Hudson, NY; Chiu, George [Cross River, NY; Cipolla, Thomas M [Katonah, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Hall, Shawn [Pleasantville, NY; Haring, Rudolf A [Cortlandt Manor, NY; Heidelberger, Philip [Cortlandt Manor, NY; Kopcsay, Gerard V [Yorktown Heights, NY; Ohmacht, Martin [Yorktown Heights, NY; Salapura, Valentina [Chappaqua, NY; Sugavanam, Krishnan [Mahopac, NY; Takken, Todd [Brewster, NY

    2010-07-20

    A massively parallel supercomputer of petaOPS-scale includes node architectures based upon System-On-a-Chip technology, where each processing node comprises a single Application Specific Integrated Circuit (ASIC) having up to four processing elements. The ASIC nodes are interconnected by multiple independent networks that optimally maximize the throughput of packet communications between nodes with minimal latency. The multiple networks may include three high-speed networks for parallel algorithm message passing including a Torus, collective network, and a Global Asynchronous network that provides global barrier and notification functions. These multiple independent networks may be collaboratively or independently utilized according to the needs or phases of an algorithm for optimizing algorithm processing performance. The use of a DMA engine is provided to facilitate message passing among the nodes without the expenditure of processing resources at the node.

  17. Parallel magnetic resonance imaging

    Larkman, David J; Nunes, Rita G

    2007-01-01

    Parallel imaging has been the single biggest innovation in magnetic resonance imaging in the last decade. The use of multiple receiver coils to augment the time consuming Fourier encoding has reduced acquisition times significantly. This increase in speed comes at a time when other approaches to acquisition time reduction were reaching engineering and human limits. A brief summary of spatial encoding in MRI is followed by an introduction to the problem parallel imaging is designed to solve. There are a large number of parallel reconstruction algorithms; this article reviews a cross-section, SENSE, SMASH, g-SMASH and GRAPPA, selected to demonstrate the different approaches. Theoretical (the g-factor) and practical (coil design) limits to acquisition speed are reviewed. The practical implementation of parallel imaging is also discussed, in particular coil calibration. How to recognize potential failure modes and their associated artefacts are shown. Well-established applications including angiography, cardiac imaging and applications using echo planar imaging are reviewed and we discuss what makes a good application for parallel imaging. Finally, active research areas where parallel imaging is being used to improve data quality by repairing artefacted images are also reviewed. (invited topical review)

  18. Parallel encoders for pixel detectors

    Nikityuk, N.M.

    1991-01-01

    A new method of fast encoding and determining the multiplicity and coordinates of fired pixels is described. A specific example construction of parallel encodes and MCC for n=49 and t=2 is given. 16 refs.; 6 figs.; 2 tabs

  19. Detection of the V1016G mutation in the voltage-gated sodium channel gene of Aedes aegypti (Diptera: Culicidae) by allele-specific PCR assay, and its distribution and effect on deltamethrin resistance in Thailand.

    Stenhouse, Steven A; Plernsub, Suriya; Yanola, Jintana; Lumjuan, Nongkran; Dantrakool, Anchalee; Choochote, Wej; Somboon, Pradya

    2013-08-30

    Resistance to pyrethroid insecticides is widespread among populations of Aedes aegypti, the main vector for the dengue virus. Several different point mutations within the voltage-gated sodium channel (VGSC) gene contribute to such resistance. A mutation at position 1016 in domain II, segment 6 of the VGSC gene in Ae. aegypti leads to a valine to glycine substitution (V1016G) that confers resistance to deltamethrin. This study developed and utilized an allele-specific PCR (AS-PCR) assay that could be used to detect the V1016G mutation. The assay was validated against a number of sequenced DNA samples of known genotype and was determined to be in complete agreement. Larvae and pupae were collected from various localities throughout Thailand. Samples were reared to adulthood and their resistance status against deltamethrin was determined by standard WHO susceptibility bioassays. Deltamethrin-resistant and susceptible insects were then genotyped for the V1016G mutation. Additionally, some samples were genotyped for a second mutation at position 1534 in domain III (F1534C) which is also known to confer pyrethroid resistance. The bioassay results revealed an overall mortality of 77.6%. Homozygous 1016G individuals survived at higher rates than either heterozygous or wild-type (1016 V) mosquitoes. The 1016G mutation was significantly and positively associated with deltamethrin resistance and was widely distributed throughout Thailand. Interestingly, wild-type 1016 V mosquitoes tested were homozygous for the 1534C mutation, and all heterozygous mosquitoes were also heterozygous for 1534C. Mutant homozygous (G/G) mosquitoes expressed the wild-type (F/F) at position 1534. However, the presence of the 1534C mutation was not associated with deltamethrin resistance. Our bioassay results indicate that all populations sampled display some degree of resistance to deltamethrin. Homozygous 1016G mosquitoes were far likelier to survive such exposure. However, resistance in some

  20. Comparison of 454 Ultra-Deep Sequencing and Allele-Specific Real-Time PCR with Regard to the Detection of Emerging Drug-Resistant Minor HIV-1 Variants after Antiretroviral Prophylaxis for Vertical Transmission.

    Andrea Hauser

    Full Text Available Pregnant HIV-infected women were screened for the development of HIV-1 drug resistance after implementation of a triple-antiretroviral transmission prophylaxis as recommended by the WHO in 2006. The study offered the opportunity to compare amplicon-based 454 ultra-deep sequencing (UDS and allele-specific real-time PCR (ASPCR for the detection of drug-resistant minor variants in the HIV-1 reverse transcriptase (RT.Plasma samples from 34 Tanzanian women were previously analysed by ASPCR for key resistance mutations in the viral RT selected by AZT, 3TC, and NVP (K70R, K103N, Y181C, M184V, T215Y/F. In this study, the RT region of the same samples was investigated by amplicon-based UDS for resistance mutations using the 454 GS FLX System.Drug-resistant HIV-variants were identified in 69% (20/29 of women by UDS and in 45% (13/29 by ASPCR. The absolute number of resistance mutations identified by UDS was twice that identified by ASPCR (45 vs 24. By UDS 14 of 24 ASPCR-detected resistance mutations were identified at the same position. The overall concordance between UDS and ASPCR was 61.0% (25/41. The proportions of variants quantified by UDS were approximately 2-3 times lower than by ASPCR. Amplicon generation from samples with viral loads below 20,000 copies/ml failed more frequently by UDS compared to ASPCR (limit of detection = 650 copies/ml, resulting in missing or insufficient sequence coverage.Both methods can provide useful information about drug-resistant minor HIV-1 variants. ASPCR has a higher sensitivity than UDS, but is restricted to single resistance mutations. In contrast, UDS is limited by its requirement for high viral loads to achieve sufficient sequence coverage, but the sequence information reveals the complete resistance patterns within the genomic region analysed. Improvements to the UDS limit of detection are in progress, and UDS could then facilitate monitoring of drug-resistant minor variants in the HIV-1 quasispecies.

  1. Parallel computation

    Jejcic, A.; Maillard, J.; Maurel, G.; Silva, J.; Wolff-Bacha, F.

    1997-01-01

    The work in the field of parallel processing has developed as research activities using several numerical Monte Carlo simulations related to basic or applied current problems of nuclear and particle physics. For the applications utilizing the GEANT code development or improvement works were done on parts simulating low energy physical phenomena like radiation, transport and interaction. The problem of actinide burning by means of accelerators was approached using a simulation with the GEANT code. A program of neutron tracking in the range of low energies up to the thermal region has been developed. It is coupled to the GEANT code and permits in a single pass the simulation of a hybrid reactor core receiving a proton burst. Other works in this field refers to simulations for nuclear medicine applications like, for instance, development of biological probes, evaluation and characterization of the gamma cameras (collimators, crystal thickness) as well as the method for dosimetric calculations. Particularly, these calculations are suited for a geometrical parallelization approach especially adapted to parallel machines of the TN310 type. Other works mentioned in the same field refer to simulation of the electron channelling in crystals and simulation of the beam-beam interaction effect in colliders. The GEANT code was also used to simulate the operation of germanium detectors designed for natural and artificial radioactivity monitoring of environment

  2. The Galley Parallel File System

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    Most current multiprocessor file systems are designed to use multiple disks in parallel, using the high aggregate bandwidth to meet the growing I/0 requirements of parallel scientific applications. Many multiprocessor file systems provide applications with a conventional Unix-like interface, allowing the application to access multiple disks transparently. This interface conceals the parallelism within the file system, increasing the ease of programmability, but making it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. In addition to providing an insufficient interface, most current multiprocessor file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic scientific multiprocessor workloads. We discuss Galley's file structure and application interface, as well as the performance advantages offered by that interface.

  3. CELLS v1.0: updated and parallelized version of an electrical scheme to simulate multiple electrified clouds and flashes over large domains

    C. Barthe

    2012-01-01

    Full Text Available The paper describes the fully parallelized electrical scheme CELLS which is suitable to simulate explicitly electrified storm systems on parallel computers. Our motivation here is to show that a cloud electricity scheme can be developed for use on large grids with complex terrain. Large computational domains are needed to perform real case meteorological simulations with many independent convective cells.

    The scheme computes the bulk electric charge attached to each cloud particle and hydrometeor. Positive and negative ions are also taken into account. Several parametrizations of the dominant non-inductive charging process are included and an inductive charging process as well. The electric field is obtained by inverting the Gauss equation with an extension to terrain-following coordinates. The new feature concerns the lightning flash scheme which is a simplified version of an older detailed sequential scheme. Flashes are composed of a bidirectional leader phase (vertical extension from the triggering point and a phase obeying a fractal law (with horizontal extension on electrically charged zones. The originality of the scheme lies in the way the branching phase is treated to get a parallel code.

    The complete electrification scheme is tested for the 10 July 1996 STERAO case and for the 21 July 1998 EULINOX case. Flash characteristics are analysed in detail and additional sensitivity experiments are performed for the STERAO case. Although the simulations were run for flat terrain conditions, they show that the model behaves well on multiprocessor computers. This opens a wide area of application for this electrical scheme with the next objective of running real meterological case on large domains.

  4. Analysis of Heat Transfer and Pressure Drop for a Gas Flowing Through a set of Multiple Parallel Flat Plates at High Temperatures

    Einstein, Thomas H.

    1961-01-01

    Equations were derived representing heat transfer and pressure drop for a gas flowing in the passages of a heater composed of a series of parallel flat plates. The plates generated heat which was transferred to the flowing gas by convection. The relatively high temperature level of this system necessitated the consideration of heat transfer between the plates by radiation. The equations were solved on an IBM 704 computer, and results were obtained for hydrogen as the working fluid for a series of cases with a gas inlet temperature of 200 R, an exit temperature of 5000 0 R, and exit Mach numbers ranging from 0.2 to O.8. The length of the heater composed of the plates ranged from 2 to 4 feet, and the spacing between the plates was varied from 0.003 to 0.01 foot. Most of the results were for a five- plate heater, but results are also given for nine plates to show the effect of increasing the number of plates. The heat generation was assumed to be identical for each plate but was varied along the length of the plates. The axial variation of power used to obtain the results presented is the so-called "2/3-cosine variation." The boundaries surrounding the set of plates, and parallel to it, were assumed adiabatic, so that all the power generated in the plates went into heating the gas. The results are presented in plots of maximum plate and maximum adiabatic wall temperatures as functions of parameters proportional to f(L/D), for the case of both laminar and turbulent flow. Here f is the Fanning friction factor and (L/D) is the length to equivalent diameter ratio of the passages in the heater. The pressure drop through the heater is presented as a function of these same parameters, the exit Mach number, and the pressure at the exit of the heater.

  5. Parallel Lines

    James G. Worner

    2017-05-01

    Full Text Available James Worner is an Australian-based writer and scholar currently pursuing a PhD at the University of Technology Sydney. His research seeks to expose masculinities lost in the shadow of Australia’s Anzac hegemony while exploring new opportunities for contemporary historiography. He is the recipient of the Doctoral Scholarship in Historical Consciousness at the university’s Australian Centre of Public History and will be hosted by the University of Bologna during 2017 on a doctoral research writing scholarship.   ‘Parallel Lines’ is one of a collection of stories, The Shapes of Us, exploring liminal spaces of modern life: class, gender, sexuality, race, religion and education. It looks at lives, like lines, that do not meet but which travel in proximity, simultaneously attracted and repelled. James’ short stories have been published in various journals and anthologies.

  6. On the Organization of Parallel Operation of Some Algorithms for Finding the Shortest Path on a Graph on a Computer System with Multiple Instruction Stream and Single Data Stream

    V. E. Podol'skii

    2015-01-01

    Full Text Available The paper considers the implementing Bellman-Ford and Lee algorithms to find the shortest graph path on a computer system with multiple instruction stream and single data stream (MISD. The MISD computer is a computer that executes commands of arithmetic-logical processing (on the CPU and commands of structures processing (on the structures processor in parallel on a single data stream. Transformation of sequential programs into the MISD programs is a labor intensity process because it requires a stream of the arithmetic-logical processing to be manually separated from that of the structures processing. Algorithms based on the processing of data structures (e.g., algorithms on graphs show high performance on a MISD computer. Bellman-Ford and Lee algorithms for finding the shortest path on a graph are representatives of these algorithms. They are applied to robotics for automatic planning of the robot movement in-situ. Modification of Bellman-Ford and Lee algorithms for finding the shortest graph path in coprocessor MISD mode and the parallel MISD modification of these algorithms were first obtained in this article. Thus, this article continues a series of studies on the transformation of sequential algorithms into MISD ones (Dijkstra and Ford-Fulkerson 's algorithms and has a pronouncedly applied nature. The article also presents the analysis results of Bellman-Ford and Lee algorithms in MISD mode. The paper formulates the basic trends of a technique for parallelization of algorithms into arithmetic-logical processing stream and structures processing stream. Among the key areas for future research, development of the mathematical approach to provide a subsequently formalized and automated process of parallelizing sequential algorithms between the CPU and structures processor is highlighted. Among the mathematical models that can be used in future studies there are graph models of algorithms (e.g., dependency graph of a program. Due to the high

  7. System performances of optical space code-division multiple-access-based fiber-optic two-dimensional parallel data link.

    Nakamura, M; Kitayama, K

    1998-05-10

    Optical space code-division multiple access is a scheme to multiplex and link data between two-dimensional processors such as smart pixels and spatial light modulators or arrays of optical sources like vertical-cavity surface-emitting lasers. We examine the multiplexing characteristics of optical space code-division multiple access by using optical orthogonal signature patterns. The probability density function of interference noise in interfering optical orthogonal signature patterns is calculated. The bit-error rate is derived from the result and plotted as a function of receiver threshold, code length, code weight, and number of users. Furthermore, we propose a prethresholding method to suppress the interference noise, and we experimentally verify that the method works effectively in improving system performance.

  8. Amplificação gênica alelo-específica na caracterização das hemoglobinas S, C e D e as interações entre elas e talassemias beta Allele-specific genic amplification in the characterization of hemoglobins S, C, D and interactions among them and with beta thalassemia

    Luciane Cristina Bertholo

    2006-08-01

    possible interactions, based on the allele-specific genic amplification (PCR-AE with the use in parallel two primers that differ at their 3’ extremities and are complementary to the normal or mutated sequences. RESULTS AND DISCUSSION: The results make evident the validation of this methodology in the characterization of these mutations, once this procedure is easy to execute,to reproduce, as well as it is possible to be applied to a significative number of samples.

  9. An open 8-channel parallel transmission coil for static and dynamic 7T MRI of the knee and ankle joints at multiple postures.

    Jin, Jin; Weber, Ewald; Destruel, Aurelien; O'Brien, Kieran; Henin, Bassem; Engstrom, Craig; Crozier, Stuart

    2018-03-01

    We present the initial in vivo imaging results of an open architecture eight-channel parallel transmission (pTx) transceive radiofrequency (RF) coil array that was designed and constructed for static and dynamic 7T MRI of the knee and ankle joints. The pTx coil has a U-shaped dual-row configuration (200 mm overall length longitudinally) that allows static and dynamic imaging of the knee and ankle joints at various postures and during active movements. This coil structure, in combination with B 1 shimming, allows flexible configuration of B 1 transmit profiles, with good homogeneity over 120-mm regions of interest. This coil enabled high-resolution gradient echo (e.g., 3D dual-echo steady state [DESS] and 3D multiecho data image combination [MEDIC]) and turbo spin echo (TSE) imaging (e.g., with proton density weighting [PDw], PDw with fat saturation, and T 1 and T 2 weightings) with local RF energy absorption rates well below regulatory limits. High-resolution 2D and 3D image series (e.g., 0.3 mm in-plane resolution for TSE, 0.47 mm isotropic for DESS and MEDIC) were obtained from the knee and ankle joints with excellent tissue contrast. Dynamic imaging during continuous knee and ankle flexion-extension cycles were successfully acquired. The new open pTx coil array provides versatility for high-quality static and dynamic MRI of the knee and ankle joints at 7T. Magn Reson Med 79:1804-1816, 2018. © 2017 International Society for Magnetic Resonance in Medicine. © 2017 International Society for Magnetic Resonance in Medicine.

  10. Streaming for Functional Data-Parallel Languages

    Madsen, Frederik Meisner

    In this thesis, we investigate streaming as a general solution to the space inefficiency commonly found in functional data-parallel programming languages. The data-parallel paradigm maps well to parallel SIMD-style hardware. However, the traditional fully materializing execution strategy...... by extending two existing data-parallel languages: NESL and Accelerate. In the extensions we map bulk operations to data-parallel streams that can evaluate fully sequential, fully parallel or anything in between. By a dataflow, piecewise parallel execution strategy, the runtime system can adjust to any target...... flattening necessitates all sub-computations to materialize at the same time. For example, naive n by n matrix multiplication requires n^3 space in NESL because the algorithm contains n^3 independent scalar multiplications. For large values of n, this is completely unacceptable. We address the problem...

  11. A new system for parallel drug screening against multiple-resistant HIV mutants based on lentiviral self-inactivating (SIN vectors and multi-colour analyses

    Prokofjeva Maria M

    2013-01-01

    Full Text Available Abstract Background Despite progress in the development of combined antiretroviral therapies (cART, HIV infection remains a significant challenge for human health. Current problems of cART include multi-drug-resistant virus variants, long-term toxicity and enormous treatment costs. Therefore, the identification of novel effective drugs is urgently needed. Methods We developed a straightforward screening approach for simultaneously evaluating the sensitivity of multiple HIV gag-pol mutants to antiviral drugs in one assay. Our technique is based on multi-colour lentiviral self-inactivating (SIN LeGO vector technology. Results We demonstrated the successful use of this approach for screening compounds against up to four HIV gag-pol variants (wild-type and three mutants simultaneously. Importantly, the technique was adapted to Biosafety Level 1 conditions by utilising ecotropic pseudotypes. This allowed upscaling to a large-scale screening protocol exploited by pharmaceutical companies in a successful proof-of-concept experiment. Conclusions The technology developed here facilitates fast screening for anti-HIV activity of individual agents from large compound libraries. Although drugs targeting gag-pol variants were used here, our approach permits screening compounds that target several different, key cellular and viral functions of the HIV life-cycle. The modular principle of the method also allows the easy exchange of various mutations in HIV sequences. In conclusion, the methodology presented here provides a valuable new approach for the identification of novel anti-HIV drugs.

  12. Parallel integer sorting with medium and fine-scale parallelism

    Dagum, Leonardo

    1993-01-01

    Two new parallel integer sorting algorithms, queue-sort and barrel-sort, are presented and analyzed in detail. These algorithms do not have optimal parallel complexity, yet they show very good performance in practice. Queue-sort designed for fine-scale parallel architectures which allow the queueing of multiple messages to the same destination. Barrel-sort is designed for medium-scale parallel architectures with a high message passing overhead. The performance results from the implementation of queue-sort on a Connection Machine CM-2 and barrel-sort on a 128 processor iPSC/860 are given. The two implementations are found to be comparable in performance but not as good as a fully vectorized bucket sort on the Cray YMP.

  13. Exploiting Symmetry on Parallel Architectures.

    Stiller, Lewis Benjamin

    1995-01-01

    This thesis describes techniques for the design of parallel programs that solve well-structured problems with inherent symmetry. Part I demonstrates the reduction of such problems to generalized matrix multiplication by a group-equivariant matrix. Fast techniques for this multiplication are described, including factorization, orbit decomposition, and Fourier transforms over finite groups. Our algorithms entail interaction between two symmetry groups: one arising at the software level from the problem's symmetry and the other arising at the hardware level from the processors' communication network. Part II illustrates the applicability of our symmetry -exploitation techniques by presenting a series of case studies of the design and implementation of parallel programs. First, a parallel program that solves chess endgames by factorization of an associated dihedral group-equivariant matrix is described. This code runs faster than previous serial programs, and discovered it a number of results. Second, parallel algorithms for Fourier transforms for finite groups are developed, and preliminary parallel implementations for group transforms of dihedral and of symmetric groups are described. Applications in learning, vision, pattern recognition, and statistics are proposed. Third, parallel implementations solving several computational science problems are described, including the direct n-body problem, convolutions arising from molecular biology, and some communication primitives such as broadcast and reduce. Some of our implementations ran orders of magnitude faster than previous techniques, and were used in the investigation of various physical phenomena.

  14. Is Monte Carlo embarrassingly parallel?

    Hoogenboom, J. E. [Delft Univ. of Technology, Mekelweg 15, 2629 JB Delft (Netherlands); Delft Nuclear Consultancy, IJsselzoom 2, 2902 LB Capelle aan den IJssel (Netherlands)

    2012-07-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  15. Is Monte Carlo embarrassingly parallel?

    Hoogenboom, J. E.

    2012-01-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  16. Safety and pharmacokinetics of single and multiple intravenous bolus doses of diclofenac sodium compared with oral diclofenac potassium 50 mg: A randomized, parallel-group, single-center study in healthy subjects.

    Munjal, Sagar; Gautam, Anirudh; Okumu, Franklin; McDowell, James; Allenby, Kent

    2016-01-01

    In a randomized, parallel-group, single-center study in 42 healthy adults, the safety and pharmacokinetic parameters of an intravenous formulation of 18.75 and 37.5 mg diclofenac sodium (DFP-08) following single- and multiple-dose bolus administration were compared with diclofenac potassium 50 mg oral tablets. Mean AUC0-inf values for a 50-mg oral tablet and an 18.75-mg intravenous formulation were similar (1308.9 [393.0]) vs 1232.4 [147.6]). As measured by the AUC, DFP-08 18.75 mg and 37.5 mg demonstrated dose proportionality for extent of exposure. One subject in each of the placebo and DFP-08 18.75-mg groups and 2 subjects in the DFP-08 37.5-mg group reported adverse events that were considered by the investigator to be related to the study drug. All were mild in intensity and did not require treatment. Two subjects in the placebo group and 1 subject in the DFP-08 18.75-mg group reported grade 1 thrombophlebitis; no subjects reported higher than grade 1 thrombophlebitis after receiving a single intravenous dose. The 18.75- and 37.5-mg doses of intravenous diclofenac (single and multiple) were well tolerated for 7 days. Additional efficacy and safety studies are required to fully characterize the product. © 2015, The American College of Clinical Pharmacology.

  17. A faster, high resolution, mtPA-GFP-based mitochondrial fusion assay acquiring kinetic data of multiple cells in parallel using confocal microscopy.

    Lovy, Alenka; Molina, Anthony J A; Cerqueira, Fernanda M; Trudeau, Kyle; Shirihai, Orian S

    2012-07-20

    exposing loaded cells (3-15 nM TMRE) to the imaging parameters that will be used in the assay (perhaps 7 stacks of 6 optical sections in a row), and assessing cell health after 2 hours. If the mitochondria appear too fragmented and cells are dying, other mitochondrial markers, such as dsRED or Mitotracker red could be used instead of TMRE. The mtPAGFP method has revealed details about mitochondrial network behavior that could not be visualized using other methods. For example, we now know that mitochondrial fusion can be full or transient, where matrix content can mix without changing the overall network morphology. Additionally, we know that the probability of fusion is independent of contact duration and organelle dimension, is influenced by organelle motility, membrane potential and history of previous fusion activity. In this manuscript, we describe a methodology for scaling up the previously published protocol using mtPAGFP and 15 nM TMRE in order to examine multiple cells at a time and improve the time efficiency of data collection without sacrificing the subcellular resolution. This has been made possible by the use of an automated microscope stage, and programmable image acquisition software. Zen software from Zeiss allows the user to mark and track several designated cells expressing mtPAGFP. Each of these cells can be photoactivated in a particular region of interest, and stacks of confocal slices can be monitored for mtPAGFP signal as well as TMRE at specified intervals. Other confocal systems could be used to perform this protocol provided there is an automated stage that is programmable, an incubator with CO2, and a means by which to photoactivate the PAGFP; either a multiphoton laser, or a 405 nm diode laser.

  18. Distributed parallel messaging for multiprocessor systems

    Chen, Dong; Heidelberger, Philip; Salapura, Valentina; Senger, Robert M; Steinmacher-Burrow, Burhard; Sugawara, Yutaka

    2013-06-04

    A method and apparatus for distributed parallel messaging in a parallel computing system. The apparatus includes, at each node of a multiprocessor network, multiple injection messaging engine units and reception messaging engine units, each implementing a DMA engine and each supporting both multiple packet injection into and multiple reception from a network, in parallel. The reception side of the messaging unit (MU) includes a switch interface enabling writing of data of a packet received from the network to the memory system. The transmission side of the messaging unit, includes switch interface for reading from the memory system when injecting packets into the network.

  19. Synchronization Techniques in Parallel Discrete Event Simulation

    Lindén, Jonatan

    2018-01-01

    Discrete event simulation is an important tool for evaluating system models in many fields of science and engineering. To improve the performance of large-scale discrete event simulations, several techniques to parallelize discrete event simulation have been developed. In parallel discrete event simulation, the work of a single discrete event simulation is distributed over multiple processing elements. A key challenge in parallel discrete event simulation is to ensure that causally dependent ...

  20. Parallel Programming with Intel Parallel Studio XE

    Blair-Chappell , Stephen

    2012-01-01

    Optimize code for multi-core processors with Intel's Parallel Studio Parallel programming is rapidly becoming a "must-know" skill for developers. Yet, where to start? This teach-yourself tutorial is an ideal starting point for developers who already know Windows C and C++ and are eager to add parallelism to their code. With a focus on applying tools, techniques, and language extensions to implement parallelism, this essential resource teaches you how to write programs for multicore and leverage the power of multicore in your programs. Sharing hands-on case studies and real-world examples, the

  1. Structured Parallel Programming Patterns for Efficient Computation

    McCool, Michael; Robison, Arch

    2012-01-01

    Programming is now parallel programming. Much as structured programming revolutionized traditional serial programming decades ago, a new kind of structured programming, based on patterns, is relevant to parallel programming today. Parallel computing experts and industry insiders Michael McCool, Arch Robison, and James Reinders describe how to design and implement maintainable and efficient parallel algorithms using a pattern-based approach. They present both theory and practice, and give detailed concrete examples using multiple programming models. Examples are primarily given using two of th

  2. Integration of paper-based microarray and time-of-flight secondary ion mass spectrometry (ToF-SIMS) for parallel detection and quantification of molecules in multiple samples automatically.

    Chu, Kuo-Jui; Chen, Po-Chun; You, Yun-Wen; Chang, Hsun-Yun; Kao, Wei-Lun; Chu, Yi-Hsuan; Wu, Chen-Yi; Shyue, Jing-Jong

    2018-04-16

    With its low-cost fabrication and ease of modification, paper-based analytical devices have developed rapidly in recent years. Microarrays allow automatic analysis of multiple samples or multiple reactions with minimal sample consumption. While cellulose paper is generally used, its high backgrounds in spectrometry outside of the visible range has limited its application to be mostly colorimetric analysis. In this work, glass-microfiber paper is used as the substrate for a microarray. The glass-microfiber is essentially chemically inert SiO x , and the lower background from this inorganic microfiber can avoid interference from organic analytes in various spectrometers. However, generally used wax printing fails to wet glass microfibers to form hydrophobic barriers. Therefore, to prepare the hydrophobic-hydrophilic pattern, the glass-microfiber paper was first modified with an octadecyltrichlorosilane (OTS) self-assembled monolayer (SAM) to make the paper hydrophobic. A hydrophilic microarray was then prepared using a CO 2 laser scriber that selectively removed the OTS layer with a designed pattern. One microliter of aqueous drops of peptides at various concentrations were then dispensed inside the round patterns where OTS SAM was removed while the surrounding area with OTS layer served as a barrier to separate each drop. The resulting specimen of multiple spots was automatically analyzed with a time-of-flight secondary ion mass spectrometer (ToF-SIMS), and all of the secondary ions were collected. Among the various cluster ions that have developed over the past decade, pulsed C 60 + was selected as the primary ion because of its high secondary ion intensity in the high mass region, its minimal alteration of the surface when operating within the static-limit and spatial resolution at the ∼μm level. In the resulting spectra, parent ions of various peptides (in the forms [M+H] + and [M+Na] + ) were readily identified for parallel detection of molecules in a mixture

  3. Practical parallel computing

    Morse, H Stephen

    1994-01-01

    Practical Parallel Computing provides information pertinent to the fundamental aspects of high-performance parallel processing. This book discusses the development of parallel applications on a variety of equipment.Organized into three parts encompassing 12 chapters, this book begins with an overview of the technology trends that converge to favor massively parallel hardware over traditional mainframes and vector machines. This text then gives a tutorial introduction to parallel hardware architectures. Other chapters provide worked-out examples of programs using several parallel languages. Thi

  4. Parallel computing works!

    Fox, Geoffrey C; Messina, Guiseppe C

    2014-01-01

    A clear illustration of how parallel computers can be successfully appliedto large-scale scientific computations. This book demonstrates how avariety of applications in physics, biology, mathematics and other scienceswere implemented on real parallel computers to produce new scientificresults. It investigates issues of fine-grained parallelism relevant forfuture supercomputers with particular emphasis on hypercube architecture. The authors describe how they used an experimental approach to configuredifferent massively parallel machines, design and implement basic systemsoftware, and develop

  5. An allele-specific polymerase chain reaction assay for the ...

    Unknown

    designated as species A, B (Green and Miles 1980), C. (Subbarao et al 1983), ... onstrated in laboratory feeding experiments on three members of An. .... in Rameshwaram Island and Sri Lanka only) are not pre- sent. For example, in most of ...

  6. An allele-specific polymerase chain reaction assay for the ...

    Unknown

    designated as species A, B (Green and Miles 1980), C. (Subbarao et al 1983), ... 3 | September 2004. O P Singh et al. 276 able to distinguish species A from species B/C when sin- gle mosquito-extract was diluted to 1/200. However such hybridization assay .... in Rameshwaram Island and Sri Lanka only) are not pre- sent.

  7. Parallel Atomistic Simulations

    HEFFELFINGER,GRANT S.

    2000-01-18

    Algorithms developed to enable the use of atomistic molecular simulation methods with parallel computers are reviewed. Methods appropriate for bonded as well as non-bonded (and charged) interactions are included. While strategies for obtaining parallel molecular simulations have been developed for the full variety of atomistic simulation methods, molecular dynamics and Monte Carlo have received the most attention. Three main types of parallel molecular dynamics simulations have been developed, the replicated data decomposition, the spatial decomposition, and the force decomposition. For Monte Carlo simulations, parallel algorithms have been developed which can be divided into two categories, those which require a modified Markov chain and those which do not. Parallel algorithms developed for other simulation methods such as Gibbs ensemble Monte Carlo, grand canonical molecular dynamics, and Monte Carlo methods for protein structure determination are also reviewed and issues such as how to measure parallel efficiency, especially in the case of parallel Monte Carlo algorithms with modified Markov chains are discussed.

  8. Parallel Application Development Using Architecture View Driven Model Transformations

    Arkin, E.; Tekinerdogan, B.

    2015-01-01

    o realize the increased need for computing performance the current trend is towards applying parallel computing in which the tasks are run in parallel on multiple nodes. On its turn we can observe the rapid increase of the scale of parallel computing platforms. This situation has led to a complexity

  9. Applications of the parallel computing system using network

    Ido, Shunji; Hasebe, Hiroki

    1994-01-01

    Parallel programming is applied to multiple processors connected in Ethernet. Data exchanges between tasks located in each processing element are realized by two ways. One is socket which is standard library on recent UNIX operating systems. Another is a network connecting software, named as Parallel Virtual Machine (PVM) which is a free software developed by ORNL, to use many workstations connected to network as a parallel computer. This paper discusses the availability of parallel computing using network and UNIX workstations and comparison between specialized parallel systems (Transputer and iPSC/860) in a Monte Carlo simulation which generally shows high parallelization ratio. (author)

  10. Parallelization in Modern C++

    CERN. Geneva

    2016-01-01

    The traditionally used and well established parallel programming models OpenMP and MPI are both targeting lower level parallelism and are meant to be as language agnostic as possible. For a long time, those models were the only widely available portable options for developing parallel C++ applications beyond using plain threads. This has strongly limited the optimization capabilities of compilers, has inhibited extensibility and genericity, and has restricted the use of those models together with other, modern higher level abstractions introduced by the C++11 and C++14 standards. The recent revival of interest in the industry and wider community for the C++ language has also spurred a remarkable amount of standardization proposals and technical specifications being developed. Those efforts however have so far failed to build a vision on how to seamlessly integrate various types of parallelism, such as iterative parallel execution, task-based parallelism, asynchronous many-task execution flows, continuation s...

  11. Parallelism in matrix computations

    Gallopoulos, Efstratios; Sameh, Ahmed H

    2016-01-01

    This book is primarily intended as a research monograph that could also be used in graduate courses for the design of parallel algorithms in matrix computations. It assumes general but not extensive knowledge of numerical linear algebra, parallel architectures, and parallel programming paradigms. The book consists of four parts: (I) Basics; (II) Dense and Special Matrix Computations; (III) Sparse Matrix Computations; and (IV) Matrix functions and characteristics. Part I deals with parallel programming paradigms and fundamental kernels, including reordering schemes for sparse matrices. Part II is devoted to dense matrix computations such as parallel algorithms for solving linear systems, linear least squares, the symmetric algebraic eigenvalue problem, and the singular-value decomposition. It also deals with the development of parallel algorithms for special linear systems such as banded ,Vandermonde ,Toeplitz ,and block Toeplitz systems. Part III addresses sparse matrix computations: (a) the development of pa...

  12. Alternative derivation of the parallel ion viscosity

    Bravenec, R.V.; Berk, H.L.; Hammer, J.H.

    1982-01-01

    A set of double-adiabatic fluid equations with additional collisional relaxation between the ion temperatures parallel and perpendicular to a magnetic field are shown to reduce to a set involving a single temperature and a parallel viscosity. This result is applied to a recently published paper [R. V. Bravenec, A. J. Lichtenberg, M. A. Leiberman, and H. L. Berk, Phys. Fluids 24, 1320 (1981)] on viscous flow in a multiple-mirror configuration

  13. A parallel buffer tree

    Sitchinava, Nodar; Zeh, Norbert

    2012-01-01

    We present the parallel buffer tree, a parallel external memory (PEM) data structure for batched search problems. This data structure is a non-trivial extension of Arge's sequential buffer tree to a private-cache multiprocessor environment and reduces the number of I/O operations by the number of...... in the optimal OhOf(psortN + K/PB) parallel I/O complexity, where K is the size of the output reported in the process and psortN is the parallel I/O complexity of sorting N elements using P processors....

  14. Parallel MR imaging.

    Deshmane, Anagha; Gulani, Vikas; Griswold, Mark A; Seiberlich, Nicole

    2012-07-01

    Parallel imaging is a robust method for accelerating the acquisition of magnetic resonance imaging (MRI) data, and has made possible many new applications of MR imaging. Parallel imaging works by acquiring a reduced amount of k-space data with an array of receiver coils. These undersampled data can be acquired more quickly, but the undersampling leads to aliased images. One of several parallel imaging algorithms can then be used to reconstruct artifact-free images from either the aliased images (SENSE-type reconstruction) or from the undersampled data (GRAPPA-type reconstruction). The advantages of parallel imaging in a clinical setting include faster image acquisition, which can be used, for instance, to shorten breath-hold times resulting in fewer motion-corrupted examinations. In this article the basic concepts behind parallel imaging are introduced. The relationship between undersampling and aliasing is discussed and two commonly used parallel imaging methods, SENSE and GRAPPA, are explained in detail. Examples of artifacts arising from parallel imaging are shown and ways to detect and mitigate these artifacts are described. Finally, several current applications of parallel imaging are presented and recent advancements and promising research in parallel imaging are briefly reviewed. Copyright © 2012 Wiley Periodicals, Inc.

  15. Parallel Algorithms and Patterns

    Robey, Robert W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-06-16

    This is a powerpoint presentation on parallel algorithms and patterns. A parallel algorithm is a well-defined, step-by-step computational procedure that emphasizes concurrency to solve a problem. Examples of problems include: Sorting, searching, optimization, matrix operations. A parallel pattern is a computational step in a sequence of independent, potentially concurrent operations that occurs in diverse scenarios with some frequency. Examples are: Reductions, prefix scans, ghost cell updates. We only touch on parallel patterns in this presentation. It really deserves its own detailed discussion which Gabe Rockefeller would like to develop.

  16. Application Portable Parallel Library

    Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott

    1995-01-01

    Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.

  17. Performance of the Galley Parallel File System

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    As the input/output (I/O) needs of parallel scientific applications increase, file systems for multiprocessors are being designed to provide applications with parallel access to multiple disks. Many parallel file systems present applications with a conventional Unix-like interface that allows the application to access multiple disks transparently. This interface conceals the parallism within the file system, which increases the ease of programmability, but makes it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. Furthermore, most current parallel file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic parallel workloads. Initial experiments, reported in this paper, indicate that Galley is capable of providing high-performance 1/O to applications the applications that rely on them. In Section 3 we describe that access data in patterns that have been observed to be common.

  18. Shared Variable Oriented Parallel Precompiler for SPMD Model

    1995-01-01

    For the moment,commercial parallel computer systems with distributed memory architecture are usually provided with parallel FORTRAN or parallel C compliers,which are just traditional sequential FORTRAN or C compilers expanded with communication statements.Programmers suffer from writing parallel programs with communication statements. The Shared Variable Oriented Parallel Precompiler (SVOPP) proposed in this paper can automatically generate appropriate communication statements based on shared variables for SPMD(Single Program Multiple Data) computation model and greatly ease the parallel programming with high communication efficiency.The core function of parallel C precompiler has been successfully verified on a transputer-based parallel computer.Its prominent performance shows that SVOPP is probably a break-through in parallel programming technique.

  19. Parallel discrete event simulation

    Overeinder, B.J.; Hertzberger, L.O.; Sloot, P.M.A.; Withagen, W.J.

    1991-01-01

    In simulating applications for execution on specific computing systems, the simulation performance figures must be known in a short period of time. One basic approach to the problem of reducing the required simulation time is the exploitation of parallelism. However, in parallelizing the simulation

  20. Parallel reservoir simulator computations

    Hemanth-Kumar, K.; Young, L.C.

    1995-01-01

    The adaptation of a reservoir simulator for parallel computations is described. The simulator was originally designed for vector processors. It performs approximately 99% of its calculations in vector/parallel mode and relative to scalar calculations it achieves speedups of 65 and 81 for black oil and EOS simulations, respectively on the CRAY C-90

  1. Totally parallel multilevel algorithms

    Frederickson, Paul O.

    1988-01-01

    Four totally parallel algorithms for the solution of a sparse linear system have common characteristics which become quite apparent when they are implemented on a highly parallel hypercube such as the CM2. These four algorithms are Parallel Superconvergent Multigrid (PSMG) of Frederickson and McBryan, Robust Multigrid (RMG) of Hackbusch, the FFT based Spectral Algorithm, and Parallel Cyclic Reduction. In fact, all four can be formulated as particular cases of the same totally parallel multilevel algorithm, which are referred to as TPMA. In certain cases the spectral radius of TPMA is zero, and it is recognized to be a direct algorithm. In many other cases the spectral radius, although not zero, is small enough that a single iteration per timestep keeps the local error within the required tolerance.

  2. Parallel computing works

    1991-10-23

    An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

  3. Massively parallel mathematical sieves

    Montry, G.R.

    1989-01-01

    The Sieve of Eratosthenes is a well-known algorithm for finding all prime numbers in a given subset of integers. A parallel version of the Sieve is described that produces computational speedups over 800 on a hypercube with 1,024 processing elements for problems of fixed size. Computational speedups as high as 980 are achieved when the problem size per processor is fixed. The method of parallelization generalizes to other sieves and will be efficient on any ensemble architecture. We investigate two highly parallel sieves using scattered decomposition and compare their performance on a hypercube multiprocessor. A comparison of different parallelization techniques for the sieve illustrates the trade-offs necessary in the design and implementation of massively parallel algorithms for large ensemble computers.

  4. Icotinib versus whole-brain irradiation in patients with EGFR-mutant non-small-cell lung cancer and multiple brain metastases (BRAIN): a multicentre, phase 3, open-label, parallel, randomised controlled trial.

    Yang, Jin-Ji; Zhou, Caicun; Huang, Yisheng; Feng, Jifeng; Lu, Sun; Song, Yong; Huang, Cheng; Wu, Gang; Zhang, Li; Cheng, Ying; Hu, Chengping; Chen, Gongyan; Zhang, Li; Liu, Xiaoqing; Yan, Hong Hong; Tan, Fen Lai; Zhong, Wenzhao; Wu, Yi-Long

    2017-09-01

    For patients with non-small-cell lung cancer (NSCLC) and multiple brain metastases, whole-brain irradiation (WBI) is a standard-of-care treatment, but its effects on neurocognition are complex and concerning. We compared the efficacy of an epidermal growth factor receptor (EGFR)-tyrosine kinase inhibitor (TKI), icotinib, versus WBI with or without chemotherapy in a phase 3 trial of patients with EGFR-mutant NSCLC and multiple brain metastases. We did a multicentre, open-label, parallel randomised controlled trial (BRAIN) at 17 hospitals in China. Eligible participants were patients with NSCLC with EGFR mutations, who were naive to treatment with EGFR-TKIs or radiotherapy, and had at least three metastatic brain lesions. We randomly assigned participants (1:1) to either icotinib 125 mg orally (three times per day) or WBI (30 Gy in ten fractions of 3 Gy) plus concurrent or sequential chemotherapy for 4-6 cycles, until unacceptable adverse events or intracranial disease progression occurred. The randomisation was done by the Chinese Thoracic Oncology Group with a web-based allocation system applying the Pocock and Simon minimisation method; groups were stratified by EGFR gene mutation status, treatment line (first line or second line), brain metastases only versus both intracranial and extracranial metastases, and presence or absence of symptoms of intracranial hypertension. Clinicians and patients were not masked to treatment assignment, but individuals involved in the data analysis did not participate in the treatments and were thus masked to allocation. Patients receiving icotinib who had intracranial progression only were switched to WBI plus either icotinib or chemotherapy until further progression; those receiving icotinib who had extracranial progression only were switched to icotinib plus chemotherapy. Patients receiving WBI who progressed were switched to icotinib until further progression. Icotinib could be continued beyond progression if a clinical benefit

  5. Parallel computing and networking; Heiretsu keisanki to network

    Asakawa, E; Tsuru, T [Japan National Oil Corp., Tokyo (Japan); Matsuoka, T [Japan Petroleum Exploration Co. Ltd., Tokyo (Japan)

    1996-05-01

    This paper describes the trend of parallel computers used in geophysical exploration. Around 1993 was the early days when the parallel computers began to be used for geophysical exploration. Classification of these computers those days was mainly MIMD (multiple instruction stream, multiple data stream), SIMD (single instruction stream, multiple data stream) and the like. Parallel computers were publicized in the 1994 meeting of the Geophysical Exploration Society as a `high precision imaging technology`. Concerning the library of parallel computers, there was a shift to PVM (parallel virtual machine) in 1993 and to MPI (message passing interface) in 1995. In addition, the compiler of FORTRAN90 was released with support implemented for data parallel and vector computers. In 1993, networks used were Ethernet, FDDI, CDDI and HIPPI. In 1995, the OC-3 products under ATM began to propagate. However, ATM remains to be an interoffice high speed network because the ATM service has not spread yet for the public network. 1 ref.

  6. Algorithms for parallel computers

    Churchhouse, R.F.

    1985-01-01

    Until relatively recently almost all the algorithms for use on computers had been designed on the (usually unstated) assumption that they were to be run on single processor, serial machines. With the introduction of vector processors, array processors and interconnected systems of mainframes, minis and micros, however, various forms of parallelism have become available. The advantage of parallelism is that it offers increased overall processing speed but it also raises some fundamental questions, including: (i) which, if any, of the existing 'serial' algorithms can be adapted for use in the parallel mode. (ii) How close to optimal can such adapted algorithms be and, where relevant, what are the convergence criteria. (iii) How can we design new algorithms specifically for parallel systems. (iv) For multi-processor systems how can we handle the software aspects of the interprocessor communications. Aspects of these questions illustrated by examples are considered in these lectures. (orig.)

  7. Parallelism and array processing

    Zacharov, V.

    1983-01-01

    Modern computing, as well as the historical development of computing, has been dominated by sequential monoprocessing. Yet there is the alternative of parallelism, where several processes may be in concurrent execution. This alternative is discussed in a series of lectures, in which the main developments involving parallelism are considered, both from the standpoint of computing systems and that of applications that can exploit such systems. The lectures seek to discuss parallelism in a historical context, and to identify all the main aspects of concurrency in computation right up to the present time. Included will be consideration of the important question as to what use parallelism might be in the field of data processing. (orig.)

  8. Portable parallel programming in a Fortran environment

    May, E.N.

    1989-01-01

    Experience using the Argonne-developed PARMACs macro package to implement a portable parallel programming environment is described. Fortran programs with intrinsic parallelism of coarse and medium granularity are easily converted to parallel programs which are portable among a number of commercially available parallel processors in the class of shared-memory bus-based and local-memory network based MIMD processors. The parallelism is implemented using standard UNIX (tm) tools and a small number of easily understood synchronization concepts (monitors and message-passing techniques) to construct and coordinate multiple cooperating processes on one or many processors. Benchmark results are presented for parallel computers such as the Alliant FX/8, the Encore MultiMax, the Sequent Balance, the Intel iPSC/2 Hypercube and a network of Sun 3 workstations. These parallel machines are typical MIMD types with from 8 to 30 processors, each rated at from 1 to 10 MIPS processing power. The demonstration code used for this work is a Monte Carlo simulation of the response to photons of a ''nearly realistic'' lead, iron and plastic electromagnetic and hadronic calorimeter, using the EGS4 code system. 6 refs., 2 figs., 2 tabs

  9. Parallel optoelectronic trinary signed-digit division

    Alam, Mohammad S.

    1999-03-01

    The trinary signed-digit (TSD) number system has been found to be very useful for parallel addition and subtraction of any arbitrary length operands in constant time. Using the TSD addition and multiplication modules as the basic building blocks, we develop an efficient algorithm for performing parallel TSD division in constant time. The proposed division technique uses one TSD subtraction and two TSD multiplication steps. An optoelectronic correlator based architecture is suggested for implementation of the proposed TSD division algorithm, which fully exploits the parallelism and high processing speed of optics. An efficient spatial encoding scheme is used to ensure better utilization of space bandwidth product of the spatial light modulators used in the optoelectronic implementation.

  10. Stranger than fiction parallel universes beguile science

    2007-01-01

    A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too. We may not be able -- at least not yet -- to prove they exist, many serious scientists say, but there are plenty of reasons to think that parallel dimensions are more than figments of eggheaded imagination.

  11. Stranger that fiction parallel universes beguile science

    2007-01-01

    Is the universe -- correction: 'our' universe -- no more than a speck of cosmic dust amid an infinite number of parallel worlds? A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too.

  12. Stranger than fiction: parallel universes beguile science

    Hautefeuille, Annie

    2007-01-01

    Is the universe-correction: 'our' universe-no more than a speck of cosmic dust amid an infinite number of parallel worlds? A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too.

  13. The STAPL Parallel Graph Library

    Harshvardhan,; Fidel, Adam; Amato, Nancy M.; Rauchwerger, Lawrence

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable

  14. Massively parallel multicanonical simulations

    Gross, Jonathan; Zierenberg, Johannes; Weigel, Martin; Janke, Wolfhard

    2018-03-01

    Generalized-ensemble Monte Carlo simulations such as the multicanonical method and similar techniques are among the most efficient approaches for simulations of systems undergoing discontinuous phase transitions or with rugged free-energy landscapes. As Markov chain methods, they are inherently serial computationally. It was demonstrated recently, however, that a combination of independent simulations that communicate weight updates at variable intervals allows for the efficient utilization of parallel computational resources for multicanonical simulations. Implementing this approach for the many-thread architecture provided by current generations of graphics processing units (GPUs), we show how it can be efficiently employed with of the order of 104 parallel walkers and beyond, thus constituting a versatile tool for Monte Carlo simulations in the era of massively parallel computing. We provide the fully documented source code for the approach applied to the paradigmatic example of the two-dimensional Ising model as starting point and reference for practitioners in the field.

  15. SPINning parallel systems software

    Matlin, O.S.; Lusk, E.; McCune, W.

    2002-01-01

    We describe our experiences in using Spin to verify parts of the Multi Purpose Daemon (MPD) parallel process management system. MPD is a distributed collection of processes connected by Unix network sockets. MPD is dynamic processes and connections among them are created and destroyed as MPD is initialized, runs user processes, recovers from faults, and terminates. This dynamic nature is easily expressible in the Spin/Promela framework but poses performance and scalability challenges. We present here the results of expressing some of the parallel algorithms of MPD and executing both simulation and verification runs with Spin

  16. Parallel programming with Python

    Palach, Jan

    2014-01-01

    A fast, easy-to-follow and clear tutorial to help you develop Parallel computing systems using Python. Along with explaining the fundamentals, the book will also introduce you to slightly advanced concepts and will help you in implementing these techniques in the real world. If you are an experienced Python programmer and are willing to utilize the available computing resources by parallelizing applications in a simple way, then this book is for you. You are required to have a basic knowledge of Python development to get the most of this book.

  17. High Performance Parallel Multigrid Algorithms for Unstructured Grids

    Frederickson, Paul O.

    1996-01-01

    We describe a high performance parallel multigrid algorithm for a rather general class of unstructured grid problems in two and three dimensions. The algorithm PUMG, for parallel unstructured multigrid, is related in structure to the parallel multigrid algorithm PSMG introduced by McBryan and Frederickson, for they both obtain a higher convergence rate through the use of multiple coarse grids. Another reason for the high convergence rate of PUMG is its smoother, an approximate inverse developed by Baumgardner and Frederickson.

  18. Parallel transposition of sparse data structures

    Wang, Hao; Liu, Weifeng; Hou, Kaixi

    2016-01-01

    Many applications in computational sciences and social sciences exploit sparsity and connectivity of acquired data. Even though many parallel sparse primitives such as sparse matrix-vector (SpMV) multiplication have been extensively studied, some other important building blocks, e.g., parallel tr...... transposition in the latest vendor-supplied library on an Intel multicore CPU platform, and the MergeTrans approach achieves on average of 3.4-fold (up to 11.7-fold) speedup on an Intel Xeon Phi many-core processor....

  19. Temporal fringe pattern analysis with parallel computing

    Tuck Wah Ng; Kar Tien Ang; Argentini, Gianluca

    2005-01-01

    Temporal fringe pattern analysis is invaluable in transient phenomena studies but necessitates long processing times. Here we describe a parallel computing strategy based on the single-program multiple-data model and hyperthreading processor technology to reduce the execution time. In a two-node cluster workstation configuration we found that execution periods were reduced by 1.6 times when four virtual processors were used. To allow even lower execution times with an increasing number of processors, the time allocated for data transfer, data read, and waiting should be minimized. Parallel computing is found here to present a feasible approach to reduce execution times in temporal fringe pattern analysis

  20. Analysis of a parallel multigrid algorithm

    Chan, Tony F.; Tuminaro, Ray S.

    1989-01-01

    The parallel multigrid algorithm of Frederickson and McBryan (1987) is considered. This algorithm uses multiple coarse-grid problems (instead of one problem) in the hope of accelerating convergence and is found to have a close relationship to traditional multigrid methods. Specifically, the parallel coarse-grid correction operator is identical to a traditional multigrid coarse-grid correction operator, except that the mixing of high and low frequencies caused by aliasing error is removed. Appropriate relaxation operators can be chosen to take advantage of this property. Comparisons between the standard multigrid and the new method are made.

  1. Use of parallel counters for triggering

    Nikityuk, N.M.

    1991-01-01

    Results of investigation of using parallel counters, majority coincidence schemes, parallel compressors for triggering in multichannel high energy spectrometers are described. Concrete examples of methods of constructing fast and economic new devices used to determine multiplicity hits t>900 registered in a hodoscopic plane and a pixel detector are given. For this purpose the author uses the syndrome coding method and cellular arrays. In addition, an effective coding matrix has been created which can be used for light signal coding. For example, such signals are supplied from scintillators to photomultipliers. 23 refs.; 21 figs

  2. Expressing Parallelism with ROOT

    Piparo, D. [CERN; Tejedor, E. [CERN; Guiraud, E. [CERN; Ganis, G. [CERN; Mato, P. [CERN; Moneta, L. [CERN; Valls Pla, X. [CERN; Canal, P. [Fermilab

    2017-11-22

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  3. Expressing Parallelism with ROOT

    Piparo, D.; Tejedor, E.; Guiraud, E.; Ganis, G.; Mato, P.; Moneta, L.; Valls Pla, X.; Canal, P.

    2017-10-01

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  4. Parallel Fast Legendre Transform

    Alves de Inda, M.; Bisseling, R.H.; Maslen, D.K.

    1998-01-01

    We discuss a parallel implementation of a fast algorithm for the discrete polynomial Legendre transform We give an introduction to the DriscollHealy algorithm using polynomial arithmetic and present experimental results on the eciency and accuracy of our implementation The algorithms were

  5. Practical parallel programming

    Bauer, Barr E

    2014-01-01

    This is the book that will teach programmers to write faster, more efficient code for parallel processors. The reader is introduced to a vast array of procedures and paradigms on which actual coding may be based. Examples and real-life simulations using these devices are presented in C and FORTRAN.

  6. Parallel hierarchical radiosity rendering

    Carter, Michael [Iowa State Univ., Ames, IA (United States)

    1993-07-01

    In this dissertation, the step-by-step development of a scalable parallel hierarchical radiosity renderer is documented. First, a new look is taken at the traditional radiosity equation, and a new form is presented in which the matrix of linear system coefficients is transformed into a symmetric matrix, thereby simplifying the problem and enabling a new solution technique to be applied. Next, the state-of-the-art hierarchical radiosity methods are examined for their suitability to parallel implementation, and scalability. Significant enhancements are also discovered which both improve their theoretical foundations and improve the images they generate. The resultant hierarchical radiosity algorithm is then examined for sources of parallelism, and for an architectural mapping. Several architectural mappings are discussed. A few key algorithmic changes are suggested during the process of making the algorithm parallel. Next, the performance, efficiency, and scalability of the algorithm are analyzed. The dissertation closes with a discussion of several ideas which have the potential to further enhance the hierarchical radiosity method, or provide an entirely new forum for the application of hierarchical methods.

  7. Parallel k-means++

    2017-04-04

    A parallelization of the k-means++ seed selection algorithm on three distinct hardware platforms: GPU, multicore CPU, and multithreaded architecture. K-means++ was developed by David Arthur and Sergei Vassilvitskii in 2007 as an extension of the k-means data clustering technique. These algorithms allow people to cluster multidimensional data, by attempting to minimize the mean distance of data points within a cluster. K-means++ improved upon traditional k-means by using a more intelligent approach to selecting the initial seeds for the clustering process. While k-means++ has become a popular alternative to traditional k-means clustering, little work has been done to parallelize this technique. We have developed original C++ code for parallelizing the algorithm on three unique hardware architectures: GPU using NVidia's CUDA/Thrust framework, multicore CPU using OpenMP, and the Cray XMT multithreaded architecture. By parallelizing the process for these platforms, we are able to perform k-means++ clustering much more quickly than it could be done before.

  8. Parallel plate detectors

    Gardes, D.; Volkov, P.

    1981-01-01

    A 5x3cm 2 (timing only) and a 15x5cm 2 (timing and position) parallel plate avalanche counters (PPAC) are considered. The theory of operation and timing resolution is given. The measurement set-up and the curves of experimental results illustrate the possibilities of the two counters [fr

  9. Parallel hierarchical global illumination

    Snell, Quinn O. [Iowa State Univ., Ames, IA (United States)

    1997-10-08

    Solving the global illumination problem is equivalent to determining the intensity of every wavelength of light in all directions at every point in a given scene. The complexity of the problem has led researchers to use approximation methods for solving the problem on serial computers. Rather than using an approximation method, such as backward ray tracing or radiosity, the authors have chosen to solve the Rendering Equation by direct simulation of light transport from the light sources. This paper presents an algorithm that solves the Rendering Equation to any desired accuracy, and can be run in parallel on distributed memory or shared memory computer systems with excellent scaling properties. It appears superior in both speed and physical correctness to recent published methods involving bidirectional ray tracing or hybrid treatments of diffuse and specular surfaces. Like progressive radiosity methods, it dynamically refines the geometry decomposition where required, but does so without the excessive storage requirements for ray histories. The algorithm, called Photon, produces a scene which converges to the global illumination solution. This amounts to a huge task for a 1997-vintage serial computer, but using the power of a parallel supercomputer significantly reduces the time required to generate a solution. Currently, Photon can be run on most parallel environments from a shared memory multiprocessor to a parallel supercomputer, as well as on clusters of heterogeneous workstations.

  10. One weird trick for parallelizing convolutional neural networks

    Krizhevsky, Alex

    2014-01-01

    I present a new way to parallelize the training of convolutional neural networks across multiple GPUs. The method scales significantly better than all alternatives when applied to modern convolutional neural networks.

  11. Parallel computing of physical maps--a comparative study in SIMD and MIMD parallelism.

    Bhandarkar, S M; Chirravuri, S; Arnold, J

    1996-01-01

    Ordering clones from a genomic library into physical maps of whole chromosomes presents a central computational problem in genetics. Chromosome reconstruction via clone ordering is usually isomorphic to the NP-complete Optimal Linear Arrangement problem. Parallel SIMD and MIMD algorithms for simulated annealing based on Markov chain distribution are proposed and applied to the problem of chromosome reconstruction via clone ordering. Perturbation methods and problem-specific annealing heuristics are proposed and described. The SIMD algorithms are implemented on a 2048 processor MasPar MP-2 system which is an SIMD 2-D toroidal mesh architecture whereas the MIMD algorithms are implemented on an 8 processor Intel iPSC/860 which is an MIMD hypercube architecture. A comparative analysis of the various SIMD and MIMD algorithms is presented in which the convergence, speedup, and scalability characteristics of the various algorithms are analyzed and discussed. On a fine-grained, massively parallel SIMD architecture with a low synchronization overhead such as the MasPar MP-2, a parallel simulated annealing algorithm based on multiple periodically interacting searches performs the best. For a coarse-grained MIMD architecture with high synchronization overhead such as the Intel iPSC/860, a parallel simulated annealing algorithm based on multiple independent searches yields the best results. In either case, distribution of clonal data across multiple processors is shown to exacerbate the tendency of the parallel simulated annealing algorithm to get trapped in a local optimum.

  12. Parallel grid population

    Wald, Ingo; Ize, Santiago

    2015-07-28

    Parallel population of a grid with a plurality of objects using a plurality of processors. One example embodiment is a method for parallel population of a grid with a plurality of objects using a plurality of processors. The method includes a first act of dividing a grid into n distinct grid portions, where n is the number of processors available for populating the grid. The method also includes acts of dividing a plurality of objects into n distinct sets of objects, assigning a distinct set of objects to each processor such that each processor determines by which distinct grid portion(s) each object in its distinct set of objects is at least partially bounded, and assigning a distinct grid portion to each processor such that each processor populates its distinct grid portion with any objects that were previously determined to be at least partially bounded by its distinct grid portion.

  13. More parallel please

    Gregersen, Frans; Josephson, Olle; Kristoffersen, Gjert

    of departure that English may be used in parallel with the various local, in this case Nordic, languages. As such, the book integrates the challenge of internationalization faced by any university with the wish to improve quality in research, education and administration based on the local language......Abstract [en] More parallel, please is the result of the work of an Inter-Nordic group of experts on language policy financed by the Nordic Council of Ministers 2014-17. The book presents all that is needed to plan, practice and revise a university language policy which takes as its point......(s). There are three layers in the text: First, you may read the extremely brief version of the in total 11 recommendations for best practice. Second, you may acquaint yourself with the extended version of the recommendations and finally, you may study the reasoning behind each of them. At the end of the text, we give...

  14. PARALLEL MOVING MECHANICAL SYSTEMS

    Florian Ion Tiberius Petrescu

    2014-09-01

    Full Text Available Normal 0 false false false EN-US X-NONE X-NONE MicrosoftInternetExplorer4 Moving mechanical systems parallel structures are solid, fast, and accurate. Between parallel systems it is to be noticed Stewart platforms, as the oldest systems, fast, solid and precise. The work outlines a few main elements of Stewart platforms. Begin with the geometry platform, kinematic elements of it, and presented then and a few items of dynamics. Dynamic primary element on it means the determination mechanism kinetic energy of the entire Stewart platforms. It is then in a record tail cinematic mobile by a method dot matrix of rotation. If a structural mottoelement consists of two moving elements which translates relative, drive train and especially dynamic it is more convenient to represent the mottoelement as a single moving components. We have thus seven moving parts (the six motoelements or feet to which is added mobile platform 7 and one fixed.

  15. Xyce parallel electronic simulator.

    Keiter, Eric R; Mei, Ting; Russo, Thomas V.; Rankin, Eric Lamont; Schiek, Richard Louis; Thornquist, Heidi K.; Fixel, Deborah A.; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R.

    2010-05-01

    This document is a reference guide to the Xyce Parallel Electronic Simulator, and is a companion document to the Xyce Users Guide. The focus of this document is (to the extent possible) exhaustively list device parameters, solver options, parser options, and other usage details of Xyce. This document is not intended to be a tutorial. Users who are new to circuit simulation are better served by the Xyce Users Guide.

  16. Stability of parallel flows

    Betchov, R

    2012-01-01

    Stability of Parallel Flows provides information pertinent to hydrodynamical stability. This book explores the stability problems that occur in various fields, including electronics, mechanics, oceanography, administration, economics, as well as naval and aeronautical engineering. Organized into two parts encompassing 10 chapters, this book starts with an overview of the general equations of a two-dimensional incompressible flow. This text then explores the stability of a laminar boundary layer and presents the equation of the inviscid approximation. Other chapters present the general equation

  17. Algorithmically specialized parallel computers

    Snyder, Lawrence; Gannon, Dennis B

    1985-01-01

    Algorithmically Specialized Parallel Computers focuses on the concept and characteristics of an algorithmically specialized computer.This book discusses the algorithmically specialized computers, algorithmic specialization using VLSI, and innovative architectures. The architectures and algorithms for digital signal, speech, and image processing and specialized architectures for numerical computations are also elaborated. Other topics include the model for analyzing generalized inter-processor, pipelined architecture for search tree maintenance, and specialized computer organization for raster

  18. Mapping robust parallel multigrid algorithms to scalable memory architectures

    Overman, Andrea; Vanrosendale, John

    1993-01-01

    The convergence rate of standard multigrid algorithms degenerates on problems with stretched grids or anisotropic operators. The usual cure for this is the use of line or plane relaxation. However, multigrid algorithms based on line and plane relaxation have limited and awkward parallelism and are quite difficult to map effectively to highly parallel architectures. Newer multigrid algorithms that overcome anisotropy through the use of multiple coarse grids rather than relaxation are better suited to massively parallel architectures because they require only simple point-relaxation smoothers. In this paper, we look at the parallel implementation of a V-cycle multiple semicoarsened grid (MSG) algorithm on distributed-memory architectures such as the Intel iPSC/860 and Paragon computers. The MSG algorithms provide two levels of parallelism: parallelism within the relaxation or interpolation on each grid and across the grids on each multigrid level. Both levels of parallelism must be exploited to map these algorithms effectively to parallel architectures. This paper describes a mapping of an MSG algorithm to distributed-memory architectures that demonstrates how both levels of parallelism can be exploited. The result is a robust and effective multigrid algorithm for distributed-memory machines.

  19. Parallelization of Subchannel Analysis Code MATRA

    Kim, Seongjin; Hwang, Daehyun; Kwon, Hyouk

    2014-01-01

    A stand-alone calculation of MATRA code used up pertinent computing time for the thermal margin calculations while a relatively considerable time is needed to solve the whole core pin-by-pin problems. In addition, it is strongly required to improve the computation speed of the MATRA code to satisfy the overall performance of the multi-physics coupling calculations. Therefore, a parallel approach to improve and optimize the computability of the MATRA code is proposed and verified in this study. The parallel algorithm is embodied in the MATRA code using the MPI communication method and the modification of the previous code structure was minimized. An improvement is confirmed by comparing the results between the single and multiple processor algorithms. The speedup and efficiency are also evaluated when increasing the number of processors. The parallel algorithm was implemented to the subchannel code MATRA using the MPI. The performance of the parallel algorithm was verified by comparing the results with those from the MATRA with the single processor. It is also noticed that the performance of the MATRA code was greatly improved by implementing the parallel algorithm for the 1/8 core and whole core problems

  20. The BLAZE language - A parallel language for scientific programming

    Mehrotra, Piyush; Van Rosendale, John

    1987-01-01

    A Pascal-like scientific programming language, BLAZE, is described. BLAZE contains array arithmetic, forall loops, and APL-style accumulation operators, which allow natural expression of fine grained parallelism. It also employs an applicative or functional procedure invocation mechanism, which makes it easy for compilers to extract coarse grained parallelism using machine specific program restructuring. Thus BLAZE should allow one to achieve highly parallel execution on multiprocessor architectures, while still providing the user with conceptually sequential control flow. A central goal in the design of BLAZE is portability across a broad range of parallel architectures. The multiple levels of parallelism present in BLAZE code, in principle, allow a compiler to extract the types of parallelism appropriate for the given architecture while neglecting the remainder. The features of BLAZE are described and it is shown how this language would be used in typical scientific programming.

  1. The BLAZE language: A parallel language for scientific programming

    Mehrotra, P.; Vanrosendale, J.

    1985-01-01

    A Pascal-like scientific programming language, Blaze, is described. Blaze contains array arithmetic, forall loops, and APL-style accumulation operators, which allow natural expression of fine grained parallelism. It also employs an applicative or functional procedure invocation mechanism, which makes it easy for compilers to extract coarse grained parallelism using machine specific program restructuring. Thus Blaze should allow one to achieve highly parallel execution on multiprocessor architectures, while still providing the user with onceptually sequential control flow. A central goal in the design of Blaze is portability across a broad range of parallel architectures. The multiple levels of parallelism present in Blaze code, in principle, allow a compiler to extract the types of parallelism appropriate for the given architecture while neglecting the remainder. The features of Blaze are described and shows how this language would be used in typical scientific programming.

  2. A randomized, double-blind, placebo-controlled, multiple-dose, parallel-group clinical trial to assess the effects of teduglutide on gastric emptying of liquids in healthy subjects.

    Berg, Jolene Kay; Kim, Eric H; Li, Benjamin; Joelsson, Bo; Youssef, Nader N

    2014-02-12

    Teduglutide, a recombinant analog of human glucagon-like peptide (GLP)-2, is a novel therapy recently approved for the treatment of adult patients with short bowel syndrome who are dependent on parenteral support. Previous studies assessing the effect of GLP-2 on gastric emptying in humans have yielded inconsistent results, with some studies showing no effect and others documenting a GLP-2-dependent delay in gastric emptying. The primary objective of this study was to assess the effect of teduglutide on gastric emptying of liquids in healthy subjects, as measured by the pharmacokinetics of acetaminophen. This double-blind, parallel-group, single-center study enrolled and randomized 36 healthy subjects (22 men, 14 women) to receive subcutaneous doses of teduglutide 4 mg or placebo (2:1 ratio; 23:13) once daily on Days 1 through 10 in the morning. Gastric emptying of a mixed nutrient liquid meal was assessed by measuring acetaminophen levels predose and at 0.25, 0.5, 0.75, 1, 1.25, 1.5, 2, 3, 3.5, 4, 5, 6, 8, 10, 12, and 14 hours after administration of 1000 mg acetaminophen on Days 0 and 10. The primary study endpoint was a pharmacokinetic analysis of acetaminophen absorption in subjects receiving teduglutide or placebo. No significant differences in gastric emptying of liquids (acetaminophen area under the concentration [AUC] vs time curve from time 0 to the last measurable concentration, AUC extrapolated to infinity, maximum concentration [Cmax], and time to Cmax) were observed on Day 10 in subjects receiving teduglutide 4 mg versus subjects receiving placebo. There were no serious adverse events (AEs), deaths, or discontinuations due to an AE reported during the study. Teduglutide 4 mg/day for 10 days does not affect gastric emptying of liquids in healthy subjects as measured by acetaminophen pharmacokinetics. No unexpected safety signals were observed. This study was registered at ClinicalTrials.gov, identifier NCT01209351.

  3. A double-blind, randomized, multiple-dose, parallel-group study to characterize the occurrence of diarrhea following two different dosing regimens of neratinib, an irreversible pan-ErbB receptor tyrosine kinase inhibitor.

    Abbas, Richat; Hug, Bruce A; Leister, Cathie; Sonnichsen, Daryl

    2012-07-01

    Neratinib, a potent, low-molecular-weight, orally administered, irreversible, pan-ErbB receptor tyrosine kinase inhibitor has antitumor activity in ErbB2 + breast cancer. The objective of this study was to characterize the onset, severity, and duration of diarrhea after administration of neratinib 240 mg once daily (QD) and 120 mg twice daily (BID) for ≤14 days in healthy subjects. A randomized, double-blind, parallel-group, inpatient study was conducted in 50 subjects given oral neratinib either 240 mg QD or 120 mg BID with food for ≤14 days. The primary endpoint was the proportion of subjects with diarrhea of at least moderate severity (grade 2; 5-7 loose stools/day). In subjects with grade 2 diarrhea, fecal analytes were determined. Pharmacokinetic profiles were characterized for neratinib on Days 1 and 7. No severe (grade 3) diarrhea was reported. By Day 4, all subjects had grade 1 diarrhea. Grade 2 diarrhea occurred in 11/22 evaluable subjects (50 % [90 % confidence interval (CI): 28-72 %]) in the QD group and 17/23 evaluable subjects (74 % [90 % CI: 52-90 %]) in the BID group (P = 0.130). In fecal analyses, 18 % tested positive for hemoglobin and 46 % revealed fecal lactoferrin. Specimen pH was neutral to slightly alkaline. In pharmacokinetic analyses, Day 1 peak plasma concentration and Day 7 steady-state exposure were higher with the QD regimen than the BID regimen. In an exploratory analysis, ABCG2 genotype showed no correlation with severity or onset of diarrhea. Incidences and onsets of at least grade 1 and at least grade 2 diarrhea were not improved on BID dosing compared with QD dosing.

  4. Resistor Combinations for Parallel Circuits.

    McTernan, James P.

    1978-01-01

    To help simplify both teaching and learning of parallel circuits, a high school electricity/electronics teacher presents and illustrates the use of tables of values for parallel resistive circuits in which total resistances are whole numbers. (MF)

  5. SOFTWARE FOR DESIGNING PARALLEL APPLICATIONS

    M. K. Bouza

    2017-01-01

    Full Text Available The object of research is the tools to support the development of parallel programs in C/C ++. The methods and software which automates the process of designing parallel applications are proposed.

  6. Linear parallel processing machines I

    Von Kunze, M

    1984-01-01

    As is well-known, non-context-free grammars for generating formal languages happen to be of a certain intrinsic computational power that presents serious difficulties to efficient parsing algorithms as well as for the development of an algebraic theory of contextsensitive languages. In this paper a framework is given for the investigation of the computational power of formal grammars, in order to start a thorough analysis of grammars consisting of derivation rules of the form aB ..-->.. A/sub 1/ ... A /sub n/ b/sub 1/...b /sub m/ . These grammars may be thought of as automata by means of parallel processing, if one considers the variables as operators acting on the terminals while reading them right-to-left. This kind of automata and their 2-dimensional programming language prove to be useful by allowing a concise linear-time algorithm for integer multiplication. Linear parallel processing machines (LP-machines) which are, in their general form, equivalent to Turing machines, include finite automata and pushdown automata (with states encoded) as special cases. Bounded LP-machines yield deterministic accepting automata for nondeterministic contextfree languages, and they define an interesting class of contextsensitive languages. A characterization of this class in terms of generating grammars is established by using derivation trees with crossings as a helpful tool. From the algebraic point of view, deterministic LP-machines are effectively represented semigroups with distinguished subsets. Concerning the dualism between generating and accepting devices of formal languages within the algebraic setting, the concept of accepting automata turns out to reduce essentially to embeddability in an effectively represented extension monoid, even in the classical cases.

  7. Parallel External Memory Graph Algorithms

    Arge, Lars Allan; Goodrich, Michael T.; Sitchinava, Nodari

    2010-01-01

    In this paper, we study parallel I/O efficient graph algorithms in the Parallel External Memory (PEM) model, one o f the private-cache chip multiprocessor (CMP) models. We study the fundamental problem of list ranking which leads to efficient solutions to problems on trees, such as computing lowest...... an optimal speedup of ¿(P) in parallel I/O complexity and parallel computation time, compared to the single-processor external memory counterparts....

  8. Parallel inter channel interaction mechanisms

    Jovic, V.; Afgan, N.; Jovic, L.

    1995-01-01

    Parallel channels interactions are examined. For experimental researches of nonstationary regimes flow in three parallel vertical channels results of phenomenon analysis and mechanisms of parallel channel interaction for adiabatic condition of one-phase fluid and two-phase mixture flow are shown. (author)

  9. Massively Parallel QCD

    Soltz, R; Vranas, P; Blumrich, M; Chen, D; Gara, A; Giampap, M; Heidelberger, P; Salapura, V; Sexton, J; Bhanot, G

    2007-01-01

    The theory of the strong nuclear force, Quantum Chromodynamics (QCD), can be numerically simulated from first principles on massively-parallel supercomputers using the method of Lattice Gauge Theory. We describe the special programming requirements of lattice QCD (LQCD) as well as the optimal supercomputer hardware architectures that it suggests. We demonstrate these methods on the BlueGene massively-parallel supercomputer and argue that LQCD and the BlueGene architecture are a natural match. This can be traced to the simple fact that LQCD is a regular lattice discretization of space into lattice sites while the BlueGene supercomputer is a discretization of space into compute nodes, and that both are constrained by requirements of locality. This simple relation is both technologically important and theoretically intriguing. The main result of this paper is the speedup of LQCD using up to 131,072 CPUs on the largest BlueGene/L supercomputer. The speedup is perfect with sustained performance of about 20% of peak. This corresponds to a maximum of 70.5 sustained TFlop/s. At these speeds LQCD and BlueGene are poised to produce the next generation of strong interaction physics theoretical results

  10. A Parallel Butterfly Algorithm

    Poulson, Jack; Demanet, Laurent; Maxwell, Nicholas; Ying, Lexing

    2014-01-01

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  11. A Parallel Butterfly Algorithm

    Poulson, Jack

    2014-02-04

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  12. Fast parallel event reconstruction

    CERN. Geneva

    2010-01-01

    On-line processing of large data volumes produced in modern HEP experiments requires using maximum capabilities of modern and future many-core CPU and GPU architectures.One of such powerful feature is a SIMD instruction set, which allows packing several data items in one register and to operate on all of them, thus achievingmore operations per clock cycle. Motivated by the idea of using the SIMD unit ofmodern processors, the KF based track fit has been adapted for parallelism, including memory optimization, numerical analysis, vectorization with inline operator overloading, and optimization using SDKs. The speed of the algorithm has been increased in 120000 times with 0.1 ms/track, running in parallel on 16 SPEs of a Cell Blade computer.  Running on a Nehalem CPU with 8 cores it shows the processing speed of 52 ns/track using the Intel Threading Building Blocks. The same KF algorithm running on an Nvidia GTX 280 in the CUDA frameworkprovi...

  13. Neoclassical parallel flow calculation in the presence of external parallel momentum sources in Heliotron J

    Nishioka, K.; Nakamura, Y. [Graduate School of Energy Science, Kyoto University, Gokasho, Uji, Kyoto 611-0011 (Japan); Nishimura, S. [National Institute for Fusion Science, 322-6 Oroshi-cho, Toki, Gifu 509-5292 (Japan); Lee, H. Y. [Korea Advanced Institute of Science and Technology, Daejeon 305-701 (Korea, Republic of); Kobayashi, S.; Mizuuchi, T.; Nagasaki, K.; Okada, H.; Minami, T.; Kado, S.; Yamamoto, S.; Ohshima, S.; Konoshima, S.; Sano, F. [Institute of Advanced Energy, Kyoto University, Gokasho, Uji, Kyoto 611-0011 (Japan)

    2016-03-15

    A moment approach to calculate neoclassical transport in non-axisymmetric torus plasmas composed of multiple ion species is extended to include the external parallel momentum sources due to unbalanced tangential neutral beam injections (NBIs). The momentum sources that are included in the parallel momentum balance are calculated from the collision operators of background particles with fast ions. This method is applied for the clarification of the physical mechanism of the neoclassical parallel ion flows and the multi-ion species effect on them in Heliotron J NBI plasmas. It is found that parallel ion flow can be determined by the balance between the parallel viscosity and the external momentum source in the region where the external source is much larger than the thermodynamic force driven source in the collisional plasmas. This is because the friction between C{sup 6+} and D{sup +} prevents a large difference between C{sup 6+} and D{sup +} flow velocities in such plasmas. The C{sup 6+} flow velocities, which are measured by the charge exchange recombination spectroscopy system, are numerically evaluated with this method. It is shown that the experimentally measured C{sup 6+} impurity flow velocities do not contradict clearly with the neoclassical estimations, and the dependence of parallel flow velocities on the magnetic field ripples is consistent in both results.

  14. Hyper-systolic matrix multiplication

    Lippert, Th.; Petkov, N.; Palazzari, P.; Schilling, K.

    A novel parallel algorithm for matrix multiplication is presented. It is based on a 1-D hyper-systolic processor abstraction. The procedure can be implemented on all types of parallel systems. (C) 2001 Elsevier Science B,V. All rights reserved.

  15. Parallel Computing in SCALE

    DeHart, Mark D.; Williams, Mark L.; Bowman, Stephen M.

    2010-01-01

    The SCALE computational architecture has remained basically the same since its inception 30 years ago, although constituent modules and capabilities have changed significantly. This SCALE concept was intended to provide a framework whereby independent codes can be linked to provide a more comprehensive capability than possible with the individual programs - allowing flexibility to address a wide variety of applications. However, the current system was designed originally for mainframe computers with a single CPU and with significantly less memory than today's personal computers. It has been recognized that the present SCALE computation system could be restructured to take advantage of modern hardware and software capabilities, while retaining many of the modular features of the present system. Preliminary work is being done to define specifications and capabilities for a more advanced computational architecture. This paper describes the state of current SCALE development activities and plans for future development. With the release of SCALE 6.1 in 2010, a new phase of evolutionary development will be available to SCALE users within the TRITON and NEWT modules. The SCALE (Standardized Computer Analyses for Licensing Evaluation) code system developed by Oak Ridge National Laboratory (ORNL) provides a comprehensive and integrated package of codes and nuclear data for a wide range of applications in criticality safety, reactor physics, shielding, isotopic depletion and decay, and sensitivity/uncertainty (S/U) analysis. Over the last three years, since the release of version 5.1 in 2006, several important new codes have been introduced within SCALE, and significant advances applied to existing codes. Many of these new features became available with the release of SCALE 6.0 in early 2009. However, beginning with SCALE 6.1, a first generation of parallel computing is being introduced. In addition to near-term improvements, a plan for longer term SCALE enhancement

  16. Parallel Polarization State Generation.

    She, Alan; Capasso, Federico

    2016-05-17

    The control of polarization, an essential property of light, is of wide scientific and technological interest. The general problem of generating arbitrary time-varying states of polarization (SOP) has always been mathematically formulated by a series of linear transformations, i.e. a product of matrices, imposing a serial architecture. Here we show a parallel architecture described by a sum of matrices. The theory is experimentally demonstrated by modulating spatially-separated polarization components of a laser using a digital micromirror device that are subsequently beam combined. This method greatly expands the parameter space for engineering devices that control polarization. Consequently, performance characteristics, such as speed, stability, and spectral range, are entirely dictated by the technologies of optical intensity modulation, including absorption, reflection, emission, and scattering. This opens up important prospects for polarization state generation (PSG) with unique performance characteristics with applications in spectroscopic ellipsometry, spectropolarimetry, communications, imaging, and security.

  17. Parallel imaging microfluidic cytometer.

    Ehrlich, Daniel J; McKenna, Brian K; Evans, James G; Belkina, Anna C; Denis, Gerald V; Sherr, David H; Cheung, Man Ching

    2011-01-01

    By adding an additional degree of freedom from multichannel flow, the parallel microfluidic cytometer (PMC) combines some of the best features of fluorescence-activated flow cytometry (FCM) and microscope-based high-content screening (HCS). The PMC (i) lends itself to fast processing of large numbers of samples, (ii) adds a 1D imaging capability for intracellular localization assays (HCS), (iii) has a high rare-cell sensitivity, and (iv) has an unusual capability for time-synchronized sampling. An inability to practically handle large sample numbers has restricted applications of conventional flow cytometers and microscopes in combinatorial cell assays, network biology, and drug discovery. The PMC promises to relieve a bottleneck in these previously constrained applications. The PMC may also be a powerful tool for finding rare primary cells in the clinic. The multichannel architecture of current PMC prototypes allows 384 unique samples for a cell-based screen to be read out in ∼6-10 min, about 30 times the speed of most current FCM systems. In 1D intracellular imaging, the PMC can obtain protein localization using HCS marker strategies at many times for the sample throughput of charge-coupled device (CCD)-based microscopes or CCD-based single-channel flow cytometers. The PMC also permits the signal integration time to be varied over a larger range than is practical in conventional flow cytometers. The signal-to-noise advantages are useful, for example, in counting rare positive cells in the most difficult early stages of genome-wide screening. We review the status of parallel microfluidic cytometry and discuss some of the directions the new technology may take. Copyright © 2011 Elsevier Inc. All rights reserved.

  18. Parallel generation of architecture on the GPU

    Steinberger, Markus

    2014-05-01

    In this paper, we present a novel approach for the parallel evaluation of procedural shape grammars on the graphics processing unit (GPU). Unlike previous approaches that are either limited in the kind of shapes they allow, the amount of parallelism they can take advantage of, or both, our method supports state of the art procedural modeling including stochasticity and context-sensitivity. To increase parallelism, we explicitly express independence in the grammar, reduce inter-rule dependencies required for context-sensitive evaluation, and introduce intra-rule parallelism. Our rule scheduling scheme avoids unnecessary back and forth between CPU and GPU and reduces round trips to slow global memory by dynamically grouping rules in on-chip shared memory. Our GPU shape grammar implementation is multiple orders of magnitude faster than the standard in CPU-based rule evaluation, while offering equal expressive power. In comparison to the state of the art in GPU shape grammar derivation, our approach is nearly 50 times faster, while adding support for geometric context-sensitivity. © 2014 The Author(s) Computer Graphics Forum © 2014 The Eurographics Association and John Wiley & Sons Ltd. Published by John Wiley & Sons Ltd.

  19. About Parallel Programming: Paradigms, Parallel Execution and Collaborative Systems

    Loredana MOCEAN

    2009-01-01

    Full Text Available In the last years, there were made efforts for delineation of a stabile and unitary frame, where the problems of logical parallel processing must find solutions at least at the level of imperative languages. The results obtained by now are not at the level of the made efforts. This paper wants to be a little contribution at these efforts. We propose an overview in parallel programming, parallel execution and collaborative systems.

  20. Scalable parallel prefix solvers for discrete ordinates transport

    Pautz, S.; Pandya, T.; Adams, M.

    2009-01-01

    The well-known 'sweep' algorithm for inverting the streaming-plus-collision term in first-order deterministic radiation transport calculations has some desirable numerical properties. However, it suffers from parallel scaling issues caused by a lack of concurrency. The maximum degree of concurrency, and thus the maximum parallelism, grows more slowly than the problem size for sweeps-based solvers. We investigate a new class of parallel algorithms that involves recasting the streaming-plus-collision problem in prefix form and solving via cyclic reduction. This method, although computationally more expensive at low levels of parallelism than the sweep algorithm, offers better theoretical scalability properties. Previous work has demonstrated this approach for one-dimensional calculations; we show how to extend it to multidimensional calculations. Notably, for multiple dimensions it appears that this approach is limited to long-characteristics discretizations; other discretizations cannot be cast in prefix form. We implement two variants of the algorithm within the radlib/SCEPTRE transport code library at Sandia National Laboratories and show results on two different massively parallel systems. Both the 'forward' and 'symmetric' solvers behave similarly, scaling well to larger degrees of parallelism then sweeps-based solvers. We do observe some issues at the highest levels of parallelism (relative to the system size) and discuss possible causes. We conclude that this approach shows good potential for future parallel systems, but the parallel scalability will depend heavily on the architecture of the communication networks of these systems. (authors)

  1. Pharmacokinetic comparison of controlled-release and immediate-release oral formulations of simvastatin in healthy Korean subjects: a randomized, open-label, parallel-group, single- and multiple-dose study.

    Jang, Seong Bok; Lee, Yoon Jung; Lim, Lay Ahyoung; Park, Kyung-Mi; Kwon, Bong-Ju; Woo, Jong Soo; Kim, Yong-Il; Park, Min Soo; Kim, Kyung Hwan; Park, Kyungsoo

    2010-01-01

    A controlled-release (CR) formulation of simvastatin was recently developed in Korea. The formulation is expected to yield a lower C(max) and similar AUC values compared with the immediate-release (IR) formulation. The goal of this study was to compare the pharmacokinetics of the new CR formulation and an IR formulation of simvastatin after single- and multiple-dose administration in healthy Korean subjects. This study was developed as part of a product development project at the request of the Korean regulatory agency. This was a randomized, open-label, parallelgroup, 2-part study. Eligible subjects were healthy male or female volunteers between the ages of 19 and 55 years and within 20% of their ideal weight. In part I, each subject received a single dose of the CR or IR formulation of simvastatin 40 mg orally (20 mg x 2 tablets) after fasting. In part II, each subject received the same dose of the CR or IR formulation for 8 consecutive days. Blood samples were obtained for 48 hours after the dose in part I and after the first and the last dose in part II. Pharmacokinetic parameters were determined for both simvastatin (the inactive prodrug) and simvastatin acid (the active moiety). An adverse event (AE) was defined as any unfavorable sign (including an abnormal laboratory finding) or symptom, regardless of whether it had a causal relationship with the study medication. Serious AEs were defined as any events that are considered life threatening, require hospitalization or prolongation of existing hospitalization, cause persistent or significant disability or incapacity, or result in congenital abnormality, birth defect, or death. AEs were determined based on patient interviews and physical examinations. Twenty-four healthy subjects (17 men, 7 women; mean [SD] age, 29 [7] years; age range, 22-50 years) were enrolled in part I, and 29 subjects (17 men, 12 women; mean age, 33 [9] years; age range, 19-55 years) were enrolled in part II. For simvastatin acid, C

  2. Climate models on massively parallel computers

    Vitart, F.; Rouvillois, P.

    1993-01-01

    First results got on massively parallel computers (Multiple Instruction Multiple Data and Simple Instruction Multiple Data) allow to consider building of coupled models with high resolutions. This would make possible simulation of thermoaline circulation and other interaction phenomena between atmosphere and ocean. The increasing of computers powers, and then the improvement of resolution will go us to revise our approximations. Then hydrostatic approximation (in ocean circulation) will not be valid when the grid mesh will be of a dimension lower than a few kilometers: We shall have to find other models. The expert appraisement got in numerical analysis at the Center of Limeil-Valenton (CEL-V) will be used again to imagine global models taking in account atmosphere, ocean, ice floe and biosphere, allowing climate simulation until a regional scale

  3. Parallelization and automatic data distribution for nuclear reactor simulations

    Liebrock, L.M. [Liebrock-Hicks Research, Calumet, MI (United States)

    1997-07-01

    Detailed attempts at realistic nuclear reactor simulations currently take many times real time to execute on high performance workstations. Even the fastest sequential machine can not run these simulations fast enough to ensure that the best corrective measure is used during a nuclear accident to prevent a minor malfunction from becoming a major catastrophe. Since sequential computers have nearly reached the speed of light barrier, these simulations will have to be run in parallel to make significant improvements in speed. In physical reactor plants, parallelism abounds. Fluids flow, controls change, and reactions occur in parallel with only adjacent components directly affecting each other. These do not occur in the sequentialized manner, with global instantaneous effects, that is often used in simulators. Development of parallel algorithms that more closely approximate the real-world operation of a reactor may, in addition to speeding up the simulations, actually improve the accuracy and reliability of the predictions generated. Three types of parallel architecture (shared memory machines, distributed memory multicomputers, and distributed networks) are briefly reviewed as targets for parallelization of nuclear reactor simulation. Various parallelization models (loop-based model, shared memory model, functional model, data parallel model, and a combined functional and data parallel model) are discussed along with their advantages and disadvantages for nuclear reactor simulation. A variety of tools are introduced for each of the models. Emphasis is placed on the data parallel model as the primary focus for two-phase flow simulation. Tools to support data parallel programming for multiple component applications and special parallelization considerations are also discussed.

  4. Parallelization and automatic data distribution for nuclear reactor simulations

    Liebrock, L.M.

    1997-01-01

    Detailed attempts at realistic nuclear reactor simulations currently take many times real time to execute on high performance workstations. Even the fastest sequential machine can not run these simulations fast enough to ensure that the best corrective measure is used during a nuclear accident to prevent a minor malfunction from becoming a major catastrophe. Since sequential computers have nearly reached the speed of light barrier, these simulations will have to be run in parallel to make significant improvements in speed. In physical reactor plants, parallelism abounds. Fluids flow, controls change, and reactions occur in parallel with only adjacent components directly affecting each other. These do not occur in the sequentialized manner, with global instantaneous effects, that is often used in simulators. Development of parallel algorithms that more closely approximate the real-world operation of a reactor may, in addition to speeding up the simulations, actually improve the accuracy and reliability of the predictions generated. Three types of parallel architecture (shared memory machines, distributed memory multicomputers, and distributed networks) are briefly reviewed as targets for parallelization of nuclear reactor simulation. Various parallelization models (loop-based model, shared memory model, functional model, data parallel model, and a combined functional and data parallel model) are discussed along with their advantages and disadvantages for nuclear reactor simulation. A variety of tools are introduced for each of the models. Emphasis is placed on the data parallel model as the primary focus for two-phase flow simulation. Tools to support data parallel programming for multiple component applications and special parallelization considerations are also discussed

  5. Parallel Framework for Cooperative Processes

    Mitică Craus

    2005-01-01

    Full Text Available This paper describes the work of an object oriented framework designed to be used in the parallelization of a set of related algorithms. The idea behind the system we are describing is to have a re-usable framework for running several sequential algorithms in a parallel environment. The algorithms that the framework can be used with have several things in common: they have to run in cycles and the work should be possible to be split between several "processing units". The parallel framework uses the message-passing communication paradigm and is organized as a master-slave system. Two applications are presented: an Ant Colony Optimization (ACO parallel algorithm for the Travelling Salesman Problem (TSP and an Image Processing (IP parallel algorithm for the Symmetrical Neighborhood Filter (SNF. The implementations of these applications by means of the parallel framework prove to have good performances: approximatively linear speedup and low communication cost.

  6. Parallel multigrid smoothing: polynomial versus Gauss-Seidel

    Adams, Mark; Brezina, Marian; Hu, Jonathan; Tuminaro, Ray

    2003-01-01

    Gauss-Seidel is often the smoother of choice within multigrid applications. In the context of unstructured meshes, however, maintaining good parallel efficiency is difficult with multiplicative iterative methods such as Gauss-Seidel. This leads us to consider alternative smoothers. We discuss the computational advantages of polynomial smoothers within parallel multigrid algorithms for positive definite symmetric systems. Two particular polynomials are considered: Chebyshev and a multilevel specific polynomial. The advantages of polynomial smoothing over traditional smoothers such as Gauss-Seidel are illustrated on several applications: Poisson's equation, thin-body elasticity, and eddy current approximations to Maxwell's equations. While parallelizing the Gauss-Seidel method typically involves a compromise between a scalable convergence rate and maintaining high flop rates, polynomial smoothers achieve parallel scalable multigrid convergence rates without sacrificing flop rates. We show that, although parallel computers are the main motivation, polynomial smoothers are often surprisingly competitive with Gauss-Seidel smoothers on serial machines

  7. Parallel multigrid smoothing: polynomial versus Gauss-Seidel

    Adams, Mark; Brezina, Marian; Hu, Jonathan; Tuminaro, Ray

    2003-07-01

    Gauss-Seidel is often the smoother of choice within multigrid applications. In the context of unstructured meshes, however, maintaining good parallel efficiency is difficult with multiplicative iterative methods such as Gauss-Seidel. This leads us to consider alternative smoothers. We discuss the computational advantages of polynomial smoothers within parallel multigrid algorithms for positive definite symmetric systems. Two particular polynomials are considered: Chebyshev and a multilevel specific polynomial. The advantages of polynomial smoothing over traditional smoothers such as Gauss-Seidel are illustrated on several applications: Poisson's equation, thin-body elasticity, and eddy current approximations to Maxwell's equations. While parallelizing the Gauss-Seidel method typically involves a compromise between a scalable convergence rate and maintaining high flop rates, polynomial smoothers achieve parallel scalable multigrid convergence rates without sacrificing flop rates. We show that, although parallel computers are the main motivation, polynomial smoothers are often surprisingly competitive with Gauss-Seidel smoothers on serial machines.

  8. Rubus: A compiler for seamless and extensible parallelism.

    Muhammad Adnan

    Full Text Available Nowadays, a typical processor may have multiple processing cores on a single chip. Furthermore, a special purpose processing unit called Graphic Processing Unit (GPU, originally designed for 2D/3D games, is now available for general purpose use in computers and mobile devices. However, the traditional programming languages which were designed to work with machines having single core CPUs, cannot utilize the parallelism available on multi-core processors efficiently. Therefore, to exploit the extraordinary processing power of multi-core processors, researchers are working on new tools and techniques to facilitate parallel programming. To this end, languages like CUDA and OpenCL have been introduced, which can be used to write code with parallelism. The main shortcoming of these languages is that programmer needs to specify all the complex details manually in order to parallelize the code across multiple cores. Therefore, the code written in these languages is difficult to understand, debug and maintain. Furthermore, to parallelize legacy code can require rewriting a significant portion of code in CUDA or OpenCL, which can consume significant time and resources. Thus, the amount of parallelism achieved is proportional to the skills of the programmer and the time spent in code optimizations. This paper proposes a new open source compiler, Rubus, to achieve seamless parallelism. The Rubus compiler relieves the programmer from manually specifying the low-level details. It analyses and transforms a sequential program into a parallel program automatically, without any user intervention. This achieves massive speedup and better utilization of the underlying hardware without a programmer's expertise in parallel programming. For five different benchmarks, on average a speedup of 34.54 times has been achieved by Rubus as compared to Java on a basic GPU having only 96 cores. Whereas, for a matrix multiplication benchmark the average execution speedup of 84

  9. Parallel Monte Carlo reactor neutronics

    Blomquist, R.N.; Brown, F.B.

    1994-01-01

    The issues affecting implementation of parallel algorithms for large-scale engineering Monte Carlo neutron transport simulations are discussed. For nuclear reactor calculations, these include load balancing, recoding effort, reproducibility, domain decomposition techniques, I/O minimization, and strategies for different parallel architectures. Two codes were parallelized and tested for performance. The architectures employed include SIMD, MIMD-distributed memory, and workstation network with uneven interactive load. Speedups linear with the number of nodes were achieved

  10. Parallel scalability of Hartree-Fock calculations

    Chow, Edmond; Liu, Xing; Smelyanskiy, Mikhail; Hammond, Jeff R.

    2015-03-01

    Quantum chemistry is increasingly performed using large cluster computers consisting of multiple interconnected nodes. For a fixed molecular problem, the efficiency of a calculation usually decreases as more nodes are used, due to the cost of communication between the nodes. This paper empirically investigates the parallel scalability of Hartree-Fock calculations. The construction of the Fock matrix and the density matrix calculation are analyzed separately. For the former, we use a parallelization of Fock matrix construction based on a static partitioning of work followed by a work stealing phase. For the latter, we use density matrix purification from the linear scaling methods literature, but without using sparsity. When using large numbers of nodes for moderately sized problems, density matrix computations are network-bandwidth bound, making purification methods potentially faster than eigendecomposition methods.

  11. A parallel robot to assist vitreoretinal surgery

    Nakano, Taiga; Sugita, Naohiko; Mitsuishi, Mamoru [University of Tokyo, School of Engineering, Tokyo (Japan); Ueta, Takashi; Tamaki, Yasuhiro [University of Tokyo, Graduate School of Medicine, Tokyo (Japan)

    2009-11-15

    This paper describes the development and evaluation of a parallel prototype robot for vitreoretinal surgery where physiological hand tremor limits performance. The manipulator was specifically designed to meet requirements such as size, precision, and sterilization; this has six-degree-of-freedom parallel architecture and provides positioning accuracy with micrometer resolution within the eye. The manipulator is controlled by an operator with a ''master manipulator'' consisting of multiple joints. Results of the in vitro experiments revealed that when compared to the manual procedure, a higher stability and accuracy of tool positioning could be achieved using the prototype robot. This microsurgical system that we have developed has superior operability as compared to traditional manual procedure and has sufficient potential to be used clinically for vitreoretinal surgery. (orig.)

  12. Anti-parallel triplexes

    Kosbar, Tamer R.; Sofan, Mamdouh A.; Waly, Mohamed A.

    2015-01-01

    about 6.1 °C when the TFO strand was modified with Z and the Watson-Crick strand with adenine-LNA (AL). The molecular modeling results showed that, in case of nucleobases Y and Z a hydrogen bond (1.69 and 1.72 Å, respectively) was formed between the protonated 3-aminopropyn-1-yl chain and one...... of the phosphate groups in Watson-Crick strand. Also, it was shown that the nucleobase Y made a good stacking and binding with the other nucleobases in the TFO and Watson-Crick duplex, respectively. In contrast, the nucleobase Z with LNA moiety was forced to twist out of plane of Watson-Crick base pair which......The phosphoramidites of DNA monomers of 7-(3-aminopropyn-1-yl)-8-aza-7-deazaadenine (Y) and 7-(3-aminopropyn-1-yl)-8-aza-7-deazaadenine LNA (Z) are synthesized, and the thermal stability at pH 7.2 and 8.2 of anti-parallel triplexes modified with these two monomers is determined. When, the anti...

  13. Parallel consensual neural networks.

    Benediktsson, J A; Sveinsson, J R; Ersoy, O K; Swain, P H

    1997-01-01

    A new type of a neural-network architecture, the parallel consensual neural network (PCNN), is introduced and applied in classification/data fusion of multisource remote sensing and geographic data. The PCNN architecture is based on statistical consensus theory and involves using stage neural networks with transformed input data. The input data are transformed several times and the different transformed data are used as if they were independent inputs. The independent inputs are first classified using the stage neural networks. The output responses from the stage networks are then weighted and combined to make a consensual decision. In this paper, optimization methods are used in order to weight the outputs from the stage networks. Two approaches are proposed to compute the data transforms for the PCNN, one for binary data and another for analog data. The analog approach uses wavelet packets. The experimental results obtained with the proposed approach show that the PCNN outperforms both a conjugate-gradient backpropagation neural network and conventional statistical methods in terms of overall classification accuracy of test data.

  14. A Parallel Particle Swarm Optimizer

    Schutte, J. F; Fregly, B .J; Haftka, R. T; George, A. D

    2003-01-01

    .... Motivated by a computationally demanding biomechanical system identification problem, we introduce a parallel implementation of a stochastic population based global optimizer, the Particle Swarm...

  15. Patterns for Parallel Software Design

    Ortega-Arjona, Jorge Luis

    2010-01-01

    Essential reading to understand patterns for parallel programming Software patterns have revolutionized the way we think about how software is designed, built, and documented, and the design of parallel software requires you to consider other particular design aspects and special skills. From clusters to supercomputers, success heavily depends on the design skills of software developers. Patterns for Parallel Software Design presents a pattern-oriented software architecture approach to parallel software design. This approach is not a design method in the classic sense, but a new way of managin

  16. Seeing or moving in parallel

    Christensen, Mark Schram; Ehrsson, H Henrik; Nielsen, Jens Bo

    2013-01-01

    a different network, involving bilateral dorsal premotor cortex (PMd), primary motor cortex, and SMA, was more active when subjects viewed parallel movements while performing either symmetrical or parallel movements. Correlations between behavioral instability and brain activity were present in right lateral...... adduction-abduction movements symmetrically or in parallel with real-time congruent or incongruent visual feedback of the movements. One network, consisting of bilateral superior and middle frontal gyrus and supplementary motor area (SMA), was more active when subjects performed parallel movements, whereas...

  17. Parallel grid generation algorithm for distributed memory computers

    Moitra, Stuti; Moitra, Anutosh

    1994-01-01

    A parallel grid-generation algorithm and its implementation on the Intel iPSC/860 computer are described. The grid-generation scheme is based on an algebraic formulation of homotopic relations. Methods for utilizing the inherent parallelism of the grid-generation scheme are described, and implementation of multiple levELs of parallelism on multiple instruction multiple data machines are indicated. The algorithm is capable of providing near orthogonality and spacing control at solid boundaries while requiring minimal interprocessor communications. Results obtained on the Intel hypercube for a blended wing-body configuration are used to demonstrate the effectiveness of the algorithm. Fortran implementations bAsed on the native programming model of the iPSC/860 computer and the Express system of software tools are reported. Computational gains in execution time speed-up ratios are given.

  18. PARALLEL IMPORT: REALITY FOR RUSSIA

    Т. А. Сухопарова

    2014-01-01

    Full Text Available Problem of parallel import is urgent question at now. Parallel import legalization in Russia is expedient. Such statement based on opposite experts opinion analysis. At the same time it’s necessary to negative consequences consider of this decision and to apply remedies to its minimization.Purchase on Elibrary.ru > Buy now

  19. Parallelization of the FLAPW method

    Canning, A.; Mannstadt, W.; Freeman, A.J.

    1999-01-01

    The FLAPW (full-potential linearized-augmented plane-wave) method is one of the most accurate first-principles methods for determining electronic and magnetic properties of crystals and surfaces. Until the present work, the FLAPW method has been limited to systems of less than about one hundred atoms due to a lack of an efficient parallel implementation to exploit the power and memory of parallel computers. In this work we present an efficient parallelization of the method by division among the processors of the plane-wave components for each state. The code is also optimized for RISC (reduced instruction set computer) architectures, such as those found on most parallel computers, making full use of BLAS (basic linear algebra subprograms) wherever possible. Scaling results are presented for systems of up to 686 silicon atoms and 343 palladium atoms per unit cell, running on up to 512 processors on a CRAY T3E parallel computer

  20. Parallelization of the FLAPW method

    Canning, A.; Mannstadt, W.; Freeman, A. J.

    2000-08-01

    The FLAPW (full-potential linearized-augmented plane-wave) method is one of the most accurate first-principles methods for determining structural, electronic and magnetic properties of crystals and surfaces. Until the present work, the FLAPW method has been limited to systems of less than about a hundred atoms due to the lack of an efficient parallel implementation to exploit the power and memory of parallel computers. In this work, we present an efficient parallelization of the method by division among the processors of the plane-wave components for each state. The code is also optimized for RISC (reduced instruction set computer) architectures, such as those found on most parallel computers, making full use of BLAS (basic linear algebra subprograms) wherever possible. Scaling results are presented for systems of up to 686 silicon atoms and 343 palladium atoms per unit cell, running on up to 512 processors on a CRAY T3E parallel supercomputer.

  1. A Massively Parallel Code for Polarization Calculations

    Akiyama, Shizuka; Höflich, Peter

    2001-03-01

    We present an implementation of our Monte-Carlo radiation transport method for rapidly expanding, NLTE atmospheres for massively parallel computers which utilizes both the distributed and shared memory models. This allows us to take full advantage of the fast communication and low latency inherent to nodes with multiple CPUs, and to stretch the limits of scalability with the number of nodes compared to a version which is based on the shared memory model. Test calculations on a local 20-node Beowulf cluster with dual CPUs showed an improved scalability by about 40%.

  2. A parallel input composite transimpedance amplifier

    Kim, D. J.; Kim, C.

    2018-01-01

    A new approach to high performance current to voltage preamplifier design is presented. The design using multiple operational amplifiers (op-amps) has a parasitic capacitance compensation network and a composite amplifier topology for fast, precision, and low noise performance. The input stage consisting of a parallel linked JFET op-amps and a high-speed bipolar junction transistor (BJT) gain stage driving the output in the composite amplifier topology, cooperating with the capacitance compensation feedback network, ensures wide bandwidth stability in the presence of input capacitance above 40 nF. The design is ideal for any two-probe measurement, including high impedance transport and scanning tunneling microscopy measurements.

  3. Practical parallel processing

    Arendt, M.L.

    1986-01-01

    ELXSI, a San Jose based computer company, was founded in January of 1979 for the purpose of developing and marketing a tightly-coupled multiple processor system. After five years ELXSI succeeded in making the first commercial installations at Digicon Geophysical, NASA-Dryden, and Sandia National Laboratories. Since that time over fifty-one systems and ninety-three processors have been installed. The commercial success of the ELXSI system 6400(TM) is due to several significant breakthroughs in computer technology including a system bus operating at 320 million bytes per second, a new Message-Based Operating System, EMBOS (TM), and a new system organization which allows for easy expansion in any dimension without changes to the operating system, the user environment, or the application programs. (Auth.)

  4. Explorations of the implementation of a parallel IDW interpolation algorithm in a Linux cluster-based parallel GIS

    Huang, Fang; Liu, Dingsheng; Tan, Xicheng; Wang, Jian; Chen, Yunping; He, Binbin

    2011-04-01

    To design and implement an open-source parallel GIS (OP-GIS) based on a Linux cluster, the parallel inverse distance weighting (IDW) interpolation algorithm has been chosen as an example to explore the working model and the principle of algorithm parallel pattern (APP), one of the parallelization patterns for OP-GIS. Based on an analysis of the serial IDW interpolation algorithm of GRASS GIS, this paper has proposed and designed a specific parallel IDW interpolation algorithm, incorporating both single process, multiple data (SPMD) and master/slave (M/S) programming modes. The main steps of the parallel IDW interpolation algorithm are: (1) the master node packages the related information, and then broadcasts it to the slave nodes; (2) each node calculates its assigned data extent along one row using the serial algorithm; (3) the master node gathers the data from all nodes; and (4) iterations continue until all rows have been processed, after which the results are outputted. According to the experiments performed in the course of this work, the parallel IDW interpolation algorithm can attain an efficiency greater than 0.93 compared with similar algorithms, which indicates that the parallel algorithm can greatly reduce processing time and maximize speed and performance.

  5. Power Consumption Optimization for Multiple Parallel Centrifugal Pumps

    Jepsen, Kasper Lund; Hansen, Leif; Mai, Christian

    2017-01-01

    Large amounts of energy is being used in a wide range of applications to transport liquid. This paper proposes a generic solution for minimizing power consumption of a generic pumping station equipped with identical variable speed pumps. The proposed solution consists of two sequential steps; fir...

  6. Parallel Detection of Multiple Biomarkers During Spaceflight, Phase I

    National Aeronautics and Space Administration — Maintaining the health of astronauts during extended spaceflight is critical to the success of the mission. Radiation Monitoring Devices, Inc. (RMD) proposes an...

  7. Communication-Avoiding Parallel Recursive Algorithms for Matrix Multiplication

    2013-05-17

    processor, and an Nvidia K20 GPU. As of November 2012, it ranked first on the TOP500 list [53], with a LINPACK score of 17.59 Tflop/s. In our... NVIDIA , Oracle, and Samsung, and support from MathWorks. We also acknowledge the support of the US DOE (grants DE- SC0003959, DE-SC0004938, DE...Data Base and a New Technique in File Sequencing. International Business Machines Company , 1966. BIBLIOGRAPHY 83 [55] J.-S. Park, M. Penner, and V. K

  8. Parallel Low-Loss Measurement of Multiple Atomic Qubits.

    Kwon, Minho; Ebert, Matthew F; Walker, Thad G; Saffman, M

    2017-11-03

    We demonstrate low-loss measurement of the hyperfine ground state of rubidium atoms by state dependent fluorescence detection in a dipole trap array of five sites. The presence of atoms and their internal states are minimally altered by utilizing circularly polarized probe light and a strictly controlled quantization axis. We achieve mean state detection fidelity of 97% without correcting for imperfect state preparation or background losses, and 98.7% when corrected. After state detection and correction for background losses, the probability of atom loss due to the state measurement is state is preserved with >98% probability.

  9. Parallel computing solution of Boltzmann neutron transport equation

    Ansah-Narh, T.

    2010-01-01

    The focus of the research was on developing parallel computing algorithm for solving Eigen-values of the Boltzmam Neutron Transport Equation (BNTE) in a slab geometry using multi-grid approach. In response to the problem of slow execution of serial computing when solving large problems, such as BNTE, the study was focused on the design of parallel computing systems which was an evolution of serial computing that used multiple processing elements simultaneously to solve complex physical and mathematical problems. Finite element method (FEM) was used for the spatial discretization scheme, while angular discretization was accomplished by expanding the angular dependence in terms of Legendre polynomials. The eigenvalues representing the multiplication factors in the BNTE were determined by the power method. MATLAB Compiler Version 4.1 (R2009a) was used to compile the MATLAB codes of BNTE. The implemented parallel algorithms were enabled with matlabpool, a Parallel Computing Toolbox function. The option UseParallel was set to 'always' and the default value of the option was 'never'. When those conditions held, the solvers computed estimated gradients in parallel. The parallel computing system was used to handle all the bottlenecks in the matrix generated from the finite element scheme and each domain of the power method generated. The parallel algorithm was implemented on a Symmetric Multi Processor (SMP) cluster machine, which had Intel 32 bit quad-core x 86 processors. Convergence rates and timings for the algorithm on the SMP cluster machine were obtained. Numerical experiments indicated the designed parallel algorithm could reach perfect speedup and had good stability and scalability. (au)

  10. Multibus-based parallel processor for simulation

    Ogrady, E. P.; Wang, C.-H.

    1983-01-01

    A Multibus-based parallel processor simulation system is described. The system is intended to serve as a vehicle for gaining hands-on experience, testing system and application software, and evaluating parallel processor performance during development of a larger system based on the horizontal/vertical-bus interprocessor communication mechanism. The prototype system consists of up to seven Intel iSBC 86/12A single-board computers which serve as processing elements, a multiple transmission controller (MTC) designed to support system operation, and an Intel Model 225 Microcomputer Development System which serves as the user interface and input/output processor. All components are interconnected by a Multibus/IEEE 796 bus. An important characteristic of the system is that it provides a mechanism for a processing element to broadcast data to other selected processing elements. This parallel transfer capability is provided through the design of the MTC and a minor modification to the iSBC 86/12A board. The operation of the MTC, the basic hardware-level operation of the system, and pertinent details about the iSBC 86/12A and the Multibus are described.

  11. A multitransputer parallel processing system (MTPPS)

    Jethra, A.K.; Pande, S.S.; Borkar, S.P.; Khare, A.N.; Ghodgaonkar, M.D.; Bairi, B.R.

    1993-01-01

    This report describes the design and implementation of a 16 node Multi Transputer Parallel Processing System(MTPPS) which is a platform for parallel program development. It is a MIMD machine based on message passing paradigm. The basic compute engine is an Inmos Transputer Ims T800-20. Transputer with local memory constitutes the processing element (NODE) of this MIMD architecture. Multiple NODES can be connected to each other in an identifiable network topology through the high speed serial links of the transputer. A Network Configuration Unit (NCU) incorporates the necessary hardware to provide software controlled network configuration. System is modularly expandable and more NODES can be added to the system to achieve the required processing power. The system is backend to the IBM-PC which has been integrated into the system to provide user I/O interface. PC resources are available to the programmer. Interface hardware between the PC and the network of transputers is INMOS compatible. Therefore, all the commercially available development software compatible to INMOS products can run on this system. While giving the details of design and implementation, this report briefly summarises MIMD Architectures, Transputer Architecture and Parallel Processing Software Development issues. LINPACK performance evaluation of the system and solutions of neutron physics and plasma physics problem have been discussed along with results. (author). 12 refs., 22 figs., 3 tabs., 3 appendixes

  12. Template based parallel checkpointing in a massively parallel computer system

    Archer, Charles Jens [Rochester, MN; Inglett, Todd Alan [Rochester, MN

    2009-01-13

    A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.

  13. Parallel education: what is it?

    Amos, Michelle Peta

    2017-01-01

    In the history of education it has long been discussed that single-sex and coeducation are the two models of education present in schools. With the introduction of parallel schools over the last 15 years, there has been very little research into this 'new model'. Many people do not understand what it means for a school to be parallel or they confuse a parallel model with co-education, due to the presence of both boys and girls within the one institution. Therefore, the main obj...

  14. Balanced, parallel operation of flashlamps

    Carder, B.M.; Merritt, B.T.

    1979-01-01

    A new energy store, the Compensated Pulsed Alternator (CPA), promises to be a cost effective substitute for capacitors to drive flashlamps that pump large Nd:glass lasers. Because the CPA is large and discrete, it will be necessary that it drive many parallel flashlamp circuits, presenting a problem in equal current distribution. Current division to +- 20% between parallel flashlamps has been achieved, but this is marginal for laser pumping. A method is presented here that provides equal current sharing to about 1%, and it includes fused protection against short circuit faults. The method was tested with eight parallel circuits, including both open-circuit and short-circuit fault tests

  15. A solution for automatic parallelization of sequential assembly code

    Kovačević Đorđe

    2013-01-01

    Full Text Available Since modern multicore processors can execute existing sequential programs only on a single core, there is a strong need for automatic parallelization of program code. Relying on existing algorithms, this paper describes one new software solution tool for parallelization of sequential assembly code. The main goal of this paper is to develop the parallelizator which reads sequential assembler code and at the output provides parallelized code for MIPS processor with multiple cores. The idea is the following: the parser translates assembler input file to program objects suitable for further processing. After that the static single assignment is done. Based on the data flow graph, the parallelization algorithm separates instructions on different cores. Once sequential code is parallelized by the parallelization algorithm, registers are allocated with the algorithm for linear allocation, and the result at the end of the program is distributed assembler code on each of the cores. In the paper we evaluate the speedup of the matrix multiplication example, which was processed by the parallelizator of assembly code. The result is almost linear speedup of code execution, which increases with the number of cores. The speed up on the two cores is 1.99, while on 16 cores the speed up is 13.88.

  16. Comparison of the pharmacokinetics of a new 30 mg modified-release tablet formulation of metoclopramide for once-a-day administration versus 10 mg immediate-release tablets: a single and multiple-dose, randomized, open-label, parallel study in healthy male subjects.

    Bernardo-Escudero, Roberto; Alonso-Campero, Rosalba; Francisco-Doce, María Teresa de Jesús; Cortés-Fuentes, Myriam; Villa-Vargas, Miriam; Angeles-Uribe, Juan

    2012-12-01

    The study aimed to assess the pharmacokinetics of a new, modified-release metoclopramide tablet, and compare it to an immediate-release tablet. A single and multiple-dose, randomized, open-label, parallel, pharmacokinetic study was conducted. Investigational products were administered to 26 healthy Hispanic Mexican male volunteers for two consecutive days: either one 30 mg modified-release tablet every 24 h, or one 10 mg immediate-release tablet every 8 h. Blood samples were collected after the first and last doses of metoclopramide. Plasma metoclopramide concentrations were determined by high-performance liquid chromatography. Safety and tolerability were assessed through vital signs measurements, clinical evaluations, and spontaneous reports from study subjects. All 26 subjects were included in the analyses [mean (SD) age: 27 (8) years, range 18-50; BMI: 23.65 (2.22) kg/m², range 18.01-27.47)]. Peak plasmatic concentrations were not statistically different with both formulations, but occurred significantly later (p 0.05)]. One adverse event was reported in the test group (diarrhea), and one in the reference group (headache). This study suggests that the 30 mg modified-release metoclopramide tablets show features compatible with slow-release formulations when compared to immediate-release tablets, and is suitable for once-a-day administration.

  17. Collectively loading programs in a multiple program multiple data environment

    Aho, Michael E.; Attinella, John E.; Gooding, Thomas M.; Gooding, Thomas M.; Miller, Samuel J.

    2016-11-08

    Techniques are disclosed for loading programs efficiently in a parallel computing system. In one embodiment, nodes of the parallel computing system receive a load description file which indicates, for each program of a multiple program multiple data (MPMD) job, nodes which are to load the program. The nodes determine, using collective operations, a total number of programs to load and a number of programs to load in parallel. The nodes further generate a class route for each program to be loaded in parallel, where the class route generated for a particular program includes only those nodes on which the program needs to be loaded. For each class route, a node is selected using a collective operation to be a load leader which accesses a file system to load the program associated with a class route and broadcasts the program via the class route to other nodes which require the program.

  18. Workspace Analysis for Parallel Robot

    Ying Sun

    2013-05-01

    Full Text Available As a completely new-type of robot, the parallel robot possesses a lot of advantages that the serial robot does not, such as high rigidity, great load-carrying capacity, small error, high precision, small self-weight/load ratio, good dynamic behavior and easy control, hence its range is extended in using domain. In order to find workspace of parallel mechanism, the numerical boundary-searching algorithm based on the reverse solution of kinematics and limitation of link length has been introduced. This paper analyses position workspace, orientation workspace of parallel robot of the six degrees of freedom. The result shows: It is a main means to increase and decrease its workspace to change the length of branch of parallel mechanism; The radius of the movement platform has no effect on the size of workspace, but will change position of workspace.

  19. "Feeling" Series and Parallel Resistances.

    Morse, Robert A.

    1993-01-01

    Equipped with drinking straws and stirring straws, a teacher can help students understand how resistances in electric circuits combine in series and in parallel. Follow-up suggestions are provided. (ZWH)

  20. Massively Parallel Finite Element Programming

    Heister, Timo

    2010-01-01

    Today\\'s large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  1. Event monitoring of parallel computations

    Gruzlikov Alexander M.

    2015-06-01

    Full Text Available The paper considers the monitoring of parallel computations for detection of abnormal events. It is assumed that computations are organized according to an event model, and monitoring is based on specific test sequences

  2. Massively Parallel Finite Element Programming

    Heister, Timo; Kronbichler, Martin; Bangerth, Wolfgang

    2010-01-01

    Today's large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  3. The STAPL Parallel Graph Library

    Harshvardhan,

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable distributed graph container and a collection of commonly used parallel graph algorithms. The library introduces pGraph pViews that separate algorithm design from the container implementation. It supports three graph processing algorithmic paradigms, level-synchronous, asynchronous and coarse-grained, and provides common graph algorithms based on them. Experimental results demonstrate improved scalability in performance and data size over existing graph libraries on more than 16,000 cores and on internet-scale graphs containing over 16 billion vertices and 250 billion edges. © Springer-Verlag Berlin Heidelberg 2013.

  4. SPRINT: A new parallel framework for R

    Scharinger Florian

    2008-12-01

    Full Text Available Abstract Background Microarray analysis allows the simultaneous measurement of thousands to millions of genes or sequences across tens to thousands of different samples. The analysis of the resulting data tests the limits of existing bioinformatics computing infrastructure. A solution to this issue is to use High Performance Computing (HPC systems, which contain many processors and more memory than desktop computer systems. Many biostatisticians use R to process the data gleaned from microarray analysis and there is even a dedicated group of packages, Bioconductor, for this purpose. However, to exploit HPC systems, R must be able to utilise the multiple processors available on these systems. There are existing modules that enable R to use multiple processors, but these are either difficult to use for the HPC novice or cannot be used to solve certain classes of problems. A method of exploiting HPC systems, using R, but without recourse to mastering parallel programming paradigms is therefore necessary to analyse genomic data to its fullest. Results We have designed and built a prototype framework that allows the addition of parallelised functions to R to enable the easy exploitation of HPC systems. The Simple Parallel R INTerface (SPRINT is a wrapper around such parallelised functions. Their use requires very little modification to existing sequential R scripts and no expertise in parallel computing. As an example we created a function that carries out the computation of a pairwise calculated correlation matrix. This performs well with SPRINT. When executed using SPRINT on an HPC resource of eight processors this computation reduces by more than three times the time R takes to complete it on one processor. Conclusion SPRINT allows the biostatistician to concentrate on the research problems rather than the computation, while still allowing exploitation of HPC systems. It is easy to use and with further development will become more useful as more

  5. Massively parallel sparse matrix function calculations with NTPoly

    Dawson, William; Nakajima, Takahito

    2018-04-01

    We present NTPoly, a massively parallel library for computing the functions of sparse, symmetric matrices. The theory of matrix functions is a well developed framework with a wide range of applications including differential equations, graph theory, and electronic structure calculations. One particularly important application area is diagonalization free methods in quantum chemistry. When the input and output of the matrix function are sparse, methods based on polynomial expansions can be used to compute matrix functions in linear time. We present a library based on these methods that can compute a variety of matrix functions. Distributed memory parallelization is based on a communication avoiding sparse matrix multiplication algorithm. OpenMP task parallellization is utilized to implement hybrid parallelization. We describe NTPoly's interface and show how it can be integrated with programs written in many different programming languages. We demonstrate the merits of NTPoly by performing large scale calculations on the K computer.

  6. Writing parallel programs that work

    CERN. Geneva

    2012-01-01

    Serial algorithms typically run inefficiently on parallel machines. This may sound like an obvious statement, but it is the root cause of why parallel programming is considered to be difficult. The current state of the computer industry is still that almost all programs in existence are serial. This talk will describe the techniques used in the Intel Parallel Studio to provide a developer with the tools necessary to understand the behaviors and limitations of the existing serial programs. Once the limitations are known the developer can refactor the algorithms and reanalyze the resulting programs with the tools in the Intel Parallel Studio to create parallel programs that work. About the speaker Paul Petersen is a Sr. Principal Engineer in the Software and Solutions Group (SSG) at Intel. He received a Ph.D. degree in Computer Science from the University of Illinois in 1993. After UIUC, he was employed at Kuck and Associates, Inc. (KAI) working on auto-parallelizing compiler (KAP), and was involved in th...

  7. Parallel algorithms for continuum dynamics

    Hicks, D.L.; Liebrock, L.M.

    1987-01-01

    Simply porting existing parallel programs to a new parallel processor may not achieve the full speedup possible; to achieve the maximum efficiency may require redesigning the parallel algorithms for the specific architecture. The authors discuss here parallel algorithms that were developed first for the HEP processor and then ported to the CRAY X-MP/4, the ELXSI/10, and the Intel iPSC/32. Focus is mainly on the most recent parallel processing results produced, i.e., those on the Intel Hypercube. The applications are simulations of continuum dynamics in which the momentum and stress gradients are important. Examples of these are inertial confinement fusion experiments, severe breaks in the coolant system of a reactor, weapons physics, shock-wave physics. Speedup efficiencies on the Intel iPSC Hypercube are very sensitive to the ratio of communication to computation. Great care must be taken in designing algorithms for this machine to avoid global communication. This is much more critical on the iPSC than it was on the three previous parallel processors

  8. A dynamic bead-based microarray for parallel DNA detection

    Sochol, R D; Lin, L; Casavant, B P; Dueck, M E; Lee, L P

    2011-01-01

    A microfluidic system has been designed and constructed by means of micromachining processes to integrate both microfluidic mixing of mobile microbeads and hydrodynamic microbead arraying capabilities on a single chip to simultaneously detect multiple bio-molecules. The prototype system has four parallel reaction chambers, which include microchannels of 18 × 50 µm 2 cross-sectional area and a microfluidic mixing section of 22 cm length. Parallel detection of multiple DNA oligonucleotide sequences was achieved via molecular beacon probes immobilized on polystyrene microbeads of 16 µm diameter. Experimental results show quantitative detection of three distinct DNA oligonucleotide sequences from the Hepatitis C viral (HCV) genome with single base-pair mismatch specificity. Our dynamic bead-based microarray offers an effective microfluidic platform to increase parallelization of reactions and improve microbead handling for various biological applications, including bio-molecule detection, medical diagnostics and drug screening

  9. Parallel universes may be more than sci-fi daydreams

    2007-01-01

    Is the universe -- correction: "our" universe -- no more than a speck of cosmic dust amid an infinite number of parallel worlds? A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians, cosmologists, and other scientists.

  10. Induction heating using induction coils in series-parallel circuits

    Matsen, Marc Rollo; Geren, William Preston; Miller, Robert James; Negley, Mark Alan; Dykstra, William Chet

    2017-11-14

    A part is inductively heated by multiple, self-regulating induction coil circuits having susceptors, coupled together in parallel and in series with an AC power supply. Each of the circuits includes a tuning capacitor that tunes the circuit to resonate at the frequency of AC power supply.

  11. Parallel graded attention in reading: A pupillometric study

    Snell, Joshua; Mathot, Sebastiaan; Mirault, Jonathan; Grainger, Jonathan

    2018-01-01

    There are roughly two lines of theory to account for recent evidence that word processing is influenced by adjacent orthographic information. One line assumes that multiple words can be processed simultaneously through a parallel graded distribution of visuo-spatial attention. The other line assumes

  12. A Parallel Algebraic Multigrid Solver on Graphics Processing Units

    Haase, Gundolf; Liebmann, Manfred; Douglas, Craig C.; Plank, Gernot

    2010-01-01

    -vector multiplication scheme underlying the PCG-AMG algorithm is presented for the many-core GPU architecture. A performance comparison of the parallel solver shows that a singe Nvidia Tesla C1060 GPU board delivers the performance of a sixteen node Infiniband cluster

  13. Parallel Task Processing on a Multicore Platform in a PC-based Control System for Parallel Kinematics

    Harald Michalik

    2009-02-01

    Full Text Available Multicore platforms are such that have one physical processor chip with multiple cores interconnected via a chip level bus. Because they deliver a greater computing power through concurrency, offer greater system density multicore platforms provide best qualifications to address the performance bottleneck encountered in PC-based control systems for parallel kinematic robots with heavy CPU-load. Heavy load control tasks are generated by new control approaches that include features like singularity prediction, structure control algorithms, vision data integration and similar tasks. In this paper we introduce the parallel task scheduling extension of a communication architecture specially tailored for the development of PC-based control of parallel kinematics. The Sche-duling is specially designed for the processing on a multicore platform. It breaks down the serial task processing of the robot control cycle and extends it with parallel task processing paths in order to enhance the overall control performance.

  14. Endpoint-based parallel data processing in a parallel active messaging interface of a parallel computer

    Archer, Charles J.; Blocksome, Michael A.; Ratterman, Joseph D.; Smith, Brian E.

    2014-08-12

    Endpoint-based parallel data processing in a parallel active messaging interface (`PAMI`) of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective operation through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.

  15. Parallel Implicit Algorithms for CFD

    Keyes, David E.

    1998-01-01

    The main goal of this project was efficient distributed parallel and workstation cluster implementations of Newton-Krylov-Schwarz (NKS) solvers for implicit Computational Fluid Dynamics (CFD.) "Newton" refers to a quadratically convergent nonlinear iteration using gradient information based on the true residual, "Krylov" to an inner linear iteration that accesses the Jacobian matrix only through highly parallelizable sparse matrix-vector products, and "Schwarz" to a domain decomposition form of preconditioning the inner Krylov iterations with primarily neighbor-only exchange of data between the processors. Prior experience has established that Newton-Krylov methods are competitive solvers in the CFD context and that Krylov-Schwarz methods port well to distributed memory computers. The combination of the techniques into Newton-Krylov-Schwarz was implemented on 2D and 3D unstructured Euler codes on the parallel testbeds that used to be at LaRC and on several other parallel computers operated by other agencies or made available by the vendors. Early implementations were made directly in Massively Parallel Integration (MPI) with parallel solvers we adapted from legacy NASA codes and enhanced for full NKS functionality. Later implementations were made in the framework of the PETSC library from Argonne National Laboratory, which now includes pseudo-transient continuation Newton-Krylov-Schwarz solver capability (as a result of demands we made upon PETSC during our early porting experiences). A secondary project pursued with funding from this contract was parallel implicit solvers in acoustics, specifically in the Helmholtz formulation. A 2D acoustic inverse problem has been solved in parallel within the PETSC framework.

  16. Second derivative parallel block backward differentiation type ...

    Second derivative parallel block backward differentiation type formulas for Stiff ODEs. ... Log in or Register to get access to full text downloads. ... and the methods are inherently parallel and can be distributed over parallel processors. They are ...

  17. A Parallel Approach to Fractal Image Compression

    Lubomir Dedera

    2004-01-01

    The paper deals with a parallel approach to coding and decoding algorithms in fractal image compressionand presents experimental results comparing sequential and parallel algorithms from the point of view of achieved bothcoding and decoding time and effectiveness of parallelization.

  18. Parallel fabrication of macroporous scaffolds.

    Dobos, Andrew; Grandhi, Taraka Sai Pavan; Godeshala, Sudhakar; Meldrum, Deirdre R; Rege, Kaushal

    2018-07-01

    Scaffolds generated from naturally occurring and synthetic polymers have been investigated in several applications because of their biocompatibility and tunable chemo-mechanical properties. Existing methods for generation of 3D polymeric scaffolds typically cannot be parallelized, suffer from low throughputs, and do not allow for quick and easy removal of the fragile structures that are formed. Current molds used in hydrogel and scaffold fabrication using solvent casting and porogen leaching are often single-use and do not facilitate 3D scaffold formation in parallel. Here, we describe a simple device and related approaches for the parallel fabrication of macroporous scaffolds. This approach was employed for the generation of macroporous and non-macroporous materials in parallel, in higher throughput and allowed for easy retrieval of these 3D scaffolds once formed. In addition, macroporous scaffolds with interconnected as well as non-interconnected pores were generated, and the versatility of this approach was employed for the generation of 3D scaffolds from diverse materials including an aminoglycoside-derived cationic hydrogel ("Amikagel"), poly(lactic-co-glycolic acid) or PLGA, and collagen. Macroporous scaffolds generated using the device were investigated for plasmid DNA binding and cell loading, indicating the use of this approach for developing materials for different applications in biotechnology. Our results demonstrate that the device-based approach is a simple technology for generating scaffolds in parallel, which can enhance the toolbox of current fabrication techniques. © 2018 Wiley Periodicals, Inc.

  19. Parallel plasma fluid turbulence calculations

    Leboeuf, J.N.; Carreras, B.A.; Charlton, L.A.; Drake, J.B.; Lynch, V.E.; Newman, D.E.; Sidikman, K.L.; Spong, D.A.

    1994-01-01

    The study of plasma turbulence and transport is a complex problem of critical importance for fusion-relevant plasmas. To this day, the fluid treatment of plasma dynamics is the best approach to realistic physics at the high resolution required for certain experimentally relevant calculations. Core and edge turbulence in a magnetic fusion device have been modeled using state-of-the-art, nonlinear, three-dimensional, initial-value fluid and gyrofluid codes. Parallel implementation of these models on diverse platforms--vector parallel (National Energy Research Supercomputer Center's CRAY Y-MP C90), massively parallel (Intel Paragon XP/S 35), and serial parallel (clusters of high-performance workstations using the Parallel Virtual Machine protocol)--offers a variety of paths to high resolution and significant improvements in real-time efficiency, each with its own advantages. The largest and most efficient calculations have been performed at the 200 Mword memory limit on the C90 in dedicated mode, where an overlap of 12 to 13 out of a maximum of 16 processors has been achieved with a gyrofluid model of core fluctuations. The richness of the physics captured by these calculations is commensurate with the increased resolution and efficiency and is limited only by the ingenuity brought to the analysis of the massive amounts of data generated

  20. Evaluating parallel optimization on transputers

    A.G. Chalmers

    2003-12-01

    Full Text Available The faster processing power of modern computers and the development of efficient algorithms have made it possible for operations researchers to tackle a much wider range of problems than ever before. Further improvements in processing speed can be achieved utilising relatively inexpensive transputers to process components of an algorithm in parallel. The Davidon-Fletcher-Powell method is one of the most successful and widely used optimisation algorithms for unconstrained problems. This paper examines the algorithm and identifies the components that can be processed in parallel. The results of some experiments with these components are presented which indicates under what conditions parallel processing with an inexpensive configuration is likely to be faster than the traditional sequential implementations. The performance of the whole algorithm with its parallel components is then compared with the original sequential algorithm. The implementation serves to illustrate the practicalities of speeding up typical OR algorithms in terms of difficulty, effort and cost. The results give an indication of the savings in time a given parallel implementation can be expected to yield.

  1. Pattern-Driven Automatic Parallelization

    Christoph W. Kessler

    1996-01-01

    Full Text Available This article describes a knowledge-based system for automatic parallelization of a wide class of sequential numerical codes operating on vectors and dense matrices, and for execution on distributed memory message-passing multiprocessors. Its main feature is a fast and powerful pattern recognition tool that locally identifies frequently occurring computations and programming concepts in the source code. This tool also works for dusty deck codes that have been "encrypted" by former machine-specific code transformations. Successful pattern recognition guides sophisticated code transformations including local algorithm replacement such that the parallelized code need not emerge from the sequential program structure by just parallelizing the loops. It allows access to an expert's knowledge on useful parallel algorithms, available machine-specific library routines, and powerful program transformations. The partially restored program semantics also supports local array alignment, distribution, and redistribution, and allows for faster and more exact prediction of the performance of the parallelized target code than is usually possible.

  2. Parallel artificial liquid membrane extraction

    Gjelstad, Astrid; Rasmussen, Knut Einar; Parmer, Marthe Petrine

    2013-01-01

    This paper reports development of a new approach towards analytical liquid-liquid-liquid membrane extraction termed parallel artificial liquid membrane extraction. A donor plate and acceptor plate create a sandwich, in which each sample (human plasma) and acceptor solution is separated by an arti......This paper reports development of a new approach towards analytical liquid-liquid-liquid membrane extraction termed parallel artificial liquid membrane extraction. A donor plate and acceptor plate create a sandwich, in which each sample (human plasma) and acceptor solution is separated...... by an artificial liquid membrane. Parallel artificial liquid membrane extraction is a modification of hollow-fiber liquid-phase microextraction, where the hollow fibers are replaced by flat membranes in a 96-well plate format....

  3. Parallel algorithms for mapping pipelined and parallel computations

    Nicol, David M.

    1988-01-01

    Many computational problems in image processing, signal processing, and scientific computing are naturally structured for either pipelined or parallel computation. When mapping such problems onto a parallel architecture it is often necessary to aggregate an obvious problem decomposition. Even in this context the general mapping problem is known to be computationally intractable, but recent advances have been made in identifying classes of problems and architectures for which optimal solutions can be found in polynomial time. Among these, the mapping of pipelined or parallel computations onto linear array, shared memory, and host-satellite systems figures prominently. This paper extends that work first by showing how to improve existing serial mapping algorithms. These improvements have significantly lower time and space complexities: in one case a published O(nm sup 3) time algorithm for mapping m modules onto n processors is reduced to an O(nm log m) time complexity, and its space requirements reduced from O(nm sup 2) to O(m). Run time complexity is further reduced with parallel mapping algorithms based on these improvements, which run on the architecture for which they create the mappings.

  4. MEDUSA - An overset grid flow solver for network-based parallel computer systems

    Smith, Merritt H.; Pallis, Jani M.

    1993-01-01

    Continuing improvement in processing speed has made it feasible to solve the Reynolds-Averaged Navier-Stokes equations for simple three-dimensional flows on advanced workstations. Combining multiple workstations into a network-based heterogeneous parallel computer allows the application of programming principles learned on MIMD (Multiple Instruction Multiple Data) distributed memory parallel computers to the solution of larger problems. An overset-grid flow solution code has been developed which uses a cluster of workstations as a network-based parallel computer. Inter-process communication is provided by the Parallel Virtual Machine (PVM) software. Solution speed equivalent to one-third of a Cray-YMP processor has been achieved from a cluster of nine commonly used engineering workstation processors. Load imbalance and communication overhead are the principal impediments to parallel efficiency in this application.

  5. Modelling and parallel calculation of a kinetic boundary layer

    Perlat, Jean Philippe

    1998-01-01

    This research thesis aims at addressing reliability and cost issues in the calculation by numeric simulation of flows in transition regime. The first step has been to reduce calculation cost and memory space for the Monte Carlo method which is known to provide performance and reliability for rarefied regimes. Vector and parallel computers allow this objective to be reached. Here, a MIMD (multiple instructions, multiple data) machine has been used which implements parallel calculation at different levels of parallelization. Parallelization procedures have been adapted, and results showed that parallelization by calculation domain decomposition was far more efficient. Due to reliability issue related to the statistic feature of Monte Carlo methods, a new deterministic model was necessary to simulate gas molecules in transition regime. New models and hyperbolic systems have therefore been studied. One is chosen which allows thermodynamic values (density, average velocity, temperature, deformation tensor, heat flow) present in Navier-Stokes equations to be determined, and the equations of evolution of thermodynamic values are described for the mono-atomic case. Numerical resolution of is reported. A kinetic scheme is developed which complies with the structure of all systems, and which naturally expresses boundary conditions. The validation of the obtained 14 moment-based model is performed on shock problems and on Couette flows [fr

  6. Cellular automata a parallel model

    Mazoyer, J

    1999-01-01

    Cellular automata can be viewed both as computational models and modelling systems of real processes. This volume emphasises the first aspect. In articles written by leading researchers, sophisticated massive parallel algorithms (firing squad, life, Fischer's primes recognition) are treated. Their computational power and the specific complexity classes they determine are surveyed, while some recent results in relation to chaos from a new dynamic systems point of view are also presented. Audience: This book will be of interest to specialists of theoretical computer science and the parallelism challenge.

  7. Parallelizing AT with MatlabMPI

    2011-01-01

    The Accelerator Toolbox (AT) is a high-level collection of tools and scripts specifically oriented toward solving problems dealing with computational accelerator physics. It is integrated into the MATLAB environment, which provides an accessible, intuitive interface for accelerator physicists, allowing researchers to focus the majority of their efforts on simulations and calculations, rather than programming and debugging difficulties. Efforts toward parallelization of AT have been put in place to upgrade its performance to modern standards of computing. We utilized the packages MatlabMPI and pMatlab, which were developed by MIT Lincoln Laboratory, to set up a message-passing environment that could be called within MATLAB, which set up the necessary pre-requisites for multithread processing capabilities. On local quad-core CPUs, we were able to demonstrate processor efficiencies of roughly 95% and speed increases of nearly 380%. By exploiting the efficacy of modern-day parallel computing, we were able to demonstrate incredibly efficient speed increments per processor in AT's beam-tracking functions. Extrapolating from prediction, we can expect to reduce week-long computation runtimes to less than 15 minutes. This is a huge performance improvement and has enormous implications for the future computing power of the accelerator physics group at SSRL. However, one of the downfalls of parringpass is its current lack of transparency; the pMatlab and MatlabMPI packages must first be well-understood by the user before the system can be configured to run the scripts. In addition, the instantiation of argument parameters requires internal modification of the source code. Thus, parringpass, cannot be directly run from the MATLAB command line, which detracts from its flexibility and user-friendliness. Future work in AT's parallelization will focus on development of external functions and scripts that can be called from within MATLAB and configured on multiple nodes, while

  8. MPI_XSTAR: MPI-based Parallelization of the XSTAR Photoionization Program

    Danehkar, Ashkbiz; Nowak, Michael A.; Lee, Julia C.; Smith, Randall K.

    2018-02-01

    We describe a program for the parallel implementation of multiple runs of XSTAR, a photoionization code that is used to predict the physical properties of an ionized gas from its emission and/or absorption lines. The parallelization program, called MPI_XSTAR, has been developed and implemented in the C++ language by using the Message Passing Interface (MPI) protocol, a conventional standard of parallel computing. We have benchmarked parallel multiprocessing executions of XSTAR, using MPI_XSTAR, against a serial execution of XSTAR, in terms of the parallelization speedup and the computing resource efficiency. Our experience indicates that the parallel execution runs significantly faster than the serial execution, however, the efficiency in terms of the computing resource usage decreases with increasing the number of processors used in the parallel computing.

  9. Design, analysis and control of cable-suspended parallel robots and its applications

    Zi, Bin

    2017-01-01

    This book provides an essential overview of the authors’ work in the field of cable-suspended parallel robots, focusing on innovative design, mechanics, control, development and applications. It presents and analyzes several typical mechanical architectures of cable-suspended parallel robots in practical applications, including the feed cable-suspended structure for super antennae, hybrid-driven-based cable-suspended parallel robots, and cooperative cable parallel manipulators for multiple mobile cranes. It also addresses the fundamental mechanics of cable-suspended parallel robots on the basis of their typical applications, including the kinematics, dynamics and trajectory tracking control of the feed cable-suspended structure for super antennae. In addition it proposes a novel hybrid-driven-based cable-suspended parallel robot that uses integrated mechanism design methods to improve the performance of traditional cable-suspended parallel robots. A comparative study on error and performance indices of hybr...

  10. Genetic heterogeneity in type 1 Gaucher disease: Multiple genotypes in Ashkenazic and non-Ashkenazic individuals

    Tsuji, Shoji; Martin, B.M.; Stubblefield, B.K.; LaMarca, M.E.; Ginns, E.I.; Barranger, J.A.

    1988-01-01

    Nucleotide sequence analysis of a genomic clone from an Ashkenazic Jewish patient with type 1 Gaucher disease revealed a single-base mutation (adenosine to guanosine transition) in exon 9 of the glucocerebrosidase gene. This change results in the amino acid substitution of serine for asparagine. Transient expression studies following oligonucleotide-directed mutagenesis of the normal cDNA confirmed that the mutation results in loss of glucocerebrosidase activity. Allele-specific hybridization with oligonucleotide probes demonstrated that this mutation was found exclusively in type 1 phenotype. None of the 6 type 2 patients, 11 type 3 patients, or 12 normal controls had this allele. In contrast, 15 of 24 type 1 patients had one allele with this mutation, and 3 others were homozygous for the mutation. Furthermore, some of the Ashkenazic Jewish type 1 patients had only one allele with this mutation, suggesting that even in this population there is allelic heterozygosity. These findings indicate that there are multiple allelic mutations responsible for type 1 Gaucher disease in both the Jewish and non-Jewish populations. Allelic-specific hybridization demonstrating this mutation in exon 9, used in conjunction with the Nci I restriction fragment length polymorphism described as a marker for neuronopathic Gaucher disease, provides a tool for diagnosis and genetic counseling that is ∼80% informative in all Gaucher patients studied

  11. Genetic heterogeneity in type 1 Gaucher disease: Multiple genotypes in Ashkenazic and non-Ashkenazic individuals

    Tsuji, Shoji; Martin, B.M.; Stubblefield, B.K.; LaMarca, M.E.; Ginns, E.I. (National Institute of Mental Health, Bethesda, MD (USA)); Barranger, J.A. (Childrens Hospital of Los Angeles, CA (USA))

    1988-04-01

    Nucleotide sequence analysis of a genomic clone from an Ashkenazic Jewish patient with type 1 Gaucher disease revealed a single-base mutation (adenosine to guanosine transition) in exon 9 of the glucocerebrosidase gene. This change results in the amino acid substitution of serine for asparagine. Transient expression studies following oligonucleotide-directed mutagenesis of the normal cDNA confirmed that the mutation results in loss of glucocerebrosidase activity. Allele-specific hybridization with oligonucleotide probes demonstrated that this mutation was found exclusively in type 1 phenotype. None of the 6 type 2 patients, 11 type 3 patients, or 12 normal controls had this allele. In contrast, 15 of 24 type 1 patients had one allele with this mutation, and 3 others were homozygous for the mutation. Furthermore, some of the Ashkenazic Jewish type 1 patients had only one allele with this mutation, suggesting that even in this population there is allelic heterozygosity. These findings indicate that there are multiple allelic mutations responsible for type 1 Gaucher disease in both the Jewish and non-Jewish populations. Allelic-specific hybridization demonstrating this mutation in exon 9, used in conjunction with the Nci I restriction fragment length polymorphism described as a marker for neuronopathic Gaucher disease, provides a tool for diagnosis and genetic counseling that is {approx}80% informative in all Gaucher patients studied.

  12. Rubus: A compiler for seamless and extensible parallelism

    Adnan, Muhammad; Aslam, Faisal; Sarwar, Syed Mansoor

    2017-01-01

    Nowadays, a typical processor may have multiple processing cores on a single chip. Furthermore, a special purpose processing unit called Graphic Processing Unit (GPU), originally designed for 2D/3D games, is now available for general purpose use in computers and mobile devices. However, the traditional programming languages which were designed to work with machines having single core CPUs, cannot utilize the parallelism available on multi-core processors efficiently. Therefore, to exploit the extraordinary processing power of multi-core processors, researchers are working on new tools and techniques to facilitate parallel programming. To this end, languages like CUDA and OpenCL have been introduced, which can be used to write code with parallelism. The main shortcoming of these languages is that programmer needs to specify all the complex details manually in order to parallelize the code across multiple cores. Therefore, the code written in these languages is difficult to understand, debug and maintain. Furthermore, to parallelize legacy code can require rewriting a significant portion of code in CUDA or OpenCL, which can consume significant time and resources. Thus, the amount of parallelism achieved is proportional to the skills of the programmer and the time spent in code optimizations. This paper proposes a new open source compiler, Rubus, to achieve seamless parallelism. The Rubus compiler relieves the programmer from manually specifying the low-level details. It analyses and transforms a sequential program into a parallel program automatically, without any user intervention. This achieves massive speedup and better utilization of the underlying hardware without a programmer’s expertise in parallel programming. For five different benchmarks, on average a speedup of 34.54 times has been achieved by Rubus as compared to Java on a basic GPU having only 96 cores. Whereas, for a matrix multiplication benchmark the average execution speedup of 84 times has been

  13. History Matching in Parallel Computational Environments

    Steven Bryant; Sanjay Srinivasan; Alvaro Barrera; Sharad Yadav

    2004-08-31

    In the probabilistic approach for history matching, the information from the dynamic data is merged with the prior geologic information in order to generate permeability models consistent with the observed dynamic data as well as the prior geology. The relationship between dynamic response data and reservoir attributes may vary in different regions of the reservoir due to spatial variations in reservoir attributes, fluid properties, well configuration, flow constrains on wells etc. This implies probabilistic approach should then update different regions of the reservoir in different ways. This necessitates delineation of multiple reservoir domains in order to increase the accuracy of the approach. The research focuses on a probabilistic approach to integrate dynamic data that ensures consistency between reservoir models developed from one stage to the next. The algorithm relies on efficient parameterization of the dynamic data integration problem and permits rapid assessment of the updated reservoir model at each stage. The report also outlines various domain decomposition schemes from the perspective of increasing the accuracy of probabilistic approach of history matching. Research progress in three important areas of the project are discussed: {lg_bullet}Validation and testing the probabilistic approach to incorporating production data in reservoir models. {lg_bullet}Development of a robust scheme for identifying reservoir regions that will result in a more robust parameterization of the history matching process. {lg_bullet}Testing commercial simulators for parallel capability and development of a parallel algorithm for history matching.

  14. A Parallel Algebraic Multigrid Solver on Graphics Processing Units

    Haase, Gundolf

    2010-01-01

    The paper presents a multi-GPU implementation of the preconditioned conjugate gradient algorithm with an algebraic multigrid preconditioner (PCG-AMG) for an elliptic model problem on a 3D unstructured grid. An efficient parallel sparse matrix-vector multiplication scheme underlying the PCG-AMG algorithm is presented for the many-core GPU architecture. A performance comparison of the parallel solver shows that a singe Nvidia Tesla C1060 GPU board delivers the performance of a sixteen node Infiniband cluster and a multi-GPU configuration with eight GPUs is about 100 times faster than a typical server CPU core. © 2010 Springer-Verlag.

  15. Memory Retrieval Given Two Independent Cues: Cue Selection or Parallel Access?

    Rickard, Timothy C.; Bajic, Daniel

    2004-01-01

    A basic but unresolved issue in the study of memory retrieval is whether multiple independent cues can be used concurrently (i.e., in parallel) to recall a single, common response. A number of empirical results, as well as potentially applicable theories, suggest that retrieval can proceed in parallel, though Rickard (1997) set forth a model that…

  16. Parallel manipulators with two end-effectors : Getting a grip on Jacobian-based stiffness analysis

    Hoevenaars, A.G.L.

    2016-01-01

    Robots that are developed for applications which require a high stiffness-over-inertia ratio, such as pick-and-place robots, machining robots, or haptic devices, are often based on parallel manipulators. Parallel manipulators connect an end-effector to an inertial base using multiple serial

  17. Parallel Sparse Matrix - Vector Product

    Alexandersen, Joe; Lazarov, Boyan Stefanov; Dammann, Bernd

    This technical report contains a case study of a sparse matrix-vector product routine, implemented for parallel execution on a compute cluster with both pure MPI and hybrid MPI-OpenMP solutions. C++ classes for sparse data types were developed and the report shows how these class can be used...

  18. [Falsified medicines in parallel trade].

    Muckenfuß, Heide

    2017-11-01

    The number of falsified medicines on the German market has distinctly increased over the past few years. In particular, stolen pharmaceutical products, a form of falsified medicines, have increasingly been introduced into the legal supply chain via parallel trading. The reasons why parallel trading serves as a gateway for falsified medicines are most likely the complex supply chains and routes of transport. It is hardly possible for national authorities to trace the history of a medicinal product that was bought and sold by several intermediaries in different EU member states. In addition, the heterogeneous outward appearance of imported and relabelled pharmaceutical products facilitates the introduction of illegal products onto the market. Official batch release at the Paul-Ehrlich-Institut offers the possibility of checking some aspects that might provide an indication of a falsified medicine. In some circumstances, this may allow the identification of falsified medicines before they come onto the German market. However, this control is only possible for biomedicinal products that have not received a waiver regarding official batch release. For improved control of parallel trade, better networking among the EU member states would be beneficial. European-wide regulations, e. g., for disclosure of the complete supply chain, would help to minimise the risks of parallel trading and hinder the marketing of falsified medicines.

  19. The parallel adult education system

    Wahlgren, Bjarne

    2015-01-01

    for competence development. The Danish university educational system includes two parallel programs: a traditional academic track (candidatus) and an alternative practice-based track (master). The practice-based program was established in 2001 and organized as part time. The total program takes half the time...

  20. Where are the parallel algorithms?

    Voigt, R. G.

    1985-01-01

    Four paradigms that can be useful in developing parallel algorithms are discussed. These include computational complexity analysis, changing the order of computation, asynchronous computation, and divide and conquer. Each is illustrated with an example from scientific computation, and it is shown that computational complexity must be used with great care or an inefficient algorithm may be selected.

  1. Parallel imaging with phase scrambling.

    Zaitsev, Maxim; Schultz, Gerrit; Hennig, Juergen; Gruetter, Rolf; Gallichan, Daniel

    2015-04-01

    Most existing methods for accelerated parallel imaging in MRI require additional data, which are used to derive information about the sensitivity profile of each radiofrequency (RF) channel. In this work, a method is presented to avoid the acquisition of separate coil calibration data for accelerated Cartesian trajectories. Quadratic phase is imparted to the image to spread the signals in k-space (aka phase scrambling). By rewriting the Fourier transform as a convolution operation, a window can be introduced to the convolved chirp function, allowing a low-resolution image to be reconstructed from phase-scrambled data without prominent aliasing. This image (for each RF channel) can be used to derive coil sensitivities to drive existing parallel imaging techniques. As a proof of concept, the quadratic phase was applied by introducing an offset to the x(2) - y(2) shim and the data were reconstructed using adapted versions of the image space-based sensitivity encoding and GeneRalized Autocalibrating Partially Parallel Acquisitions algorithms. The method is demonstrated in a phantom (1 × 2, 1 × 3, and 2 × 2 acceleration) and in vivo (2 × 2 acceleration) using a 3D gradient echo acquisition. Phase scrambling can be used to perform parallel imaging acceleration without acquisition of separate coil calibration data, demonstrated here for a 3D-Cartesian trajectory. Further research is required to prove the applicability to other 2D and 3D sampling schemes. © 2014 Wiley Periodicals, Inc.

  2. Default Parallels Plesk Panel Page

    services that small businesses want and need. Our software includes key building blocks of cloud service virtualized servers Service Provider Products Parallels® Automation Hosting, SaaS, and cloud computing , the leading hosting automation software. You see this page because there is no Web site at this

  3. Parallel plate transmission line transformer

    Voeten, S.J.; Brussaard, G.J.H.; Pemen, A.J.M.

    2011-01-01

    A Transmission Line Transformer (TLT) can be used to transform high-voltage nanosecond pulses. These transformers rely on the fact that the length of the pulse is shorter than the transmission lines used. This allows connecting the transmission lines in parallel at the input and in series at the

  4. Matpar: Parallel Extensions for MATLAB

    Springer, P. L.

    1998-01-01

    Matpar is a set of client/server software that allows a MATLAB user to take advantage of a parallel computer for very large problems. The user can replace calls to certain built-in MATLAB functions with calls to Matpar functions.

  5. Massively parallel quantum computer simulator

    De Raedt, K.; Michielsen, K.; De Raedt, H.; Trieu, B.; Arnold, G.; Richter, M.; Lippert, Th.; Watanabe, H.; Ito, N.

    2007-01-01

    We describe portable software to simulate universal quantum computers on massive parallel Computers. We illustrate the use of the simulation software by running various quantum algorithms on different computer architectures, such as a IBM BlueGene/L, a IBM Regatta p690+, a Hitachi SR11000/J1, a Cray

  6. Parallel computing: numerics, applications, and trends

    Trobec, Roman; Vajteršic, Marián; Zinterhof, Peter

    2009-01-01

    ... and/or distributed systems. The contributions to this book are focused on topics most concerned in the trends of today's parallel computing. These range from parallel algorithmics, programming, tools, network computing to future parallel computing. Particular attention is paid to parallel numerics: linear algebra, differential equations, numerica...

  7. Experiments with parallel algorithms for combinatorial problems

    G.A.P. Kindervater (Gerard); H.W.J.M. Trienekens

    1985-01-01

    textabstractIn the last decade many models for parallel computation have been proposed and many parallel algorithms have been developed. However, few of these models have been realized and most of these algorithms are supposed to run on idealized, unrealistic parallel machines. The parallel machines

  8. CS-Studio Scan System Parallelization

    Kasemir, Kay [ORNL; Pearson, Matthew R [ORNL

    2015-01-01

    For several years, the Control System Studio (CS-Studio) Scan System has successfully automated the operation of beam lines at the Oak Ridge National Laboratory (ORNL) High Flux Isotope Reactor (HFIR) and Spallation Neutron Source (SNS). As it is applied to additional beam lines, we need to support simultaneous adjustments of temperatures or motor positions. While this can be implemented via virtual motors or similar logic inside the Experimental Physics and Industrial Control System (EPICS) Input/Output Controllers (IOCs), doing so requires a priori knowledge of experimenters requirements. By adding support for the parallel control of multiple process variables (PVs) to the Scan System, we can better support ad hoc automation of experiments that benefit from such simultaneous PV adjustments.

  9. Parallel R-matrix computation

    Heggarty, J.W.

    1999-06-01

    For almost thirty years, sequential R-matrix computation has been used by atomic physics research groups, from around the world, to model collision phenomena involving the scattering of electrons or positrons with atomic or molecular targets. As considerable progress has been made in the understanding of fundamental scattering processes, new data, obtained from more complex calculations, is of current interest to experimentalists. Performing such calculations, however, places considerable demands on the computational resources to be provided by the target machine, in terms of both processor speed and memory requirement. Indeed, in some instances the computational requirements are so great that the proposed R-matrix calculations are intractable, even when utilising contemporary classic supercomputers. Historically, increases in the computational requirements of R-matrix computation were accommodated by porting the problem codes to a more powerful classic supercomputer. Although this approach has been successful in the past, it is no longer considered to be a satisfactory solution due to the limitations of current (and future) Von Neumann machines. As a consequence, there has been considerable interest in the high performance multicomputers, that have emerged over the last decade which appear to offer the computational resources required by contemporary R-matrix research. Unfortunately, developing codes for these machines is not as simple a task as it was to develop codes for successive classic supercomputers. The difficulty arises from the considerable differences in the computing models that exist between the two types of machine and results in the programming of multicomputers to be widely acknowledged as a difficult, time consuming and error-prone task. Nevertheless, unless parallel R-matrix computation is realised, important theoretical and experimental atomic physics research will continue to be hindered. This thesis describes work that was undertaken in

  10. The numerical parallel computing of photon transport

    Huang Qingnan; Liang Xiaoguang; Zhang Lifa

    1998-12-01

    The parallel computing of photon transport is investigated, the parallel algorithm and the parallelization of programs on parallel computers both with shared memory and with distributed memory are discussed. By analyzing the inherent law of the mathematics and physics model of photon transport according to the structure feature of parallel computers, using the strategy of 'to divide and conquer', adjusting the algorithm structure of the program, dissolving the data relationship, finding parallel liable ingredients and creating large grain parallel subtasks, the sequential computing of photon transport into is efficiently transformed into parallel and vector computing. The program was run on various HP parallel computers such as the HY-1 (PVP), the Challenge (SMP) and the YH-3 (MPP) and very good parallel speedup has been gotten

  11. Automatic Parallelization Tool: Classification of Program Code for Parallel Computing

    Mustafa Basthikodi

    2016-04-01

    Full Text Available Performance growth of single-core processors has come to a halt in the past decade, but was re-enabled by the introduction of parallelism in processors. Multicore frameworks along with Graphical Processing Units empowered to enhance parallelism broadly. Couples of compilers are updated to developing challenges forsynchronization and threading issues. Appropriate program and algorithm classifications will have advantage to a great extent to the group of software engineers to get opportunities for effective parallelization. In present work we investigated current species for classification of algorithms, in that related work on classification is discussed along with the comparison of issues that challenges the classification. The set of algorithms are chosen which matches the structure with different issues and perform given task. We have tested these algorithms utilizing existing automatic species extraction toolsalong with Bones compiler. We have added functionalities to existing tool, providing a more detailed characterization. The contributions of our work include support for pointer arithmetic, conditional and incremental statements, user defined types, constants and mathematical functions. With this, we can retain significant data which is not captured by original speciesof algorithms. We executed new theories into the device, empowering automatic characterization of program code.

  12. Multi-lane detection based on multiple vanishing points detection

    Li, Chuanxiang; Nie, Yiming; Dai, Bin; Wu, Tao

    2015-03-01

    Lane detection plays a significant role in Advanced Driver Assistance Systems (ADAS) for intelligent vehicles. In this paper we present a multi-lane detection method based on multiple vanishing points detection. A new multi-lane model assumes that a single lane, which has two approximately parallel boundaries, may not parallel to others on road plane. Non-parallel lanes associate with different vanishing points. A biological plausibility model is used to detect multiple vanishing points and fit lane model. Experimental results show that the proposed method can detect both parallel lanes and non-parallel lanes.

  13. Fluid dynamics parallel computer development at NASA Langley Research Center

    Townsend, James C.; Zang, Thomas A.; Dwoyer, Douglas L.

    1987-01-01

    To accomplish more detailed simulations of highly complex flows, such as the transition to turbulence, fluid dynamics research requires computers much more powerful than any available today. Only parallel processing on multiple-processor computers offers hope for achieving the required effective speeds. Looking ahead to the use of these machines, the fluid dynamicist faces three issues: algorithm development for near-term parallel computers, architecture development for future computer power increases, and assessment of possible advantages of special purpose designs. Two projects at NASA Langley address these issues. Software development and algorithm exploration is being done on the FLEX/32 Parallel Processing Research Computer. New architecture features are being explored in the special purpose hardware design of the Navier-Stokes Computer. These projects are complementary and are producing promising results.

  14. Parallel and distributed processing in power system simulation and control

    Falcao, Djalma M [Universidade Federal, Rio de Janeiro, RJ (Brazil). Coordenacao dos Programas de Pos-graduacao de Engenharia

    1994-12-31

    Recent advances in computer technology will certainly have a great impact in the methodologies used in power system expansion and operational planning as well as in real-time control. Parallel and distributed processing are among the new technologies that present great potential for application in these areas. Parallel computers use multiple functional or processing units to speed up computation while distributed processing computer systems are collection of computers joined together by high speed communication networks having many objectives and advantages. The paper presents some ideas for the use of parallel and distributed processing in power system simulation and control. It also comments on some of the current research work in these topics and presents a summary of the work presently being developed at COPPE. (author) 53 refs., 2 figs.

  15. Structural synthesis of parallel robots

    Gogu, Grigore

    This book represents the fifth part of a larger work dedicated to the structural synthesis of parallel robots. The originality of this work resides in the fact that it combines new formulae for mobility, connectivity, redundancy and overconstraints with evolutionary morphology in a unified structural synthesis approach that yields interesting and innovative solutions for parallel robotic manipulators.  This is the first book on robotics that presents solutions for coupled, decoupled, uncoupled, fully-isotropic and maximally regular robotic manipulators with Schönflies motions systematically generated by using the structural synthesis approach proposed in Part 1.  Overconstrained non-redundant/overactuated/redundantly actuated solutions with simple/complex limbs are proposed. Many solutions are presented here for the first time in the literature. The author had to make a difficult and challenging choice between protecting these solutions through patents and releasing them directly into the public domain. T...

  16. GPU Parallel Bundle Block Adjustment

    ZHENG Maoteng

    2017-09-01

    Full Text Available To deal with massive data in photogrammetry, we introduce the GPU parallel computing technology. The preconditioned conjugate gradient and inexact Newton method are also applied to decrease the iteration times while solving the normal equation. A brand new workflow of bundle adjustment is developed to utilize GPU parallel computing technology. Our method can avoid the storage and inversion of the big normal matrix, and compute the normal matrix in real time. The proposed method can not only largely decrease the memory requirement of normal matrix, but also largely improve the efficiency of bundle adjustment. It also achieves the same accuracy as the conventional method. Preliminary experiment results show that the bundle adjustment of a dataset with about 4500 images and 9 million image points can be done in only 1.5 minutes while achieving sub-pixel accuracy.

  17. A tandem parallel plate analyzer

    Hamada, Y.; Fujisawa, A.; Iguchi, H.; Nishizawa, A.; Kawasumi, Y.

    1996-11-01

    By a new modification of a parallel plate analyzer the second-order focus is obtained in an arbitrary injection angle. This kind of an analyzer with a small injection angle will have an advantage of small operational voltage, compared to the Proca and Green analyzer where the injection angle is 30 degrees. Thus, the newly proposed analyzer will be very useful for the precise energy measurement of high energy particles in MeV range. (author)

  18. High-speed parallel counter

    Gus'kov, B.N.; Kalinnikov, V.A.; Krastev, V.R.; Maksimov, A.N.; Nikityuk, N.M.

    1985-01-01

    This paper describes a high-speed parallel counter that contains 31 inputs and 15 outputs and is implemented by integrated circuits of series 500. The counter is designed for fast sampling of events according to the number of particles that pass simultaneously through the hodoscopic plane of the detector. The minimum delay of the output signals relative to the input is 43 nsec. The duration of the output signals can be varied from 75 to 120 nsec

  19. An anthropologist in parallel structure

    Noelle Molé Liston

    2016-08-01

    Full Text Available The essay examines the parallels between Molé Liston’s studies on labor and precarity in Italy and the United States’ anthropology job market. Probing the way economic shift reshaped the field of anthropology of Europe in the late 2000s, the piece explores how the neoliberalization of the American academy increased the value in studying the hardships and daily lives of non-western populations in Europe.

  20. Combinatorics of spreads and parallelisms

    Johnson, Norman

    2010-01-01

    Partitions of Vector Spaces Quasi-Subgeometry Partitions Finite Focal-SpreadsGeneralizing André SpreadsThe Going Up Construction for Focal-SpreadsSubgeometry Partitions Subgeometry and Quasi-Subgeometry Partitions Subgeometries from Focal-SpreadsExtended André SubgeometriesKantor's Flag-Transitive DesignsMaximal Additive Partial SpreadsSubplane Covered Nets and Baer Groups Partial Desarguesian t-Parallelisms Direct Products of Affine PlanesJha-Johnson SL(2,

  1. New algorithms for parallel MRI

    Anzengruber, S; Ramlau, R; Bauer, F; Leitao, A

    2008-01-01

    Magnetic Resonance Imaging with parallel data acquisition requires algorithms for reconstructing the patient's image from a small number of measured lines of the Fourier domain (k-space). In contrast to well-known algorithms like SENSE and GRAPPA and its flavors we consider the problem as a non-linear inverse problem. However, in order to avoid cost intensive derivatives we will use Landweber-Kaczmarz iteration and in order to improve the overall results some additional sparsity constraints.

  2. Wakefield calculations on parallel computers

    Schoessow, P.

    1990-01-01

    The use of parallelism in the solution of wakefield problems is illustrated for two different computer architectures (SIMD and MIMD). Results are given for finite difference codes which have been implemented on a Connection Machine and an Alliant FX/8 and which are used to compute wakefields in dielectric loaded structures. Benchmarks on code performance are presented for both cases. 4 refs., 3 figs., 2 tabs

  3. Aspects of computation on asynchronous parallel processors

    Wright, M.

    1989-01-01

    The increasing availability of asynchronous parallel processors has provided opportunities for original and useful work in scientific computing. However, the field of parallel computing is still in a highly volatile state, and researchers display a wide range of opinion about many fundamental questions such as models of parallelism, approaches for detecting and analyzing parallelism of algorithms, and tools that allow software developers and users to make effective use of diverse forms of complex hardware. This volume collects the work of researchers specializing in different aspects of parallel computing, who met to discuss the framework and the mechanics of numerical computing. The far-reaching impact of high-performance asynchronous systems is reflected in the wide variety of topics, which include scientific applications (e.g. linear algebra, lattice gauge simulation, ordinary and partial differential equations), models of parallelism, parallel language features, task scheduling, automatic parallelization techniques, tools for algorithm development in parallel environments, and system design issues

  4. Parallel processing of genomics data

    Agapito, Giuseppe; Guzzi, Pietro Hiram; Cannataro, Mario

    2016-10-01

    The availability of high-throughput experimental platforms for the analysis of biological samples, such as mass spectrometry, microarrays and Next Generation Sequencing, have made possible to analyze a whole genome in a single experiment. Such platforms produce an enormous volume of data per single experiment, thus the analysis of this enormous flow of data poses several challenges in term of data storage, preprocessing, and analysis. To face those issues, efficient, possibly parallel, bioinformatics software needs to be used to preprocess and analyze data, for instance to highlight genetic variation associated with complex diseases. In this paper we present a parallel algorithm for the parallel preprocessing and statistical analysis of genomics data, able to face high dimension of data and resulting in good response time. The proposed system is able to find statistically significant biological markers able to discriminate classes of patients that respond to drugs in different ways. Experiments performed on real and synthetic genomic datasets show good speed-up and scalability.

  5. Multiple Perspectives / Multiple Readings

    Simon Biggs

    2005-01-01

    Full Text Available People experience things from their own physical point of view. What they see is usually a function of where they are and what physical attitude they adopt relative to the subject. With augmented vision (periscopes, mirrors, remote cameras, etc we are able to see things from places where we are not present. With time-shifting technologies, such as the video recorder, we can also see things from the past; a time and a place we may never have visited.In recent artistic work I have been exploring the implications of digital technology, interactivity and internet connectivity that allow people to not so much space/time-shift their visual experience of things but rather see what happens when everybody is simultaneously able to see what everybody else can see. This is extrapolated through the remote networking of sites that are actual installation spaces; where the physical movements of viewers in the space generate multiple perspectives, linked to other similar sites at remote locations or to other viewers entering the shared data-space through a web based version of the work.This text explores the processes involved in such a practice and reflects on related questions regarding the non-singularity of being and the sense of self as linked to time and place.

  6. Construction of a digital elevation model: methods and parallelization

    Mazzoni, Christophe

    1995-01-01

    The aim of this work is to reduce the computation time needed to produce the Digital Elevation Models (DEM) by using a parallel machine. It is made in collaboration between the French 'Institut Geographique National' (IGN) and the Laboratoire d'Electronique de Technologie et d'Instrumentation (LETI) of the French Atomic Energy Commission (CEA). The IGN has developed a system which provides DEM that is used to produce topographic maps. The kernel of this system is the correlator, a software which automatically matches pairs of homologous points of a stereo-pair of photographs. Nevertheless the correlator is expensive In computing time. In order to reduce computation time and to produce the DEM with same accuracy that the actual system, we have parallelized the IGN's correlator on the OPENVISION system. This hardware solution uses a SIMD (Single Instruction Multiple Data) parallel machine SYMPATI-2, developed by the LETI that is involved in parallel architecture and image processing. Our analysis of the implementation has demonstrated the difficulty of efficient coupling between scalar and parallel structure. So we propose solutions to reinforce this coupling. In order to accelerate more the processing we evaluate SYMPHONIE, a SIMD calculator, successor of SYMPATI-2. On an other hand, we developed a multi-agent approach for what a MIMD (Multiple Instruction, Multiple Data) architecture is available. At last, we describe a Multi-SIMD architecture that conciliates our two approaches. This architecture offers a capacity to apprehend efficiently multi-level treatment image. It is flexible by its modularity, and its communication network supplies reliability that interest sensible systems. (author) [fr

  7. Evidence for parallel consolidation of motion direction and orientation into visual short-term memory.

    Rideaux, Reuben; Apthorp, Deborah; Edwards, Mark

    2015-02-12

    Recent findings have indicated the capacity to consolidate multiple items into visual short-term memory in parallel varies as a function of the type of information. That is, while color can be consolidated in parallel, evidence suggests that orientation cannot. Here we investigated the capacity to consolidate multiple motion directions in parallel and reexamined this capacity using orientation. This was achieved by determining the shortest exposure duration necessary to consolidate a single item, then examining whether two items, presented simultaneously, could be consolidated in that time. The results show that parallel consolidation of direction and orientation information is possible, and that parallel consolidation of direction appears to be limited to two. Additionally, we demonstrate the importance of adequate separation between feature intervals used to define items when attempting to consolidate in parallel, suggesting that when multiple items are consolidated in parallel, as opposed to serially, the resolution of representations suffer. Finally, we used facilitation of spatial attention to show that the deterioration of item resolution occurs during parallel consolidation, as opposed to storage. © 2015 ARVO.

  8. Hydraulic Profiling of a Parallel Channel Type Reactor Core

    Seo, Kyong-Won; Hwang, Dae-Hyun; Lee, Chung-Chan

    2006-01-01

    An advanced reactor core which consisted of closed multiple parallel channels was optimized to maximize the thermal margin of the core. The closed multiple parallel channel configurations have different characteristics to the open channels of conventional PWRs. The channels, usually assemblies, are isolated hydraulically from each other and there is no cross flow between channels. The distribution of inlet flow rate between channels is a very important design parameter in the core because distribution of inlet flow is directly proportional to a margin for a certain hydraulic parameter. The thermal hydraulic parameter may be the boiling margin, maximum fuel temperature, and critical heat flux. The inlet flow distribution of the core was optimized for the boiling margins by grouping the inlet orifices by several hydraulic regions. The procedure is called a hydraulic profiling

  9. Parallel Computation of the Jacobian Matrix for Nonlinear Equation Solvers Using MATLAB

    Rose, Geoffrey K.; Nguyen, Duc T.; Newman, Brett A.

    2017-01-01

    Demonstrating speedup for parallel code on a multicore shared memory PC can be challenging in MATLAB due to underlying parallel operations that are often opaque to the user. This can limit potential for improvement of serial code even for the so-called embarrassingly parallel applications. One such application is the computation of the Jacobian matrix inherent to most nonlinear equation solvers. Computation of this matrix represents the primary bottleneck in nonlinear solver speed such that commercial finite element (FE) and multi-body-dynamic (MBD) codes attempt to minimize computations. A timing study using MATLAB's Parallel Computing Toolbox was performed for numerical computation of the Jacobian. Several approaches for implementing parallel code were investigated while only the single program multiple data (spmd) method using composite objects provided positive results. Parallel code speedup is demonstrated but the goal of linear speedup through the addition of processors was not achieved due to PC architecture.

  10. Parallel Computational Intelligence-Based Multi-Camera Surveillance System

    Orts-Escolano, Sergio; Garcia-Rodriguez, Jose; Morell, Vicente; Cazorla, Miguel; Azorin-Lopez, Jorge; García-Chamizo, Juan Manuel

    2014-01-01

    In this work, we present a multi-camera surveillance system based on the use of self-organizing neural networks to represent events on video. The system processes several tasks in parallel using GPUs (graphic processor units). It addresses multiple vision tasks at various levels, such as segmentation, representation or characterization, analysis and monitoring of the movement. These features allow the construction of a robust representation of the environment and interpret the behavior of mob...

  11. Overview of the Force Scientific Parallel Language

    Gita Alaghband

    1994-01-01

    Full Text Available The Force parallel programming language designed for large-scale shared-memory multiprocessors is presented. The language provides a number of parallel constructs as extensions to the ordinary Fortran language and is implemented as a two-level macro preprocessor to support portability across shared memory multiprocessors. The global parallelism model on which the Force is based provides a powerful parallel language. The parallel constructs, generic synchronization, and freedom from process management supported by the Force has resulted in structured parallel programs that are ported to the many multiprocessors on which the Force is implemented. Two new parallel constructs for looping and functional decomposition are discussed. Several programming examples to illustrate some parallel programming approaches using the Force are also presented.

  12. Automatic Loop Parallelization via Compiler Guided Refactoring

    Larsen, Per; Ladelsky, Razya; Lidman, Jacob

    For many parallel applications, performance relies not on instruction-level parallelism, but on loop-level parallelism. Unfortunately, many modern applications are written in ways that obstruct automatic loop parallelization. Since we cannot identify sufficient parallelization opportunities...... for these codes in a static, off-line compiler, we developed an interactive compilation feedback system that guides the programmer in iteratively modifying application source, thereby improving the compiler’s ability to generate loop-parallel code. We use this compilation system to modify two sequential...... benchmarks, finding that the code parallelized in this way runs up to 8.3 times faster on an octo-core Intel Xeon 5570 system and up to 12.5 times faster on a quad-core IBM POWER6 system. Benchmark performance varies significantly between the systems. This suggests that semi-automatic parallelization should...

  13. Parallel kinematics type, kinematics, and optimal design

    Liu, Xin-Jun

    2014-01-01

    Parallel Kinematics- Type, Kinematics, and Optimal Design presents the results of 15 year's research on parallel mechanisms and parallel kinematics machines. This book covers the systematic classification of parallel mechanisms (PMs) as well as providing a large number of mechanical architectures of PMs available for use in practical applications. It focuses on the kinematic design of parallel robots. One successful application of parallel mechanisms in the field of machine tools, which is also called parallel kinematics machines, has been the emerging trend in advanced machine tools. The book describes not only the main aspects and important topics in parallel kinematics, but also references novel concepts and approaches, i.e. type synthesis based on evolution, performance evaluation and optimization based on screw theory, singularity model taking into account motion and force transmissibility, and others.   This book is intended for researchers, scientists, engineers and postgraduates or above with interes...

  14. Applied Parallel Computing Industrial Computation and Optimization

    Madsen, Kaj; NA NA NA Olesen, Dorte

    Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)......Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)...

  15. Distributed Memory Parallel Computing with SEAWAT

    Verkaik, J.; Huizer, S.; van Engelen, J.; Oude Essink, G.; Ram, R.; Vuik, K.

    2017-12-01

    Fresh groundwater reserves in coastal aquifers are threatened by sea-level rise, extreme weather conditions, increasing urbanization and associated groundwater extraction rates. To counteract these threats, accurate high-resolution numerical models are required to optimize the management of these precious reserves. The major model drawbacks are long run times and large memory requirements, limiting the predictive power of these models. Distributed memory parallel computing is an efficient technique for reducing run times and memory requirements, where the problem is divided over multiple processor cores. A new Parallel Krylov Solver (PKS) for SEAWAT is presented. PKS has recently been applied to MODFLOW and includes Conjugate Gradient (CG) and Biconjugate Gradient Stabilized (BiCGSTAB) linear accelerators. Both accelerators are preconditioned by an overlapping additive Schwarz preconditioner in a way that: a) subdomains are partitioned using Recursive Coordinate Bisection (RCB) load balancing, b) each subdomain uses local memory only and communicates with other subdomains by Message Passing Interface (MPI) within the linear accelerator, c) it is fully integrated in SEAWAT. Within SEAWAT, the PKS-CG solver replaces the Preconditioned Conjugate Gradient (PCG) solver for solving the variable-density groundwater flow equation and the PKS-BiCGSTAB solver replaces the Generalized Conjugate Gradient (GCG) solver for solving the advection-diffusion equation. PKS supports the third-order Total Variation Diminishing (TVD) scheme for computing advection. Benchmarks were performed on the Dutch national supercomputer (https://userinfo.surfsara.nl/systems/cartesius) using up to 128 cores, for a synthetic 3D Henry model (100 million cells) and the real-life Sand Engine model ( 10 million cells). The Sand Engine model was used to investigate the potential effect of the long-term morphological evolution of a large sand replenishment and climate change on fresh groundwater resources

  16. Parallel algorithms and cluster computing

    Hoffmann, Karl Heinz

    2007-01-01

    This book presents major advances in high performance computing as well as major advances due to high performance computing. It contains a collection of papers in which results achieved in the collaboration of scientists from computer science, mathematics, physics, and mechanical engineering are presented. From the science problems to the mathematical algorithms and on to the effective implementation of these algorithms on massively parallel and cluster computers we present state-of-the-art methods and technology as well as exemplary results in these fields. This book shows that problems which seem superficially distinct become intimately connected on a computational level.

  17. Parallel computation of rotating flows

    Lundin, Lars Kristian; Barker, Vincent A.; Sørensen, Jens Nørkær

    1999-01-01

    This paper deals with the simulation of 3‐D rotating flows based on the velocity‐vorticity formulation of the Navier‐Stokes equations in cylindrical coordinates. The governing equations are discretized by a finite difference method. The solution is advanced to a new time level by a two‐step process...... is that of solving a singular, large, sparse, over‐determined linear system of equations, and the iterative method CGLS is applied for this purpose. We discuss some of the mathematical and numerical aspects of this procedure and report on the performance of our software on a wide range of parallel computers. Darbe...

  18. A four-element based transposon system for allele specific tagging ...

    The four-element based construct would include both Ds and dSpm along with relevant marker genes and .... subsequent generations after transposition has taken place so as to ..... Cardon G H, Frey M, Seadler H and Gierl A 1993 Definition.

  19. DQB1*06:02 allele-specific expression varies by allelic dosage, not narcolepsy status

    Weiner Lachmi, Karin; Lin, Ling; Kornum, Birgitte Rahbek

    2012-01-01

    The association of narcolepsy-cataplexy, a sleep disorder caused by the loss of hypocretin/orexin neurons in the hypothalamus, with DQA1*01:02-DQB1*06:02 is one of the tightest known single-allele human leukocyte antigen (HLA) associations. In this study, we explored genome-wide expression...

  20. Differential ACE expression among tissues in allele-specific Wistar rat lines

    Kamilic, Jelena; Lely, A. Titia; van Goor, Harry; Buikema, Hendrik; Tent, Hilde; Navis, Gerjan J.; Korstanje, Ron

    In humans, the insertion/deletion polymorphism in the angiotensin converting enzyme (ACE) gene accounts for half of the variance in plasma ACE activity. The deletion allele is associated with high plasma ACE activity, cardiovascular disease, and renal disease. In rat, a similar association is found

  1. Mutations in the p53 homolog p63: allele-specific developmental syndromes in humans.

    Bokhoven, J.H.L.M. van; McKeon, F.

    2002-01-01

    p63 is the most recently discovered but most ancient member of the p53 family. In marked contrast to p53, p63 is highly expressed in embryonic ectoderm and in the basal, regenerative layers of many epithelial tissues in the adult. The p63-knockout mouse dies at birth and lacks limbs, epidermis,

  2. Aberrant allele-specific replication, independent of parental origin, in blood cells of cancer patients

    Dotan, Zohar A; Dotan, Aviva; Ramon, Jacob; Avivi, Lydia

    2008-01-01

    Allelic counterparts of biallelically expressed genes display an epigenetic symmetry normally manifested by synchronous replication, different from genes subjected to monoallelic expression, which normally are characterized by an asynchronous mode of replication (well exemplified by the SNRPN imprinted locus). Malignancy was documented to be associated with gross modifications in the inherent replication-timing coordination between allelic counterparts of imprinted genes as well as of biallelically expressed loci. The cancer-related allelic replication timing aberrations are non-disease specific and appear in peripheral blood cells of cancer patients, including those with solid tumors. As such they offer potential blood markers for non-invasive cancer test. The present study was aimed to gain some insight into the mechanism leading to the replication timing alterations of genes in blood lymphocytes of cancer patients. Peripheral blood samples derived from patients with prostate cancer were chosen to represent the cancerous status, and samples taken from patients with no cancer but with benign prostate hyperplasia were used to portray the normal status. Fluorescence In Situ Hybridization (FISH) replication assay, applied to phytohemagglutinin (PHA)-stimulated blood lymphocytes, was used to evaluate the temporal order (either synchronous or asynchronous) of genes in the patients' cells. We demonstrated that: (i) the aberrant epigenetic profile, as delineated by the cancer status, is a reversible modification, evidenced by our ability to restore the normal patterns of replication in three unrelated loci (CEN15, SNRPN and RB1) by introducing an archetypical demethylating agent, 5-azacytidine; (ii) following the rehabilitating effect of demethylation, an imprinted gene (SNRPN) retains its original parental imprint; and (iii) the choice of an allele between early or late replication in the aberrant asynchronous replication, delineated by the cancer status, is not random but is independent of the parental origin. The non-disease specific aberrant epigenetic profile displayed in peripheral blood cells of patients with a solid tumour (unlike genetic aberrations) can be reversed, by an epigenetic drug applied in vitro, to the normal. It appears that the cancerous status differentiates between two allelic counterparts in a non-random manner, but independent of the parental origin

  3. Allele-specific characterization of alanine: glyoxylate aminotransferase variants associated with primary hyperoxaluria.

    Melissa D Lage

    Full Text Available Primary Hyperoxaluria Type 1 (PH1 is a rare autosomal recessive kidney stone disease caused by deficiency of the peroxisomal enzyme alanine: glyoxylate aminotransferase (AGT, which is involved in glyoxylate detoxification. Over 75 different missense mutations in AGT have been found associated with PH1. While some of the mutations have been found to affect enzyme activity, stability, and/or localization, approximately half of these mutations are completely uncharacterized. In this study, we sought to systematically characterize AGT missense mutations associated with PH1. To facilitate analysis, we used two high-throughput yeast-based assays: one that assesses AGT specific activity, and one that assesses protein stability. Approximately 30% of PH1-associated missense mutations are found in conjunction with a minor allele polymorphic variant, which can interact to elicit complex effects on protein stability and trafficking. To better understand this allele interaction, we functionally characterized each of 34 mutants on both the major (wild-type and minor allele backgrounds, identifying mutations that synergize with the minor allele. We classify these mutants into four distinct categories depending on activity/stability results in the different alleles. Twelve mutants were found to display reduced activity in combination with the minor allele, compared with the major allele background. When mapped on the AGT dimer structure, these mutants reveal localized regions of the protein that appear particularly sensitive to interactions with the minor allele variant. While the majority of the deleterious effects on activity in the minor allele can be attributed to synergistic interaction affecting protein stability, we identify one mutation, E274D, that appears to specifically affect activity when in combination with the minor allele.

  4. Genetic analysis of gamma-ray mutagenesis in yeast. II. Allele-specific control of mutagenesis

    McKee, R.H.; Lawrence, C.W.

    1979-01-01

    We find that partially different sets of gene functions are required for the production of different kinds of mutations induced by 60 Co γ rays in Saccharomyces cerevisiae. This observation is very similar to others made previously with respect to uv mutagenesis and confirms the conclusion that such distinctive patterns of genetic control reflect properties of the test alleles and their genetic locations, rather than the kinds of lesions required to revert them. The data also support the model of mutagenic repair outlined in the first paper of this series in which partially different sets of gene functions are required for the production of different kinds of mutations, the formation of mutations at different genetic sites and the induction of mutations by different mutagens

  5. Sequenza: allele-specific copy number and mutation profiles from tumor sequencing data

    Favero, Francesco; Joshi, Tejal; Marquard, Andrea Marion

    2015-01-01

    : Comparison between Sequenza/exome and SNP/ASCAT revealed strong correlation in cellularity (Pearson’s r = 0.90) and ploidy estimates (r = 0.42, or r = 0.94 after manual inspecting alternative solutions). This performance was noticeably superior to previously published algorithms. In addition, in artificial...

  6. Rapid identification of capybara (Hydrochaeris hydrochaeris through allele-specific PCR

    Flávio Henrique-Silva

    2005-07-01

    Full Text Available The capybara is the largest rodent in the world and is widely distributed throughout Central and South America.  It is an animal of economic interest due to the pleasant flavor of its meat and higher protein content in comparison  to beef and pork meat.  The hide, hair and fat also have economic advantages. Thus,  as an animal with such high economic potential, it is the target of hunters, even though  hunting capybara is prohibited by law in Brazil.   Due to their  similarities,  capybara meat  is easily confused with  pork  meat.   This  occurs  upon  the apprehension of the  meat  from hunters, as well as in some restaurants that serve capybara meat that was slaughtered clandestinely. In both cases, when the meat is confiscated, those responsible for the crimes claim it is pork meat,  hindering  the enforcement of the law. A practical  course was ministered  to undergraduate biology students enrolled in the elective course Introduction to Genetic  Engineering  at Federal  University  of Sao Carlos (UFSCar, Sao Paulo  State, Brazil.  The  objective  of the  course was to establish  and  apply  a Polymerase  Chain  Reaction  (PCR assay to identify capybara meat and discriminate it in relation  to other types of meat,  including pork. Primers  were designed based  on 12S rRNA,  transthyretin and  growth  hormone  receptor  genes.  The primers generated  capybara specific fragments  of approximately 220, 290 and 330 bp for transthyretin,12S rRNA  and  growth  hormone  receptor,  respectively.   The  duplexes  developed  in the  present work can be used effectively to discriminate capybara meat  from other  animals,  contributing to combating predatory capybara hunting. The results were extensively discussed and the students have contributed to written a paper  to be submitted to a publication.

  7. Ploidy mosaicism and allele-specific gene expression differences in the allopolyploid Squalius alburnoides

    Matos Isa

    2011-12-01

    Full Text Available Abstract Background Squalius alburnoides is an Iberian cyprinid fish resulting from an interspecific hybridisation between Squalius pyrenaicus females (P genome and males of an unknown Anaecypris hispanica-like species (A genome. S. alburnoides is an allopolyploid hybridogenetic complex, which makes it a likely candidate for ploidy mosaicism occurrence, and is also an interesting model to address questions about gene expression regulation and genomic interactions. Indeed, it was previously suggested that in S. alburnoides triploids (PAA composition silencing of one of the three alleles (mainly of the P allele occurs. However, not a whole haplome is inactivated but a more or less random inactivation of alleles varying between individuals and even between organs of the same fish was seen. In this work we intended to correlate expression differences between individuals and/or between organs to the occurrence of mosaicism, evaluating if mosaics could explain previous observations and its impact on the assessment of gene expression patterns. Results To achieve our goal, we developed flow cytometry and cell sorting protocols for this system generating more homogenous cellular and transcriptional samples. With this set-up we detected 10% ploidy mosaicism within the S. alburnoides complex, and determined the allelic expression profiles of ubiquitously expressed genes (rpl8; gapdh and β-actin in cells from liver and kidney of mosaic and non-mosaic individuals coming from different rivers over a wide geographic range. Conclusions Ploidy mosaicism occurs sporadically within the S. alburnoides complex, but in a frequency significantly higher than reported for other organisms. Moreover, we could exclude the influence of this phenomenon on the detection of variable allelic expression profiles of ubiquitously expressed genes (rpl8; gapdh and β-actin in cells from liver and kidney of triploid individuals. Finally, we determined that the expression patterns previously detected only in a narrow geographic range is not a local restricted phenomenon but is pervasive in rivers where S. pyrenaicus is sympatric with S. alburnoides. We discuss mechanisms that could lead to the formation of mosaic S. alburnoides and hypothesise about a relaxation of the mechanisms that impose a tight control over mitosis and ploidy control in mixoploids.

  8. The parallel volume at large distances

    Kampf, Jürgen

    In this paper we examine the asymptotic behavior of the parallel volume of planar non-convex bodies as the distance tends to infinity. We show that the difference between the parallel volume of the convex hull of a body and the parallel volume of the body itself tends to . This yields a new proof...... for the fact that a planar body can only have polynomial parallel volume, if it is convex. Extensions to Minkowski spaces and random sets are also discussed....

  9. The parallel volume at large distances

    Kampf, Jürgen

    In this paper we examine the asymptotic behavior of the parallel volume of planar non-convex bodies as the distance tends to infinity. We show that the difference between the parallel volume of the convex hull of a body and the parallel volume of the body itself tends to 0. This yields a new proof...... for the fact that a planar body can only have polynomial parallel volume, if it is convex. Extensions to Minkowski spaces and random sets are also discussed....

  10. Multiple sclerosis

    ... indwelling catheter Osteoporosis or thinning of the bones Pressure sores Side effects of medicines used to treat the ... Daily bowel care program Multiple sclerosis - discharge Preventing pressure ulcers Swallowing problems Images Multiple sclerosis MRI of the ...

  11. A Parallel Approach to Fractal Image Compression

    Lubomir Dedera

    2004-01-01

    Full Text Available The paper deals with a parallel approach to coding and decoding algorithms in fractal image compressionand presents experimental results comparing sequential and parallel algorithms from the point of view of achieved bothcoding and decoding time and effectiveness of parallelization.

  12. An Introduction to Parallel Computation R

    How are they programmed? This article provides an introduction. A parallel computer is a network of processors built for ... and have been used to solve problems much faster than a single ... in parallel computer design is to select an organization which ..... The most ambitious approach to parallel computing is to develop.

  13. Comparison of parallel viscosity with neoclassical theory

    Ida, K.; Nakajima, N.

    1996-04-01

    Toroidal rotation profiles are measured with charge exchange spectroscopy for the plasma heated with tangential NBI in CHS heliotron/torsatron device to estimate parallel viscosity. The parallel viscosity derived from the toroidal rotation velocity shows good agreement with the neoclassical parallel viscosity plus the perpendicular viscosity. (μ perpendicular = 2 m 2 /s). (author)

  14. Advances in randomized parallel computing

    Rajasekaran, Sanguthevar

    1999-01-01

    The technique of randomization has been employed to solve numerous prob­ lems of computing both sequentially and in parallel. Examples of randomized algorithms that are asymptotically better than their deterministic counterparts in solving various fundamental problems abound. Randomized algorithms have the advantages of simplicity and better performance both in theory and often in practice. This book is a collection of articles written by renowned experts in the area of randomized parallel computing. A brief introduction to randomized algorithms In the aflalysis of algorithms, at least three different measures of performance can be used: the best case, the worst case, and the average case. Often, the average case run time of an algorithm is much smaller than the worst case. 2 For instance, the worst case run time of Hoare's quicksort is O(n ), whereas its average case run time is only O( n log n). The average case analysis is conducted with an assumption on the input space. The assumption made to arrive at t...

  15. Xyce parallel electronic simulator design.

    Thornquist, Heidi K.; Rankin, Eric Lamont; Mei, Ting; Schiek, Richard Louis; Keiter, Eric Richard; Russo, Thomas V.

    2010-09-01

    This document is the Xyce Circuit Simulator developer guide. Xyce has been designed from the 'ground up' to be a SPICE-compatible, distributed memory parallel circuit simulator. While it is in many respects a research code, Xyce is intended to be a production simulator. As such, having software quality engineering (SQE) procedures in place to insure a high level of code quality and robustness are essential. Version control, issue tracking customer support, C++ style guildlines and the Xyce release process are all described. The Xyce Parallel Electronic Simulator has been under development at Sandia since 1999. Historically, Xyce has mostly been funded by ASC, the original focus of Xyce development has primarily been related to circuits for nuclear weapons. However, this has not been the only focus and it is expected that the project will diversify. Like many ASC projects, Xyce is a group development effort, which involves a number of researchers, engineers, scientists, mathmaticians and computer scientists. In addition to diversity of background, it is to be expected on long term projects for there to be a certain amount of staff turnover, as people move on to different projects. As a result, it is very important that the project maintain high software quality standards. The point of this document is to formally document a number of the software quality practices followed by the Xyce team in one place. Also, it is hoped that this document will be a good source of information for new developers.

  16. Improving parallel imaging by jointly reconstructing multi-contrast data.

    Bilgic, Berkin; Kim, Tae Hyung; Liao, Congyu; Manhard, Mary Kate; Wald, Lawrence L; Haldar, Justin P; Setsompop, Kawin

    2018-08-01

    To develop parallel imaging techniques that simultaneously exploit coil sensitivity encoding, image phase prior information, similarities across multiple images, and complementary k-space sampling for highly accelerated data acquisition. We introduce joint virtual coil (JVC)-generalized autocalibrating partially parallel acquisitions (GRAPPA) to jointly reconstruct data acquired with different contrast preparations, and show its application in 2D, 3D, and simultaneous multi-slice (SMS) acquisitions. We extend the joint parallel imaging concept to exploit limited support and smooth phase constraints through Joint (J-) LORAKS formulation. J-LORAKS allows joint parallel imaging from limited autocalibration signal region, as well as permitting partial Fourier sampling and calibrationless reconstruction. We demonstrate highly accelerated 2D balanced steady-state free precession with phase cycling, SMS multi-echo spin echo, 3D multi-echo magnetization-prepared rapid gradient echo, and multi-echo gradient recalled echo acquisitions in vivo. Compared to conventional GRAPPA, proposed joint acquisition/reconstruction techniques provide more than 2-fold reduction in reconstruction error. JVC-GRAPPA takes advantage of additional spatial encoding from phase information and image similarity, and employs different sampling patterns across acquisitions. J-LORAKS achieves a more parsimonious low-rank representation of local k-space by considering multiple images as additional coils. Both approaches provide dramatic improvement in artifact and noise mitigation over conventional single-contrast parallel imaging reconstruction. Magn Reson Med 80:619-632, 2018. © 2018 International Society for Magnetic Resonance in Medicine. © 2018 International Society for Magnetic Resonance in Medicine.

  17. PDDP, A Data Parallel Programming Model

    Karen H. Warren

    1996-01-01

    Full Text Available PDDP, the parallel data distribution preprocessor, is a data parallel programming model for distributed memory parallel computers. PDDP implements high-performance Fortran-compatible data distribution directives and parallelism expressed by the use of Fortran 90 array syntax, the FORALL statement, and the WHERE construct. Distributed data objects belong to a global name space; other data objects are treated as local and replicated on each processor. PDDP allows the user to program in a shared memory style and generates codes that are portable to a variety of parallel machines. For interprocessor communication, PDDP uses the fastest communication primitives on each platform.

  18. Parallelization of quantum molecular dynamics simulation code

    Kato, Kaori; Kunugi, Tomoaki; Shibahara, Masahiko; Kotake, Susumu

    1998-02-01

    A quantum molecular dynamics simulation code has been developed for the analysis of the thermalization of photon energies in the molecule or materials in Kansai Research Establishment. The simulation code is parallelized for both Scalar massively parallel computer (Intel Paragon XP/S75) and Vector parallel computer (Fujitsu VPP300/12). Scalable speed-up has been obtained with a distribution to processor units by division of particle group in both parallel computers. As a result of distribution to processor units not only by particle group but also by the particles calculation that is constructed with fine calculations, highly parallelization performance is achieved in Intel Paragon XP/S75. (author)

  19. Implementation and performance of parallelized elegant

    Wang, Y.; Borland, M.

    2008-01-01

    The program elegant is widely used for design and modeling of linacs for free-electron lasers and energy recovery linacs, as well as storage rings and other applications. As part of a multi-year effort, we have parallelized many aspects of the code, including single-particle dynamics, wakefields, and coherent synchrotron radiation. We report on the approach used for gradual parallelization, which proved very beneficial in getting parallel features into the hands of users quickly. We also report details of parallelization of collective effects. Finally, we discuss performance of the parallelized code in various applications.

  20. MULTIPLE OBJECTS

    A. A. Bosov

    2015-04-01

    Full Text Available Purpose. The development of complicated techniques of production and management processes, information systems, computer science, applied objects of systems theory and others requires improvement of mathematical methods, new approaches for researches of application systems. And the variety and diversity of subject systems makes necessary the development of a model that generalizes the classical sets and their development – sets of sets. Multiple objects unlike sets are constructed by multiple structures and represented by the structure and content. The aim of the work is the analysis of multiple structures, generating multiple objects, the further development of operations on these objects in application systems. Methodology. To achieve the objectives of the researches, the structure of multiple objects represents as constructive trio, consisting of media, signatures and axiomatic. Multiple object is determined by the structure and content, as well as represented by hybrid superposition, composed of sets, multi-sets, ordered sets (lists and heterogeneous sets (sequences, corteges. Findings. In this paper we study the properties and characteristics of the components of hybrid multiple objects of complex systems, proposed assessments of their complexity, shown the rules of internal and external operations on objects of implementation. We introduce the relation of arbitrary order over multiple objects, we define the description of functions and display on objects of multiple structures. Originality.In this paper we consider the development of multiple structures, generating multiple objects.Practical value. The transition from the abstract to the subject of multiple structures requires the transformation of the system and multiple objects. Transformation involves three successive stages: specification (binding to the domain, interpretation (multiple sites and particularization (goals. The proposed describe systems approach based on hybrid sets

  1. Parallelization of 2-D lattice Boltzmann codes

    Suzuki, Soichiro; Kaburaki, Hideo; Yokokawa, Mitsuo.

    1996-03-01

    Lattice Boltzmann (LB) codes to simulate two dimensional fluid flow are developed on vector parallel computer Fujitsu VPP500 and scalar parallel computer Intel Paragon XP/S. While a 2-D domain decomposition method is used for the scalar parallel LB code, a 1-D domain decomposition method is used for the vector parallel LB code to be vectorized along with the axis perpendicular to the direction of the decomposition. High parallel efficiency of 95.1% by the vector parallel calculation on 16 processors with 1152x1152 grid and 88.6% by the scalar parallel calculation on 100 processors with 800x800 grid are obtained. The performance models are developed to analyze the performance of the LB codes. It is shown by our performance models that the execution speed of the vector parallel code is about one hundred times faster than that of the scalar parallel code with the same number of processors up to 100 processors. We also analyze the scalability in keeping the available memory size of one processor element at maximum. Our performance model predicts that the execution time of the vector parallel code increases about 3% on 500 processors. Although the 1-D domain decomposition method has in general a drawback in the interprocessor communication, the vector parallel LB code is still suitable for the large scale and/or high resolution simulations. (author)

  2. Parallelization of 2-D lattice Boltzmann codes

    Suzuki, Soichiro; Kaburaki, Hideo; Yokokawa, Mitsuo

    1996-03-01

    Lattice Boltzmann (LB) codes to simulate two dimensional fluid flow are developed on vector parallel computer Fujitsu VPP500 and scalar parallel computer Intel Paragon XP/S. While a 2-D domain decomposition method is used for the scalar parallel LB code, a 1-D domain decomposition method is used for the vector parallel LB code to be vectorized along with the axis perpendicular to the direction of the decomposition. High parallel efficiency of 95.1% by the vector parallel calculation on 16 processors with 1152x1152 grid and 88.6% by the scalar parallel calculation on 100 processors with 800x800 grid are obtained. The performance models are developed to analyze the performance of the LB codes. It is shown by our performance models that the execution speed of the vector parallel code is about one hundred times faster than that of the scalar parallel code with the same number of processors up to 100 processors. We also analyze the scalability in keeping the available memory size of one processor element at maximum. Our performance model predicts that the execution time of the vector parallel code increases about 3% on 500 processors. Although the 1-D domain decomposition method has in general a drawback in the interprocessor communication, the vector parallel LB code is still suitable for the large scale and/or high resolution simulations. (author).

  3. Systematic approach for deriving feasible mappings of parallel algorithms to parallel computing platforms

    Arkin, Ethem; Tekinerdogan, Bedir; Imre, Kayhan M.

    2017-01-01

    The need for high-performance computing together with the increasing trend from single processor to parallel computer architectures has leveraged the adoption of parallel computing. To benefit from parallel computing power, usually parallel algorithms are defined that can be mapped and executed

  4. Experiences in Data-Parallel Programming

    Terry W. Clark

    1997-01-01

    Full Text Available To efficiently parallelize a scientific application with a data-parallel compiler requires certain structural properties in the source program, and conversely, the absence of others. A recent parallelization effort of ours reinforced this observation and motivated this correspondence. Specifically, we have transformed a Fortran 77 version of GROMOS, a popular dusty-deck program for molecular dynamics, into Fortran D, a data-parallel dialect of Fortran. During this transformation we have encountered a number of difficulties that probably are neither limited to this particular application nor do they seem likely to be addressed by improved compiler technology in the near future. Our experience with GROMOS suggests a number of points to keep in mind when developing software that may at some time in its life cycle be parallelized with a data-parallel compiler. This note presents some guidelines for engineering data-parallel applications that are compatible with Fortran D or High Performance Fortran compilers.

  5. Massively parallel diffuse optical tomography

    Sandusky, John V.; Pitts, Todd A.

    2017-09-05

    Diffuse optical tomography systems and methods are described herein. In a general embodiment, the diffuse optical tomography system comprises a plurality of sensor heads, the plurality of sensor heads comprising respective optical emitter systems and respective sensor systems. A sensor head in the plurality of sensors heads is caused to act as an illuminator, such that its optical emitter system transmits a transillumination beam towards a portion of a sample. Other sensor heads in the plurality of sensor heads act as observers, detecting portions of the transillumination beam that radiate from the sample in the fields of view of the respective sensory systems of the other sensor heads. Thus, sensor heads in the plurality of sensors heads generate sensor data in parallel.

  6. Embodied and Distributed Parallel DJing.

    Cappelen, Birgitta; Andersson, Anders-Petter

    2016-01-01

    Everyone has a right to take part in cultural events and activities, such as music performances and music making. Enforcing that right, within Universal Design, is often limited to a focus on physical access to public areas, hearing aids etc., or groups of persons with special needs performing in traditional ways. The latter might be people with disabilities, being musicians playing traditional instruments, or actors playing theatre. In this paper we focus on the innovative potential of including people with special needs, when creating new cultural activities. In our project RHYME our goal was to create health promoting activities for children with severe disabilities, by developing new musical and multimedia technologies. Because of the users' extreme demands and rich contribution, we ended up creating both a new genre of musical instruments and a new art form. We call this new art form Embodied and Distributed Parallel DJing, and the new genre of instruments for Empowering Multi-Sensorial Things.

  7. Device for balancing parallel strings

    Mashikian, Matthew S.

    1985-01-01

    A battery plant is described which features magnetic circuit means in association with each of the battery strings in the battery plant for balancing the electrical current flow through the battery strings by equalizing the voltage across each of the battery strings. Each of the magnetic circuit means generally comprises means for sensing the electrical current flow through one of the battery strings, and a saturable reactor having a main winding connected electrically in series with the battery string, a bias winding connected to a source of alternating current and a control winding connected to a variable source of direct current controlled by the sensing means. Each of the battery strings is formed by a plurality of batteries connected electrically in series, and these battery strings are connected electrically in parallel across common bus conductors.

  8. Parallel computing in enterprise modeling.

    Goldsby, Michael E.; Armstrong, Robert C.; Shneider, Max S.; Vanderveen, Keith; Ray, Jaideep; Heath, Zach; Allan, Benjamin A.

    2008-08-01

    This report presents the results of our efforts to apply high-performance computing to entity-based simulations with a multi-use plugin for parallel computing. We use the term 'Entity-based simulation' to describe a class of simulation which includes both discrete event simulation and agent based simulation. What simulations of this class share, and what differs from more traditional models, is that the result sought is emergent from a large number of contributing entities. Logistic, economic and social simulations are members of this class where things or people are organized or self-organize to produce a solution. Entity-based problems never have an a priori ergodic principle that will greatly simplify calculations. Because the results of entity-based simulations can only be realized at scale, scalable computing is de rigueur for large problems. Having said that, the absence of a spatial organizing principal makes the decomposition of the problem onto processors problematic. In addition, practitioners in this domain commonly use the Java programming language which presents its own problems in a high-performance setting. The plugin we have developed, called the Parallel Particle Data Model, overcomes both of these obstacles and is now being used by two Sandia frameworks: the Decision Analysis Center, and the Seldon social simulation facility. While the ability to engage U.S.-sized problems is now available to the Decision Analysis Center, this plugin is central to the success of Seldon. Because Seldon relies on computationally intensive cognitive sub-models, this work is necessary to achieve the scale necessary for realistic results. With the recent upheavals in the financial markets, and the inscrutability of terrorist activity, this simulation domain will likely need a capability with ever greater fidelity. High-performance computing will play an important part in enabling that greater fidelity.

  9. Parallel ecological networks in ecosystems

    Olff, Han; Alonso, David; Berg, Matty P.; Eriksson, B. Klemens; Loreau, Michel; Piersma, Theunis; Rooney, Neil

    2009-01-01

    In ecosystems, species interact with other species directly and through abiotic factors in multiple ways, often forming complex networks of various types of ecological interaction. Out of this suite of interactions, predator-prey interactions have received most attention. The resulting food webs,

  10. Plastic collapse behavior for thin tube with two parallel cracks

    Moon, Seong In; Chang, Yoon Suk; Kim, Young Jin; Lee, Jin Ho; Song, Myung Ho; Choi, Young Hwan; Kim, Joung Soo

    2004-01-01

    The current plugging criterion is known to be too conservative for some locations and types of defects. Many defects detected during in-service inspection take on the form of multiple cracks at the top of tube sheet but there is no reliable plugging criterion for the steam generator tubes with multiple cracks. Most of the previous studies on multiple cracks are confined to elastic analyses and only few studies have been done on the steam generator tubes failed by plastic collapse. Therefore, it is necessary to develop models which can be used to estimate the failure behavior of steam generator tubes with multiple cracks. The objective of this study is to verify the applicability of the optimum local failure prediction models proposed in the previous study. For this, plastic collapse tests are performed with the tube specimens containing two parallel through-wall cracks. The plastic collapse load of the steam generator tubes containing two parallel through-wall cracks are also estimated by using the proposed optimum global failure model and the applicability is investigated by comparing the estimated results with the experimental results. Also, the interaction effect between two cracks was evaluated to explain the plastic collapse behavior

  11. Massively Parallel Dimension Independent Adaptive Metropolis

    Chen, Yuxin

    2015-05-14

    This work considers black-box Bayesian inference over high-dimensional parameter spaces. The well-known and widely respected adaptive Metropolis (AM) algorithm is extended herein to asymptotically scale uniformly with respect to the underlying parameter dimension, by respecting the variance, for Gaussian targets. The result- ing algorithm, referred to as the dimension-independent adaptive Metropolis (DIAM) algorithm, also shows improved performance with respect to adaptive Metropolis on non-Gaussian targets. This algorithm is further improved, and the possibility of probing high-dimensional targets is enabled, via GPU-accelerated numerical libraries and periodically synchronized concurrent chains (justified a posteriori). Asymptoti- cally in dimension, this massively parallel dimension-independent adaptive Metropolis (MPDIAM) GPU implementation exhibits a factor of four improvement versus the CPU-based Intel MKL version alone, which is itself already a factor of three improve- ment versus the serial version. The scaling to multiple CPUs and GPUs exhibits a form of strong scaling in terms of the time necessary to reach a certain convergence criterion, through a combination of longer time per sample batch (weak scaling) and yet fewer necessary samples to convergence. This is illustrated by e ciently sampling from several Gaussian and non-Gaussian targets for dimension d 1000.

  12. Parallel algorithms for online trackfinding at PANDA

    Bianchi, Ludovico; Ritman, James; Stockmanns, Tobias [IKP, Forschungszentrum Juelich GmbH (Germany); Herten, Andreas [JSC, Forschungszentrum Juelich GmbH (Germany); Collaboration: PANDA-Collaboration

    2016-07-01

    The PANDA experiment, one of the four scientific pillars of the FAIR facility currently in construction in Darmstadt, is a next-generation particle detector that will study collisions of antiprotons with beam momenta of 1.5-15 GeV/c on a fixed proton target. Because of the broad physics scope and the similar signature of signal and background events, PANDA's strategy for data acquisition is to continuously record data from the whole detector and use this global information to perform online event reconstruction and filtering. A real-time rejection factor of up to 1000 must be achieved to match the incoming data rate for offline storage, making all components of the data processing system computationally very challenging. Online particle track identification and reconstruction is an essential step, since track information is used as input in all following phases. Online tracking algorithms must ensure a delicate balance between high tracking efficiency and quality, and minimal computational footprint. For this reason, a massively parallel solution exploiting multiple Graphic Processing Units (GPUs) is under investigation. The talk presents the core concepts of the algorithms being developed for primary trackfinding, along with details of their implementation on GPUs.

  13. OpenMP parallelization of a gridded SWAT (SWATG)

    Zhang, Ying; Hou, Jinliang; Cao, Yongpan; Gu, Juan; Huang, Chunlin

    2017-12-01

    Large-scale, long-term and high spatial resolution simulation is a common issue in environmental modeling. A Gridded Hydrologic Response Unit (HRU)-based Soil and Water Assessment Tool (SWATG) that integrates grid modeling scheme with different spatial representations also presents such problems. The time-consuming problem affects applications of very high resolution large-scale watershed modeling. The OpenMP (Open Multi-Processing) parallel application interface is integrated with SWATG (called SWATGP) to accelerate grid modeling based on the HRU level. Such parallel implementation takes better advantage of the computational power of a shared memory computer system. We conducted two experiments at multiple temporal and spatial scales of hydrological modeling using SWATG and SWATGP on a high-end server. At 500-m resolution, SWATGP was found to be up to nine times faster than SWATG in modeling over a roughly 2000 km2 watershed with 1 CPU and a 15 thread configuration. The study results demonstrate that parallel models save considerable time relative to traditional sequential simulation runs. Parallel computations of environmental models are beneficial for model applications, especially at large spatial and temporal scales and at high resolutions. The proposed SWATGP model is thus a promising tool for large-scale and high-resolution water resources research and management in addition to offering data fusion and model coupling ability.

  14. Exact parallel maximum clique algorithm for general and protein graphs.

    Depolli, Matjaž; Konc, Janez; Rozman, Kati; Trobec, Roman; Janežič, Dušanka

    2013-09-23

    A new exact parallel maximum clique algorithm MaxCliquePara, which finds the maximum clique (the fully connected subgraph) in undirected general and protein graphs, is presented. First, a new branch and bound algorithm for finding a maximum clique on a single computer core, which builds on ideas presented in two published state of the art sequential algorithms is implemented. The new sequential MaxCliqueSeq algorithm is faster than the reference algorithms on both DIMACS benchmark graphs as well as on protein-derived product graphs used for protein structural comparisons. Next, the MaxCliqueSeq algorithm is parallelized by splitting the branch-and-bound search tree to multiple cores, resulting in MaxCliquePara algorithm. The ability to exploit all cores efficiently makes the new parallel MaxCliquePara algorithm markedly superior to other tested algorithms. On a 12-core computer, the parallelization provides up to 2 orders of magnitude faster execution on the large DIMACS benchmark graphs and up to an order of magnitude faster execution on protein product graphs. The algorithms are freely accessible on http://commsys.ijs.si/~matjaz/maxclique.

  15. Tuning HDF5 subfiling performance on parallel file systems

    Byna, Suren [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Chaarawi, Mohamad [Intel Corp. (United States); Koziol, Quincey [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Mainzer, John [The HDF Group (United States); Willmore, Frank [The HDF Group (United States)

    2017-05-12

    Subfiling is a technique used on parallel file systems to reduce locking and contention issues when multiple compute nodes interact with the same storage target node. Subfiling provides a compromise between the single shared file approach that instigates the lock contention problems on parallel file systems and having one file per process, which results in generating a massive and unmanageable number of files. In this paper, we evaluate and tune the performance of recently implemented subfiling feature in HDF5. In specific, we explain the implementation strategy of subfiling feature in HDF5, provide examples of using the feature, and evaluate and tune parallel I/O performance of this feature with parallel file systems of the Cray XC40 system at NERSC (Cori) that include a burst buffer storage and a Lustre disk-based storage. We also evaluate I/O performance on the Cray XC30 system, Edison, at NERSC. Our results show performance benefits of 1.2X to 6X performance advantage with subfiling compared to writing a single shared HDF5 file. We present our exploration of configurations, such as the number of subfiles and the number of Lustre storage targets to storing files, as optimization parameters to obtain superior I/O performance. Based on this exploration, we discuss recommendations for achieving good I/O performance as well as limitations with using the subfiling feature.

  16. Semi-coarsening multigrid methods for parallel computing

    Jones, J.E.

    1996-12-31

    Standard multigrid methods are not well suited for problems with anisotropic coefficients which can occur, for example, on grids that are stretched to resolve a boundary layer. There are several different modifications of the standard multigrid algorithm that yield efficient methods for anisotropic problems. In the paper, we investigate the parallel performance of these multigrid algorithms. Multigrid algorithms which work well for anisotropic problems are based on line relaxation and/or semi-coarsening. In semi-coarsening multigrid algorithms a grid is coarsened in only one of the coordinate directions unlike standard or full-coarsening multigrid algorithms where a grid is coarsened in each of the coordinate directions. When both semi-coarsening and line relaxation are used, the resulting multigrid algorithm is robust and automatic in that it requires no knowledge of the nature of the anisotropy. This is the basic multigrid algorithm whose parallel performance we investigate in the paper. The algorithm is currently being implemented on an IBM SP2 and its performance is being analyzed. In addition to looking at the parallel performance of the basic semi-coarsening algorithm, we present algorithmic modifications with potentially better parallel efficiency. One modification reduces the amount of computational work done in relaxation at the expense of using multiple coarse grids. This modification is also being implemented with the aim of comparing its performance to that of the basic semi-coarsening algorithm.

  17. Compiler Technology for Parallel Scientific Computation

    Can Özturan

    1994-01-01

    Full Text Available There is a need for compiler technology that, given the source program, will generate efficient parallel codes for different architectures with minimal user involvement. Parallel computation is becoming indispensable in solving large-scale problems in science and engineering. Yet, the use of parallel computation is limited by the high costs of developing the needed software. To overcome this difficulty we advocate a comprehensive approach to the development of scalable architecture-independent software for scientific computation based on our experience with equational programming language (EPL. Our approach is based on a program decomposition, parallel code synthesis, and run-time support for parallel scientific computation. The program decomposition is guided by the source program annotations provided by the user. The synthesis of parallel code is based on configurations that describe the overall computation as a set of interacting components. Run-time support is provided by the compiler-generated code that redistributes computation and data during object program execution. The generated parallel code is optimized using techniques of data alignment, operator placement, wavefront determination, and memory optimization. In this article we discuss annotations, configurations, parallel code generation, and run-time support suitable for parallel programs written in the functional parallel programming language EPL and in Fortran.

  18. Computer-Aided Parallelizer and Optimizer

    Jin, Haoqiang

    2011-01-01

    The Computer-Aided Parallelizer and Optimizer (CAPO) automates the insertion of compiler directives (see figure) to facilitate parallel processing on Shared Memory Parallel (SMP) machines. While CAPO currently is integrated seamlessly into CAPTools (developed at the University of Greenwich, now marketed as ParaWise), CAPO was independently developed at Ames Research Center as one of the components for the Legacy Code Modernization (LCM) project. The current version takes serial FORTRAN programs, performs interprocedural data dependence analysis, and generates OpenMP directives. Due to the widely supported OpenMP standard, the generated OpenMP codes have the potential to run on a wide range of SMP machines. CAPO relies on accurate interprocedural data dependence information currently provided by CAPTools. Compiler directives are generated through identification of parallel loops in the outermost level, construction of parallel regions around parallel loops and optimization of parallel regions, and insertion of directives with automatic identification of private, reduction, induction, and shared variables. Attempts also have been made to identify potential pipeline parallelism (implemented with point-to-point synchronization). Although directives are generated automatically, user interaction with the tool is still important for producing good parallel codes. A comprehensive graphical user interface is included for users to interact with the parallelization process.

  19. Computationally efficient implementation of combustion chemistry in parallel PDF calculations

    Lu Liuyan; Lantz, Steven R.; Ren Zhuyin; Pope, Stephen B.

    2009-01-01

    In parallel calculations of combustion processes with realistic chemistry, the serial in situ adaptive tabulation (ISAT) algorithm [S.B. Pope, Computationally efficient implementation of combustion chemistry using in situ adaptive tabulation, Combustion Theory and Modelling, 1 (1997) 41-63; L. Lu, S.B. Pope, An improved algorithm for in situ adaptive tabulation, Journal of Computational Physics 228 (2009) 361-386] substantially speeds up the chemistry calculations on each processor. To improve the parallel efficiency of large ensembles of such calculations in parallel computations, in this work, the ISAT algorithm is extended to the multi-processor environment, with the aim of minimizing the wall clock time required for the whole ensemble. Parallel ISAT strategies are developed by combining the existing serial ISAT algorithm with different distribution strategies, namely purely local processing (PLP), uniformly random distribution (URAN), and preferential distribution (PREF). The distribution strategies enable the queued load redistribution of chemistry calculations among processors using message passing. They are implemented in the software x2f m pi, which is a Fortran 95 library for facilitating many parallel evaluations of a general vector function. The relative performance of the parallel ISAT strategies is investigated in different computational regimes via the PDF calculations of multiple partially stirred reactors burning methane/air mixtures. The results show that the performance of ISAT with a fixed distribution strategy strongly depends on certain computational regimes, based on how much memory is available and how much overlap exists between tabulated information on different processors. No one fixed strategy consistently achieves good performance in all the regimes. Therefore, an adaptive distribution strategy, which blends PLP, URAN and PREF, is devised and implemented. It yields consistently good performance in all regimes. In the adaptive parallel

  20. Multiple sclerosis

    Grunwald, I.Q.; Kuehn, A.L.; Backens, M.; Papanagiotou, P.; Shariat, K.; Kostopoulos, P.

    2008-01-01

    Multiple sclerosis is the most common chronic inflammatory disease of myelin with interspersed lesions in the white matter of the central nervous system. Magnetic resonance imaging (MRI) plays a key role in the diagnosis and monitoring of white matter diseases. This article focuses on key findings in multiple sclerosis as detected by MRI. (orig.) [de

  1. Parallel processing for fluid dynamics applications

    Johnson, G.M.

    1989-01-01

    The impact of parallel processing on computational science and, in particular, on computational fluid dynamics is growing rapidly. In this paper, particular emphasis is given to developments which have occurred within the past two years. Parallel processing is defined and the reasons for its importance in high-performance computing are reviewed. Parallel computer architectures are classified according to the number and power of their processing units, their memory, and the nature of their connection scheme. Architectures which show promise for fluid dynamics applications are emphasized. Fluid dynamics problems are examined for parallelism inherent at the physical level. CFD algorithms and their mappings onto parallel architectures are discussed. Several example are presented to document the performance of fluid dynamics applications on present-generation parallel processing devices

  2. Design considerations for parallel graphics libraries

    Crockett, Thomas W.

    1994-01-01

    Applications which run on parallel supercomputers are often characterized by massive datasets. Converting these vast collections of numbers to visual form has proven to be a powerful aid to comprehension. For a variety of reasons, it may be desirable to provide this visual feedback at runtime. One way to accomplish this is to exploit the available parallelism to perform graphics operations in place. In order to do this, we need appropriate parallel rendering algorithms and library interfaces. This paper provides a tutorial introduction to some of the issues which arise in designing parallel graphics libraries and their underlying rendering algorithms. The focus is on polygon rendering for distributed memory message-passing systems. We illustrate our discussion with examples from PGL, a parallel graphics library which has been developed on the Intel family of parallel systems.

  3. A parallelized three-dimensional cellular automaton model for grain growth during additive manufacturing

    Lian, Yanping; Lin, Stephen; Yan, Wentao; Liu, Wing Kam; Wagner, Gregory J.

    2018-01-01

    In this paper, a parallelized 3D cellular automaton computational model is developed to predict grain morphology for solidification of metal during the additive manufacturing process. Solidification phenomena are characterized by highly localized events, such as the nucleation and growth of multiple grains. As a result, parallelization requires careful treatment of load balancing between processors as well as interprocess communication in order to maintain a high parallel efficiency. We give a detailed summary of the formulation of the model, as well as a description of the communication strategies implemented to ensure parallel efficiency. Scaling tests on a representative problem with about half a billion cells demonstrate parallel efficiency of more than 80% on 8 processors and around 50% on 64; loss of efficiency is attributable to load imbalance due to near-surface grain nucleation in this test problem. The model is further demonstrated through an additive manufacturing simulation with resulting grain structures showing reasonable agreement with those observed in experiments.

  4. A parallelized three-dimensional cellular automaton model for grain growth during additive manufacturing

    Lian, Yanping; Lin, Stephen; Yan, Wentao; Liu, Wing Kam; Wagner, Gregory J.

    2018-05-01

    In this paper, a parallelized 3D cellular automaton computational model is developed to predict grain morphology for solidification of metal during the additive manufacturing process. Solidification phenomena are characterized by highly localized events, such as the nucleation and growth of multiple grains. As a result, parallelization requires careful treatment of load balancing between processors as well as interprocess communication in order to maintain a high parallel efficiency. We give a detailed summary of the formulation of the model, as well as a description of the communication strategies implemented to ensure parallel efficiency. Scaling tests on a representative problem with about half a billion cells demonstrate parallel efficiency of more than 80% on 8 processors and around 50% on 64; loss of efficiency is attributable to load imbalance due to near-surface grain nucleation in this test problem. The model is further demonstrated through an additive manufacturing simulation with resulting grain structures showing reasonable agreement with those observed in experiments.

  5. OpenMP Issues Arising in the Development of Parallel BLAS and LAPACK Libraries

    C. Addison

    2003-01-01

    Full Text Available Dense linear algebra libraries need to cope efficiently with a range of input problem sizes and shapes. Inherently this means that parallel implementations have to exploit parallelism wherever it is present. While OpenMP allows relatively fine grain parallelism to be exploited in a shared memory environment it currently lacks features to make it easy to partition computation over multiple array indices or to overlap sequential and parallel computations. The inherent flexible nature of shared memory paradigms such as OpenMP poses other difficulties when it becomes necessary to optimise performance across successive parallel library calls. Notions borrowed from distributed memory paradigms, such as explicit data distributions help address some of these problems, but the focus on data rather than work distribution appears misplaced in an SMP context.

  6. A language for data-parallel and task parallel programming dedicated to multi-SIMD computers. Contributions to hydrodynamic simulation with lattice gases

    Pic, Marc Michel

    1995-01-01

    Parallel programming covers task-parallelism and data-parallelism. Many problems need both parallelisms. Multi-SIMD computers allow hierarchical approach of these parallelisms. The T++ language, based on C++, is dedicated to exploit Multi-SIMD computers using a programming paradigm which is an extension of array-programming to tasks managing. Our language introduced array of independent tasks to achieve separately (MIMD), on subsets of processors of identical behaviour (SIMD), in order to translate the hierarchical inclusion of data-parallelism in task-parallelism. To manipulate in a symmetrical way tasks and data we propose meta-operations which have the same behaviour on tasks arrays and on data arrays. We explain how to implement this language on our parallel computer SYMPHONIE in order to profit by the locally-shared memory, by the hardware virtualization, and by the multiplicity of communications networks. We analyse simultaneously a typical application of such architecture. Finite elements scheme for Fluid mechanic needs powerful parallel computers and requires large floating points abilities. Lattice gases is an alternative to such simulations. Boolean lattice bases are simple, stable, modular, need to floating point computation, but include numerical noise. Boltzmann lattice gases present large precision of computation, but needs floating points and are only locally stable. We propose a new scheme, called multi-bit, who keeps the advantages of each boolean model to which it is applied, with large numerical precision and reduced noise. Experiments on viscosity, physical behaviour, noise reduction and spurious invariants are shown and implementation techniques for parallel Multi-SIMD computers detailed. (author) [fr

  7. Parallel processing from applications to systems

    Moldovan, Dan I

    1993-01-01

    This text provides one of the broadest presentations of parallelprocessing available, including the structure of parallelprocessors and parallel algorithms. The emphasis is on mappingalgorithms to highly parallel computers, with extensive coverage ofarray and multiprocessor architectures. Early chapters provideinsightful coverage on the analysis of parallel algorithms andprogram transformations, effectively integrating a variety ofmaterial previously scattered throughout the literature. Theory andpractice are well balanced across diverse topics in this concisepresentation. For exceptional cla

  8. Parallel processing for artificial intelligence 1

    Kanal, LN; Kumar, V; Suttner, CB

    1994-01-01

    Parallel processing for AI problems is of great current interest because of its potential for alleviating the computational demands of AI procedures. The articles in this book consider parallel processing for problems in several areas of artificial intelligence: image processing, knowledge representation in semantic networks, production rules, mechanization of logic, constraint satisfaction, parsing of natural language, data filtering and data mining. The publication is divided into six sections. The first addresses parallel computing for processing and understanding images. The second discus

  9. A survey of parallel multigrid algorithms

    Chan, Tony F.; Tuminaro, Ray S.

    1987-01-01

    A typical multigrid algorithm applied to well-behaved linear-elliptic partial-differential equations (PDEs) is described. Criteria for designing and evaluating parallel algorithms are presented. Before evaluating the performance of some parallel multigrid algorithms, consideration is given to some theoretical complexity results for solving PDEs in parallel and for executing the multigrid algorithm. The effect of mapping and load imbalance on the partial efficiency of the algorithm is studied.

  10. Refinement of Parallel and Reactive Programs

    Back, R. J. R.

    1992-01-01

    We show how to apply the refinement calculus to stepwise refinement of parallel and reactive programs. We use action systems as our basic program model. Action systems are sequential programs which can be implemented in a parallel fashion. Hence refinement calculus methods, originally developed for sequential programs, carry over to the derivation of parallel programs. Refinement of reactive programs is handled by data refinement techniques originally developed for the sequential refinement c...

  11. PPOOLEX experiments with two parallel blowdown pipes

    Laine, J.; Puustinen, M.; Raesaenen, A. (Lappeenranta Univ. of Technology, Nuclear Safety Research Unit (Finland))

    2011-01-15

    This report summarizes the results of the experiments with two transparent blowdown pipes carried out with the scaled down PPOOLEX test facility designed and constructed at Lappeenranta University of Technology. Steam was blown into the dry well compartment and from there through either one or two vertical transparent blowdown pipes to the condensation pool. Five experiments with one pipe and six with two parallel pipes were carried out. The main purpose of the experiments was to study loads caused by chugging (rapid condensation) while steam is discharged into the condensation pool filled with sub-cooled water. The PPOOLEX test facility is a closed stainless steel vessel divided into two compartments, dry well and wet well. In the experiments the initial temperature of the condensation pool water varied from 12 deg. C to 55 deg. C, the steam flow rate from 40 g/s to 1 300 g/s and the temperature of incoming steam from 120 deg. C to 185 deg. C. In the experiments with only one transparent blowdown pipe chugging phenomenon didn't occur as intensified as in the preceding experiments carried out with a DN200 stainless steel pipe. With the steel blowdown pipe even 10 times higher pressure pulses were registered inside the pipe. Meanwhile, loads registered in the pool didn't indicate significant differences between the steel and polycarbonate pipe experiments. In the experiments with two transparent blowdown pipes, the steamwater interface moved almost synchronously up and down inside both pipes. Chugging was stronger than in the one pipe experiments and even two times higher loads were measured inside the pipes. The loads at the blowdown pipe outlet were approximately the same as in the one pipe cases. Other registered loads around the pool were about 50-100 % higher than with one pipe. The experiments with two parallel blowdown pipes gave contradictory results compared to the earlier studies dealing with chugging loads in case of multiple pipes. Contributing

  12. Parallel processing approach to transform-based image coding

    Normile, James O.; Wright, Dan; Chu, Ken; Yeh, Chia L.

    1991-06-01

    This paper describes a flexible parallel processing architecture designed for use in real time video processing. The system consists of floating point DSP processors connected to each other via fast serial links, each processor has access to a globally shared memory. A multiple bus architecture in combination with a dual ported memory allows communication with a host control processor. The system has been applied to prototyping of video compression and decompression algorithms. The decomposition of transform based algorithms for decompression into a form suitable for parallel processing is described. A technique for automatic load balancing among the processors is developed and discussed, results ar presented with image statistics and data rates. Finally techniques for accelerating the system throughput are analyzed and results from the application of one such modification described.

  13. Concurrent, parallel, multiphysics coupling in the FACETS project

    Cary, J R; Carlsson, J A; Hakim, A H; Kruger, S E; Miah, M; Pletzer, A; Shasharina, S [Tech-X Corporation, 5621 Arapahoe Avenue, Suite A, Boulder, CO 80303 (United States); Candy, J; Groebner, R J [General Atomics (United States); Cobb, J; Fahey, M R [Oak Ridge National Laboratory (United States); Cohen, R H; Epperly, T [Lawrence Livermore National Laboratory (United States); Estep, D J [Colorado State University (United States); Krasheninnikov, S [University of California at San Diego (United States); Malony, A D [ParaTools, Inc (United States); McCune, D C [Princeton Plasma Physics Laboratory (United States); McInnes, L; Balay, S [Argonne National Laboratory (United States); Pankin, A, E-mail: cary@txcorp.co [Lehigh University (United States)

    2009-07-01

    FACETS (Framework Application for Core-Edge Transport Simulations), is now in its third year. The FACETS team has developed a framework for concurrent coupling of parallel computational physics for use on Leadership Class Facilities (LCFs). In the course of the last year, FACETS has tackled many of the difficult problems of moving to parallel, integrated modeling by developing algorithms for coupled systems, extracting legacy applications as components, modifying them to run on LCFs, and improving the performance of all components. The development of FACETS abides by rigorous engineering standards, including cross platform build and test systems, with the latter covering regression, performance, and visualization. In addition, FACETS has demonstrated the ability to incorporate full turbulence computations for the highest fidelity transport computations. Early indications are that the framework, using such computations, scales to multiple tens of thousands of processors. These accomplishments were a result of an interdisciplinary collaboration among computational physics, computer scientists and applied mathematicians on the team.

  14. Parallel Geometries in Geant4 foundation and recent enhancements

    Apostolakis, J; Cosmo, G; Howard, A; Ivanchenko, V; Verderi, M

    2009-01-01

    The Geant4 software toolkit simulates the passage of particles through matter. It is utilized in high energy and nuclear physics experiments, in medical physics and space applications. For many applications it is necessary to measure particle fluxes and radiation doses in parts of the setup where there are complex structures. To undertake this in a flexible way, Geant4 has tools to create and use additional, parallel, geometrical hierarchies within a single application. A separate, parallel geometry can be used for each one amongst shower parameterization, event biasing, scoring of radiation, and/or the creation of hits in detailed readout structures. We describe the existing basic capabilities of the Geant4 toolkit to create multiple geometries and the recent major enhancements undertaken to streamline, enhance and extend these. New functionality enables Geant4 developers to offer new embedded schemes for scoring (requiring no user C++ code); has simplified the implementation of processes or capabilities usi...

  15. Parallel Prediction of Stock Volatility

    Priscilla Jenq

    2017-10-01

    Full Text Available Volatility is a measurement of the risk of financial products. A stock will hit new highs and lows over time and if these highs and lows fluctuate wildly, then it is considered a high volatile stock. Such a stock is considered riskier than a stock whose volatility is low. Although highly volatile stocks are riskier, the returns that they generate for investors can be quite high. Of course, with a riskier stock also comes the chance of losing money and yielding negative returns. In this project, we will use historic stock data to help us forecast volatility. Since the financial industry usually uses S&P 500 as the indicator of the market, we will use S&P 500 as a benchmark to compute the risk. We will also use artificial neural networks as a tool to predict volatilities for a specific time frame that will be set when we configure this neural network. There have been reports that neural networks with different numbers of layers and different numbers of hidden nodes may generate varying results. In fact, we may be able to find the best configuration of a neural network to compute volatilities. We will implement this system using the parallel approach. The system can be used as a tool for investors to allocating and hedging assets.

  16. Vectoring of parallel synthetic jets

    Berk, Tim; Ganapathisubramani, Bharathram; Gomit, Guillaume

    2015-11-01

    A pair of parallel synthetic jets can be vectored by applying a phase difference between the two driving signals. The resulting jet can be merged or bifurcated and either vectored towards the actuator leading in phase or the actuator lagging in phase. In the present study, the influence of phase difference and Strouhal number on the vectoring behaviour is examined experimentally. Phase-locked vorticity fields, measured using Particle Image Velocimetry (PIV), are used to track vortex pairs. The physical mechanisms that explain the diversity in vectoring behaviour are observed based on the vortex trajectories. For a fixed phase difference, the vectoring behaviour is shown to be primarily influenced by pinch-off time of vortex rings generated by the synthetic jets. Beyond a certain formation number, the pinch-off timescale becomes invariant. In this region, the vectoring behaviour is determined by the distance between subsequent vortex rings. We acknowledge the financial support from the European Research Council (ERC grant agreement no. 277472).

  17. A Soft Parallel Kinematic Mechanism.

    White, Edward L; Case, Jennifer C; Kramer-Bottiglio, Rebecca

    2018-02-01

    In this article, we describe a novel holonomic soft robotic structure based on a parallel kinematic mechanism. The design is based on the Stewart platform, which uses six sensors and actuators to achieve full six-degree-of-freedom motion. Our design is much less complex than a traditional platform, since it replaces the 12 spherical and universal joints found in a traditional Stewart platform with a single highly deformable elastomer body and flexible actuators. This reduces the total number of parts in the system and simplifies the assembly process. Actuation is achieved through coiled-shape memory alloy actuators. State observation and feedback is accomplished through the use of capacitive elastomer strain gauges. The main structural element is an elastomer joint that provides antagonistic force. We report the response of the actuators and sensors individually, then report the response of the complete assembly. We show that the completed robotic system is able to achieve full position control, and we discuss the limitations associated with using responsive material actuators. We believe that control demonstrated on a single body in this work could be extended to chains of such bodies to create complex soft robots.

  18. Existence of parallel spinors on non-simply-connected Riemannian manifolds

    McInnes, B.

    1997-04-01

    It is well known, and important for applications, that Ricci-flat Riemannian manifolds of non-generic holonomy always admit a parallel [covariant constant] spinor if they are simply connected. The non-simply-connected case is much more subtle, however. We show that a parallel spinor can still be found in this case provided that the [real] dimension is not a multiple of four, and provided that the spin structure is carefully chosen. (author). 10 refs

  19. Productive Parallel Programming: The PCN Approach

    Ian Foster

    1992-01-01

    Full Text Available We describe the PCN programming system, focusing on those features designed to improve the productivity of scientists and engineers using parallel supercomputers. These features include a simple notation for the concise specification of concurrent algorithms, the ability to incorporate existing Fortran and C code into parallel applications, facilities for reusing parallel program components, a portable toolkit that allows applications to be developed on a workstation or small parallel computer and run unchanged on supercomputers, and integrated debugging and performance analysis tools. We survey representative scientific applications and identify problem classes for which PCN has proved particularly useful.

  20. High performance parallel I/O

    Prabhat

    2014-01-01

    Gain Critical Insight into the Parallel I/O EcosystemParallel I/O is an integral component of modern high performance computing (HPC), especially in storing and processing very large datasets to facilitate scientific discovery. Revealing the state of the art in this field, High Performance Parallel I/O draws on insights from leading practitioners, researchers, software architects, developers, and scientists who shed light on the parallel I/O ecosystem.The first part of the book explains how large-scale HPC facilities scope, configure, and operate systems, with an emphasis on choices of I/O har

  1. Parallel, Rapid Diffuse Optical Tomography of Breast

    Yodh, Arjun

    2001-01-01

    During the last year we have experimentally and computationally investigated rapid acquisition and analysis of informationally dense diffuse optical data sets in the parallel plate compressed breast geometry...

  2. Parallel, Rapid Diffuse Optical Tomography of Breast

    Yodh, Arjun

    2002-01-01

    During the last year we have experimentally and computationally investigated rapid acquisition and analysis of informationally dense diffuse optical data sets in the parallel plate compressed breast geometry...

  3. Parallel auto-correlative statistics with VTK.

    Pebay, Philippe Pierre; Bennett, Janine Camille

    2013-08-01

    This report summarizes existing statistical engines in VTK and presents both the serial and parallel auto-correlative statistics engines. It is a sequel to [PT08, BPRT09b, PT09, BPT09, PT10] which studied the parallel descriptive, correlative, multi-correlative, principal component analysis, contingency, k-means, and order statistics engines. The ease of use of the new parallel auto-correlative statistics engine is illustrated by the means of C++ code snippets and algorithm verification is provided. This report justifies the design of the statistics engines with parallel scalability in mind, and provides scalability and speed-up analysis results for the autocorrelative statistics engine.

  4. Conformal pure radiation with parallel rays

    Leistner, Thomas; Paweł Nurowski

    2012-01-01

    We define pure radiation metrics with parallel rays to be n-dimensional pseudo-Riemannian metrics that admit a parallel null line bundle K and whose Ricci tensor vanishes on vectors that are orthogonal to K. We give necessary conditions in terms of the Weyl, Cotton and Bach tensors for a pseudo-Riemannian metric to be conformal to a pure radiation metric with parallel rays. Then, we derive conditions in terms of the tractor calculus that are equivalent to the existence of a pure radiation metric with parallel rays in a conformal class. We also give analogous results for n-dimensional pseudo-Riemannian pp-waves. (paper)

  5. Compiling Scientific Programs for Scalable Parallel Systems

    Kennedy, Ken

    2001-01-01

    ...). The research performed in this project included new techniques for recognizing implicit parallelism in sequential programs, a powerful and precise set-based framework for analysis and transformation...

  6. Parallel thermal radiation transport in two dimensions

    Smedley-Stevenson, R.P.; Ball, S.R.

    2003-01-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  7. Parallel Algorithms for the Exascale Era

    Robey, Robert W. [Los Alamos National Laboratory

    2016-10-19

    New parallel algorithms are needed to reach the Exascale level of parallelism with millions of cores. We look at some of the research developed by students in projects at LANL. The research blends ideas from the early days of computing while weaving in the fresh approach brought by students new to the field of high performance computing. We look at reproducibility of global sums and why it is important to parallel computing. Next we look at how the concept of hashing has led to the development of more scalable algorithms suitable for next-generation parallel computers. Nearly all of this work has been done by undergraduates and published in leading scientific journals.

  8. Parallel thermal radiation transport in two dimensions

    Smedley-Stevenson, R.P.; Ball, S.R. [AWE Aldermaston (United Kingdom)

    2003-07-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  9. Wing-Body Aeroelasticity Using Finite-Difference Fluid/Finite-Element Structural Equations on Parallel Computers

    Byun, Chansup; Guruswamy, Guru P.; Kutler, Paul (Technical Monitor)

    1994-01-01

    In recent years significant advances have been made for parallel computers in both hardware and software. Now parallel computers have become viable tools in computational mechanics. Many application codes developed on conventional computers have been modified to benefit from parallel computers. Significant speedups in some areas have been achieved by parallel computations. For single-discipline use of both fluid dynamics and structural dynamics, computations have been made on wing-body configurations using parallel computers. However, only a limited amount of work has been completed in combining these two disciplines for multidisciplinary applications. The prime reason is the increased level of complication associated with a multidisciplinary approach. In this work, procedures to compute aeroelasticity on parallel computers using direct coupling of fluid and structural equations will be investigated for wing-body configurations. The parallel computer selected for computations is an Intel iPSC/860 computer which is a distributed-memory, multiple-instruction, multiple data (MIMD) computer with 128 processors. In this study, the computational efficiency issues of parallel integration of both fluid and structural equations will be investigated in detail. The fluid and structural domains will be modeled using finite-difference and finite-element approaches, respectively. Results from the parallel computer will be compared with those from the conventional computers using a single processor. This study will provide an efficient computational tool for the aeroelastic analysis of wing-body structures on MIMD type parallel computers.

  10. Analysis of Retransmission Policies for Parallel Data Transmission

    I. A. Halepoto

    2018-06-01

    Full Text Available Stream control transmission protocol (SCTP is a transport layer protocol, which is efficient, reliable, and connection-oriented as compared to transmission control protocol (TCP and user datagram protocol (UDP. Additionally, SCTP has more innovative features like multihoming, multistreaming and unordered delivery. With multihoming, SCTP establishes multiple paths between a sender and receiver. However, it only uses the primary path for data transmission and the secondary path (or paths for fault tolerance. Concurrent multipath transfer extension of SCTP (CMT-SCTP allows a sender to transmit data in parallel over multiple paths, which increases the overall transmission throughput. Parallel data transmission is beneficial for higher data rates. Parallel transmission or connection is also good in services such as video streaming where if one connection is occupied with errors the transmission continues on alternate links. With parallel transmission, the unordered data packets arrival is very common at receiver. The receiver has to wait until the missing data packets arrive, causing performance degradation while using CMT-SCTP. In order to reduce the transmission delay at the receiver, CMT-SCTP uses intelligent retransmission polices to immediately retransmit the missing packets. The retransmission policies used by CMT-SCTP are RTX-SSTHRESH, RTX-LOSSRATE and RTX-CWND. The main objective of this paper is the performance analysis of the retransmission policies. This paper evaluates RTX-SSTHRESH, RTX-LOSSRATE and RTX-CWND. Simulations are performed on the Network Simulator 2. In the simulations with various scenarios and parameters, it is observed that the RTX-LOSSRATE is a suitable policy.

  11. A High-Performance Parallel FDTD Method Enhanced by Using SSE Instruction Set

    Dau-Chyrh Chang

    2012-01-01

    Full Text Available We introduce a hardware acceleration technique for the parallel finite difference time domain (FDTD method using the SSE (streaming (single instruction multiple data SIMD extensions instruction set. The implementation of SSE instruction set to parallel FDTD method has achieved the significant improvement on the simulation performance. The benchmarks of the SSE acceleration on both the multi-CPU workstation and computer cluster have demonstrated the advantages of (vector arithmetic logic unit VALU acceleration over GPU acceleration. Several engineering applications are employed to demonstrate the performance of parallel FDTD method enhanced by SSE instruction set.

  12. Optimization under uncertainty of parallel nonlinear energy sinks

    Boroson, Ethan; Missoum, Samy; Mattei, Pierre-Olivier; Vergez, Christophe

    2017-04-01

    Nonlinear Energy Sinks (NESs) are a promising technique for passively reducing the amplitude of vibrations. Through nonlinear stiffness properties, a NES is able to passively and irreversibly absorb energy. Unlike the traditional Tuned Mass Damper (TMD), NESs do not require a specific tuning and absorb energy over a wider range of frequencies. Nevertheless, they are still only efficient over a limited range of excitations. In order to mitigate this limitation and maximize the efficiency range, this work investigates the optimization of multiple NESs configured in parallel. It is well known that the efficiency of a NES is extremely sensitive to small perturbations in loading conditions or design parameters. In fact, the efficiency of a NES has been shown to be nearly discontinuous in the neighborhood of its activation threshold. For this reason, uncertainties must be taken into account in the design optimization of NESs. In addition, the discontinuities require a specific treatment during the optimization process. In this work, the objective of the optimization is to maximize the expected value of the efficiency of NESs in parallel. The optimization algorithm is able to tackle design variables with uncertainty (e.g., nonlinear stiffness coefficients) as well as aleatory variables such as the initial velocity of the main system. The optimal design of several parallel NES configurations for maximum mean efficiency is investigated. Specifically, NES nonlinear stiffness properties, considered random design variables, are optimized for cases with 1, 2, 3, 4, 5, and 10 NESs in parallel. The distributions of efficiency for the optimal parallel configurations are compared to distributions of efficiencies of non-optimized NESs. It is observed that the optimization enables a sharp increase in the mean value of efficiency while reducing the corresponding variance, thus leading to more robust NES designs.

  13. Multiple homicides.

    Copeland, A R

    1989-09-01

    A study of multiple homicides or multiple deaths involving a solitary incident of violence by another individual was performed on the case files of the Office of the Medical Examiner of Metropolitan Dade County in Miami, Florida, during 1983-1987. A total of 107 multiple homicides were studied: 88 double, 17 triple, one quadruple, and one quintuple. The 236 victims were analyzed regarding age, race, sex, cause of death, toxicologic data, perpetrator, locale of the incident, and reason for the incident. This article compares this type of slaying with other types of homicide including those perpetrated by serial killers. Suggestions for future research in this field are offered.

  14. Balancing Structure for Multiple Generator

    LUPU Ciprian

    2014-05-01

    Full Text Available This paper presents a strategy to (rebalance a multiple generator control system structure on maintaining the global output in case of load and functioning disturbances. Applicability is proved on a control structure of the two and three sources connected in parallel to produce energy, a situation that has been encountered more and more these days especially in the renewable energy industry (wind, solar and small generators etc.

  15. Parallelization of the Physical-Space Statistical Analysis System (PSAS)

    Larson, J. W.; Guo, J.; Lyster, P. M.

    1999-01-01

    Atmospheric data assimilation is a method of combining observations with model forecasts to produce a more accurate description of the atmosphere than the observations or forecast alone can provide. Data assimilation plays an increasingly important role in the study of climate and atmospheric chemistry. The NASA Data Assimilation Office (DAO) has developed the Goddard Earth Observing System Data Assimilation System (GEOS DAS) to create assimilated datasets. The core computational components of the GEOS DAS include the GEOS General Circulation Model (GCM) and the Physical-space Statistical Analysis System (PSAS). The need for timely validation of scientific enhancements to the data assimilation system poses computational demands that are best met by distributed parallel software. PSAS is implemented in Fortran 90 using object-based design principles. The analysis portions of the code solve two equations. The first of these is the "innovation" equation, which is solved on the unstructured observation grid using a preconditioned conjugate gradient (CG) method. The "analysis" equation is a transformation from the observation grid back to a structured grid, and is solved by a direct matrix-vector multiplication. Use of a factored-operator formulation reduces the computational complexity of both the CG solver and the matrix-vector multiplication, rendering the matrix-vector multiplications as a successive product of operators on a vector. Sparsity is introduced to these operators by partitioning the observations using an icosahedral decomposition scheme. PSAS builds a large (approx. 128MB) run-time database of parameters used in the calculation of these operators. Implementing a message passing parallel computing paradigm into an existing yet developing computational system as complex as PSAS is nontrivial. One of the technical challenges is balancing the requirements for computational reproducibility with the need for high performance. The problem of computational

  16. Parallel Computing for Brain Simulation.

    Pastur-Romay, L A; Porto-Pazos, A B; Cedron, F; Pazos, A

    2017-01-01

    The human brain is the most complex system in the known universe, it is therefore one of the greatest mysteries. It provides human beings with extraordinary abilities. However, until now it has not been understood yet how and why most of these abilities are produced. For decades, researchers have been trying to make computers reproduce these abilities, focusing on both understanding the nervous system and, on processing data in a more efficient way than before. Their aim is to make computers process information similarly to the brain. Important technological developments and vast multidisciplinary projects have allowed creating the first simulation with a number of neurons similar to that of a human brain. This paper presents an up-to-date review about the main research projects that are trying to simulate and/or emulate the human brain. They employ different types of computational models using parallel computing: digital models, analog models and hybrid models. This review includes the current applications of these works, as well as future trends. It is focused on various works that look for advanced progress in Neuroscience and still others which seek new discoveries in Computer Science (neuromorphic hardware, machine learning techniques). Their most outstanding characteristics are summarized and the latest advances and future plans are presented. In addition, this review points out the importance of considering not only neurons: Computational models of the brain should also include glial cells, given the proven importance of astrocytes in information processing. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  17. Parallel implementation of many-body mean-field equations

    Chinn, C.R.; Umar, A.S.; Vallieres, M.; Strayer, M.R.

    1994-01-01

    We describe the numerical methods used to solve the system of stiff, nonlinear partial differential equations resulting from the Hartree-Fock description of many-particle quantum systems, as applied to the structure of the nucleus. The solutions are performed on a three-dimensional Cartesian lattice. Discretization is achieved through the lattice basis-spline collocation method, in which quantum-state vectors and coordinate-space operators are expressed in terms of basis-spline functions on a spatial lattice. All numerical procedures reduce to a series of matrix-vector multiplications and other elementary operations, which we perform on a number of different computing architectures, including the Intel Paragon and the Intel iPSC/860 hypercube. Parallelization is achieved through a combination of mechanisms employing the Gram-Schmidt procedure, broadcasts, global operations, and domain decomposition of state vectors. We discuss the approach to the problems of limited node memory and node-to-node communication overhead inherent in using distributed-memory, multiple-instruction, multiple-data stream parallel computers. An algorithm was developed to reduce the communication overhead by pipelining some of the message passing procedures

  18. Multiple Sclerosis

    Multiple sclerosis (MS) is a nervous system disease that affects your brain and spinal cord. It damages the myelin sheath, the material that surrounds and protects your nerve cells. This damage slows down ...

  19. Multiple myeloma.

    Collins, Conor D

    2012-02-01

    Advances in the imaging and treatment of multiple myeloma have occurred over the past decade. This article summarises the current status and highlights how an understanding of both is necessary for optimum management.

  20. Multiple mononeuropathy

    ... with multiple mononeuropathy are prone to new nerve injuries at pressure points such as the knees and elbows. They should avoid putting pressure on these areas, for example, by not leaning on the elbows, crossing the knees, ...

  1. High-Performance Psychometrics: The Parallel-E Parallel-M Algorithm for Generalized Latent Variable Models. Research Report. ETS RR-16-34

    von Davier, Matthias

    2016-01-01

    This report presents results on a parallel implementation of the expectation-maximization (EM) algorithm for multidimensional latent variable models. The developments presented here are based on code that parallelizes both the E step and the M step of the parallel-E parallel-M algorithm. Examples presented in this report include item response…

  2. The language parallel Pascal and other aspects of the massively parallel processor

    Reeves, A. P.; Bruner, J. D.

    1982-01-01

    A high level language for the Massively Parallel Processor (MPP) was designed. This language, called Parallel Pascal, is described in detail. A description of the language design, a description of the intermediate language, Parallel P-Code, and details for the MPP implementation are included. Formal descriptions of Parallel Pascal and Parallel P-Code are given. A compiler was developed which converts programs in Parallel Pascal into the intermediate Parallel P-Code language. The code generator to complete the compiler for the MPP is being developed independently. A Parallel Pascal to Pascal translator was also developed. The architecture design for a VLSI version of the MPP was completed with a description of fault tolerant interconnection networks. The memory arrangement aspects of the MPP are discussed and a survey of other high level languages is given.

  3. Parallelization of ultrasonic field simulations for non destructive testing

    Lambert, Jason

    2015-01-01

    The Non Destructive Testing field increasingly uses simulation. It is used at every step of the whole control process of an industrial part, from speeding up control development to helping experts understand results. During this thesis, a fast ultrasonic field simulation tool dedicated to the computation of an ultrasonic field radiated by a phase array probe in an isotropic specimen has been developed. During this thesis, a simulation tool dedicated to the fast computation of an ultrasonic field radiated by a phased array probe in an isotropic specimen has been developed. Its performance enables an interactive usage. To benefit from the commonly available parallel architectures, a regular model (aimed at removing divergent branching) derived from the generic CIVA model has been developed. First, a reference implementation was developed to validate this model against CIVA results, and to analyze its performance behaviour before optimization. The resulting code has been optimized for three kinds of parallel architectures commonly available in workstations: general purpose processors (GPP), many-core co-processors (Intel MIC) and graphics processing units (nVidia GPU). On the GPP and the MIC, the algorithm was reorganized and implemented to benefit from both parallelism levels, multithreading and vector instructions. On the GPU, the multiple steps of field computing have been divided in multiple successive CUDA kernels. Moreover, libraries dedicated to each architecture were used to speedup Fast Fourier Transforms, Intel MKL on GPP and MIC and nVidia cuFFT on GPU. Performance and hardware adequation of the produced codes were thoroughly studied for each architecture. On multiple realistic control configurations, interactive performance was reached. Perspectives to address more complex configurations were drawn. Finally, the integration and the industrialization of this code in the commercial NDT platform CIVA is discussed. (author) [fr

  4. Parallel Boltzmann machines : a mathematical model

    Zwietering, P.J.; Aarts, E.H.L.

    1991-01-01

    A mathematical model is presented for the description of parallel Boltzmann machines. The framework is based on the theory of Markov chains and combines a number of previously known results into one generic model. It is argued that parallel Boltzmann machines maximize a function consisting of a

  5. The convergence of parallel Boltzmann machines

    Zwietering, P.J.; Aarts, E.H.L.; Eckmiller, R.; Hartmann, G.; Hauske, G.

    1990-01-01

    We discuss the main results obtained in a study of a mathematical model of synchronously parallel Boltzmann machines. We present supporting evidence for the conjecture that a synchronously parallel Boltzmann machine maximizes a consensus function that consists of a weighted sum of the regular

  6. Customizable Memory Schemes for Data Parallel Architectures

    Gou, C.

    2011-01-01

    Memory system efficiency is crucial for any processor to achieve high performance, especially in the case of data parallel machines. Processing capabilities of parallel lanes will be wasted, when data requests are not accomplished in a sustainable and timely manner. Irregular vector memory accesses

  7. Parallel Narrative Structure in Paul Harding's "Tinkers"

    Çirakli, Mustafa Zeki

    2014-01-01

    The present paper explores the implications of parallel narrative structure in Paul Harding's "Tinkers" (2009). Besides primarily recounting the two sets of parallel narratives, "Tinkers" also comprises of seemingly unrelated fragments such as excerpts from clock repair manuals and diaries. The main stories, however, told…

  8. Streaming nested data parallelism on multicores

    Madsen, Frederik Meisner; Filinski, Andrzej

    2016-01-01

    The paradigm of nested data parallelism (NDP) allows a variety of semi-regular computation tasks to be mapped onto SIMD-style hardware, including GPUs and vector units. However, some care is needed to keep down space consumption in situations where the available parallelism may vastly exceed...

  9. Bayer image parallel decoding based on GPU

    Hu, Rihui; Xu, Zhiyong; Wei, Yuxing; Sun, Shaohua

    2012-11-01

    In the photoelectrical tracking system, Bayer image is decompressed in traditional method, which is CPU-based. However, it is too slow when the images become large, for example, 2K×2K×16bit. In order to accelerate the Bayer image decoding, this paper introduces a parallel speedup method for NVIDA's Graphics Processor Unit (GPU) which supports CUDA architecture. The decoding procedure can be divided into three parts: the first is serial part, the second is task-parallelism part, and the last is data-parallelism part including inverse quantization, inverse discrete wavelet transform (IDWT) as well as image post-processing part. For reducing the execution time, the task-parallelism part is optimized by OpenMP techniques. The data-parallelism part could advance its efficiency through executing on the GPU as CUDA parallel program. The optimization techniques include instruction optimization, shared memory access optimization, the access memory coalesced optimization and texture memory optimization. In particular, it can significantly speed up the IDWT by rewriting the 2D (Tow-dimensional) serial IDWT into 1D parallel IDWT. Through experimenting with 1K×1K×16bit Bayer image, data-parallelism part is 10 more times faster than CPU-based implementation. Finally, a CPU+GPU heterogeneous decompression system was designed. The experimental result shows that it could achieve 3 to 5 times speed increase compared to the CPU serial method.

  10. Parallelization of TMVA Machine Learning Algorithms

    Hajili, Mammad

    2017-01-01

    This report reflects my work on Parallelization of TMVA Machine Learning Algorithms integrated to ROOT Data Analysis Framework during summer internship at CERN. The report consists of 4 impor- tant part - data set used in training and validation, algorithms that multiprocessing applied on them, parallelization techniques and re- sults of execution time changes due to number of workers.

  11. 17 CFR 12.24 - Parallel proceedings.

    2010-04-01

    ...) Definition. For purposes of this section, a parallel proceeding shall include: (1) An arbitration proceeding... the receivership includes the resolution of claims made by customers; or (3) A petition filed under... any of the foregoing with knowledge of a parallel proceeding shall promptly notify the Commission, by...

  12. Parallel S/sub n/ iteration schemes

    Wienke, B.R.; Hiromoto, R.E.

    1986-01-01

    The iterative, multigroup, discrete ordinates (S/sub n/) technique for solving the linear transport equation enjoys widespread usage and appeal. Serial iteration schemes and numerical algorithms developed over the years provide a timely framework for parallel extension. On the Denelcor HEP, the authors investigate three parallel iteration schemes for solving the one-dimensional S/sub n/ transport equation. The multigroup representation and serial iteration methods are also reviewed. This analysis represents a first attempt to extend serial S/sub n/ algorithms to parallel environments and provides good baseline estimates on ease of parallel implementation, relative algorithm efficiency, comparative speedup, and some future directions. The authors examine ordered and chaotic versions of these strategies, with and without concurrent rebalance and diffusion acceleration. Two strategies efficiently support high degrees of parallelization and appear to be robust parallel iteration techniques. The third strategy is a weaker parallel algorithm. Chaotic iteration, difficult to simulate on serial machines, holds promise and converges faster than ordered versions of the schemes. Actual parallel speedup and efficiency are high and payoff appears substantial

  13. Parallel Computing Strategies for Irregular Algorithms

    Biswas, Rupak; Oliker, Leonid; Shan, Hongzhang; Biegel, Bryan (Technical Monitor)

    2002-01-01

    Parallel computing promises several orders of magnitude increase in our ability to solve realistic computationally-intensive problems, but relies on their efficient mapping and execution on large-scale multiprocessor architectures. Unfortunately, many important applications are irregular and dynamic in nature, making their effective parallel implementation a daunting task. Moreover, with the proliferation of parallel architectures and programming paradigms, the typical scientist is faced with a plethora of questions that must be answered in order to obtain an acceptable parallel implementation of the solution algorithm. In this paper, we consider three representative irregular applications: unstructured remeshing, sparse matrix computations, and N-body problems, and parallelize them using various popular programming paradigms on a wide spectrum of computer platforms ranging from state-of-the-art supercomputers to PC clusters. We present the underlying problems, the solution algorithms, and the parallel implementation strategies. Smart load-balancing, partitioning, and ordering techniques are used to enhance parallel performance. Overall results demonstrate the complexity of efficiently parallelizing irregular algorithms.

  14. Parallel fuzzy connected image segmentation on GPU

    Zhuge, Ying; Cao, Yong; Udupa, Jayaram K.; Miller, Robert W.

    2011-01-01

    Purpose: Image segmentation techniques using fuzzy connectedness (FC) principles have shown their effectiveness in segmenting a variety of objects in several large applications. However, one challenge in these algorithms has been their excessive computational requirements when processing large image datasets. Nowadays, commodity graphics hardware provides a highly parallel computing environment. In this paper, the authors present a parallel fuzzy connected image segmentation algorithm impleme...

  15. Non-Cartesian parallel imaging reconstruction.

    Wright, Katherine L; Hamilton, Jesse I; Griswold, Mark A; Gulani, Vikas; Seiberlich, Nicole

    2014-11-01

    Non-Cartesian parallel imaging has played an important role in reducing data acquisition time in MRI. The use of non-Cartesian trajectories can enable more efficient coverage of k-space, which can be leveraged to reduce scan times. These trajectories can be undersampled to achieve even faster scan times, but the resulting images may contain aliasing artifacts. Just as Cartesian parallel imaging can be used to reconstruct images from undersampled Cartesian data, non-Cartesian parallel imaging methods can mitigate aliasing artifacts by using additional spatial encoding information in the form of the nonhomogeneous sensitivities of multi-coil phased arrays. This review will begin with an overview of non-Cartesian k-space trajectories and their sampling properties, followed by an in-depth discussion of several selected non-Cartesian parallel imaging algorithms. Three representative non-Cartesian parallel imaging methods will be described, including Conjugate Gradient SENSE (CG SENSE), non-Cartesian generalized autocalibrating partially parallel acquisition (GRAPPA), and Iterative Self-Consistent Parallel Imaging Reconstruction (SPIRiT). After a discussion of these three techniques, several potential promising clinical applications of non-Cartesian parallel imaging will be covered. © 2014 Wiley Periodicals, Inc.

  16. Parallel Algorithms for Groebner-Basis Reduction

    1987-09-25

    22209 ELEMENT NO. NO. NO. ACCESSION NO. 11. TITLE (Include Security Classification) * PARALLEL ALGORITHMS FOR GROEBNER -BASIS REDUCTION 12. PERSONAL...All other editions are obsolete. Productivity Engineering in the UNIXt Environment p Parallel Algorithms for Groebner -Basis Reduction Technical Report

  17. Parallel knock-out schemes in networks

    Broersma, H.J.; Fomin, F.V.; Woeginger, G.J.

    2004-01-01

    We consider parallel knock-out schemes, a procedure on graphs introduced by Lampert and Slater in 1997 in which each vertex eliminates exactly one of its neighbors in each round. We are considering cases in which after a finite number of rounds, where the minimimum number is called the parallel

  18. Building a parallel file system simulator

    Molina-Estolano, E; Maltzahn, C; Brandt, S A; Bent, J

    2009-01-01

    Parallel file systems are gaining in popularity in high-end computing centers as well as commercial data centers. High-end computing systems are expected to scale exponentially and to pose new challenges to their storage scalability in terms of cost and power. To address these challenges scientists and file system designers will need a thorough understanding of the design space of parallel file systems. Yet there exist few systematic studies of parallel file system behavior at petabyte- and exabyte scale. An important reason is the significant cost of getting access to large-scale hardware to test parallel file systems. To contribute to this understanding we are building a parallel file system simulator that can simulate parallel file systems at very large scale. Our goal is to simulate petabyte-scale parallel file systems on a small cluster or even a single machine in reasonable time and fidelity. With this simulator, file system experts will be able to tune existing file systems for specific workloads, scientists and file system deployment engineers will be able to better communicate workload requirements, file system designers and researchers will be able to try out design alternatives and innovations at scale, and instructors will be able to study very large-scale parallel file system behavior in the class room. In this paper we describe our approach and provide preliminary results that are encouraging both in terms of fidelity and simulation scalability.

  19. Parallel Architectures for Planetary Exploration Requirements (PAPER)

    Cezzar, Ruknet

    1993-01-01

    The project's main contributions have been in the area of student support. Throughout the project, at least one, in some cases two, undergraduate students have been supported. By working with the project, these students gained valuable knowledge involving the scientific research project, including the not-so-pleasant reporting requirements to the funding agencies. The other important contribution was towards the establishment of a graduate program in computer science at Hampton University. Primarily, the PAPER project has served as the main research basis in seeking funds from other agencies, such as the National Science Foundation, for establishing a research infrastructure in the department. In technical areas, especially in the first phase, we believe the trip to Jet Propulsion Laboratory, and gathering together all the pertinent information involving experimental computer architectures aimed for planetary explorations was very helpful. Indeed, if this effort is to be revived in the future due to congressional funding for planetary explorations, say an unmanned mission to Mars, our interim report will be an important starting point. In other technical areas, our simulator has pinpointed and highlighted several important performance issues related to the design of operating system kernels for MIMD machines. In particular, the critical issue of how the kernel itself will run in parallel on a multiple-processor system has been addressed through the various ready list organization and access policies. In the area of neural computing, our main contribution was an introductory tutorial package to familiarize the researchers at NASA with this new and promising field zone axes (20). Finally, we have introduced the notion of reversibility in programming systems which may find applications in various areas of space research.

  20. Broadcasting a message in a parallel computer

    Berg, Jeremy E [Rochester, MN; Faraj, Ahmad A [Rochester, MN

    2011-08-02

    Methods, systems, and products are disclosed for broadcasting a message in a parallel computer. The parallel computer includes a plurality of compute nodes connected together using a data communications network. The data communications network optimized for point to point data communications and is characterized by at least two dimensions. The compute nodes are organized into at least one operational group of compute nodes for collective parallel operations of the parallel computer. One compute node of the operational group assigned to be a logical root. Broadcasting a message in a parallel computer includes: establishing a Hamiltonian path along all of the compute nodes in at least one plane of the data communications network and in the operational group; and broadcasting, by the logical root to the remaining compute nodes, the logical root's message along the established Hamiltonian path.