WorldWideScience

Sample records for lines substantially parallel

  1. Parallel Lines

    James G. Worner

    2017-05-01

    Full Text Available James Worner is an Australian-based writer and scholar currently pursuing a PhD at the University of Technology Sydney. His research seeks to expose masculinities lost in the shadow of Australia’s Anzac hegemony while exploring new opportunities for contemporary historiography. He is the recipient of the Doctoral Scholarship in Historical Consciousness at the university’s Australian Centre of Public History and will be hosted by the University of Bologna during 2017 on a doctoral research writing scholarship.   ‘Parallel Lines’ is one of a collection of stories, The Shapes of Us, exploring liminal spaces of modern life: class, gender, sexuality, race, religion and education. It looks at lives, like lines, that do not meet but which travel in proximity, simultaneously attracted and repelled. James’ short stories have been published in various journals and anthologies.

  2. Parallel plate transmission line transformer

    Voeten, S.J.; Brussaard, G.J.H.; Pemen, A.J.M.

    2011-01-01

    A Transmission Line Transformer (TLT) can be used to transform high-voltage nanosecond pulses. These transformers rely on the fact that the length of the pulse is shorter than the transmission lines used. This allows connecting the transmission lines in parallel at the input and in series at the

  3. Parallel inhomogeneity and the Alfven resonance. 1: Open field lines

    Hansen, P. J.; Harrold, B. G.

    1994-01-01

    In light of a recent demonstration of the general nonexistence of a singularity at the Alfven resonance in cold, ideal, linearized magnetohydrodynamics, we examine the effect of a small density gradient parallel to uniform, open ambient magnetic field lines. To lowest order, energy deposition is quantitatively unaffected but occurs continuously over a thickened layer. This effect is illustrated in a numerical analysis of a plasma sheet boundary layer model with perfectly absorbing boundary conditions. Consequences of the results are discussed, both for the open field line approximation and for the ensuing closed field line analysis.

  4. Highly parallel line-based image coding for many cores.

    Peng, Xiulian; Xu, Jizheng; Zhou, You; Wu, Feng

    2012-01-01

    Computers are developing along with a new trend from the dual-core and quad-core processors to ones with tens or even hundreds of cores. Multimedia, as one of the most important applications in computers, has an urgent need to design parallel coding algorithms for compression. Taking intraframe/image coding as a start point, this paper proposes a pure line-by-line coding scheme (LBLC) to meet the need. In LBLC, an input image is processed line by line sequentially, and each line is divided into small fixed-length segments. The compression of all segments from prediction to entropy coding is completely independent and concurrent at many cores. Results on a general-purpose computer show that our scheme can get a 13.9 times speedup with 15 cores at the encoder and a 10.3 times speedup at the decoder. Ideally, such near-linear speeding relation with the number of cores can be kept for more than 100 cores. In addition to the high parallelism, the proposed scheme can perform comparatively or even better than the H.264 high profile above middle bit rates. At near-lossless coding, it outperforms H.264 more than 10 dB. At lossless coding, up to 14% bit-rate reduction is observed compared with H.264 lossless coding at the high 4:4:4 profile.

  5. Model-driven product line engineering for mapping parallel algorithms to parallel computing platforms

    Arkin, Ethem; Tekinerdogan, Bedir

    2016-01-01

    Mapping parallel algorithms to parallel computing platforms requires several activities such as the analysis of the parallel algorithm, the definition of the logical configuration of the platform, the mapping of the algorithm to the logical configuration platform and the implementation of the

  6. Parallel field line and stream line tracing algorithms for space physics applications

    Toth, G.; de Zeeuw, D.; Monostori, G.

    2004-05-01

    Field line and stream line tracing is required in various space physics applications, such as the coupling of the global magnetosphere and inner magnetosphere models, the coupling of the solar energetic particle and heliosphere models, or the modeling of comets, where the multispecies chemical equations are solved along stream lines of a steady state solution obtained with single fluid MHD model. Tracing a vector field is an inherently serial process, which is difficult to parallelize. This is especially true when the data corresponding to the vector field is distributed over a large number of processors. We designed algorithms for the various applications, which scale well to a large number of processors. In the first algorithm the computational domain is divided into blocks. Each block is on a single processor. The algorithm folows the vector field inside the blocks, and calculates a mapping of the block surfaces. The blocks communicate the values at the coinciding surfaces, and the results are interpolated. Finally all block surfaces are defined and values inside the blocks are obtained. In the second algorithm all processors start integrating along the vector field inside the accessible volume. When the field line leaves the local subdomain, the position and other information is stored in a buffer. Periodically the processors exchange the buffers, and continue integration of the field lines until they reach a boundary. At that point the results are sent back to the originating processor. Efficiency is achieved by a careful phasing of computation and communication. In the third algorithm the results of a steady state simulation are stored on a hard drive. The vector field is contained in blocks. All processors read in all the grid and vector field data and the stream lines are integrated in parallel. If a stream line enters a block, which has already been integrated, the results can be interpolated. By a clever ordering of the blocks the execution speed can be

  7. 'Ogura'-based 'CMS' lines with different nuclear backgrounds of cabbage revealed substantial diversity at morphological and molecular levels.

    Parkash, Chander; Kumar, Sandeep; Singh, Rajender; Kumar, Ajay; Kumar, Satish; Dey, Shyam Sundar; Bhatia, Reeta; Kumar, Raj

    2018-01-01

    A comprehensive study on characterization and genetic diversity analysis was carried out in 16 'Ogura'-based 'CMS' lines of cabbage using 14 agro-morphological traits and 29 SSR markers. Agro-morphological characterization depicted considerable variations for different horticultural traits studied. The genotype, ZHA-2, performed better for most of the economically important quantitative traits. Further, gross head weight (0.76), head length (0.60) and head width (0.83) revealed significant positive correlation with net head weight. Dendrogram based on 10 quantitative traits exhibited considerable diversity among different CMS lines and principle component analysis (PCA) indicated that net and gross head weight, and head length and width are the main components of divergence between 16 CMS lines of cabbage. In molecular study, a total of 58 alleles were amplified by 29 SSR primers, averaging to 2.0 alleles in each locus. High mean values of Shannon's Information index (0.62), expected (0.45) and observed (0.32) heterozygosity and polymorphic information content (0.35) depicted substantial polymorphism. Dendrogram based on Jaccard's similarity coefficient constructed two major groups and eight sub-groups, which revealed substantial diversity among different CMS lines. In overall, based on agro-morphological and molecular studies genotype RRMA, ZHA-2 and RCA were found most divergent. Hence, they have immense potential in future breeding programs for the high-yielding hybrid development in cabbage.

  8. Parallel transport of long mean-free-path plasma along open magnetic field lines: Parallel heat flux

    Guo Zehua; Tang Xianzhu

    2012-01-01

    In a long mean-free-path plasma where temperature anisotropy can be sustained, the parallel heat flux has two components with one associated with the parallel thermal energy and the other the perpendicular thermal energy. Due to the large deviation of the distribution function from local Maxwellian in an open field line plasma with low collisionality, the conventional perturbative calculation of the parallel heat flux closure in its local or non-local form is no longer applicable. Here, a non-perturbative calculation is presented for a collisionless plasma in a two-dimensional flux expander bounded by absorbing walls. Specifically, closures of previously unfamiliar form are obtained for ions and electrons, which relate two distinct components of the species parallel heat flux to the lower order fluid moments such as density, parallel flow, parallel and perpendicular temperatures, and the field quantities such as the magnetic field strength and the electrostatic potential. The plasma source and boundary condition at the absorbing wall enter explicitly in the closure calculation. Although the closure calculation does not take into account wave-particle interactions, the results based on passing orbits from steady-state collisionless drift-kinetic equation show remarkable agreement with fully kinetic-Maxwell simulations. As an example of the physical implications of the theory, the parallel heat flux closures are found to predict a surprising observation in the kinetic-Maxwell simulation of the 2D magnetic flux expander problem, where the parallel heat flux of the parallel thermal energy flows from low to high parallel temperature region.

  9. A 32-channel lattice transmission line array for parallel transmit and receive MRI at 7 tesla.

    Adriany, Gregor; Auerbach, Edward J; Snyder, Carl J; Gözübüyük, Ark; Moeller, Steen; Ritter, Johannes; Van de Moortele, Pierre-François; Vaughan, Tommy; Uğurbil, Kâmil

    2010-06-01

    Transmit and receive RF coil arrays have proven to be particularly beneficial for ultra-high-field MR. Transmit coil arrays enable such techniques as B(1) (+) shimming to substantially improve transmit B(1) homogeneity compared to conventional volume coil designs, and receive coil arrays offer enhanced parallel imaging performance and SNR. Concentric coil arrangements hold promise for developing transceiver arrays incorporating large numbers of coil elements. At magnetic field strengths of 7 tesla and higher where the Larmor frequencies of interest can exceed 300 MHz, the coil array design must also overcome the problem of the coil conductor length approaching the RF wavelength. In this study, a novel concentric arrangement of resonance elements built from capacitively-shortened half-wavelength transmission lines is presented. This approach was utilized to construct an array with whole-brain coverage using 16 transceiver elements and 16 receive-only elements, resulting in a coil with a total of 16 transmit and 32 receive channels. (c) 2010 Wiley-Liss, Inc.

  10. Integrated configurable equipment selection and line balancing for mass production with serial-parallel machining systems

    Battaïa, Olga; Dolgui, Alexandre; Guschinsky, Nikolai; Levin, Genrikh

    2014-10-01

    Solving equipment selection and line balancing problems together allows better line configurations to be reached and avoids local optimal solutions. This article considers jointly these two decision problems for mass production lines with serial-parallel workplaces. This study was motivated by the design of production lines based on machines with rotary or mobile tables. Nevertheless, the results are more general and can be applied to assembly and production lines with similar structures. The designers' objectives and the constraints are studied in order to suggest a relevant mathematical model and an efficient optimization approach to solve it. A real case study is used to validate the model and the developed approach.

  11. On-line event reconstruction using a parallel in-memory data base

    Argante, E; Van der Stok, P D V; Willers, Ian Malcolm

    1995-01-01

    PORS is a system designed for on-line event reconstruction in high energy physics (HEP) experiments. It uses the CPREAD reconstruction program. Central to the system is a parallel in-memory database which is used as communication medium between parallel workers. A farming control structure is implemented with PORS in a natural way. The database provides structured storage of data with a short life time. PORS serves as a case study for the construction of a methodology on how to apply parallel...

  12. Passing in Command Line Arguments and Parallel Cluster/Multicore Batching in R with batch.

    Hoffmann, Thomas J

    2011-03-01

    It is often useful to rerun a command line R script with some slight change in the parameters used to run it - a new set of parameters for a simulation, a different dataset to process, etc. The R package batch provides a means to pass in multiple command line options, including vectors of values in the usual R format, easily into R. The same script can be setup to run things in parallel via different command line arguments. The R package batch also provides a means to simplify this parallel batching by allowing one to use R and an R-like syntax for arguments to spread a script across a cluster or local multicore/multiprocessor computer, with automated syntax for several popular cluster types. Finally it provides a means to aggregate the results together of multiple processes run on a cluster.

  13. Line filter design of parallel interleaved VSCs for high power wind energy conversion systems

    Gohil, Ghanshyamsinh Vijaysinh; Bede, Lorand; Teodorescu, Remus

    2015-01-01

    The Voltage Source Converters (VSCs) are often connected in parallel in a Wind Energy Conversion System (WECS) to match the high power rating of the modern wind turbines. The effect of the interleaved carriers on the harmonic performance of the parallel connected VSCs is analyzed in this paper...... limit. In order to achieve the desired filter performance with optimal values of the filter parameters, the use of a LC trap branch with the conventional LCL filter is proposed. The expressions for the resonant frequencies of the proposed line filter are derived and used in the design to selectively...

  14. Parallel diffusion calculation for the PHAETON on-line multiprocessor computer

    Collart, J.M.; Fedon-Magnaud, C.; Lautard, J.J.

    1987-04-01

    The aim of the PHAETON project is the design of an on-line computer in order to increase the immediate knowledge of the main operating and safety parameters in power plants. A significant stage is the computation of the three dimensional flux distribution. For cost and safety reason a computer based on a parallel microprocessor architecture has been studied. This paper presents a first approach to parallelized three dimensional diffusion calculation. A computing software has been written and built in a four processors demonstrator. We present the realization in progress, concerning the final equipment. 8 refs

  15. Critical care medicine as a distinct product line with substantial financial profitability: the role of business planning.

    Bekes, Carolyn E; Dellinger, R Phillip; Brooks, Daniel; Edmondson, Robert; Olivia, Christopher T; Parrillo, Joseph E

    2004-05-01

    As academic health centers face increasing financial pressures, they have adopted a more businesslike approach to planning, particularly for discrete "product" or clinical service lines. Since critical care typically has been viewed as a service provided by a hospital, and not a product line, business plans have not historically been developed to expand and promote critical care. The major focus when examining the finances of critical care has been cost reduction, not business development. We hypothesized that a critical care business plan can be developed and analyzed like other more typical product lines and that such a critical care product line can be profitable for an institution. In-depth analysis of critical care including business planning for critical care services. Regional academic health center in southern New Jersey. None. As part of an overall business planning process directed by the Board of Trustees, the critical care product line was identified by isolating revenue, expenses, and profitability associated with critical care patients. We were able to identify the major sources ("value chain") of critical care patients: the emergency room, patients who are admitted for other problems but spend time in a critical care unit, and patients transferred to our intensive care units from other hospitals. The greatest opportunity to expand the product line comes from increasing the referrals from other hospitals. A methodology was developed to identify the revenue and expenses associated with critical care, based on the analysis of past experience. With this model, we were able to demonstrate a positive contribution margin of dollar 7 million per year related to patients transferred to the institution primarily for critical care services. This can be seen as the profit related to the product line segment of critical care. There was an additional positive contribution margin of dollar 5.8 million attributed to the critical care portion of the hospital stay of

  16. Cubic systems with invariant affine straight lines of total parallel multiplicity seven

    Alexandru Suba

    2013-12-01

    Full Text Available In this article, we study the planar cubic differential systems with invariant affine straight lines of total parallel multiplicity seven. We classify these system according to their geometric properties encoded in the configurations of invariant straight lines. We show that there are only 17 different topological phase portraits in the Poincar\\'e disc associated to this family of cubic systems up to a reversal of the sense of their orbits, and we provide representatives of every class modulo an affine change of variables and rescaling of the time variable.

  17. Transmission line theory for long plasma production by radio frequency discharges between parallel-plate electrodes

    Nonaka, S.

    1991-01-01

    In order to seek for a radio frequency (RF) eigen-mode of waves in producing a plasma between a pair of long dielectric-covered parallel-plate RF electrodes, this paper analyzed all normal modes propagating along the electrodes by solving Maxwell's equations. The result showed that only an odd surface wave mode will produce the plasma in usual experimental conditions, which will become a basic transmission line theory when use of such long electrodes for on-line mass-production of amorphous silicon solar cells

  18. Implementation of a microcomputer based distance relay for parallel transmission lines

    Phadke, A.G.; Jihuang, L.

    1986-01-01

    Distance relaying for parallel transmission lines is a difficult application problem with conventional phase and ground distance relays. It is known that for cross-country faults involving dissimilar phases and ground, three phase tripping may result. This paper summarizes a newly developed microcomputer based relay which is capable of classifying the cross-country fault correctly. The paper describes the principle of operation and results of laboratory tests of this relay

  19. Parallel Hough Transform-Based Straight Line Detection and Its FPGA Implementation in Embedded Vision

    Nam Ling

    2013-07-01

    Full Text Available Hough Transform has been widely used for straight line detection in low-definition and still images, but it suffers from execution time and resource requirements. Field Programmable Gate Arrays (FPGA provide a competitive alternative for hardware acceleration to reap tremendous computing performance. In this paper, we propose a novel parallel Hough Transform (PHT and FPGA architecture-associated framework for real-time straight line detection in high-definition videos. A resource-optimized Canny edge detection method with enhanced non-maximum suppression conditions is presented to suppress most possible false edges and obtain more accurate candidate edge pixels for subsequent accelerated computation. Then, a novel PHT algorithm exploiting spatial angle-level parallelism is proposed to upgrade computational accuracy by improving the minimum computational step. Moreover, the FPGA based multi-level pipelined PHT architecture optimized by spatial parallelism ensures real-time computation for 1,024 × 768 resolution videos without any off-chip memory consumption. This framework is evaluated on ALTERA DE2-115 FPGA evaluation platform at a maximum frequency of 200 MHz, and it can calculate straight line parameters in 15.59 ms on the average for one frame. Qualitative and quantitative evaluation results have validated the system performance regarding data throughput, memory bandwidth, resource, speed and robustness.

  20. Parallel Hough Transform-based straight line detection and its FPGA implementation in embedded vision.

    Lu, Xiaofeng; Song, Li; Shen, Sumin; He, Kang; Yu, Songyu; Ling, Nam

    2013-07-17

    Hough Transform has been widely used for straight line detection in low-definition and still images, but it suffers from execution time and resource requirements. Field Programmable Gate Arrays (FPGA) provide a competitive alternative for hardware acceleration to reap tremendous computing performance. In this paper, we propose a novel parallel Hough Transform (PHT) and FPGA architecture-associated framework for real-time straight line detection in high-definition videos. A resource-optimized Canny edge detection method with enhanced non-maximum suppression conditions is presented to suppress most possible false edges and obtain more accurate candidate edge pixels for subsequent accelerated computation. Then, a novel PHT algorithm exploiting spatial angle-level parallelism is proposed to upgrade computational accuracy by improving the minimum computational step. Moreover, the FPGA based multi-level pipelined PHT architecture optimized by spatial parallelism ensures real-time computation for 1,024 × 768 resolution videos without any off-chip memory consumption. This framework is evaluated on ALTERA DE2-115 FPGA evaluation platform at a maximum frequency of 200 MHz, and it can calculate straight line parameters in 15.59 ms on the average for one frame. Qualitative and quantitative evaluation results have validated the system performance regarding data throughput, memory bandwidth, resource, speed and robustness.

  1. Parallel superconducting strip-line detectors: reset behaviour in the single-strip switch regime

    Casaburi, A; Heath, R M; Tanner, M G; Hadfield, R H; Cristiano, R; Ejrnaes, M; Nappi, C

    2014-01-01

    Superconducting strip-line detectors (SSLDs) are an important emerging technology for the detection of single molecules in time-of-flight mass spectrometry (TOF-MS). We present an experimental investigation of a SSLD laid out in a parallel configuration, designed to address selected single strip-lines operating in the single-strip switch regime. Fast laser pulses were tightly focused onto the device, allowing controllable nucleation of a resistive region at a specific location and study of the subsequent device response dynamics. We observed that in this regime, although the strip-line returns to the superconducting state after triggering, no effective recovery of the bias current occurs, in qualitative agreement with a phenomenological circuit simulation that we performed. Moreover, from theoretical considerations and by looking at the experimental pulse amplitude distribution histogram, we have the first confirmation of the fact that the phenomenological London model governs the current redistribution in these large area devices also after detection events. (paper)

  2. Line-plane broadcasting in a data communications network of a parallel computer

    Archer, Charles J.; Berg, Jeremy E.; Blocksome, Michael A.; Smith, Brian E.

    2010-06-08

    Methods, apparatus, and products are disclosed for line-plane broadcasting in a data communications network of a parallel computer, the parallel computer comprising a plurality of compute nodes connected together through the network, the network optimized for point to point data communications and characterized by at least a first dimension, a second dimension, and a third dimension, that include: initiating, by a broadcasting compute node, a broadcast operation, including sending a message to all of the compute nodes along an axis of the first dimension for the network; sending, by each compute node along the axis of the first dimension, the message to all of the compute nodes along an axis of the second dimension for the network; and sending, by each compute node along the axis of the second dimension, the message to all of the compute nodes along an axis of the third dimension for the network.

  3. Data Parallel Line Relaxation (DPLR) Code User Manual: Acadia - Version 4.01.1

    Wright, Michael J.; White, Todd; Mangini, Nancy

    2009-01-01

    Data-Parallel Line Relaxation (DPLR) code is a computational fluid dynamic (CFD) solver that was developed at NASA Ames Research Center to help mission support teams generate high-value predictive solutions for hypersonic flow field problems. The DPLR Code Package is an MPI-based, parallel, full three-dimensional Navier-Stokes CFD solver with generalized models for finite-rate reaction kinetics, thermal and chemical non-equilibrium, accurate high-temperature transport coefficients, and ionized flow physics incorporated into the code. DPLR also includes a large selection of generalized realistic surface boundary conditions and links to enable loose coupling with external thermal protection system (TPS) material response and shock layer radiation codes.

  4. Wideband Dual-Polarization Patch Antenna Array With Parallel Strip Line Balun Feeding

    Zhang, Jin; Lin, Xianqi; Nie, Liying

    2016-01-01

    A wideband dual-polarization patch antenna array is proposed in this letter. The array is fed by a parallel strip line balun, which is adopted to generate 180° phase shift in a wide frequency range. In addition, this balun has simple structure, very small phase shift error, and good ports isolati...... is higher than 30 dB. The simulation and measurement turns out to be similar. This antenna array can be used in TD-LTE base stations, and the design methods are also useful to other wideband microstrip antennas....

  5. Airport object extraction based on visual attention mechanism and parallel line detection

    Lv, Jing; Lv, Wen; Zhang, Libao

    2017-10-01

    Target extraction is one of the important aspects in remote sensing image analysis and processing, which has wide applications in images compression, target tracking, target recognition and change detection. Among different targets, airport has attracted more and more attention due to its significance in military and civilian. In this paper, we propose a novel and reliable airport object extraction model combining visual attention mechanism and parallel line detection algorithm. First, a novel saliency analysis model for remote sensing images with airport region is proposed to complete statistical saliency feature analysis. The proposed model can precisely extract the most salient region and preferably suppress the background interference. Then, the prior geometric knowledge is analyzed and airport runways contained two parallel lines with similar length are detected efficiently. Finally, we use the improved Otsu threshold segmentation method to segment and extract the airport regions from the salient map of remote sensing images. The experimental results demonstrate that the proposed model outperforms existing saliency analysis models and shows good performance in the detection of the airport.

  6. A study of parallelism of the occlusal plane and ala-tragus line.

    Sadr, Katayoun; Sadr, Makan

    2009-01-01

    Orientation of the occlusal plane is one of the most important clinical procedures in prostho-dontic rehabilitation of edentulous patients. The aim of this study was to define the best posterior reference point of ala-tragus line for orientation of occlusal plane for complete denture fabrication. Fifty-three dental students (27 females and 26 males) with complete natural dentition and Angel's Class I occlusal relationship were selected. The subjects were photographed in natural head position while clenching on a Fox plane. After tracing the photographs, the angles between the following lines were measured: the occlusal plane (Fox plane) and the superior border of ala-tragus, the occlusal plane (Fox plane) and the middle of ala-tragus as well as the occlusal plane (Fox plane) and the inferior border of ala-tragus. Descriptive statistics, one sample t-test and independent t-test were used. P value less than 0.05 was considered significant. There was no parallelism between the occlusal plane and ala-tragus line with three different posterior ends and one sample t-test showed that the angles between them were significantly different from zero (pplane. The superior border of the tragus is suggested as the posterior reference for ala-tragus line.

  7. Parallel Evolution under Chemotherapy Pressure in 29 Breast Cancer Cell Lines Results in Dissimilar Mechanisms of Resistance

    Tegze, Balint; Szallasi, Zoltan Imre; Haltrich, Iren

    2012-01-01

    Background: Developing chemotherapy resistant cell lines can help to identify markers of resistance. Instead of using a panel of highly heterogeneous cell lines, we assumed that truly robust and convergent pattern of resistance can be identified in multiple parallel engineered derivatives of only...

  8. Establishing Substantial Equivalence: Transcriptomics

    Baudo, María Marcela; Powers, Stephen J.; Mitchell, Rowan A. C.; Shewry, Peter R.

    Regulatory authorities in Western Europe require transgenic crops to be substantially equivalent to conventionally bred forms if they are to be approved for commercial production. One way to establish substantial equivalence is to compare the transcript profiles of developing grain and other tissues of transgenic and conventionally bred lines, in order to identify any unintended effects of the transformation process. We present detailed protocols for transcriptomic comparisons of developing wheat grain and leaf material, and illustrate their use by reference to our own studies of lines transformed to express additional gluten protein genes controlled by their own endosperm-specific promoters. The results show that the transgenes present in these lines (which included those encoding marker genes) did not have any significant unpredicted effects on the expression of endogenous genes and that the transgenic plants were therefore substantially equivalent to the corresponding parental lines.

  9. Multi-objective optimization algorithms for mixed model assembly line balancing problem with parallel workstations

    Masoud Rabbani

    2016-12-01

    Full Text Available This paper deals with mixed model assembly line (MMAL balancing problem of type-I. In MMALs several products are made on an assembly line while the similarity of these products is so high. As a result, it is possible to assemble several types of products simultaneously without any additional setup times. The problem has some particular features such as parallel workstations and precedence constraints in dynamic periods in which each period also effects on its next period. The research intends to reduce the number of workstations and maximize the workload smoothness between workstations. Dynamic periods are used to determine all variables in different periods to achieve efficient solutions. A non-dominated sorting genetic algorithm (NSGA-II and multi-objective particle swarm optimization (MOPSO are used to solve the problem. The proposed model is validated with GAMS software for small size problem and the performance of the foregoing algorithms is compared with each other based on some comparison metrics. The NSGA-II outperforms MOPSO with respect to some comparison metrics used in this paper, but in other metrics MOPSO is better than NSGA-II. Finally, conclusion and future research is provided.

  10. Parallel evolution under chemotherapy pressure in 29 breast cancer cell lines results in dissimilar mechanisms of resistance.

    Bálint Tegze

    Full Text Available BACKGROUND: Developing chemotherapy resistant cell lines can help to identify markers of resistance. Instead of using a panel of highly heterogeneous cell lines, we assumed that truly robust and convergent pattern of resistance can be identified in multiple parallel engineered derivatives of only a few parental cell lines. METHODS: Parallel cell populations were initiated for two breast cancer cell lines (MDA-MB-231 and MCF-7 and these were treated independently for 18 months with doxorubicin or paclitaxel. IC50 values against 4 chemotherapy agents were determined to measure cross-resistance. Chromosomal instability and karyotypic changes were determined by cytogenetics. TaqMan RT-PCR measurements were performed for resistance-candidate genes. Pgp activity was measured by FACS. RESULTS: All together 16 doxorubicin- and 13 paclitaxel-treated cell lines were developed showing 2-46 fold and 3-28 fold increase in resistance, respectively. The RT-PCR and FACS analyses confirmed changes in tubulin isofom composition, TOP2A and MVP expression and activity of transport pumps (ABCB1, ABCG2. Cytogenetics showed less chromosomes but more structural aberrations in the resistant cells. CONCLUSION: We surpassed previous studies by parallel developing a massive number of cell lines to investigate chemoresistance. While the heterogeneity caused evolution of multiple resistant clones with different resistance characteristics, the activation of only a few mechanisms were sufficient in one cell line to achieve resistance.

  11. Characterization of DNA repair phenotypes of Xeroderma pigmentosum cell lines by a paralleled in vitro test

    Raffin, A.L.

    2009-06-01

    DNA is constantly damaged modifying the genetic information for which it encodes. Several cellular mechanisms as the Base Excision Repair (BER) and the Nucleotide Excision Repair (NER) allow recovering the right DNA sequence. The Xeroderma pigmentosum is a disease characterised by a deficiency in the NER pathway. The aim of this study was to propose an efficient and fast test for the diagnosis of this disease as an alternative to the currently available UDS test. DNA repair activities of XP cell lines were quantified using in vitro miniaturized and paralleled tests in order to establish DNA repair phenotypes of XPA and XPC deficient cells. The main advantage of the tests used in this study is the simultaneous measurement of excision or excision synthesis (ES) of several lesions by only one cellular extract. We showed on one hand that the relative ES of the different lesions depend strongly on the protein concentration of the nuclear extract tested. Working at high protein concentration allowed discriminating the XP phenotype versus the control one, whereas it was impossible under a certain concentration's threshold. On the other hand, while the UVB irradiation of control cells stimulated their repair activities, this effect was not observed in XP cells. This study brings new information on the XPA and XPC protein roles during BER and NER and underlines the complexity of the regulations of DNA repair processes. (author)

  12. Slit shaped microwave induced atmospheric pressure plasma based on a parallel plate transmission line resonator

    Kang, S. K.; Seo, Y. S.; Lee, H. Wk; Aman-ur-Rehman; Kim, G. C.; Lee, J. K.

    2011-11-01

    A new type of microwave-excited atmospheric pressure plasma source, based on the principle of parallel plate transmission line resonator, is developed for the treatment of large areas in biomedical applications such as skin treatment and wound healing. A stable plasma of 20 mm width is sustained by a small microwave power source operated at a frequency of 700 MHz and a gas flow rate of 0.9 slm. Plasma impedance and plasma density of this plasma source are estimated by fitting the calculated reflection coefficient to the measured one. The estimated plasma impedance shows a decreasing trend while estimated plasma density shows an increasing trend with the increase in the input power. Plasma uniformity is confirmed by temperature and optical emission distribution measurements. Plasma temperature is sustained at less than 40 °C and abundant amounts of reactive species, which are important agents for bacteria inactivation, are detected over the entire plasma region. Large area treatment ability of this newly developed device is verified through bacteria inactivation experiment using E. coli. Sterilization experiment shows a large bacterial killing mark of 25 mm for a plasma treatment time of 10 s.

  13. Design of a chemical batch plant : a study of dedicated parallel lines with intermediate storage and the plant performance

    Verbiest, Floor; Cornelissens, Trijntje; Springael, Johan

    2016-01-01

    Abstract: Production plants worldwide face huge challenges in satisfying high service levels and outperforming competition. These challenges require appropriate strategic decisions on plant design and production strategies. In this paper, we focus on multiproduct chemical batch plants, which are typically equipped with multiple production lines and intermediate storage tanks. First we extend the existing MI(N) LP design models with the concept of parallel production lines, and optimise the as...

  14. Substantially parallel flux uncluttered rotor machines

    Hsu, John S.

    2012-12-11

    A permanent magnet-less and brushless synchronous system includes a stator that generates a magnetic rotating field when sourced by polyphase alternating currents. An uncluttered rotor is positioned within the magnetic rotating field and is spaced apart from the stator. An excitation core is spaced apart from the stator and the uncluttered rotor and magnetically couples the uncluttered rotor. The brushless excitation source generates a magnet torque by inducing magnetic poles near an outer peripheral surface of the uncluttered rotor, and the stator currents also generate a reluctance torque by a reaction of the difference between the direct and quadrature magnetic paths of the uncluttered rotor. The system can be used either as a motor or a generator

  15. Advanced mathematical on-line analysis in nuclear experiments. Usage of parallel computing CUDA routines in standard root analysis

    Grzeszczuk, A.; Kowalski, S.

    2015-04-01

    Compute Unified Device Architecture (CUDA) is a parallel computing platform developed by Nvidia for increase speed of graphics by usage of parallel mode for processes calculation. The success of this solution has opened technology General-Purpose Graphic Processor Units (GPGPUs) for applications not coupled with graphics. The GPGPUs system can be applying as effective tool for reducing huge number of data for pulse shape analysis measures, by on-line recalculation or by very quick system of compression. The simplified structure of CUDA system and model of programming based on example Nvidia GForce GTX580 card are presented by our poster contribution in stand-alone version and as ROOT application.

  16. Real-Time Straight-Line Detection for XGA-Size Videos by Hough Transform with Parallelized Voting Procedures.

    Guan, Jungang; An, Fengwei; Zhang, Xiangyu; Chen, Lei; Mattausch, Hans Jürgen

    2017-01-30

    The Hough Transform (HT) is a method for extracting straight lines from an edge image. The main limitations of the HT for usage in actual applications are computation time and storage requirements. This paper reports a hardware architecture for HT implementation on a Field Programmable Gate Array (FPGA) with parallelized voting procedure. The 2-dimensional accumulator array, namely the Hough space in parametric form (ρ, θ), for computing the strength of each line by a voting mechanism is mapped on a 1-dimensional array with regular increments of θ. Then, this Hough space is divided into a number of parallel parts. The computation of (ρ, θ) for the edge pixels and the voting procedure for straight-line determination are therefore executable in parallel. In addition, a synchronized initialization for the Hough space further increases the speed of straight-line detection, so that XGA video processing becomes possible. The designed prototype system has been synthesized on a DE4 platform with a Stratix-IV FPGA device. In the application of road-lane detection, the average processing speed of this HT implementation is 5.4ms per XGA-frame at 200 MHz working frequency.

  17. Parallelism at Cern: real-time and off-line applications in the GP-MIMD2 project

    Calafiura, P.

    1997-01-01

    A wide range of general purpose high-energy physics applications, ranging from Monte Carlo simulation to data acquisition, from interactive data analysis to on-line filtering, have been ported, or developed, and run in parallel on IBM SP-2 and Meiko CS-2 CERN large multi-processor machines. The ESPRIT project GP-MIMD2 has been a catalyst for the interest in parallel computing at CERN. The project provided the 128 processor Meiko CS-2 system that is now succesfully integrated in the CERN computing environment. The CERN experiment NA48 was involved in the GP-MIMD2 project since the beginning. NA48 physicists run, as part of their day-to-day work, simulation and analysis programs parallelized using the message passing interface MPI. The CS-2 is also a vital component of the experiment data acquisition system and will be used to calibrate in real-time the 13000 channels liquid krypton calorimeter. (orig.)

  18. Attempt to identify the functional areas of the cerebral cortex on CT slices parallel to the orbito-meatal line

    Tanabe, Hirotaka; Okuda, Junichiro; Nishikawa, Takashi; Nishimura, Tsuyoshi (Osaka Univ. (Japan). Faculty of Medicine); Shiraishi, Junzo

    1982-06-01

    In order to identify the functional brain areas, such as Broca's area, on computed tomography slices parallel to the orbito-meatal line, the numbers of Brodmann's cortical mapping were shown on a diagram of representative brain sections parallel to the orbito-meatal line. Also, we described a method, using cerebral sulci as anatomical landmarks, for projecting lesions shown by CT scan onto the lateral brain diagram. The procedures were as follows. The distribution of lesions on CT slices was determined by the identification of major cerebral sulci and fissures, such as the Sylvian fissure, the central sulcus, and the superior frontal sulcus. Those lesions were then projected onto the lateral diagram by comparing each CT slice with the horizontal diagrams of brain sections. The method was demonstrated in three cases developing neuropsychological symptoms.

  19. 600 GHz resonant mode in a parallel array of Josephson tunnel junctions connected by superconducting microstrip lines

    Kaplunenko, V. K.; Larsen, Britt Hvolbæk; Mygind, Jesper

    1994-01-01

    on experimental and numerical investigations of a resonant step observed at a voltage corresponding to 600 GHz in the dc current-voltage characteristic of a parallel array of 20 identical small NbAl2O3Nb Josephson junctions interconnected by short sections of superconducting microstrip line. The junctions...... are mutually phase locked due to collective interaction with the line sections excited close to the half wavelength resonance. The phase locking range can be adjusted by means of an external dc magnetic field and the step size varies periodically with the magnetic field. The largest step corresponds...

  20. Advanced mathematical on-line analysis in nuclear experiments. Usage of parallel computing CUDA routines in standard root analysis

    Grzeszczuk A.

    2015-01-01

    Full Text Available Compute Unified Device Architecture (CUDA is a parallel computing platform developed by Nvidia for increase speed of graphics by usage of parallel mode for processes calculation. The success of this solution has opened technology General-Purpose Graphic Processor Units (GPGPUs for applications not coupled with graphics. The GPGPUs system can be applying as effective tool for reducing huge number of data for pulse shape analysis measures, by on-line recalculation or by very quick system of compression. The simplified structure of CUDA system and model of programming based on example Nvidia GForce GTX580 card are presented by our poster contribution in stand-alone version and as ROOT application.

  1. Parallel mRNA, proteomics and miRNA expression analysis in cell line models of the intestine.

    O'Sullivan, Finbarr; Keenan, Joanne; Aherne, Sinead; O'Neill, Fiona; Clarke, Colin; Henry, Michael; Meleady, Paula; Breen, Laura; Barron, Niall; Clynes, Martin; Horgan, Karina; Doolan, Padraig; Murphy, Richard

    2017-11-07

    To identify miRNA-regulated proteins differentially expressed between Caco2 and HT-29: two principal cell line models of the intestine. Exponentially growing Caco-2 and HT-29 cells were harvested and prepared for mRNA, miRNA and proteomic profiling. mRNA microarray profiling analysis was carried out using the Affymetrix GeneChip Human Gene 1.0 ST array. miRNA microarray profiling analysis was carried out using the Affymetrix Genechip miRNA 3.0 array. Quantitative Label-free LC-MS/MS proteomic analysis was performed using a Dionex Ultimate 3000 RSLCnano system coupled to a hybrid linear ion trap/Orbitrap mass spectrometer. Peptide identities were validated in Proteome Discoverer 2.1 and were subsequently imported into Progenesis QI software for further analysis. Hierarchical cluster analysis for all three parallel datasets (miRNA, proteomics, mRNA) was conducted in the R software environment using the Euclidean distance measure and Ward's clustering algorithm. The prediction of miRNA and oppositely correlated protein/mRNA interactions was performed using TargetScan 6.1. GO biological process, molecular function and cellular component enrichment analysis was carried out for the DE miRNA, protein and mRNA lists via the Pathway Studio 11.3 Web interface using their Mammalian database. Differential expression (DE) profiling comparing the intestinal cell lines HT-29 and Caco-2 identified 1795 Genes, 168 Proteins and 160 miRNAs as DE between the two cell lines. At the gene level, 1084 genes were upregulated and 711 were downregulated in the Caco-2 cell line relative to the HT-29 cell line. At the protein level, 57 proteins were found to be upregulated and 111 downregulated in the Caco-2 cell line relative to the HT-29 cell line. Finally, at the miRNAs level, 104 were upregulated and 56 downregulated in the Caco-2 cell line relative to the HT-29 cell line. Gene ontology (GO) analysis of the DE mRNA identified cell adhesion, migration and ECM organization, cellular lipid

  2. Operating system design of parallel computer for on-line management of nuclear pressurised water reactor cores

    Gougam, F.

    1991-04-01

    This study is part of the PHAETON project which aims at increasing the knowledge of safety parameters of PWR core and reducing operating margins during the reactor cycle. The on-line system associates a simulator process to compute the three dimensional flux distribution and an acquisition process of reactor core parameters from the central instrumentation. The 3D flux calculation is the most time consuming. So, for cost and safety reasons, the PHAETON project proposes an approach which is to parallelize the 3D diffusion calculation and to use a computer based on parallel processor architecture. This paper presents the design of the operating system on which the application is executed. The routine interface proposed, includes the main operations necessary for programming a real time and parallel application. The primitives include: task management, data transfer, synchronisation by event signalling and by using the rendez-vous mechanisms. The primitives which are proposed use standard softwares like real-time kernel and UNIX operating system [fr

  3. A New Track Reconstruction Algorithm suitable for Parallel Processing based on Hit Triplets and Broken Lines

    Schöning André

    2016-01-01

    Full Text Available Track reconstruction in high track multiplicity environments at current and future high rate particle physics experiments is a big challenge and very time consuming. The search for track seeds and the fitting of track candidates are usually the most time consuming steps in the track reconstruction. Here, a new and fast track reconstruction method based on hit triplets is proposed which exploits a three-dimensional fit model including multiple scattering and hit uncertainties from the very start, including the search for track seeds. The hit triplet based reconstruction method assumes a homogeneous magnetic field which allows to give an analytical solutions for the triplet fit result. This method is highly parallelizable, needs fewer operations than other standard track reconstruction methods and is therefore ideal for the implementation on parallel computing architectures. The proposed track reconstruction algorithm has been studied in the context of the Mu3e-experiment and a typical LHC experiment.

  4. Geometrical reasoning in the primary school, the case of parallel lines

    Sinclair, Nathalie; Jones, Keith

    2009-01-01

    During the primary school years, children are typically expected to develop ways of explaining their mathematical reasoning. This paper reports on ideas developed during an analysis of data from a project which involved young children (aged 5-7 years old) in a whole-class situation using dynamic geometry software (specifically Sketchpad). The focus is a classroom episode in which the children try to decide whether two lines that they know continue (but cannot see all of the continuation) will...

  5. VERY STRONG EMISSION-LINE GALAXIES IN THE WFC3 INFRARED SPECTROSCOPIC PARALLEL SURVEY AND IMPLICATIONS FOR HIGH-REDSHIFT GALAXIES

    Atek, H.; Colbert, J.; Shim, H. [Spitzer Science Center, Caltech, Pasadena, CA 91125 (United States); Siana, B.; Bridge, C. [Department of Astronomy, Caltech, Pasadena, CA 91125 (United States); Scarlata, C. [Department of Astronomy, University of Minnesota-Twin Cities, Minneapolis, MN 55455 (United States); Malkan, M.; Ross, N. R. [Department of Physics and Astronomy, University of California, Los Angeles, CA (United States); McCarthy, P.; Dressler, A.; Hathi, N. P. [Observatories of the Carnegie Institution for Science, Pasadena, CA 91101 (United States); Teplitz, H. [Infrared Processing and Analysis Center, Caltech, Pasadena, CA 91125 (United States); Henry, A.; Martin, C. [Department of Physics, University of California, Santa Barbara, CA 93106 (United States); Bunker, A. J. [Department of Physics, University of Oxford, Denys Wilkinson Building, Keble Road, Oxford OX1 3RH (United Kingdom); Fosbury, R. A. E. [Space Telescope-European Coordinating Facility, Garching bei Muenchen (Germany)

    2011-12-20

    The WFC3 Infrared Spectroscopic Parallel Survey uses the Hubble Space Telescope (HST) infrared grism capabilities to obtain slitless spectra of thousands of galaxies over a wide redshift range including the peak of star formation history of the universe. We select a population of very strong emission-line galaxies with rest-frame equivalent widths (EWs) higher than 200 A. A total of 176 objects are found over the redshift range 0.35 < z < 2.3 in the 180 arcmin{sup 2} area that we have analyzed so far. This population consists of young and low-mass starbursts with high specific star formation rates (sSFR). After spectroscopic follow-up of one of these galaxies with Keck/Low Resolution Imaging Spectrometer, we report the detection at z = 0.7 of an extremely metal-poor galaxy with 12 + log(O/H) =7.47 {+-} 0.11. After estimating the active galactic nucleus fraction in the sample, we show that the high-EW galaxies have higher sSFR than normal star-forming galaxies at any redshift. We find that the nebular emission lines can substantially affect the total broadband flux density with a median brightening of 0.3 mag, with some examples of line contamination producing brightening of up to 1 mag. We show that the presence of strong emission lines in low-z galaxies can mimic the color-selection criteria used in the z {approx} 8 dropout surveys. In order to effectively remove low-redshift interlopers, deep optical imaging is needed, at least 1 mag deeper than the bands in which the objects are detected. Without deep optical data, most of the interlopers cannot be ruled out in the wide shallow HST imaging surveys. Finally, we empirically demonstrate that strong nebular lines can lead to an overestimation of the mass and the age of galaxies derived from fitting of their spectral energy distribution (SED). Without removing emission lines, the age and the stellar mass estimates are overestimated by a factor of 2 on average and up to a factor of 10 for the high-EW galaxies

  6. Establishing Substantial Equivalence: Proteomics

    Lovegrove, Alison; Salt, Louise; Shewry, Peter R.

    Wheat is a major crop in world agriculture and is consumed after processing into a range of food products. It is therefore of great importance to determine the consequences (intended and unintended) of transgenesis in wheat and whether genetically modified lines are substantially equivalent to those produced by conventional plant breeding. Proteomic analysis is one of several approaches which can be used to address these questions. Two-dimensional PAGE (2D PAGE) remains the most widely available method for proteomic analysis, but is notoriously difficult to reproduce between laboratories. We therefore describe methods which have been developed as standard operating procedures in our laboratory to ensure the reproducibility of proteomic analyses of wheat using 2D PAGE analysis of grain proteins.

  7. High-voltage isolation transformer for sub-nanosecond rise time pulses constructed with annular parallel-strip transmission lines.

    Homma, Akira

    2011-07-01

    A novel annular parallel-strip transmission line was devised to construct high-voltage high-speed pulse isolation transformers. The transmission lines can easily realize stable high-voltage operation and good impedance matching between primary and secondary circuits. The time constant for the step response of the transformer was calculated by introducing a simple low-frequency equivalent circuit model. Results show that the relation between the time constant and low-cut-off frequency of the transformer conforms to the theory of the general first-order linear time-invariant system. Results also show that the test transformer composed of the new transmission lines can transmit about 600 ps rise time pulses across the dc potential difference of more than 150 kV with insertion loss of -2.5 dB. The measured effective time constant of 12 ns agreed exactly with the theoretically predicted value. For practical applications involving the delivery of synchronized trigger signals to a dc high-voltage electron gun station, the transformer described in this paper exhibited advantages over methods using fiber optic cables for the signal transfer system. This transformer has no jitter or breakdown problems that invariably occur in active circuit components.

  8. A germ cell determinant reveals parallel pathways for germ line development in Caenorhabditis elegans.

    Mainpal, Rana; Nance, Jeremy; Yanowitz, Judith L

    2015-10-15

    Despite the central importance of germ cells for transmission of genetic material, our understanding of the molecular programs that control primordial germ cell (PGC) specification and differentiation are limited. Here, we present findings that X chromosome NonDisjunction factor-1 (XND-1), known for its role in regulating meiotic crossover formation, is an early determinant of germ cell fates in Caenorhabditis elegans. xnd-1 mutant embryos display a novel 'one PGC' phenotype as a result of G2 cell cycle arrest of the P4 blastomere. Larvae and adults display smaller germ lines and reduced brood size consistent with a role for XND-1 in germ cell proliferation. Maternal XND-1 proteins are found in the P4 lineage and are exclusively localized to the nucleus in PGCs, Z2 and Z3. Zygotic XND-1 turns on shortly thereafter, at the ∼300-cell stage, making XND-1 the earliest zygotically expressed gene in worm PGCs. Strikingly, a subset of xnd-1 mutants lack germ cells, a phenotype shared with nos-2, a member of the conserved Nanos family of germline determinants. We generated a nos-2 null allele and show that nos-2; xnd-1 double mutants display synthetic sterility. Further removal of nos-1 leads to almost complete sterility, with the vast majority of animals without germ cells. Sterility in xnd-1 mutants is correlated with an increase in transcriptional activation-associated histone modification and aberrant expression of somatic transgenes. Together, these data strongly suggest that xnd-1 defines a new branch for PGC development that functions redundantly with nos-2 and nos-1 to promote germline fates by maintaining transcriptional quiescence and regulating germ cell proliferation. © 2015. Published by The Company of Biologists Ltd.

  9. Micromachined silicon parallel acoustic delay lines as time-delayed ultrasound detector array for real-time photoacoustic tomography

    Cho, Y.; Chang, C.-C.; Wang, L. V.; Zou, J.

    2016-02-01

    This paper reports the development of a new 16-channel parallel acoustic delay line (PADL) array for real-time photoacoustic tomography (PAT). The PADLs were directly fabricated from single-crystalline silicon substrates using deep reactive ion etching. Compared with other acoustic delay lines (e.g., optical fibers), the micromachined silicon PADLs offer higher acoustic transmission efficiency, smaller form factor, easier assembly, and mass production capability. To demonstrate its real-time photoacoustic imaging capability, the silicon PADL array was interfaced with one single-element ultrasonic transducer followed by one channel of data acquisition electronics to receive 16 channels of photoacoustic signals simultaneously. A PAT image of an optically-absorbing target embedded in an optically-scattering phantom was reconstructed, which matched well with the actual size of the imaged target. Because the silicon PADL array allows a signal-to-channel reduction ratio of 16:1, it could significantly simplify the design and construction of ultrasonic receivers for real-time PAT.

  10. Micromachined silicon parallel acoustic delay lines as time-delayed ultrasound detector array for real-time photoacoustic tomography

    Cho, Y; Chang, C-C; Zou, J; Wang, L V

    2016-01-01

    This paper reports the development of a new 16-channel parallel acoustic delay line (PADL) array for real-time photoacoustic tomography (PAT). The PADLs were directly fabricated from single-crystalline silicon substrates using deep reactive ion etching. Compared with other acoustic delay lines (e.g., optical fibers), the micromachined silicon PADLs offer higher acoustic transmission efficiency, smaller form factor, easier assembly, and mass production capability. To demonstrate its real-time photoacoustic imaging capability, the silicon PADL array was interfaced with one single-element ultrasonic transducer followed by one channel of data acquisition electronics to receive 16 channels of photoacoustic signals simultaneously. A PAT image of an optically-absorbing target embedded in an optically-scattering phantom was reconstructed, which matched well with the actual size of the imaged target. Because the silicon PADL array allows a signal-to-channel reduction ratio of 16:1, it could significantly simplify the design and construction of ultrasonic receivers for real-time PAT. (paper)

  11. An attempt to identify the functional areas of the cerebral cortex on CT slices parallel to the orbito-meatal line

    Tanabe, Hirotaka; Okuda, Junichiro; Nishikawa, Takashi; Nishimura, Tsuyoshi; Shiraishi, Junzo.

    1982-01-01

    In order to identify the functional brain areas, such as Broca's area, on computed tomography slices parallel to the orbito-meatal line, the numbers of Brodmann's cortical mapping were shown on a diagram of representative brain sections parallel to the orbito-meatal line. Also, we described a method, using cerebral sulci as anatomical landmarks, for projecting lesions shown by CT scan onto the lateral brain diagram. The procedures were as follows. The distribution of lesions on CT slices was determined by the identification of major cerebral sulci and fissures, such as the Sylvian fissure, the central sulcus, and the superior frontal sulcus. Those lesions were then projected onto the lateral diagram by comparing each CT slice with the horizontal diagrams of brain sections. The method was demonstrated in three cases developing neuropsychological symptoms. (author)

  12. GPU-based, parallel-line, omni-directional integration of measured acceleration field to obtain the 3D pressure distribution

    Wang, Jin; Zhang, Cao; Katz, Joseph

    2016-11-01

    A PIV based method to reconstruct the volumetric pressure field by direct integration of the 3D material acceleration directions has been developed. Extending the 2D virtual-boundary omni-directional method (Omni2D, Liu & Katz, 2013), the new 3D parallel-line omni-directional method (Omni3D) integrates the material acceleration along parallel lines aligned in multiple directions. Their angles are set by a spherical virtual grid. The integration is parallelized on a Tesla K40c GPU, which reduced the computing time from three hours to one minute for a single realization. To validate its performance, this method is utilized to calculate the 3D pressure fields in isotropic turbulence and channel flow using the JHU DNS Databases (http://turbulence.pha.jhu.edu). Both integration of the DNS acceleration as well as acceleration from synthetic 3D particles are tested. Results are compared to other method, e.g. solution to the Pressure Poisson Equation (e.g. PPE, Ghaemi et al., 2012) with Bernoulli based Dirichlet boundary conditions, and the Omni2D method. The error in Omni3D prediction is uniformly low, and its sensitivity to acceleration errors is local. It agrees with the PPE/Bernoulli prediction away from the Dirichlet boundary. The Omni3D method is also applied to experimental data obtained using tomographic PIV, and results are correlated with deformation of a compliant wall. ONR.

  13. MOEA based design of decentralized controllers for LFC of interconnected power systems with nonlinearities, AC-DC parallel tie-lines and SMES units

    Ganapathy, S.; Velusami, S.

    2010-01-01

    A new design of Multi-Objective Evolutionary Algorithm based decentralized controllers for load-frequency control of interconnected power systems with Governor Dead Band and Generation Rate Constraint nonlinearities, AC-DC parallel tie-lines and Superconducting Magnetic Energy Storage (SMES) units, is proposed in this paper. The HVDC link is used as system interconnection in parallel with AC tie-line to effectively damp the frequency oscillations of AC system while the SMES unit provides bulk energy storage and release, thereby achieving combined benefits. The proposed controller satisfies two main objectives, namely, minimum Integral Squared Error of the system output and maximum closed-loop stability of the system. Simulation studies are conducted on a two area interconnected power system with nonlinearities, AC-DC tie-lines and SMES units. Results indicate that the proposed controller improves the transient responses and guarantees the closed-loop stability of the overall system even in the presence of system nonlinearities and with parameter changes.

  14. Similarities and differences between helminth parasites and cancer cell lines in shaping human monocytes: Insights into parallel mechanisms of immune evasion.

    Prakash Babu Narasimhan

    2018-04-01

    Full Text Available A number of features at the host-parasite interface are reminiscent of those that are also observed at the host-tumor interface. Both cancer cells and parasites establish a tissue microenvironment that allows for immune evasion and may reflect functional alterations of various innate cells. Here, we investigated how the phenotype and function of human monocytes is altered by exposure to cancer cell lines and if these functional and phenotypic alterations parallel those induced by exposure to helminth parasites. Thus, human monocytes were exposed to three different cancer cell lines (breast, ovarian, or glioblastoma or to live microfilariae (mf of Brugia malayi-a causative agent of lymphatic filariasis. After 2 days of co-culture, monocytes exposed to cancer cell lines showed markedly upregulated expression of M1-associated (TNF-α, IL-1β, M2-associated (CCL13, CD206, Mreg-associated (IL-10, TGF-β, and angiogenesis associated (MMP9, VEGF genes. Similar to cancer cell lines, but less dramatically, mf altered the mRNA expression of IL-1β, CCL13, TGM2 and MMP9. When surface expression of the inhibitory ligands PDL1 and PDL2 was assessed, monocytes exposed to both cancer cell lines and to live mf significantly upregulated PDL1 and PDL2 expression. In contrast to exposure to mf, exposure to cancer cell lines increased the phagocytic ability of monocytes and reduced their ability to induce T cell proliferation and to expand Granzyme A+ CD8+ T cells. Our data suggest that despite the fact that helminth parasites and cancer cell lines are extraordinarily disparate, they share the ability to alter the phenotype of human monocytes.

  15. Similarities and differences between helminth parasites and cancer cell lines in shaping human monocytes: Insights into parallel mechanisms of immune evasion.

    Narasimhan, Prakash Babu; Akabas, Leor; Tariq, Sameha; Huda, Naureen; Bennuru, Sasisekhar; Sabzevari, Helen; Hofmeister, Robert; Nutman, Thomas B; Tolouei Semnani, Roshanak

    2018-04-01

    A number of features at the host-parasite interface are reminiscent of those that are also observed at the host-tumor interface. Both cancer cells and parasites establish a tissue microenvironment that allows for immune evasion and may reflect functional alterations of various innate cells. Here, we investigated how the phenotype and function of human monocytes is altered by exposure to cancer cell lines and if these functional and phenotypic alterations parallel those induced by exposure to helminth parasites. Thus, human monocytes were exposed to three different cancer cell lines (breast, ovarian, or glioblastoma) or to live microfilariae (mf) of Brugia malayi-a causative agent of lymphatic filariasis. After 2 days of co-culture, monocytes exposed to cancer cell lines showed markedly upregulated expression of M1-associated (TNF-α, IL-1β), M2-associated (CCL13, CD206), Mreg-associated (IL-10, TGF-β), and angiogenesis associated (MMP9, VEGF) genes. Similar to cancer cell lines, but less dramatically, mf altered the mRNA expression of IL-1β, CCL13, TGM2 and MMP9. When surface expression of the inhibitory ligands PDL1 and PDL2 was assessed, monocytes exposed to both cancer cell lines and to live mf significantly upregulated PDL1 and PDL2 expression. In contrast to exposure to mf, exposure to cancer cell lines increased the phagocytic ability of monocytes and reduced their ability to induce T cell proliferation and to expand Granzyme A+ CD8+ T cells. Our data suggest that despite the fact that helminth parasites and cancer cell lines are extraordinarily disparate, they share the ability to alter the phenotype of human monocytes.

  16. On-line electrochemistry-bioaffinity screening with parallel HR-LC-MS for the generation and characterization of modified p38α kinase inhibitors.

    Falck, David; de Vlieger, Jon S B; Giera, Martin; Honing, Maarten; Irth, Hubertus; Niessen, Wilfried M A; Kool, Jeroen

    2012-04-01

    In this study, an integrated approach is developed for the formation, identification and biological characterization of electrochemical conversion products of p38α mitogen-activated protein kinase inhibitors. This work demonstrates the hyphenation of an electrochemical reaction cell with a continuous-flow bioaffinity assay and parallel LC-HR-MS. Competition of the formed products with a tracer (SKF-86002) that shows fluorescence enhancement in the orthosteric binding site of the p38α kinase is the readout for bioaffinity. Parallel HR-MS(n) experiments provided information on the identity of binders and non-binders. Finally, the data produced with this on-line system were compared to electrochemical conversion products generated off-line. The electrochemical conversion of 1-{6-chloro-5-[(2R,5S)-4-(4-fluorobenzyl)-2,5-dimethylpiperazine-1-carbonyl]-3aH-indol-3-yl}-2-morpholinoethane-1,2-dione resulted in eight products, three of which showed bioaffinity in the continuous-flow p38α bioaffinity assay used. Electrochemical conversion of BIRB796 resulted, amongst others, in the formation of the reactive quinoneimine structure and its corresponding hydroquinone. Both products were detected in the p38α bioaffinity assay, which indicates binding to the p38α kinase.

  17. Stability of arsenic peptides in plant extracts: off-line versus on-line parallel elemental and molecular mass spectrometric detection for liquid chromatographic separation.

    Bluemlein, Katharina; Raab, Andrea; Feldmann, Jörg

    2009-01-01

    The instability of metal and metalloid complexes during analytical processes has always been an issue of an uncertainty regarding their speciation in plant extracts. Two different speciation protocols were compared regarding the analysis of arsenic phytochelatin (As(III)PC) complexes in fresh plant material. As the final step for separation/detection both methods used RP-HPLC simultaneously coupled to ICP-MS and ES-MS. However, one method was the often used off-line approach using two-dimensional separation, i.e. a pre-cleaning step using size-exclusion chromatography with subsequent fraction collection and freeze-drying prior to the analysis using RP-HPLC-ICP-MS and/or ES-MS. This approach revealed that less than 2% of the total arsenic was bound to peptides such as phytochelatins in the root extract of an arsenate exposed Thunbergia alata, whereas the direct on-line method showed that 83% of arsenic was bound to peptides, mainly as As(III)PC(3) and (GS)As(III)PC(2). Key analytical factors were identified which destabilise the As(III)PCs. The low pH of the mobile phase (0.1% formic acid) using RP-HPLC-ICP-MS/ES-MS stabilises the arsenic peptide complexes in the plant extract as well as the free peptide concentration, as shown by the kinetic disintegration study of the model compound As(III)(GS)(3) at pH 2.2 and 3.8. But only short half-lives of only a few hours were determined for the arsenic glutathione complex. Although As(III)PC(3) showed a ten times higher half-life (23 h) in a plant extract, the pre-cleaning step with subsequent fractionation in a mobile phase of pH 5.6 contributes to the destabilisation of the arsenic peptides in the off-line method. Furthermore, it was found that during a freeze-drying process more than 90% of an As(III)PC(3) complex and smaller free peptides such as PC(2) and PC(3) can be lost. Although the two-dimensional off-line method has been used successfully for other metal complexes, it is concluded here that the fractionation and

  18. Study on Distribution Reliability with Parallel and On-site Distributed Generation Considering Protection Miscoordination and Tie Line

    Chaitusaney, Surachai; Yokoyama, Akihiko

    In distribution system, Distributed Generation (DG) is expected to improve the system reliability as its backup generation. However, DG contribution in fault current may cause the loss of the existing protection coordination, e.g. recloser-fuse coordination and breaker-breaker coordination. This problem can drastically deteriorate the system reliability, and it is more serious and complicated when there are several DG sources in the system. Hence, the above conflict in reliability aspect unavoidably needs a detailed investigation before the installation or enhancement of DG is done. The model of composite DG fault current is proposed to find the threshold beyond which existing protection coordination is lost. Cases of protection miscoordination are described, together with their consequences. Since a distribution system may be tied with another system, the issues of tie line and on-site DG are integrated into this study. Reliability indices are evaluated and compared in the distribution reliability test system RBTS Bus 2.

  19. Power-flow control and stability enhancement of four parallel-operated offshore wind farms using a line-commutated HVDC link

    Wang, Li; Wang, Kuo-Hua; Lee, Wei-Jen

    2010-01-01

    This paper presents an effective control scheme using a line-commutated high-voltage direct-current (HVDC) link with a designed rectifier current regulator (RCR) to simultaneously perform both power-fluctuation mitigation and damping improvement of four parallel-operated 80-MW offshore wind farms...... delivering generated power to a large utility grid. The proposed RCR of the HVDC link is designed by using modal control theory to contribute adequate damping to the studied four offshore wind farms under various wind speeds. A systematic analysis using a frequency-domain approach based on eigenvalue...... characteristics to the studied offshore wind farms under various wind speeds but also effectively mitigate power fluctuations of the offshore wind farms under wind-speed disturbance conditions....

  20. LINES

    Minas Bakalchev

    2015-10-01

    Full Text Available The perception of elements in a system often creates their interdependence, interconditionality, and suppression. The lines from a basic geometrical element have become the model of a reductive world based on isolation according to certain criteria such as function, structure, and social organization. Their traces are experienced in the contemporary world as fragments or ruins of a system of domination of an assumed hierarchical unity. How can one release oneself from such dependence or determinism? How can the lines become less “systematic” and forms more autonomous, and less reductive? How is a form released from modernistic determinism on the new controversial ground? How can these elements or forms of representation become forms of action in the present complex world? In this paper, the meaning of lines through the ideas of Le Corbusier, Leonidov, Picasso, and Hitchcock is presented. Spatial research was made through a series of examples arising from the projects of the architectural studio “Residential Transformations”, which was a backbone for mapping the possibilities ranging from playfulness to exactness, as tactics of transformation in the different contexts of the contemporary world.

  1. Decreased CSF-flow artefacts in T2 imaging of the cervical spine with periodically rotated overlapping parallel lines with enhanced reconstruction (PROPELLER/BLADE)

    Ragoschke-Schumm, Andreas; Schmidt, Peter; Mayer, Thomas E.; Schumm, Julia; Reimann, Georg; Mentzel, Hans-Joachim; Kaiser, Werner A.

    2011-01-01

    The cervical spine is prone to artefacts in T2 MR-imaging due to patient movements and cerebrospinal fluid flow. The periodically rotated overlapping parallel lines with enhanced reconstruction (PROPELLER/BLADE) acquisition method was developed to reduce motion artefacts. We sought to determine if T2-BLADE is superior to T2-TSE with conventional k-space reading. Twenty-five patients were examined using a 1.5 T MR-scanner. T2-weighted imaging of the cervical spine in sagittal and axial orientation using conventional or BLADE k-space reading was performed. Spinal cord, subarachnoid space, vertebrae and discs were evaluated by two independent observers using a scale from 0 (non-diagnostic) to 3 (excellent). Interobserver correlation was assessed as Cohen's kappa. Results of Mann-Whitney U test with p < 0.05 were regarded as significant. Furthermore, the investigators were asked for subjective evaluation in consensus. Overall interobserver accuracy of κ = 0.91 was obtained. Comparison of sagittal images showed better values for all investigated structures in T2-BLADE: spinal cord (TSE/BLADE: 1.52/2.04; p < 0.001), subarachnoid space (1.36/2.06; p < 0.001) and vertebrae/discs (1.66/2.86; p < 0.001). Comparison of axial images showed better values in T2-BLADE for spinal cord (1.68/1.86; p = 0.149) and vertebrae/discs (1.0/1.96: p < 0.001) while subarachnoid space was better to be evaluated in conventional T2-TSE (1.94/1.12; p < 0.001). In sagittal orientation, motion- and CSF-flow artefacts were reduced in T2-BLADE. In axial orientation, however, CSF-flow artefacts were pronounced in T2-BLADE. The image quality of the sagittal T2-BLADE sequences was significantly better than the T2-TSE and acquired in less time. In axial orientation, increased CSF-flow artefacts may reduce accuracy of structures in the subarachnoid space. (orig.)

  2. Parallel rendering

    Crockett, Thomas W.

    1995-01-01

    This article provides a broad introduction to the subject of parallel rendering, encompassing both hardware and software systems. The focus is on the underlying concepts and the issues which arise in the design of parallel rendering algorithms and systems. We examine the different types of parallelism and how they can be applied in rendering applications. Concepts from parallel computing, such as data decomposition, task granularity, scalability, and load balancing, are considered in relation to the rendering problem. We also explore concepts from computer graphics, such as coherence and projection, which have a significant impact on the structure of parallel rendering algorithms. Our survey covers a number of practical considerations as well, including the choice of architectural platform, communication and memory requirements, and the problem of image assembly and display. We illustrate the discussion with numerous examples from the parallel rendering literature, representing most of the principal rendering methods currently used in computer graphics.

  3. Parallel computations

    1982-01-01

    Parallel Computations focuses on parallel computation, with emphasis on algorithms used in a variety of numerical and physical applications and for many different types of parallel computers. Topics covered range from vectorization of fast Fourier transforms (FFTs) and of the incomplete Cholesky conjugate gradient (ICCG) algorithm on the Cray-1 to calculation of table lookups and piecewise functions. Single tridiagonal linear systems and vectorized computation of reactive flow are also discussed.Comprised of 13 chapters, this volume begins by classifying parallel computers and describing techn

  4. Parallel algorithms

    Casanova, Henri; Robert, Yves

    2008-01-01

    ""…The authors of the present book, who have extensive credentials in both research and instruction in the area of parallelism, present a sound, principled treatment of parallel algorithms. … This book is very well written and extremely well designed from an instructional point of view. … The authors have created an instructive and fascinating text. The book will serve researchers as well as instructors who need a solid, readable text for a course on parallelism in computing. Indeed, for anyone who wants an understandable text from which to acquire a current, rigorous, and broad vi

  5. Problems with earth fault detecting relays assigned to parallel cables or overhead lines; Probleme bei der Erdschlussortung mit wattmetrischen Erdschlussrichtungsrelais bei parallelen Kabeln oder Leitungen

    Birkner, P.; Foerg, R. [Lech-Elektrizitaetswerke AG, Augsburg (Germany)

    1998-06-29

    For practical conditions one can find currents in underground electrical conductors like cable coverings earthed on both sides. As an example these currents are due to the alternating current system of the railroad or to the alternating current system of a Peterson coil, that tries to find a minimum resistance way from the transformer station to the place of the earth fault. Currents like these create a series voltage in the cable by inductive coupling. The voltage depends on the type and the length of the cable. The series voltages of all three phases form a zero sequence system. Taking into consideration that two cable systems running parallel to another, under certain circumstances it is possible to achieve a circulating zero sequence current. Additionally there is a shift voltage between the neutral point and the earth in the case of an earth fault in another place in the grid. The combination of these two factors can cause a malfunction of the earth fault detecting relays that are assigned to the parallel cable system. (orig.) [Deutsch] Im Erdreich vorhandene elektrische Leiter, z.B. die beidseitig geerdeten Schirme von Energiekabeln, werden in der Praxis nicht selten von Stroemen beaufschlagt. Dabei kann es sich z.B. auch um den Wechselstrom einer Petersenspule, der sich im Erdschlussfall einen widerstandsminimierten Weg vom Umspannwerk zur Fehlerstelle sucht, handeln. Ueber induktive Einkopplung entsteht im Leiter des Kabels eine Laengsspannung. Deren Hoehe ist vom Kabeltyp und der Kabellaenge abhaengig. Liegt als Netzkonfiguration eine Doppelleitung vor, die parallel betrieben wird, so koennen sich unter gewissen Randbedingungen kreisende Nullstroeme ausbilden. Diese wiederum koennen bei Vorhandensein einer Verlagerungsspannung zu einem Fehlansprechen von wattmetrischen Erdschlussrichtungsrelais fuehren. (orig.)

  6. PORTA: A three-dimensional multilevel radiative transfer code for modeling the intensity and polarization of spectral lines with massively parallel computers

    Štěpán, Jiří; Trujillo Bueno, J.

    2013-01-01

    Roč. 557, September (2013), A143/1-A143/15 ISSN 0004-6361 R&D Projects: GA ČR GPP209/12/P741 Grant - others:EU(XE) COST action MP1104 Institutional support: RVO:67985815 Keywords : line formation * magnetic fields * numerical methods Subject RIV: BN - Astronomy, Celestial Mechanics, Astrophysics Impact factor: 4.479, year: 2013

  7. Identification of phenylbutyrate-generated metabolites in Huntington disease patients using parallel liquid chromatography/electrochemical array/mass spectrometry and off-line tandem mass spectrometry.

    Ebbel, Erika N; Leymarie, Nancy; Schiavo, Susan; Sharma, Swati; Gevorkian, Sona; Hersch, Steven; Matson, Wayne R; Costello, Catherine E

    2010-04-15

    Oral sodium phenylbutyrate (SPB) is currently under investigation as a histone deacetylation (HDAC) inhibitor in Huntington disease (HD). Ongoing studies indicate that symptoms related to HD genetic abnormalities decrease with SPB therapy. In a recently reported safety and tolerability study of SPB in HD, we analyzed overall chromatographic patterns from a method that employs gradient liquid chromatography with series electrochemical array, ultraviolet (UV), and fluorescence (LCECA/UV/F) for measuring SPB and its metabolite phenylacetate (PA). We found that plasma and urine from SPB-treated patients yielded individual-specific patterns of approximately 20 metabolites that may provide a means for the selection of subjects for extended trials of SPB. The structural identification of these metabolites is of critical importance because their characterization will facilitate understanding the mechanisms of drug action and possible side effects. We have now developed an iterative process with LCECA, parallel LCECA/LCMS, and high-performance tandem MS for metabolite characterization. Here we report the details of this method and its use for identification of 10 plasma and urinary metabolites in treated subjects, including indole species in urine that are not themselves metabolites of SPB. Thus, this approach contributes to understanding metabolic pathways that differ among HD patients being treated with SPB. Copyright 2010 Elsevier Inc. All rights reserved.

  8. Parallel computation

    Jejcic, A.; Maillard, J.; Maurel, G.; Silva, J.; Wolff-Bacha, F.

    1997-01-01

    The work in the field of parallel processing has developed as research activities using several numerical Monte Carlo simulations related to basic or applied current problems of nuclear and particle physics. For the applications utilizing the GEANT code development or improvement works were done on parts simulating low energy physical phenomena like radiation, transport and interaction. The problem of actinide burning by means of accelerators was approached using a simulation with the GEANT code. A program of neutron tracking in the range of low energies up to the thermal region has been developed. It is coupled to the GEANT code and permits in a single pass the simulation of a hybrid reactor core receiving a proton burst. Other works in this field refers to simulations for nuclear medicine applications like, for instance, development of biological probes, evaluation and characterization of the gamma cameras (collimators, crystal thickness) as well as the method for dosimetric calculations. Particularly, these calculations are suited for a geometrical parallelization approach especially adapted to parallel machines of the TN310 type. Other works mentioned in the same field refer to simulation of the electron channelling in crystals and simulation of the beam-beam interaction effect in colliders. The GEANT code was also used to simulate the operation of germanium detectors designed for natural and artificial radioactivity monitoring of environment

  9. 40 CFR Appendix C to Part 112 - Substantial Harm Criteria

    2010-07-01

    ... to Part 112 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER PROGRAMS OIL POLLUTION PREVENTION Pt. 112, App. C Appendix C to Part 112—Substantial Harm Criteria 1.0Introduction The..., except in the Gulf of Mexico. In the Gulf of Mexico, it means the area shoreward of the lines of...

  10. 24 CFR 902.79 - Substantial default.

    2010-04-01

    ... 24 Housing and Urban Development 4 2010-04-01 2010-04-01 false Substantial default. 902.79 Section... PUBLIC HOUSING ASSESSMENT SYSTEM PHAS Incentives and Remedies § 902.79 Substantial default. (a) Events or conditions that constitute substantial default. The following events or conditions shall constitute...

  11. Processing semblances induced through inter-postsynaptic functional LINKs, presumed biological parallels of K-lines proposed for building artificial intelligence

    Kunjumon I Vadakkan

    2011-07-01

    Full Text Available The internal sensation of memory, which is available only to the owner of an individual nervous system, is difficult to analyze for its basic elements of operation. We hypothesize that associative learning induces the formation of functional LINK between the postsynapses. During memory retrieval, the activation of either postsynapse re-activates the functional LINK evoking a semblance of sensory activity arriving at its opposite postsynapse, nature of which defines the basic unit of virtual internal sensation - namely, semblion. Neuronal networks that undergo continuous oscillatory activity at certain levels of their organization induce semblions enabling the system to continuously learn, self-organize, and demonstrate instantiation, features that can be utilized for developing artificial intelligence (AI. Suitability of the inter-postsynaptic functional LINKs to meet the expectations of Minsky’s K-lines, basic elements of a memory theory generated to develop AI and methods to replicate semblances outside the nervous system are explained.

  12. Processing Semblances Induced through Inter-Postsynaptic Functional LINKs, Presumed Biological Parallels of K-Lines Proposed for Building Artificial Intelligence

    Vadakkan, Kunjumon I.

    2011-01-01

    The internal sensation of memory, which is available only to the owner of an individual nervous system, is difficult to analyze for its basic elements of operation. We hypothesize that associative learning induces the formation of functional LINK between the postsynapses. During memory retrieval, the activation of either postsynapse re-activates the functional LINK evoking a semblance of sensory activity arriving at its opposite postsynapse, nature of which defines the basic unit of internal sensation – namely, the semblion. In neuronal networks that undergo continuous oscillatory activity at certain levels of their organization re-activation of functional LINKs is expected to induce semblions, enabling the system to continuously learn, self-organize, and demonstrate instantiation, features that can be utilized for developing artificial intelligence (AI). This paper also explains suitability of the inter-postsynaptic functional LINKs to meet the expectations of Minsky’s K-lines, basic elements of a memory theory generated to develop AI and methods to replicate semblances outside the nervous system. PMID:21845180

  13. Parallel R

    McCallum, Ethan

    2011-01-01

    It's tough to argue with R as a high-quality, cross-platform, open source statistical software product-unless you're in the business of crunching Big Data. This concise book introduces you to several strategies for using R to analyze large datasets. You'll learn the basics of Snow, Multicore, Parallel, and some Hadoop-related tools, including how to find them, how to use them, when they work well, and when they don't. With these packages, you can overcome R's single-threaded nature by spreading work across multiple CPUs, or offloading work to multiple machines to address R's memory barrier.

  14. Parallel Solid-Phase Synthesis Using a New Diethylsilylacetylenic Linker and Leading to Mestranol Derivatives with Potent Antiproliferative Activities on Multiple Cancer Cell Lines.

    Dutour, Raphael; Maltais, Rene; Perreault, Martin; Roy, Jenny; Poirier, Donald

    2018-03-07

    RM-133 belongs to a new family of aminosteroid derivatives demonstrating interesting anticancer properties, as confirmed in vivo in four mouse cancer xenograft models. However, the metabolic stability of RM-133 needs to be improved. After investigation, the replacement of its androstane scaffold by a more stable estrane scaffold led to the development of the mestranol derivative RM-581. Using solid-phase strategy involving five steps, we quickly synthesized a series of RM-581 analogs using the recently-developed diethylsilyl acetylenic linker. To establish structure-activity relationships, we then investigated their antiproliferative potency on a panel of cancer cell lines from various cancers (breast, prostate, ovarian and pancreatic). Some of the mestranol derivatives have shown in vitro anticancer activities that are close to, or better than those observed for RM-581. Compound 23, a mestranol derivative having a ((3,5-dimethylbenzoyl)-L-prolyl)piperazine side chain at position C2, was found to be active as an antiproliferative agent (IC50 = 0.38 ± 0.34 to 3.17 ± 0.10 µM) and to be twice as active as RM-581 on LNCaP, PC-3, MCF-7, PANC-1 and OVCAR-3 cancer cells (IC50 = 0.56 ± 0.30, 0.89 ± 0.63, 1.36 ± 0.31, 2.47 ± 0.91 and 3.17 ± 0.10 µM, respectively). Easily synthesized in good yields by both solid-phase organic synthesis and classic solution-phase chemistry, this promising candidate could be used as an antiproliferative agent on a variety of cancers, notably pancreatic and ovarian cancers, both having very bad prognoses. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  15. Characterization of DNA repair phenotypes of Xeroderma pigmentosum cell lines by a paralleled in vitro test; Phenotypage de la reparation de l'ADN de lignees Xeroderma pigmentosum, par un test in vitro multiparametrique

    Raffin, A.L.

    2009-06-15

    DNA is constantly damaged modifying the genetic information for which it encodes. Several cellular mechanisms as the Base Excision Repair (BER) and the Nucleotide Excision Repair (NER) allow recovering the right DNA sequence. The Xeroderma pigmentosum is a disease characterised by a deficiency in the NER pathway. The aim of this study was to propose an efficient and fast test for the diagnosis of this disease as an alternative to the currently available UDS test. DNA repair activities of XP cell lines were quantified using in vitro miniaturized and paralleled tests in order to establish DNA repair phenotypes of XPA and XPC deficient cells. The main advantage of the tests used in this study is the simultaneous measurement of excision or excision synthesis (ES) of several lesions by only one cellular extract. We showed on one hand that the relative ES of the different lesions depend strongly on the protein concentration of the nuclear extract tested. Working at high protein concentration allowed discriminating the XP phenotype versus the control one, whereas it was impossible under a certain concentration's threshold. On the other hand, while the UVB irradiation of control cells stimulated their repair activities, this effect was not observed in XP cells. This study brings new information on the XPA and XPC protein roles during BER and NER and underlines the complexity of the regulations of DNA repair processes. (author)

  16. Automatic Loop Parallelization via Compiler Guided Refactoring

    Larsen, Per; Ladelsky, Razya; Lidman, Jacob

    For many parallel applications, performance relies not on instruction-level parallelism, but on loop-level parallelism. Unfortunately, many modern applications are written in ways that obstruct automatic loop parallelization. Since we cannot identify sufficient parallelization opportunities...... for these codes in a static, off-line compiler, we developed an interactive compilation feedback system that guides the programmer in iteratively modifying application source, thereby improving the compiler’s ability to generate loop-parallel code. We use this compilation system to modify two sequential...... benchmarks, finding that the code parallelized in this way runs up to 8.3 times faster on an octo-core Intel Xeon 5570 system and up to 12.5 times faster on a quad-core IBM POWER6 system. Benchmark performance varies significantly between the systems. This suggests that semi-automatic parallelization should...

  17. Increased greenhouse effect substantiated through measurements

    Skartveit, Arvid

    2001-01-01

    The article presents studies on the greenhouse effect which substantiates the results from satellite measurements during the period 1970 - 1997. These show an increased effect due to increase in the concentration of the climatic gases CO 2 , methane, CFC-11 and CFC-12 in the atmosphere

  18. 21 CFR 514.4 - Substantial evidence.

    2010-04-01

    ... adequate and well-controlled studies, such as a study in a target species, study in laboratory animals... and conditions of use. Substantial evidence of effectiveness of a new animal drug shall demonstrate that the new animal drug is effective for each intended use and associated conditions of use for and...

  19. Toward More Substantial Theories of Language Acquisition

    Jenson, Cinnamon Ann

    2015-01-01

    Cognitive linguists argue that certain sets of knowledge of language are innate. However, critics have argued that the theoretical concept of "innateness" should be eliminated since it is ambiguous and insubstantial. In response, I aim to strengthen theories of language acquisition and identify ways to make them more substantial. I…

  20. SESOTHO trial ("Switch Either near Suppression Or THOusand") - switch to second-line versus WHO-guided standard of care for unsuppressed patients on first-line ART with viremia below 1000 copies/mL: protocol of a multicenter, parallel-group, open-label, randomized clinical trial in Lesotho, Southern Africa.

    Amstutz, Alain; Nsakala, Bienvenu Lengo; Vanobberghen, Fiona; Muhairwe, Josephine; Glass, Tracy Renée; Achieng, Beatrice; Sepeka, Mamorena; Tlali, Katleho; Sao, Lebohang; Thin, Kyaw; Klimkait, Thomas; Battegay, Manuel; Labhardt, Niklaus Daniel

    2018-02-12

    The World Health Organization (WHO) recommends viral load (VL) measurement as the preferred monitoring strategy for HIV-infected individuals on antiretroviral therapy (ART) in resource-limited settings. The new WHO guidelines 2016 continue to define virologic failure as two consecutive VL ≥1000 copies/mL (at least 3 months apart) despite good adherence, triggering switch to second-line therapy. However, the threshold of 1000 copies/mL for defining virologic failure is based on low-quality evidence. Observational studies have shown that individuals with low-level viremia (measurable but below 1000 copies/mL) are at increased risk for accumulation of resistance mutations and subsequent virologic failure. The SESOTHO trial assesses a lower threshold for switch to second-line ART in patients with sustained unsuppressed VL. In this multicenter, parallel-group, open-label, randomized controlled trial conducted in Lesotho, patients on first-line ART with two consecutive unsuppressed VL measurements ≥100 copies/mL, where the second VL is between 100 and 999 copies/mL, will either be switched to second-line ART immediately (intervention group) or not be switched (standard of care, according to WHO guidelines). The primary endpoint is viral resuppression (VL < 50 copies/mL) 9 months after randomization. We will enrol 80 patients, giving us 90% power to detect a difference of 35% in viral resuppression between the groups (assuming two-sided 5% alpha error). For our primary analysis, we will use a modified intention-to-treat set, with those lost to care, death, or crossed over considered failure to resuppress, and using logistic regression models adjusted for the prespecified stratification variables. The SESOTHO trial challenges the current WHO guidelines, assessing an alternative, lower VL threshold for patients with unsuppressed VL on first-line ART. This trial will provide data to inform future WHO guidelines on VL thresholds to recommend switch to second-line ART

  1. Fast parallel event reconstruction

    CERN. Geneva

    2010-01-01

    On-line processing of large data volumes produced in modern HEP experiments requires using maximum capabilities of modern and future many-core CPU and GPU architectures.One of such powerful feature is a SIMD instruction set, which allows packing several data items in one register and to operate on all of them, thus achievingmore operations per clock cycle. Motivated by the idea of using the SIMD unit ofmodern processors, the KF based track fit has been adapted for parallelism, including memory optimization, numerical analysis, vectorization with inline operator overloading, and optimization using SDKs. The speed of the algorithm has been increased in 120000 times with 0.1 ms/track, running in parallel on 16 SPEs of a Cell Blade computer.  Running on a Nehalem CPU with 8 cores it shows the processing speed of 52 ns/track using the Intel Threading Building Blocks. The same KF algorithm running on an Nvidia GTX 280 in the CUDA frameworkprovi...

  2. Parallel heater system for subsurface formations

    Harris, Christopher Kelvin [Houston, TX; Karanikas, John Michael [Houston, TX; Nguyen, Scott Vinh [Houston, TX

    2011-10-25

    A heating system for a subsurface formation is disclosed. The system includes a plurality of substantially horizontally oriented or inclined heater sections located in a hydrocarbon containing layer in the formation. At least a portion of two of the heater sections are substantially parallel to each other. The ends of at least two of the heater sections in the layer are electrically coupled to a substantially horizontal, or inclined, electrical conductor oriented substantially perpendicular to the ends of the at least two heater sections.

  3. Conformal pure radiation with parallel rays

    Leistner, Thomas; Paweł Nurowski

    2012-01-01

    We define pure radiation metrics with parallel rays to be n-dimensional pseudo-Riemannian metrics that admit a parallel null line bundle K and whose Ricci tensor vanishes on vectors that are orthogonal to K. We give necessary conditions in terms of the Weyl, Cotton and Bach tensors for a pseudo-Riemannian metric to be conformal to a pure radiation metric with parallel rays. Then, we derive conditions in terms of the tractor calculus that are equivalent to the existence of a pure radiation metric with parallel rays in a conformal class. We also give analogous results for n-dimensional pseudo-Riemannian pp-waves. (paper)

  4. Parallel Programming with Intel Parallel Studio XE

    Blair-Chappell , Stephen

    2012-01-01

    Optimize code for multi-core processors with Intel's Parallel Studio Parallel programming is rapidly becoming a "must-know" skill for developers. Yet, where to start? This teach-yourself tutorial is an ideal starting point for developers who already know Windows C and C++ and are eager to add parallelism to their code. With a focus on applying tools, techniques, and language extensions to implement parallelism, this essential resource teaches you how to write programs for multicore and leverage the power of multicore in your programs. Sharing hands-on case studies and real-world examples, the

  5. Substantiating the Incurred but not Reported Reserve

    Georgeta Vintilã

    2009-12-01

    Full Text Available In order to handle past and future liability taken by insurance contracts concluded, any insurance company must constitute and maintain technical reserves. Substantiating technical reserves is done through actuarial methods and its over-evaluation or under-evaluation influence solvency and financial performance of the insurance companies, in the sense of reducing solvency through over-evaluating reserves and, respectively, influencing profit (hence of outstanding tax through under-evaluating reserves. An important reserve for insurance companies is represented by the incurred but not reported reserve, as it allows the estimation of the liability the company may confront in the future, generated by events occurred in the past, which are not currently known in the present but will be reported in the future.

  6. New algorithms for parallel MRI

    Anzengruber, S; Ramlau, R; Bauer, F; Leitao, A

    2008-01-01

    Magnetic Resonance Imaging with parallel data acquisition requires algorithms for reconstructing the patient's image from a small number of measured lines of the Fourier domain (k-space). In contrast to well-known algorithms like SENSE and GRAPPA and its flavors we consider the problem as a non-linear inverse problem. However, in order to avoid cost intensive derivatives we will use Landweber-Kaczmarz iteration and in order to improve the overall results some additional sparsity constraints.

  7. Language constructs for modular parallel programs

    Foster, I.

    1996-03-01

    We describe programming language constructs that facilitate the application of modular design techniques in parallel programming. These constructs allow us to isolate resource management and processor scheduling decisions from the specification of individual modules, which can themselves encapsulate design decisions concerned with concurrence, communication, process mapping, and data distribution. This approach permits development of libraries of reusable parallel program components and the reuse of these components in different contexts. In particular, alternative mapping strategies can be explored without modifying other aspects of program logic. We describe how these constructs are incorporated in two practical parallel programming languages, PCN and Fortran M. Compilers have been developed for both languages, allowing experimentation in substantial applications.

  8. Practical parallel computing

    Morse, H Stephen

    1994-01-01

    Practical Parallel Computing provides information pertinent to the fundamental aspects of high-performance parallel processing. This book discusses the development of parallel applications on a variety of equipment.Organized into three parts encompassing 12 chapters, this book begins with an overview of the technology trends that converge to favor massively parallel hardware over traditional mainframes and vector machines. This text then gives a tutorial introduction to parallel hardware architectures. Other chapters provide worked-out examples of programs using several parallel languages. Thi

  9. Parallel sorting algorithms

    Akl, Selim G

    1985-01-01

    Parallel Sorting Algorithms explains how to use parallel algorithms to sort a sequence of items on a variety of parallel computers. The book reviews the sorting problem, the parallel models of computation, parallel algorithms, and the lower bounds on the parallel sorting problems. The text also presents twenty different algorithms, such as linear arrays, mesh-connected computers, cube-connected computers. Another example where algorithm can be applied is on the shared-memory SIMD (single instruction stream multiple data stream) computers in which the whole sequence to be sorted can fit in the

  10. Substantial nitrogen pollution embedded in international trade

    Oita, Azusa; Malik, Arunima; Kanemoto, Keiichiro; Geschke, Arne; Nishijima, Shota; Lenzen, Manfred

    2016-02-01

    Anthropogenic emissions of reactive nitrogen to the atmosphere and water bodies can damage human health and ecosystems. As a measure of a nation’s contribution to this potential damage, a country’s nitrogen footprint has been defined as the quantity of reactive nitrogen emitted during the production, consumption and transportation of commodities consumed within that country, whether those commodities are produced domestically or internationally. Here we use global emissions databases, a global nitrogen cycle model, and a global input-output database of domestic and international trade to calculate the nitrogen footprints for 188 countries as the sum of emissions of ammonia, nitrogen oxides and nitrous oxide to the atmosphere, and of nitrogen potentially exportable to water bodies. Per-capita footprints range from under 7 kg N yr-1 in some developing countries to over 100 kg N yr-1 in some wealthy nations. Consumption in China, India, the United States and Brazil is responsible for 46% of global emissions. Roughly a quarter of the global nitrogen footprint is from commodities that were traded across country borders. The main net exporters have significant agricultural, food and textile exports, and are often developing countries, whereas important net importers are almost exclusively developed economies. We conclude that substantial local nitrogen pollution is driven by demand from consumers in other countries.

  11. Introduction to parallel programming

    Brawer, Steven

    1989-01-01

    Introduction to Parallel Programming focuses on the techniques, processes, methodologies, and approaches involved in parallel programming. The book first offers information on Fortran, hardware and operating system models, and processes, shared memory, and simple parallel programs. Discussions focus on processes and processors, joining processes, shared memory, time-sharing with multiple processors, hardware, loops, passing arguments in function/subroutine calls, program structure, and arithmetic expressions. The text then elaborates on basic parallel programming techniques, barriers and race

  12. Parallel computing works!

    Fox, Geoffrey C; Messina, Guiseppe C

    2014-01-01

    A clear illustration of how parallel computers can be successfully appliedto large-scale scientific computations. This book demonstrates how avariety of applications in physics, biology, mathematics and other scienceswere implemented on real parallel computers to produce new scientificresults. It investigates issues of fine-grained parallelism relevant forfuture supercomputers with particular emphasis on hypercube architecture. The authors describe how they used an experimental approach to configuredifferent massively parallel machines, design and implement basic systemsoftware, and develop

  13. Substantial Union or Substantial Distinction of Mind and Body in Descartes' Metaphysics

    Fahime Jamei

    2009-01-01

    Full Text Available According to Descartes’ metaphysics there are two different kinds of substances in the world of creatures: “thinking substance” and “extended substance” or soul and matter. In Descartes’ philosophy the soul is equal to the mind and considered as a “thinking substance”. This immaterial substance is the essence of the human being. Body, being considered as a “matter“, is an “extended substance” and entirely distinct from the soul. The soul, therefore, exists and may be known prior to body and, not being corporeal, can exist after human death. Hence, Descartes can prove the immortality of human soul in the framework of the principle of substantial distinction. On the other hand, as a physiologist and psychologist, Descartes indeed believes in mind-body union, so that some causal interactions between mind and body show their substantial union. In this essay, the authors show that Descartes faces a serious problem in combining substantial union of mind and body with their substantial distinction; despite of his efforts in introducing the idea of pineal gland, the problem remains unsolved. Therefore it seems that as he cannot dispense with his only reason for proving the immortality of human soul, he has to hold the mind-body distinction theory in his metaphysics. Indeed, Descartes prefers to support the distinction theory rather than union theory in confronting a thesis and an antithesis stating one of two theories

  14. Substantial :union: or Substantial Distinction of Mind and Body in Descartes\\' Metaphysics

    f Jamei

    2009-06-01

    Full Text Available According to Descartes’ metaphysics there are two different kinds of substances in the world of creatures: “thinking substance” and “extended substance” or soul and matter. In Descartes’ philosophy the soul is equal to the mind and considered as a “thinking substance”. This immaterial substance is the essence of the human being. Body, being considered as a “matter“, is an “extended substance” and entirely distinct from the soul. The soul, therefore, exists and may be known prior to body and, not being corporeal, can exist after human death. Hence, Descartes can prove the immortality of human soul in the framework of the principle of substantial distinction. On the other hand, as a physiologist and psychologist, Descartes indeed believes in mind-body :union:, so that some causal interactions between mind and body show their substantial :union:. In this essay, the authors show that Descartes faces a serious problem in combining substantial :union: of mind and body with their substantial distinction despite of his efforts in introducing the idea of pineal gland, the problem remains unsolved. Therefore it seems that as he cannot dispense with his only reason for proving the immortality of human soul, he has to hold the mind-body distinction theory in his metaphysics. Indeed, Descartes prefers to support the distinction theory rather than :union: theory in confronting a thesis and an antithesis stating one of two theories.

  15. Series Transmission Line Transformer

    Buckles, Robert A.; Booth, Rex; Yen, Boris T.

    2004-06-29

    A series transmission line transformer is set forth which includes two or more of impedance matched sets of at least two transmissions lines such as shielded cables, connected in parallel at one end ans series at the other in a cascading fashion. The cables are wound about a magnetic core. The series transmission line transformer (STLT) which can provide for higher impedance ratios and bandwidths, which is scalable, and which is of simpler design and construction.

  16. Parallel S/sub n/ iteration schemes

    Wienke, B.R.; Hiromoto, R.E.

    1986-01-01

    The iterative, multigroup, discrete ordinates (S/sub n/) technique for solving the linear transport equation enjoys widespread usage and appeal. Serial iteration schemes and numerical algorithms developed over the years provide a timely framework for parallel extension. On the Denelcor HEP, the authors investigate three parallel iteration schemes for solving the one-dimensional S/sub n/ transport equation. The multigroup representation and serial iteration methods are also reviewed. This analysis represents a first attempt to extend serial S/sub n/ algorithms to parallel environments and provides good baseline estimates on ease of parallel implementation, relative algorithm efficiency, comparative speedup, and some future directions. The authors examine ordered and chaotic versions of these strategies, with and without concurrent rebalance and diffusion acceleration. Two strategies efficiently support high degrees of parallelization and appear to be robust parallel iteration techniques. The third strategy is a weaker parallel algorithm. Chaotic iteration, difficult to simulate on serial machines, holds promise and converges faster than ordered versions of the schemes. Actual parallel speedup and efficiency are high and payoff appears substantial

  17. Parallel Atomistic Simulations

    HEFFELFINGER,GRANT S.

    2000-01-18

    Algorithms developed to enable the use of atomistic molecular simulation methods with parallel computers are reviewed. Methods appropriate for bonded as well as non-bonded (and charged) interactions are included. While strategies for obtaining parallel molecular simulations have been developed for the full variety of atomistic simulation methods, molecular dynamics and Monte Carlo have received the most attention. Three main types of parallel molecular dynamics simulations have been developed, the replicated data decomposition, the spatial decomposition, and the force decomposition. For Monte Carlo simulations, parallel algorithms have been developed which can be divided into two categories, those which require a modified Markov chain and those which do not. Parallel algorithms developed for other simulation methods such as Gibbs ensemble Monte Carlo, grand canonical molecular dynamics, and Monte Carlo methods for protein structure determination are also reviewed and issues such as how to measure parallel efficiency, especially in the case of parallel Monte Carlo algorithms with modified Markov chains are discussed.

  18. Parallelization in Modern C++

    CERN. Geneva

    2016-01-01

    The traditionally used and well established parallel programming models OpenMP and MPI are both targeting lower level parallelism and are meant to be as language agnostic as possible. For a long time, those models were the only widely available portable options for developing parallel C++ applications beyond using plain threads. This has strongly limited the optimization capabilities of compilers, has inhibited extensibility and genericity, and has restricted the use of those models together with other, modern higher level abstractions introduced by the C++11 and C++14 standards. The recent revival of interest in the industry and wider community for the C++ language has also spurred a remarkable amount of standardization proposals and technical specifications being developed. Those efforts however have so far failed to build a vision on how to seamlessly integrate various types of parallelism, such as iterative parallel execution, task-based parallelism, asynchronous many-task execution flows, continuation s...

  19. Parallelism in matrix computations

    Gallopoulos, Efstratios; Sameh, Ahmed H

    2016-01-01

    This book is primarily intended as a research monograph that could also be used in graduate courses for the design of parallel algorithms in matrix computations. It assumes general but not extensive knowledge of numerical linear algebra, parallel architectures, and parallel programming paradigms. The book consists of four parts: (I) Basics; (II) Dense and Special Matrix Computations; (III) Sparse Matrix Computations; and (IV) Matrix functions and characteristics. Part I deals with parallel programming paradigms and fundamental kernels, including reordering schemes for sparse matrices. Part II is devoted to dense matrix computations such as parallel algorithms for solving linear systems, linear least squares, the symmetric algebraic eigenvalue problem, and the singular-value decomposition. It also deals with the development of parallel algorithms for special linear systems such as banded ,Vandermonde ,Toeplitz ,and block Toeplitz systems. Part III addresses sparse matrix computations: (a) the development of pa...

  20. 20 CFR 416.1075 - Finding of substantial failure.

    2010-04-01

    ... 20 Employees' Benefits 2 2010-04-01 2010-04-01 false Finding of substantial failure. 416.1075... AGED, BLIND, AND DISABLED Determinations of Disability Substantial Failure § 416.1075 Finding of substantial failure. A finding of substantial failure with respect to a State may not be made unless and until...

  1. 20 CFR 404.1675 - Finding of substantial failure.

    2010-04-01

    ... 20 Employees' Benefits 2 2010-04-01 2010-04-01 false Finding of substantial failure. 404.1675... DISABILITY INSURANCE (1950- ) Determinations of Disability Substantial Failure § 404.1675 Finding of substantial failure. A finding of substantial failure with respect to a State may not be made unless and until...

  2. Parallel Monte Carlo Search for Hough Transform

    Lopes, Raul H. C.; Franqueira, Virginia N. L.; Reid, Ivan D.; Hobson, Peter R.

    2017-10-01

    We investigate the problem of line detection in digital image processing and in special how state of the art algorithms behave in the presence of noise and whether CPU efficiency can be improved by the combination of a Monte Carlo Tree Search, hierarchical space decomposition, and parallel computing. The starting point of the investigation is the method introduced in 1962 by Paul Hough for detecting lines in binary images. Extended in the 1970s to the detection of space forms, what came to be known as Hough Transform (HT) has been proposed, for example, in the context of track fitting in the LHC ATLAS and CMS projects. The Hough Transform transfers the problem of line detection, for example, into one of optimization of the peak in a vote counting process for cells which contain the possible points of candidate lines. The detection algorithm can be computationally expensive both in the demands made upon the processor and on memory. Additionally, it can have a reduced effectiveness in detection in the presence of noise. Our first contribution consists in an evaluation of the use of a variation of the Radon Transform as a form of improving theeffectiveness of line detection in the presence of noise. Then, parallel algorithms for variations of the Hough Transform and the Radon Transform for line detection are introduced. An algorithm for Parallel Monte Carlo Search applied to line detection is also introduced. Their algorithmic complexities are discussed. Finally, implementations on multi-GPU and multicore architectures are discussed.

  3. A parallel buffer tree

    Sitchinava, Nodar; Zeh, Norbert

    2012-01-01

    We present the parallel buffer tree, a parallel external memory (PEM) data structure for batched search problems. This data structure is a non-trivial extension of Arge's sequential buffer tree to a private-cache multiprocessor environment and reduces the number of I/O operations by the number of...... in the optimal OhOf(psortN + K/PB) parallel I/O complexity, where K is the size of the output reported in the process and psortN is the parallel I/O complexity of sorting N elements using P processors....

  4. Parallel MR imaging.

    Deshmane, Anagha; Gulani, Vikas; Griswold, Mark A; Seiberlich, Nicole

    2012-07-01

    Parallel imaging is a robust method for accelerating the acquisition of magnetic resonance imaging (MRI) data, and has made possible many new applications of MR imaging. Parallel imaging works by acquiring a reduced amount of k-space data with an array of receiver coils. These undersampled data can be acquired more quickly, but the undersampling leads to aliased images. One of several parallel imaging algorithms can then be used to reconstruct artifact-free images from either the aliased images (SENSE-type reconstruction) or from the undersampled data (GRAPPA-type reconstruction). The advantages of parallel imaging in a clinical setting include faster image acquisition, which can be used, for instance, to shorten breath-hold times resulting in fewer motion-corrupted examinations. In this article the basic concepts behind parallel imaging are introduced. The relationship between undersampling and aliasing is discussed and two commonly used parallel imaging methods, SENSE and GRAPPA, are explained in detail. Examples of artifacts arising from parallel imaging are shown and ways to detect and mitigate these artifacts are described. Finally, several current applications of parallel imaging are presented and recent advancements and promising research in parallel imaging are briefly reviewed. Copyright © 2012 Wiley Periodicals, Inc.

  5. Parallel Algorithms and Patterns

    Robey, Robert W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-06-16

    This is a powerpoint presentation on parallel algorithms and patterns. A parallel algorithm is a well-defined, step-by-step computational procedure that emphasizes concurrency to solve a problem. Examples of problems include: Sorting, searching, optimization, matrix operations. A parallel pattern is a computational step in a sequence of independent, potentially concurrent operations that occurs in diverse scenarios with some frequency. Examples are: Reductions, prefix scans, ghost cell updates. We only touch on parallel patterns in this presentation. It really deserves its own detailed discussion which Gabe Rockefeller would like to develop.

  6. Application Portable Parallel Library

    Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott

    1995-01-01

    Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.

  7. Parallel discrete event simulation

    Overeinder, B.J.; Hertzberger, L.O.; Sloot, P.M.A.; Withagen, W.J.

    1991-01-01

    In simulating applications for execution on specific computing systems, the simulation performance figures must be known in a short period of time. One basic approach to the problem of reducing the required simulation time is the exploitation of parallelism. However, in parallelizing the simulation

  8. Parallel reservoir simulator computations

    Hemanth-Kumar, K.; Young, L.C.

    1995-01-01

    The adaptation of a reservoir simulator for parallel computations is described. The simulator was originally designed for vector processors. It performs approximately 99% of its calculations in vector/parallel mode and relative to scalar calculations it achieves speedups of 65 and 81 for black oil and EOS simulations, respectively on the CRAY C-90

  9. A position sensitive parallel plate avalanche counter

    Lombardi, M.; Tan Jilian; Potenza, R.; D'amico, V.

    1986-01-01

    A position sensitive parallel plate avalanche counter with a distributed constant delay-line-cathode (PSAC) is described. The strips formed on the printed board were served as the cathode and the delay line for readout of signals. The detector (PSAC) was operated in isobutane gas at the pressure range from 10 to 20 torr. The position resolution is better than 1 mm and the time resolution is about 350 ps, for 252 Cf fission-spectrum source

  10. Totally parallel multilevel algorithms

    Frederickson, Paul O.

    1988-01-01

    Four totally parallel algorithms for the solution of a sparse linear system have common characteristics which become quite apparent when they are implemented on a highly parallel hypercube such as the CM2. These four algorithms are Parallel Superconvergent Multigrid (PSMG) of Frederickson and McBryan, Robust Multigrid (RMG) of Hackbusch, the FFT based Spectral Algorithm, and Parallel Cyclic Reduction. In fact, all four can be formulated as particular cases of the same totally parallel multilevel algorithm, which are referred to as TPMA. In certain cases the spectral radius of TPMA is zero, and it is recognized to be a direct algorithm. In many other cases the spectral radius, although not zero, is small enough that a single iteration per timestep keeps the local error within the required tolerance.

  11. Parallel computing works

    1991-10-23

    An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

  12. Massively parallel mathematical sieves

    Montry, G.R.

    1989-01-01

    The Sieve of Eratosthenes is a well-known algorithm for finding all prime numbers in a given subset of integers. A parallel version of the Sieve is described that produces computational speedups over 800 on a hypercube with 1,024 processing elements for problems of fixed size. Computational speedups as high as 980 are achieved when the problem size per processor is fixed. The method of parallelization generalizes to other sieves and will be efficient on any ensemble architecture. We investigate two highly parallel sieves using scattered decomposition and compare their performance on a hypercube multiprocessor. A comparison of different parallelization techniques for the sieve illustrates the trade-offs necessary in the design and implementation of massively parallel algorithms for large ensemble computers.

  13. Parallel graded attention in reading: A pupillometric study

    Snell, Joshua; Mathot, Sebastiaan; Mirault, Jonathan; Grainger, Jonathan

    2018-01-01

    There are roughly two lines of theory to account for recent evidence that word processing is influenced by adjacent orthographic information. One line assumes that multiple words can be processed simultaneously through a parallel graded distribution of visuo-spatial attention. The other line assumes

  14. 19 CFR 10.7 - Substantial containers or holders.

    2010-04-01

    ... 19 Customs Duties 1 2010-04-01 2010-04-01 false Substantial containers or holders. 10.7 Section 10... Exported and Returned § 10.7 Substantial containers or holders. (a) Substantial containers or holders... domestic products exported and returned. When such containers or holders are imported not containing or...

  15. 20 CFR 604.6 - Conformity and substantial compliance.

    2010-04-01

    ... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Conformity and substantial compliance. 604.6... FOR ELIGIBILITY FOR UNEMPLOYMENT COMPENSATION § 604.6 Conformity and substantial compliance. (a) In... for the administration of its UC program. (b) Resolving Issues of Conformity and Substantial...

  16. Algorithms for parallel computers

    Churchhouse, R.F.

    1985-01-01

    Until relatively recently almost all the algorithms for use on computers had been designed on the (usually unstated) assumption that they were to be run on single processor, serial machines. With the introduction of vector processors, array processors and interconnected systems of mainframes, minis and micros, however, various forms of parallelism have become available. The advantage of parallelism is that it offers increased overall processing speed but it also raises some fundamental questions, including: (i) which, if any, of the existing 'serial' algorithms can be adapted for use in the parallel mode. (ii) How close to optimal can such adapted algorithms be and, where relevant, what are the convergence criteria. (iii) How can we design new algorithms specifically for parallel systems. (iv) For multi-processor systems how can we handle the software aspects of the interprocessor communications. Aspects of these questions illustrated by examples are considered in these lectures. (orig.)

  17. Parallelism and array processing

    Zacharov, V.

    1983-01-01

    Modern computing, as well as the historical development of computing, has been dominated by sequential monoprocessing. Yet there is the alternative of parallelism, where several processes may be in concurrent execution. This alternative is discussed in a series of lectures, in which the main developments involving parallelism are considered, both from the standpoint of computing systems and that of applications that can exploit such systems. The lectures seek to discuss parallelism in a historical context, and to identify all the main aspects of concurrency in computation right up to the present time. Included will be consideration of the important question as to what use parallelism might be in the field of data processing. (orig.)

  18. Parallel magnetic resonance imaging

    Larkman, David J; Nunes, Rita G

    2007-01-01

    Parallel imaging has been the single biggest innovation in magnetic resonance imaging in the last decade. The use of multiple receiver coils to augment the time consuming Fourier encoding has reduced acquisition times significantly. This increase in speed comes at a time when other approaches to acquisition time reduction were reaching engineering and human limits. A brief summary of spatial encoding in MRI is followed by an introduction to the problem parallel imaging is designed to solve. There are a large number of parallel reconstruction algorithms; this article reviews a cross-section, SENSE, SMASH, g-SMASH and GRAPPA, selected to demonstrate the different approaches. Theoretical (the g-factor) and practical (coil design) limits to acquisition speed are reviewed. The practical implementation of parallel imaging is also discussed, in particular coil calibration. How to recognize potential failure modes and their associated artefacts are shown. Well-established applications including angiography, cardiac imaging and applications using echo planar imaging are reviewed and we discuss what makes a good application for parallel imaging. Finally, active research areas where parallel imaging is being used to improve data quality by repairing artefacted images are also reviewed. (invited topical review)

  19. The STAPL Parallel Graph Library

    Harshvardhan,; Fidel, Adam; Amato, Nancy M.; Rauchwerger, Lawrence

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable

  20. Parallel implementation of the PHOENIX generalized stellar atmosphere program. II. Wavelength parallelization

    Baron, E.; Hauschildt, Peter H.

    1998-01-01

    We describe an important addition to the parallel implementation of our generalized nonlocal thermodynamic equilibrium (NLTE) stellar atmosphere and radiative transfer computer program PHOENIX. In a previous paper in this series we described data and task parallel algorithms we have developed for radiative transfer, spectral line opacity, and NLTE opacity and rate calculations. These algorithms divided the work spatially or by spectral lines, that is, distributing the radial zones, individual spectral lines, or characteristic rays among different processors and employ, in addition, task parallelism for logically independent functions (such as atomic and molecular line opacities). For finite, monotonic velocity fields, the radiative transfer equation is an initial value problem in wavelength, and hence each wavelength point depends upon the previous one. However, for sophisticated NLTE models of both static and moving atmospheres needed to accurately describe, e.g., novae and supernovae, the number of wavelength points is very large (200,000 - 300,000) and hence parallelization over wavelength can lead both to considerable speedup in calculation time and the ability to make use of the aggregate memory available on massively parallel supercomputers. Here, we describe an implementation of a pipelined design for the wavelength parallelization of PHOENIX, where the necessary data from the processor working on a previous wavelength point is sent to the processor working on the succeeding wavelength point as soon as it is known. Our implementation uses a MIMD design based on a relatively small number of standard message passing interface (MPI) library calls and is fully portable between serial and parallel computers. copyright 1998 The American Astronomical Society

  1. Eccentric vision : adverse interactions between line segments

    Andriessen, J.J.; Bouma, H.

    1976-01-01

    The paper deals with adverse interactions between line stimuli ineccentric vision. Bothcontrast thresholdandjust noticeable difference of slanthave been measured for a test line as a function of the distance from a number of surrounding lines. Test lines were either parallel or perpendicular to the

  2. Massively parallel multicanonical simulations

    Gross, Jonathan; Zierenberg, Johannes; Weigel, Martin; Janke, Wolfhard

    2018-03-01

    Generalized-ensemble Monte Carlo simulations such as the multicanonical method and similar techniques are among the most efficient approaches for simulations of systems undergoing discontinuous phase transitions or with rugged free-energy landscapes. As Markov chain methods, they are inherently serial computationally. It was demonstrated recently, however, that a combination of independent simulations that communicate weight updates at variable intervals allows for the efficient utilization of parallel computational resources for multicanonical simulations. Implementing this approach for the many-thread architecture provided by current generations of graphics processing units (GPUs), we show how it can be efficiently employed with of the order of 104 parallel walkers and beyond, thus constituting a versatile tool for Monte Carlo simulations in the era of massively parallel computing. We provide the fully documented source code for the approach applied to the paradigmatic example of the two-dimensional Ising model as starting point and reference for practitioners in the field.

  3. SPINning parallel systems software

    Matlin, O.S.; Lusk, E.; McCune, W.

    2002-01-01

    We describe our experiences in using Spin to verify parts of the Multi Purpose Daemon (MPD) parallel process management system. MPD is a distributed collection of processes connected by Unix network sockets. MPD is dynamic processes and connections among them are created and destroyed as MPD is initialized, runs user processes, recovers from faults, and terminates. This dynamic nature is easily expressible in the Spin/Promela framework but poses performance and scalability challenges. We present here the results of expressing some of the parallel algorithms of MPD and executing both simulation and verification runs with Spin

  4. Parallel programming with Python

    Palach, Jan

    2014-01-01

    A fast, easy-to-follow and clear tutorial to help you develop Parallel computing systems using Python. Along with explaining the fundamentals, the book will also introduce you to slightly advanced concepts and will help you in implementing these techniques in the real world. If you are an experienced Python programmer and are willing to utilize the available computing resources by parallelizing applications in a simple way, then this book is for you. You are required to have a basic knowledge of Python development to get the most of this book.

  5. Simulation Exploration through Immersive Parallel Planes: Preprint

    Brunhart-Lupo, Nicholas; Bush, Brian W.; Gruchalla, Kenny; Smith, Steve

    2016-03-01

    We present a visualization-driven simulation system that tightly couples systems dynamics simulations with an immersive virtual environment to allow analysts to rapidly develop and test hypotheses in a high-dimensional parameter space. To accomplish this, we generalize the two-dimensional parallel-coordinates statistical graphic as an immersive 'parallel-planes' visualization for multivariate time series emitted by simulations running in parallel with the visualization. In contrast to traditional parallel coordinate's mapping the multivariate dimensions onto coordinate axes represented by a series of parallel lines, we map pairs of the multivariate dimensions onto a series of parallel rectangles. As in the case of parallel coordinates, each individual observation in the dataset is mapped to a polyline whose vertices coincide with its coordinate values. Regions of the rectangles can be 'brushed' to highlight and select observations of interest: a 'slider' control allows the user to filter the observations by their time coordinate. In an immersive virtual environment, users interact with the parallel planes using a joystick that can select regions on the planes, manipulate selection, and filter time. The brushing and selection actions are used to both explore existing data as well as to launch additional simulations corresponding to the visually selected portions of the input parameter space. As soon as the new simulations complete, their resulting observations are displayed in the virtual environment. This tight feedback loop between simulation and immersive analytics accelerates users' realization of insights about the simulation and its output.

  6. Simulation Exploration through Immersive Parallel Planes

    Brunhart-Lupo, Nicholas J [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Bush, Brian W [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Gruchalla, Kenny M [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Smith, Steve [Los Alamos Visualization Associates

    2017-05-25

    We present a visualization-driven simulation system that tightly couples systems dynamics simulations with an immersive virtual environment to allow analysts to rapidly develop and test hypotheses in a high-dimensional parameter space. To accomplish this, we generalize the two-dimensional parallel-coordinates statistical graphic as an immersive 'parallel-planes' visualization for multivariate time series emitted by simulations running in parallel with the visualization. In contrast to traditional parallel coordinate's mapping the multivariate dimensions onto coordinate axes represented by a series of parallel lines, we map pairs of the multivariate dimensions onto a series of parallel rectangles. As in the case of parallel coordinates, each individual observation in the dataset is mapped to a polyline whose vertices coincide with its coordinate values. Regions of the rectangles can be 'brushed' to highlight and select observations of interest: a 'slider' control allows the user to filter the observations by their time coordinate. In an immersive virtual environment, users interact with the parallel planes using a joystick that can select regions on the planes, manipulate selection, and filter time. The brushing and selection actions are used to both explore existing data as well as to launch additional simulations corresponding to the visually selected portions of the input parameter space. As soon as the new simulations complete, their resulting observations are displayed in the virtual environment. This tight feedback loop between simulation and immersive analytics accelerates users' realization of insights about the simulation and its output.

  7. Net-Zero Building Technologies Create Substantial Energy Savings -

    only an estimated 1% of commercial buildings are built to net-zero energy criteria. One reason for this Continuum Magazine | NREL Net-Zero Building Technologies Create Substantial Energy Savings Net -Zero Building Technologies Create Substantial Energy Savings Researchers work to package and share step

  8. 19 CFR 134.35 - Articles substantially changed by manufacture.

    2010-04-01

    ... 19 Customs Duties 1 2010-04-01 2010-04-01 false Articles substantially changed by manufacture. 134... substantially changed by manufacture. (a) Articles other than goods of a NAFTA country. An article used in the United States in manufacture which results in an article having a name, character, or use differing from...

  9. Expressing Parallelism with ROOT

    Piparo, D. [CERN; Tejedor, E. [CERN; Guiraud, E. [CERN; Ganis, G. [CERN; Mato, P. [CERN; Moneta, L. [CERN; Valls Pla, X. [CERN; Canal, P. [Fermilab

    2017-11-22

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  10. Expressing Parallelism with ROOT

    Piparo, D.; Tejedor, E.; Guiraud, E.; Ganis, G.; Mato, P.; Moneta, L.; Valls Pla, X.; Canal, P.

    2017-10-01

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  11. Parallel Fast Legendre Transform

    Alves de Inda, M.; Bisseling, R.H.; Maslen, D.K.

    1998-01-01

    We discuss a parallel implementation of a fast algorithm for the discrete polynomial Legendre transform We give an introduction to the DriscollHealy algorithm using polynomial arithmetic and present experimental results on the eciency and accuracy of our implementation The algorithms were

  12. Practical parallel programming

    Bauer, Barr E

    2014-01-01

    This is the book that will teach programmers to write faster, more efficient code for parallel processors. The reader is introduced to a vast array of procedures and paradigms on which actual coding may be based. Examples and real-life simulations using these devices are presented in C and FORTRAN.

  13. Parallel hierarchical radiosity rendering

    Carter, Michael [Iowa State Univ., Ames, IA (United States)

    1993-07-01

    In this dissertation, the step-by-step development of a scalable parallel hierarchical radiosity renderer is documented. First, a new look is taken at the traditional radiosity equation, and a new form is presented in which the matrix of linear system coefficients is transformed into a symmetric matrix, thereby simplifying the problem and enabling a new solution technique to be applied. Next, the state-of-the-art hierarchical radiosity methods are examined for their suitability to parallel implementation, and scalability. Significant enhancements are also discovered which both improve their theoretical foundations and improve the images they generate. The resultant hierarchical radiosity algorithm is then examined for sources of parallelism, and for an architectural mapping. Several architectural mappings are discussed. A few key algorithmic changes are suggested during the process of making the algorithm parallel. Next, the performance, efficiency, and scalability of the algorithm are analyzed. The dissertation closes with a discussion of several ideas which have the potential to further enhance the hierarchical radiosity method, or provide an entirely new forum for the application of hierarchical methods.

  14. Parallel universes beguile science

    2007-01-01

    A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too. We may not be able -- as least not yet -- to prove they exist, many serious scientists say, but there are plenty of reasons to think that parallel dimensions are more than figments of eggheaded imagination.

  15. Parallel k-means++

    2017-04-04

    A parallelization of the k-means++ seed selection algorithm on three distinct hardware platforms: GPU, multicore CPU, and multithreaded architecture. K-means++ was developed by David Arthur and Sergei Vassilvitskii in 2007 as an extension of the k-means data clustering technique. These algorithms allow people to cluster multidimensional data, by attempting to minimize the mean distance of data points within a cluster. K-means++ improved upon traditional k-means by using a more intelligent approach to selecting the initial seeds for the clustering process. While k-means++ has become a popular alternative to traditional k-means clustering, little work has been done to parallelize this technique. We have developed original C++ code for parallelizing the algorithm on three unique hardware architectures: GPU using NVidia's CUDA/Thrust framework, multicore CPU using OpenMP, and the Cray XMT multithreaded architecture. By parallelizing the process for these platforms, we are able to perform k-means++ clustering much more quickly than it could be done before.

  16. Parallel plate detectors

    Gardes, D.; Volkov, P.

    1981-01-01

    A 5x3cm 2 (timing only) and a 15x5cm 2 (timing and position) parallel plate avalanche counters (PPAC) are considered. The theory of operation and timing resolution is given. The measurement set-up and the curves of experimental results illustrate the possibilities of the two counters [fr

  17. Parallel hierarchical global illumination

    Snell, Quinn O. [Iowa State Univ., Ames, IA (United States)

    1997-10-08

    Solving the global illumination problem is equivalent to determining the intensity of every wavelength of light in all directions at every point in a given scene. The complexity of the problem has led researchers to use approximation methods for solving the problem on serial computers. Rather than using an approximation method, such as backward ray tracing or radiosity, the authors have chosen to solve the Rendering Equation by direct simulation of light transport from the light sources. This paper presents an algorithm that solves the Rendering Equation to any desired accuracy, and can be run in parallel on distributed memory or shared memory computer systems with excellent scaling properties. It appears superior in both speed and physical correctness to recent published methods involving bidirectional ray tracing or hybrid treatments of diffuse and specular surfaces. Like progressive radiosity methods, it dynamically refines the geometry decomposition where required, but does so without the excessive storage requirements for ray histories. The algorithm, called Photon, produces a scene which converges to the global illumination solution. This amounts to a huge task for a 1997-vintage serial computer, but using the power of a parallel supercomputer significantly reduces the time required to generate a solution. Currently, Photon can be run on most parallel environments from a shared memory multiprocessor to a parallel supercomputer, as well as on clusters of heterogeneous workstations.

  18. (Nearly) portable PIC code for parallel computers

    Decyk, V.K.

    1993-01-01

    As part of the Numerical Tokamak Project, the author has developed a (nearly) portable, one dimensional version of the GCPIC algorithm for particle-in-cell codes on parallel computers. This algorithm uses a spatial domain decomposition for the fields, and passes particles from one domain to another as the particles move spatially. With only minor changes, the code has been run in parallel on the Intel Delta, the Cray C-90, the IBM ES/9000 and a cluster of workstations. After a line by line translation into cmfortran, the code was also run on the CM-200. Impressive speeds have been achieved, both on the Intel Delta and the Cray C-90, around 30 nanoseconds per particle per time step. In addition, the author was able to isolate the data management modules, so that the physics modules were not changed much from their sequential version, and the data management modules can be used as open-quotes black boxes.close quotes

  19. Mapping robust parallel multigrid algorithms to scalable memory architectures

    Overman, Andrea; Vanrosendale, John

    1993-01-01

    The convergence rate of standard multigrid algorithms degenerates on problems with stretched grids or anisotropic operators. The usual cure for this is the use of line or plane relaxation. However, multigrid algorithms based on line and plane relaxation have limited and awkward parallelism and are quite difficult to map effectively to highly parallel architectures. Newer multigrid algorithms that overcome anisotropy through the use of multiple coarse grids rather than relaxation are better suited to massively parallel architectures because they require only simple point-relaxation smoothers. In this paper, we look at the parallel implementation of a V-cycle multiple semicoarsened grid (MSG) algorithm on distributed-memory architectures such as the Intel iPSC/860 and Paragon computers. The MSG algorithms provide two levels of parallelism: parallelism within the relaxation or interpolation on each grid and across the grids on each multigrid level. Both levels of parallelism must be exploited to map these algorithms effectively to parallel architectures. This paper describes a mapping of an MSG algorithm to distributed-memory architectures that demonstrates how both levels of parallelism can be exploited. The result is a robust and effective multigrid algorithm for distributed-memory machines.

  20. Digital parallel-to-series pulse-train converter

    Hussey, J.

    1971-01-01

    Circuit converts number represented as two level signal on n-bit lines to series of pulses on one of two lines, depending on sign of number. Converter accepts parallel binary input data and produces number of output pulses equal to number represented by input data.

  1. Methodology for substantiation of the fast reactor fuel element serviceability

    Tsykanov, V.A.; Maershin, A.A.

    1988-01-01

    Methodological aspects of fast reactor fuel element serviceability substantiation are presented. The choice of the experimental program and strategies of its realization to solve the problem set in short time, taking into account available experimental means, are substantiated. Factors determining fuel element serviceability depending on parameters and operational conditions are considered. The methodological approach recommending separate studing of the factors, which points to the possibility of data acquisition, required for the development of calculational models and substantiation of fuel element serviceability in pilot and experimental reactors, is described. It is shown that the special-purpose data are more useful for the substantiation of fuel element serviceability and analytical method development than unsubstantial and expensive complex tests of fuel elements and fuel assemblies, which should be conducted only at final stages for the improvement of the structure on the whole

  2. Nuclear industry in a country with a substantial oil reserve

    Alvarez, R.; Castillo, H.; Costa, D.; Galan, I.; Martinez, M.

    1981-01-01

    The importance of the development of a nuclear industry in a country like Mexico, with a substantial oil reserve is analyzed, taking into account the technical, economical, political, ecological and social aspects of the problem. (author)

  3. Parallel grid population

    Wald, Ingo; Ize, Santiago

    2015-07-28

    Parallel population of a grid with a plurality of objects using a plurality of processors. One example embodiment is a method for parallel population of a grid with a plurality of objects using a plurality of processors. The method includes a first act of dividing a grid into n distinct grid portions, where n is the number of processors available for populating the grid. The method also includes acts of dividing a plurality of objects into n distinct sets of objects, assigning a distinct set of objects to each processor such that each processor determines by which distinct grid portion(s) each object in its distinct set of objects is at least partially bounded, and assigning a distinct grid portion to each processor such that each processor populates its distinct grid portion with any objects that were previously determined to be at least partially bounded by its distinct grid portion.

  4. Ultrascalable petaflop parallel supercomputer

    Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton On Hudson, NY; Chiu, George [Cross River, NY; Cipolla, Thomas M [Katonah, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Hall, Shawn [Pleasantville, NY; Haring, Rudolf A [Cortlandt Manor, NY; Heidelberger, Philip [Cortlandt Manor, NY; Kopcsay, Gerard V [Yorktown Heights, NY; Ohmacht, Martin [Yorktown Heights, NY; Salapura, Valentina [Chappaqua, NY; Sugavanam, Krishnan [Mahopac, NY; Takken, Todd [Brewster, NY

    2010-07-20

    A massively parallel supercomputer of petaOPS-scale includes node architectures based upon System-On-a-Chip technology, where each processing node comprises a single Application Specific Integrated Circuit (ASIC) having up to four processing elements. The ASIC nodes are interconnected by multiple independent networks that optimally maximize the throughput of packet communications between nodes with minimal latency. The multiple networks may include three high-speed networks for parallel algorithm message passing including a Torus, collective network, and a Global Asynchronous network that provides global barrier and notification functions. These multiple independent networks may be collaboratively or independently utilized according to the needs or phases of an algorithm for optimizing algorithm processing performance. The use of a DMA engine is provided to facilitate message passing among the nodes without the expenditure of processing resources at the node.

  5. More parallel please

    Gregersen, Frans; Josephson, Olle; Kristoffersen, Gjert

    of departure that English may be used in parallel with the various local, in this case Nordic, languages. As such, the book integrates the challenge of internationalization faced by any university with the wish to improve quality in research, education and administration based on the local language......Abstract [en] More parallel, please is the result of the work of an Inter-Nordic group of experts on language policy financed by the Nordic Council of Ministers 2014-17. The book presents all that is needed to plan, practice and revise a university language policy which takes as its point......(s). There are three layers in the text: First, you may read the extremely brief version of the in total 11 recommendations for best practice. Second, you may acquaint yourself with the extended version of the recommendations and finally, you may study the reasoning behind each of them. At the end of the text, we give...

  6. PARALLEL MOVING MECHANICAL SYSTEMS

    Florian Ion Tiberius Petrescu

    2014-09-01

    Full Text Available Normal 0 false false false EN-US X-NONE X-NONE MicrosoftInternetExplorer4 Moving mechanical systems parallel structures are solid, fast, and accurate. Between parallel systems it is to be noticed Stewart platforms, as the oldest systems, fast, solid and precise. The work outlines a few main elements of Stewart platforms. Begin with the geometry platform, kinematic elements of it, and presented then and a few items of dynamics. Dynamic primary element on it means the determination mechanism kinetic energy of the entire Stewart platforms. It is then in a record tail cinematic mobile by a method dot matrix of rotation. If a structural mottoelement consists of two moving elements which translates relative, drive train and especially dynamic it is more convenient to represent the mottoelement as a single moving components. We have thus seven moving parts (the six motoelements or feet to which is added mobile platform 7 and one fixed.

  7. Xyce parallel electronic simulator.

    Keiter, Eric R; Mei, Ting; Russo, Thomas V.; Rankin, Eric Lamont; Schiek, Richard Louis; Thornquist, Heidi K.; Fixel, Deborah A.; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R.

    2010-05-01

    This document is a reference guide to the Xyce Parallel Electronic Simulator, and is a companion document to the Xyce Users Guide. The focus of this document is (to the extent possible) exhaustively list device parameters, solver options, parser options, and other usage details of Xyce. This document is not intended to be a tutorial. Users who are new to circuit simulation are better served by the Xyce Users Guide.

  8. Stability of parallel flows

    Betchov, R

    2012-01-01

    Stability of Parallel Flows provides information pertinent to hydrodynamical stability. This book explores the stability problems that occur in various fields, including electronics, mechanics, oceanography, administration, economics, as well as naval and aeronautical engineering. Organized into two parts encompassing 10 chapters, this book starts with an overview of the general equations of a two-dimensional incompressible flow. This text then explores the stability of a laminar boundary layer and presents the equation of the inviscid approximation. Other chapters present the general equation

  9. Algorithmically specialized parallel computers

    Snyder, Lawrence; Gannon, Dennis B

    1985-01-01

    Algorithmically Specialized Parallel Computers focuses on the concept and characteristics of an algorithmically specialized computer.This book discusses the algorithmically specialized computers, algorithmic specialization using VLSI, and innovative architectures. The architectures and algorithms for digital signal, speech, and image processing and specialized architectures for numerical computations are also elaborated. Other topics include the model for analyzing generalized inter-processor, pipelined architecture for search tree maintenance, and specialized computer organization for raster

  10. The ongoing investigation of high performance parallel computing in HEP

    Peach, Kenneth J; Böck, R K; Dobinson, Robert W; Hansroul, M; Norton, Alan Robert; Willers, Ian Malcolm; Baud, J P; Carminati, F; Gagliardi, F; McIntosh, E; Metcalf, M; Robertson, L; CERN. Geneva. Detector Research and Development Committee

    1993-01-01

    Past and current exploitation of parallel computing in High Energy Physics is summarized and a list of R & D projects in this area is presented. The applicability of new parallel hardware and software to physics problems is investigated, in the light of the requirements for computing power of LHC experiments and the current trends in the computer industry. Four main themes are discussed (possibilities for a finer grain of parallelism; fine-grain communication mechanism; usable parallel programming environment; different programming models and architectures, using standard commercial products). Parallel computing technology is potentially of interest for offline and vital for real time applications in LHC. A substantial investment in applications development and evaluation of state of the art hardware and software products is needed. A solid development environment is required at an early stage, before mainline LHC program development begins.

  11. Parallel GPU implementation of iterative PCA algorithms.

    Andrecut, M

    2009-11-01

    Principal component analysis (PCA) is a key statistical technique for multivariate data analysis. For large data sets, the common approach to PCA computation is based on the standard NIPALS-PCA algorithm, which unfortunately suffers from loss of orthogonality, and therefore its applicability is usually limited to the estimation of the first few components. Here we present an algorithm based on Gram-Schmidt orthogonalization (called GS-PCA), which eliminates this shortcoming of NIPALS-PCA. Also, we discuss the GPU (Graphics Processing Unit) parallel implementation of both NIPALS-PCA and GS-PCA algorithms. The numerical results show that the GPU parallel optimized versions, based on CUBLAS (NVIDIA), are substantially faster (up to 12 times) than the CPU optimized versions based on CBLAS (GNU Scientific Library).

  12. Automated Parallel Capillary Electrophoretic System

    Li, Qingbo; Kane, Thomas E.; Liu, Changsheng; Sonnenschein, Bernard; Sharer, Michael V.; Kernan, John R.

    2000-02-22

    An automated electrophoretic system is disclosed. The system employs a capillary cartridge having a plurality of capillary tubes. The cartridge has a first array of capillary ends projecting from one side of a plate. The first array of capillary ends are spaced apart in substantially the same manner as the wells of a microtitre tray of standard size. This allows one to simultaneously perform capillary electrophoresis on samples present in each of the wells of the tray. The system includes a stacked, dual carousel arrangement to eliminate cross-contamination resulting from reuse of the same buffer tray on consecutive executions from electrophoresis. The system also has a gel delivery module containing a gel syringe/a stepper motor or a high pressure chamber with a pump to quickly and uniformly deliver gel through the capillary tubes. The system further includes a multi-wavelength beam generator to generate a laser beam which produces a beam with a wide range of wavelengths. An off-line capillary reconditioner thoroughly cleans a capillary cartridge to enable simultaneous execution of electrophoresis with another capillary cartridge. The streamlined nature of the off-line capillary reconditioner offers the advantage of increased system throughput with a minimal increase in system cost.

  13. Strong contributions from vertical triads to helix-partner preferences in parallel coiled coils.

    Steinkruger, Jay D; Bartlett, Gail J; Woolfson, Derek N; Gellman, Samuel H

    2012-09-26

    Pairing preferences in heterodimeric coiled coils are determined by complementarities among side chains that pack against one another at the helix-helix interface. However, relationships between dimer stability and interfacial residue identity are not fully understood. In the context of the "knobs-into-holes" (KIH) packing pattern, one can identify two classes of interactions between side chains from different helices: "lateral", in which a line connecting the adjacent side chains is perpendicular to the helix axes, and "vertical", in which the connecting line is parallel to the helix axes. We have previously analyzed vertical interactions in antiparallel coiled coils and found that one type of triad constellation (a'-a-a') exerts a strong effect on pairing preferences, while the other type of triad (d'-d-d') has relatively little impact on pairing tendencies. Here, we ask whether vertical interactions (d'-a-d') influence pairing in parallel coiled-coil dimers. Our results indicate that vertical interactions can exert a substantial impact on pairing specificity, and that the influence of the d'-a-d' triad depends on the lateral a' contact within the local KIH motif. Structure-informed bioinformatic analyses of protein sequences reveal trends consistent with the thermodynamic data derived from our experimental model system in suggesting that heterotriads involving Leu and Ile are preferred over homotriads involving Leu and Ile.

  14. Resistor Combinations for Parallel Circuits.

    McTernan, James P.

    1978-01-01

    To help simplify both teaching and learning of parallel circuits, a high school electricity/electronics teacher presents and illustrates the use of tables of values for parallel resistive circuits in which total resistances are whole numbers. (MF)

  15. SOFTWARE FOR DESIGNING PARALLEL APPLICATIONS

    M. K. Bouza

    2017-01-01

    Full Text Available The object of research is the tools to support the development of parallel programs in C/C ++. The methods and software which automates the process of designing parallel applications are proposed.

  16. Parallel constraint satisfaction in memory-based decisions.

    Glöckner, Andreas; Hodges, Sara D

    2011-01-01

    Three studies sought to investigate decision strategies in memory-based decisions and to test the predictions of the parallel constraint satisfaction (PCS) model for decision making (Glöckner & Betsch, 2008). Time pressure was manipulated and the model was compared against simple heuristics (take the best and equal weight) and a weighted additive strategy. From PCS we predicted that fast intuitive decision making is based on compensatory information integration and that decision time increases and confidence decreases with increasing inconsistency in the decision task. In line with these predictions we observed a predominant usage of compensatory strategies under all time-pressure conditions and even with decision times as short as 1.7 s. For a substantial number of participants, choices and decision times were best explained by PCS, but there was also evidence for use of simple heuristics. The time-pressure manipulation did not significantly affect decision strategies. Overall, the results highlight intuitive, automatic processes in decision making and support the idea that human information-processing capabilities are less severely bounded than often assumed.

  17. Parallelized Seeded Region Growing Using CUDA

    Seongjin Park

    2014-01-01

    Full Text Available This paper presents a novel method for parallelizing the seeded region growing (SRG algorithm using Compute Unified Device Architecture (CUDA technology, with intention to overcome the theoretical weakness of SRG algorithm of its computation time being directly proportional to the size of a segmented region. The segmentation performance of the proposed CUDA-based SRG is compared with SRG implementations on single-core CPUs, quad-core CPUs, and shader language programming, using synthetic datasets and 20 body CT scans. Based on the experimental results, the CUDA-based SRG outperforms the other three implementations, advocating that it can substantially assist the segmentation during massive CT screening tests.

  18. Illustrative Line Styles for Flow Visualization

    Everts, Maarten H.; Bekker, Hendrik; Roerdink, Jos B. T. M.; Isenberg, Tobias

    2011-01-01

    We present a flexible illustrative line style model for the visualization of streamline data. Our model partitions view-oriented line strips into parallel bands whose basic visual properties can be controlled independently. We thus extend previous line stylization techniques specifically for

  19. Flow Visualization using Illustrative Line Styles

    Everts, Maarten H.; Bekker, Hendrik; Roerdink, Jos B. T. M.; Isenberg, Tobias; Bekker, Paulus

    2011-01-01

    We present a flexible illustrative line style model for the visualization of streamline data. Our model partitions view- oriented line strips into parallel bands whose basic visual properties can be controlled independently. We thus extend previous line stylization techniques specifically for

  20. Parallel Implementation of the Multi-Dimensional Spectral Code SPECT3D on large 3D grids.

    Golovkin, Igor E.; Macfarlane, Joseph J.; Woodruff, Pamela R.; Pereyra, Nicolas A.

    2006-10-01

    The multi-dimensional collisional-radiative, spectral analysis code SPECT3D can be used to study radiation from complex plasmas. SPECT3D can generate instantaneous and time-gated images and spectra, space-resolved and streaked spectra, which makes it a valuable tool for post-processing hydrodynamics calculations and direct comparison between simulations and experimental data. On large three dimensional grids, transporting radiation along lines of sight (LOS) requires substantial memory and CPU resources. Currently, the parallel option in SPECT3D is based on parallelization over photon frequencies and allows for a nearly linear speed-up for a variety of problems. In addition, we are introducing a new parallel mechanism that will greatly reduce memory requirements. In the new implementation, spatial domain decomposition will be utilized allowing transport along a LOS to be performed only on the mesh cells the LOS crosses. The ability to operate on a fraction of the grid is crucial for post-processing the results of large-scale three-dimensional hydrodynamics simulations. We will present a parallel implementation of the code and provide a scalability study performed on a Linux cluster.

  1. Parallel External Memory Graph Algorithms

    Arge, Lars Allan; Goodrich, Michael T.; Sitchinava, Nodari

    2010-01-01

    In this paper, we study parallel I/O efficient graph algorithms in the Parallel External Memory (PEM) model, one o f the private-cache chip multiprocessor (CMP) models. We study the fundamental problem of list ranking which leads to efficient solutions to problems on trees, such as computing lowest...... an optimal speedup of ¿(P) in parallel I/O complexity and parallel computation time, compared to the single-processor external memory counterparts....

  2. Parallel inter channel interaction mechanisms

    Jovic, V.; Afgan, N.; Jovic, L.

    1995-01-01

    Parallel channels interactions are examined. For experimental researches of nonstationary regimes flow in three parallel vertical channels results of phenomenon analysis and mechanisms of parallel channel interaction for adiabatic condition of one-phase fluid and two-phase mixture flow are shown. (author)

  3. Massively Parallel QCD

    Soltz, R; Vranas, P; Blumrich, M; Chen, D; Gara, A; Giampap, M; Heidelberger, P; Salapura, V; Sexton, J; Bhanot, G

    2007-01-01

    The theory of the strong nuclear force, Quantum Chromodynamics (QCD), can be numerically simulated from first principles on massively-parallel supercomputers using the method of Lattice Gauge Theory. We describe the special programming requirements of lattice QCD (LQCD) as well as the optimal supercomputer hardware architectures that it suggests. We demonstrate these methods on the BlueGene massively-parallel supercomputer and argue that LQCD and the BlueGene architecture are a natural match. This can be traced to the simple fact that LQCD is a regular lattice discretization of space into lattice sites while the BlueGene supercomputer is a discretization of space into compute nodes, and that both are constrained by requirements of locality. This simple relation is both technologically important and theoretically intriguing. The main result of this paper is the speedup of LQCD using up to 131,072 CPUs on the largest BlueGene/L supercomputer. The speedup is perfect with sustained performance of about 20% of peak. This corresponds to a maximum of 70.5 sustained TFlop/s. At these speeds LQCD and BlueGene are poised to produce the next generation of strong interaction physics theoretical results

  4. A Parallel Butterfly Algorithm

    Poulson, Jack; Demanet, Laurent; Maxwell, Nicholas; Ying, Lexing

    2014-01-01

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  5. A Parallel Butterfly Algorithm

    Poulson, Jack

    2014-02-04

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  6. A parallel line sieve for the GNFS Algorithm

    Sameh Daoud; Ibrahim Gad

    2014-01-01

    RSA is one of the most important public key cryptosystems for information security. The security of RSA depends on Integer factorization problem, it relies on the difficulty of factoring large integers. Much research has gone into problem of factoring a large number. Due to advances in factoring algorithms and advances in computing hardware the size of the number that can be factorized increases exponentially year by year. The General Number Field Sieve algorithm (GNFS) is currently the best ...

  7. SESOTHO trial (“Switch Either near Suppression Or THOusand”) – switch to second-line versus WHO-guided standard of care for unsuppressed patients on first-line ART with viremia below 1000 copies/mL: protocol of a multicenter, parallel-group, open-label, randomized clinical trial in Lesotho, Southern Africa

    Amstutz, Alain; Nsakala, Bienvenu Lengo; Vanobberghen, Fiona; Muhairwe, Josephine; Glass, Tracy Renée; Achieng, Beatrice; Sepeka, Mamorena; Tlali, Katleho; Sao, Lebohang; Thin, Kyaw; Klimkait, Thomas; Battegay, Manuel; Labhardt, Niklaus Daniel

    2018-01-01

    Background The World Health Organization (WHO) recommends viral load (VL) measurement as the preferred monitoring strategy for HIV-infected individuals on antiretroviral therapy (ART) in resource-limited settings. The new WHO guidelines 2016 continue to define virologic failure as two consecutive VL ≥1000 copies/mL (at least 3 months apart) despite good adherence, triggering switch to second-line therapy. However, the threshold of 1000 copies/mL for defining virologic failure is based on low-...

  8. SUBSTANTIATION OF SOLUTIONS PERTAINING TO COMPLEX PRODUCTION RECONSTRUCTION

    V. Y. Gurinovich

    2011-01-01

    Full Text Available While taking an example of reconstruction and modernization of OJSC «Construction and Mounting Trust No.16, Novopolotsk» the paper substantiates solutions for complex production reconstruction. The production reconstruction is divided in three start-up facilities ensuring continuous finished-product output.

  9. 26 CFR 1.528-4 - Substantiality test.

    2010-04-01

    ... 26 Internal Revenue 7 2010-04-01 2010-04-01 true Substantiality test. 1.528-4 Section 1.528-4 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED... residence. Units which are used for purposes auxiliary to residential use (such as laundry areas, swimming...

  10. Engineering Substantially Prolonged Human Lifespans: Biotechnological Enhancement and Ethics

    Derkx, P.H.J.M.

    2006-01-01

    Substantial extension of the human lifespan has recently become a subject of lively debate. One reason for this is the completion in 2001 of the Human Genome Project and the experimental avenues for biogerontological research it has opened. Another is recent theoretical progress in biogerontology.

  11. 76 FR 63846 - Substantially Underserved Trust Areas (SUTA)

    2011-10-14

    ... Cost Rural Communities; 10.861, Public Television Station Digital Transition Grant Program; 10.862... rule do not impose substantial unreimbursed direct compliance costs on Indian tribal, Alaska native, or native Hawaiian governments and sovereign institutions or have tribal implications that preempt tribal...

  12. Xyce parallel electronic simulator : reference guide.

    Mei, Ting; Rankin, Eric Lamont; Thornquist, Heidi K.; Santarelli, Keith R.; Fixel, Deborah A.; Coffey, Todd Stirling; Russo, Thomas V.; Schiek, Richard Louis; Warrender, Christina E.; Keiter, Eric Richard; Pawlowski, Roger Patrick

    2011-05-01

    This document is a reference guide to the Xyce Parallel Electronic Simulator, and is a companion document to the Xyce Users Guide. The focus of this document is (to the extent possible) exhaustively list device parameters, solver options, parser options, and other usage details of Xyce. This document is not intended to be a tutorial. Users who are new to circuit simulation are better served by the Xyce Users Guide. The Xyce Parallel Electronic Simulator has been written to support, in a rigorous manner, the simulation needs of the Sandia National Laboratories electrical designers. It is targeted specifically to run on large-scale parallel computing platforms but also runs well on a variety of architectures including single processor workstations. It also aims to support a variety of devices and models specific to Sandia needs. This document is intended to complement the Xyce Users Guide. It contains comprehensive, detailed information about a number of topics pertinent to the usage of Xyce. Included in this document is a netlist reference for the input-file commands and elements supported within Xyce; a command line reference, which describes the available command line arguments for Xyce; and quick-references for users of other circuit codes, such as Orcad's PSpice and Sandia's ChileSPICE.

  13. Parallel Computing in SCALE

    DeHart, Mark D.; Williams, Mark L.; Bowman, Stephen M.

    2010-01-01

    The SCALE computational architecture has remained basically the same since its inception 30 years ago, although constituent modules and capabilities have changed significantly. This SCALE concept was intended to provide a framework whereby independent codes can be linked to provide a more comprehensive capability than possible with the individual programs - allowing flexibility to address a wide variety of applications. However, the current system was designed originally for mainframe computers with a single CPU and with significantly less memory than today's personal computers. It has been recognized that the present SCALE computation system could be restructured to take advantage of modern hardware and software capabilities, while retaining many of the modular features of the present system. Preliminary work is being done to define specifications and capabilities for a more advanced computational architecture. This paper describes the state of current SCALE development activities and plans for future development. With the release of SCALE 6.1 in 2010, a new phase of evolutionary development will be available to SCALE users within the TRITON and NEWT modules. The SCALE (Standardized Computer Analyses for Licensing Evaluation) code system developed by Oak Ridge National Laboratory (ORNL) provides a comprehensive and integrated package of codes and nuclear data for a wide range of applications in criticality safety, reactor physics, shielding, isotopic depletion and decay, and sensitivity/uncertainty (S/U) analysis. Over the last three years, since the release of version 5.1 in 2006, several important new codes have been introduced within SCALE, and significant advances applied to existing codes. Many of these new features became available with the release of SCALE 6.0 in early 2009. However, beginning with SCALE 6.1, a first generation of parallel computing is being introduced. In addition to near-term improvements, a plan for longer term SCALE enhancement

  14. Parallel Polarization State Generation.

    She, Alan; Capasso, Federico

    2016-05-17

    The control of polarization, an essential property of light, is of wide scientific and technological interest. The general problem of generating arbitrary time-varying states of polarization (SOP) has always been mathematically formulated by a series of linear transformations, i.e. a product of matrices, imposing a serial architecture. Here we show a parallel architecture described by a sum of matrices. The theory is experimentally demonstrated by modulating spatially-separated polarization components of a laser using a digital micromirror device that are subsequently beam combined. This method greatly expands the parameter space for engineering devices that control polarization. Consequently, performance characteristics, such as speed, stability, and spectral range, are entirely dictated by the technologies of optical intensity modulation, including absorption, reflection, emission, and scattering. This opens up important prospects for polarization state generation (PSG) with unique performance characteristics with applications in spectroscopic ellipsometry, spectropolarimetry, communications, imaging, and security.

  15. Parallel imaging microfluidic cytometer.

    Ehrlich, Daniel J; McKenna, Brian K; Evans, James G; Belkina, Anna C; Denis, Gerald V; Sherr, David H; Cheung, Man Ching

    2011-01-01

    By adding an additional degree of freedom from multichannel flow, the parallel microfluidic cytometer (PMC) combines some of the best features of fluorescence-activated flow cytometry (FCM) and microscope-based high-content screening (HCS). The PMC (i) lends itself to fast processing of large numbers of samples, (ii) adds a 1D imaging capability for intracellular localization assays (HCS), (iii) has a high rare-cell sensitivity, and (iv) has an unusual capability for time-synchronized sampling. An inability to practically handle large sample numbers has restricted applications of conventional flow cytometers and microscopes in combinatorial cell assays, network biology, and drug discovery. The PMC promises to relieve a bottleneck in these previously constrained applications. The PMC may also be a powerful tool for finding rare primary cells in the clinic. The multichannel architecture of current PMC prototypes allows 384 unique samples for a cell-based screen to be read out in ∼6-10 min, about 30 times the speed of most current FCM systems. In 1D intracellular imaging, the PMC can obtain protein localization using HCS marker strategies at many times for the sample throughput of charge-coupled device (CCD)-based microscopes or CCD-based single-channel flow cytometers. The PMC also permits the signal integration time to be varied over a larger range than is practical in conventional flow cytometers. The signal-to-noise advantages are useful, for example, in counting rare positive cells in the most difficult early stages of genome-wide screening. We review the status of parallel microfluidic cytometry and discuss some of the directions the new technology may take. Copyright © 2011 Elsevier Inc. All rights reserved.

  16. Improved magnetic field line design for TMX

    Logan, B.G.; Baldwin, D.E.; Foote, J.H.; Chargin, A.K.; Hinkle, R.E.; Hussung, R.O.; Damm, C.C.

    1977-01-01

    Optimization of the currents in the TMX magnet set leads to a field line configuration which has a central solenoidal region uniform in parallel B parallel to within 10 percent over a 2m length. The field design has sufficient flexibility to meet all three physics objectives of the TMX experiment

  17. EXPERIMENTAL SUBSTANTIATION OF ATTRIBUTIONAL STYLE IN TRAFFIC POLICE OFFICERS’ BEHAVIOR

    Olga Vladimirovna Merkusheva

    2016-02-01

    Full Text Available The article prepared for the ‘Psychological and Methodological Aspects of Professional Personnel Training at Internal Affairs Bodies of the Russian Federation’ scholar tradition looks at the applied aspect of handling security issues by traffic police officers. It presents the results of empirical study of the employees’ psychological characteristics conducted to substantiate their attributional style, which determines the specific character as well as the qua-lity and reliability of traffic police officers’ performance. Traffic police officers’ attributional style of behavior is researched with the help of attributional cognitive techniques for hazard phenomena detection as well as social perception and attribution technique and is viewed as the most stable entity, which reflects the employee’s personal and professional focus.Goal: to provide empirical substantiation of attributional style in traffic police officers’ behavior as an integral personality characteristic that ensures timely detection of hazard phenomena.

  18. About Parallel Programming: Paradigms, Parallel Execution and Collaborative Systems

    Loredana MOCEAN

    2009-01-01

    Full Text Available In the last years, there were made efforts for delineation of a stabile and unitary frame, where the problems of logical parallel processing must find solutions at least at the level of imperative languages. The results obtained by now are not at the level of the made efforts. This paper wants to be a little contribution at these efforts. We propose an overview in parallel programming, parallel execution and collaborative systems.

  19. Substantiation of the road toll for heavy transport vehicles

    Burmaka, N.; Chernykh, A.

    2010-01-01

    The existing and possible additional sources of developing state and local road funds of Ukraine have been considered. The formula for calculating monthly road toll for heavy transport vehicles has been proposed. This formula includes the payment rate per every kilometer of distance, the vehicle capacity utilization factor and the run with the load. The payment rate per every kilometer of distance for transport vehicles depending on the allowed total weight has been substantiated. The given r...

  20. Patients' substantialization of disease, the hybrid symptom and metaphysical care.

    Pârvan, Alexandra

    2015-06-01

    In the context of current scholarship concerned with facilitating integration between the biomedical and the patient-centred models of care, the article suggests that disease brings about an ontological disruption in patients, which is not directly addressed in either model, and may interfere with treatment and therapy outcomes if not met with a type of care termed here as 'metaphysical'. The receipt of diagnosis and medical care can give patients the sense that they are ontologically diminished, or less of a human, and along with physicians' approaches to and discourses about disease, may prompt them to seek ontological restoration or security in the same way as psychologically traumatized patients sometimes do: by treating the disease and/or the experience of harm associated with it as a thing that exists per se. I call this 'substantialization' of disease (or harm) and draw on Augustine's theory of non-substantial deficiencies (physiological and moral) and on Plato's and Plotinus's different takes on such defects in order to discuss what substantialization can do for patients. Based on literature that examines patients' ways of talking about and living with their disease, I speculate that substantialization can generate a 'hybrid symptom', consisting in patterns of exercising agency which may predispose to non-adherence. Ways in which physicians could provide metaphysical care are proposed, along with an understanding of chronic patients as hybrid ontological and agentic units, which draws on theories of enactive cognition. I opine that metaphysical care may facilitate integration between the depersonalized and personalized models of care. © 2014 John Wiley & Sons, Ltd.

  1. A multi-objective approach to the assignment of stock keeping units to unidirectional picking lines

    Le Roux, G. J.

    2017-05-01

    Full Text Available An order picking system in a distribution centre consisting of parallel unidirectional picking lines is considered. The objectives are to minimise the walking distance of the pickers, the largest volume of stock on a picking line over all picking lines, the number of small packages, and the total penalty incurred for late distributions. The problem is formulated as a multi-objective multiple knapsack problem that is not solvable in a realistic time. Population-based algorithms, including the artificial bee colony algorithm and the genetic algorithm, are also implemented. The results obtained from all algorithms indicate a substantial improvement on all objectives relative to historical assignments. The genetic algorithm delivers the best performance.

  2. Massive hybrid parallelism for fully implicit multiphysics

    Gaston, D. R.; Permann, C. J.; Andrs, D.; Peterson, J. W.

    2013-01-01

    As hardware advances continue to modify the supercomputing landscape, traditional scientific software development practices will become more outdated, ineffective, and inefficient. The process of rewriting/retooling existing software for new architectures is a Sisyphean task, and results in substantial hours of development time, effort, and money. Software libraries which provide an abstraction of the resources provided by such architectures are therefore essential if the computational engineering and science communities are to continue to flourish in this modern computing environment. The Multiphysics Object Oriented Simulation Environment (MOOSE) framework enables complex multiphysics analysis tools to be built rapidly by scientists, engineers, and domain specialists, while also allowing them to both take advantage of current HPC architectures, and efficiently prepare for future supercomputer designs. MOOSE employs a hybrid shared-memory and distributed-memory parallel model and provides a complete and consistent interface for creating multiphysics analysis tools. In this paper, a brief discussion of the mathematical algorithms underlying the framework and the internal object-oriented hybrid parallel design are given. Representative massively parallel results from several applications areas are presented, and a brief discussion of future areas of research for the framework are provided. (authors)

  3. Massive hybrid parallelism for fully implicit multiphysics

    Gaston, D. R.; Permann, C. J.; Andrs, D.; Peterson, J. W. [Idaho National Laboratory, 2525 N. Fremont Ave., Idaho Falls, ID 83415 (United States)

    2013-07-01

    As hardware advances continue to modify the supercomputing landscape, traditional scientific software development practices will become more outdated, ineffective, and inefficient. The process of rewriting/retooling existing software for new architectures is a Sisyphean task, and results in substantial hours of development time, effort, and money. Software libraries which provide an abstraction of the resources provided by such architectures are therefore essential if the computational engineering and science communities are to continue to flourish in this modern computing environment. The Multiphysics Object Oriented Simulation Environment (MOOSE) framework enables complex multiphysics analysis tools to be built rapidly by scientists, engineers, and domain specialists, while also allowing them to both take advantage of current HPC architectures, and efficiently prepare for future supercomputer designs. MOOSE employs a hybrid shared-memory and distributed-memory parallel model and provides a complete and consistent interface for creating multiphysics analysis tools. In this paper, a brief discussion of the mathematical algorithms underlying the framework and the internal object-oriented hybrid parallel design are given. Representative massively parallel results from several applications areas are presented, and a brief discussion of future areas of research for the framework are provided. (authors)

  4. MASSIVE HYBRID PARALLELISM FOR FULLY IMPLICIT MULTIPHYSICS

    Cody J. Permann; David Andrs; John W. Peterson; Derek R. Gaston

    2013-05-01

    As hardware advances continue to modify the supercomputing landscape, traditional scientific software development practices will become more outdated, ineffective, and inefficient. The process of rewriting/retooling existing software for new architectures is a Sisyphean task, and results in substantial hours of development time, effort, and money. Software libraries which provide an abstraction of the resources provided by such architectures are therefore essential if the computational engineering and science communities are to continue to flourish in this modern computing environment. The Multiphysics Object Oriented Simulation Environment (MOOSE) framework enables complex multiphysics analysis tools to be built rapidly by scientists, engineers, and domain specialists, while also allowing them to both take advantage of current HPC architectures, and efficiently prepare for future supercomputer designs. MOOSE employs a hybrid shared-memory and distributed-memory parallel model and provides a complete and consistent interface for creating multiphysics analysis tools. In this paper, a brief discussion of the mathematical algorithms underlying the framework and the internal object-oriented hybrid parallel design are given. Representative massively parallel results from several applications areas are presented, and a brief discussion of future areas of research for the framework are provided.

  5. Parallel Framework for Cooperative Processes

    Mitică Craus

    2005-01-01

    Full Text Available This paper describes the work of an object oriented framework designed to be used in the parallelization of a set of related algorithms. The idea behind the system we are describing is to have a re-usable framework for running several sequential algorithms in a parallel environment. The algorithms that the framework can be used with have several things in common: they have to run in cycles and the work should be possible to be split between several "processing units". The parallel framework uses the message-passing communication paradigm and is organized as a master-slave system. Two applications are presented: an Ant Colony Optimization (ACO parallel algorithm for the Travelling Salesman Problem (TSP and an Image Processing (IP parallel algorithm for the Symmetrical Neighborhood Filter (SNF. The implementations of these applications by means of the parallel framework prove to have good performances: approximatively linear speedup and low communication cost.

  6. Juvenile prison in parallel legislation

    Lutovac Mitar

    2016-01-01

    Full Text Available The need for punishment of juveniles occurred from the time when there was no clear line separating them from the adult criminal population. At the same time, the evolution of the juvenile punishment is not in itself involve substantial changes to their criminal status. On the contrary, the status of minors in society did not show serious differences regarding the status of young adults, as well as the adult elderly. On the other hand, on the ground of their punishment is recorded deviations that go in the direction of application of mild corporal punishment. Closing the minor was performed in a physically separate parts of the general penal institutions with the use of a lower degree of restrictions while serving juvenile prison. Due to the different treatment of minors during the evolution of their criminal status leads to their different treatment in comparative law. That is why we are witnessing the existence of numerous differences in the juvenile punishment in some countries in the world. On the European continent there is a wide range of different legal solutions when it comes to punishing juveniles. There are considerable differences in the procedure pronouncing juvenile prison and in particular penal treatment of juveniles in penitentiary institutions. For these reasons, the author has decided to show the basic statutory provisions in the part that relates to the issue of punishment of minors in the legislation of individual countries.

  7. Parallel Monte Carlo reactor neutronics

    Blomquist, R.N.; Brown, F.B.

    1994-01-01

    The issues affecting implementation of parallel algorithms for large-scale engineering Monte Carlo neutron transport simulations are discussed. For nuclear reactor calculations, these include load balancing, recoding effort, reproducibility, domain decomposition techniques, I/O minimization, and strategies for different parallel architectures. Two codes were parallelized and tested for performance. The architectures employed include SIMD, MIMD-distributed memory, and workstation network with uneven interactive load. Speedups linear with the number of nodes were achieved

  8. Anti-parallel triplexes

    Kosbar, Tamer R.; Sofan, Mamdouh A.; Waly, Mohamed A.

    2015-01-01

    about 6.1 °C when the TFO strand was modified with Z and the Watson-Crick strand with adenine-LNA (AL). The molecular modeling results showed that, in case of nucleobases Y and Z a hydrogen bond (1.69 and 1.72 Å, respectively) was formed between the protonated 3-aminopropyn-1-yl chain and one...... of the phosphate groups in Watson-Crick strand. Also, it was shown that the nucleobase Y made a good stacking and binding with the other nucleobases in the TFO and Watson-Crick duplex, respectively. In contrast, the nucleobase Z with LNA moiety was forced to twist out of plane of Watson-Crick base pair which......The phosphoramidites of DNA monomers of 7-(3-aminopropyn-1-yl)-8-aza-7-deazaadenine (Y) and 7-(3-aminopropyn-1-yl)-8-aza-7-deazaadenine LNA (Z) are synthesized, and the thermal stability at pH 7.2 and 8.2 of anti-parallel triplexes modified with these two monomers is determined. When, the anti...

  9. Parallel consensual neural networks.

    Benediktsson, J A; Sveinsson, J R; Ersoy, O K; Swain, P H

    1997-01-01

    A new type of a neural-network architecture, the parallel consensual neural network (PCNN), is introduced and applied in classification/data fusion of multisource remote sensing and geographic data. The PCNN architecture is based on statistical consensus theory and involves using stage neural networks with transformed input data. The input data are transformed several times and the different transformed data are used as if they were independent inputs. The independent inputs are first classified using the stage neural networks. The output responses from the stage networks are then weighted and combined to make a consensual decision. In this paper, optimization methods are used in order to weight the outputs from the stage networks. Two approaches are proposed to compute the data transforms for the PCNN, one for binary data and another for analog data. The analog approach uses wavelet packets. The experimental results obtained with the proposed approach show that the PCNN outperforms both a conjugate-gradient backpropagation neural network and conventional statistical methods in terms of overall classification accuracy of test data.

  10. A Parallel Particle Swarm Optimizer

    Schutte, J. F; Fregly, B .J; Haftka, R. T; George, A. D

    2003-01-01

    .... Motivated by a computationally demanding biomechanical system identification problem, we introduce a parallel implementation of a stochastic population based global optimizer, the Particle Swarm...

  11. Patterns for Parallel Software Design

    Ortega-Arjona, Jorge Luis

    2010-01-01

    Essential reading to understand patterns for parallel programming Software patterns have revolutionized the way we think about how software is designed, built, and documented, and the design of parallel software requires you to consider other particular design aspects and special skills. From clusters to supercomputers, success heavily depends on the design skills of software developers. Patterns for Parallel Software Design presents a pattern-oriented software architecture approach to parallel software design. This approach is not a design method in the classic sense, but a new way of managin

  12. Seeing or moving in parallel

    Christensen, Mark Schram; Ehrsson, H Henrik; Nielsen, Jens Bo

    2013-01-01

    a different network, involving bilateral dorsal premotor cortex (PMd), primary motor cortex, and SMA, was more active when subjects viewed parallel movements while performing either symmetrical or parallel movements. Correlations between behavioral instability and brain activity were present in right lateral...... adduction-abduction movements symmetrically or in parallel with real-time congruent or incongruent visual feedback of the movements. One network, consisting of bilateral superior and middle frontal gyrus and supplementary motor area (SMA), was more active when subjects performed parallel movements, whereas...

  13. Energy flow of electric dipole radiation in between parallel mirrors

    Xu, Zhangjin; Arnoldus, Henk F.

    2017-11-01

    We have studied the energy flow patterns of the radiation emitted by an electric dipole located in between parallel mirrors. It appears that the field lines of the Poynting vector (the flow lines of energy) can have very intricate structures, including many singularities and vortices. The flow line patterns depend on the distance between the mirrors, the distance of the dipole to one of the mirrors and the angle of oscillation of the dipole moment with respect to the normal of the mirror surfaces. Already for the simplest case of a dipole moment oscillating perpendicular to the mirrors, singularities appear at regular intervals along the direction of propagation (parallel to the mirrors). For a parallel dipole, vortices appear in the neighbourhood of the dipole. For a dipole oscillating under a finite angle with the surface normal, the radiating tends to swirl around the dipole before travelling off parallel to the mirrors. For relatively large mirror separations, vortices appear in the pattern. When the dipole is off-centred with respect to the midway point between the mirrors, the flow line structure becomes even more complicated, with numerous vortices in the pattern, and tiny loops near the dipole. We have also investigated the locations of the vortices and singularities, and these can be found without any specific knowledge about the flow lines. This provides an independent means of studying the propagation of dipole radiation between mirrors.

  14. Substantial increase in acceleration potential of pyroelectric crystals

    Tornow, W.; Lynam, S. M.; Shafroth, S. M.

    2010-01-01

    We report on a substantial increase in the acceleration potential achieved with a LiTaO 3 pyroelectric crystal. With a single 2.5 cm diameter and 2.5 cm long z-cut crystal without electric field-enhancing nanotip we produced positive ion beams with maximal energies between 300 and 310 keV during the cooling phase when the crystal was exposed to 5 mTorr of deuterium gas. These values are about a factor of 2 larger than previously obtained with single pyroelectric crystals.

  15. COMPOSITIONAL AND SUBSTANTIAL STRUCTURE OF THE MEDICAL DOCUMENT: FORMATION STAGES

    Romashova Olga Vladimirovna

    2015-03-01

    Full Text Available The article deals with the compositional and substantial structure of the ambulatory medical record, or "case history", which has being formed for a long time. The author allocates the three main periods in the formation of this medical document: the first period (the beginning of the 19th century – 1920s is connected with the origin and formation; the second period (1920-1980s is marked by emergence of the normative legal acts regulating registration and maintaining; the third period (1980s – up to the present is associated with the cancellation of regulations and the introduction of the new order of the Ministry of Health of the USSR that changed the document's form and name. It is determined that the composition of the case history consists of the title page and the main part. The following processes take place in the course of ambulatory medical record's formation: strengthening formalization, increase in the number of pattern text fragments, increase in the text's volume, and the implementation of bigger number of functions. The author reveals the main (informative and cumulative, accounting and additional (scientific, controlling, legal, financial functions of the document. The implementation of these functions is reflected in the compositional and substantial structure of the document text and is conditioned by a number of extralinguistic factors.

  16. CS-Studio Scan System Parallelization

    Kasemir, Kay [ORNL; Pearson, Matthew R [ORNL

    2015-01-01

    For several years, the Control System Studio (CS-Studio) Scan System has successfully automated the operation of beam lines at the Oak Ridge National Laboratory (ORNL) High Flux Isotope Reactor (HFIR) and Spallation Neutron Source (SNS). As it is applied to additional beam lines, we need to support simultaneous adjustments of temperatures or motor positions. While this can be implemented via virtual motors or similar logic inside the Experimental Physics and Industrial Control System (EPICS) Input/Output Controllers (IOCs), doing so requires a priori knowledge of experimenters requirements. By adding support for the parallel control of multiple process variables (PVs) to the Scan System, we can better support ad hoc automation of experiments that benefit from such simultaneous PV adjustments.

  17. Boosted protease inhibitor monotherapy versus boosted protease inhibitor plus lamivudine dual therapy as second-line maintenance treatment for HIV-1-infected patients in sub-Saharan Africa (ANRS12 286/MOBIDIP): a multicentre, randomised, parallel, open-label, superiority trial.

    Ciaffi, Laura; Koulla-Shiro, Sinata; Sawadogo, Adrien Bruno; Ndour, Cheik Tidiane; Eymard-Duvernay, Sabrina; Mbouyap, Pretty Rosereine; Ayangma, Liliane; Zoungrana, Jacques; Gueye, Ndeye Fatou Ngom; Diallo, Mohamadou; Izard, Suzanne; Bado, Guillaume; Kane, Coumba Toure; Aghokeng, Avelin Fobang; Peeters, Martine; Girard, Pierre Marie; Le Moing, Vincent; Reynes, Jacques; Delaporte, Eric

    2017-09-01

    Despite satisfactory efficacy of WHO-recommended second-line antiretroviral treatment for patients with HIV in low-income countries, the need for simplified, low-cost, and less-toxic maintenance strategies remains high. We compared boosted protease inhibitor monotherapy with dual therapy with boosted protease inhibitor plus lamivudine in patients on second-line antiretrovial therapy (ART). We did a multicentre, randomised, parallel, open-label, superiority, trial in the HIV services of five hospitals in sub-Saharan Africa (Yaoundé, Cameroon; Dakar, Senegal; and Bobo Dioulasso, Burkina Faso). We recruited patients from the long-term, post-trial cohort of the ANRS 12169/2LADY study that compared the efficacy of three second-line combinations based on boosted protease inhibitors. Participants for our study were HIV-1 infected with multiple mutations including M184V, at first-line failure, aged 18 years and older, on boosted protease inhibitor plus two nucleoside reverse transcriptase inhibitors (NRTI) for at least 48 weeks with at least 48 weeks follow-up in the 2LADY trial, with two viral load measurements of less than 200 copies per mL in the previous 6 months, CD4 counts of more than 100 cells per μL, adherence of at least 90%, and no change to ART in the past 3 months. We randomly assigned participants (1:1) to receive either monotherapy with their boosted protease inhibitor (once-daily darunavir 800 mg [two 400 mg tablets] boosted with ritonavir 100 mg [one tablet] or coformulation of lopinavir 200 mg with ritonavir 50 mg [two tablets taken twice per day]) or to boosted protease inhibitor plus once-daily lamivudine 300 mg (one 300 mg tablet or two 150 mg tablets). Computer-generated randomisation was stratified by study site and viral load at screening (treatment allocation was not masked from clinicians or patients]. Patients had follow-up visits at weeks 4 and 12, and every 3 months until 96 weeks; if viral load exceeded 500 copies per mL at any visit, NRTI

  18. PARALLEL IMPORT: REALITY FOR RUSSIA

    Т. А. Сухопарова

    2014-01-01

    Full Text Available Problem of parallel import is urgent question at now. Parallel import legalization in Russia is expedient. Such statement based on opposite experts opinion analysis. At the same time it’s necessary to negative consequences consider of this decision and to apply remedies to its minimization.Purchase on Elibrary.ru > Buy now

  19. The Galley Parallel File System

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    Most current multiprocessor file systems are designed to use multiple disks in parallel, using the high aggregate bandwidth to meet the growing I/0 requirements of parallel scientific applications. Many multiprocessor file systems provide applications with a conventional Unix-like interface, allowing the application to access multiple disks transparently. This interface conceals the parallelism within the file system, increasing the ease of programmability, but making it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. In addition to providing an insufficient interface, most current multiprocessor file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic scientific multiprocessor workloads. We discuss Galley's file structure and application interface, as well as the performance advantages offered by that interface.

  20. Parallelization of the FLAPW method

    Canning, A.; Mannstadt, W.; Freeman, A.J.

    1999-01-01

    The FLAPW (full-potential linearized-augmented plane-wave) method is one of the most accurate first-principles methods for determining electronic and magnetic properties of crystals and surfaces. Until the present work, the FLAPW method has been limited to systems of less than about one hundred atoms due to a lack of an efficient parallel implementation to exploit the power and memory of parallel computers. In this work we present an efficient parallelization of the method by division among the processors of the plane-wave components for each state. The code is also optimized for RISC (reduced instruction set computer) architectures, such as those found on most parallel computers, making full use of BLAS (basic linear algebra subprograms) wherever possible. Scaling results are presented for systems of up to 686 silicon atoms and 343 palladium atoms per unit cell, running on up to 512 processors on a CRAY T3E parallel computer

  1. Parallelization of the FLAPW method

    Canning, A.; Mannstadt, W.; Freeman, A. J.

    2000-08-01

    The FLAPW (full-potential linearized-augmented plane-wave) method is one of the most accurate first-principles methods for determining structural, electronic and magnetic properties of crystals and surfaces. Until the present work, the FLAPW method has been limited to systems of less than about a hundred atoms due to the lack of an efficient parallel implementation to exploit the power and memory of parallel computers. In this work, we present an efficient parallelization of the method by division among the processors of the plane-wave components for each state. The code is also optimized for RISC (reduced instruction set computer) architectures, such as those found on most parallel computers, making full use of BLAS (basic linear algebra subprograms) wherever possible. Scaling results are presented for systems of up to 686 silicon atoms and 343 palladium atoms per unit cell, running on up to 512 processors on a CRAY T3E parallel supercomputer.

  2. Vectorization and parallelization of a production reactor assembly code

    Vujic, J.L.; Martin, W.R.; Michigan Univ., Ann Arbor, MI

    1991-01-01

    In order to use efficiently the new features of supercomputers, production codes, usually written 10 -20 years ago, must be tailored for modern computer architectures. We have chosen to optimize the CPM-2 code, a production reactor assembly code based on the collision probability transport method. Substantial speedup in the execution times was obtained with the parallel/vector version of the CPM-2 code. In addition, we have developed a new transfer probability method, which removes some of the modelling limitations of the collision probability method encoded in the CPM-2 code, and can fully utilize the parallel/vector architecture of a multiprocessor IBM 3090. (author)

  3. PIMR: Parallel and Integrated Matching for Raw Data.

    Li, Zhenghao; Yang, Junying; Zhao, Jiaduo; Han, Peng; Chai, Zhi

    2016-01-02

    With the trend of high-resolution imaging, computational costs of image matching have substantially increased. In order to find the compromise between accuracy and computation in real-time applications, we bring forward a fast and robust matching algorithm, named parallel and integrated matching for raw data (PIMR). This algorithm not only effectively utilizes the color information of raw data, but also designs a parallel and integrated framework to shorten the time-cost in the demosaicing stage. Experiments show that compared to existing state-of-the-art methods, the proposed algorithm yields a comparable recognition rate, while the total time-cost of imaging and matching is significantly reduced.

  4. Vectorization and parallelization of a production reactor assembly code

    Vujic, J.L.; Martin, W.R.

    1991-01-01

    In order to efficiently use new features of supercomputers, production codes, usually written 10 - 20 years ago, must be tailored for modern computer architectures. We have chosen to optimize the CPM-2 code, a production reactor assembly code based on the collision probability transport method. Substantial speedups in the execution times were obtained with the parallel/vector version of the CPM-2 code. In addition, we have developed a new transfer probability method, which removes some of the modelling limitations of the collision probability method encoded in the CPM-2 code, and can fully utilize parallel/vector architecture of a multiprocessor IBM 3090. (author)

  5. Clozapine-induced dysphagia with secondary substantial weight loss.

    Osman, Mugtaba; Devadas, Vekneswaran

    2016-08-19

    Dysphagia is listed as a 'rare' side effect following clozapine treatment. In this case report, we describe how significant clozapine-induced dysphagia has led to significant reduction of nutritional intake with subsequent substantial weight loss. An 18-year-old single man with an established diagnosis of treatment-resistant paranoid schizophrenia recovered well on a therapeutic dose of clozapine. However, he was noted to lose weight significantly (up to 20% of his original weight) as the dose was uptitrated. This was brought about by development of dysphagia, likely to be due to clozapine. Addition of nutritional supplementary liquids and initiation of a modified behavioural dietary/swallowing programme, while repeatedly mastering the Mendelsohn manoeuvre technique, alleviated the swallowing difficulties and restored his weight. 2016 BMJ Publishing Group Ltd.

  6. Monte Carlo reactor calculation with substantially reduced number of cycles

    Lee, M. J.; Joo, H. G.; Lee, D.; Smith, K.

    2012-01-01

    A new Monte Carlo (MC) eigenvalue calculation scheme that substantially reduces the number of cycles is introduced with the aid of coarse mesh finite difference (CMFD) formulation. First, it is confirmed in terms of pin power errors that using extremely many particles resulting in short active cycles is beneficial even in the conventional MC scheme although wasted operations in inactive cycles cannot be reduced with more particles. A CMFD-assisted MC scheme is introduced as an effort to reduce the number of inactive cycles and the fast convergence behavior and reduced inter-cycle effect of the CMFD assisted MC calculation is investigated in detail. As a practical means of providing a good initial fission source distribution, an assembly based few-group condensation and homogenization scheme is introduced and it is shown that efficient MC eigenvalue calculations with fewer than 20 total cycles (including inactive cycles) are possible for large power reactor problems. (authors)

  7. Is Monte Carlo embarrassingly parallel?

    Hoogenboom, J. E. [Delft Univ. of Technology, Mekelweg 15, 2629 JB Delft (Netherlands); Delft Nuclear Consultancy, IJsselzoom 2, 2902 LB Capelle aan den IJssel (Netherlands)

    2012-07-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  8. Is Monte Carlo embarrassingly parallel?

    Hoogenboom, J. E.

    2012-01-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  9. Parallel integer sorting with medium and fine-scale parallelism

    Dagum, Leonardo

    1993-01-01

    Two new parallel integer sorting algorithms, queue-sort and barrel-sort, are presented and analyzed in detail. These algorithms do not have optimal parallel complexity, yet they show very good performance in practice. Queue-sort designed for fine-scale parallel architectures which allow the queueing of multiple messages to the same destination. Barrel-sort is designed for medium-scale parallel architectures with a high message passing overhead. The performance results from the implementation of queue-sort on a Connection Machine CM-2 and barrel-sort on a 128 processor iPSC/860 are given. The two implementations are found to be comparable in performance but not as good as a fully vectorized bucket sort on the Cray YMP.

  10. Template based parallel checkpointing in a massively parallel computer system

    Archer, Charles Jens [Rochester, MN; Inglett, Todd Alan [Rochester, MN

    2009-01-13

    A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.

  11. Parallel education: what is it?

    Amos, Michelle Peta

    2017-01-01

    In the history of education it has long been discussed that single-sex and coeducation are the two models of education present in schools. With the introduction of parallel schools over the last 15 years, there has been very little research into this 'new model'. Many people do not understand what it means for a school to be parallel or they confuse a parallel model with co-education, due to the presence of both boys and girls within the one institution. Therefore, the main obj...

  12. Balanced, parallel operation of flashlamps

    Carder, B.M.; Merritt, B.T.

    1979-01-01

    A new energy store, the Compensated Pulsed Alternator (CPA), promises to be a cost effective substitute for capacitors to drive flashlamps that pump large Nd:glass lasers. Because the CPA is large and discrete, it will be necessary that it drive many parallel flashlamp circuits, presenting a problem in equal current distribution. Current division to +- 20% between parallel flashlamps has been achieved, but this is marginal for laser pumping. A method is presented here that provides equal current sharing to about 1%, and it includes fused protection against short circuit faults. The method was tested with eight parallel circuits, including both open-circuit and short-circuit fault tests

  13. Parallelization of the model-based iterative reconstruction algorithm DIRA

    Oertenberg, A.; Sandborg, M.; Alm Carlsson, G.; Malusek, A.; Magnusson, M.

    2016-01-01

    New paradigms for parallel programming have been devised to simplify software development on multi-core processors and many-core graphical processing units (GPU). Despite their obvious benefits, the parallelization of existing computer programs is not an easy task. In this work, the use of the Open Multiprocessing (OpenMP) and Open Computing Language (OpenCL) frameworks is considered for the parallelization of the model-based iterative reconstruction algorithm DIRA with the aim to significantly shorten the code's execution time. Selected routines were parallelized using OpenMP and OpenCL libraries; some routines were converted from MATLAB to C and optimised. Parallelization of the code with the OpenMP was easy and resulted in an overall speedup of 15 on a 16-core computer. Parallelization with OpenCL was more difficult owing to differences between the central processing unit and GPU architectures. The resulting speedup was substantially lower than the theoretical peak performance of the GPU; the cause was explained. (authors)

  14. High Line

    Kiib, Hans

    2015-01-01

    At just over 10 meters above street level, the High Line extends three kilometers through three districts of Southwestern Manhattan in New York. It consists of simple steel construction, and previously served as an elevated rail line connection between Penn Station on 34th Street and the many....... The High Line project has been carried out as part of an open conversion strategy. The result is a remarkable urban architectural project, which works as a catalyst for the urban development of Western Manhattan. The greater project includes the restoration and reuse of many old industrial buildings...

  15. Complementary sex determination substantially increases extinction proneness of haplodiploid populations.

    Zayed, Amro; Packer, Laurence

    2005-07-26

    The role of genetic factors in extinction is firmly established for diploid organisms, but haplodiploids have been considered immune to genetic load impacts because deleterious alleles are readily purged in haploid males. However, we show that single-locus complementary sex determination ancestral to the haplodiploid Hymenoptera (ants, bees, and wasps) imposes a substantial genetic load through homozygosity at the sex locus that results in the production of inviable or sterile diploid males. Using stochastic modeling, we have discovered that diploid male production (DMP) can initiate a rapid and previously uncharacterized extinction vortex. The extinction rate in haplodiploid populations with DMP is an order of magnitude greater than in its absence under realistic but conservative demographic parameter values. Furthermore, DMP alone can elevate the base extinction risk in haplodiploids by over an order of magnitude higher than that caused by inbreeding depression in threatened diploids. Thus, contrary to previous expectations, haplodiploids are more, rather than less, prone to extinction for genetic reasons. Our findings necessitate a fundamental shift in approaches to the conservation and population biology of these ecologically and economically crucial insects.

  16. Substantial soil organic carbon retention along floodplains of mountain streams

    Sutfin, Nicholas A.; Wohl, Ellen

    2017-07-01

    Small, snowmelt-dominated mountain streams have the potential to store substantial organic carbon in floodplain sediment because of high inputs of particulate organic matter, relatively lower temperatures compared with lowland regions, and potential for increased moisture conditions. This work (i) quantifies mean soil organic carbon (OC) content along 24 study reaches in the Colorado Rocky Mountains using 660 soil samples, (ii) identifies potential controls of OC content based on soil properties and spatial position with respect to the channel, and (iii) and examines soil properties and OC across various floodplain geomorphic features in the study area. Stepwise multiple linear regression (adjusted r2 = 0.48, p sample depth, percent sand, distance from the channel, and relative elevation from the channel are significant predictors of OC content in the study area. Principle component analysis indicates limited separation between geomorphic floodplain features based on predictors of OC content. A lack of significant differences among floodplain features suggests that the systematic random sampling employed in this study can capture the variability of OC across floodplains in the study area. Mean floodplain OC (6.3 ± 0.3%) is more variable but on average greater than values in uplands (1.5 ± 0.08% to 2.2 ± 0.14%) of the Colorado Front Range and higher than published values from floodplains in other regions, particularly those of larger rivers.

  17. Substantial proportion of global streamflow less than three months old

    Jasechko, Scott; Kirchner, James W.; Welker, Jeffrey M.; McDonnell, Jeffrey J.

    2016-02-01

    Biogeochemical cycles, contaminant transport and chemical weathering are regulated by the speed at which precipitation travels through landscapes and reaches streams. Streamflow is a mixture of young and old precipitation, but the global proportions of these young and old components are not known. Here we analyse seasonal cycles of oxygen isotope ratios in rain, snow and streamflow compiled from 254 watersheds around the world, and calculate the fraction of streamflow that is derived from precipitation that fell within the past two or three months. This young streamflow accounts for about a third of global river discharge, and comprises at least 5% of discharge in about 90% of the catchments we investigated. We conclude that, although typical catchments have mean transit times of years or even decades, they nonetheless can rapidly transmit substantial fractions of soluble contaminant inputs to streams. Young streamflow is less prevalent in steeper landscapes, which suggests they are characterized by deeper vertical infiltration. Because young streamflow is derived from less than 0.1% of global groundwater storage, we conclude that this thin veneer of aquifer storage will have a disproportionate influence on stream water quality.

  18. Substantially Evolutionary Theorizing in Designing Software-Intensive Systems

    Petr Sosnin

    2018-04-01

    Full Text Available Useful inheritances from scientific experience open perspective ways for increasing the degree of success in designing of systems with software. One such way is a search and build applied theory that takes into account the nature of design and the specificity of software engineering. This paper presents a substantially evolutionary approach to creating the project theories, the application of which leads to positive effects that are traditionally expected from theorizing. Any implementation of the approach is based on a reflection by designers of an operational space of designing onto a semantic memory of a question-answer type. One of the results of such reflection is a system of question-answer nets, the nodes of which register facts of interactions of designers with accessible experience. A set of such facts is used by designers for creating and using the theory that belongs to the new subclass of Grounded Theories. This sub-class is oriented on organizationally behavioral features of a project’s work based on design thinking, automated mental imagination, and thought experimenting that facilitate increasing the degree of controlled intellectualization in the design process and, correspondingly, increasing the degree of success in the development of software-intensive systems.

  19. Parallel asynchronous systems and image processing algorithms

    Coon, D. D.; Perera, A. G. U.

    1989-01-01

    A new hardware approach to implementation of image processing algorithms is described. The approach is based on silicon devices which would permit an independent analog processing channel to be dedicated to evey pixel. A laminar architecture consisting of a stack of planar arrays of the device would form a two-dimensional array processor with a 2-D array of inputs located directly behind a focal plane detector array. A 2-D image data stream would propagate in neuronlike asynchronous pulse coded form through the laminar processor. Such systems would integrate image acquisition and image processing. Acquisition and processing would be performed concurrently as in natural vision systems. The research is aimed at implementation of algorithms, such as the intensity dependent summation algorithm and pyramid processing structures, which are motivated by the operation of natural vision systems. Implementation of natural vision algorithms would benefit from the use of neuronlike information coding and the laminar, 2-D parallel, vision system type architecture. Besides providing a neural network framework for implementation of natural vision algorithms, a 2-D parallel approach could eliminate the serial bottleneck of conventional processing systems. Conversion to serial format would occur only after raw intensity data has been substantially processed. An interesting challenge arises from the fact that the mathematical formulation of natural vision algorithms does not specify the means of implementation, so that hardware implementation poses intriguing questions involving vision science.

  20. Workspace Analysis for Parallel Robot

    Ying Sun

    2013-05-01

    Full Text Available As a completely new-type of robot, the parallel robot possesses a lot of advantages that the serial robot does not, such as high rigidity, great load-carrying capacity, small error, high precision, small self-weight/load ratio, good dynamic behavior and easy control, hence its range is extended in using domain. In order to find workspace of parallel mechanism, the numerical boundary-searching algorithm based on the reverse solution of kinematics and limitation of link length has been introduced. This paper analyses position workspace, orientation workspace of parallel robot of the six degrees of freedom. The result shows: It is a main means to increase and decrease its workspace to change the length of branch of parallel mechanism; The radius of the movement platform has no effect on the size of workspace, but will change position of workspace.

  1. "Feeling" Series and Parallel Resistances.

    Morse, Robert A.

    1993-01-01

    Equipped with drinking straws and stirring straws, a teacher can help students understand how resistances in electric circuits combine in series and in parallel. Follow-up suggestions are provided. (ZWH)

  2. Parallel encoders for pixel detectors

    Nikityuk, N.M.

    1991-01-01

    A new method of fast encoding and determining the multiplicity and coordinates of fired pixels is described. A specific example construction of parallel encodes and MCC for n=49 and t=2 is given. 16 refs.; 6 figs.; 2 tabs

  3. Massively Parallel Finite Element Programming

    Heister, Timo

    2010-01-01

    Today\\'s large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  4. Event monitoring of parallel computations

    Gruzlikov Alexander M.

    2015-06-01

    Full Text Available The paper considers the monitoring of parallel computations for detection of abnormal events. It is assumed that computations are organized according to an event model, and monitoring is based on specific test sequences

  5. Massively Parallel Finite Element Programming

    Heister, Timo; Kronbichler, Martin; Bangerth, Wolfgang

    2010-01-01

    Today's large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  6. The STAPL Parallel Graph Library

    Harshvardhan,

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable distributed graph container and a collection of commonly used parallel graph algorithms. The library introduces pGraph pViews that separate algorithm design from the container implementation. It supports three graph processing algorithmic paradigms, level-synchronous, asynchronous and coarse-grained, and provides common graph algorithms based on them. Experimental results demonstrate improved scalability in performance and data size over existing graph libraries on more than 16,000 cores and on internet-scale graphs containing over 16 billion vertices and 250 billion edges. © Springer-Verlag Berlin Heidelberg 2013.

  7. Writing parallel programs that work

    CERN. Geneva

    2012-01-01

    Serial algorithms typically run inefficiently on parallel machines. This may sound like an obvious statement, but it is the root cause of why parallel programming is considered to be difficult. The current state of the computer industry is still that almost all programs in existence are serial. This talk will describe the techniques used in the Intel Parallel Studio to provide a developer with the tools necessary to understand the behaviors and limitations of the existing serial programs. Once the limitations are known the developer can refactor the algorithms and reanalyze the resulting programs with the tools in the Intel Parallel Studio to create parallel programs that work. About the speaker Paul Petersen is a Sr. Principal Engineer in the Software and Solutions Group (SSG) at Intel. He received a Ph.D. degree in Computer Science from the University of Illinois in 1993. After UIUC, he was employed at Kuck and Associates, Inc. (KAI) working on auto-parallelizing compiler (KAP), and was involved in th...

  8. Exploiting Symmetry on Parallel Architectures.

    Stiller, Lewis Benjamin

    1995-01-01

    This thesis describes techniques for the design of parallel programs that solve well-structured problems with inherent symmetry. Part I demonstrates the reduction of such problems to generalized matrix multiplication by a group-equivariant matrix. Fast techniques for this multiplication are described, including factorization, orbit decomposition, and Fourier transforms over finite groups. Our algorithms entail interaction between two symmetry groups: one arising at the software level from the problem's symmetry and the other arising at the hardware level from the processors' communication network. Part II illustrates the applicability of our symmetry -exploitation techniques by presenting a series of case studies of the design and implementation of parallel programs. First, a parallel program that solves chess endgames by factorization of an associated dihedral group-equivariant matrix is described. This code runs faster than previous serial programs, and discovered it a number of results. Second, parallel algorithms for Fourier transforms for finite groups are developed, and preliminary parallel implementations for group transforms of dihedral and of symmetric groups are described. Applications in learning, vision, pattern recognition, and statistics are proposed. Third, parallel implementations solving several computational science problems are described, including the direct n-body problem, convolutions arising from molecular biology, and some communication primitives such as broadcast and reduce. Some of our implementations ran orders of magnitude faster than previous techniques, and were used in the investigation of various physical phenomena.

  9. Parallel algorithms for continuum dynamics

    Hicks, D.L.; Liebrock, L.M.

    1987-01-01

    Simply porting existing parallel programs to a new parallel processor may not achieve the full speedup possible; to achieve the maximum efficiency may require redesigning the parallel algorithms for the specific architecture. The authors discuss here parallel algorithms that were developed first for the HEP processor and then ported to the CRAY X-MP/4, the ELXSI/10, and the Intel iPSC/32. Focus is mainly on the most recent parallel processing results produced, i.e., those on the Intel Hypercube. The applications are simulations of continuum dynamics in which the momentum and stress gradients are important. Examples of these are inertial confinement fusion experiments, severe breaks in the coolant system of a reactor, weapons physics, shock-wave physics. Speedup efficiencies on the Intel iPSC Hypercube are very sensitive to the ratio of communication to computation. Great care must be taken in designing algorithms for this machine to avoid global communication. This is much more critical on the iPSC than it was on the three previous parallel processors

  10. World lines.

    Waser Jürgen; Fuchs Raphael; Ribicic Hrvoje; Schindler Benjamin; Blöschl Günther; Gröller Eduard

    2010-01-01

    In this paper we present World Lines as a novel interactive visualization that provides complete control over multiple heterogeneous simulation runs. In many application areas decisions can only be made by exploring alternative scenarios. The goal of the suggested approach is to support users in this decision making process. In this setting the data domain is extended to a set of alternative worlds where only one outcome will actually happen. World Lines integrate simulation visualization and...

  11. New theory of arrhythmia. Conceptual substantiation of arrhythmia mechanisms

    Vladimir I. Ermoshkin

    2015-11-01

    Full Text Available Aims A new attempt is made to substantiate the concept of the mechanism of arrhythmia and sudden cardiac death. Methods The paper is based on a theoretical analysis of special literature, personal experience of participation in conferences and discussions with leading Russian cardiologists. Results We have succeeded in demonstrating the fact that researchers ignore the fact that cardiomyocytes can be excited by mechanical pulses, when considering the arrhythmia mechanisms. We have conducted trials using the Cardiocode device. Under stress in a human, opened may be large and small arteriovenous anastomoses, via which blood under high pressure is ejected into veins. It leads to pressure surges in arteries and veins. The vena cava dilates, its tonus increases. In some cases, the pulse waves travel via anastomoses along the vena cava walls to the atria and the ventricles. An above-threshold concentration of tensions from mechanical pulses may excite cardiomyocytes from different points of the myocardium, disturbing the sinus rhythm. As a result, extrasystoles, tachycardia attacks, blocking of blood circulation in the peripheral segments of the venous arterial networks, edemata, thrombosis and metabolism disorders appear. Arrhythmia, tachycardia attacks and concomitant myocardial ischemia lead to progression of heart fibrosis. Such changes increase the probability of fibrillations and sudden cardiac death. Conclusion Unhealthy lifestyle, the presence of opening and not properly closing anastomoses may provoke a number of diseases. To avoid the cardiac arrhythmia attacks and prevent SCD, it is necessary to suppress travel of the mechanical waves within the following circuitry: aorta – artery – anastomosis – vein – vena cava – atria – ventricles. The travel of the mechanical waves within the same vessel circuitry explains the fact that the fixed couplings under extrasystoles are observed, and the beat-to-beat RR intervals under tachycardia

  12. How common is substantial weight gain after pregnancy?

    Sumithran, Priya; Houlihan, Christine; Shub, Alexis; Churilov, Leonid; Pritchard, Natasha; Price, Sarah; Ekinci, Elif; Proietto, Joseph; Permezel, Michael

    2017-11-20

    Although population-based studies indicate that on average, women gain 1-2kg between pregnancies, women with obesity often attribute its development to childbearing. There is little contemporary data available regarding how commonly this occurs, particularly in women of different body mass index (BMI) categories. The aim of this study was to examine inter-pregnancy weight changes among women at a tertiary obstetric hospital in Melbourne, Australia. This was a retrospective review of data from the Birthing Outcomes System electronic record of 19,617 women aged 20 years or older, who delivered at least two consecutive singleton infants at ≥37 weeks' gestation at Mercy Hospital for Women between December 1994 and December 2015. A logistic regression model was used to assess the relationship between gain of ≥4kg/m 2 between pregnancies and maternal BMI category in the first pregnancy, adjusting for covariates of maternal age, inter-pregnancy interval, and socioeconomic status. Gain of ≥4kg/m 2 between the first two pregnancies occurred in 7.5% of normal weight women, 10.5% of overweight women, and 13.4% of women with obesity. One in five women who were normal weight in their first pregnancy increased to overweight or obese BMI categories in their second pregnancy. Substantial weight gain in relation to pregnancy affects a considerable proportion of women. Since inter-pregnancy weight gain is associated with several complications in the next pregnancy and longer term, avoiding excessive weight gain during and between pregnancies may prevent adverse health consequences in mothers and offspring. Copyright © 2017 Asia Oceania Association for the Study of Obesity. Published by Elsevier Ltd. All rights reserved.

  13. Three-dimensional parallel vortex rings in Bose-Einstein condensates

    Crasovan, Lucian-Cornel; Perez-Garcia, Victor M.; Danaila, Ionut; Mihalache, Dumitru; Torner, Lluis

    2004-01-01

    We construct three-dimensional structures of topological defects hosted in trapped wave fields, in the form of vortex stars, vortex cages, parallel vortex lines, perpendicular vortex rings, and parallel vortex rings, and we show that the latter exist as robust stationary, collective states of nonrotating Bose-Einstein condensates. We discuss the stability properties of excited states containing several parallel vortex rings hosted by the condensate, including their dynamical and structural stability

  14. Parallel heat transport in integrable and chaotic magnetic fields

    Castillo-Negrete, D. del; Chacon, L. [Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831-8071 (United States)

    2012-05-15

    The study of transport in magnetized plasmas is a problem of fundamental interest in controlled fusion, space plasmas, and astrophysics research. Three issues make this problem particularly challenging: (i) The extreme anisotropy between the parallel (i.e., along the magnetic field), {chi}{sub ||} , and the perpendicular, {chi}{sub Up-Tack }, conductivities ({chi}{sub ||} /{chi}{sub Up-Tack} may exceed 10{sup 10} in fusion plasmas); (ii) Nonlocal parallel transport in the limit of small collisionality; and (iii) Magnetic field lines chaos which in general complicates (and may preclude) the construction of magnetic field line coordinates. Motivated by these issues, we present a Lagrangian Green's function method to solve the local and non-local parallel transport equation applicable to integrable and chaotic magnetic fields in arbitrary geometry. The method avoids by construction the numerical pollution issues of grid-based algorithms. The potential of the approach is demonstrated with nontrivial applications to integrable (magnetic island), weakly chaotic (Devil's staircase), and fully chaotic magnetic field configurations. For the latter, numerical solutions of the parallel heat transport equation show that the effective radial transport, with local and non-local parallel closures, is non-diffusive, thus casting doubts on the applicability of quasilinear diffusion descriptions. General conditions for the existence of non-diffusive, multivalued flux-gradient relations in the temperature evolution are derived.

  15. Endpoint-based parallel data processing in a parallel active messaging interface of a parallel computer

    Archer, Charles J.; Blocksome, Michael A.; Ratterman, Joseph D.; Smith, Brian E.

    2014-08-12

    Endpoint-based parallel data processing in a parallel active messaging interface (`PAMI`) of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective operation through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.

  16. Parallelization Issues and Particle-In Codes.

    Elster, Anne Cathrine

    1994-01-01

    "Everything should be made as simple as possible, but not simpler." Albert Einstein. The field of parallel scientific computing has concentrated on parallelization of individual modules such as matrix solvers and factorizers. However, many applications involve several interacting modules. Our analyses of a particle-in-cell code modeling charged particles in an electric field, show that these accompanying dependencies affect data partitioning and lead to new parallelization strategies concerning processor, memory and cache utilization. Our test-bed, a KSR1, is a distributed memory machine with a globally shared addressing space. However, most of the new methods presented hold generally for hierarchical and/or distributed memory systems. We introduce a novel approach that uses dual pointers on the local particle arrays to keep the particle locations automatically partially sorted. Complexity and performance analyses with accompanying KSR benchmarks, have been included for both this scheme and for the traditional replicated grids approach. The latter approach maintains load-balance with respect to particles. However, our results demonstrate it fails to scale properly for problems with large grids (say, greater than 128-by-128) running on as few as 15 KSR nodes, since the extra storage and computation time associated with adding the grid copies, becomes significant. Our grid partitioning scheme, although harder to implement, does not need to replicate the whole grid. Consequently, it scales well for large problems on highly parallel systems. It may, however, require load balancing schemes for non-uniform particle distributions. Our dual pointer approach may facilitate this through dynamically partitioned grids. We also introduce hierarchical data structures that store neighboring grid-points within the same cache -line by reordering the grid indexing. This alignment produces a 25% savings in cache-hits for a 4-by-4 cache. A consideration of the input data's effect on

  17. Parallel Implicit Algorithms for CFD

    Keyes, David E.

    1998-01-01

    The main goal of this project was efficient distributed parallel and workstation cluster implementations of Newton-Krylov-Schwarz (NKS) solvers for implicit Computational Fluid Dynamics (CFD.) "Newton" refers to a quadratically convergent nonlinear iteration using gradient information based on the true residual, "Krylov" to an inner linear iteration that accesses the Jacobian matrix only through highly parallelizable sparse matrix-vector products, and "Schwarz" to a domain decomposition form of preconditioning the inner Krylov iterations with primarily neighbor-only exchange of data between the processors. Prior experience has established that Newton-Krylov methods are competitive solvers in the CFD context and that Krylov-Schwarz methods port well to distributed memory computers. The combination of the techniques into Newton-Krylov-Schwarz was implemented on 2D and 3D unstructured Euler codes on the parallel testbeds that used to be at LaRC and on several other parallel computers operated by other agencies or made available by the vendors. Early implementations were made directly in Massively Parallel Integration (MPI) with parallel solvers we adapted from legacy NASA codes and enhanced for full NKS functionality. Later implementations were made in the framework of the PETSC library from Argonne National Laboratory, which now includes pseudo-transient continuation Newton-Krylov-Schwarz solver capability (as a result of demands we made upon PETSC during our early porting experiences). A secondary project pursued with funding from this contract was parallel implicit solvers in acoustics, specifically in the Helmholtz formulation. A 2D acoustic inverse problem has been solved in parallel within the PETSC framework.

  18. Second derivative parallel block backward differentiation type ...

    Second derivative parallel block backward differentiation type formulas for Stiff ODEs. ... Log in or Register to get access to full text downloads. ... and the methods are inherently parallel and can be distributed over parallel processors. They are ...

  19. A Parallel Approach to Fractal Image Compression

    Lubomir Dedera

    2004-01-01

    The paper deals with a parallel approach to coding and decoding algorithms in fractal image compressionand presents experimental results comparing sequential and parallel algorithms from the point of view of achieved bothcoding and decoding time and effectiveness of parallelization.

  20. A novel two-level dynamic parallel data scheme for large 3-D SN calculations

    Sjoden, G.E.; Shedlock, D.; Haghighat, A.; Yi, C.

    2005-01-01

    We introduce a new dynamic parallel memory optimization scheme for executing large scale 3-D discrete ordinates (Sn) simulations on distributed memory parallel computers. In order for parallel transport codes to be truly scalable, they must use parallel data storage, where only the variables that are locally computed are locally stored. Even with parallel data storage for the angular variables, cumulative storage requirements for large discrete ordinates calculations can be prohibitive. To address this problem, Memory Tuning has been implemented into the PENTRAN 3-D parallel discrete ordinates code as an optimized, two-level ('large' array, 'small' array) parallel data storage scheme. Memory Tuning can be described as the process of parallel data memory optimization. Memory Tuning dynamically minimizes the amount of required parallel data in allocated memory on each processor using a statistical sampling algorithm. This algorithm is based on the integral average and standard deviation of the number of fine meshes contained in each coarse mesh in the global problem. Because PENTRAN only stores the locally computed problem phase space, optimal two-level memory assignments can be unique on each node, depending upon the parallel decomposition used (hybrid combinations of angular, energy, or spatial). As demonstrated in the two large discrete ordinates models presented (a storage cask and an OECD MOX Benchmark), Memory Tuning can save a substantial amount of memory per parallel processor, allowing one to accomplish very large scale Sn computations. (authors)

  1. Substantial equivalence--an appropriate paradigm for the safety assessment of genetically modified foods?

    Kuiper, Harry A.; Kleter, Gijs A.; Noteborn, Hub P.J.M.; Kok, Esther J.

    2002-01-01

    Safety assessment of genetically modified food crops is based on the concept of substantial equivalence, developed by OECD and further elaborated by FAO/WHO. The concept embraces a comparative approach to identify possible differences between the genetically modified food and its traditional comparator, which is considered to be safe. The concept is not a safety assessment in itself, it identifies hazards but does not assess them. The outcome of the comparative exercise will further guide the safety assessment, which may include (immuno)toxicological and biochemical testing. Application of the concept of substantial equivalence may encounter practical difficulties: (i) the availability of near-isogenic parental lines to compare the genetically modified food with; (ii) limited availability of methods for the detection of (un)intended effects resulting from the genetic modification; and (iii) limited information on natural variations in levels of relevant crop constituents. In order to further improve the methodology for identification of unintended effects, new 'profiling' methods are recommended. Such methods will allow for the screening of potential changes in the modified host organism at different integration levels, i.e. at the genome level, during gene expression and protein translation, and at the level of cellular metabolism

  2. MPI_XSTAR: MPI-based Parallelization of the XSTAR Photoionization Program

    Danehkar, Ashkbiz; Nowak, Michael A.; Lee, Julia C.; Smith, Randall K.

    2018-02-01

    We describe a program for the parallel implementation of multiple runs of XSTAR, a photoionization code that is used to predict the physical properties of an ionized gas from its emission and/or absorption lines. The parallelization program, called MPI_XSTAR, has been developed and implemented in the C++ language by using the Message Passing Interface (MPI) protocol, a conventional standard of parallel computing. We have benchmarked parallel multiprocessing executions of XSTAR, using MPI_XSTAR, against a serial execution of XSTAR, in terms of the parallelization speedup and the computing resource efficiency. Our experience indicates that the parallel execution runs significantly faster than the serial execution, however, the efficiency in terms of the computing resource usage decreases with increasing the number of processors used in the parallel computing.

  3. Parallel transport in ideal magnetohydrodynamics and applications to resistive wall modes

    Finn, J.M.; Gerwin, R.A.

    1996-01-01

    It is shown that in magnetohydrodynamics (MHD) with an ideal Ohm close-quote s law, in the presence of parallel heat flux, density gradient, temperature gradient, and parallel compression, but in the absence of perpendicular compressibility, there is an exact cancellation of the parallel transport terms. This cancellation is due to the fact that magnetic flux is advected in the presence of an ideal Ohm close-quote s law, and therefore parallel transport of temperature and density gives the same result as perpendicular advection of the same quantities. Discussions are also presented regarding parallel viscosity and parallel velocity shear, and the generalization to toroidal geometry. These results suggest that a correct generalization of the Hammett endash Perkins fluid operator [G. W. Hammett and F. W. Perkins, Phys. Rev. Lett. 64, 3019 (1990)] to simulate Landau damping for electromagnetic modes must give an operator that acts on the dynamics parallel to the perturbed magnetic field lines. copyright 1996 American Institute of Physics

  4. Parallel fabrication of macroporous scaffolds.

    Dobos, Andrew; Grandhi, Taraka Sai Pavan; Godeshala, Sudhakar; Meldrum, Deirdre R; Rege, Kaushal

    2018-07-01

    Scaffolds generated from naturally occurring and synthetic polymers have been investigated in several applications because of their biocompatibility and tunable chemo-mechanical properties. Existing methods for generation of 3D polymeric scaffolds typically cannot be parallelized, suffer from low throughputs, and do not allow for quick and easy removal of the fragile structures that are formed. Current molds used in hydrogel and scaffold fabrication using solvent casting and porogen leaching are often single-use and do not facilitate 3D scaffold formation in parallel. Here, we describe a simple device and related approaches for the parallel fabrication of macroporous scaffolds. This approach was employed for the generation of macroporous and non-macroporous materials in parallel, in higher throughput and allowed for easy retrieval of these 3D scaffolds once formed. In addition, macroporous scaffolds with interconnected as well as non-interconnected pores were generated, and the versatility of this approach was employed for the generation of 3D scaffolds from diverse materials including an aminoglycoside-derived cationic hydrogel ("Amikagel"), poly(lactic-co-glycolic acid) or PLGA, and collagen. Macroporous scaffolds generated using the device were investigated for plasmid DNA binding and cell loading, indicating the use of this approach for developing materials for different applications in biotechnology. Our results demonstrate that the device-based approach is a simple technology for generating scaffolds in parallel, which can enhance the toolbox of current fabrication techniques. © 2018 Wiley Periodicals, Inc.

  5. Parallel plasma fluid turbulence calculations

    Leboeuf, J.N.; Carreras, B.A.; Charlton, L.A.; Drake, J.B.; Lynch, V.E.; Newman, D.E.; Sidikman, K.L.; Spong, D.A.

    1994-01-01

    The study of plasma turbulence and transport is a complex problem of critical importance for fusion-relevant plasmas. To this day, the fluid treatment of plasma dynamics is the best approach to realistic physics at the high resolution required for certain experimentally relevant calculations. Core and edge turbulence in a magnetic fusion device have been modeled using state-of-the-art, nonlinear, three-dimensional, initial-value fluid and gyrofluid codes. Parallel implementation of these models on diverse platforms--vector parallel (National Energy Research Supercomputer Center's CRAY Y-MP C90), massively parallel (Intel Paragon XP/S 35), and serial parallel (clusters of high-performance workstations using the Parallel Virtual Machine protocol)--offers a variety of paths to high resolution and significant improvements in real-time efficiency, each with its own advantages. The largest and most efficient calculations have been performed at the 200 Mword memory limit on the C90 in dedicated mode, where an overlap of 12 to 13 out of a maximum of 16 processors has been achieved with a gyrofluid model of core fluctuations. The richness of the physics captured by these calculations is commensurate with the increased resolution and efficiency and is limited only by the ingenuity brought to the analysis of the massive amounts of data generated

  6. Evaluating parallel optimization on transputers

    A.G. Chalmers

    2003-12-01

    Full Text Available The faster processing power of modern computers and the development of efficient algorithms have made it possible for operations researchers to tackle a much wider range of problems than ever before. Further improvements in processing speed can be achieved utilising relatively inexpensive transputers to process components of an algorithm in parallel. The Davidon-Fletcher-Powell method is one of the most successful and widely used optimisation algorithms for unconstrained problems. This paper examines the algorithm and identifies the components that can be processed in parallel. The results of some experiments with these components are presented which indicates under what conditions parallel processing with an inexpensive configuration is likely to be faster than the traditional sequential implementations. The performance of the whole algorithm with its parallel components is then compared with the original sequential algorithm. The implementation serves to illustrate the practicalities of speeding up typical OR algorithms in terms of difficulty, effort and cost. The results give an indication of the savings in time a given parallel implementation can be expected to yield.

  7. Pattern-Driven Automatic Parallelization

    Christoph W. Kessler

    1996-01-01

    Full Text Available This article describes a knowledge-based system for automatic parallelization of a wide class of sequential numerical codes operating on vectors and dense matrices, and for execution on distributed memory message-passing multiprocessors. Its main feature is a fast and powerful pattern recognition tool that locally identifies frequently occurring computations and programming concepts in the source code. This tool also works for dusty deck codes that have been "encrypted" by former machine-specific code transformations. Successful pattern recognition guides sophisticated code transformations including local algorithm replacement such that the parallelized code need not emerge from the sequential program structure by just parallelizing the loops. It allows access to an expert's knowledge on useful parallel algorithms, available machine-specific library routines, and powerful program transformations. The partially restored program semantics also supports local array alignment, distribution, and redistribution, and allows for faster and more exact prediction of the performance of the parallelized target code than is usually possible.

  8. Silver linings.

    Bultas, Margaret W; Pohlman, Shawn

    2014-01-01

    The purpose of this interpretive phenomenological study was to gain a better understanding of the experiences of 11 mothers of preschool children with autism spectrum disorder (ASD). Mothers were interviewed three times over a 6 week period. Interviews were analyzed using interpretive methods. This manuscript highlights one particular theme-a positive perspective mothers described as the "silver lining." This "silver lining" represents optimism despite the adversities associated with parenting a child with ASD. A deeper understanding of this side of mothering children with ASD may help health care providers improve rapport, communication, and result in more authentic family centered care. Copyright © 2014 Elsevier Inc. All rights reserved.

  9. Parallelizing AT with MatlabMPI

    2011-01-01

    The Accelerator Toolbox (AT) is a high-level collection of tools and scripts specifically oriented toward solving problems dealing with computational accelerator physics. It is integrated into the MATLAB environment, which provides an accessible, intuitive interface for accelerator physicists, allowing researchers to focus the majority of their efforts on simulations and calculations, rather than programming and debugging difficulties. Efforts toward parallelization of AT have been put in place to upgrade its performance to modern standards of computing. We utilized the packages MatlabMPI and pMatlab, which were developed by MIT Lincoln Laboratory, to set up a message-passing environment that could be called within MATLAB, which set up the necessary pre-requisites for multithread processing capabilities. On local quad-core CPUs, we were able to demonstrate processor efficiencies of roughly 95% and speed increases of nearly 380%. By exploiting the efficacy of modern-day parallel computing, we were able to demonstrate incredibly efficient speed increments per processor in AT's beam-tracking functions. Extrapolating from prediction, we can expect to reduce week-long computation runtimes to less than 15 minutes. This is a huge performance improvement and has enormous implications for the future computing power of the accelerator physics group at SSRL. However, one of the downfalls of parringpass is its current lack of transparency; the pMatlab and MatlabMPI packages must first be well-understood by the user before the system can be configured to run the scripts. In addition, the instantiation of argument parameters requires internal modification of the source code. Thus, parringpass, cannot be directly run from the MATLAB command line, which detracts from its flexibility and user-friendliness. Future work in AT's parallelization will focus on development of external functions and scripts that can be called from within MATLAB and configured on multiple nodes, while

  10. Conceptual design of multiple parallel switching controller

    Ugolini, D.; Yoshikawa, S.; Ozawa, K.

    1996-01-01

    This paper discusses the conceptual design and the development of a preliminary model of a multiple parallel switching (MPS) controller. The introduction of several advanced controllers has widened and improved the control capability of nonlinear dynamical systems. However, it is not possible to uniquely define a controller that always outperforms the others, and, in many situations, the controller providing the best control action depends on the operating conditions and on the intrinsic properties and behavior of the controlled dynamical system. The desire to combine the control action of several controllers with the purpose to continuously attain the best control action has motivated the development of the MPS controller. The MPS controller consists of a number of single controllers acting in parallel and of an artificial intelligence (AI) based selecting mechanism. The AI selecting mechanism analyzes the output of each controller and implements the one providing the best control performance. An inherent property of the MPS controller is the possibility to discard unreliable controllers while still being able to perform the control action. To demonstrate the feasibility and the capability of the MPS controller the simulation of the on-line operation control of a fast breeder reactor (FBR) evaporator is presented. (author)

  11. Cosmic Shear With ACS Pure Parallels

    Rhodes, Jason

    2002-07-01

    Small distortions in the shapes of background galaxies by foreground mass provide a powerful method of directly measuring the amount and distribution of dark matter. Several groups have recently detected this weak lensing by large-scale structure, also called cosmic shear. The high resolution and sensitivity of HST/ACS provide a unique opportunity to measure cosmic shear accurately on small scales. Using 260 parallel orbits in Sloan textiti {F775W} we will measure for the first time: beginlistosetlength sep0cm setlengthemsep0cm setlengthopsep0cm em the cosmic shear variance on scales Omega_m^0.5, with signal-to-noise {s/n} 20, and the mass density Omega_m with s/n=4. They will be done at small angular scales where non-linear effects dominate the power spectrum, providing a test of the gravitational instability paradigm for structure formation. Measurements on these scales are not possible from the ground, because of the systematic effects induced by PSF smearing from seeing. Having many independent lines of sight reduces the uncertainty due to cosmic variance, making parallel observations ideal.

  12. Parallel artificial liquid membrane extraction

    Gjelstad, Astrid; Rasmussen, Knut Einar; Parmer, Marthe Petrine

    2013-01-01

    This paper reports development of a new approach towards analytical liquid-liquid-liquid membrane extraction termed parallel artificial liquid membrane extraction. A donor plate and acceptor plate create a sandwich, in which each sample (human plasma) and acceptor solution is separated by an arti......This paper reports development of a new approach towards analytical liquid-liquid-liquid membrane extraction termed parallel artificial liquid membrane extraction. A donor plate and acceptor plate create a sandwich, in which each sample (human plasma) and acceptor solution is separated...... by an artificial liquid membrane. Parallel artificial liquid membrane extraction is a modification of hollow-fiber liquid-phase microextraction, where the hollow fibers are replaced by flat membranes in a 96-well plate format....

  13. Parallel algorithms for mapping pipelined and parallel computations

    Nicol, David M.

    1988-01-01

    Many computational problems in image processing, signal processing, and scientific computing are naturally structured for either pipelined or parallel computation. When mapping such problems onto a parallel architecture it is often necessary to aggregate an obvious problem decomposition. Even in this context the general mapping problem is known to be computationally intractable, but recent advances have been made in identifying classes of problems and architectures for which optimal solutions can be found in polynomial time. Among these, the mapping of pipelined or parallel computations onto linear array, shared memory, and host-satellite systems figures prominently. This paper extends that work first by showing how to improve existing serial mapping algorithms. These improvements have significantly lower time and space complexities: in one case a published O(nm sup 3) time algorithm for mapping m modules onto n processors is reduced to an O(nm log m) time complexity, and its space requirements reduced from O(nm sup 2) to O(m). Run time complexity is further reduced with parallel mapping algorithms based on these improvements, which run on the architecture for which they create the mappings.

  14. Collisionless reconnection: magnetic field line interaction

    R. A. Treumann

    2012-10-01

    Full Text Available Magnetic field lines are quantum objects carrying one quantum Φ0 = 2πh/e of magnetic flux and have finite radius λm. Here we argue that they possess a very specific dynamical interaction. Parallel field lines reject each other. When confined to a certain area they form two-dimensional lattices of hexagonal structure. We estimate the filling factor of such an area. Anti-parallel field lines, on the other hand, attract each other. We identify the physical mechanism as being due to the action of the gauge potential field, which we determine quantum mechanically for two parallel and two anti-parallel field lines. The distortion of the quantum electrodynamic vacuum causes a cloud of virtual pairs. We calculate the virtual pair production rate from quantum electrodynamics and estimate the virtual pair cloud density, pair current and Lorentz force density acting on the field lines via the pair cloud. These properties of field line dynamics become important in collisionless reconnection, consistently explaining why and how reconnection can spontaneously set on in the field-free centre of a current sheet below the electron-inertial scale.

  15. 26 CFR 1.132-4 - Line of business limitation.

    2010-04-01

    ... athletic facilities. (iii) Performance of substantial services in more than one line of business. An... one line of business, such lines of business will be treated as a single line of business where and to... business. For example, assume that on the same premises an employer sells both women's apparel and jewelry...

  16. Cellular automata a parallel model

    Mazoyer, J

    1999-01-01

    Cellular automata can be viewed both as computational models and modelling systems of real processes. This volume emphasises the first aspect. In articles written by leading researchers, sophisticated massive parallel algorithms (firing squad, life, Fischer's primes recognition) are treated. Their computational power and the specific complexity classes they determine are surveyed, while some recent results in relation to chaos from a new dynamic systems point of view are also presented. Audience: This book will be of interest to specialists of theoretical computer science and the parallelism challenge.

  17. Substantial Research Secures the Blue Future for our Blue Plant

    Moustafa Abdel Maksoud

    2016-06-01

    Full Text Available Earth, the blue planet, is our home, and seas and oceans cover more than 70% of its surface. As the earth’s population rapidly increases and available resources decrease, seas and oceans can play a key role in assuring the long-term survival of humankind. Renewable maritime energy has huge potential to provide a considerable part of the earth’s population with decarbonised electricity generation systems. Renewable maritime energy is very flexible and can be harvested above the water’s free surface by using offshore wind turbines, on the water’s surface by using wave energy converters or below the water’s surface by using current or tidal turbines. The supposed conflict between environmental protection measures and economic interests is neither viable nor reasonable. Renewable maritime energy can be the motor for considerable substantial economic growth for many maritime regions and therefore for society at large. The fastest growing sector of renewable maritime energy is offshore wind. The annual report of the European Wind Energy Association from the year 2015 confirms the growing relevance of the offshore wind industry. In 2015, the total installed and grid-connected capacity of wind power was 12,800 MW in the EU and 6,013.4 MW in Germany. 38% of the 2015 annual installation in Germany was offshore, accounting for a capacity of 2,282.4 MW. However, there are a limited number of available installation sites in shallow water, meaning that there is an urgent need to develop new offshore structures for water depths greater than 50m. The persistent trend towards deeper waters has encouraged the offshore wind industry to look for floating wind turbine structures and larger turbines. Floating wind turbine technologies are at an early stage of development and many technical and economic challenges will still need to be faced. Nonetheless, intensive research activities and the employment of advanced technologies are the key factors in

  18. Parallel Sparse Matrix - Vector Product

    Alexandersen, Joe; Lazarov, Boyan Stefanov; Dammann, Bernd

    This technical report contains a case study of a sparse matrix-vector product routine, implemented for parallel execution on a compute cluster with both pure MPI and hybrid MPI-OpenMP solutions. C++ classes for sparse data types were developed and the report shows how these class can be used...

  19. [Falsified medicines in parallel trade].

    Muckenfuß, Heide

    2017-11-01

    The number of falsified medicines on the German market has distinctly increased over the past few years. In particular, stolen pharmaceutical products, a form of falsified medicines, have increasingly been introduced into the legal supply chain via parallel trading. The reasons why parallel trading serves as a gateway for falsified medicines are most likely the complex supply chains and routes of transport. It is hardly possible for national authorities to trace the history of a medicinal product that was bought and sold by several intermediaries in different EU member states. In addition, the heterogeneous outward appearance of imported and relabelled pharmaceutical products facilitates the introduction of illegal products onto the market. Official batch release at the Paul-Ehrlich-Institut offers the possibility of checking some aspects that might provide an indication of a falsified medicine. In some circumstances, this may allow the identification of falsified medicines before they come onto the German market. However, this control is only possible for biomedicinal products that have not received a waiver regarding official batch release. For improved control of parallel trade, better networking among the EU member states would be beneficial. European-wide regulations, e. g., for disclosure of the complete supply chain, would help to minimise the risks of parallel trading and hinder the marketing of falsified medicines.

  20. The parallel adult education system

    Wahlgren, Bjarne

    2015-01-01

    for competence development. The Danish university educational system includes two parallel programs: a traditional academic track (candidatus) and an alternative practice-based track (master). The practice-based program was established in 2001 and organized as part time. The total program takes half the time...

  1. Where are the parallel algorithms?

    Voigt, R. G.

    1985-01-01

    Four paradigms that can be useful in developing parallel algorithms are discussed. These include computational complexity analysis, changing the order of computation, asynchronous computation, and divide and conquer. Each is illustrated with an example from scientific computation, and it is shown that computational complexity must be used with great care or an inefficient algorithm may be selected.

  2. Parallel imaging with phase scrambling.

    Zaitsev, Maxim; Schultz, Gerrit; Hennig, Juergen; Gruetter, Rolf; Gallichan, Daniel

    2015-04-01

    Most existing methods for accelerated parallel imaging in MRI require additional data, which are used to derive information about the sensitivity profile of each radiofrequency (RF) channel. In this work, a method is presented to avoid the acquisition of separate coil calibration data for accelerated Cartesian trajectories. Quadratic phase is imparted to the image to spread the signals in k-space (aka phase scrambling). By rewriting the Fourier transform as a convolution operation, a window can be introduced to the convolved chirp function, allowing a low-resolution image to be reconstructed from phase-scrambled data without prominent aliasing. This image (for each RF channel) can be used to derive coil sensitivities to drive existing parallel imaging techniques. As a proof of concept, the quadratic phase was applied by introducing an offset to the x(2) - y(2) shim and the data were reconstructed using adapted versions of the image space-based sensitivity encoding and GeneRalized Autocalibrating Partially Parallel Acquisitions algorithms. The method is demonstrated in a phantom (1 × 2, 1 × 3, and 2 × 2 acceleration) and in vivo (2 × 2 acceleration) using a 3D gradient echo acquisition. Phase scrambling can be used to perform parallel imaging acceleration without acquisition of separate coil calibration data, demonstrated here for a 3D-Cartesian trajectory. Further research is required to prove the applicability to other 2D and 3D sampling schemes. © 2014 Wiley Periodicals, Inc.

  3. Default Parallels Plesk Panel Page

    services that small businesses want and need. Our software includes key building blocks of cloud service virtualized servers Service Provider Products Parallels® Automation Hosting, SaaS, and cloud computing , the leading hosting automation software. You see this page because there is no Web site at this

  4. Matpar: Parallel Extensions for MATLAB

    Springer, P. L.

    1998-01-01

    Matpar is a set of client/server software that allows a MATLAB user to take advantage of a parallel computer for very large problems. The user can replace calls to certain built-in MATLAB functions with calls to Matpar functions.

  5. Massively parallel quantum computer simulator

    De Raedt, K.; Michielsen, K.; De Raedt, H.; Trieu, B.; Arnold, G.; Richter, M.; Lippert, Th.; Watanabe, H.; Ito, N.

    2007-01-01

    We describe portable software to simulate universal quantum computers on massive parallel Computers. We illustrate the use of the simulation software by running various quantum algorithms on different computer architectures, such as a IBM BlueGene/L, a IBM Regatta p690+, a Hitachi SR11000/J1, a Cray

  6. Parallel computing: numerics, applications, and trends

    Trobec, Roman; Vajteršic, Marián; Zinterhof, Peter

    2009-01-01

    ... and/or distributed systems. The contributions to this book are focused on topics most concerned in the trends of today's parallel computing. These range from parallel algorithmics, programming, tools, network computing to future parallel computing. Particular attention is paid to parallel numerics: linear algebra, differential equations, numerica...

  7. Experiments with parallel algorithms for combinatorial problems

    G.A.P. Kindervater (Gerard); H.W.J.M. Trienekens

    1985-01-01

    textabstractIn the last decade many models for parallel computation have been proposed and many parallel algorithms have been developed. However, few of these models have been realized and most of these algorithms are supposed to run on idealized, unrealistic parallel machines. The parallel machines

  8. Parallel R-matrix computation

    Heggarty, J.W.

    1999-06-01

    For almost thirty years, sequential R-matrix computation has been used by atomic physics research groups, from around the world, to model collision phenomena involving the scattering of electrons or positrons with atomic or molecular targets. As considerable progress has been made in the understanding of fundamental scattering processes, new data, obtained from more complex calculations, is of current interest to experimentalists. Performing such calculations, however, places considerable demands on the computational resources to be provided by the target machine, in terms of both processor speed and memory requirement. Indeed, in some instances the computational requirements are so great that the proposed R-matrix calculations are intractable, even when utilising contemporary classic supercomputers. Historically, increases in the computational requirements of R-matrix computation were accommodated by porting the problem codes to a more powerful classic supercomputer. Although this approach has been successful in the past, it is no longer considered to be a satisfactory solution due to the limitations of current (and future) Von Neumann machines. As a consequence, there has been considerable interest in the high performance multicomputers, that have emerged over the last decade which appear to offer the computational resources required by contemporary R-matrix research. Unfortunately, developing codes for these machines is not as simple a task as it was to develop codes for successive classic supercomputers. The difficulty arises from the considerable differences in the computing models that exist between the two types of machine and results in the programming of multicomputers to be widely acknowledged as a difficult, time consuming and error-prone task. Nevertheless, unless parallel R-matrix computation is realised, important theoretical and experimental atomic physics research will continue to be hindered. This thesis describes work that was undertaken in

  9. The numerical parallel computing of photon transport

    Huang Qingnan; Liang Xiaoguang; Zhang Lifa

    1998-12-01

    The parallel computing of photon transport is investigated, the parallel algorithm and the parallelization of programs on parallel computers both with shared memory and with distributed memory are discussed. By analyzing the inherent law of the mathematics and physics model of photon transport according to the structure feature of parallel computers, using the strategy of 'to divide and conquer', adjusting the algorithm structure of the program, dissolving the data relationship, finding parallel liable ingredients and creating large grain parallel subtasks, the sequential computing of photon transport into is efficiently transformed into parallel and vector computing. The program was run on various HP parallel computers such as the HY-1 (PVP), the Challenge (SMP) and the YH-3 (MPP) and very good parallel speedup has been gotten

  10. Effects of parallel electron dynamics on plasma blob transport

    Angus, Justin R.; Krasheninnikov, Sergei I. [University of California, San Diego, 9500 Gilman Drive, La Jolla, California 92093 (United States); Umansky, Maxim V. [Lawrence Livermore National Laboratory, 7000 East Avenue, Livermore, California 94550 (United States)

    2012-08-15

    The 3D effects on sheath connected plasma blobs that result from parallel electron dynamics are studied by allowing for the variation of blob density and potential along the magnetic field line and using collisional Ohm's law to model the parallel current density. The parallel current density from linear sheath theory, typically used in the 2D model, is implemented as parallel boundary conditions. This model includes electrostatic 3D effects, such as resistive drift waves and blob spinning, while retaining all of the fundamental 2D physics of sheath connected plasma blobs. If the growth time of unstable drift waves is comparable to the 2D advection time scale of the blob, then the blob's density gradient will be depleted resulting in a much more diffusive blob with little radial motion. Furthermore, blob profiles that are initially varying along the field line drive the potential to a Boltzmann relation that spins the blob and thereby acts as an addition sink of the 2D potential. Basic dimensionless parameters are presented to estimate the relative importance of these two 3D effects. The deviation of blob dynamics from that predicted by 2D theory in the appropriate limits of these parameters is demonstrated by a direct comparison of 2D and 3D seeded blob simulations.

  11. Semi-coarsening multigrid methods for parallel computing

    Jones, J.E.

    1996-12-31

    Standard multigrid methods are not well suited for problems with anisotropic coefficients which can occur, for example, on grids that are stretched to resolve a boundary layer. There are several different modifications of the standard multigrid algorithm that yield efficient methods for anisotropic problems. In the paper, we investigate the parallel performance of these multigrid algorithms. Multigrid algorithms which work well for anisotropic problems are based on line relaxation and/or semi-coarsening. In semi-coarsening multigrid algorithms a grid is coarsened in only one of the coordinate directions unlike standard or full-coarsening multigrid algorithms where a grid is coarsened in each of the coordinate directions. When both semi-coarsening and line relaxation are used, the resulting multigrid algorithm is robust and automatic in that it requires no knowledge of the nature of the anisotropy. This is the basic multigrid algorithm whose parallel performance we investigate in the paper. The algorithm is currently being implemented on an IBM SP2 and its performance is being analyzed. In addition to looking at the parallel performance of the basic semi-coarsening algorithm, we present algorithmic modifications with potentially better parallel efficiency. One modification reduces the amount of computational work done in relaxation at the expense of using multiple coarse grids. This modification is also being implemented with the aim of comparing its performance to that of the basic semi-coarsening algorithm.

  12. Automatic Parallelization Tool: Classification of Program Code for Parallel Computing

    Mustafa Basthikodi

    2016-04-01

    Full Text Available Performance growth of single-core processors has come to a halt in the past decade, but was re-enabled by the introduction of parallelism in processors. Multicore frameworks along with Graphical Processing Units empowered to enhance parallelism broadly. Couples of compilers are updated to developing challenges forsynchronization and threading issues. Appropriate program and algorithm classifications will have advantage to a great extent to the group of software engineers to get opportunities for effective parallelization. In present work we investigated current species for classification of algorithms, in that related work on classification is discussed along with the comparison of issues that challenges the classification. The set of algorithms are chosen which matches the structure with different issues and perform given task. We have tested these algorithms utilizing existing automatic species extraction toolsalong with Bones compiler. We have added functionalities to existing tool, providing a more detailed characterization. The contributions of our work include support for pointer arithmetic, conditional and incremental statements, user defined types, constants and mathematical functions. With this, we can retain significant data which is not captured by original speciesof algorithms. We executed new theories into the device, empowering automatic characterization of program code.

  13. Assignment of stock keeping units to parallel undirectional picking

    Matthews, Jason

    2015-05-01

    Full Text Available An order picking system consisting of a number of parallel unidirectional picking lines is investigated. Stock keeping units (SKUs that are grouped by product type into distributions (DBNs are assigned daily to available picking lines. A mathematical programming formulation and its relaxations is presented. A greedy insertion and a greedy phased insertion are further introduced to obtain feasible results within usable computation times for all test cases. The walking distance of the pickers was shown to decrease by about 22 per cent compared with the current assignment approach. However, product handling and operational risk increases.

  14. Dynamic Line Rating Oncor Electric Delivery Smart Grid Program

    Johnson, Justin; Smith, Cale; Young, Mike; Donohoo, Ken; Owen, Ross; Clark, Eddit; Espejo, Raul; Aivaliotis, Sandy; Stelmak, Ron; Mohr, Ron; Barba, Cristian; Gonzalez, Guillermo; Malkin, Stuart; Dimitrova, Vessela; Ragsdale, Gary; Mitchem, Sean; Jeirath, Nakul; Loomis, Joe; Trevino, Gerardo; Syracuse, Steve; Hurst, Neil; Mereness, Matt; Johnson, Chad; Bivens, Carrie

    2013-05-04

    Electric transmission lines are the lifeline of the electric utility industry, delivering its product from source to consumer. This critical infrastructure is often constrained such that there is inadequate capacity on existing transmission lines to efficiently deliver the power to meet demand in certain areas or to transport energy from high-generation areas to high-consumption regions. When this happens, the cost of the energy rises; more costly sources of power are used to meet the demand or the system operates less reliably. These economic impacts are known as congestion, and they can amount to substantial dollars for any time frame of reference: hour, day or year. There are several solutions to the transmission constraint problem, including: construction of new generation, construction of new transmission facilities, rebuilding and reconductoring of existing transmission assets, and Dynamic Line Rating (DLR). All of these options except DLR are capital intensive, have long lead times and often experience strong public and regulatory opposition. The Smart Grid Demonstration Program (SGDP) project co-funded by the Department of Energy (DOE) and Oncor Electric Delivery Company developed and deployed the most extensive and advanced DLR installation to demonstrate that DLR technology is capable of resolving many transmission capacity constraint problems with a system that is reliable, safe and very cost competitive. The SGDP DLR deployment is the first application of DLR technology to feed transmission line real-time dynamic ratings directly into the system operation’s State Estimator and load dispatch program, which optimizes the matching of generation with load demand on a security, reliability and economic basis. The integrated Dynamic Line Rating (iDLR)1 collects transmission line parameters at remote locations on the lines, calculates the real-time line rating based on the equivalent conductor temperature, ambient temperature and influence of wind and solar

  15. production lines

    Jingshan Li

    2000-01-01

    Full Text Available In this work, serial production lines with finished goods buffers operating in the pull regime are considered. The machines are assumed to obey Bernoulli reliability model. The problem of satisfying customers demand is addressed. The level of demand satisfaction is quantified by the due-time performance (DTP, which is defined as the probability to ship to the customer a required number of parts during a fixed time interval. Within this scenario, the definitions of DTP bottlenecks are introduced and a method for their identification is developed.

  16. An Integrated Inductor for Parallel Interleaved VSCs and PWM Schemes for Flux Minimization

    Gohil, Ghanshyamsinh Vijaysinh; Bede, Lorand; Teodorescu, Remus

    2015-01-01

    The interleaving of the carrier signals of the parallel Voltage Source Converters (VSCs) can reduce the harmonic content in the resultant switched output voltages. As a result, the size of the line filter inductor can be reduced. However, in addition to the line filter, an inductive filter is often...

  17. Structural synthesis of parallel robots

    Gogu, Grigore

    This book represents the fifth part of a larger work dedicated to the structural synthesis of parallel robots. The originality of this work resides in the fact that it combines new formulae for mobility, connectivity, redundancy and overconstraints with evolutionary morphology in a unified structural synthesis approach that yields interesting and innovative solutions for parallel robotic manipulators.  This is the first book on robotics that presents solutions for coupled, decoupled, uncoupled, fully-isotropic and maximally regular robotic manipulators with Schönflies motions systematically generated by using the structural synthesis approach proposed in Part 1.  Overconstrained non-redundant/overactuated/redundantly actuated solutions with simple/complex limbs are proposed. Many solutions are presented here for the first time in the literature. The author had to make a difficult and challenging choice between protecting these solutions through patents and releasing them directly into the public domain. T...

  18. GPU Parallel Bundle Block Adjustment

    ZHENG Maoteng

    2017-09-01

    Full Text Available To deal with massive data in photogrammetry, we introduce the GPU parallel computing technology. The preconditioned conjugate gradient and inexact Newton method are also applied to decrease the iteration times while solving the normal equation. A brand new workflow of bundle adjustment is developed to utilize GPU parallel computing technology. Our method can avoid the storage and inversion of the big normal matrix, and compute the normal matrix in real time. The proposed method can not only largely decrease the memory requirement of normal matrix, but also largely improve the efficiency of bundle adjustment. It also achieves the same accuracy as the conventional method. Preliminary experiment results show that the bundle adjustment of a dataset with about 4500 images and 9 million image points can be done in only 1.5 minutes while achieving sub-pixel accuracy.

  19. A tandem parallel plate analyzer

    Hamada, Y.; Fujisawa, A.; Iguchi, H.; Nishizawa, A.; Kawasumi, Y.

    1996-11-01

    By a new modification of a parallel plate analyzer the second-order focus is obtained in an arbitrary injection angle. This kind of an analyzer with a small injection angle will have an advantage of small operational voltage, compared to the Proca and Green analyzer where the injection angle is 30 degrees. Thus, the newly proposed analyzer will be very useful for the precise energy measurement of high energy particles in MeV range. (author)

  20. High-speed parallel counter

    Gus'kov, B.N.; Kalinnikov, V.A.; Krastev, V.R.; Maksimov, A.N.; Nikityuk, N.M.

    1985-01-01

    This paper describes a high-speed parallel counter that contains 31 inputs and 15 outputs and is implemented by integrated circuits of series 500. The counter is designed for fast sampling of events according to the number of particles that pass simultaneously through the hodoscopic plane of the detector. The minimum delay of the output signals relative to the input is 43 nsec. The duration of the output signals can be varied from 75 to 120 nsec

  1. An anthropologist in parallel structure

    Noelle Molé Liston

    2016-08-01

    Full Text Available The essay examines the parallels between Molé Liston’s studies on labor and precarity in Italy and the United States’ anthropology job market. Probing the way economic shift reshaped the field of anthropology of Europe in the late 2000s, the piece explores how the neoliberalization of the American academy increased the value in studying the hardships and daily lives of non-western populations in Europe.

  2. Combinatorics of spreads and parallelisms

    Johnson, Norman

    2010-01-01

    Partitions of Vector Spaces Quasi-Subgeometry Partitions Finite Focal-SpreadsGeneralizing André SpreadsThe Going Up Construction for Focal-SpreadsSubgeometry Partitions Subgeometry and Quasi-Subgeometry Partitions Subgeometries from Focal-SpreadsExtended André SubgeometriesKantor's Flag-Transitive DesignsMaximal Additive Partial SpreadsSubplane Covered Nets and Baer Groups Partial Desarguesian t-Parallelisms Direct Products of Affine PlanesJha-Johnson SL(2,

  3. Wakefield calculations on parallel computers

    Schoessow, P.

    1990-01-01

    The use of parallelism in the solution of wakefield problems is illustrated for two different computer architectures (SIMD and MIMD). Results are given for finite difference codes which have been implemented on a Connection Machine and an Alliant FX/8 and which are used to compute wakefields in dielectric loaded structures. Benchmarks on code performance are presented for both cases. 4 refs., 3 figs., 2 tabs

  4. Aspects of computation on asynchronous parallel processors

    Wright, M.

    1989-01-01

    The increasing availability of asynchronous parallel processors has provided opportunities for original and useful work in scientific computing. However, the field of parallel computing is still in a highly volatile state, and researchers display a wide range of opinion about many fundamental questions such as models of parallelism, approaches for detecting and analyzing parallelism of algorithms, and tools that allow software developers and users to make effective use of diverse forms of complex hardware. This volume collects the work of researchers specializing in different aspects of parallel computing, who met to discuss the framework and the mechanics of numerical computing. The far-reaching impact of high-performance asynchronous systems is reflected in the wide variety of topics, which include scientific applications (e.g. linear algebra, lattice gauge simulation, ordinary and partial differential equations), models of parallelism, parallel language features, task scheduling, automatic parallelization techniques, tools for algorithm development in parallel environments, and system design issues

  5. Parallel processing of genomics data

    Agapito, Giuseppe; Guzzi, Pietro Hiram; Cannataro, Mario

    2016-10-01

    The availability of high-throughput experimental platforms for the analysis of biological samples, such as mass spectrometry, microarrays and Next Generation Sequencing, have made possible to analyze a whole genome in a single experiment. Such platforms produce an enormous volume of data per single experiment, thus the analysis of this enormous flow of data poses several challenges in term of data storage, preprocessing, and analysis. To face those issues, efficient, possibly parallel, bioinformatics software needs to be used to preprocess and analyze data, for instance to highlight genetic variation associated with complex diseases. In this paper we present a parallel algorithm for the parallel preprocessing and statistical analysis of genomics data, able to face high dimension of data and resulting in good response time. The proposed system is able to find statistically significant biological markers able to discriminate classes of patients that respond to drugs in different ways. Experiments performed on real and synthetic genomic datasets show good speed-up and scalability.

  6. Line facilities outline

    1998-08-01

    This book deals with line facilities. The contents of this book are outline line of wire telecommunication ; development of line, classification of section of line and theory of transmission of line, cable line ; structure of line, line of cable in town, line out of town, domestic cable and other lines, Optical communication ; line of optical cable, transmission method, measurement of optical communication and cable of the sea bottom, Equipment of telecommunication line ; telecommunication line facilities and telecommunication of public works, construction of cable line and maintenance and Regulation of line equipment ; regulation on technique, construction and maintenance.

  7. Flux-line-cutting losses in type-II superconductors

    Clem, J.R.

    1982-01-01

    Energy dissipation associated with flux-line cutting (intersection and cross-joining of adjacent nonparallel vortices) is considered theoretically. The flux-line-cutting contribution to the dissipation per unit volume, arising from mutual annihilation of transverse magnetic flux, is identified as J/sub parallel/xE/sub parallel/, where J/sub parallel/ and E/sub parallel/ are the components of the current density and the electric field parallel to the magnetic induction. The dynamical behavior of the magnetic structure at the flux-line-cutting threshold is shown to be governed by a special critical-state model similar to that proposed by previous authors. The resulting flux-line-cutting critical-state model, characterized in planar geometry by a parallel critical current density J/sub c/parallel or a critical angle gradient k/sub c/, is used to calculate predicted hysteretic ac flux-line-cutting losses in type-II superconductors in which the flux pinning is weak. The relation of the theory to previous experiments is discussed

  8. Large rainfall changes consistently projected over substantial areas of tropical land

    Chadwick, Robin; Good, Peter; Martin, Gill; Rowell, David P.

    2016-02-01

    Many tropical countries are exceptionally vulnerable to changes in rainfall patterns, with floods or droughts often severely affecting human life and health, food and water supplies, ecosystems and infrastructure. There is widespread disagreement among climate model projections of how and where rainfall will change over tropical land at the regional scales relevant to impacts, with different models predicting the position of current tropical wet and dry regions to shift in different ways. Here we show that despite uncertainty in the location of future rainfall shifts, climate models consistently project that large rainfall changes will occur for a considerable proportion of tropical land over the twenty-first century. The area of semi-arid land affected by large changes under a higher emissions scenario is likely to be greater than during even the most extreme regional wet or dry periods of the twentieth century, such as the Sahel drought of the late 1960s to 1990s. Substantial changes are projected to occur by mid-century--earlier than previously expected--and to intensify in line with global temperature rise. Therefore, current climate projections contain quantitative, decision-relevant information on future regional rainfall changes, particularly with regard to climate change mitigation policy.

  9. Overview of the Force Scientific Parallel Language

    Gita Alaghband

    1994-01-01

    Full Text Available The Force parallel programming language designed for large-scale shared-memory multiprocessors is presented. The language provides a number of parallel constructs as extensions to the ordinary Fortran language and is implemented as a two-level macro preprocessor to support portability across shared memory multiprocessors. The global parallelism model on which the Force is based provides a powerful parallel language. The parallel constructs, generic synchronization, and freedom from process management supported by the Force has resulted in structured parallel programs that are ported to the many multiprocessors on which the Force is implemented. Two new parallel constructs for looping and functional decomposition are discussed. Several programming examples to illustrate some parallel programming approaches using the Force are also presented.

  10. Parallel kinematics type, kinematics, and optimal design

    Liu, Xin-Jun

    2014-01-01

    Parallel Kinematics- Type, Kinematics, and Optimal Design presents the results of 15 year's research on parallel mechanisms and parallel kinematics machines. This book covers the systematic classification of parallel mechanisms (PMs) as well as providing a large number of mechanical architectures of PMs available for use in practical applications. It focuses on the kinematic design of parallel robots. One successful application of parallel mechanisms in the field of machine tools, which is also called parallel kinematics machines, has been the emerging trend in advanced machine tools. The book describes not only the main aspects and important topics in parallel kinematics, but also references novel concepts and approaches, i.e. type synthesis based on evolution, performance evaluation and optimization based on screw theory, singularity model taking into account motion and force transmissibility, and others.   This book is intended for researchers, scientists, engineers and postgraduates or above with interes...

  11. Applied Parallel Computing Industrial Computation and Optimization

    Madsen, Kaj; NA NA NA Olesen, Dorte

    Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)......Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)...

  12. Bottom lines

    Dabbs, F.

    1998-01-01

    The technological challenges facing the Alberta oilpatch in 1998 were reviewed. Attention was drawn to the fallacy of certain neo-conservative views, held by some members of the provincial government. This view holds that technology development is not a public concern and should be left entirely to the private sector. It is argued that rather than leaving it to the private sector, the government should play a strong leadership role. The government should do this by setting the agenda, by providing substantial and patient funding, by initiating partnerships with the private sector and by being a committed partner and catalyst in technology transfer. In support of this contention, the achievements of the Alberta Oil Sands Technology and Research Authority (AOSTRA) and of the Alberta Research Council were cite., These include the work on horizontal drilling, top-drive rigs, advanced exploration software, oilsands mining innovations, the steam-assisted gravity drainage (SAGD) technology, and the Taciuk process for bitumen recovery. All of these technologies, developed by government-industry funding, have contributed immeasurably to the Alberta economic miracle. Public expenditures on funding these research projects were but a fraction of the billions of dollars that have been, and will continue to be, returned to the people of Alberta in licence fees and royalties. In 1998, in the wake of the Kyoto Conference, the need for government leadership in energy research is even greater than it has been in the past. Nothing less than the viability of the fossil fuel industry and the Alberta economy depend upon finding solutions to the production and processing of greenhouse gas emissions

  13. Parallel algorithms and cluster computing

    Hoffmann, Karl Heinz

    2007-01-01

    This book presents major advances in high performance computing as well as major advances due to high performance computing. It contains a collection of papers in which results achieved in the collaboration of scientists from computer science, mathematics, physics, and mechanical engineering are presented. From the science problems to the mathematical algorithms and on to the effective implementation of these algorithms on massively parallel and cluster computers we present state-of-the-art methods and technology as well as exemplary results in these fields. This book shows that problems which seem superficially distinct become intimately connected on a computational level.

  14. Parallel computation of rotating flows

    Lundin, Lars Kristian; Barker, Vincent A.; Sørensen, Jens Nørkær

    1999-01-01

    This paper deals with the simulation of 3‐D rotating flows based on the velocity‐vorticity formulation of the Navier‐Stokes equations in cylindrical coordinates. The governing equations are discretized by a finite difference method. The solution is advanced to a new time level by a two‐step process...... is that of solving a singular, large, sparse, over‐determined linear system of equations, and the iterative method CGLS is applied for this purpose. We discuss some of the mathematical and numerical aspects of this procedure and report on the performance of our software on a wide range of parallel computers. Darbe...

  15. The parallel volume at large distances

    Kampf, Jürgen

    In this paper we examine the asymptotic behavior of the parallel volume of planar non-convex bodies as the distance tends to infinity. We show that the difference between the parallel volume of the convex hull of a body and the parallel volume of the body itself tends to . This yields a new proof...... for the fact that a planar body can only have polynomial parallel volume, if it is convex. Extensions to Minkowski spaces and random sets are also discussed....

  16. The parallel volume at large distances

    Kampf, Jürgen

    In this paper we examine the asymptotic behavior of the parallel volume of planar non-convex bodies as the distance tends to infinity. We show that the difference between the parallel volume of the convex hull of a body and the parallel volume of the body itself tends to 0. This yields a new proof...... for the fact that a planar body can only have polynomial parallel volume, if it is convex. Extensions to Minkowski spaces and random sets are also discussed....

  17. Computationally efficient implementation of combustion chemistry in parallel PDF calculations

    Lu Liuyan; Lantz, Steven R.; Ren Zhuyin; Pope, Stephen B.

    2009-01-01

    In parallel calculations of combustion processes with realistic chemistry, the serial in situ adaptive tabulation (ISAT) algorithm [S.B. Pope, Computationally efficient implementation of combustion chemistry using in situ adaptive tabulation, Combustion Theory and Modelling, 1 (1997) 41-63; L. Lu, S.B. Pope, An improved algorithm for in situ adaptive tabulation, Journal of Computational Physics 228 (2009) 361-386] substantially speeds up the chemistry calculations on each processor. To improve the parallel efficiency of large ensembles of such calculations in parallel computations, in this work, the ISAT algorithm is extended to the multi-processor environment, with the aim of minimizing the wall clock time required for the whole ensemble. Parallel ISAT strategies are developed by combining the existing serial ISAT algorithm with different distribution strategies, namely purely local processing (PLP), uniformly random distribution (URAN), and preferential distribution (PREF). The distribution strategies enable the queued load redistribution of chemistry calculations among processors using message passing. They are implemented in the software x2f m pi, which is a Fortran 95 library for facilitating many parallel evaluations of a general vector function. The relative performance of the parallel ISAT strategies is investigated in different computational regimes via the PDF calculations of multiple partially stirred reactors burning methane/air mixtures. The results show that the performance of ISAT with a fixed distribution strategy strongly depends on certain computational regimes, based on how much memory is available and how much overlap exists between tabulated information on different processors. No one fixed strategy consistently achieves good performance in all the regimes. Therefore, an adaptive distribution strategy, which blends PLP, URAN and PREF, is devised and implemented. It yields consistently good performance in all regimes. In the adaptive parallel

  18. 20 CFR 654.13 - Determination of areas of substantial unemployment.

    2010-04-01

    ... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Determination of areas of substantial unemployment. 654.13 Section 654.13 Employees' Benefits EMPLOYMENT AND TRAINING ADMINISTRATION, DEPARTMENT OF... 10582 § 654.13 Determination of areas of substantial unemployment. An area of substantial unemployment...

  19. A Parallel Approach to Fractal Image Compression

    Lubomir Dedera

    2004-01-01

    Full Text Available The paper deals with a parallel approach to coding and decoding algorithms in fractal image compressionand presents experimental results comparing sequential and parallel algorithms from the point of view of achieved bothcoding and decoding time and effectiveness of parallelization.

  20. Parallel Computing Using Web Servers and "Servlets".

    Lo, Alfred; Bloor, Chris; Choi, Y. K.

    2000-01-01

    Describes parallel computing and presents inexpensive ways to implement a virtual parallel computer with multiple Web servers. Highlights include performance measurement of parallel systems; models for using Java and intranet technology including single server, multiple clients and multiple servers, single client; and a comparison of CGI (common…

  1. An Introduction to Parallel Computation R

    How are they programmed? This article provides an introduction. A parallel computer is a network of processors built for ... and have been used to solve problems much faster than a single ... in parallel computer design is to select an organization which ..... The most ambitious approach to parallel computing is to develop.

  2. Comparison of parallel viscosity with neoclassical theory

    Ida, K.; Nakajima, N.

    1996-04-01

    Toroidal rotation profiles are measured with charge exchange spectroscopy for the plasma heated with tangential NBI in CHS heliotron/torsatron device to estimate parallel viscosity. The parallel viscosity derived from the toroidal rotation velocity shows good agreement with the neoclassical parallel viscosity plus the perpendicular viscosity. (μ perpendicular = 2 m 2 /s). (author)

  3. Advances in randomized parallel computing

    Rajasekaran, Sanguthevar

    1999-01-01

    The technique of randomization has been employed to solve numerous prob­ lems of computing both sequentially and in parallel. Examples of randomized algorithms that are asymptotically better than their deterministic counterparts in solving various fundamental problems abound. Randomized algorithms have the advantages of simplicity and better performance both in theory and often in practice. This book is a collection of articles written by renowned experts in the area of randomized parallel computing. A brief introduction to randomized algorithms In the aflalysis of algorithms, at least three different measures of performance can be used: the best case, the worst case, and the average case. Often, the average case run time of an algorithm is much smaller than the worst case. 2 For instance, the worst case run time of Hoare's quicksort is O(n ), whereas its average case run time is only O( n log n). The average case analysis is conducted with an assumption on the input space. The assumption made to arrive at t...

  4. Xyce parallel electronic simulator design.

    Thornquist, Heidi K.; Rankin, Eric Lamont; Mei, Ting; Schiek, Richard Louis; Keiter, Eric Richard; Russo, Thomas V.

    2010-09-01

    This document is the Xyce Circuit Simulator developer guide. Xyce has been designed from the 'ground up' to be a SPICE-compatible, distributed memory parallel circuit simulator. While it is in many respects a research code, Xyce is intended to be a production simulator. As such, having software quality engineering (SQE) procedures in place to insure a high level of code quality and robustness are essential. Version control, issue tracking customer support, C++ style guildlines and the Xyce release process are all described. The Xyce Parallel Electronic Simulator has been under development at Sandia since 1999. Historically, Xyce has mostly been funded by ASC, the original focus of Xyce development has primarily been related to circuits for nuclear weapons. However, this has not been the only focus and it is expected that the project will diversify. Like many ASC projects, Xyce is a group development effort, which involves a number of researchers, engineers, scientists, mathmaticians and computer scientists. In addition to diversity of background, it is to be expected on long term projects for there to be a certain amount of staff turnover, as people move on to different projects. As a result, it is very important that the project maintain high software quality standards. The point of this document is to formally document a number of the software quality practices followed by the Xyce team in one place. Also, it is hoped that this document will be a good source of information for new developers.

  5. Quadrupole singlet focusing for achromatic parallel-to-parallel devices

    Brown, K.L.

    1983-01-01

    A first order achromatic magnetic deflection system for use in conjunction with a charged particle accelerator is realized from a stepped gap magnet wherein charged particles propagating through the system are subject to at least two adjacent homogeneous magnetic fields in adjacent regions in traversing one-half of a symmetric trajectory through the system. A quadrupole singlet element Q of adjustable focal length disposed substantially at the entrance plane of such a symmetric system makes possible the coincidence of the waists of the beam in both the vertical (transverse) and (radial) bending planes. (author)

  6. VT Digital Line Graph Miscellaneous Transmission Lines

    Vermont Center for Geographic Information — (Link to Metadata) This datalayer is comprised of Miscellaineous Transmission Lines. Digital line graph (DLG) data are digital representations of cartographic...

  7. Same-source parallel implementation of the PSU/NCAR MM5

    Michalakes, J.

    1997-12-31

    The Pennsylvania State/National Center for Atmospheric Research Mesoscale Model is a limited-area model of atmospheric systems, now in its fifth generation, MM5. Designed and maintained for vector and shared-memory parallel architectures, the official version of MM5 does not run on message-passing distributed memory (DM) parallel computers. The authors describe a same-source parallel implementation of the PSU/NCAR MM5 using FLIC, the Fortran Loop and Index Converter. The resulting source is nearly line-for-line identical with the original source code. The result is an efficient distributed memory parallel option to MM5 that can be seamlessly integrated into the official version.

  8. A microeconomic scheduler for parallel computers

    Stoica, Ion; Abdel-Wahab, Hussein; Pothen, Alex

    1995-01-01

    We describe a scheduler based on the microeconomic paradigm for scheduling on-line a set of parallel jobs in a multiprocessor system. In addition to the classical objectives of increasing the system throughput and reducing the response time, we consider fairness in allocating system resources among the users, and providing the user with control over the relative performances of his jobs. We associate with every user a savings account in which he receives money at a constant rate. When a user wants to run a job, he creates an expense account for that job to which he transfers money from his savings account. The job uses the funds in its expense account to obtain the system resources it needs for execution. The share of the system resources allocated to the user is directly related to the rate at which the user receives money; the rate at which the user transfers money into a job expense account controls the job's performance. We prove that starvation is not possible in our model. Simulation results show that our scheduler improves both system and user performances in comparison with two different variable partitioning policies. It is also shown to be effective in guaranteeing fairness and providing control over the performance of jobs.

  9. PDDP, A Data Parallel Programming Model

    Karen H. Warren

    1996-01-01

    Full Text Available PDDP, the parallel data distribution preprocessor, is a data parallel programming model for distributed memory parallel computers. PDDP implements high-performance Fortran-compatible data distribution directives and parallelism expressed by the use of Fortran 90 array syntax, the FORALL statement, and the WHERE construct. Distributed data objects belong to a global name space; other data objects are treated as local and replicated on each processor. PDDP allows the user to program in a shared memory style and generates codes that are portable to a variety of parallel machines. For interprocessor communication, PDDP uses the fastest communication primitives on each platform.

  10. Parallelization of quantum molecular dynamics simulation code

    Kato, Kaori; Kunugi, Tomoaki; Shibahara, Masahiko; Kotake, Susumu

    1998-02-01

    A quantum molecular dynamics simulation code has been developed for the analysis of the thermalization of photon energies in the molecule or materials in Kansai Research Establishment. The simulation code is parallelized for both Scalar massively parallel computer (Intel Paragon XP/S75) and Vector parallel computer (Fujitsu VPP300/12). Scalable speed-up has been obtained with a distribution to processor units by division of particle group in both parallel computers. As a result of distribution to processor units not only by particle group but also by the particles calculation that is constructed with fine calculations, highly parallelization performance is achieved in Intel Paragon XP/S75. (author)

  11. Implementation and performance of parallelized elegant

    Wang, Y.; Borland, M.

    2008-01-01

    The program elegant is widely used for design and modeling of linacs for free-electron lasers and energy recovery linacs, as well as storage rings and other applications. As part of a multi-year effort, we have parallelized many aspects of the code, including single-particle dynamics, wakefields, and coherent synchrotron radiation. We report on the approach used for gradual parallelization, which proved very beneficial in getting parallel features into the hands of users quickly. We also report details of parallelization of collective effects. Finally, we discuss performance of the parallelized code in various applications.

  12. Increasing phylogenetic resolution at low taxonomic levels using massively parallel sequencing of chloroplast genomes

    Matthew Parks; Richard Cronn; Aaron Liston

    2009-01-01

    We reconstruct the infrageneric phylogeny of Pinus from 37 nearly-complete chloroplast genomes (average 109 kilobases each of an approximately 120 kilobase genome) generated using multiplexed massively parallel sequencing. We found that 30/33 ingroup nodes resolved wlth > 95-percent bootstrap support; this is a substantial improvement relative...

  13. A parallel algorithm for 3D dislocation dynamics

    Wang Zhiqiang; Ghoniem, Nasr; Swaminarayan, Sriram; LeSar, Richard

    2006-01-01

    Dislocation dynamics (DD), a discrete dynamic simulation method in which dislocations are the fundamental entities, is a powerful tool for investigation of plasticity, deformation and fracture of materials at the micron length scale. However, severe computational difficulties arising from complex, long-range interactions between these curvilinear line defects limit the application of DD in the study of large-scale plastic deformation. We present here the development of a parallel algorithm for accelerated computer simulations of DD. By representing dislocations as a 3D set of dislocation particles, we show here that the problem of an interacting ensemble of dislocations can be converted to a problem of a particle ensemble, interacting with a long-range force field. A grid using binary space partitioning is constructed to keep track of node connectivity across domains. We demonstrate the computational efficiency of the parallel micro-plasticity code and discuss how O(N) methods map naturally onto the parallel data structure. Finally, we present results from applications of the parallel code to deformation in single crystal fcc metals

  14. User-friendly parallelization of GAUDI applications with Python

    Mato, Pere; Smith, Eoin

    2010-01-01

    GAUDI is a software framework in C++ used to build event data processing applications using a set of standard components with well-defined interfaces. Simulation, high-level trigger, reconstruction, and analysis programs used by several experiments are developed using GAUDI. These applications can be configured and driven by simple Python scripts. Given the fact that a considerable amount of existing software has been developed using serial methodology, and has existed in some cases for many years, implementation of parallelisation techniques at the framework level may offer a way of exploiting current multi-core technologies to maximize performance and reduce latencies without re-writing thousands/millions of lines of code. In the solution we have developed, the parallelization techniques are introduced to the high level Python scripts which configure and drive the applications, such that the core C++ application code requires no modification, and that end users need make only minimal changes to their scripts. The developed solution leverages from existing generic Python modules that support parallel processing. Naturally, the parallel version of a given program should produce results consistent with its serial execution. The evaluation of several prototypes incorporating various parallelization techniques are presented and discussed.

  15. User-friendly parallelization of GAUDI applications with Python

    Mato, Pere; Smith, Eoin, E-mail: pere.mato@cern.c [PH Department, CERN, 1211 Geneva 23 (Switzerland)

    2010-04-01

    GAUDI is a software framework in C++ used to build event data processing applications using a set of standard components with well-defined interfaces. Simulation, high-level trigger, reconstruction, and analysis programs used by several experiments are developed using GAUDI. These applications can be configured and driven by simple Python scripts. Given the fact that a considerable amount of existing software has been developed using serial methodology, and has existed in some cases for many years, implementation of parallelisation techniques at the framework level may offer a way of exploiting current multi-core technologies to maximize performance and reduce latencies without re-writing thousands/millions of lines of code. In the solution we have developed, the parallelization techniques are introduced to the high level Python scripts which configure and drive the applications, such that the core C++ application code requires no modification, and that end users need make only minimal changes to their scripts. The developed solution leverages from existing generic Python modules that support parallel processing. Naturally, the parallel version of a given program should produce results consistent with its serial execution. The evaluation of several prototypes incorporating various parallelization techniques are presented and discussed.

  16. Parallelization of 2-D lattice Boltzmann codes

    Suzuki, Soichiro; Kaburaki, Hideo; Yokokawa, Mitsuo.

    1996-03-01

    Lattice Boltzmann (LB) codes to simulate two dimensional fluid flow are developed on vector parallel computer Fujitsu VPP500 and scalar parallel computer Intel Paragon XP/S. While a 2-D domain decomposition method is used for the scalar parallel LB code, a 1-D domain decomposition method is used for the vector parallel LB code to be vectorized along with the axis perpendicular to the direction of the decomposition. High parallel efficiency of 95.1% by the vector parallel calculation on 16 processors with 1152x1152 grid and 88.6% by the scalar parallel calculation on 100 processors with 800x800 grid are obtained. The performance models are developed to analyze the performance of the LB codes. It is shown by our performance models that the execution speed of the vector parallel code is about one hundred times faster than that of the scalar parallel code with the same number of processors up to 100 processors. We also analyze the scalability in keeping the available memory size of one processor element at maximum. Our performance model predicts that the execution time of the vector parallel code increases about 3% on 500 processors. Although the 1-D domain decomposition method has in general a drawback in the interprocessor communication, the vector parallel LB code is still suitable for the large scale and/or high resolution simulations. (author)

  17. Parallelization of 2-D lattice Boltzmann codes

    Suzuki, Soichiro; Kaburaki, Hideo; Yokokawa, Mitsuo

    1996-03-01

    Lattice Boltzmann (LB) codes to simulate two dimensional fluid flow are developed on vector parallel computer Fujitsu VPP500 and scalar parallel computer Intel Paragon XP/S. While a 2-D domain decomposition method is used for the scalar parallel LB code, a 1-D domain decomposition method is used for the vector parallel LB code to be vectorized along with the axis perpendicular to the direction of the decomposition. High parallel efficiency of 95.1% by the vector parallel calculation on 16 processors with 1152x1152 grid and 88.6% by the scalar parallel calculation on 100 processors with 800x800 grid are obtained. The performance models are developed to analyze the performance of the LB codes. It is shown by our performance models that the execution speed of the vector parallel code is about one hundred times faster than that of the scalar parallel code with the same number of processors up to 100 processors. We also analyze the scalability in keeping the available memory size of one processor element at maximum. Our performance model predicts that the execution time of the vector parallel code increases about 3% on 500 processors. Although the 1-D domain decomposition method has in general a drawback in the interprocessor communication, the vector parallel LB code is still suitable for the large scale and/or high resolution simulations. (author).

  18. Systematic approach for deriving feasible mappings of parallel algorithms to parallel computing platforms

    Arkin, Ethem; Tekinerdogan, Bedir; Imre, Kayhan M.

    2017-01-01

    The need for high-performance computing together with the increasing trend from single processor to parallel computer architectures has leveraged the adoption of parallel computing. To benefit from parallel computing power, usually parallel algorithms are defined that can be mapped and executed

  19. Experiences in Data-Parallel Programming

    Terry W. Clark

    1997-01-01

    Full Text Available To efficiently parallelize a scientific application with a data-parallel compiler requires certain structural properties in the source program, and conversely, the absence of others. A recent parallelization effort of ours reinforced this observation and motivated this correspondence. Specifically, we have transformed a Fortran 77 version of GROMOS, a popular dusty-deck program for molecular dynamics, into Fortran D, a data-parallel dialect of Fortran. During this transformation we have encountered a number of difficulties that probably are neither limited to this particular application nor do they seem likely to be addressed by improved compiler technology in the near future. Our experience with GROMOS suggests a number of points to keep in mind when developing software that may at some time in its life cycle be parallelized with a data-parallel compiler. This note presents some guidelines for engineering data-parallel applications that are compatible with Fortran D or High Performance Fortran compilers.

  20. Streaming for Functional Data-Parallel Languages

    Madsen, Frederik Meisner

    In this thesis, we investigate streaming as a general solution to the space inefficiency commonly found in functional data-parallel programming languages. The data-parallel paradigm maps well to parallel SIMD-style hardware. However, the traditional fully materializing execution strategy...... by extending two existing data-parallel languages: NESL and Accelerate. In the extensions we map bulk operations to data-parallel streams that can evaluate fully sequential, fully parallel or anything in between. By a dataflow, piecewise parallel execution strategy, the runtime system can adjust to any target...... flattening necessitates all sub-computations to materialize at the same time. For example, naive n by n matrix multiplication requires n^3 space in NESL because the algorithm contains n^3 independent scalar multiplications. For large values of n, this is completely unacceptable. We address the problem...

  1. An Integrated Inductor For Parallel Interleaved Three-Phase Voltage Source Converters

    Gohil, Ghanshyamsinh Vijaysinh; Bede, Lorand; Teodorescu, Remus

    2016-01-01

    Three phase Voltage Source Converters (VSCs) are often connected in parallel to realize high current output converter system. The harmonic quality of the resultant switched output voltage can be improved by interleaving the carrier signals of these parallel connected VSCs. As a result, the line...... of the state-of-the-art filtering solution. The performance of the integrated inductor is also verified by the experimental measurements....

  2. Massively parallel diffuse optical tomography

    Sandusky, John V.; Pitts, Todd A.

    2017-09-05

    Diffuse optical tomography systems and methods are described herein. In a general embodiment, the diffuse optical tomography system comprises a plurality of sensor heads, the plurality of sensor heads comprising respective optical emitter systems and respective sensor systems. A sensor head in the plurality of sensors heads is caused to act as an illuminator, such that its optical emitter system transmits a transillumination beam towards a portion of a sample. Other sensor heads in the plurality of sensor heads act as observers, detecting portions of the transillumination beam that radiate from the sample in the fields of view of the respective sensory systems of the other sensor heads. Thus, sensor heads in the plurality of sensors heads generate sensor data in parallel.

  3. Embodied and Distributed Parallel DJing.

    Cappelen, Birgitta; Andersson, Anders-Petter

    2016-01-01

    Everyone has a right to take part in cultural events and activities, such as music performances and music making. Enforcing that right, within Universal Design, is often limited to a focus on physical access to public areas, hearing aids etc., or groups of persons with special needs performing in traditional ways. The latter might be people with disabilities, being musicians playing traditional instruments, or actors playing theatre. In this paper we focus on the innovative potential of including people with special needs, when creating new cultural activities. In our project RHYME our goal was to create health promoting activities for children with severe disabilities, by developing new musical and multimedia technologies. Because of the users' extreme demands and rich contribution, we ended up creating both a new genre of musical instruments and a new art form. We call this new art form Embodied and Distributed Parallel DJing, and the new genre of instruments for Empowering Multi-Sensorial Things.

  4. Device for balancing parallel strings

    Mashikian, Matthew S.

    1985-01-01

    A battery plant is described which features magnetic circuit means in association with each of the battery strings in the battery plant for balancing the electrical current flow through the battery strings by equalizing the voltage across each of the battery strings. Each of the magnetic circuit means generally comprises means for sensing the electrical current flow through one of the battery strings, and a saturable reactor having a main winding connected electrically in series with the battery string, a bias winding connected to a source of alternating current and a control winding connected to a variable source of direct current controlled by the sensing means. Each of the battery strings is formed by a plurality of batteries connected electrically in series, and these battery strings are connected electrically in parallel across common bus conductors.

  5. Linear parallel processing machines I

    Von Kunze, M

    1984-01-01

    As is well-known, non-context-free grammars for generating formal languages happen to be of a certain intrinsic computational power that presents serious difficulties to efficient parsing algorithms as well as for the development of an algebraic theory of contextsensitive languages. In this paper a framework is given for the investigation of the computational power of formal grammars, in order to start a thorough analysis of grammars consisting of derivation rules of the form aB ..-->.. A/sub 1/ ... A /sub n/ b/sub 1/...b /sub m/ . These grammars may be thought of as automata by means of parallel processing, if one considers the variables as operators acting on the terminals while reading them right-to-left. This kind of automata and their 2-dimensional programming language prove to be useful by allowing a concise linear-time algorithm for integer multiplication. Linear parallel processing machines (LP-machines) which are, in their general form, equivalent to Turing machines, include finite automata and pushdown automata (with states encoded) as special cases. Bounded LP-machines yield deterministic accepting automata for nondeterministic contextfree languages, and they define an interesting class of contextsensitive languages. A characterization of this class in terms of generating grammars is established by using derivation trees with crossings as a helpful tool. From the algebraic point of view, deterministic LP-machines are effectively represented semigroups with distinguished subsets. Concerning the dualism between generating and accepting devices of formal languages within the algebraic setting, the concept of accepting automata turns out to reduce essentially to embeddability in an effectively represented extension monoid, even in the classical cases.

  6. Parallel computing in enterprise modeling.

    Goldsby, Michael E.; Armstrong, Robert C.; Shneider, Max S.; Vanderveen, Keith; Ray, Jaideep; Heath, Zach; Allan, Benjamin A.

    2008-08-01

    This report presents the results of our efforts to apply high-performance computing to entity-based simulations with a multi-use plugin for parallel computing. We use the term 'Entity-based simulation' to describe a class of simulation which includes both discrete event simulation and agent based simulation. What simulations of this class share, and what differs from more traditional models, is that the result sought is emergent from a large number of contributing entities. Logistic, economic and social simulations are members of this class where things or people are organized or self-organize to produce a solution. Entity-based problems never have an a priori ergodic principle that will greatly simplify calculations. Because the results of entity-based simulations can only be realized at scale, scalable computing is de rigueur for large problems. Having said that, the absence of a spatial organizing principal makes the decomposition of the problem onto processors problematic. In addition, practitioners in this domain commonly use the Java programming language which presents its own problems in a high-performance setting. The plugin we have developed, called the Parallel Particle Data Model, overcomes both of these obstacles and is now being used by two Sandia frameworks: the Decision Analysis Center, and the Seldon social simulation facility. While the ability to engage U.S.-sized problems is now available to the Decision Analysis Center, this plugin is central to the success of Seldon. Because Seldon relies on computationally intensive cognitive sub-models, this work is necessary to achieve the scale necessary for realistic results. With the recent upheavals in the financial markets, and the inscrutability of terrorist activity, this simulation domain will likely need a capability with ever greater fidelity. High-performance computing will play an important part in enabling that greater fidelity.

  7. Research on Control Strategy of Complex Systems through VSC-HVDC Grid Parallel Device

    Xue Mei-Juan

    2014-07-01

    Full Text Available After the completion of grid parallel, the device can turn to be UPFC, STATCOM, SSSC, research on the conversion circuit and transform method by corresponding switching operation. Accomplish the grid parallel and comprehensive control of the tie-line and stable operation and control functions of grid after parallel. Defines the function select operation switch matrix and grid parallel system branch variable, forming a switch matrix to achieve corresponding function of the composite system. Formed a criterion of the selection means to choose control strategy according to the switch matrix, to accomplish corresponding function. Put the grid parallel, STATCOM, SSSC and UPFC together as a system, improve the stable operation and flexible control of the power system.

  8. The U-line line balancing problem

    Miltenburg, G.J.; Wijngaard, J.

    1994-01-01

    The traditional line balancing (LB) problem considers a production line in which stations are arranged consecutively in a line. A balance is determined by grouping tasks into stations while moving forward (or backward) through a precedence network. Recently many production lines are being arranged

  9. A parallel algorithm for filtering gravitational waves from coalescing binaries

    Sathyaprakash, B.S.; Dhurandhar, S.V.

    1992-10-01

    Coalescing binary stars are perhaps the most promising sources for the observation of gravitational waves with laser interferometric gravity wave detectors. The waveform from these sources can be predicted with sufficient accuracy for matched filtering techniques to be applied. In this paper we present a parallel algorithm for detecting signals from coalescing compact binaries by the method of matched filtering. We also report the details of its implementation on a 256-node connection machine consisting of a network of transputers. The results of our analysis indicate that parallel processing is a promising approach to on-line analysis of data from gravitational wave detectors to filter out coalescing binary signals. The algorithm described is quite general in that the kernel of the algorithm is applicable to any set of matched filters. (author). 15 refs, 4 figs

  10. Compiler Technology for Parallel Scientific Computation

    Can Özturan

    1994-01-01

    Full Text Available There is a need for compiler technology that, given the source program, will generate efficient parallel codes for different architectures with minimal user involvement. Parallel computation is becoming indispensable in solving large-scale problems in science and engineering. Yet, the use of parallel computation is limited by the high costs of developing the needed software. To overcome this difficulty we advocate a comprehensive approach to the development of scalable architecture-independent software for scientific computation based on our experience with equational programming language (EPL. Our approach is based on a program decomposition, parallel code synthesis, and run-time support for parallel scientific computation. The program decomposition is guided by the source program annotations provided by the user. The synthesis of parallel code is based on configurations that describe the overall computation as a set of interacting components. Run-time support is provided by the compiler-generated code that redistributes computation and data during object program execution. The generated parallel code is optimized using techniques of data alignment, operator placement, wavefront determination, and memory optimization. In this article we discuss annotations, configurations, parallel code generation, and run-time support suitable for parallel programs written in the functional parallel programming language EPL and in Fortran.

  11. Computer-Aided Parallelizer and Optimizer

    Jin, Haoqiang

    2011-01-01

    The Computer-Aided Parallelizer and Optimizer (CAPO) automates the insertion of compiler directives (see figure) to facilitate parallel processing on Shared Memory Parallel (SMP) machines. While CAPO currently is integrated seamlessly into CAPTools (developed at the University of Greenwich, now marketed as ParaWise), CAPO was independently developed at Ames Research Center as one of the components for the Legacy Code Modernization (LCM) project. The current version takes serial FORTRAN programs, performs interprocedural data dependence analysis, and generates OpenMP directives. Due to the widely supported OpenMP standard, the generated OpenMP codes have the potential to run on a wide range of SMP machines. CAPO relies on accurate interprocedural data dependence information currently provided by CAPTools. Compiler directives are generated through identification of parallel loops in the outermost level, construction of parallel regions around parallel loops and optimization of parallel regions, and insertion of directives with automatic identification of private, reduction, induction, and shared variables. Attempts also have been made to identify potential pipeline parallelism (implemented with point-to-point synchronization). Although directives are generated automatically, user interaction with the tool is still important for producing good parallel codes. A comprehensive graphical user interface is included for users to interact with the parallelization process.

  12. Implementation of PHENIX trigger algorithms on massively parallel computers

    Petridis, A.N.; Wohn, F.K.

    1995-01-01

    The event selection requirements of contemporary high energy and nuclear physics experiments are met by the introduction of on-line trigger algorithms which identify potentially interesting events and reduce the data acquisition rate to levels that are manageable by the electronics. Such algorithms being parallel in nature can be simulated off-line using massively parallel computers. The PHENIX experiment intends to investigate the possible existence of a new phase of matter called the quark gluon plasma which has been theorized to have existed in very early stages of the evolution of the universe by studying collisions of heavy nuclei at ultra-relativistic energies. Such interactions can also reveal important information regarding the structure of the nucleus and mandate a thorough investigation of the simpler proton-nucleus collisions at the same energies. The complexity of PHENIX events and the need to analyze and also simulate them at rates similar to the data collection ones imposes enormous computation demands. This work is a first effort to implement PHENIX trigger algorithms on parallel computers and to study the feasibility of using such machines to run the complex programs necessary for the simulation of the PHENIX detector response. Fine and coarse grain approaches have been studied and evaluated. Depending on the application the performance of a massively parallel computer can be much better or much worse than that of a serial workstation. A comparison between single instruction and multiple instruction computers is also made and possible applications of the single instruction machines to high energy and nuclear physics experiments are outlined. copyright 1995 American Institute of Physics

  13. Automatic Thread-Level Parallelization in the Chombo AMR Library

    Christen, Matthias; Keen, Noel; Ligocki, Terry; Oliker, Leonid; Shalf, John; Van Straalen, Brian; Williams, Samuel

    2011-05-26

    The increasing on-chip parallelism has some substantial implications for HPC applications. Currently, hybrid programming models (typically MPI+OpenMP) are employed for mapping software to the hardware in order to leverage the hardware?s architectural features. In this paper, we present an approach that automatically introduces thread level parallelism into Chombo, a parallel adaptive mesh refinement framework for finite difference type PDE solvers. In Chombo, core algorithms are specified in the ChomboFortran, a macro language extension to F77 that is part of the Chombo framework. This domain-specific language forms an already used target language for an automatic migration of the large number of existing algorithms into a hybrid MPI+OpenMP implementation. It also provides access to the auto-tuning methodology that enables tuning certain aspects of an algorithm to hardware characteristics. Performance measurements are presented for a few of the most relevant kernels with respect to a specific application benchmark using this technique as well as benchmark results for the entire application. The kernel benchmarks show that, using auto-tuning, up to a factor of 11 in performance was gained with 4 threads with respect to the serial reference implementation.

  14. Parallel solutions of the two-group neutron diffusion equations

    Zee, K.S.; Turinsky, P.J.

    1987-01-01

    Recent efforts to adapt various numerical solution algorithms to parallel computer architectures have addressed the possibility of substantially reducing the running time of few-group neutron diffusion calculations. The authors have developed an efficient iterative parallel algorithm and an associated computer code for the rapid solution of the finite difference method representation of the two-group neutron diffusion equations on the CRAY X/MP-48 supercomputer having multi-CPUs and vector pipelines. For realistic simulation of light water reactor cores, the code employees a macroscopic depletion model with trace capability for selected fission product transients and critical boron. In addition to this, moderator and fuel temperature feedback models are also incorporated into the code. The validity of the physics models used in the code were benchmarked against qualified codes and proved accurate. This work is an extension of previous work in that various feedback effects are accounted for in the system; the entire code is structured to accommodate extensive vectorization; and an additional parallelism by multitasking is achieved not only for the solution of the matrix equations associated with the inner iterations but also for the other segments of the code, e.g., outer iterations

  15. Increasing the reach of forensic genetics with massively parallel sequencing.

    Budowle, Bruce; Schmedes, Sarah E; Wendt, Frank R

    2017-09-01

    The field of forensic genetics has made great strides in the analysis of biological evidence related to criminal and civil matters. More so, the discipline has set a standard of performance and quality in the forensic sciences. The advent of massively parallel sequencing will allow the field to expand its capabilities substantially. This review describes the salient features of massively parallel sequencing and how it can impact forensic genetics. The features of this technology offer increased number and types of genetic markers that can be analyzed, higher throughput of samples, and the capability of targeting different organisms, all by one unifying methodology. While there are many applications, three are described where massively parallel sequencing will have immediate impact: molecular autopsy, microbial forensics and differentiation of monozygotic twins. The intent of this review is to expose the forensic science community to the potential enhancements that have or are soon to arrive and demonstrate the continued expansion the field of forensic genetics and its service in the investigation of legal matters.

  16. Massively parallel algorithms for trace-driven cache simulations

    Nicol, David M.; Greenberg, Albert G.; Lubachevsky, Boris D.

    1991-01-01

    Trace driven cache simulation is central to computer design. A trace is a very long sequence of reference lines from main memory. At the t(exp th) instant, reference x sub t is hashed into a set of cache locations, the contents of which are then compared with x sub t. If at the t sup th instant x sub t is not present in the cache, then it is said to be a miss, and is loaded into the cache set, possibly forcing the replacement of some other memory line, and making x sub t present for the (t+1) sup st instant. The problem of parallel simulation of a subtrace of N references directed to a C line cache set is considered, with the aim of determining which references are misses and related statistics. A simulation method is presented for the Least Recently Used (LRU) policy, which regradless of the set size C runs in time O(log N) using N processors on the exclusive read, exclusive write (EREW) parallel model. A simpler LRU simulation algorithm is given that runs in O(C log N) time using N/log N processors. Timings are presented of the second algorithm's implementation on the MasPar MP-1, a machine with 16384 processors. A broad class of reference based line replacement policies are considered, which includes LRU as well as the Least Frequently Used and Random replacement policies. A simulation method is presented for any such policy that on any trace of length N directed to a C line set runs in the O(C log N) time with high probability using N processors on the EREW model. The algorithms are simple, have very little space overhead, and are well suited for SIMD implementation.

  17. Information Technology & Applications Corporation v. United States: An Interested Party's "Substantial Chance" at APA Standing

    Slicker, Christina

    2003-01-01

    .... Building on CICA's "interested party" definition with Information Technology's refinement of "substantial chance" rule, the Federal Circuit has effectively translated "APA standing" into the language...

  18. 20 CFR 416.972 - What we mean by substantial gainful activity.

    2010-04-01

    ... INCOME FOR THE AGED, BLIND, AND DISABLED Determining Disability and Blindness Substantial Gainful... consider activities like taking care of yourself, household tasks, hobbies, therapy, school attendance...

  19. Parallel processing for fluid dynamics applications

    Johnson, G.M.

    1989-01-01

    The impact of parallel processing on computational science and, in particular, on computational fluid dynamics is growing rapidly. In this paper, particular emphasis is given to developments which have occurred within the past two years. Parallel processing is defined and the reasons for its importance in high-performance computing are reviewed. Parallel computer architectures are classified according to the number and power of their processing units, their memory, and the nature of their connection scheme. Architectures which show promise for fluid dynamics applications are emphasized. Fluid dynamics problems are examined for parallelism inherent at the physical level. CFD algorithms and their mappings onto parallel architectures are discussed. Several example are presented to document the performance of fluid dynamics applications on present-generation parallel processing devices

  20. Design considerations for parallel graphics libraries

    Crockett, Thomas W.

    1994-01-01

    Applications which run on parallel supercomputers are often characterized by massive datasets. Converting these vast collections of numbers to visual form has proven to be a powerful aid to comprehension. For a variety of reasons, it may be desirable to provide this visual feedback at runtime. One way to accomplish this is to exploit the available parallelism to perform graphics operations in place. In order to do this, we need appropriate parallel rendering algorithms and library interfaces. This paper provides a tutorial introduction to some of the issues which arise in designing parallel graphics libraries and their underlying rendering algorithms. The focus is on polygon rendering for distributed memory message-passing systems. We illustrate our discussion with examples from PGL, a parallel graphics library which has been developed on the Intel family of parallel systems.

  1. Coiled transmission line pulse generators

    McDonald, Kenneth Fox

    2010-11-09

    Methods and apparatus are provided for fabricating and constructing solid dielectric "Coiled Transmission Line" pulse generators in radial or axial coiled geometries. The pour and cure fabrication process enables a wide variety of geometries and form factors. The volume between the conductors is filled with liquid blends of monomers, polymers, oligomers, and/or cross-linkers and dielectric powders; and then cured to form high field strength and high dielectric constant solid dielectric transmission lines that intrinsically produce ideal rectangular high voltage pulses when charged and switched into matched impedance loads. Voltage levels may be increased by Marx and/or Blumlein principles incorporating spark gap or, preferentially, solid state switches (such as optically triggered thyristors) which produce reliable, high repetition rate operation. Moreover, these Marxed pulse generators can be DC charged and do not require additional pulse forming circuitry, pulse forming lines, transformers, or an a high voltage spark gap output switch. The apparatus accommodates a wide range of voltages, impedances, pulse durations, pulse repetition rates, and duty cycles. The resulting mobile or flight platform friendly cylindrical geometric configuration is much more compact, light-weight, and robust than conventional linear geometries, or pulse generators constructed from conventional components. Installing additional circuitry may accommodate optional pulse shape improvements. The Coiled Transmission Lines can also be connected in parallel to decrease the impedance, or in series to increase the pulse length.

  2. Synchronization Techniques in Parallel Discrete Event Simulation

    Lindén, Jonatan

    2018-01-01

    Discrete event simulation is an important tool for evaluating system models in many fields of science and engineering. To improve the performance of large-scale discrete event simulations, several techniques to parallelize discrete event simulation have been developed. In parallel discrete event simulation, the work of a single discrete event simulation is distributed over multiple processing elements. A key challenge in parallel discrete event simulation is to ensure that causally dependent ...

  3. Parallel processing from applications to systems

    Moldovan, Dan I

    1993-01-01

    This text provides one of the broadest presentations of parallelprocessing available, including the structure of parallelprocessors and parallel algorithms. The emphasis is on mappingalgorithms to highly parallel computers, with extensive coverage ofarray and multiprocessor architectures. Early chapters provideinsightful coverage on the analysis of parallel algorithms andprogram transformations, effectively integrating a variety ofmaterial previously scattered throughout the literature. Theory andpractice are well balanced across diverse topics in this concisepresentation. For exceptional cla

  4. Parallel processing for artificial intelligence 1

    Kanal, LN; Kumar, V; Suttner, CB

    1994-01-01

    Parallel processing for AI problems is of great current interest because of its potential for alleviating the computational demands of AI procedures. The articles in this book consider parallel processing for problems in several areas of artificial intelligence: image processing, knowledge representation in semantic networks, production rules, mechanization of logic, constraint satisfaction, parsing of natural language, data filtering and data mining. The publication is divided into six sections. The first addresses parallel computing for processing and understanding images. The second discus

  5. A survey of parallel multigrid algorithms

    Chan, Tony F.; Tuminaro, Ray S.

    1987-01-01

    A typical multigrid algorithm applied to well-behaved linear-elliptic partial-differential equations (PDEs) is described. Criteria for designing and evaluating parallel algorithms are presented. Before evaluating the performance of some parallel multigrid algorithms, consideration is given to some theoretical complexity results for solving PDEs in parallel and for executing the multigrid algorithm. The effect of mapping and load imbalance on the partial efficiency of the algorithm is studied.

  6. Refinement of Parallel and Reactive Programs

    Back, R. J. R.

    1992-01-01

    We show how to apply the refinement calculus to stepwise refinement of parallel and reactive programs. We use action systems as our basic program model. Action systems are sequential programs which can be implemented in a parallel fashion. Hence refinement calculus methods, originally developed for sequential programs, carry over to the derivation of parallel programs. Refinement of reactive programs is handled by data refinement techniques originally developed for the sequential refinement c...

  7. Interactive animation of fault-tolerant parallel algorithms

    Apgar, S.W.

    1992-02-01

    Animation of algorithms makes understanding them intuitively easier. This paper describes the software tool Raft (Robust Animator of Fault Tolerant Algorithms). The Raft system allows the user to animate a number of parallel algorithms which achieve fault tolerant execution. In particular, we use it to illustrate the key Write-All problem. It has an extensive user-interface which allows a choice of the number of processors, the number of elements in the Write-All array, and the adversary to control the processor failures. The novelty of the system is that the interface allows the user to create new on-line adversaries as the algorithm executes.

  8. Parallel algorithms for finding cliques in a graph

    Szabo, S

    2011-01-01

    A clique is a subgraph in a graph that is complete in the sense that each two of its nodes are connected by an edge. Finding cliques in a given graph is an important procedure in discrete mathematical modeling. The paper will show how concepts such as splitting partitions, quasi coloring, node and edge dominance are related to clique search problems. In particular we will discuss the connection with parallel clique search algorithms. These concepts also suggest practical guide lines to inspect a given graph before starting a large scale search.

  9. Computing NLTE Opacities -- Node Level Parallel Calculation

    Holladay, Daniel [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2015-09-11

    Presentation. The goal: to produce a robust library capable of computing reasonably accurate opacities inline with the assumption of LTE relaxed (non-LTE). Near term: demonstrate acceleration of non-LTE opacity computation. Far term (if funded): connect to application codes with in-line capability and compute opacities. Study science problems. Use efficient algorithms that expose many levels of parallelism and utilize good memory access patterns for use on advanced architectures. Portability to multiple types of hardware including multicore processors, manycore processors such as KNL, GPUs, etc. Easily coupled to radiation hydrodynamics and thermal radiative transfer codes.

  10. Systematic test on fast time resolution parallel plate avalanche counter

    Chen Yu; Li Guangwu; Gu Xianbao; Chen Yanchao; Zhang Gang; Zhang Wenhui; Yan Guohong

    2011-01-01

    Systematic test on each detect unit of parallel plate avalanche counter (PPAC) used in the fission multi-parameter measurement was performed with a 241 Am α source to get the time resolution and position resolution. The detectors work at 600 Pa flowing isobutane and with-600 V on cathode. The time resolution was got by TOF method and the position resolution was got by delay line method. The time resolution of detect units is better than 400 ps, and the position resolution is 6 mm. The results show that the demand of measurement is fully covered. (authors)

  11. Parallel Prediction of Stock Volatility

    Priscilla Jenq

    2017-10-01

    Full Text Available Volatility is a measurement of the risk of financial products. A stock will hit new highs and lows over time and if these highs and lows fluctuate wildly, then it is considered a high volatile stock. Such a stock is considered riskier than a stock whose volatility is low. Although highly volatile stocks are riskier, the returns that they generate for investors can be quite high. Of course, with a riskier stock also comes the chance of losing money and yielding negative returns. In this project, we will use historic stock data to help us forecast volatility. Since the financial industry usually uses S&P 500 as the indicator of the market, we will use S&P 500 as a benchmark to compute the risk. We will also use artificial neural networks as a tool to predict volatilities for a specific time frame that will be set when we configure this neural network. There have been reports that neural networks with different numbers of layers and different numbers of hidden nodes may generate varying results. In fact, we may be able to find the best configuration of a neural network to compute volatilities. We will implement this system using the parallel approach. The system can be used as a tool for investors to allocating and hedging assets.

  12. Vectoring of parallel synthetic jets

    Berk, Tim; Ganapathisubramani, Bharathram; Gomit, Guillaume

    2015-11-01

    A pair of parallel synthetic jets can be vectored by applying a phase difference between the two driving signals. The resulting jet can be merged or bifurcated and either vectored towards the actuator leading in phase or the actuator lagging in phase. In the present study, the influence of phase difference and Strouhal number on the vectoring behaviour is examined experimentally. Phase-locked vorticity fields, measured using Particle Image Velocimetry (PIV), are used to track vortex pairs. The physical mechanisms that explain the diversity in vectoring behaviour are observed based on the vortex trajectories. For a fixed phase difference, the vectoring behaviour is shown to be primarily influenced by pinch-off time of vortex rings generated by the synthetic jets. Beyond a certain formation number, the pinch-off timescale becomes invariant. In this region, the vectoring behaviour is determined by the distance between subsequent vortex rings. We acknowledge the financial support from the European Research Council (ERC grant agreement no. 277472).

  13. A Soft Parallel Kinematic Mechanism.

    White, Edward L; Case, Jennifer C; Kramer-Bottiglio, Rebecca

    2018-02-01

    In this article, we describe a novel holonomic soft robotic structure based on a parallel kinematic mechanism. The design is based on the Stewart platform, which uses six sensors and actuators to achieve full six-degree-of-freedom motion. Our design is much less complex than a traditional platform, since it replaces the 12 spherical and universal joints found in a traditional Stewart platform with a single highly deformable elastomer body and flexible actuators. This reduces the total number of parts in the system and simplifies the assembly process. Actuation is achieved through coiled-shape memory alloy actuators. State observation and feedback is accomplished through the use of capacitive elastomer strain gauges. The main structural element is an elastomer joint that provides antagonistic force. We report the response of the actuators and sensors individually, then report the response of the complete assembly. We show that the completed robotic system is able to achieve full position control, and we discuss the limitations associated with using responsive material actuators. We believe that control demonstrated on a single body in this work could be extended to chains of such bodies to create complex soft robots.

  14. Productive Parallel Programming: The PCN Approach

    Ian Foster

    1992-01-01

    Full Text Available We describe the PCN programming system, focusing on those features designed to improve the productivity of scientists and engineers using parallel supercomputers. These features include a simple notation for the concise specification of concurrent algorithms, the ability to incorporate existing Fortran and C code into parallel applications, facilities for reusing parallel program components, a portable toolkit that allows applications to be developed on a workstation or small parallel computer and run unchanged on supercomputers, and integrated debugging and performance analysis tools. We survey representative scientific applications and identify problem classes for which PCN has proved particularly useful.

  15. High performance parallel I/O

    Prabhat

    2014-01-01

    Gain Critical Insight into the Parallel I/O EcosystemParallel I/O is an integral component of modern high performance computing (HPC), especially in storing and processing very large datasets to facilitate scientific discovery. Revealing the state of the art in this field, High Performance Parallel I/O draws on insights from leading practitioners, researchers, software architects, developers, and scientists who shed light on the parallel I/O ecosystem.The first part of the book explains how large-scale HPC facilities scope, configure, and operate systems, with an emphasis on choices of I/O har

  16. Parallel, Rapid Diffuse Optical Tomography of Breast

    Yodh, Arjun

    2001-01-01

    During the last year we have experimentally and computationally investigated rapid acquisition and analysis of informationally dense diffuse optical data sets in the parallel plate compressed breast geometry...

  17. Parallel, Rapid Diffuse Optical Tomography of Breast

    Yodh, Arjun

    2002-01-01

    During the last year we have experimentally and computationally investigated rapid acquisition and analysis of informationally dense diffuse optical data sets in the parallel plate compressed breast geometry...

  18. Parallel auto-correlative statistics with VTK.

    Pebay, Philippe Pierre; Bennett, Janine Camille

    2013-08-01

    This report summarizes existing statistical engines in VTK and presents both the serial and parallel auto-correlative statistics engines. It is a sequel to [PT08, BPRT09b, PT09, BPT09, PT10] which studied the parallel descriptive, correlative, multi-correlative, principal component analysis, contingency, k-means, and order statistics engines. The ease of use of the new parallel auto-correlative statistics engine is illustrated by the means of C++ code snippets and algorithm verification is provided. This report justifies the design of the statistics engines with parallel scalability in mind, and provides scalability and speed-up analysis results for the autocorrelative statistics engine.

  19. Compiling Scientific Programs for Scalable Parallel Systems

    Kennedy, Ken

    2001-01-01

    ...). The research performed in this project included new techniques for recognizing implicit parallelism in sequential programs, a powerful and precise set-based framework for analysis and transformation...

  20. Parallel thermal radiation transport in two dimensions

    Smedley-Stevenson, R.P.; Ball, S.R.

    2003-01-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  1. Parallel Algorithms for the Exascale Era

    Robey, Robert W. [Los Alamos National Laboratory

    2016-10-19

    New parallel algorithms are needed to reach the Exascale level of parallelism with millions of cores. We look at some of the research developed by students in projects at LANL. The research blends ideas from the early days of computing while weaving in the fresh approach brought by students new to the field of high performance computing. We look at reproducibility of global sums and why it is important to parallel computing. Next we look at how the concept of hashing has led to the development of more scalable algorithms suitable for next-generation parallel computers. Nearly all of this work has been done by undergraduates and published in leading scientific journals.

  2. Parallel thermal radiation transport in two dimensions

    Smedley-Stevenson, R.P.; Ball, S.R. [AWE Aldermaston (United Kingdom)

    2003-07-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  3. Structured Parallel Programming Patterns for Efficient Computation

    McCool, Michael; Robison, Arch

    2012-01-01

    Programming is now parallel programming. Much as structured programming revolutionized traditional serial programming decades ago, a new kind of structured programming, based on patterns, is relevant to parallel programming today. Parallel computing experts and industry insiders Michael McCool, Arch Robison, and James Reinders describe how to design and implement maintainable and efficient parallel algorithms using a pattern-based approach. They present both theory and practice, and give detailed concrete examples using multiple programming models. Examples are primarily given using two of th

  4. Parallelization of a hydrological model using the message passing interface

    Wu, Yiping; Li, Tiejian; Sun, Liqun; Chen, Ji

    2013-01-01

    With the increasing knowledge about the natural processes, hydrological models such as the Soil and Water Assessment Tool (SWAT) are becoming larger and more complex with increasing computation time. Additionally, other procedures such as model calibration, which may require thousands of model iterations, can increase running time and thus further reduce rapid modeling and analysis. Using the widely-applied SWAT as an example, this study demonstrates how to parallelize a serial hydrological model in a Windows® environment using a parallel programing technology—Message Passing Interface (MPI). With a case study, we derived the optimal values for the two parameters (the number of processes and the corresponding percentage of work to be distributed to the master process) of the parallel SWAT (P-SWAT) on an ordinary personal computer and a work station. Our study indicates that model execution time can be reduced by 42%–70% (or a speedup of 1.74–3.36) using multiple processes (two to five) with a proper task-distribution scheme (between the master and slave processes). Although the computation time cost becomes lower with an increasing number of processes (from two to five), this enhancement becomes less due to the accompanied increase in demand for message passing procedures between the master and all slave processes. Our case study demonstrates that the P-SWAT with a five-process run may reach the maximum speedup, and the performance can be quite stable (fairly independent of a project size). Overall, the P-SWAT can help reduce the computation time substantially for an individual model run, manual and automatic calibration procedures, and optimization of best management practices. In particular, the parallelization method we used and the scheme for deriving the optimal parameters in this study can be valuable and easily applied to other hydrological or environmental models.

  5. Examination of Speed Contribution of Parallelization for Several Fingerprint Pre-Processing Algorithms

    GORGUNOGLU, S.

    2014-05-01

    Full Text Available In analysis of minutiae based fingerprint systems, fingerprints needs to be pre-processed. The pre-processing is carried out to enhance the quality of the fingerprint and to obtain more accurate minutiae points. Reducing the pre-processing time is important for identification and verification in real time systems and especially for databases holding large fingerprints information. Parallel processing and parallel CPU computing can be considered as distribution of processes over multi core processor. This is done by using parallel programming techniques. Reducing the execution time is the main objective in parallel processing. In this study, pre-processing of minutiae based fingerprint system is implemented by parallel processing on multi core computers using OpenMP and on graphics processor using CUDA to improve execution time. The execution times and speedup ratios are compared with the one that of single core processor. The results show that by using parallel processing, execution time is substantially improved. The improvement ratios obtained for different pre-processing algorithms allowed us to make suggestions on the more suitable approaches for parallelization.

  6. 26 CFR 1.42-7 - Substantially bond-financed buildings. [Reserved

    2010-04-01

    ... 26 Internal Revenue 1 2010-04-01 2010-04-01 true Substantially bond-financed buildings. [Reserved] 1.42-7 Section 1.42-7 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY INCOME TAX INCOME TAXES Credits Against Tax § 1.42-7 Substantially bond-financed buildings. [Reserved] ...

  7. 20 CFR 416.1080 - Notice of right to hearing on proposed finding of substantial failure.

    2010-04-01

    ... agency has substantially failed to meet our standards unless the State submits a written request for a... finding of substantial failure. 416.1080 Section 416.1080 Employees' Benefits SOCIAL SECURITY ADMINISTRATION SUPPLEMENTAL SECURITY INCOME FOR THE AGED, BLIND, AND DISABLED Determinations of Disability...

  8. 20 CFR 404.1680 - Notice of right to hearing on proposed finding of substantial failure.

    2010-04-01

    ... agency has substantially failed to meet our standards unless the State submits a written request for a... finding of substantial failure. 404.1680 Section 404.1680 Employees' Benefits SOCIAL SECURITY..., following the mandatory performance support period and the 3-month adjustment period, a State agency again...

  9. Detection of Dew-Point by substantial Raman Band Frequency Jumps (A new Method)

    Hansen, Susanne Brunsgaard; Berg, Rolf W.; Stenby, Erling Halfdan

    Detection of Dew-Point by substantial Raman Band Frequency Jumps (A new Method). See poster at http://www.kemi.dtu.dk/~ajo/rolf/jumps.pdf......Detection of Dew-Point by substantial Raman Band Frequency Jumps (A new Method). See poster at http://www.kemi.dtu.dk/~ajo/rolf/jumps.pdf...

  10. Substantiation of Optimum Stoker Number in Potassium Production by Game Method

    B. A. Bogatov

    2005-01-01

    Full Text Available Dump formation process of hard halite waste in potassium production has been analyzed in the paper. The paper shows high cost of a salt dump formation and an actuality to substantiate number of ОШ-1900-110/150 stokers that are operating simultaneously. A game theory method has been used to substantiate a solution of the problem.

  11. 32 CFR 48.302 - Substantiating evidence regarding dependency and age of dependents.

    2010-07-01

    ... 32 National Defense 1 2010-07-01 2010-07-01 false Substantiating evidence regarding dependency and age of dependents. 48.302 Section 48.302 National Defense Department of Defense OFFICE OF THE... Designation of Beneficiaries § 48.302 Substantiating evidence regarding dependency and age of dependents. At...

  12. Parallel Computing for Brain Simulation.

    Pastur-Romay, L A; Porto-Pazos, A B; Cedron, F; Pazos, A

    2017-01-01

    The human brain is the most complex system in the known universe, it is therefore one of the greatest mysteries. It provides human beings with extraordinary abilities. However, until now it has not been understood yet how and why most of these abilities are produced. For decades, researchers have been trying to make computers reproduce these abilities, focusing on both understanding the nervous system and, on processing data in a more efficient way than before. Their aim is to make computers process information similarly to the brain. Important technological developments and vast multidisciplinary projects have allowed creating the first simulation with a number of neurons similar to that of a human brain. This paper presents an up-to-date review about the main research projects that are trying to simulate and/or emulate the human brain. They employ different types of computational models using parallel computing: digital models, analog models and hybrid models. This review includes the current applications of these works, as well as future trends. It is focused on various works that look for advanced progress in Neuroscience and still others which seek new discoveries in Computer Science (neuromorphic hardware, machine learning techniques). Their most outstanding characteristics are summarized and the latest advances and future plans are presented. In addition, this review points out the importance of considering not only neurons: Computational models of the brain should also include glial cells, given the proven importance of astrocytes in information processing. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  13. Electric Mars: A Large Trans-Terminator Electric Potential Drop on Closed Magnetic Field Lines Above Utopia Planitia

    Collinson, Glyn; Mitchell, David; Xu, Shaosui; Glocer, Alex; Grebowsky, Joseph; Hara, Takuya; Lillis, Robert; Espley, Jared; Mazelle, Christian; Sauvaud, Jean-Andre

    2017-01-01

    Abstract Parallel electric fields and their associated electric potential structures play a crucial role inionospheric-magnetospheric interactions at any planet. Although there is abundant evidence that parallel electric fields play key roles in Martian ionospheric outflow and auroral electron acceleration, the fields themselves are challenging to directly measure due to their relatively weak nature. Using measurements by the Solar Wind Electron Analyzer instrument aboard the NASA Mars Atmosphere and Volatile EvolutioN(MAVEN) Mars Scout, we present the discovery and measurement of a substantial (Phi) Mars 7.7 +/-0.6 V) parallel electric potential drop on closed magnetic field lines spanning the terminator from day to night above the great impact basin of Utopia Planitia, a region largely free of crustal magnetic fields. A survey of the previous 26 orbits passing over a range of longitudes revealed similar signatures on seven orbits, with a mean potential drop (Phi) Mars of 10.9 +/- 0.8 V, suggestive that although trans-terminator electric fields of comparable strength are not ubiquitous, they may be common, at least at these northerly latitudes.

  14. Electric Mars: A large trans-terminator electric potential drop on closed magnetic field lines above Utopia Planitia

    Collinson, Glyn; Mitchell, David; Xu, Shaosui; Glocer, Alex; Grebowsky, Joseph; Hara, Takuya; Lillis, Robert; Espley, Jared; Mazelle, Christian; Sauvaud, Jean-André; Fedorov, Andrey; Liemohn, Mike; Andersson, Laila; Jakosky, Bruce

    2017-02-01

    Parallel electric fields and their associated electric potential structures play a crucial role in ionospheric-magnetospheric interactions at any planet. Although there is abundant evidence that parallel electric fields play key roles in Martian ionospheric outflow and auroral electron acceleration, the fields themselves are challenging to directly measure due to their relatively weak nature. Using measurements by the Solar Wind Electron Analyzer instrument aboard the NASA Mars Atmosphere and Volatile EvolutioN (MAVEN) Mars Scout, we present the discovery and measurement of a substantial (ΦMars=7.7 ± 0.6 V) parallel electric potential drop on closed magnetic field lines spanning the terminator from day to night above the great impact basin of Utopia Planitia, a region largely free of crustal magnetic fields. A survey of the previous 26 orbits passing over a range of longitudes revealed similar signatures on seven orbits, with a mean potential drop (ΦMars) of 10.9 ± 0.8 V, suggestive that although trans-terminator electric fields of comparable strength are not ubiquitous, they may be common, at least at these northerly latitudes.

  15. High-Performance Psychometrics: The Parallel-E Parallel-M Algorithm for Generalized Latent Variable Models. Research Report. ETS RR-16-34

    von Davier, Matthias

    2016-01-01

    This report presents results on a parallel implementation of the expectation-maximization (EM) algorithm for multidimensional latent variable models. The developments presented here are based on code that parallelizes both the E step and the M step of the parallel-E parallel-M algorithm. Examples presented in this report include item response…

  16. The language parallel Pascal and other aspects of the massively parallel processor

    Reeves, A. P.; Bruner, J. D.

    1982-01-01

    A high level language for the Massively Parallel Processor (MPP) was designed. This language, called Parallel Pascal, is described in detail. A description of the language design, a description of the intermediate language, Parallel P-Code, and details for the MPP implementation are included. Formal descriptions of Parallel Pascal and Parallel P-Code are given. A compiler was developed which converts programs in Parallel Pascal into the intermediate Parallel P-Code language. The code generator to complete the compiler for the MPP is being developed independently. A Parallel Pascal to Pascal translator was also developed. The architecture design for a VLSI version of the MPP was completed with a description of fault tolerant interconnection networks. The memory arrangement aspects of the MPP are discussed and a survey of other high level languages is given.

  17. Parallel Boltzmann machines : a mathematical model

    Zwietering, P.J.; Aarts, E.H.L.

    1991-01-01

    A mathematical model is presented for the description of parallel Boltzmann machines. The framework is based on the theory of Markov chains and combines a number of previously known results into one generic model. It is argued that parallel Boltzmann machines maximize a function consisting of a

  18. The convergence of parallel Boltzmann machines

    Zwietering, P.J.; Aarts, E.H.L.; Eckmiller, R.; Hartmann, G.; Hauske, G.

    1990-01-01

    We discuss the main results obtained in a study of a mathematical model of synchronously parallel Boltzmann machines. We present supporting evidence for the conjecture that a synchronously parallel Boltzmann machine maximizes a consensus function that consists of a weighted sum of the regular

  19. Customizable Memory Schemes for Data Parallel Architectures

    Gou, C.

    2011-01-01

    Memory system efficiency is crucial for any processor to achieve high performance, especially in the case of data parallel machines. Processing capabilities of parallel lanes will be wasted, when data requests are not accomplished in a sustainable and timely manner. Irregular vector memory accesses

  20. Parallel Narrative Structure in Paul Harding's "Tinkers"

    Çirakli, Mustafa Zeki

    2014-01-01

    The present paper explores the implications of parallel narrative structure in Paul Harding's "Tinkers" (2009). Besides primarily recounting the two sets of parallel narratives, "Tinkers" also comprises of seemingly unrelated fragments such as excerpts from clock repair manuals and diaries. The main stories, however, told…

  1. Streaming nested data parallelism on multicores

    Madsen, Frederik Meisner; Filinski, Andrzej

    2016-01-01

    The paradigm of nested data parallelism (NDP) allows a variety of semi-regular computation tasks to be mapped onto SIMD-style hardware, including GPUs and vector units. However, some care is needed to keep down space consumption in situations where the available parallelism may vastly exceed...

  2. Bayer image parallel decoding based on GPU

    Hu, Rihui; Xu, Zhiyong; Wei, Yuxing; Sun, Shaohua

    2012-11-01

    In the photoelectrical tracking system, Bayer image is decompressed in traditional method, which is CPU-based. However, it is too slow when the images become large, for example, 2K×2K×16bit. In order to accelerate the Bayer image decoding, this paper introduces a parallel speedup method for NVIDA's Graphics Processor Unit (GPU) which supports CUDA architecture. The decoding procedure can be divided into three parts: the first is serial part, the second is task-parallelism part, and the last is data-parallelism part including inverse quantization, inverse discrete wavelet transform (IDWT) as well as image post-processing part. For reducing the execution time, the task-parallelism part is optimized by OpenMP techniques. The data-parallelism part could advance its efficiency through executing on the GPU as CUDA parallel program. The optimization techniques include instruction optimization, shared memory access optimization, the access memory coalesced optimization and texture memory optimization. In particular, it can significantly speed up the IDWT by rewriting the 2D (Tow-dimensional) serial IDWT into 1D parallel IDWT. Through experimenting with 1K×1K×16bit Bayer image, data-parallelism part is 10 more times faster than CPU-based implementation. Finally, a CPU+GPU heterogeneous decompression system was designed. The experimental result shows that it could achieve 3 to 5 times speed increase compared to the CPU serial method.

  3. Parallelization of TMVA Machine Learning Algorithms

    Hajili, Mammad

    2017-01-01

    This report reflects my work on Parallelization of TMVA Machine Learning Algorithms integrated to ROOT Data Analysis Framework during summer internship at CERN. The report consists of 4 impor- tant part - data set used in training and validation, algorithms that multiprocessing applied on them, parallelization techniques and re- sults of execution time changes due to number of workers.

  4. 17 CFR 12.24 - Parallel proceedings.

    2010-04-01

    ...) Definition. For purposes of this section, a parallel proceeding shall include: (1) An arbitration proceeding... the receivership includes the resolution of claims made by customers; or (3) A petition filed under... any of the foregoing with knowledge of a parallel proceeding shall promptly notify the Commission, by...

  5. Parallel Computing Strategies for Irregular Algorithms

    Biswas, Rupak; Oliker, Leonid; Shan, Hongzhang; Biegel, Bryan (Technical Monitor)

    2002-01-01

    Parallel computing promises several orders of magnitude increase in our ability to solve realistic computationally-intensive problems, but relies on their efficient mapping and execution on large-scale multiprocessor architectures. Unfortunately, many important applications are irregular and dynamic in nature, making their effective parallel implementation a daunting task. Moreover, with the proliferation of parallel architectures and programming paradigms, the typical scientist is faced with a plethora of questions that must be answered in order to obtain an acceptable parallel implementation of the solution algorithm. In this paper, we consider three representative irregular applications: unstructured remeshing, sparse matrix computations, and N-body problems, and parallelize them using various popular programming paradigms on a wide spectrum of computer platforms ranging from state-of-the-art supercomputers to PC clusters. We present the underlying problems, the solution algorithms, and the parallel implementation strategies. Smart load-balancing, partitioning, and ordering techniques are used to enhance parallel performance. Overall results demonstrate the complexity of efficiently parallelizing irregular algorithms.

  6. Parallel fuzzy connected image segmentation on GPU

    Zhuge, Ying; Cao, Yong; Udupa, Jayaram K.; Miller, Robert W.

    2011-01-01

    Purpose: Image segmentation techniques using fuzzy connectedness (FC) principles have shown their effectiveness in segmenting a variety of objects in several large applications. However, one challenge in these algorithms has been their excessive computational requirements when processing large image datasets. Nowadays, commodity graphics hardware provides a highly parallel computing environment. In this paper, the authors present a parallel fuzzy connected image segmentation algorithm impleme...

  7. Non-Cartesian parallel imaging reconstruction.

    Wright, Katherine L; Hamilton, Jesse I; Griswold, Mark A; Gulani, Vikas; Seiberlich, Nicole

    2014-11-01

    Non-Cartesian parallel imaging has played an important role in reducing data acquisition time in MRI. The use of non-Cartesian trajectories can enable more efficient coverage of k-space, which can be leveraged to reduce scan times. These trajectories can be undersampled to achieve even faster scan times, but the resulting images may contain aliasing artifacts. Just as Cartesian parallel imaging can be used to reconstruct images from undersampled Cartesian data, non-Cartesian parallel imaging methods can mitigate aliasing artifacts by using additional spatial encoding information in the form of the nonhomogeneous sensitivities of multi-coil phased arrays. This review will begin with an overview of non-Cartesian k-space trajectories and their sampling properties, followed by an in-depth discussion of several selected non-Cartesian parallel imaging algorithms. Three representative non-Cartesian parallel imaging methods will be described, including Conjugate Gradient SENSE (CG SENSE), non-Cartesian generalized autocalibrating partially parallel acquisition (GRAPPA), and Iterative Self-Consistent Parallel Imaging Reconstruction (SPIRiT). After a discussion of these three techniques, several potential promising clinical applications of non-Cartesian parallel imaging will be covered. © 2014 Wiley Periodicals, Inc.

  8. Parallel Algorithms for Groebner-Basis Reduction

    1987-09-25

    22209 ELEMENT NO. NO. NO. ACCESSION NO. 11. TITLE (Include Security Classification) * PARALLEL ALGORITHMS FOR GROEBNER -BASIS REDUCTION 12. PERSONAL...All other editions are obsolete. Productivity Engineering in the UNIXt Environment p Parallel Algorithms for Groebner -Basis Reduction Technical Report

  9. Parallel knock-out schemes in networks

    Broersma, H.J.; Fomin, F.V.; Woeginger, G.J.

    2004-01-01

    We consider parallel knock-out schemes, a procedure on graphs introduced by Lampert and Slater in 1997 in which each vertex eliminates exactly one of its neighbors in each round. We are considering cases in which after a finite number of rounds, where the minimimum number is called the parallel

  10. Building a parallel file system simulator

    Molina-Estolano, E; Maltzahn, C; Brandt, S A; Bent, J

    2009-01-01

    Parallel file systems are gaining in popularity in high-end computing centers as well as commercial data centers. High-end computing systems are expected to scale exponentially and to pose new challenges to their storage scalability in terms of cost and power. To address these challenges scientists and file system designers will need a thorough understanding of the design space of parallel file systems. Yet there exist few systematic studies of parallel file system behavior at petabyte- and exabyte scale. An important reason is the significant cost of getting access to large-scale hardware to test parallel file systems. To contribute to this understanding we are building a parallel file system simulator that can simulate parallel file systems at very large scale. Our goal is to simulate petabyte-scale parallel file systems on a small cluster or even a single machine in reasonable time and fidelity. With this simulator, file system experts will be able to tune existing file systems for specific workloads, scientists and file system deployment engineers will be able to better communicate workload requirements, file system designers and researchers will be able to try out design alternatives and innovations at scale, and instructors will be able to study very large-scale parallel file system behavior in the class room. In this paper we describe our approach and provide preliminary results that are encouraging both in terms of fidelity and simulation scalability.

  11. Application of parallelized software architecture to an autonomous ground vehicle

    Shakya, Rahul; Wright, Adam; Shin, Young Ho; Momin, Orko; Petkovsek, Steven; Wortman, Paul; Gautam, Prasanna; Norton, Adam

    2011-01-01

    This paper presents improvements made to Q, an autonomous ground vehicle designed to participate in the Intelligent Ground Vehicle Competition (IGVC). For the 2010 IGVC, Q was upgraded with a new parallelized software architecture and a new vision processor. Improvements were made to the power system reducing the number of batteries required for operation from six to one. In previous years, a single state machine was used to execute the bulk of processing activities including sensor interfacing, data processing, path planning, navigation algorithms and motor control. This inefficient approach led to poor software performance and made it difficult to maintain or modify. For IGVC 2010, the team implemented a modular parallel architecture using the National Instruments (NI) LabVIEW programming language. The new architecture divides all the necessary tasks - motor control, navigation, sensor data collection, etc. into well-organized components that execute in parallel, providing considerable flexibility and facilitating efficient use of processing power. Computer vision is used to detect white lines on the ground and determine their location relative to the robot. With the new vision processor and some optimization of the image processing algorithm used last year, two frames can be acquired and processed in 70ms. With all these improvements, Q placed 2nd in the autonomous challenge.

  12. Real-time trajectory optimization on parallel processors

    Psiaki, Mark L.

    1993-01-01

    A parallel algorithm has been developed for rapidly solving trajectory optimization problems. The goal of the work has been to develop an algorithm that is suitable to do real-time, on-line optimal guidance through repeated solution of a trajectory optimization problem. The algorithm has been developed on an INTEL iPSC/860 message passing parallel processor. It uses a zero-order-hold discretization of a continuous-time problem and solves the resulting nonlinear programming problem using a custom-designed augmented Lagrangian nonlinear programming algorithm. The algorithm achieves parallelism of function, derivative, and search direction calculations through the principle of domain decomposition applied along the time axis. It has been encoded and tested on 3 example problems, the Goddard problem, the acceleration-limited, planar minimum-time to the origin problem, and a National Aerospace Plane minimum-fuel ascent guidance problem. Execution times as fast as 118 sec of wall clock time have been achieved for a 128-stage Goddard problem solved on 32 processors. A 32-stage minimum-time problem has been solved in 151 sec on 32 processors. A 32-stage National Aerospace Plane problem required 2 hours when solved on 32 processors. A speed-up factor of 7.2 has been achieved by using 32-nodes instead of 1-node to solve a 64-stage Goddard problem.

  13. Broadcasting a message in a parallel computer

    Berg, Jeremy E [Rochester, MN; Faraj, Ahmad A [Rochester, MN

    2011-08-02

    Methods, systems, and products are disclosed for broadcasting a message in a parallel computer. The parallel computer includes a plurality of compute nodes connected together using a data communications network. The data communications network optimized for point to point data communications and is characterized by at least two dimensions. The compute nodes are organized into at least one operational group of compute nodes for collective parallel operations of the parallel computer. One compute node of the operational group assigned to be a logical root. Broadcasting a message in a parallel computer includes: establishing a Hamiltonian path along all of the compute nodes in at least one plane of the data communications network and in the operational group; and broadcasting, by the logical root to the remaining compute nodes, the logical root's message along the established Hamiltonian path.

  14. Advanced parallel processing with supercomputer architectures

    Hwang, K.

    1987-01-01

    This paper investigates advanced parallel processing techniques and innovative hardware/software architectures that can be applied to boost the performance of supercomputers. Critical issues on architectural choices, parallel languages, compiling techniques, resource management, concurrency control, programming environment, parallel algorithms, and performance enhancement methods are examined and the best answers are presented. The authors cover advanced processing techniques suitable for supercomputers, high-end mainframes, minisupers, and array processors. The coverage emphasizes vectorization, multitasking, multiprocessing, and distributed computing. In order to achieve these operation modes, parallel languages, smart compilers, synchronization mechanisms, load balancing methods, mapping parallel algorithms, operating system functions, application library, and multidiscipline interactions are investigated to ensure high performance. At the end, they assess the potentials of optical and neural technologies for developing future supercomputers

  15. Differences Between Distributed and Parallel Systems

    Brightwell, R.; Maccabe, A.B.; Rissen, R.

    1998-10-01

    Distributed systems have been studied for twenty years and are now coming into wider use as fast networks and powerful workstations become more readily available. In many respects a massively parallel computer resembles a network of workstations and it is tempting to port a distributed operating system to such a machine. However, there are significant differences between these two environments and a parallel operating system is needed to get the best performance out of a massively parallel system. This report characterizes the differences between distributed systems, networks of workstations, and massively parallel systems and analyzes the impact of these differences on operating system design. In the second part of the report, we introduce Puma, an operating system specifically developed for massively parallel systems. We describe Puma portals, the basic building blocks for message passing paradigms implemented on top of Puma, and show how the differences observed in the first part of the report have influenced the design and implementation of Puma.

  16. Parallel-In-Time For Moving Meshes

    Falgout, R. D. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Manteuffel, T. A. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Southworth, B. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Schroder, J. B. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2016-02-04

    With steadily growing computational resources available, scientists must develop e ective ways to utilize the increased resources. High performance, highly parallel software has be- come a standard. However until recent years parallelism has focused primarily on the spatial domain. When solving a space-time partial di erential equation (PDE), this leads to a sequential bottleneck in the temporal dimension, particularly when taking a large number of time steps. The XBraid parallel-in-time library was developed as a practical way to add temporal parallelism to existing se- quential codes with only minor modi cations. In this work, a rezoning-type moving mesh is applied to a di usion problem and formulated in a parallel-in-time framework. Tests and scaling studies are run using XBraid and demonstrate excellent results for the simple model problem considered herein.

  17. Parallel programming with Easy Java Simulations

    Esquembre, F.; Christian, W.; Belloni, M.

    2018-01-01

    Nearly all of today's processors are multicore, and ideally programming and algorithm development utilizing the entire processor should be introduced early in the computational physics curriculum. Parallel programming is often not introduced because it requires a new programming environment and uses constructs that are unfamiliar to many teachers. We describe how we decrease the barrier to parallel programming by using a java-based programming environment to treat problems in the usual undergraduate curriculum. We use the easy java simulations programming and authoring tool to create the program's graphical user interface together with objects based on those developed by Kaminsky [Building Parallel Programs (Course Technology, Boston, 2010)] to handle common parallel programming tasks. Shared-memory parallel implementations of physics problems, such as time evolution of the Schrödinger equation, are available as source code and as ready-to-run programs from the AAPT-ComPADRE digital library.

  18. A parallel solver for huge dense linear systems

    Badia, J. M.; Movilla, J. L.; Climente, J. I.; Castillo, M.; Marqués, M.; Mayo, R.; Quintana-Ortí, E. S.; Planelles, J.

    2011-11-01

    HDSS (Huge Dense Linear System Solver) is a Fortran Application Programming Interface (API) to facilitate the parallel solution of very large dense systems to scientists and engineers. The API makes use of parallelism to yield an efficient solution of the systems on a wide range of parallel platforms, from clusters of processors to massively parallel multiprocessors. It exploits out-of-core strategies to leverage the secondary memory in order to solve huge linear systems O(100.000). The API is based on the parallel linear algebra library PLAPACK, and on its Out-Of-Core (OOC) extension POOCLAPACK. Both PLAPACK and POOCLAPACK use the Message Passing Interface (MPI) as the communication layer and BLAS to perform the local matrix operations. The API provides a friendly interface to the users, hiding almost all the technical aspects related to the parallel execution of the code and the use of the secondary memory to solve the systems. In particular, the API can automatically select the best way to store and solve the systems, depending of the dimension of the system, the number of processes and the main memory of the platform. Experimental results on several parallel platforms report high performance, reaching more than 1 TFLOP with 64 cores to solve a system with more than 200 000 equations and more than 10 000 right-hand side vectors. New version program summaryProgram title: Huge Dense System Solver (HDSS) Catalogue identifier: AEHU_v1_1 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEHU_v1_1.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 87 062 No. of bytes in distributed program, including test data, etc.: 1 069 110 Distribution format: tar.gz Programming language: Fortran90, C Computer: Parallel architectures: multiprocessors, computer clusters Operating system

  19. Parallel hyperbolic PDE simulation on clusters: Cell versus GPU

    Rostrup, Scott; De Sterck, Hans

    2010-12-01

    :http://cpc.cs.qub.ac.uk/summaries/AEGY_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GPL v3 No. of lines in distributed program, including test data, etc.: 59 168 No. of bytes in distributed program, including test data, etc.: 453 409 Distribution format: tar.gz Programming language: C, CUDA Computer: Parallel Computing Clusters. Individual compute nodes may consist of x86 CPU, Cell processor, or x86 CPU with attached NVIDIA GPU accelerator. Operating system: Linux Has the code been vectorised or parallelized?: Yes. Tested on 1-128 x86 CPU cores, 1-32 Cell Processors, and 1-32 NVIDIA GPUs. RAM: Tested on Problems requiring up to 4 GB per compute node. Classification: 12 External routines: MPI, CUDA, IBM Cell SDK Nature of problem: MPI-parallel simulation of Shallow Water equations using high-resolution 2D hyperbolic equation solver on regular Cartesian grids for x86 CPU, Cell Processor, and NVIDIA GPU using CUDA. Solution method: SWsolver provides 3 implementations of a high-resolution 2D Shallow Water equation solver on regular Cartesian grids, for CPU, Cell Processor, and NVIDIA GPU. Each implementation uses MPI to divide work across a parallel computing cluster. Additional comments: Sub-program numdiff is used for the test run.

  20. Conductance of auroral magnetic field lines

    Weimer, D.R.; Gurnett, D.A.; Goertz, C.K.

    1986-01-01

    DE-1 high-resolution double-probe electric-field data and simultaneous magnetic-field measurements are reported for two 1981 events with large electric fields which reversed over short distances. The data are presented graphically and analyzed in detail. A field-line conductance of about 1 nmho/sq m is determined for both upward and downward currents, and the ionospheric conductivity is shown, in the short-wavelength limit, to have little effect on the relationship between the (N-S) electric and (E-W) magnetic fields above the potential drop parallel to the magnetic-field lines. The results are found to be consistent with a linear relationship between the field-aligned current density and the parallel potential drop. 14 references

  1. Parallel Branch-and-Bound Methods for the Job Shop Scheduling

    Clausen, Jens; Perregaard, Michael

    1998-01-01

    Job-shop scheduling (JSS) problems are among the more difficult to solve in the class of NP-complete problems. The only successful approach has been branch-and-bound based algorithms, but such algorithms depend heavily on good bound functions. Much work has been done to identify such functions...... for the JSS problem, but with limited success. Even with recent methods, it is still not possible to solve problems substantially larger than 10 machines and 10 jobs. In the current study, we focus on parallel methods for solving JSS problems. We implement two different parallel branch-and-bound algorithms...

  2. Integrated variable projection approach (IVAPA) for parallel magnetic resonance imaging.

    Zhang, Qiao; Sheng, Jinhua

    2012-10-01

    Parallel magnetic resonance imaging (pMRI) is a fast method which requires algorithms for the reconstructing image from a small number of measured k-space lines. The accurate estimation of the coil sensitivity functions is still a challenging problem in parallel imaging. The joint estimation of the coil sensitivity functions and the desired image has recently been proposed to improve the situation by iteratively optimizing both the coil sensitivity functions and the image reconstruction. It regards both the coil sensitivities and the desired images as unknowns to be solved for jointly. In this paper, we propose an integrated variable projection approach (IVAPA) for pMRI, which integrates two individual processing steps (coil sensitivity estimation and image reconstruction) into a single processing step to improve the accuracy of the coil sensitivity estimation using the variable projection approach. The method is demonstrated to be able to give an optimal solution with considerably reduced artifacts for high reduction factors and a low number of auto-calibration signal (ACS) lines, and our implementation has a fast convergence rate. The performance of the proposed method is evaluated using a set of in vivo experiment data. Copyright © 2012 Elsevier Ltd. All rights reserved.

  3. Managing first-line failure.

    Cooper, David A

    2014-01-01

    WHO standard of care for failure of a first regimen, usually 2N(t)RTI's and an NNRTI, consists of a ritonavir-boosted protease inhibitor with a change in N(t)RTI's. Until recently, there was no evidence to support these recommendations which were based on expert opinion. Two large randomized clinical trials, SECOND LINE and EARNEST both showed excellent response rates (>80%) for the WHO standard of care and indicated that a novel regimen of a boosted protease inhibitor with an integrase inhibitor had equal efficacy with no difference in toxicity. In EARNEST, a third arm consisting of induction with the combined protease and integrase inhibitor followed by protease inhibitor monotherapy maintenance was inferior and led to substantial (20%) protease inhibitor resistance. These studies confirm the validity of the current recommendations of WHO and point to a novel public health approach of using two new classes for second line when standard first-line therapy has failed, which avoids resistance genotyping. Notwithstanding, adherence must be stressed in those failing first-line treatments. Protease inhibitor monotherapy is not suitable for a public health approach in low- and middle-income countries.

  4. Superconducting coherence in a vortex line liquid

    Chen, T.; Teitel, S.

    1995-01-01

    We carry out simulations of the anisotropic uniformly frustrated 3d XY model, as a model for vortex line fluctuations in high T c superconductors. We compute the phase diagram as a function of temperature and anisotropy, for a fixed applied magnetic field B. We find two distinct phase transitions. Upon heating, there is first a lower T c perpendicular where the vortex line lattice melts and super-conducting coherence perpendicular to the applied magnetic field vanishes. At a higher T cz , within the vortex line liquid, superconducting coherence parallel to the applied magnetic field vanishes. For finite anisotropy, both T c perpendicular and T cz lie well below the crossover from the vortex line liquid to the normal state

  5. Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks

    Hesham Mostafa

    2017-09-01

    Full Text Available Artificial neural networks (ANNs trained using backpropagation are powerful learning architectures that have achieved state-of-the-art performance in various benchmarks. Significant effort has been devoted to developing custom silicon devices to accelerate inference in ANNs. Accelerating the training phase, however, has attracted relatively little attention. In this paper, we describe a hardware-efficient on-line learning technique for feedforward multi-layer ANNs that is based on pipelined backpropagation. Learning is performed in parallel with inference in the forward pass, removing the need for an explicit backward pass and requiring no extra weight lookup. By using binary state variables in the feedforward network and ternary errors in truncated-error backpropagation, the need for any multiplications in the forward and backward passes is removed, and memory requirements for the pipelining are drastically reduced. Further reduction in addition operations owing to the sparsity in the forward neural and backpropagating error signal paths contributes to highly efficient hardware implementation. For proof-of-concept validation, we demonstrate on-line learning of MNIST handwritten digit classification on a Spartan 6 FPGA interfacing with an external 1Gb DDR2 DRAM, that shows small degradation in test error performance compared to an equivalently sized binary ANN trained off-line using standard back-propagation and exact errors. Our results highlight an attractive synergy between pipelined backpropagation and binary-state networks in substantially reducing computation and memory requirements, making pipelined on-line learning practical in deep networks.

  6. Hardware-Efficient On-line Learning through Pipelined Truncated-Error Backpropagation in Binary-State Networks.

    Mostafa, Hesham; Pedroni, Bruno; Sheik, Sadique; Cauwenberghs, Gert

    2017-01-01

    Artificial neural networks (ANNs) trained using backpropagation are powerful learning architectures that have achieved state-of-the-art performance in various benchmarks. Significant effort has been devoted to developing custom silicon devices to accelerate inference in ANNs. Accelerating the training phase, however, has attracted relatively little attention. In this paper, we describe a hardware-efficient on-line learning technique for feedforward multi-layer ANNs that is based on pipelined backpropagation. Learning is performed in parallel with inference in the forward pass, removing the need for an explicit backward pass and requiring no extra weight lookup. By using binary state variables in the feedforward network and ternary errors in truncated-error backpropagation, the need for any multiplications in the forward and backward passes is removed, and memory requirements for the pipelining are drastically reduced. Further reduction in addition operations owing to the sparsity in the forward neural and backpropagating error signal paths contributes to highly efficient hardware implementation. For proof-of-concept validation, we demonstrate on-line learning of MNIST handwritten digit classification on a Spartan 6 FPGA interfacing with an external 1Gb DDR2 DRAM, that shows small degradation in test error performance compared to an equivalently sized binary ANN trained off-line using standard back-propagation and exact errors. Our results highlight an attractive synergy between pipelined backpropagation and binary-state networks in substantially reducing computation and memory requirements, making pipelined on-line learning practical in deep networks.

  7. A directly heated electron beam line source

    Iqbal, M.; Masood, K.; Rafiq, M.; Chaudhry, M.A.

    2002-05-01

    A 140-mm cathode length, Electron Beam Line Source with a high degree of focusing of the beam is constructed. The design principles and basic characteristic considerations for electron beam line source consists of parallel plate electrode geometric array as well as a beam power of 35kW are worked out. The dimensions of the beam at the work site are 1.25xl00mm. The gun is designed basically for the study of evaporation and deposition characteristic of refractory metals for laboratory use. However, it may be equally used for melting and casting of these metals. (author)

  8. Performance assessment of the SIMFAP parallel cluster at IFIN-HH Bucharest

    Adam, Gh.; Adam, S.; Ayriyan, A.; Dushanov, E.; Hayryan, E.; Korenkov, V.; Lutsenko, A.; Mitsyn, V.; Sapozhnikova, T.; Sapozhnikov, A; Streltsova, O.; Buzatu, F.; Dulea, M.; Vasile, I.; Sima, A.; Visan, C.; Busa, J.; Pokorny, I.

    2008-01-01

    Performance assessment and case study outputs of the parallel SIMFAP cluster at IFIN-HH Bucharest point to its effective and reliable operation. A comparison with results on the supercomputing system in LIT-JINR Dubna adds insight on resource allocation for problem solving by parallel computing. The solution of models asking for very large numbers of knots in the discretization mesh needs the migration to high performance computing based on parallel cluster architectures. The acquisition of ready-to-use parallel computing facilities being beyond limited budgetary resources, the solution at IFIN-HH was to buy the hardware and the inter-processor network, and to implement by own efforts the open software concerning both the operating system and the parallel computing standard. The present paper provides a report demonstrating the successful solution of these tasks. The implementation of the well-known HPL (High Performance LINPACK) Benchmark points to the effective and reliable operation of the cluster. The comparison of HPL outputs obtained on parallel clusters of different magnitudes shows that there is an optimum range of the order N of the linear algebraic system over which a given parallel cluster provides optimum parallel solutions. For the SIMFAP cluster, this range can be inferred to correspond to about 1 to 2 x 10 4 linear algebraic equations. For an algorithm of polynomial complexity N α the task sharing among p processors within a parallel solution mainly follows an (N/p)α behaviour under peak performance achievement. Thus, while the problem complexity remains the same, a substantial decrease of the coefficient of the leading order of the polynomial complexity is achieved. (authors)

  9. Vertical Line Nodes in the Superconducting Gap Structure of Sr_{2}RuO_{4}

    E. Hassinger

    2017-03-01

    Full Text Available There is strong experimental evidence that the superconductor Sr_{2}RuO_{4} has a chiral p-wave order parameter. This symmetry does not require that the associated gap has nodes, yet specific heat, ultrasound, and thermal conductivity measurements indicate the presence of nodes in the superconducting gap structure of Sr_{2}RuO_{4}. Theoretical scenarios have been proposed to account for the existence of deep minima or accidental nodes (minima tuned to zero or below by material parameters within a p-wave state. Other scenarios propose chiral d-wave and f-wave states, with horizontal and vertical line nodes, respectively. To elucidate the nodal structure of the gap, it is essential to know whether the lines of nodes (or minima are vertical (parallel to the tetragonal c axis or horizontal (perpendicular to the c axis. Here, we report thermal conductivity measurements on single crystals of Sr_{2}RuO_{4} down to 50 mK for currents parallel and perpendicular to the c axis. We find that there is substantial quasiparticle transport in the T=0 limit for both current directions. A magnetic field H immediately excites quasiparticles with velocities both in the basal plane and in the c direction. Our data down to T_{c}/30 and down to H_{c2}/100 show no evidence that the nodes are in fact deep minima. Relative to the normal state, the thermal conductivity of the superconducting state is found to be very similar for the two current directions, from H=0 to H=H_{c2}. These findings show that the gap structure of Sr_{2}RuO_{4} consists of vertical line nodes. This rules out a chiral d-wave state. Given that the c-axis dispersion (warping of the Fermi surface in Sr_{2}RuO_{4} varies strongly from sheet to sheet, the small a-c anisotropy suggests that the line nodes are present on all three sheets of the Fermi surface. If imposed by symmetry, vertical line nodes would be inconsistent with a p-wave order parameter for Sr_{2}RuO_{4}. To reconcile the gap structure

  10. Two-state ion heating at quasi-parallel shocks

    Thomsen, M.F.; Gosling, J.T.; Bame, S.J.; Onsager, T.G.; Russell, C.T.

    1990-01-01

    In a previous study of ion heating at quasi-parallel shocks, the authors showed a case in which the ion distributions downstream from the shock alternated between a cooler, denser, core/shoulder type and a hotter, less dense, more Maxwellian type. In this paper they further document the alternating occurrence of two different ion states downstream from several quasi-parallel shocks. Three separate lines of evidence are presented to show that the two states are not related in an evolutionary sense, but rather both are produced alternately at the shock: (1) the asymptotic downstream plasma parameters (density, ion temperature, and flow speed) are intermediate between those characterizing the two different states closer to the shock, suggesting that the asymptotic state is produced by a mixing of the two initial states; (2) examples of apparently interpenetrating (i.e., mixing) distributions can be found during transitions from one state to the other; and (3) examples of both types of distributions can be found at actual crossings of the shock ramp. The alternation between the two different types of ion distribution provides direct observational support for the idea that the dissipative dynamics of at least some quasi-parallel shocks is non-stationary and cyclic in nature, as demonstrated by recent numerical simulations. Typical cycle times between intervals of similar ion heating states are ∼2 upstream ion gyroperiods. Both the simulations and the in situ observations indicate that a process of coherent ion reflection is commonly an important part of the dissipation at quasi-parallel shocks

  11. Workshop on Radio Recombination Lines

    1980-01-01

    Since their first detection 15 years ago, radio recombination lines from several elements have been observed in a wide variety of objects including HII regions, planetary nebulae, molecular clouds, the diffuse interstellar medium, and recently, other galaxies. The observations span almost the entire range from 0.1 to 100 GHz, and employ both single­ djsh and aperture synthesis techniques. The theory of radio recombination lines has also advanced strongly, to the point where it is perhaps one of the best-understood in astro­ physics. In a parallel development, it has become possible over the last decade to study these same highly-excited atoms in the laboratory; this work provides further confirmation of the theoretical framework. However there has been continuing controversy over the astrophysical interpre­ tation of radio recombination line observations, especially regarding the role of stimulated emission. A workshop was held in Ottawa on 24-25 August, 1979, bringing together many of the active scientist...

  12. SUBSTANTIAL EXCEPTIONS AND (DELIMITATIONS OF THE POWERS OF THE JUDGES ON CIVIL PROCEDURAL LAW

    Igor Raatz

    2017-08-01

    Full Text Available This essay aims to unveil the role of the substantial exceptions on delimiting the powers of the judges on civil procedural law, especially regarding the ex officio judicial activity. This way, under a phenomenological method and based on a vision of guarantee of rights on procedural law, the article offers a brief explanation of the question concerning the content of the object under litigation and its role of (delimiting the powers of the judges. The work hypothesis is the addition of the substantial exceptions among the content of the object under litigation, along with the claim itself and the cause of action. The results lie on the premise that, by the substantial exceptions, the defendant extends the object under litigation – which is formed dynamically on civil procedure. The conclusion points towards the idea that the substantial exceptions act in a way of limiting the ex officio judicial activity on civil procedural law

  13. Main results of substantiation of the ecological safety of the Novovoronezh NPP-2 design

    Kopytov, I.I.; Kocher'yan, V.M.; Leonov, S.V.; Chionov, V.G.; Ehrnestova, L.S.

    2005-01-01

    Paper presents the results of the efforts to determine both the actual (hydrochemical, hydrobiological, geobotanical, soil, radiological) and the predicted parameters of the region ecology derived when substantiating the ecological safety of the Novovoronezh NPP-2 design [ru

  14. Portable parallel programming in a Fortran environment

    May, E.N.

    1989-01-01

    Experience using the Argonne-developed PARMACs macro package to implement a portable parallel programming environment is described. Fortran programs with intrinsic parallelism of coarse and medium granularity are easily converted to parallel programs which are portable among a number of commercially available parallel processors in the class of shared-memory bus-based and local-memory network based MIMD processors. The parallelism is implemented using standard UNIX (tm) tools and a small number of easily understood synchronization concepts (monitors and message-passing techniques) to construct and coordinate multiple cooperating processes on one or many processors. Benchmark results are presented for parallel computers such as the Alliant FX/8, the Encore MultiMax, the Sequent Balance, the Intel iPSC/2 Hypercube and a network of Sun 3 workstations. These parallel machines are typical MIMD types with from 8 to 30 processors, each rated at from 1 to 10 MIPS processing power. The demonstration code used for this work is a Monte Carlo simulation of the response to photons of a ''nearly realistic'' lead, iron and plastic electromagnetic and hadronic calorimeter, using the EGS4 code system. 6 refs., 2 figs., 2 tabs

  15. Performance of the Galley Parallel File System

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    As the input/output (I/O) needs of parallel scientific applications increase, file systems for multiprocessors are being designed to provide applications with parallel access to multiple disks. Many parallel file systems present applications with a conventional Unix-like interface that allows the application to access multiple disks transparently. This interface conceals the parallism within the file system, which increases the ease of programmability, but makes it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. Furthermore, most current parallel file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic parallel workloads. Initial experiments, reported in this paper, indicate that Galley is capable of providing high-performance 1/O to applications the applications that rely on them. In Section 3 we describe that access data in patterns that have been observed to be common.

  16. Cumulative stress and substantiated maltreatment: the importance of caregiver vulnerability and adult partner violence.

    Wekerle, Christine; Wall, Anne-Marie; Leung, Eman; Trocmé, Nico

    2007-04-01

    Our goal is to assess the effect of caregiver vulnerabilities, singly and in combination, on the substantiation of child abuse (physical, sexual) and neglect, while controlling for relevant background variables. We test the moderator role of adult partner violence in qualifying the relationship between caregiver vulnerabilities and maltreatment substantiation. Secondary analyses of the 1998 Canadian Incidence Study of Reported Child Maltreatment (CIS) are used to predict child protective service investigation substantiation versus non-substantiation from a range of caregiver vulnerability factors. Involvement in partner violence was examined as a moderator in the relation between caregiver vulnerabilities and maltreatment substantiation. The CIS is an epidemiological survey of first-reported cases to child protective services, using a random sample of child welfare agencies across Canada. Child welfare workers completed a research form on the child, primary caregiver, family, perpetrator, severity and type of maltreatment, as well as services and court outcomes. All maltreatment classifications were assigned according to the Canadian legal definition of child abuse and neglect. Hierarchical logistic regression analyses were used, with stepped entry of: (1) demographic factors, socioeconomic disadvantage, and caregiver's own history of maltreatment; (2) caregiver vulnerability factors; (3) involvement in partner violence; (4) the interaction between caregiver vulnerability and partner violence. Caregiver substance abuse was found to be the single most potent kind of caregiver vulnerability in predicting maltreatment substantiation. When the total number of vulnerabilities was used as the predictor, prediction across all types of maltreatment increased, especially for substantiated neglect. Analyses also showed that the presence of partner violence in the home exacerbated the effect of caregiver vulnerability on substantiation. The total number of caregiver

  17. Optimization approaches to mpi and area merging-based parallel buffer algorithm

    Junfu Fan

    Full Text Available On buffer zone construction, the rasterization-based dilation method inevitably introduces errors, and the double-sided parallel line method involves a series of complex operations. In this paper, we proposed a parallel buffer algorithm based on area merging and MPI (Message Passing Interface to improve the performances of buffer analyses on processing large datasets. Experimental results reveal that there are three major performance bottlenecks which significantly impact the serial and parallel buffer construction efficiencies, including the area merging strategy, the task load balance method and the MPI inter-process results merging strategy. Corresponding optimization approaches involving tree-like area merging strategy, the vertex number oriented parallel task partition method and the inter-process results merging strategy were suggested to overcome these bottlenecks. Experiments were carried out to examine the performance efficiency of the optimized parallel algorithm. The estimation results suggested that the optimization approaches could provide high performance and processing ability for buffer construction in a cluster parallel environment. Our method could provide insights into the parallelization of spatial analysis algorithm.

  18. The kpx, a program analyzer for parallelization

    Matsuyama, Yuji; Orii, Shigeo; Ota, Toshiro; Kume, Etsuo; Aikawa, Hiroshi.

    1997-03-01

    The kpx is a program analyzer, developed as a common technological basis for promoting parallel processing. The kpx consists of three tools. The first is ktool, that shows how much execution time is spent in program segments. The second is ptool, that shows parallelization overhead on the Paragon system. The last is xtool, that shows parallelization overhead on the VPP system. The kpx, designed to work for any FORTRAN cord on any UNIX computer, is confirmed to work well after testing on Paragon, SP2, SR2201, VPP500, VPP300, Monte-4, SX-4 and T90. (author)

  19. Synchronization Of Parallel Discrete Event Simulations

    Steinman, Jeffrey S.

    1992-01-01

    Adaptive, parallel, discrete-event-simulation-synchronization algorithm, Breathing Time Buckets, developed in Synchronous Parallel Environment for Emulation and Discrete Event Simulation (SPEEDES) operating system. Algorithm allows parallel simulations to process events optimistically in fluctuating time cycles that naturally adapt while simulation in progress. Combines best of optimistic and conservative synchronization strategies while avoiding major disadvantages. Algorithm processes events optimistically in time cycles adapting while simulation in progress. Well suited for modeling communication networks, for large-scale war games, for simulated flights of aircraft, for simulations of computer equipment, for mathematical modeling, for interactive engineering simulations, and for depictions of flows of information.

  20. Multistage parallel-serial time averaging filters

    Theodosiou, G.E.

    1980-01-01

    Here, a new time averaging circuit design, the 'parallel filter' is presented, which can reduce the time jitter, introduced in time measurements using counters of large dimensions. This parallel filter could be considered as a single stage unit circuit which can be repeated an arbitrary number of times in series, thus providing a parallel-serial filter type as a result. The main advantages of such a filter over a serial one are much less electronic gate jitter and time delay for the same amount of total time uncertainty reduction. (orig.)

  1. Implementations of BLAST for parallel computers.

    Jülich, A

    1995-02-01

    The BLAST sequence comparison programs have been ported to a variety of parallel computers-the shared memory machine Cray Y-MP 8/864 and the distributed memory architectures Intel iPSC/860 and nCUBE. Additionally, the programs were ported to run on workstation clusters. We explain the parallelization techniques and consider the pros and cons of these methods. The BLAST programs are very well suited for parallelization for a moderate number of processors. We illustrate our results using the program blastp as an example. As input data for blastp, a 799 residue protein query sequence and the protein database PIR were used.

  2. Speedup predictions on large scientific parallel programs

    Williams, E.; Bobrowicz, F.

    1985-01-01

    How much speedup can we expect for large scientific parallel programs running on supercomputers. For insight into this problem we extend the parallel processing environment currently existing on the Cray X-MP (a shared memory multiprocessor with at most four processors) to a simulated N-processor environment, where N greater than or equal to 1. Several large scientific parallel programs from Los Alamos National Laboratory were run in this simulated environment, and speedups were predicted. A speedup of 14.4 on 16 processors was measured for one of the three most used codes at the Laboratory

  3. Distributed parallel messaging for multiprocessor systems

    Chen, Dong; Heidelberger, Philip; Salapura, Valentina; Senger, Robert M; Steinmacher-Burrow, Burhard; Sugawara, Yutaka

    2013-06-04

    A method and apparatus for distributed parallel messaging in a parallel computing system. The apparatus includes, at each node of a multiprocessor network, multiple injection messaging engine units and reception messaging engine units, each implementing a DMA engine and each supporting both multiple packet injection into and multiple reception from a network, in parallel. The reception side of the messaging unit (MU) includes a switch interface enabling writing of data of a packet received from the network to the memory system. The transmission side of the messaging unit, includes switch interface for reading from the memory system when injecting packets into the network.

  4. Massively parallel Fokker-Planck code ALLAp

    Batishcheva, A.A.; Krasheninnikov, S.I.; Craddock, G.G.; Djordjevic, V.

    1996-01-01

    The recently developed for workstations Fokker-Planck code ALLA simulates the temporal evolution of 1V, 2V and 1D2V collisional edge plasmas. In this work we present the results of code parallelization on the CRI T3D massively parallel platform (ALLAp version). Simultaneously we benchmark the 1D2V parallel vesion against an analytic self-similar solution of the collisional kinetic equation. This test is not trivial as it demands a very strong spatial temperature and density variation within the simulation domain. (orig.)

  5. Parallel processing of dose calculation for external photon beam therapy

    Kunieda, Etsuo; Ando, Yutaka; Tsukamoto, Nobuhiro; Ito, Hisao; Kubo, Atsushi

    1994-01-01

    We implemented external photon beam dose calculation programs into a parallel processor system consisting of Transputers, 32-bit processors especially suitable for multi-processor configuration. Two network conformations, binary-tree and pipeline, were evaluated for rectangular and irregular field dose calculation algorithms. Although computation speed increased in proportion to the number of CPU, substantial overhead caused by inter-processor communication occurred when a smaller computation load was delivered to each processor. On the other hand, for irregular field calculation, which requires more computation capability for each calculation point, the communication overhead was still less even when more than 50 processors were involved. Real-time responses could be expected for more complex algorithms by increasing the number of processors. (author)

  6. Study on Parallel Processing for Efficient Flexible Multibody Analysis based on Subsystem Synthesis Method

    Han, Jong-Boo; Song, Hajun; Kim, Sung-Soo [Chungnam Nat’l Univ., Daejeon (Korea, Republic of)

    2017-06-15

    Flexible multibody simulations are widely used in the industry to design mechanical systems. In flexible multibody dynamics, deformation coordinates are described either relatively in the body reference frame that is floating in the space or in the inertial reference frame. Moreover, these deformation coordinates are generated based on the discretization of the body according to the finite element approach. Therefore, the formulation of the flexible multibody system always deals with a huge number of degrees of freedom and the numerical solution methods require a substantial amount of computational time. Parallel computational methods are a solution for efficient computation. However, most of the parallel computational methods are focused on the efficient solution of large-sized linear equations. For multibody analysis, we need to develop an efficient formulation that could be suitable for parallel computation. In this paper, we developed a subsystem synthesis method for a flexible multibody system and proposed efficient parallel computational schemes based on the OpenMP API in order to achieve efficient computation. Simulations of a rotating blade system, which consists of three identical blades, were carried out with two different parallel computational schemes. Actual CPU times were measured to investigate the efficiency of the proposed parallel schemes.

  7. A comparison of energetic ions in the plasma depletion layer and the quasi-parallel magnetosheath

    Fuselier, Stephen A.

    1994-01-01

    Energetic ion spectra measured by the Active Magnetospheric Particle Tracer Explorers/Charge Composition Explorer (AMPTE/CCE) downstream from the Earth's quasi-parallel bow shock (in the quasi-parallel magnetosheath) and in the plasma depletion layer are compared. In the latter region, energetic ions are from a single source, leakage of magnetospheric ions across the magnetopause and into the plasma depletion layer. In the former region, both the magnetospheric source and shock acceleration of the thermal solar wind population at the quasi-parallel shock can contribute to the energetic ion spectra. The relative strengths of these two energetic ion sources are determined through the comparison of spectra from the two regions. It is found that magnetospheric leakage can provide an upper limit of 35% of the total energetic H(+) population in the quasi-parallel magnetosheath near the magnetopause in the energy range from approximately 10 to approximately 80 keV/e and substantially less than this limit for the energetic He(2+) population. The rest of the energetic H(+) population and nearly all of the energetic He(2+) population are accelerated out of the thermal solar wind population through shock acceleration processes. By comparing the energetic and thermal He(2+) and H(+) populations in the quasi-parallel magnetosheath, it is found that the quasi-parallel bow shock is 2 to 3 times more efficient at accelerating He(2+) than H(+). This result is consistent with previous estimates from shock acceleration theory and simulati ons.

  8. Parallel Processing of Images in Mobile Devices using BOINC

    Curiel, Mariela; Calle, David F.; Santamaría, Alfredo S.; Suarez, David F.; Flórez, Leonardo

    2018-04-01

    Medical image processing helps health professionals make decisions for the diagnosis and treatment of patients. Since some algorithms for processing images require substantial amounts of resources, one could take advantage of distributed or parallel computing. A mobile grid can be an adequate computing infrastructure for this problem. A mobile grid is a grid that includes mobile devices as resource providers. In a previous step of this research, we selected BOINC as the infrastructure to build our mobile grid. However, parallel processing of images in mobile devices poses at least two important challenges: the execution of standard libraries for processing images and obtaining adequate performance when compared to desktop computers grids. By the time we started our research, the use of BOINC in mobile devices also involved two issues: a) the execution of programs in mobile devices required to modify the code to insert calls to the BOINC API, and b) the division of the image among the mobile devices as well as its merging required additional code in some BOINC components. This article presents answers to these four challenges.

  9. Parallel Processing of Images in Mobile Devices using BOINC

    Curiel Mariela

    2018-04-01

    Full Text Available Medical image processing helps health professionals make decisions for the diagnosis and treatment of patients. Since some algorithms for processing images require substantial amounts of resources, one could take advantage of distributed or parallel computing. A mobile grid can be an adequate computing infrastructure for this problem. A mobile grid is a grid that includes mobile devices as resource providers. In a previous step of this research, we selected BOINC as the infrastructure to build our mobile grid. However, parallel processing of images in mobile devices poses at least two important challenges: the execution of standard libraries for processing images and obtaining adequate performance when compared to desktop computers grids. By the time we started our research, the use of BOINC in mobile devices also involved two issues: a the execution of programs in mobile devices required to modify the code to insert calls to the BOINC API, and b the division of the image among the mobile devices as well as its merging required additional code in some BOINC components. This article presents answers to these four challenges.

  10. Massively Parallel Computing: A Sandia Perspective

    Dosanjh, Sudip S.; Greenberg, David S.; Hendrickson, Bruce; Heroux, Michael A.; Plimpton, Steve J.; Tomkins, James L.; Womble, David E.

    1999-05-06

    The computing power available to scientists and engineers has increased dramatically in the past decade, due in part to progress in making massively parallel computing practical and available. The expectation for these machines has been great. The reality is that progress has been slower than expected. Nevertheless, massively parallel computing is beginning to realize its potential for enabling significant break-throughs in science and engineering. This paper provides a perspective on the state of the field, colored by the authors' experiences using large scale parallel machines at Sandia National Laboratories. We address trends in hardware, system software and algorithms, and we also offer our view of the forces shaping the parallel computing industry.

  11. Parallel generation of architecture on the GPU

    Steinberger, Markus; Kenzel, Michael; Kainz, Bernhard K.; Mü ller, Jö rg; Wonka, Peter; Schmalstieg, Dieter

    2014-01-01

    they can take advantage of, or both, our method supports state of the art procedural modeling including stochasticity and context-sensitivity. To increase parallelism, we explicitly express independence in the grammar, reduce inter-rule dependencies

  12. New high voltage parallel plate analyzer

    Hamada, Y.; Kawasumi, Y.; Masai, K.; Iguchi, H.; Fujisawa, A.; Abe, Y.

    1992-01-01

    A new modification on the parallel plate analyzer for 500 keV heavy ions to eliminate the effect of the intense UV and visible radiations, is successfully conducted. Its principle and results are discussed. (author)

  13. Parallel data encryption with RSA algorithm

    Неретин, А. А.

    2016-01-01

    In this paper a parallel RSA algorithm with preliminary shuffling of source text was presented.Dependence of an encryption speed on the number of encryption nodes has been analysed, The proposed algorithm was implemented on C# language.

  14. Data parallel sorting for particle simulation

    Dagum, Leonardo

    1992-01-01

    Sorting on a parallel architecture is a communications intensive event which can incur a high penalty in applications where it is required. In the case of particle simulation, only integer sorting is necessary, and sequential implementations easily attain the minimum performance bound of O (N) for N particles. Parallel implementations, however, have to cope with the parallel sorting problem which, in addition to incurring a heavy communications cost, can make the minimun performance bound difficult to attain. This paper demonstrates how the sorting problem in a particle simulation can be reduced to a merging problem, and describes an efficient data parallel algorithm to solve this merging problem in a particle simulation. The new algorithm is shown to be optimal under conditions usual for particle simulation, and its fieldwise implementation on the Connection Machine is analyzed in detail. The new algorithm is about four times faster than a fieldwise implementation of radix sort on the Connection Machine.

  15. Parallel debt in the Serbian finance law

    Kuzman Miloš

    2014-01-01

    Full Text Available The purpose of this paper is to present the mechanism of parallel debt in the Serbian financial law. While considering whether the mechanism of parallel debt exists under the Serbian law, the Anglo-Saxon mechanism of trust is represented. Hence it is explained why the mechanism of trust is not allowed under the Serbian law. Further on, the mechanism of parallel debt is introduced as well as a debate on permissibility of its cause in the Serbian law. Comparative legal arguments about this issue are also presented in this paper. In conclusion, the author suggests that on the basis of the conclusions drawn in this paper, the parallel debt mechanism is to be declared admissible if it is ever taken into consideration by the Serbian courts.

  16. Parallel Monte Carlo simulation of aerosol dynamics

    Zhou, K.; He, Z.; Xiao, M.; Zhang, Z.

    2014-01-01

    is simulated with a stochastic method (Marcus-Lushnikov stochastic process). Operator splitting techniques are used to synthesize the deterministic and stochastic parts in the algorithm. The algorithm is parallelized using the Message Passing Interface (MPI

  17. Stranger than fiction: parallel universes beguile science

    2007-01-01

    We may not be able - at least not yet - to prove they exist, many serious scientists say, but there are plenty of reasons to think that parallel dimensions are more than figments of effeaded imagination. (1/2 page)

  18. Parallel computation of nondeterministic algorithms in VLSI

    Hortensius, P D

    1987-01-01

    This work examines parallel VLSI implementations of nondeterministic algorithms. It is demonstrated that conventional pseudorandom number generators are unsuitable for highly parallel applications. Efficient parallel pseudorandom sequence generation can be accomplished using certain classes of elementary one-dimensional cellular automata. The pseudorandom numbers appear in parallel on each clock cycle. Extensive study of the properties of these new pseudorandom number generators is made using standard empirical random number tests, cycle length tests, and implementation considerations. Furthermore, it is shown these particular cellular automata can form the basis of efficient VLSI architectures for computations involved in the Monte Carlo simulation of both the percolation and Ising models from statistical mechanics. Finally, a variation on a Built-In Self-Test technique based upon cellular automata is presented. These Cellular Automata-Logic-Block-Observation (CALBO) circuits improve upon conventional design for testability circuitry.

  19. Adapting algorithms to massively parallel hardware

    Sioulas, Panagiotis

    2016-01-01

    In the recent years, the trend in computing has shifted from delivering processors with faster clock speeds to increasing the number of cores per processor. This marks a paradigm shift towards parallel programming in which applications are programmed to exploit the power provided by multi-cores. Usually there is gain in terms of the time-to-solution and the memory footprint. Specifically, this trend has sparked an interest towards massively parallel systems that can provide a large number of processors, and possibly computing nodes, as in the GPUs and MPPAs (Massively Parallel Processor Arrays). In this project, the focus was on two distinct computing problems: k-d tree searches and track seeding cellular automata. The goal was to adapt the algorithms to parallel systems and evaluate their performance in different cases.

  20. Implementing Shared Memory Parallelism in MCBEND

    Bird Adam

    2017-01-01

    Full Text Available MCBEND is a general purpose radiation transport Monte Carlo code from AMEC Foster Wheelers’s ANSWERS® Software Service. MCBEND is well established in the UK shielding community for radiation shielding and dosimetry assessments. The existing MCBEND parallel capability effectively involves running the same calculation on many processors. This works very well except when the memory requirements of a model restrict the number of instances of a calculation that will fit on a machine. To more effectively utilise parallel hardware OpenMP has been used to implement shared memory parallelism in MCBEND. This paper describes the reasoning behind the choice of OpenMP, notes some of the challenges of multi-threading an established code such as MCBEND and assesses the performance of the parallel method implemented in MCBEND.

  1. Domain decomposition methods and parallel computing

    Meurant, G.

    1991-01-01

    In this paper, we show how to efficiently solve large linear systems on parallel computers. These linear systems arise from discretization of scientific computing problems described by systems of partial differential equations. We show how to get a discrete finite dimensional system from the continuous problem and the chosen conjugate gradient iterative algorithm is briefly described. Then, the different kinds of parallel architectures are reviewed and their advantages and deficiencies are emphasized. We sketch the problems found in programming the conjugate gradient method on parallel computers. For this algorithm to be efficient on parallel machines, domain decomposition techniques are introduced. We give results of numerical experiments showing that these techniques allow a good rate of convergence for the conjugate gradient algorithm as well as computational speeds in excess of a billion of floating point operations per second. (author). 5 refs., 11 figs., 2 tabs., 1 inset

  2. 6th International Parallel Tools Workshop

    Brinkmann, Steffen; Gracia, José; Resch, Michael; Nagel, Wolfgang

    2013-01-01

    The latest advances in the High Performance Computing hardware have significantly raised the level of available compute performance. At the same time, the growing hardware capabilities of modern supercomputing architectures have caused an increasing complexity of the parallel application development. Despite numerous efforts to improve and simplify parallel programming, there is still a lot of manual debugging and  tuning work required. This process  is supported by special software tools, facilitating debugging, performance analysis, and optimization and thus  making a major contribution to the development of  robust and efficient parallel software. This book introduces a selection of the tools, which were presented and discussed at the 6th International Parallel Tools Workshop, held in Stuttgart, Germany, 25-26 September 2012.

  3. Parallel processor programs in the Federal Government

    Schneck, P. B.; Austin, D.; Squires, S. L.; Lehmann, J.; Mizell, D.; Wallgren, K.

    1985-01-01

    In 1982, a report dealing with the nation's research needs in high-speed computing called for increased access to supercomputing resources for the research community, research in computational mathematics, and increased research in the technology base needed for the next generation of supercomputers. Since that time a number of programs addressing future generations of computers, particularly parallel processors, have been started by U.S. government agencies. The present paper provides a description of the largest government programs in parallel processing. Established in fiscal year 1985 by the Institute for Defense Analyses for the National Security Agency, the Supercomputing Research Center will pursue research to advance the state of the art in supercomputing. Attention is also given to the DOE applied mathematical sciences research program, the NYU Ultracomputer project, the DARPA multiprocessor system architectures program, NSF research on multiprocessor systems, ONR activities in parallel computing, and NASA parallel processor projects.

  4. Density functional theory and parallel processing

    Ward, R.C.; Geist, G.A.; Butler, W.H.

    1987-01-01

    The authors demonstrate a method for obtaining the ground state energies and charge densities of a system of atoms described within density functional theory using simulated annealing on a parallel computer

  5. High performance parallel computers for science

    Nash, T.; Areti, H.; Atac, R.; Biel, J.; Cook, A.; Deppe, J.; Edel, M.; Fischler, M.; Gaines, I.; Hance, R.

    1989-01-01

    This paper reports that Fermilab's Advanced Computer Program (ACP) has been developing cost effective, yet practical, parallel computers for high energy physics since 1984. The ACP's latest developments are proceeding in two directions. A Second Generation ACP Multiprocessor System for experiments will include $3500 RISC processors each with performance over 15 VAX MIPS. To support such high performance, the new system allows parallel I/O, parallel interprocess communication, and parallel host processes. The ACP Multi-Array Processor, has been developed for theoretical physics. Each $4000 node is a FORTRAN or C programmable pipelined 20 Mflops (peak), 10 MByte single board computer. These are plugged into a 16 port crossbar switch crate which handles both inter and intra crate communication. The crates are connected in a hypercube. Site oriented applications like lattice gauge theory are supported by system software called CANOPY, which makes the hardware virtually transparent to users. A 256 node, 5 GFlop, system is under construction

  6. Cable line engineering

    Jang, Hak Sin; Kim, Sin Yeong

    1998-02-01

    This book is about cable line engineering. It is comprised of nine chapters, which deals with summary of cable communication such as way, process of cable communication and optical communication, Line constant of transmission on primary constant, reflection and crosstalk, communication cable line of types like flat cable, coaxial cable and loaded cable, Install of communication line with types and facility of aerial line, construction method of communication line facility, Measurement of communication line, Carrier communication of summary, PCM communication with Introduction, regeneration relay system sampling and quantization and Electric communication service and general information network with mobile communication technique and satellite communication system.

  7. Child protection decisions to substantiate hospital child protection teams' reports of suspected maltreatment.

    Jedwab, Merav; Benbenishty, Rami; Chen, Wendy; Glasser, Saralee; Siegal, Gil; Lerner-Geva, Liat

    2015-02-01

    The present study focuses on the way child protection officers (CPOs) in Israel assess suspected abuse and neglect (SCAN) reports made by hospital child protection teams (CPTs), to determine whether the alleged maltreatment is substantiated. The study was conducted in six medical centers and included 358 reports investigated by CPOs for SCAN. A structured questionnaire was completed by hospital CPTs to capture all relevant information on each child referred to the CPTs. Structured phone interviews were conducted with each of the CPOs who received a CPT report. Bivariate associations and multivariate logistic regressions were conducted to estimate the substantiation rate of cases reported by CPTs and the types of maltreatment substantiated, as well as to identify case characteristics of the child and the family that were associated with the CPOs' substantiation decision. CPO follow-up investigations revealed a substantiation rate of 53.5%. The maltreatment type most commonly substantiated was neglect. The case characteristics associated with substantiation included socio-demographic background, parents' health and functioning, previous contact with social services, characteristics of the hospital referral, medical findings and an assessment of the parents' behaviors. The findings of the study highlighted the importance of cooperation between the health and welfare services and the policy makers. This cooperation is essential for identifying early signs of maltreatment. Enhanced cooperation and effective information transfer between various professionals would help prevent or at least reduce the recurrence of maltreatment and would ensure that the children and their families are treated appropriately. Copyright © 2014 Elsevier Ltd. All rights reserved.

  8. Final Report: Migration Mechanisms for Large-scale Parallel Applications

    Jason Nieh

    2009-10-30

    Process migration is the ability to transfer a process from one machine to another. It is a useful facility in distributed computing environments, especially as computing devices become more pervasive and Internet access becomes more ubiquitous. The potential benefits of process migration, among others, are fault resilience by migrating processes off of faulty hosts, data access locality by migrating processes closer to the data, better system response time by migrating processes closer to users, dynamic load balancing by migrating processes to less loaded hosts, and improved service availability and administration by migrating processes before host maintenance so that applications can continue to run with minimal downtime. Although process migration provides substantial potential benefits and many approaches have been considered, achieving transparent process migration functionality has been difficult in practice. To address this problem, our work has designed, implemented, and evaluated new and powerful transparent process checkpoint-restart and migration mechanisms for desktop, server, and parallel applications that operate across heterogeneous cluster and mobile computing environments. A key aspect of this work has been to introduce lightweight operating system virtualization to provide processes with private, virtual namespaces that decouple and isolate processes from dependencies on the host operating system instance. This decoupling enables processes to be transparently checkpointed and migrated without modifying, recompiling, or relinking applications or the operating system. Building on this lightweight operating system virtualization approach, we have developed novel technologies that enable (1) coordinated, consistent checkpoint-restart and migration of multiple processes, (2) fast checkpointing of process and file system state to enable restart of multiple parallel execution environments and time travel, (3) process migration across heterogeneous

  9. Massively parallel evolutionary computation on GPGPUs

    Tsutsui, Shigeyoshi

    2013-01-01

    Evolutionary algorithms (EAs) are metaheuristics that learn from natural collective behavior and are applied to solve optimization problems in domains such as scheduling, engineering, bioinformatics, and finance. Such applications demand acceptable solutions with high-speed execution using finite computational resources. Therefore, there have been many attempts to develop platforms for running parallel EAs using multicore machines, massively parallel cluster machines, or grid computing environments. Recent advances in general-purpose computing on graphics processing units (GPGPU) have opened u

  10. .NET 4.5 parallel extensions

    Freeman, Bryan

    2013-01-01

    This book contains practical recipes on everything you will need to create task-based parallel programs using C#, .NET 4.5, and Visual Studio. The book is packed with illustrated code examples to create scalable programs.This book is intended to help experienced C# developers write applications that leverage the power of modern multicore processors. It provides the necessary knowledge for an experienced C# developer to work with .NET parallelism APIs. Previous experience of writing multithreaded applications is not necessary.

  11. Alternative derivation of the parallel ion viscosity

    Bravenec, R.V.; Berk, H.L.; Hammer, J.H.

    1982-01-01

    A set of double-adiabatic fluid equations with additional collisional relaxation between the ion temperatures parallel and perpendicular to a magnetic field are shown to reduce to a set involving a single temperature and a parallel viscosity. This result is applied to a recently published paper [R. V. Bravenec, A. J. Lichtenberg, M. A. Leiberman, and H. L. Berk, Phys. Fluids 24, 1320 (1981)] on viscous flow in a multiple-mirror configuration

  12. Acoustic simulation in architecture with parallel algorithm

    Li, Xiaohong; Zhang, Xinrong; Li, Dan

    2004-03-01

    In allusion to complexity of architecture environment and Real-time simulation of architecture acoustics, a parallel radiosity algorithm was developed. The distribution of sound energy in scene is solved with this method. And then the impulse response between sources and receivers at frequency segment, which are calculated with multi-process, are combined into whole frequency response. The numerical experiment shows that parallel arithmetic can improve the acoustic simulating efficiency of complex scene.

  13. PARALLEL SOLUTION METHODS OF PARTIAL DIFFERENTIAL EQUATIONS

    Korhan KARABULUT

    1998-03-01

    Full Text Available Partial differential equations arise in almost all fields of science and engineering. Computer time spent in solving partial differential equations is much more than that of in any other problem class. For this reason, partial differential equations are suitable to be solved on parallel computers that offer great computation power. In this study, parallel solution to partial differential equations with Jacobi, Gauss-Siedel, SOR (Succesive OverRelaxation and SSOR (Symmetric SOR algorithms is studied.

  14. Current distribution characteristics of superconducting parallel circuits

    Mori, K.; Suzuki, Y.; Hara, N.; Kitamura, M.; Tominaka, T.

    1994-01-01

    In order to increase the current carrying capacity of the current path of the superconducting magnet system, the portion of parallel circuits such as insulated multi-strand cables or parallel persistent current switches (PCS) are made. In superconducting parallel circuits of an insulated multi-strand cable or a parallel persistent current switch (PCS), the current distribution during the current sweep, the persistent mode, and the quench process were investigated. In order to measure the current distribution, two methods were used. (1) Each strand was surrounded with a pure iron core with the air gap. In the air gap, a Hall probe was located. The accuracy of this method was deteriorated by the magnetic hysteresis of iron. (2) The Rogowski coil without iron was used for the current measurement of each path in a 4-parallel PCS. As a result, it was shown that the current distribution characteristics of a parallel PCS is very similar to that of an insulated multi-strand cable for the quench process

  15. Parallel processing of structural integrity analysis codes

    Swami Prasad, P.; Dutta, B.K.; Kushwaha, H.S.

    1996-01-01

    Structural integrity analysis forms an important role in assessing and demonstrating the safety of nuclear reactor components. This analysis is performed using analytical tools such as Finite Element Method (FEM) with the help of digital computers. The complexity of the problems involved in nuclear engineering demands high speed computation facilities to obtain solutions in reasonable amount of time. Parallel processing systems such as ANUPAM provide an efficient platform for realising the high speed computation. The development and implementation of software on parallel processing systems is an interesting and challenging task. The data and algorithm structure of the codes plays an important role in exploiting the parallel processing system capabilities. Structural analysis codes based on FEM can be divided into two categories with respect to their implementation on parallel processing systems. The first category codes such as those used for harmonic analysis, mechanistic fuel performance codes need not require the parallelisation of individual modules of the codes. The second category of codes such as conventional FEM codes require parallelisation of individual modules. In this category, parallelisation of equation solution module poses major difficulties. Different solution schemes such as domain decomposition method (DDM), parallel active column solver and substructuring method are currently used on parallel processing systems. Two codes, FAIR and TABS belonging to each of these categories have been implemented on ANUPAM. The implementation details of these codes and the performance of different equation solvers are highlighted. (author). 5 refs., 12 figs., 1 tab

  16. The concept of parallel input/output processing for an electron linac

    Emoto, Takashi

    1993-01-01

    The instrumentation of and the control system for the PNC 10 MeV CW electron linac are described. A new concept of parallel input/output processing for the linac has been introduced. It is based on a substantial number of input/output processors(IOP) using beam control and diagnostics. The flexibility and simplicity of hardware/software are significant advantages with this scheme. (author)

  17. A massively-parallel electronic-structure calculations based on real-space density functional theory

    Iwata, Jun-Ichi; Takahashi, Daisuke; Oshiyama, Atsushi; Boku, Taisuke; Shiraishi, Kenji; Okada, Susumu; Yabana, Kazuhiro

    2010-01-01

    Based on the real-space finite-difference method, we have developed a first-principles density functional program that efficiently performs large-scale calculations on massively-parallel computers. In addition to efficient parallel implementation, we also implemented several computational improvements, substantially reducing the computational costs of O(N 3 ) operations such as the Gram-Schmidt procedure and subspace diagonalization. Using the program on a massively-parallel computer cluster with a theoretical peak performance of several TFLOPS, we perform electronic-structure calculations for a system consisting of over 10,000 Si atoms, and obtain a self-consistent electronic-structure in a few hundred hours. We analyze in detail the costs of the program in terms of computation and of inter-node communications to clarify the efficiency, the applicability, and the possibility for further improvements.

  18. DOE-EPRI On-Line Monitoring Implementation Guidelines

    E. Davis, R. Bickford

    2003-01-01

    Industry and EPRI experience at several plants has shown on-line monitoring to be very effective in identifying out-of-calibration instrument channels or indications of equipment-degradation problems. The EPRI implementation project for on-line monitoring has demonstrated the feasibility of on-line monitoring at several participating nuclear plants. The results have been very encouraging, and substantial progress is anticipated in the coming years

  19. GRADSPMHD: A parallel MHD code based on the SPH formalism

    Vanaverbeke, S.; Keppens, R.; Poedts, S.

    2014-03-01

    We present GRADSPMHD, a completely Lagrangian parallel magnetohydrodynamics code based on the SPH formalism. The implementation of the equations of SPMHD in the “GRAD-h” formalism assembles known results, including the derivation of the discretized MHD equations from a variational principle, the inclusion of time-dependent artificial viscosity, resistivity and conductivity terms, as well as the inclusion of a mixed hyperbolic/parabolic correction scheme for satisfying the ∇ṡB→ constraint on the magnetic field. The code uses a tree-based formalism for neighbor finding and can optionally use the tree code for computing the self-gravity of the plasma. The structure of the code closely follows the framework of our parallel GRADSPH FORTRAN 90 code which we added previously to the CPC program library. We demonstrate the capabilities of GRADSPMHD by running 1, 2, and 3 dimensional standard benchmark tests and we find good agreement with previous work done by other researchers. The code is also applied to the problem of simulating the magnetorotational instability in 2.5D shearing box tests as well as in global simulations of magnetized accretion disks. We find good agreement with available results on this subject in the literature. Finally, we discuss the performance of the code on a parallel supercomputer with distributed memory architecture. Catalogue identifier: AERP_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AERP_v1_0.html Program obtainable from: CPC Program Library, Queen’s University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 620503 No. of bytes in distributed program, including test data, etc.: 19837671 Distribution format: tar.gz Programming language: FORTRAN 90/MPI. Computer: HPC cluster. Operating system: Unix. Has the code been vectorized or parallelized?: Yes, parallelized using MPI. RAM: ˜30 MB for a

  20. Parallel Architectures and Parallel Algorithms for Integrated Vision Systems. Ph.D. Thesis

    Choudhary, Alok Nidhi

    1989-01-01

    Computer vision is regarded as one of the most complex and computationally intensive problems. An integrated vision system (IVS) is a system that uses vision algorithms from all levels of processing to perform for a high level application (e.g., object recognition). An IVS normally involves algorithms from low level, intermediate level, and high level vision. Designing parallel architectures for vision systems is of tremendous interest to researchers. Several issues are addressed in parallel architectures and parallel algorithms for integrated vision systems.

  1. Controlled Compact High Voltage Power Lines

    Postolati V.

    2016-04-01

    Full Text Available Nowadays modern overhead transmission lines (OHL constructions having several significant differences from conventional ones are being used in power grids more and more widely. Implementation of compact overhead lines equipped with FACTS devices, including phase angle regulator settings (compact controlled OHL, appears to be one of the most effective ways of power grid development. Compact controlled AC HV OHL represent a new generation of power transmission lines embodying recent advanced achievements in design solutions, including towers and insulation, together with interconnection schemes and control systems. Results of comprehensive research and development in relation to 110–500kV compact controlled power transmission lines together with theoretical basis, substantiation, and methodological approaches to their practical application are presented in the present paper.

  2. AUTOMATIC RAILWAY POWER LINE EXTRACTION USING MOBILE LASER SCANNING DATA

    S. Zhang

    2016-06-01

    Full Text Available Research on power line extraction technology using mobile laser point clouds has important practical significance on railway power lines patrol work. In this paper, we presents a new method for automatic extracting railway power line from MLS (Mobile Laser Scanning data. Firstly, according to the spatial structure characteristics of power-line and trajectory, the significant data is segmented piecewise. Then, use the self-adaptive space region growing method to extract power lines parallel with rails. Finally use PCA (Principal Components Analysis combine with information entropy theory method to judge a section of the power line whether is junction or not and which type of junction it belongs to. The least squares fitting algorithm is introduced to model the power line. An evaluation of the proposed method over a complicated railway point clouds acquired by a RIEGL VMX450 MLS system shows that the proposed method is promising.

  3. Concurrent computation of attribute filters on shared memory parallel machines

    Wilkinson, Michael H.F.; Gao, Hui; Hesselink, Wim H.; Jonker, Jan-Eppo; Meijster, Arnold

    2008-01-01

    Morphological attribute filters have not previously been parallelized mainly because they are both global and nonseparable. We propose a parallel algorithm that achieves efficient parallelism for a large class of attribute filters, including attribute openings, closings, thinnings, and thickenings,

  4. A task parallel implementation of fast multipole methods

    Taura, Kenjiro; Nakashima, Jun; Yokota, Rio; Maruyama, Naoya

    2012-01-01

    This paper describes a task parallel implementation of ExaFMM, an open source implementation of fast multipole methods (FMM), using a lightweight task parallel library MassiveThreads. Although there have been many attempts on parallelizing FMM

  5. Parallel phase model : a programming model for high-end parallel machines with manycores.

    Wu, Junfeng (Syracuse University, Syracuse, NY); Wen, Zhaofang; Heroux, Michael Allen; Brightwell, Ronald Brian

    2009-04-01

    This paper presents a parallel programming model, Parallel Phase Model (PPM), for next-generation high-end parallel machines based on a distributed memory architecture consisting of a networked cluster of nodes with a large number of cores on each node. PPM has a unified high-level programming abstraction that facilitates the design and implementation of parallel algorithms to exploit both the parallelism of the many cores and the parallelism at the cluster level. The programming abstraction will be suitable for expressing both fine-grained and coarse-grained parallelism. It includes a few high-level parallel programming language constructs that can be added as an extension to an existing (sequential or parallel) programming language such as C; and the implementation of PPM also includes a light-weight runtime library that runs on top of an existing network communication software layer (e.g. MPI). Design philosophy of PPM and details of the programming abstraction are also presented. Several unstructured applications that inherently require high-volume random fine-grained data accesses have been implemented in PPM with very promising results.

  6. The relation between reconnected flux, the parallel electric field, and the reconnection rate in a three-dimensional kinetic simulation of magnetic reconnection

    Wendel, D. E.; Olson, D. K.; Hesse, M.; Kuznetsova, M.; Adrian, M. L.; Aunai, N.; Karimabadi, H.; Daughton, W.

    2013-01-01

    We investigate the distribution of parallel electric fields and their relationship to the location and rate of magnetic reconnection in a large particle-in-cell simulation of 3D turbulent magnetic reconnection with open boundary conditions. The simulation's guide field geometry inhibits the formation of simple topological features such as null points. Therefore, we derive the location of potential changes in magnetic connectivity by finding the field lines that experience a large relative change between their endpoints, i.e., the quasi-separatrix layer. We find a good correspondence between the locus of changes in magnetic connectivity or the quasi-separatrix layer and the map of large gradients in the integrated parallel electric field (or quasi-potential). Furthermore, we investigate the distribution of the parallel electric field along the reconnecting field lines. We find the reconnection rate is controlled by only the low-amplitude, zeroth and first–order trends in the parallel electric field while the contribution from fluctuations of the parallel electric field, such as electron holes, is negligible. The results impact the determination of reconnection sites and reconnection rates in models and in situ spacecraft observations of 3D turbulent reconnection. It is difficult through direct observation to isolate the loci of the reconnection parallel electric field amidst the large amplitude fluctuations. However, we demonstrate that a positive slope of the running sum of the parallel electric field along the field line as a function of field line length indicates where reconnection is occurring along the field line

  7. Parallel evolutionary computation in bioinformatics applications.

    Pinho, Jorge; Sobral, João Luis; Rocha, Miguel

    2013-05-01

    A large number of optimization problems within the field of Bioinformatics require methods able to handle its inherent complexity (e.g. NP-hard problems) and also demand increased computational efforts. In this context, the use of parallel architectures is a necessity. In this work, we propose ParJECoLi, a Java based library that offers a large set of metaheuristic methods (such as Evolutionary Algorithms) and also addresses the issue of its efficient execution on a wide range of parallel architectures. The proposed approach focuses on the easiness of use, making the adaptation to distinct parallel environments (multicore, cluster, grid) transparent to the user. Indeed, this work shows how the development of the optimization library can proceed independently of its adaptation for several architectures, making use of Aspect-Oriented Programming. The pluggable nature of parallelism related modules allows the user to easily configure its environment, adding parallelism modules to the base source code when needed. The performance of the platform is validated with two case studies within biological model optimization. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  8. Parallelization of Subchannel Analysis Code MATRA

    Kim, Seongjin; Hwang, Daehyun; Kwon, Hyouk

    2014-01-01

    A stand-alone calculation of MATRA code used up pertinent computing time for the thermal margin calculations while a relatively considerable time is needed to solve the whole core pin-by-pin problems. In addition, it is strongly required to improve the computation speed of the MATRA code to satisfy the overall performance of the multi-physics coupling calculations. Therefore, a parallel approach to improve and optimize the computability of the MATRA code is proposed and verified in this study. The parallel algorithm is embodied in the MATRA code using the MPI communication method and the modification of the previous code structure was minimized. An improvement is confirmed by comparing the results between the single and multiple processor algorithms. The speedup and efficiency are also evaluated when increasing the number of processors. The parallel algorithm was implemented to the subchannel code MATRA using the MPI. The performance of the parallel algorithm was verified by comparing the results with those from the MATRA with the single processor. It is also noticed that the performance of the MATRA code was greatly improved by implementing the parallel algorithm for the 1/8 core and whole core problems

  9. Improvement of Parallel Algorithm for MATRA Code

    Kim, Seong-Jin; Seo, Kyong-Won; Kwon, Hyouk; Hwang, Dae-Hyun

    2014-01-01

    The feasibility study to parallelize the MATRA code was conducted in KAERI early this year. As a result, a parallel algorithm for the MATRA code has been developed to decrease a considerably required computing time to solve a bigsize problem such as a whole core pin-by-pin problem of a general PWR reactor and to improve an overall performance of the multi-physics coupling calculations. It was shown that the performance of the MATRA code was greatly improved by implementing the parallel algorithm using MPI communication. For problems of a 1/8 core and whole core for SMART reactor, a speedup was evaluated as about 10 when the numbers of used processor were 25. However, it was also shown that the performance deteriorated as the axial node number increased. In this paper, the procedure of a communication between processors is optimized to improve the previous parallel algorithm.. To improve the performance deterioration of the parallelized MATRA code, the communication algorithm between processors was newly presented. It was shown that the speedup was improved and stable regardless of the axial node number

  10. Iteration schemes for parallelizing models of superconductivity

    Gray, P.A. [Michigan State Univ., East Lansing, MI (United States)

    1996-12-31

    The time dependent Lawrence-Doniach model, valid for high fields and high values of the Ginzburg-Landau parameter, is often used for studying vortex dynamics in layered high-T{sub c} superconductors. When solving these equations numerically, the added degrees of complexity due to the coupling and nonlinearity of the model often warrant the use of high-performance computers for their solution. However, the interdependence between the layers can be manipulated so as to allow parallelization of the computations at an individual layer level. The reduced parallel tasks may then be solved independently using a heterogeneous cluster of networked workstations connected together with Parallel Virtual Machine (PVM) software. Here, this parallelization of the model is discussed and several computational implementations of varying degrees of parallelism are presented. Computational results are also given which contrast properties of convergence speed, stability, and consistency of these implementations. Included in these results are models involving the motion of vortices due to an applied current and pinning effects due to various material properties.

  11. Complex researches on substantiation of construction and seismic stability of large dams in seismic region

    Negmatullaev, S.Kh.; Yasunov, P.A.

    2001-01-01

    This article is devoted to complex researches on substantiation of construction and seismic stability of large dams (Nurec hydroelectric power station) in seismic region. Geological, seismological, model, and engineering investigations are discussed in this work. At construction of Nurec hydroelectric power station the rich experience is accumulated. This experience can be used in analogous seismically active regions at construction similar hydroelectric power stations.

  12. Do substantial BMI reduction episodes among Swedish schoolchildren have any impact on their final height?

    Nilsen, Bente B; Yngve, Agneta; Werner, Bo

    2018-02-06

    This study investigated whether substantial body mass index (BMI) reductions in Swedish schoolchildren aged seven years to 19 years, caused by disease, healthy or unhealthy behaviour, had any impact on their final height. We used height and weight data on 6572 subjects from two nationally representative longitudinal samples of Swedish children born in 1973 and 1981. These provided information on their final height and any BMI reduction episodes. Of the 6572 subjects (50.9% boys), among individuals with information on final height, 1118 had a BMI reduction of 5% and BMI reduction of 10% or more. On a group level, there was no statistically significant difference in the final height of individuals with BMI reductions of 10% or more and those without. The findings were independent of age and the subject's BMI at the start of the reduction episode. However, there were a number of cases where a substantial BMI reduction probably had an impact on the subject's final height. Our study found no evidence that a substantial BMI reduction had any impact on final height on a group level, but further analyses of specific case studies are necessary to determine whether substantial BMI reduction might have an impact on final height. ©2018 Foundation Acta Paediatrica. Published by John Wiley & Sons Ltd.

  13. Substantial secondary organic aerosol formation in a coniferous forest: observations of both day- and nighttime chemistry

    A. K. Y. Lee; J. P. D. Abbatt; W. R. Leaitch; S.-M. Li; S. J. Sjostedt; S. J. Sjostedt; J. J. B. Wentzell; J. Liggio; A. M. Macdonald

    2016-01-01

    Substantial biogenic secondary organic aerosol (BSOA) formation was investigated in a coniferous forest mountain region at Whistler, British Columbia. A largely biogenic aerosol growth episode was observed, providing a unique opportunity to investigate BSOA formation chemistry in a forested environment with limited influence from anthropogenic emissions. Positive matrix factorization of aerosol mass spectrometry (AMS) measurement identifie...

  14. 76 FR 41424 - Finding of Substantial Inadequacy of Implementation Plan; Call for Iowa State Implementation Plan...

    2011-07-14

    ... modeling requirement for this SIP call should allow for the use of the modeling protocol developed by Iowa... is making a finding that the Iowa State Implementation Plan (SIP) is substantially inadequate to... ) in Muscatine County, Iowa. The specific SIP deficiencies needing revision are described below. EPA is...

  15. 77 FR 65169 - Extension of Certain Timber Sale Contracts; Finding of Substantial Overriding Public Interest

    2012-10-25

    ... that are in breach, or (6) contracts when the purchaser's processing facility has not operated during... DEPARTMENT OF AGRICULTURE Forest Service Extension of Certain Timber Sale Contracts; Finding of Substantial Overriding Public Interest AGENCY: Forest Service, USDA. ACTION: Notice of contract extensions...

  16. 26 CFR 1.6662-4 - Substantial understatement of income tax.

    2010-04-01

    ... arrangement does not have as its principal purpose the avoidance or evasion of Federal income tax solely as a... 26 Internal Revenue 13 2010-04-01 2010-04-01 false Substantial understatement of income tax. 1... (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES Additions to the Tax, Additional Amounts, and Assessable...

  17. 75 FR 69454 - Emergency Homeowners' Loan Program: Notice of Allocation of Funding for Substantially Similar...

    2010-11-12

    ... assistance and provide security for such repayment by recordation of a HUD mortgage as a junior lien on the... current first lien mortgage loan payments for which the homeowner remains responsible. IV. Submission of.... Nondiscrimination and Civil Rights Requirements States operating existing programs that provide substantially...

  18. 75 FR 27504 - Substantial Product Hazard List: Hand-Held Hair Dryers

    2010-05-17

    ... immersion during their use. Section 15(a) of the CPSA defines ``substantial product hazard'' to include, a....'' Hand-held hair dryers routinely contain open-coil heating elements that are, in essence, uninsulated..., bathtub, or lavatory). The proposed rule would define ``hand-held hair dryer'' as ``an electrical...

  19. 76 FR 32215 - Agency Information Collection Activities; Proposed Collection; Comment Request; Substantiation...

    2011-06-03

    ... of such disease in the United States [(U.S.)], describes the role of a nutrient or dietary ingredient... substantiation for their claims in the scientific literature. The time it takes to assemble the necessary... based on emerging science, where conducting literature searches and understanding the literature takes...

  20. Theoretical-and-Methodological Substantiation of Multilingual Model Activity in Kazakhstan Higher School Education System

    Ospanova, Bikesh Revovna; Azimbayeva, Zhanat Amantayevna; Timokhina, Tatyana Vladimirovna; Seydakhmetova, Zergul Koblandiyevna

    2016-01-01

    The need of implementing the model of professional development in training an expert in the conditions of multilingualism is considered. The possibility of using the multilingual approach in the context of present day education with the use of innovative technologies of training is substantiated, the definition of "multilingual…

  1. 78 FR 889 - Finding of Substantial Inadequacy of Implementation Plan; Call for California State...

    2013-01-07

    ... Resources Defense Council; and Physicians for Social Responsibility--Los Angeles, (``environmental and... ENVIRONMENTAL PROTECTION AGENCY 40 CFR Part 52 [EPA-R09-OAR-2012-0721; FRL-9767-3] Finding of Substantial Inadequacy of Implementation Plan; Call for California State Implementation Plan Revision; South...

  2. 26 CFR 1.6662-5T - Substantial and gross valuation misstatements under chapter 1 (temporary).

    2010-04-01

    ..., DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES Additions to the Tax, Additional... persons). (iii) Property. For purposes of this section, the term property refers to both tangible and... adjustments. For rules relating to the penalty imposed with respect to a substantial or gross valuation...

  3. Domestic violence, parental substance misuse and the decision to substantiate child maltreatment.

    Victor, Bryan G; Grogan-Kaylor, Andrew; Ryan, Joseph P; Perron, Brian E; Gilbert, Terri Ticknor

    2018-05-01

    Families that experience domestic violence and parental substance misuse are disproportionately involved with the child welfare system. Prior research suggests that child protective services (CPS) caseworkers are more likely to substantiate maltreatment allegations when domestic violence and parental substance misuse are identified during the investigation, pointing to one possible mechanism for this disproportionate involvement. While previous studies have relied on nationally representative data sets, the current study used administrative records from a large Midwestern child welfare agency that accounts for state-level variation in child welfare policy and practice. A total of 501,060 substantiation decisions made between 2009 and 2013 were examined to assess the influence of caseworker-perceived domestic violence and parental substance misuse on the decision to substantiate reported maltreatment. Results from multilevel modeling suggest that the identification of domestic violence and parental substance misuse during an investigation significantly increased the probability that an allegation would be substantiated. The implication of these findings for child welfare practice are considered in light of the fact that many child welfare agencies do not consider exposure to domestic violence and parental substance misuse in and of themselves to constitute child maltreatment. Copyright © 2018 Elsevier Ltd. All rights reserved.

  4. 26 CFR 301.6229(c)(2)-1T - Substantial omission of income (temporary).

    2010-04-01

    ....6229(c)(2)-1T Substantial omission of income (temporary). (a) Partnership return—(1) General rule. (i) If any partnership omits from the gross income stated in its return an amount properly includible... omitted from gross income if information sufficient to apprise the Commissioner of the nature and amount...

  5. 26 CFR 1.162-17 - Reporting and substantiation of certain business expenses of employees.

    2010-04-01

    ... Individuals and Corporations § 1.162-17 Reporting and substantiation of certain business expenses of employees.... The term “ordinary and necessary business expenses” means only those expenses which are ordinary and..., reimbursements, or otherwise, exceeds the ordinary and necessary business expenses paid or incurred by the...

  6. 34 CFR 692.41 - What standards may a State use to determine substantial financial need?

    2010-07-01

    ... PARTNERSHIP PROGRAM Leveraging Educational Assistance Partnership Program How Does A State Select Students... establishes that are approved by the Secretary. A State may define substantial financial need in terms of... “independent student” as defined under section 480(d) of the HEA. However, for good cause shown, the Secretary...

  7. VT Electric Transmission Line Corridors - corridor lines

    Vermont Center for Geographic Information — (Link to Metadata) The ELTRN layer depicts electric transmission line corridors in Vermont. Various methods have been used to digitize features. The data layer...

  8. The parallel dynamics of drift wave turbulence in the WEGA stellarator

    Marsen, S; Endler, M; Otte, M; Wagner, F, E-mail: stefan.marsen@ipp.mpg.d [Max-Planck-Institut fuer Plasmaphysik, EURATOM Association, Wendelsteinstrasse 1, 17491 Greifswald (Germany)

    2009-08-15

    The three-dimensional structure of turbulence in the edge (inside the last closed flux surface) of the WEGA stellarator is studied focusing on the parallel dynamics. WEGA as a small stellarator with moderate plasma parameters offers the opportunity to study turbulence with Langmuir probes providing high spatial and temporal resolution. Multiple probes with radial, poloidal and toroidal resolution are used to measure density fluctuations. Correlation analysis is used to reconstruct a 3D picture of turbulent structures. We find that these structures originate predominantly on the low field side and have a three-dimensional character with a finite averaged parallel wavenumber. The ratio between the parallel and perpendicular wavenumber component is in the order of 10{sup -2}. The parallel dynamics are compared at magnetic inductions of 57 and 500 mT. At 500 mT, the parallel wavelength is in the order of the field line connection length 2{pi}R{iota}-bar. A frequency resolved measure of k{sub ||}/k{sub {theta}} shows a constant ratio in this case. At 57 mT the observed k{sub ||} is much smaller than at 500 mT. However, the observed small average value is due to an averaging over positive and negative components pointing parallel and antiparallel to the magnetic field vector.

  9. Preliminary Study on the Enhancement of Reconstruction Speed for Emission Computed Tomography Using Parallel Processing

    Park, Min Jae; Lee, Jae Sung; Kim, Soo Mee; Kang, Ji Yeon; Lee, Dong Soo; Park, Kwang Suk

    2009-01-01

    Conventional image reconstruction uses simplified physical models of projection. However, real physics, for example 3D reconstruction, takes too long time to process all the data in clinic and is unable in a common reconstruction machine because of the large memory for complex physical models. We suggest the realistic distributed memory model of fast-reconstruction using parallel processing on personal computers to enable large-scale technologies. The preliminary tests for the possibility on virtual machines and various performance test on commercial super computer, Tachyon were performed. Expectation maximization algorithm with common 2D projection and realistic 3D line of response were tested. Since the process time was getting slower (max 6 times) after a certain iteration, optimization for compiler was performed to maximize the efficiency of parallelization. Parallel processing of a program on multiple computers was available on Linux with MPICH and NFS. We verified that differences between parallel processed image and single processed image at the same iterations were under the significant digits of floating point number, about 6 bit. Double processors showed good efficiency (1.96 times) of parallel computing. Delay phenomenon was solved by vectorization method using SSE. Through the study, realistic parallel computing system in clinic was established to be able to reconstruct by plenty of memory using the realistic physical models which was impossible to simplify

  10. New design of an RSFQ parallel multiply-accumulate unit

    Kataeva, Irina; Engseth, Henrik; Kidiyarova-Shevchenko, Anna

    2006-01-01

    The multiply-accumulate unit (MAC) is a central component of a successive interference canceller, an advanced receiver for W-CDMA base stations. A 4 x 4 two's complement fixed point RSFQ MAC with rounding to 5 bits has been simulated using VHDL, and maximum performance is equal to 24 GMACS (giga-multiply-accumulates per second). The clock distribution network has been re-designed from a linear ripple to a binary tree network in order to eliminate the data dependence of the clock propagation speed and reduce the number of Josephson junctions in clock lines. The 4 x 4 bit MAC has been designed for the HYPRES 4.5 kA cm -2 process and its components have been experimentally tested at low frequency: the 5-bit combiner, using an exhaustive test pattern, had margins on DC bias voltage of ± 18%, and the 4 x 4 parallel multiplier had margins equal to ± 2%

  11. Non-Almost Periodicity of Parallel Transports for Homogeneous Connections

    Brunnemann, Johannes; Fleischhack, Christian

    2012-01-01

    Let A be the affine space of all connections in an SU(2) principal fibre bundle over ℝ 3 . The set of homogeneous isotropic connections forms a line l in A. We prove that the parallel transports for general, non-straight paths in the base manifold do not depend almost periodically on l. Consequently, the embedding l ↪ A does not continuously extend to an embedding l-bar ↪ A-bar of the respective compactifications. Here, the Bohr compactification l-bar corresponds to the configuration space of homogeneous isotropic loop quantum cosmology and A-bar to that of loop quantum gravity. Analogous results are given for the anisotropic case.

  12. Microprocessor event analysis in parallel with Camac data acquisition

    Cords, D.; Eichler, R.; Riege, H.

    1981-01-01

    The Plessey MIPROC-16 microprocessor (16 bits, 250 ns execution time) has been connected to a Camac System (GEC-ELLIOTT System Crate) and shares the Camac access with a Nord-1OS computer. Interfaces have been designed and tested for execution of Camac cycles, communication with the Nord-1OS computer and DMA-transfer from Camac to the MIPROC-16 memory. The system is used in the JADE data-acquisition-system at PETRA where it receives the data from the detector in parallel with the Nord-1OS computer via DMA through the indirect-data-channel mode. The microprocessor performs an on-line analysis of events and the result of various checks is appended to the event. In case of spurious triggers or clear beam gas events, the Nord-1OS buffer will be reset and the event omitted from further processing. (orig.)

  13. Microprocessor event analysis in parallel with CAMAC data acquisition

    Cords, D; Riege, H

    1981-01-01

    The Plessey MIPROC-16 microprocessor (16 bits, 250 ns execution time) has been connected to a CAMAC System (GEC-ELLIOTT System Crate) and shares the CAMAC access with a Nord-10S computer. Interfaces have been designed and tested for execution of CAMAC cycles, communication with the Nord-10S computer and DMA-transfer from CAMAC to the MIPROC-16 memory. The system is used in the JADE data-acquisition-system at PETRA where it receives the data from the detector in parallel with the Nord-10S computer via DMA through the indirect-data-channel mode. The microprocessor performs an on-line analysis of events and the results of various checks is appended to the event. In case of spurious triggers or clear beam gas events, the Nord-10S buffer will be reset and the event omitted from further processing. (5 refs).

  14. Mouse-tracking evidence for parallel anticipatory option evaluation.

    Cranford, Edward A; Moss, Jarrod

    2017-12-23

    In fast-paced, dynamic tasks, the ability to anticipate the future outcome of a sequence of events is crucial to quickly selecting an appropriate course of action among multiple alternative options. There are two classes of theories that describe how anticipation occurs. Serial theories assume options are generated and evaluated one at a time, in order of quality, whereas parallel theories assume simultaneous generation and evaluation. The present research examined the option evaluation process during a task designed to be analogous to prior anticipation tasks, but within the domain of narrative text comprehension. Prior research has relied on indirect, off-line measurement of the option evaluation process during anticipation tasks. Because the movement of the hand can provide a window into underlying cognitive processes, online metrics such as continuous mouse tracking provide more fine-grained measurements of cognitive processing as it occurs in real time. In this study, participants listened to three-sentence stories and predicted the protagonists' final action by moving a mouse toward one of three possible options. Each story was presented with either one (control condition) or two (distractor condition) plausible ending options. Results seem most consistent with a parallel option evaluation process because initial mouse trajectories deviated further from the best option in the distractor condition compared to the control condition. It is difficult to completely rule out all possible serial processing accounts, although the results do place constraints on the time frame in which a serial processing explanation must operate.

  15. Cosmic Shear With ACS Pure Parallels. Targeted Portion.

    Rhodes, Jason

    2002-07-01

    Small distortions in the shapes of background galaxies by foreground mass provide a powerful method of directly measuring the amount and distribution of dark matter. Several groups have recently detected this weak lensing by large-scale structure, also called cosmic shear. The high resolution and sensitivity of HST/ACS provide a unique opportunity to measure cosmic shear accurately on small scales. Using 260 parallel orbits in Sloan i {F775W} we will measure for the first time: the cosmic shear variance on scales Omega_m^0.5, with signal-to-noise {s/n} 20, and the mass density Omega_m with s/n=4. They will be done at small angular scales where non-linear effects dominate the power spectrum, providing a test of the gravitational instability paradigm for structure formation. Measurements on these scales are not possible from the ground, because of the systematic effects induced by PSF smearing from seeing. Having many independent lines of sight reduces the uncertainty due to cosmic variance, making parallel observations ideal.

  16. Parallel visualization on leadership computing resources

    Peterka, T; Ross, R B [Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, IL 60439 (United States); Shen, H-W [Department of Computer Science and Engineering, Ohio State University, Columbus, OH 43210 (United States); Ma, K-L [Department of Computer Science, University of California at Davis, Davis, CA 95616 (United States); Kendall, W [Department of Electrical Engineering and Computer Science, University of Tennessee at Knoxville, Knoxville, TN 37996 (United States); Yu, H, E-mail: tpeterka@mcs.anl.go [Sandia National Laboratories, California, Livermore, CA 94551 (United States)

    2009-07-01

    Changes are needed in the way that visualization is performed, if we expect the analysis of scientific data to be effective at the petascale and beyond. By using similar techniques as those used to parallelize simulations, such as parallel I/O, load balancing, and effective use of interprocess communication, the supercomputers that compute these datasets can also serve as analysis and visualization engines for them. Our team is assessing the feasibility of performing parallel scientific visualization on some of the most powerful computational resources of the U.S. Department of Energy's National Laboratories in order to pave the way for analyzing the next generation of computational results. This paper highlights some of the conclusions of that research.

  17. Parallelization of ITOUGH2 using PVM

    Finsterle, Stefan

    1998-01-01

    ITOUGH2 inversions are computationally intensive because the forward problem must be solved many times to evaluate the objective function for different parameter combinations or to numerically calculate sensitivity coefficients. Most of these forward runs are independent from each other and can therefore be performed in parallel. Message passing based on the Parallel Virtual Machine (PVM) system has been implemented into ITOUGH2 to enable parallel processing of ITOUGH2 jobs on a heterogeneous network of Unix workstations. This report describes the PVM system and its implementation into ITOUGH2. Instructions are given for installing PVM, compiling ITOUGH2-PVM for use on a workstation cluster, the preparation of an 1.TOUGH2 input file under PVM, and the execution of an ITOUGH2-PVM application. Examples are discussed, demonstrating the use of ITOUGH2-PVM

  18. Distributed Parallel Architecture for "Big Data"

    Catalin BOJA

    2012-01-01

    Full Text Available This paper is an extension to the "Distributed Parallel Architecture for Storing and Processing Large Datasets" paper presented at the WSEAS SEPADS’12 conference in Cambridge. In its original version the paper went over the benefits of using a distributed parallel architecture to store and process large datasets. This paper analyzes the problem of storing, processing and retrieving meaningful insight from petabytes of data. It provides a survey on current distributed and parallel data processing technologies and, based on them, will propose an architecture that can be used to solve the analyzed problem. In this version there is more emphasis put on distributed files systems and the ETL processes involved in a distributed environment.

  19. Java parallel secure stream for grid computing

    Chen, J.; Akers, W.; Chen, Y.; Watson, W.

    2001-01-01

    The emergence of high speed wide area networks makes grid computing a reality. However grid applications that need reliable data transfer still have difficulties to achieve optimal TCP performance due to network tuning of TCP window size to improve the bandwidth and to reduce latency on a high speed wide area network. The authors present a pure Java package called JPARSS (Java Parallel Secure Stream) that divides data into partitions that are sent over several parallel Java streams simultaneously and allows Java or Web applications to achieve optimal TCP performance in a gird environment without the necessity of tuning the TCP window size. Several experimental results are provided to show that using parallel stream is more effective than tuning TCP window size. In addition X.509 certificate based single sign-on mechanism and SSL based connection establishment are integrated into this package. Finally a few applications using this package will be discussed

  20. Applications of Parallel Processing in Mobile Banking

    2007-01-01

    Full Text Available The future of mobile banking will be represented by such applications that support mobile, Internet banking and EFT (Electronic Funds Transfer transactions in a single user interface. In such a way, the mobile banking will be able to cover all the types of applications demanded at the market level. The parallel processing of credit card bank transactions could be performed with the help of a grid network. Excluding some limitations, the grid processing offers huge opportunities to exploit the parallelism. For this reason, a lot of applications of waiting queues in grid processing were developed in the last years. Grid networks represent a distinctive and very modern field of the parallel and distributed processing.