WorldWideScience

Sample records for performing parallel enzyme

  1. Implementation and performance of parallelized elegant

    International Nuclear Information System (INIS)

    Wang, Y.; Borland, M.

    2008-01-01

    The program elegant is widely used for design and modeling of linacs for free-electron lasers and energy recovery linacs, as well as storage rings and other applications. As part of a multi-year effort, we have parallelized many aspects of the code, including single-particle dynamics, wakefields, and coherent synchrotron radiation. We report on the approach used for gradual parallelization, which proved very beneficial in getting parallel features into the hands of users quickly. We also report details of parallelization of collective effects. Finally, we discuss performance of the parallelized code in various applications.

  2. High performance parallel I/O

    CERN Document Server

    Prabhat

    2014-01-01

    Gain Critical Insight into the Parallel I/O EcosystemParallel I/O is an integral component of modern high performance computing (HPC), especially in storing and processing very large datasets to facilitate scientific discovery. Revealing the state of the art in this field, High Performance Parallel I/O draws on insights from leading practitioners, researchers, software architects, developers, and scientists who shed light on the parallel I/O ecosystem.The first part of the book explains how large-scale HPC facilities scope, configure, and operate systems, with an emphasis on choices of I/O har

  3. High performance parallel computers for science

    International Nuclear Information System (INIS)

    Nash, T.; Areti, H.; Atac, R.; Biel, J.; Cook, A.; Deppe, J.; Edel, M.; Fischler, M.; Gaines, I.; Hance, R.

    1989-01-01

    This paper reports that Fermilab's Advanced Computer Program (ACP) has been developing cost effective, yet practical, parallel computers for high energy physics since 1984. The ACP's latest developments are proceeding in two directions. A Second Generation ACP Multiprocessor System for experiments will include $3500 RISC processors each with performance over 15 VAX MIPS. To support such high performance, the new system allows parallel I/O, parallel interprocess communication, and parallel host processes. The ACP Multi-Array Processor, has been developed for theoretical physics. Each $4000 node is a FORTRAN or C programmable pipelined 20 Mflops (peak), 10 MByte single board computer. These are plugged into a 16 port crossbar switch crate which handles both inter and intra crate communication. The crates are connected in a hypercube. Site oriented applications like lattice gauge theory are supported by system software called CANOPY, which makes the hardware virtually transparent to users. A 256 node, 5 GFlop, system is under construction

  4. Performance of the Galley Parallel File System

    Science.gov (United States)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    As the input/output (I/O) needs of parallel scientific applications increase, file systems for multiprocessors are being designed to provide applications with parallel access to multiple disks. Many parallel file systems present applications with a conventional Unix-like interface that allows the application to access multiple disks transparently. This interface conceals the parallism within the file system, which increases the ease of programmability, but makes it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. Furthermore, most current parallel file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic parallel workloads. Initial experiments, reported in this paper, indicate that Galley is capable of providing high-performance 1/O to applications the applications that rely on them. In Section 3 we describe that access data in patterns that have been observed to be common.

  5. Performance studies of the parallel VIM code

    International Nuclear Information System (INIS)

    Shi, B.; Blomquist, R.N.

    1996-01-01

    In this paper, the authors evaluate the performance of the parallel version of the VIM Monte Carlo code on the IBM SPx at the High Performance Computing Research Facility at ANL. Three test problems with contrasting computational characteristics were used to assess effects in performance. A statistical method for estimating the inefficiencies due to load imbalance and communication is also introduced. VIM is a large scale continuous energy Monte Carlo radiation transport program and was parallelized using history partitioning, the master/worker approach, and p4 message passing library. Dynamic load balancing is accomplished when the master processor assigns chunks of histories to workers that have completed a previously assigned task, accommodating variations in the lengths of histories, processor speeds, and worker loads. At the end of each batch (generation), the fission sites and tallies are sent from each worker to the master process, contributing to the parallel inefficiency. All communications are between master and workers, and are serial. The SPx is a scalable 128-node parallel supercomputer with high-performance Omega switches of 63 microsec latency and 35 MBytes/sec bandwidth. For uniform and reproducible performance, they used only the 120 identical regular processors (IBM RS/6000) and excluded the remaining eight planet nodes, which may be loaded by other's jobs

  6. Flexibility and Performance of Parallel File Systems

    Science.gov (United States)

    Kotz, David; Nieuwejaar, Nils

    1996-01-01

    As we gain experience with parallel file systems, it becomes increasingly clear that a single solution does not suit all applications. For example, it appears to be impossible to find a single appropriate interface, caching policy, file structure, or disk-management strategy. Furthermore, the proliferation of file-system interfaces and abstractions make applications difficult to port. We propose that the traditional functionality of parallel file systems be separated into two components: a fixed core that is standard on all platforms, encapsulating only primitive abstractions and interfaces, and a set of high-level libraries to provide a variety of abstractions and application-programmer interfaces (API's). We present our current and next-generation file systems as examples of this structure. Their features, such as a three-dimensional file structure, strided read and write interfaces, and I/O-node programs, are specifically designed with the flexibility and performance necessary to support a wide range of applications.

  7. Analysis of parallel computing performance of the code MCNP

    International Nuclear Information System (INIS)

    Wang Lei; Wang Kan; Yu Ganglin

    2006-01-01

    Parallel computing can reduce the running time of the code MCNP effectively. With the MPI message transmitting software, MCNP5 can achieve its parallel computing on PC cluster with Windows operating system. Parallel computing performance of MCNP is influenced by factors such as the type, the complexity level and the parameter configuration of the computing problem. This paper analyzes the parallel computing performance of MCNP regarding with these factors and gives measures to improve the MCNP parallel computing performance. (authors)

  8. Multitasking TORT Under UNICOS: Parallel Performance Models and Measurements

    International Nuclear Information System (INIS)

    Azmy, Y.Y.; Barnett, D.A.

    1999-01-01

    The existing parallel algorithms in the TORT discrete ordinates were updated to function in a UNI-COS environment. A performance model for the parallel overhead was derived for the existing algorithms. The largest contributors to the parallel overhead were identified and a new algorithm was developed. A parallel overhead model was also derived for the new algorithm. The results of the comparison of parallel performance models were compared to applications of the code to two TORT standard test problems and a large production problem. The parallel performance models agree well with the measured parallel overhead

  9. Multitasking TORT under UNICOS: Parallel performance models and measurements

    International Nuclear Information System (INIS)

    Barnett, A.; Azmy, Y.Y.

    1999-01-01

    The existing parallel algorithms in the TORT discrete ordinates code were updated to function in a UNICOS environment. A performance model for the parallel overhead was derived for the existing algorithms. The largest contributors to the parallel overhead were identified and a new algorithm was developed. A parallel overhead model was also derived for the new algorithm. The results of the comparison of parallel performance models were compared to applications of the code to two TORT standard test problems and a large production problem. The parallel performance models agree well with the measured parallel overhead

  10. Performance Analysis of Parallel Mathematical Subroutine library PARCEL

    International Nuclear Information System (INIS)

    Yamada, Susumu; Shimizu, Futoshi; Kobayashi, Kenichi; Kaburaki, Hideo; Kishida, Norio

    2000-01-01

    The parallel mathematical subroutine library PARCEL (Parallel Computing Elements) has been developed by Japan Atomic Energy Research Institute for easy use of typical parallelized mathematical codes in any application problems on distributed parallel computers. The PARCEL includes routines for linear equations, eigenvalue problems, pseudo-random number generation, and fast Fourier transforms. It is shown that the results of performance for linear equations routines exhibit good parallelization efficiency on vector, as well as scalar, parallel computers. A comparison of the efficiency results with the PETSc (Portable Extensible Tool kit for Scientific Computations) library has been reported. (author)

  11. High performance parallel backprojection on FPGA

    Energy Technology Data Exchange (ETDEWEB)

    Pfanner, Florian; Knaup, Michael; Kachelriess, Marc [Erlangen-Nuernberg Univ., Erlangen (Germany). Inst. of Medical Physics (IMP)

    2011-07-01

    Reconstruction of tomographic images, i.e., images from a Computed Tomography scanner, is a very time consuming issue. The most calculation power is needed for the backprojection step. A closer inspection shows that the algorithm for backprojection is easy to parallelize. FPGAs are able to execute many operations in the same time, so a highly parallel algorithm is a requirement for a powerful acceleration. For data flow rate maximization, we realized the backprojection in a pipelined structure with data throughput of one clock cycle. Due the hardware limitations of the FPGA, it is not possible to reconstruct the image as a whole. So it is necessary to split up the image and reconstruct these parts separately. Despite that, a reconstruction of 512 projections into a 5122 image is calculated within 13 ms on a Virtex 5 FPGA. To save hardware resources we use fixed point arithmetic with an accuracy of 23 bit for calculation. A comparison of the result image and an image, calculated with floating point arithmetic on CPU, shows that there are no differences between these images. (orig.)

  12. High Performance Parallel Multigrid Algorithms for Unstructured Grids

    Science.gov (United States)

    Frederickson, Paul O.

    1996-01-01

    We describe a high performance parallel multigrid algorithm for a rather general class of unstructured grid problems in two and three dimensions. The algorithm PUMG, for parallel unstructured multigrid, is related in structure to the parallel multigrid algorithm PSMG introduced by McBryan and Frederickson, for they both obtain a higher convergence rate through the use of multiple coarse grids. Another reason for the high convergence rate of PUMG is its smoother, an approximate inverse developed by Baumgardner and Frederickson.

  13. Circuit mismatch influence on performance of paralleling silicon carbide MOSFETs

    DEFF Research Database (Denmark)

    Li, Helong; Munk-Nielsen, Stig; Pham, Cam

    2014-01-01

    This paper focuses on circuit mismatch influence on performance of paralleling SiC MOSFETs. Power circuit mismatch and gate driver mismatch influences are analyzed in detail. Simulation and experiment results show the influence of circuit mismatch and verify the analysis. This paper aims to give...... suggestions on paralleling discrete SiC MOSFETs and designing layout of power modules with paralleled SiC MOSFETs dies....

  14. Designing a High Performance Parallel Personal Cluster

    OpenAIRE

    Kapanova, K. G.; Sellier, J. M.

    2016-01-01

    Today, many scientific and engineering areas require high performance computing to perform computationally intensive experiments. For example, many advances in transport phenomena, thermodynamics, material properties, computational chemistry and physics are possible only because of the availability of such large scale computing infrastructures. Yet many challenges are still open. The cost of energy consumption, cooling, competition for resources have been some of the reasons why the scientifi...

  15. The Performance of an Object-Oriented, Parallel Operating System

    Directory of Open Access Journals (Sweden)

    David R. Kohr, Jr.

    1994-01-01

    Full Text Available The nascent and rapidly evolving state of parallel systems often leaves parallel application developers at the mercy of inefficient, inflexible operating system software. Given the relatively primitive state of parallel systems software, maximizing the performance of parallel applications not only requires judicious tuning of the application software, but occasionally, the replacement of specific system software modules with others that can more readily respond to the imposed pattern of resource demands. To assess the feasibility of application and performance tuning via malleable system software and to understand the performance penalties for detailed operating system performance data capture, we describe a set of performance instrumentation techniques for parallel, object-oriented operating systems and a set of performance experiments with Choices, an experimental, object-oriented operating system designed for use with parallel sys- tems. These performance experiments show that (a the performance overhead for operating system data capture is modest, (b the penalty for malleable, object-oriented operating systems is negligible, but (c techniques are needed to strictly enforce adherence of implementation to design if operating system modules are to be replaced.

  16. Misleading Performance Claims in Parallel Computations

    Energy Technology Data Exchange (ETDEWEB)

    Bailey, David H.

    2009-05-29

    In a previous humorous note entitled 'Twelve Ways to Fool the Masses,' I outlined twelve common ways in which performance figures for technical computer systems can be distorted. In this paper and accompanying conference talk, I give a reprise of these twelve 'methods' and give some actual examples that have appeared in peer-reviewed literature in years past. I then propose guidelines for reporting performance, the adoption of which would raise the level of professionalism and reduce the level of confusion, not only in the world of device simulation but also in the larger arena of technical computing.

  17. Implementation of a high performance parallel finite element micromagnetics package

    International Nuclear Information System (INIS)

    Scholz, W.; Suess, D.; Dittrich, R.; Schrefl, T.; Tsiantos, V.; Forster, H.; Fidler, J.

    2004-01-01

    A new high performance scalable parallel finite element micromagnetics package has been implemented. It includes solvers for static energy minimization, time integration of the Landau-Lifshitz-Gilbert equation, and the nudged elastic band method

  18. Identification of parallel and divergent optimization solutions for homologous metabolic enzymes

    Directory of Open Access Journals (Sweden)

    Robert F. Standaert

    2018-06-01

    Full Text Available Metabolic pathway assembly typically involves the expression of enzymes from multiple organisms in a single heterologous host. Ensuring that each enzyme functions effectively can be challenging, since many potential factors can disrupt proper pathway flux. Here, we compared the performance of two enzyme homologs in a pathway engineered to allow Escherichia coli to grow on 4-hydroxybenzoate (4-HB, a byproduct of lignocellulosic biomass deconstruction. Single chromosomal copies of the 4-HB 3-monooxygenase genes pobA and praI, from Pseudomonas putida KT2440 and Paenibacillus sp. JJ-1B, respectively, were introduced into a strain able to metabolize protocatechuate (PCA, the oxidation product of 4-HB. Neither enzyme initially supported consistent growth on 4-HB. Experimental evolution was used to identify mutations that improved pathway activity. For both enzymes, silent mRNA mutations were identified that increased enzyme expression. With pobA, duplication of the genes for PCA metabolism allowed growth on 4-HB. However, with praI, growth required a mutation in the 4-HB/PCA transporter pcaK that increased intracellular concentrations of 4-HB, suggesting that flux through PraI was limiting. These findings demonstrate the value of directed evolution strategies to rapidly identify and overcome diverse factors limiting enzyme activity. Keywords: Lignin, Protocatechuate, Experimental evolution

  19. Identification of parallel and divergent optimization solutions for homologous metabolic enzymes.

    Science.gov (United States)

    Standaert, Robert F; Giannone, Richard J; Michener, Joshua K

    2018-06-01

    Metabolic pathway assembly typically involves the expression of enzymes from multiple organisms in a single heterologous host. Ensuring that each enzyme functions effectively can be challenging, since many potential factors can disrupt proper pathway flux. Here, we compared the performance of two enzyme homologs in a pathway engineered to allow Escherichia coli to grow on 4-hydroxybenzoate (4-HB), a byproduct of lignocellulosic biomass deconstruction. Single chromosomal copies of the 4-HB 3-monooxygenase genes pobA and praI , from Pseudomonas putida KT2440 and Paenibacillus sp. JJ-1B, respectively, were introduced into a strain able to metabolize protocatechuate (PCA), the oxidation product of 4-HB. Neither enzyme initially supported consistent growth on 4-HB. Experimental evolution was used to identify mutations that improved pathway activity. For both enzymes, silent mRNA mutations were identified that increased enzyme expression. With pobA , duplication of the genes for PCA metabolism allowed growth on 4-HB. However, with praI , growth required a mutation in the 4-HB/PCA transporter pcaK that increased intracellular concentrations of 4-HB, suggesting that flux through PraI was limiting. These findings demonstrate the value of directed evolution strategies to rapidly identify and overcome diverse factors limiting enzyme activity.

  20. Microwave tomography global optimization, parallelization and performance evaluation

    CERN Document Server

    Noghanian, Sima; Desell, Travis; Ashtari, Ali

    2014-01-01

    This book provides a detailed overview on the use of global optimization and parallel computing in microwave tomography techniques. The book focuses on techniques that are based on global optimization and electromagnetic numerical methods. The authors provide parallelization techniques on homogeneous and heterogeneous computing architectures on high performance and general purpose futuristic computers. The book also discusses the multi-level optimization technique, hybrid genetic algorithm and its application in breast cancer imaging.

  1. A high performance parallel approach to medical imaging

    International Nuclear Information System (INIS)

    Frieder, G.; Frieder, O.; Stytz, M.R.

    1988-01-01

    Research into medical imaging using general purpose parallel processing architectures is described and a review of the performance of previous medical imaging machines is provided. Results demonstrating that general purpose parallel architectures can achieve performance comparable to other, specialized, medical imaging machine architectures is presented. A new back-to-front hidden-surface removal algorithm is described. Results demonstrating the computational savings obtained by using the modified back-to-front hidden-surface removal algorithm are presented. Performance figures for forming a full-scale medical image on a mesh interconnected multiprocessor are presented

  2. Performance of Air Pollution Models on Massively Parallel Computers

    DEFF Research Database (Denmark)

    Brown, John; Hansen, Per Christian; Wasniewski, Jerzy

    1996-01-01

    To compare the performance and use of three massively parallel SIMD computers, we implemented a large air pollution model on the computers. Using a realistic large-scale model, we gain detailed insight about the performance of the three computers when used to solve large-scale scientific problems...

  3. Tuning HDF5 subfiling performance on parallel file systems

    Energy Technology Data Exchange (ETDEWEB)

    Byna, Suren [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Chaarawi, Mohamad [Intel Corp. (United States); Koziol, Quincey [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Mainzer, John [The HDF Group (United States); Willmore, Frank [The HDF Group (United States)

    2017-05-12

    Subfiling is a technique used on parallel file systems to reduce locking and contention issues when multiple compute nodes interact with the same storage target node. Subfiling provides a compromise between the single shared file approach that instigates the lock contention problems on parallel file systems and having one file per process, which results in generating a massive and unmanageable number of files. In this paper, we evaluate and tune the performance of recently implemented subfiling feature in HDF5. In specific, we explain the implementation strategy of subfiling feature in HDF5, provide examples of using the feature, and evaluate and tune parallel I/O performance of this feature with parallel file systems of the Cray XC40 system at NERSC (Cori) that include a burst buffer storage and a Lustre disk-based storage. We also evaluate I/O performance on the Cray XC30 system, Edison, at NERSC. Our results show performance benefits of 1.2X to 6X performance advantage with subfiling compared to writing a single shared HDF5 file. We present our exploration of configurations, such as the number of subfiles and the number of Lustre storage targets to storing files, as optimization parameters to obtain superior I/O performance. Based on this exploration, we discuss recommendations for achieving good I/O performance as well as limitations with using the subfiling feature.

  4. Distributed and parallel approach for handle and perform huge datasets

    Science.gov (United States)

    Konopko, Joanna

    2015-12-01

    Big Data refers to the dynamic, large and disparate volumes of data comes from many different sources (tools, machines, sensors, mobile devices) uncorrelated with each others. It requires new, innovative and scalable technology to collect, host and analytically process the vast amount of data. Proper architecture of the system that perform huge data sets is needed. In this paper, the comparison of distributed and parallel system architecture is presented on the example of MapReduce (MR) Hadoop platform and parallel database platform (DBMS). This paper also analyzes the problem of performing and handling valuable information from petabytes of data. The both paradigms: MapReduce and parallel DBMS are described and compared. The hybrid architecture approach is also proposed and could be used to solve the analyzed problem of storing and processing Big Data.

  5. Routing performance analysis and optimization within a massively parallel computer

    Science.gov (United States)

    Archer, Charles Jens; Peters, Amanda; Pinnow, Kurt Walter; Swartz, Brent Allen

    2013-04-16

    An apparatus, program product and method optimize the operation of a massively parallel computer system by, in part, receiving actual performance data concerning an application executed by the plurality of interconnected nodes, and analyzing the actual performance data to identify an actual performance pattern. A desired performance pattern may be determined for the application, and an algorithm may be selected from among a plurality of algorithms stored within a memory, the algorithm being configured to achieve the desired performance pattern based on the actual performance data.

  6. Performance modeling of parallel algorithms for solving neutron diffusion problems

    International Nuclear Information System (INIS)

    Azmy, Y.Y.; Kirk, B.L.

    1995-01-01

    Neutron diffusion calculations are the most common computational methods used in the design, analysis, and operation of nuclear reactors and related activities. Here, mathematical performance models are developed for the parallel algorithm used to solve the neutron diffusion equation on message passing and shared memory multiprocessors represented by the Intel iPSC/860 and the Sequent Balance 8000, respectively. The performance models are validated through several test problems, and these models are used to estimate the performance of each of the two considered architectures in situations typical of practical applications, such as fine meshes and a large number of participating processors. While message passing computers are capable of producing speedup, the parallel efficiency deteriorates rapidly as the number of processors increases. Furthermore, the speedup fails to improve appreciably for massively parallel computers so that only small- to medium-sized message passing multiprocessors offer a reasonable platform for this algorithm. In contrast, the performance model for the shared memory architecture predicts very high efficiency over a wide range of number of processors reasonable for this architecture. Furthermore, the model efficiency of the Sequent remains superior to that of the hypercube if its model parameters are adjusted to make its processors as fast as those of the iPSC/860. It is concluded that shared memory computers are better suited for this parallel algorithm than message passing computers

  7. Kinematic Analysis and Performance Evaluation of Novel PRS Parallel Mechanism

    Science.gov (United States)

    Balaji, K.; Khan, B. Shahul Hamid

    2018-02-01

    In this paper, a 3 DoF (Degree of Freedom) novel PRS (Prismatic-Revolute- Spherical) type parallel mechanisms has been designed and presented. The combination of striaght and arc type linkages for 3 DOF parallel mechanism is introduced for the first time. The performances of the mechanisms are evaluated based on the indices such as Minimum Singular Value (MSV), Condition Number (CN), Local Conditioning Index (LCI), Kinematic Configuration Index (KCI) and Global Conditioning Index (GCI). The overall reachable workspace of all mechanisms are presented. The kinematic measure, dexterity measure and workspace analysis for all the mechanism have been evaluated and compared.

  8. Design of high-performance parallelized gene predictors in MATLAB.

    Science.gov (United States)

    Rivard, Sylvain Robert; Mailloux, Jean-Gabriel; Beguenane, Rachid; Bui, Hung Tien

    2012-04-10

    This paper proposes a method of implementing parallel gene prediction algorithms in MATLAB. The proposed designs are based on either Goertzel's algorithm or on FFTs and have been implemented using varying amounts of parallelism on a central processing unit (CPU) and on a graphics processing unit (GPU). Results show that an implementation using a straightforward approach can require over 4.5 h to process 15 million base pairs (bps) whereas a properly designed one could perform the same task in less than five minutes. In the best case, a GPU implementation can yield these results in 57 s. The present work shows how parallelism can be used in MATLAB for gene prediction in very large DNA sequences to produce results that are over 270 times faster than a conventional approach. This is significant as MATLAB is typically overlooked due to its apparent slow processing time even though it offers a convenient environment for bioinformatics. From a practical standpoint, this work proposes two strategies for accelerating genome data processing which rely on different parallelization mechanisms. Using a CPU, the work shows that direct access to the MEX function increases execution speed and that the PARFOR construct should be used in order to take full advantage of the parallelizable Goertzel implementation. When the target is a GPU, the work shows that data needs to be segmented into manageable sizes within the GFOR construct before processing in order to minimize execution time.

  9. Parallel file system performances in fusion data storage

    International Nuclear Information System (INIS)

    Iannone, F.; Podda, S.; Bracco, G.; Manduchi, G.; Maslennikov, A.; Migliori, S.; Wolkersdorfer, K.

    2012-01-01

    High I/O flow rates, up to 10 GB/s, are required in large fusion Tokamak experiments like ITER where hundreds of nodes store simultaneously large amounts of data acquired during the plasma discharges. Typical network topologies such as linear arrays (systolic), rings, meshes (2-D arrays), tori (3-D arrays), trees, butterfly, hypercube in combination with high speed data transports like Infiniband or 10G-Ethernet, are the main areas in which the effort to overcome the so-called parallel I/O bottlenecks is most focused. The high I/O flow rates were modelled in an emulated testbed based on the parallel file systems such as Lustre and GPFS, commonly used in High Performance Computing. The test runs on High Performance Computing–For Fusion (8640 cores) and ENEA CRESCO (3392 cores) supercomputers. Message Passing Interface based applications were developed to emulate parallel I/O on Lustre and GPFS using data archival and access solutions like MDSPLUS and Universal Access Layer. These methods of data storage organization are widely diffused in nuclear fusion experiments and are being developed within the EFDA Integrated Tokamak Modelling – Task Force; the authors tried to evaluate their behaviour in a realistic emulation setup.

  10. Parallel file system performances in fusion data storage

    Energy Technology Data Exchange (ETDEWEB)

    Iannone, F., E-mail: francesco.iannone@enea.it [Associazione EURATOM-ENEA sulla Fusione, C.R.ENEA Frascati, via E.Fermi, 45 - 00044 Frascati, Rome (Italy); Podda, S.; Bracco, G. [ENEA Information Communication Tecnologies, Lungotevere Thaon di Revel, 76 - 00196 Rome (Italy); Manduchi, G. [Associazione EURATOM-ENEA sulla Fusione, Consorzio RFX, Corso Stati Uniti, 4 - 35127 Padua (Italy); Maslennikov, A. [CASPUR Inter-University Consortium for the Application of Super-Computing for Research, via dei Tizii, 6b - 00185 Rome (Italy); Migliori, S. [ENEA Information Communication Tecnologies, Lungotevere Thaon di Revel, 76 - 00196 Rome (Italy); Wolkersdorfer, K. [Juelich Supercomputing Centre-FZJ, D-52425 Juelich (Germany)

    2012-12-15

    High I/O flow rates, up to 10 GB/s, are required in large fusion Tokamak experiments like ITER where hundreds of nodes store simultaneously large amounts of data acquired during the plasma discharges. Typical network topologies such as linear arrays (systolic), rings, meshes (2-D arrays), tori (3-D arrays), trees, butterfly, hypercube in combination with high speed data transports like Infiniband or 10G-Ethernet, are the main areas in which the effort to overcome the so-called parallel I/O bottlenecks is most focused. The high I/O flow rates were modelled in an emulated testbed based on the parallel file systems such as Lustre and GPFS, commonly used in High Performance Computing. The test runs on High Performance Computing-For Fusion (8640 cores) and ENEA CRESCO (3392 cores) supercomputers. Message Passing Interface based applications were developed to emulate parallel I/O on Lustre and GPFS using data archival and access solutions like MDSPLUS and Universal Access Layer. These methods of data storage organization are widely diffused in nuclear fusion experiments and are being developed within the EFDA Integrated Tokamak Modelling - Task Force; the authors tried to evaluate their behaviour in a realistic emulation setup.

  11. Automatic performance tuning of parallel and accelerated seismic imaging kernels

    KAUST Repository

    Haberdar, Hakan

    2014-01-01

    With the increased complexity and diversity of mainstream high performance computing systems, significant effort is required to tune parallel applications in order to achieve the best possible performance for each particular platform. This task becomes more and more challenging and requiring a larger set of skills. Automatic performance tuning is becoming a must for optimizing applications such as Reverse Time Migration (RTM) widely used in seismic imaging for oil and gas exploration. An empirical search based auto-tuning approach is applied to the MPI communication operations of the parallel isotropic and tilted transverse isotropic kernels. The application of auto-tuning using the Abstract Data and Communication Library improved the performance of the MPI communications as well as developer productivity by providing a higher level of abstraction. Keeping productivity in mind, we opted toward pragma based programming for accelerated computation on latest accelerated architectures such as GPUs using the fairly new OpenACC standard. The same auto-tuning approach is also applied to the OpenACC accelerated seismic code for optimizing the compute intensive kernel of the Reverse Time Migration application. The application of such technique resulted in an improved performance of the original code and its ability to adapt to different execution environments.

  12. HVI Ballistic Performance Characterization of Non-Parallel Walls

    Science.gov (United States)

    Bohl, William; Miller, Joshua; Christiansen, Eric

    2012-01-01

    The Double-Wall, "Whipple" Shield [1] has been the subject of many hypervelocity impact studies and has proven to be an effective shield system for Micro-Meteoroid and Orbital Debris (MMOD) impacts for spacecraft. The US modules of the International Space Station (ISS), with their "bumper shields" offset from their pressure holding rear walls provide good examples of effective on-orbit use of the double wall shield. The concentric cylinder shield configuration with its large radius of curvature relative to separation distance is easily and effectively represented for testing and analysis as a system of two parallel plates. The parallel plate double wall configuration has been heavily tested and characterized for shield performance for normal and oblique impacts for the ISS and other programs. The double wall shield and principally similar Stuffed Whipple Shield are very common shield types for MMOD protection. However, in some locations with many spacecraft designs, the rear wall cannot be modeled as being parallel or concentric with the outer bumper wall. As represented in Figure 1, there is an included angle between the two walls. And, with a cylindrical outer wall, the effective included angle constantly changes. This complicates assessment of critical spacecraft components located within outer spacecraft walls when using software tools such as NASA's BumperII. In addition, the validity of the risk assessment comes into question when using the standard double wall shield equations, especially since verification testing of every set of double wall included angles is impossible.

  13. Performance of a parallel plate volume calorimeter prototype

    International Nuclear Information System (INIS)

    Arefiev, A.; Bencze, Gy.L.; Bizzeti, A.; Choumilov, E.; Civinini, C; D'Alessandro, R.; Ferrando, A.; Fouz, M.C.; Iglesias, A.; Ivochkin, V.; Josa, M.I.; Malinin, A.; Meschini, M.; Misyura, S.; Pojidaev, V.; Salicio, J.M.; Sikler, F.

    1995-01-01

    An iron/gas parallel plate volume calorimeter prototype, working in the avalanche mode, has been tested using electrons of 20 to 150 GeV/c momentum with high voltages varying from 5400 to 5600 V (electric fields ranging from 36 to 37 KV/cm), and a gas mixture of CF4/CO, (80/20%). The collected charge was measured as a function of the high voltage and of the electron energy. The energy resolution was also measured. Comparisons are made with Monte-Carlo predictions. Agreement between data and simulation allows the calculation of the expected performance of a full size calorimeter. (Author)

  14. Performance of a parallel plate volume calorimeter prototype

    International Nuclear Information System (INIS)

    Arefiev, A.; Bencze, G.L.; Bizzeti, A.

    1995-09-01

    An iron/gas parallel plate volume calorimeter prototype, working in the avalanche mode, has been tested using electrons of 20 to 150 GeV/c momentum with high voltages varying from 5400 to 5600 V (electric fields ranging from 36 to 37 KV/cm), and a gas mixture of CF 4 /CO 2 (80/20%). The collected charge was measured as a function of the high voltage and of the electron energy. The energy resolution was also measured. Comparisons are made with Monte-Carlo predictions. Agreement between data and simulation allows the calculation of the expected performance of a full size calorimeter

  15. Massively parallel performance of neutron transport response matrix algorithms

    International Nuclear Information System (INIS)

    Hanebutte, U.R.; Lewis, E.E.

    1993-01-01

    Massively parallel red/black response matrix algorithms for the solution of within-group neutron transport problems are implemented on the Connection Machines-2, 200 and 5. The response matrices are dericed from the diamond-differences and linear-linear nodal discrete ordinate and variational nodal P 3 approximations. The unaccelerated performance of the iterative procedure is examined relative to the maximum rated performances of the machines. The effects of processor partitions size, of virtual processor ratio and of problems size are examined in detail. For the red/black algorithm, the ratio of inter-node communication to computing times is found to be quite small, normally of the order of ten percent or less. Performance increases with problems size and with virtual processor ratio, within the memeory per physical processor limitation. Algorithm adaptation to courser grain machines is straight-forward, with total computing time being virtually inversely proportional to the number of physical processors. (orig.)

  16. Enzyme

    Science.gov (United States)

    Enzymes are complex proteins that cause a specific chemical change in all parts of the body. For ... use them. Blood clotting is another example of enzymes at work. Enzymes are needed for all body ...

  17. Overview of Parallel Platforms for Common High Performance Computing

    Directory of Open Access Journals (Sweden)

    T. Fryza

    2012-04-01

    Full Text Available The paper deals with various parallel platforms used for high performance computing in the signal processing domain. More precisely, the methods exploiting the multicores central processing units such as message passing interface and OpenMP are taken into account. The properties of the programming methods are experimentally proved in the application of a fast Fourier transform and a discrete cosine transform and they are compared with the possibilities of MATLAB's built-in functions and Texas Instruments digital signal processors with very long instruction word architectures. New FFT and DCT implementations were proposed and tested. The implementation phase was compared with CPU based computing methods and with possibilities of the Texas Instruments digital signal processing library on C6747 floating-point DSPs. The optimal combination of computing methods in the signal processing domain and new, fast routines' implementation is proposed as well.

  18. Data access performance through parallelization and vectored access. Some results

    International Nuclear Information System (INIS)

    Furano, F; Hanushevsky, A

    2008-01-01

    High Energy Physics data processing and analysis applications typically deal with the problem of accessing and processing data at high speed. Recent studies, development and test work have shown that the latencies due to data access can often be hidden by parallelizing them with the data processing, thus giving the ability to have applications which process remote data with a high level of efficiency. Techniques and algorithms able to reach this result have been implemented in the client side of the Scalla/xrootd system, and in this contribution we describe the results of some tests done in order to compare their performance and characteristics. These techniques, if used together with multiple streams data access, can also be effective in allowing to efficiently and transparently deal with data repositories accessible via a Wide Area Network

  19. Performance evaluation for compressible flow calculations on five parallel computers of different architectures

    International Nuclear Information System (INIS)

    Kimura, Toshiya.

    1997-03-01

    A two-dimensional explicit Euler solver has been implemented for five MIMD parallel computers of different machine architectures in Center for Promotion of Computational Science and Engineering of Japan Atomic Energy Research Institute. These parallel computers are Fujitsu VPP300, NEC SX-4, CRAY T94, IBM SP2, and Hitachi SR2201. The code was parallelized by several parallelization methods, and a typical compressible flow problem has been calculated for different grid sizes changing the number of processors. Their effective performances for parallel calculations, such as calculation speed, speed-up ratio and parallel efficiency, have been investigated and evaluated. The communication time among processors has been also measured and evaluated. As a result, the differences on the performance and the characteristics between vector-parallel and scalar-parallel computers can be pointed, and it will present the basic data for efficient use of parallel computers and for large scale CFD simulations on parallel computers. (author)

  20. Parallel Performance Optimizations on Unstructured Mesh-based Simulations

    Energy Technology Data Exchange (ETDEWEB)

    Sarje, Abhinav; Song, Sukhyun; Jacobsen, Douglas; Huck, Kevin; Hollingsworth, Jeffrey; Malony, Allen; Williams, Samuel; Oliker, Leonid

    2015-01-01

    © The Authors. Published by Elsevier B.V. This paper addresses two key parallelization challenges the unstructured mesh-based ocean modeling code, MPAS-Ocean, which uses a mesh based on Voronoi tessellations: (1) load imbalance across processes, and (2) unstructured data access patterns, that inhibit intra- and inter-node performance. Our work analyzes the load imbalance due to naive partitioning of the mesh, and develops methods to generate mesh partitioning with better load balance and reduced communication. Furthermore, we present methods that minimize both inter- and intranode data movement and maximize data reuse. Our techniques include predictive ordering of data elements for higher cache efficiency, as well as communication reduction approaches. We present detailed performance data when running on thousands of cores using the Cray XC30 supercomputer and show that our optimization strategies can exceed the original performance by over 2×. Additionally, many of these solutions can be broadly applied to a wide variety of unstructured grid-based computations.

  1. Evaluation method for the drying performance of enzyme containing formulations

    DEFF Research Database (Denmark)

    Sloth, Jakob; Bach, P.; Jensen, Anker Degn

    2008-01-01

    A method is presented for fast and cheap evaluation of the performance of enzyme containing formulations in terms of preserving the highest enzyme activity during spray drying. The method is based on modeling the kinetics of the thermal inactivation reaction which occurs during the drying process....... Relevant kinetic parameters are determined from differential scanning calorimeter (DSC) experiments and the model is used to simulate the severity of the inactivation reaction for temperatures and moisture levels relevant for spray drying. After conducting experiments and subsequent simulations...... for a number of different formulations it may be deduced which formulation performs best. This is illustrated by a formulation design study where 4 different enzyme containing formulations are evaluated. The method is validated by comparison to pilot scale spray dryer experiments....

  2. Effects of supplemental microbial phytase enzyme on performance ...

    African Journals Online (AJOL)

    This experiment was conducted to investigate the effects of supplemental phytase in a corn-wheatsoybean meal basal diet on phosphorus (P) digestibility and performance of broiler chicks. 378 one-day old broiler chicks (Ross 308) were allocated to 3×3 factorial arrangements with three levels of phytase enzyme (0, 500 ...

  3. Simple, parallel, high-performance virtual machines for extreme computations

    International Nuclear Information System (INIS)

    Chokoufe Nejad, Bijan; Ohl, Thorsten; Reuter, Jurgen

    2014-11-01

    We introduce a high-performance virtual machine (VM) written in a numerically fast language like Fortran or C to evaluate very large expressions. We discuss the general concept of how to perform computations in terms of a VM and present specifically a VM that is able to compute tree-level cross sections for any number of external legs, given the corresponding byte code from the optimal matrix element generator, O'Mega. Furthermore, this approach allows to formulate the parallel computation of a single phase space point in a simple and obvious way. We analyze hereby the scaling behaviour with multiple threads as well as the benefits and drawbacks that are introduced with this method. Our implementation of a VM can run faster than the corresponding native, compiled code for certain processes and compilers, especially for very high multiplicities, and has in general runtimes in the same order of magnitude. By avoiding the tedious compile and link steps, which may fail for source code files of gigabyte sizes, new processes or complex higher order corrections that are currently out of reach could be evaluated with a VM given enough computing power.

  4. The ongoing investigation of high performance parallel computing in HEP

    CERN Document Server

    Peach, Kenneth J; Böck, R K; Dobinson, Robert W; Hansroul, M; Norton, Alan Robert; Willers, Ian Malcolm; Baud, J P; Carminati, F; Gagliardi, F; McIntosh, E; Metcalf, M; Robertson, L; CERN. Geneva. Detector Research and Development Committee

    1993-01-01

    Past and current exploitation of parallel computing in High Energy Physics is summarized and a list of R & D projects in this area is presented. The applicability of new parallel hardware and software to physics problems is investigated, in the light of the requirements for computing power of LHC experiments and the current trends in the computer industry. Four main themes are discussed (possibilities for a finer grain of parallelism; fine-grain communication mechanism; usable parallel programming environment; different programming models and architectures, using standard commercial products). Parallel computing technology is potentially of interest for offline and vital for real time applications in LHC. A substantial investment in applications development and evaluation of state of the art hardware and software products is needed. A solid development environment is required at an early stage, before mainline LHC program development begins.

  5. Language interoperability for high-performance parallel scientific components

    International Nuclear Information System (INIS)

    Elliot, N; Kohn, S; Smolinski, B

    1999-01-01

    With the increasing complexity and interdisciplinary nature of scientific applications, code reuse is becoming increasingly important in scientific computing. One method for facilitating code reuse is the use of components technologies, which have been used widely in industry. However, components have only recently worked their way into scientific computing. Language interoperability is an important underlying technology for these component architectures. In this paper, we present an approach to language interoperability for a high-performance parallel, component architecture being developed by the Common Component Architecture (CCA) group. Our approach is based on Interface Definition Language (IDL) techniques. We have developed a Scientific Interface Definition Language (SIDL), as well as bindings to C and Fortran. We have also developed a SIDL compiler and run-time library support for reference counting, reflection, object management, and exception handling (Babel). Results from using Babel to call a standard numerical solver library (written in C) from C and Fortran show that the cost of using Babel is minimal, where as the savings in development time and the benefits of object-oriented development support for C and Fortran far outweigh the costs

  6. Performance of broilers fed enzyme-supplemented tigernut ...

    African Journals Online (AJOL)

    A feeding trial was set up to study the effects of replacing maize with tigernut meal (TGN) at 0, 33.33, 66.67 and 100 per cent levels, with 0.10 per cent enzyme supplementation of all levels, on performance characteristics and carcass yield in broiler chicken for 8 weeks (56 days). A total of 200 Anak-2000 breed of broilers ...

  7. Effect of Barley and Enzyme on Performance, Carcass, Enzyme Activity and Digestion Parameters of Broilers

    Directory of Open Access Journals (Sweden)

    majid kalantar

    2016-04-01

    Ross-308 broiler chickens were allocated randomly to 3 treatments with 5 replicates using a CRD statistical design. Treatments were included control, barley and barley+ enzyme. The experimental diets were formulated to have similar contents of crude protein, metabolizable energy, total non-starch polysaccharides (NSP. Results and Discussion According to the results, effect of barley with or without enzyme on growth performance at starter, grower and the entire period and also on carcass characteristics, pancreas enzyme activity and measures of ileal acidity and viscosity at the age of 42 were significant (P

  8. Performing a local reduction operation on a parallel computer

    Science.gov (United States)

    Blocksome, Michael A.; Faraj, Daniel A.

    2012-12-11

    A parallel computer including compute nodes, each including two reduction processing cores, a network write processing core, and a network read processing core, each processing core assigned an input buffer. Copying, in interleaved chunks by the reduction processing cores, contents of the reduction processing cores' input buffers to an interleaved buffer in shared memory; copying, by one of the reduction processing cores, contents of the network write processing core's input buffer to shared memory; copying, by another of the reduction processing cores, contents of the network read processing core's input buffer to shared memory; and locally reducing in parallel by the reduction processing cores: the contents of the reduction processing core's input buffer; every other interleaved chunk of the interleaved buffer; the copied contents of the network write processing core's input buffer; and the copied contents of the network read processing core's input buffer.

  9. High-Performance Psychometrics: The Parallel-E Parallel-M Algorithm for Generalized Latent Variable Models. Research Report. ETS RR-16-34

    Science.gov (United States)

    von Davier, Matthias

    2016-01-01

    This report presents results on a parallel implementation of the expectation-maximization (EM) algorithm for multidimensional latent variable models. The developments presented here are based on code that parallelizes both the E step and the M step of the parallel-E parallel-M algorithm. Examples presented in this report include item response…

  10. Integration experiences and performance studies of A COTS parallel archive systems

    Energy Technology Data Exchange (ETDEWEB)

    Chen, Hsing-bung [Los Alamos National Laboratory; Scott, Cody [Los Alamos National Laboratory; Grider, Bary [Los Alamos National Laboratory; Torres, Aaron [Los Alamos National Laboratory; Turley, Milton [Los Alamos National Laboratory; Sanchez, Kathy [Los Alamos National Laboratory; Bremer, John [Los Alamos National Laboratory

    2010-01-01

    Current and future Archive Storage Systems have been asked to (a) scale to very high bandwidths, (b) scale in metadata performance, (c) support policy-based hierarchical storage management capability, (d) scale in supporting changing needs of very large data sets, (e) support standard interface, and (f) utilize commercial-off-the-shelf(COTS) hardware. Parallel file systems have been asked to do the same thing but at one or more orders of magnitude faster in performance. Archive systems continue to move closer to file systems in their design due to the need for speed and bandwidth, especially metadata searching speeds such as more caching and less robust semantics. Currently the number of extreme highly scalable parallel archive solutions is very small especially those that will move a single large striped parallel disk file onto many tapes in parallel. We believe that a hybrid storage approach of using COTS components and innovative software technology can bring new capabilities into a production environment for the HPC community much faster than the approach of creating and maintaining a complete end-to-end unique parallel archive software solution. In this paper, we relay our experience of integrating a global parallel file system and a standard backup/archive product with a very small amount of additional code to provide a scalable, parallel archive. Our solution has a high degree of overlap with current parallel archive products including (a) doing parallel movement to/from tape for a single large parallel file, (b) hierarchical storage management, (c) ILM features, (d) high volume (non-single parallel file) archives for backup/archive/content management, and (e) leveraging all free file movement tools in Linux such as copy, move, ls, tar, etc. We have successfully applied our working COTS Parallel Archive System to the current world's first petaflop/s computing system, LANL's Roadrunner, and demonstrated its capability to address requirements of

  11. Integration experiments and performance studies of a COTS parallel archive system

    Energy Technology Data Exchange (ETDEWEB)

    Chen, Hsing-bung [Los Alamos National Laboratory; Scott, Cody [Los Alamos National Laboratory; Grider, Gary [Los Alamos National Laboratory; Torres, Aaron [Los Alamos National Laboratory; Turley, Milton [Los Alamos National Laboratory; Sanchez, Kathy [Los Alamos National Laboratory; Bremer, John [Los Alamos National Laboratory

    2010-06-16

    Current and future Archive Storage Systems have been asked to (a) scale to very high bandwidths, (b) scale in metadata performance, (c) support policy-based hierarchical storage management capability, (d) scale in supporting changing needs of very large data sets, (e) support standard interface, and (f) utilize commercial-off-the-shelf (COTS) hardware. Parallel file systems have been asked to do the same thing but at one or more orders of magnitude faster in performance. Archive systems continue to move closer to file systems in their design due to the need for speed and bandwidth, especially metadata searching speeds such as more caching and less robust semantics. Currently the number of extreme highly scalable parallel archive solutions is very small especially those that will move a single large striped parallel disk file onto many tapes in parallel. We believe that a hybrid storage approach of using COTS components and innovative software technology can bring new capabilities into a production environment for the HPC community much faster than the approach of creating and maintaining a complete end-to-end unique parallel archive software solution. In this paper, we relay our experience of integrating a global parallel file system and a standard backup/archive product with a very small amount of additional code to provide a scalable, parallel archive. Our solution has a high degree of overlap with current parallel archive products including (a) doing parallel movement to/from tape for a single large parallel file, (b) hierarchical storage management, (c) ILM features, (d) high volume (non-single parallel file) archives for backup/archive/content management, and (e) leveraging all free file movement tools in Linux such as copy, move, Is, tar, etc. We have successfully applied our working COTS Parallel Archive System to the current world's first petafiop/s computing system, LANL's Roadrunner machine, and demonstrated its capability to address

  12. Performance analysis of job scheduling policies in parallel supercomputing environments

    Energy Technology Data Exchange (ETDEWEB)

    Naik, V.K.; Squillante, M.S. [IBM T.J. Watson Research Center, Yorktown Heights, NY (United States); Setia, S.K. [George Mason Univ., Fairfax, VA (United States). Dept. of Computer Science

    1993-12-31

    In this paper the authors analyze three general classes of scheduling policies under a workload typical of largescale scientific computing. These policies differ in the manner in which processors are partitioned among the jobs as well as the way in which jobs are prioritized for execution on the partitions. Their results indicate that existing static schemes do not perform well under varying workloads. Adaptive policies tend to make better scheduling decisions, but their ability to adjust to workload changes is limited. Dynamic partitioning policies, on the other hand, yield the best performance and can be tuned to provide desired performance differences among jobs with varying resource demands.

  13. Automatic performance tuning of parallel and accelerated seismic imaging kernels

    KAUST Repository

    Haberdar, Hakan; Siddiqui, Shahzeb; Feki, Saber

    2014-01-01

    the performance of the MPI communications as well as developer productivity by providing a higher level of abstraction. Keeping productivity in mind, we opted toward pragma based programming for accelerated computation on latest accelerated architectures

  14. Performance Analysis of a Threshold-Based Parallel Multiple Beam Selection Scheme for WDM FSO Systems

    KAUST Repository

    Nam, Sung Sik; Alouini, Mohamed-Slim; Ko, Young-Chai

    2018-01-01

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme for a free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has occurred

  15. Optimized Parallel Discrete Event Simulation (PDES) for High Performance Computing (HPC) Clusters

    National Research Council Canada - National Science Library

    Abu-Ghazaleh, Nael

    2005-01-01

    The aim of this project was to study the communication subsystem performance of state of the art optimistic simulator Synchronous Parallel Environment for Emulation and Discrete-Event Simulation (SPEEDES...

  16. Performance of a Sequential and Parallel Computational Fluid Dynamic (CFD) Solver on a Missile Body Configuration

    National Research Council Canada - National Science Library

    Hisley, Dixie

    1999-01-01

    .... The goals of this report are: (1) to investigate the performance of message passing and loop level parallelization techniques, as they were implemented in the computational fluid dynamics (CFD...

  17. A High-Performance Parallel FDTD Method Enhanced by Using SSE Instruction Set

    Directory of Open Access Journals (Sweden)

    Dau-Chyrh Chang

    2012-01-01

    Full Text Available We introduce a hardware acceleration technique for the parallel finite difference time domain (FDTD method using the SSE (streaming (single instruction multiple data SIMD extensions instruction set. The implementation of SSE instruction set to parallel FDTD method has achieved the significant improvement on the simulation performance. The benchmarks of the SSE acceleration on both the multi-CPU workstation and computer cluster have demonstrated the advantages of (vector arithmetic logic unit VALU acceleration over GPU acceleration. Several engineering applications are employed to demonstrate the performance of parallel FDTD method enhanced by SSE instruction set.

  18. Parallel performance of TORT on the CRAY J90: Model and measurement

    International Nuclear Information System (INIS)

    Barnett, A.; Azmy, Y.Y.

    1997-10-01

    A limitation on the parallel performance of TORT on the CRAY J90 is the amount of extra work introduced by the multitasking algorithm itself. The extra work beyond that of the serial version of the code, called overhead, arises from the synchronization of the parallel tasks and the accumulation of results by the master task. The goal of recent updates to TORT was to reduce the time consumed by these activities. To help understand which components of the multitasking algorithm contribute significantly to the overhead, a parallel performance model was constructed and compared to measurements of actual timings of the code

  19. Analysis for Parallel Execution without Performing Hardware/Software Co-simulation

    OpenAIRE

    Muhammad Rashid

    2014-01-01

    Hardware/software co-simulation improves the performance of embedded applications by executing the applications on a virtual platform before the actual hardware is available in silicon. However, the virtual platform of the target architecture is often not available during early stages of the embedded design flow. Consequently, analysis for parallel execution without performing hardware/software co-simulation is required. This article presents an analysis methodology for parallel execution of ...

  20. A Tool for Performance Modeling of Parallel Programs

    Directory of Open Access Journals (Sweden)

    J.A. González

    2003-01-01

    Full Text Available Current performance prediction analytical models try to characterize the performance behavior of actual machines through a small set of parameters. In practice, substantial deviations are observed. These differences are due to factors as memory hierarchies or network latency. A natural approach is to associate a different proportionality constant with each basic block, and analogously, to associate different latencies and bandwidths with each "communication block". Unfortunately, to use this approach implies that the evaluation of parameters must be done for each algorithm. This is a heavy task, implying experiment design, timing, statistics, pattern recognition and multi-parameter fitting algorithms. Software support is required. We present a compiler that takes as source a C program annotated with complexity formulas and produces as output an instrumented code. The trace files obtained from the execution of the resulting code are analyzed with an interactive interpreter, giving us, among other information, the values of those parameters.

  1. NREL Discovers Enzyme Domains that Dramatically Improve Performance | News

    Science.gov (United States)

    of genomics data to find better enzymes, based on their genetic sequence alone. "In 10 years, it on these enzymes that can be targeted via genetic engineering to help break down cellulose faster to Decker, "At the time, tools for genetic engineering in Trichoderma were very limited, but we

  2. Performance study of a cluster calculation; parallelization and application under geant4

    International Nuclear Information System (INIS)

    Trabelsi, Abir

    2007-01-01

    This work concretizes the final studies project for engineering computer sciences, it is archived within the national center of nuclear sciences and technology. The project consists in studying the performance of a set of machines in order to determine the best architecture to assemble them in a cluster. As well as the parallelism and the parallel implementation of GEANT4, as a tool of simulation. The realisation of this project consists on : 1) programming with C++ and executing the two benchmarks P MV and PMM on each station; 2) Interpreting this result in order to show the best architecture of the cluster; 3) parallelism with TOP-C the two benchmarks; 4) Executing the two Top-C versions on the cluster; 5) Generalizing this results; 6)parallelism et executing the parallel version of GEANT4. (Author). 14 refs

  3. On the Performance of the Python Programming Language for Serial and Parallel Scientific Computations

    Directory of Open Access Journals (Sweden)

    Xing Cai

    2005-01-01

    Full Text Available This article addresses the performance of scientific applications that use the Python programming language. First, we investigate several techniques for improving the computational efficiency of serial Python codes. Then, we discuss the basic programming techniques in Python for parallelizing serial scientific applications. It is shown that an efficient implementation of the array-related operations is essential for achieving good parallel performance, as for the serial case. Once the array-related operations are efficiently implemented, probably using a mixed-language implementation, good serial and parallel performance become achievable. This is confirmed by a set of numerical experiments. Python is also shown to be well suited for writing high-level parallel programs.

  4. Performance assessment of the SIMFAP parallel cluster at IFIN-HH Bucharest

    International Nuclear Information System (INIS)

    Adam, Gh.; Adam, S.; Ayriyan, A.; Dushanov, E.; Hayryan, E.; Korenkov, V.; Lutsenko, A.; Mitsyn, V.; Sapozhnikova, T.; Sapozhnikov, A; Streltsova, O.; Buzatu, F.; Dulea, M.; Vasile, I.; Sima, A.; Visan, C.; Busa, J.; Pokorny, I.

    2008-01-01

    Performance assessment and case study outputs of the parallel SIMFAP cluster at IFIN-HH Bucharest point to its effective and reliable operation. A comparison with results on the supercomputing system in LIT-JINR Dubna adds insight on resource allocation for problem solving by parallel computing. The solution of models asking for very large numbers of knots in the discretization mesh needs the migration to high performance computing based on parallel cluster architectures. The acquisition of ready-to-use parallel computing facilities being beyond limited budgetary resources, the solution at IFIN-HH was to buy the hardware and the inter-processor network, and to implement by own efforts the open software concerning both the operating system and the parallel computing standard. The present paper provides a report demonstrating the successful solution of these tasks. The implementation of the well-known HPL (High Performance LINPACK) Benchmark points to the effective and reliable operation of the cluster. The comparison of HPL outputs obtained on parallel clusters of different magnitudes shows that there is an optimum range of the order N of the linear algebraic system over which a given parallel cluster provides optimum parallel solutions. For the SIMFAP cluster, this range can be inferred to correspond to about 1 to 2 x 10 4 linear algebraic equations. For an algorithm of polynomial complexity N α the task sharing among p processors within a parallel solution mainly follows an (N/p)α behaviour under peak performance achievement. Thus, while the problem complexity remains the same, a substantial decrease of the coefficient of the leading order of the polynomial complexity is achieved. (authors)

  5. Improving matrix-vector product performance and multi-level preconditioning for the parallel PCG package

    Energy Technology Data Exchange (ETDEWEB)

    McLay, R.T.; Carey, G.F.

    1996-12-31

    In this study we consider parallel solution of sparse linear systems arising from discretized PDE`s. As part of our continuing work on our parallel PCG Solver package, we have made improvements in two areas. The first is improving the performance of the matrix-vector product. Here on regular finite-difference grids, we are able to use the cache memory more efficiently for smaller domains or where there are multiple degrees of freedom. The second problem of interest in the present work is the construction of preconditioners in the context of the parallel PCG solver we are developing. Here the problem is partitioned over a set of processors subdomains and the matrix-vector product for PCG is carried out in parallel for overlapping grid subblocks. For problems of scaled speedup, the actual rate of convergence of the unpreconditioned system deteriorates as the mesh is refined. Multigrid and subdomain strategies provide a logical approach to resolving the problem. We consider the parallel trade-offs between communication and computation and provide a complexity analysis of a representative algorithm. Some preliminary calculations using the parallel package and comparisons with other preconditioners are provided together with parallel performance results.

  6. High performance parallel computers for science: New developments at the Fermilab advanced computer program

    International Nuclear Information System (INIS)

    Nash, T.; Areti, H.; Atac, R.

    1988-08-01

    Fermilab's Advanced Computer Program (ACP) has been developing highly cost effective, yet practical, parallel computers for high energy physics since 1984. The ACP's latest developments are proceeding in two directions. A Second Generation ACP Multiprocessor System for experiments will include $3500 RISC processors each with performance over 15 VAX MIPS. To support such high performance, the new system allows parallel I/O, parallel interprocess communication, and parallel host processes. The ACP Multi-Array Processor, has been developed for theoretical physics. Each $4000 node is a FORTRAN or C programmable pipelined 20 MFlops (peak), 10 MByte single board computer. These are plugged into a 16 port crossbar switch crate which handles both inter and intra crate communication. The crates are connected in a hypercube. Site oriented applications like lattice gauge theory are supported by system software called CANOPY, which makes the hardware virtually transparent to users. A 256 node, 5 GFlop, system is under construction. 10 refs., 7 figs

  7. 10th International Workshop on Parallel Tools for High Performance Computing

    CERN Document Server

    Gracia, José; Hilbrich, Tobias; Knüpfer, Andreas; Resch, Michael; Nagel, Wolfgang

    2017-01-01

    This book presents the proceedings of the 10th International Parallel Tools Workshop, held October 4-5, 2016 in Stuttgart, Germany – a forum to discuss the latest advances in parallel tools. High-performance computing plays an increasingly important role for numerical simulation and modelling in academic and industrial research. At the same time, using large-scale parallel systems efficiently is becoming more difficult. A number of tools addressing parallel program development and analysis have emerged from the high-performance computing community over the last decade, and what may have started as collection of small helper script has now matured to production-grade frameworks. Powerful user interfaces and an extensive body of documentation allow easy usage by non-specialists.

  8. Linking Hydrolysis Performance to Trichoderma reesei Cellulolytic Enzyme Profile

    DEFF Research Database (Denmark)

    Lehmann, Linda Olkjær; Petersen, Nanna; I. Jørgensen, Christian

    2016-01-01

    Trichoderma reesei expresses a large number of enzymes involved in lignocellulose hydrolysis and the mechanism of how these enzymes work together is too complex to study by traditional methods, e.g. by spiking with single enzymes and monitoring hydrolysis performance. In this study a multivariate...... approach, partial least squares regression, was used to see if it could help explain the correlation between enzyme profile and hydrolysis performance. Diverse enzyme mixtures were produced by Trichoderma reesei Rut-C30 by exploiting various fermentation conditions and used for hydrolysis of washed...

  9. PERFORMANCE EVALUATION OF OR1200 PROCESSOR WITH EVOLUTIONARY PARALLEL HPRC USING GEP

    Directory of Open Access Journals (Sweden)

    R. Maheswari

    2012-04-01

    Full Text Available In this fast computing era, most of the embedded system requires more computing power to complete the complex function/ task at the lesser amount of time. One way to achieve this is by boosting up the processor performance which allows processor core to run faster. This paper presents a novel technique of increasing the performance by parallel HPRC (High Performance Reconfigurable Computing in the CPU/DSP (Digital Signal Processor unit of OR1200 (Open Reduced Instruction Set Computer (RISC 1200 using Gene Expression Programming (GEP an evolutionary programming model. OR1200 is a soft-core RISC processor of the Intellectual Property cores that can efficiently run any modern operating system. In the manufacturing process of OR1200 a parallel HPRC is placed internally in the Integer Execution Pipeline unit of the CPU/DSP core to increase the performance. The GEP Parallel HPRC is activated /deactivated by triggering the signals i HPRC_Gene_Start ii HPRC_Gene_End. A Verilog HDL(Hardware Description language functional code for Gene Expression Programming parallel HPRC is developed and synthesised using XILINX ISE in the former part of the work and a CoreMark processor core benchmark is used to test the performance of the OR1200 soft core in the later part of the work. The result of the implementation ensures the overall speed-up increased to 20.59% by GEP based parallel HPRC in the execution unit of OR1200.

  10. The Parallel Algorithm Based on Genetic Algorithm for Improving the Performance of Cognitive Radio

    Directory of Open Access Journals (Sweden)

    Liu Miao

    2018-01-01

    Full Text Available The intercarrier interference (ICI problem of cognitive radio (CR is severe. In this paper, the machine learning algorithm is used to obtain the optimal interference subcarriers of an unlicensed user (un-LU. Masking the optimal interference subcarriers can suppress the ICI of CR. Moreover, the parallel ICI suppression algorithm is designed to improve the calculation speed and meet the practical requirement of CR. Simulation results show that the data transmission rate threshold of un-LU can be set, the data transmission quality of un-LU can be ensured, the ICI of a licensed user (LU is suppressed, and the bit error rate (BER performance of LU is improved by implementing the parallel suppression algorithm. The ICI problem of CR is solved well by the new machine learning algorithm. The computing performance of the algorithm is improved by designing a new parallel structure and the communication performance of CR is enhanced.

  11. 7th International Workshop on Parallel Tools for High Performance Computing

    CERN Document Server

    Gracia, José; Nagel, Wolfgang; Resch, Michael

    2014-01-01

    Current advances in High Performance Computing (HPC) increasingly impact efficient software development workflows. Programmers for HPC applications need to consider trends such as increased core counts, multiple levels of parallelism, reduced memory per core, and I/O system challenges in order to derive well performing and highly scalable codes. At the same time, the increasing complexity adds further sources of program defects. While novel programming paradigms and advanced system libraries provide solutions for some of these challenges, appropriate supporting tools are indispensable. Such tools aid application developers in debugging, performance analysis, or code optimization and therefore make a major contribution to the development of robust and efficient parallel software. This book introduces a selection of the tools presented and discussed at the 7th International Parallel Tools Workshop, held in Dresden, Germany, September 3-4, 2013.  

  12. Impact of temperature on performance of series and parallel connected mono-crystalline silicon solar cells

    Directory of Open Access Journals (Sweden)

    Subhash Chander

    2015-11-01

    Full Text Available This paper presents a study on impact of temperature on the performance of series and parallel connected mono-crystalline silicon (mono-Si solar cell employing solar simulator. The experiment was carried out at constant light intensity 550 W/m2with cell temperature in the range 25–60 oC for single, series and parallel connected mono-Si solar cells. The performance parameters like open circuit voltage, maximum power, fill factor and efficiency are found to decrease with cell temperature while the short circuit current is observed to increase. The experimental results reveal that silicon solar cells connected in series and parallel combinations follow the Kirchhoff’s laws and the temperature has a significant effect on the performance parameters of solar cell.

  13. Performance of MPI parallel processing implemented by MCNP5/ MCNPX for criticality benchmark problems

    International Nuclear Information System (INIS)

    Mark Dennis Usang; Mohd Hairie Rabir; Mohd Amin Sharifuldin Salleh; Mohamad Puad Abu

    2012-01-01

    MPI parallelism are implemented on a SUN Workstation for running MCNPX and on the High Performance Computing Facility (HPC) for running MCNP5. 23 input less obtained from MCNP Criticality Validation Suite are utilized for the purpose of evaluating the amount of speed up achievable by using the parallel capabilities of MPI. More importantly, we will study the economics of using more processors and the type of problem where the performance gain are obvious. This is important to enable better practices of resource sharing especially for the HPC facilities processing time. Future endeavours in this direction might even reveal clues for best MCNP5/ MCNPX coding practices for optimum performance of MPI parallelisms. (author)

  14. 8th International Workshop on Parallel Tools for High Performance Computing

    CERN Document Server

    Gracia, José; Knüpfer, Andreas; Resch, Michael; Nagel, Wolfgang

    2015-01-01

    Numerical simulation and modelling using High Performance Computing has evolved into an established technique in academic and industrial research. At the same time, the High Performance Computing infrastructure is becoming ever more complex. For instance, most of the current top systems around the world use thousands of nodes in which classical CPUs are combined with accelerator cards in order to enhance their compute power and energy efficiency. This complexity can only be mastered with adequate development and optimization tools. Key topics addressed by these tools include parallelization on heterogeneous systems, performance optimization for CPUs and accelerators, debugging of increasingly complex scientific applications, and optimization of energy usage in the spirit of green IT. This book represents the proceedings of the 8th International Parallel Tools Workshop, held October 1-2, 2014 in Stuttgart, Germany – which is a forum to discuss the latest advancements in the parallel tools.

  15. High performance statistical computing with parallel R: applications to biology and climate modelling

    International Nuclear Information System (INIS)

    Samatova, Nagiza F; Branstetter, Marcia; Ganguly, Auroop R; Hettich, Robert; Khan, Shiraj; Kora, Guruprasad; Li, Jiangtian; Ma, Xiaosong; Pan, Chongle; Shoshani, Arie; Yoginath, Srikanth

    2006-01-01

    Ultrascale computing and high-throughput experimental technologies have enabled the production of scientific data about complex natural phenomena. With this opportunity, comes a new problem - the massive quantities of data so produced. Answers to fundamental questions about the nature of those phenomena remain largely hidden in the produced data. The goal of this work is to provide a scalable high performance statistical data analysis framework to help scientists perform interactive analyses of these raw data to extract knowledge. Towards this goal we have been developing an open source parallel statistical analysis package, called Parallel R, that lets scientists employ a wide range of statistical analysis routines on high performance shared and distributed memory architectures without having to deal with the intricacies of parallelizing these routines

  16. Reproductive performance of female goats fed life-enzyme ...

    African Journals Online (AJOL)

    Direct-fed-microbes (DFM) (life-enzyme) was prepared in a traditional setting using Zymomonas mobilis (bacteria from palm sap) to ferment sawdust. The result revealed an improvement in the nutrient content of the sawdust and its feed values (protein, fibre etc.), and the feed usage efficiency. The reproductive ...

  17. Performance response and egg qualities of laying birds fed enzyme ...

    African Journals Online (AJOL)

    Theperformance response and egg qualities o laying birds fed enzyme supplemented PKC diets asreplacement for maize was investigated wth 210, 20 week old layng pullets of Dominant Black strain at the Teaching and Research Farm of the Delta State University, Asaba Campus, Nigeria. The birds which ust come into ...

  18. Parameters that affect parallel processing for computational electromagnetic simulation codes on high performance computing clusters

    Science.gov (United States)

    Moon, Hongsik

    What is the impact of multicore and associated advanced technologies on computational software for science? Most researchers and students have multicore laptops or desktops for their research and they need computing power to run computational software packages. Computing power was initially derived from Central Processing Unit (CPU) clock speed. That changed when increases in clock speed became constrained by power requirements. Chip manufacturers turned to multicore CPU architectures and associated technological advancements to create the CPUs for the future. Most software applications benefited by the increased computing power the same way that increases in clock speed helped applications run faster. However, for Computational ElectroMagnetics (CEM) software developers, this change was not an obvious benefit - it appeared to be a detriment. Developers were challenged to find a way to correctly utilize the advancements in hardware so that their codes could benefit. The solution was parallelization and this dissertation details the investigation to address these challenges. Prior to multicore CPUs, advanced computer technologies were compared with the performance using benchmark software and the metric was FLoting-point Operations Per Seconds (FLOPS) which indicates system performance for scientific applications that make heavy use of floating-point calculations. Is FLOPS an effective metric for parallelized CEM simulation tools on new multicore system? Parallel CEM software needs to be benchmarked not only by FLOPS but also by the performance of other parameters related to type and utilization of the hardware, such as CPU, Random Access Memory (RAM), hard disk, network, etc. The codes need to be optimized for more than just FLOPs and new parameters must be included in benchmarking. In this dissertation, the parallel CEM software named High Order Basis Based Integral Equation Solver (HOBBIES) is introduced. This code was developed to address the needs of the

  19. CUDA/GPU Technology : Parallel Programming For High Performance Scientific Computing

    OpenAIRE

    YUHENDRA; KUZE, Hiroaki; JOSAPHAT, Tetuko Sri Sumantyo

    2009-01-01

    [ABSTRACT]Graphics processing units (GP Us) originally designed for computer video cards have emerged as the most powerful chip in a high-performance workstation. In the high performance computation capabilities, graphic processing units (GPU) lead to much more powerful performance than conventional CPUs by means of parallel processing. In 2007, the birth of Compute Unified Device Architecture (CUDA) and CUDA-enabled GPUs by NVIDIA Corporation brought a revolution in the general purpose GPU a...

  20. Pthreads vs MPI Parallel Performance of Angular-Domain Decomposed S

    International Nuclear Information System (INIS)

    Azmy, Y.Y.; Barnett, D.A.

    2000-01-01

    Two programming models for parallelizing the Angular Domain Decomposition (ADD) of the discrete ordinates (S n ) approximation of the neutron transport equation are examined. These are the shared memory model based on the POSIX threads (Pthreads) standard, and the message passing model based on the Message Passing Interface (MPI) standard. These standard libraries are available on most multiprocessor platforms thus making the resulting parallel codes widely portable. The question is: on a fixed platform, and for a particular code solving a given test problem, which of the two programming models delivers better parallel performance? Such comparison is possible on Symmetric Multi-Processors (SMP) architectures in which several CPUs physically share a common memory, and in addition are capable of emulating message passing functionality. Implementation of the two-dimensional,(S n ), Arbitrarily High Order Transport (AHOT) code for solving neutron transport problems using these two parallelization models is described. Measured parallel performance of each model on the COMPAQ AlphaServer 8400 and the SGI Origin 2000 platforms is described, and comparison of the observed speedup for the two programming models is reported. For the case presented in this paper it appears that the MPI implementation scales better than the Pthreads implementation on both platforms

  1. PARALLEL IMPLEMENTATION OF CROSS-LAYER OPTIMIZATION - A PERFORMANCE EVALUATION BASED ON SWARM INTELLIGENCE

    Directory of Open Access Journals (Sweden)

    Vanaja Gokul

    2012-01-01

    Full Text Available In distributed systems real time optimizations need to be performed dynamically for better utilization of the network resources. Real time optimizations can be performed effectively by using Cross Layer Optimization (CLO within the network operating system. This paper presents the performance evaluation of Cross Layer Optimization (CLO in comparison with the traditional approach of Single-Layer Optimization (SLO. In the parallel implementation of the approaches the experimental study carried out indicates that the CLO results in a significant improvement in network utilization when compared to SLO. A variant of the Particle Swarm Optimization technique that utilizes Digital Pheromones (PSODP for better performance has been used here. A significantly higher speed up in performance was observed from the parallel implementation of CLO that used PSODP on a cluster of nodes.

  2. Performance of DS-CDMA systems with optimal hard-decision parallel interference cancellation

    NARCIS (Netherlands)

    Hofstad, van der R.W.; Klok, M.J.

    2003-01-01

    We study a multiuser detection system for code-division multiple access (CDMA). We show that applying multistage hard-decision parallel interference cancellation (HD-PIC) significantly improves performance compared to the matched filter system. In (multistage) HD-PIC, estimates of the interfering

  3. Parallel performance of the angular versus spatial domain decomposition for discrete ordinates transport methods

    International Nuclear Information System (INIS)

    Fischer, J.W.; Azmy, Y.Y.

    2003-01-01

    A previously reported parallel performance model for Angular Domain Decomposition (ADD) of the Discrete Ordinates method for solving multidimensional neutron transport problems is revisited for further validation. Three communication schemes: native MPI, the bucket algorithm, and the distributed bucket algorithm, are included in the validation exercise that is successfully conducted on a Beowulf cluster. The parallel performance model is comprised of three components: serial, parallel, and communication. The serial component is largely independent of the number of participating processors, P, while the parallel component decreases like 1/P. These two components are independent of the communication scheme, in contrast with the communication component that typically increases with P in a manner highly dependent on the global reduced algorithm. Correct trends for each component and each communication scheme were measured for the Arbitrarily High Order Transport (AHOT) code, thus validating the performance models. Furthermore, extensive experiments illustrate the superiority of the bucket algorithm. The primary question addressed in this research is: for a given problem size, which domain decomposition method, angular or spatial, is best suited to parallelize Discrete Ordinates methods on a specific computational platform? We address this question for three-dimensional applications via parallel performance models that include parameters specifying the problem size and system performance: the above-mentioned ADD, and a previously constructed and validated Spatial Domain Decomposition (SDD) model. We conclude that for large problems the parallel component dwarfs the communication component even on moderately large numbers of processors. The main advantages of SDD are: (a) scalability to higher numbers of processors of the order of the number of computational cells; (b) smaller memory requirement; (c) better performance than ADD on high-end platforms and large number of

  4. Performance Improvement of Shunt Active Power Filter With Dual Parallel Topology

    DEFF Research Database (Denmark)

    Asiminoaei, Lucian; Lascu, Cristian; Blaabjerg, Frede

    2007-01-01

    loop and the other is in a feedforward loop for harmonic compensation. Thus, both active power filters bring their own characteristic advantages, i.e., the feedback filter improves the steady-state performance of the harmonic mitigation and the feedforward filter improves the dynamic response. Another......This paper describes the control and parallel operation of two active power filters (APFs). Possible parallel operation situations of two APFs are investigated, and then the proposed topology is analyzed. The filters are coupled in a combined topology in which one filter is connected in a feedback...

  5. High performance parallelism pearls 2 multicore and many-core programming approaches

    CERN Document Server

    Jeffers, Jim

    2015-01-01

    High Performance Parallelism Pearls Volume 2 offers another set of examples that demonstrate how to leverage parallelism. Similar to Volume 1, the techniques included here explain how to use processors and coprocessors with the same programming - illustrating the most effective ways to combine Xeon Phi coprocessors with Xeon and other multicore processors. The book includes examples of successful programming efforts, drawn from across industries and domains such as biomed, genetics, finance, manufacturing, imaging, and more. Each chapter in this edited work includes detailed explanations of t

  6. Performance Assessment in a Heat Exchanger Tube with Opposite/Parallel Wing Twisted Tapes

    Directory of Open Access Journals (Sweden)

    S. Eiamsa-ard

    2015-02-01

    Full Text Available The thermohydraulic performance in a tube containing a modified twisted tape with alternate-axes and wing arrangements is reported. This work aims to investigate the effects of wing arrangements (opposite (O and parallel (P wings at different wing shapes (triangle (Tri, rectangular (Rec, and trapezoidal (Tra wings and on the thermohydraulic performance characteristics. The obtained results show that wing twisted tapes with all wing shape arrangements (O-Tri/O-Rec/O-Tra/P-Tri/P-Rec/P-Tra give superior thermohydraulic performance and heat transfer rate to the typical twisted tape. In addition, the tapes with opposite wing arrangement of O-Tra, O-Rec, and O-Tri give superior thermohydraulic performances to those with parallel wing arrangement of P-Tra, P-Rec, and P-Tri around 2.7%, 3.5%, and 3.2%, respectively.

  7. Performing an allreduce operation on a plurality of compute nodes of a parallel computer

    Science.gov (United States)

    Faraj, Ahmad [Rochester, MN

    2012-04-17

    Methods, apparatus, and products are disclosed for performing an allreduce operation on a plurality of compute nodes of a parallel computer. Each compute node includes at least two processing cores. Each processing core has contribution data for the allreduce operation. Performing an allreduce operation on a plurality of compute nodes of a parallel computer includes: establishing one or more logical rings among the compute nodes, each logical ring including at least one processing core from each compute node; performing, for each logical ring, a global allreduce operation using the contribution data for the processing cores included in that logical ring, yielding a global allreduce result for each processing core included in that logical ring; and performing, for each compute node, a local allreduce operation using the global allreduce results for each processing core on that compute node.

  8. Development Of A Parallel Performance Model For The THOR Neutral Particle Transport Code

    Energy Technology Data Exchange (ETDEWEB)

    Yessayan, Raffi; Azmy, Yousry; Schunert, Sebastian

    2017-02-01

    The THOR neutral particle transport code enables simulation of complex geometries for various problems from reactor simulations to nuclear non-proliferation. It is undergoing a thorough V&V requiring computational efficiency. This has motivated various improvements including angular parallelization, outer iteration acceleration, and development of peripheral tools. For guiding future improvements to the code’s efficiency, better characterization of its parallel performance is useful. A parallel performance model (PPM) can be used to evaluate the benefits of modifications and to identify performance bottlenecks. Using INL’s Falcon HPC, the PPM development incorporates an evaluation of network communication behavior over heterogeneous links and a functional characterization of the per-cell/angle/group runtime of each major code component. After evaluating several possible sources of variability, this resulted in a communication model and a parallel portion model. The former’s accuracy is bounded by the variability of communication on Falcon while the latter has an error on the order of 1%.

  9. Kemari: A Portable High Performance Fortran System for Distributed Memory Parallel Processors

    Directory of Open Access Journals (Sweden)

    T. Kamachi

    1997-01-01

    Full Text Available We have developed a compilation system which extends High Performance Fortran (HPF in various aspects. We support the parallelization of well-structured problems with loop distribution and alignment directives similar to HPF's data distribution directives. Such directives give both additional control to the user and simplify the compilation process. For the support of unstructured problems, we provide directives for dynamic data distribution through user-defined mappings. The compiler also allows integration of message-passing interface (MPI primitives. The system is part of a complete programming environment which also comprises a parallel debugger and a performance monitor and analyzer. After an overview of the compiler, we describe the language extensions and related compilation mechanisms in detail. Performance measurements demonstrate the compiler's applicability to a variety of application classes.

  10. The influence of flow maldistribution on the performance of inhomogeneous parallel plate heat exhangers

    DEFF Research Database (Denmark)

    Nielsen, Kaspar Kirstein; Engelbrecht, Kurt; Bahl, Christian R.H.

    2013-01-01

    of 50 random stacks having equal average channel thicknesses with 20 channels each are used to provide a statistical base. The standard deviation of the stacks is varied as are the flow rate (Reynolds number) and the thermal conductivity of the solid heat exchanger material. It is found that the heat...... transfer performance of inhomogeneous stacks of parallel plates may be reduced significantly due to the maldistribution of the fluid flow compared to the ideal homogeneous case. The individual channels experience different flow velocities and this further induces an inter-channel thermal cross talk.......The heat transfer performance of inhomogeneous parallel plate heat exchangers in transient operation is investigated using an established model. A performance parameter, denoted the Nusselt-scaling factor, is used as benchmark and calculated using a well-established single blow technique. A sample...

  11. Computational Performance of a Parallelized Three-Dimensional High-Order Spectral Element Toolbox

    Science.gov (United States)

    Bosshard, Christoph; Bouffanais, Roland; Clémençon, Christian; Deville, Michel O.; Fiétier, Nicolas; Gruber, Ralf; Kehtari, Sohrab; Keller, Vincent; Latt, Jonas

    In this paper, a comprehensive performance review of an MPI-based high-order three-dimensional spectral element method C++ toolbox is presented. The focus is put on the performance evaluation of several aspects with a particular emphasis on the parallel efficiency. The performance evaluation is analyzed with help of a time prediction model based on a parameterization of the application and the hardware resources. A tailor-made CFD computation benchmark case is introduced and used to carry out this review, stressing the particular interest for clusters with up to 8192 cores. Some problems in the parallel implementation have been detected and corrected. The theoretical complexities with respect to the number of elements, to the polynomial degree, and to communication needs are correctly reproduced. It is concluded that this type of code has a nearly perfect speed up on machines with thousands of cores, and is ready to make the step to next-generation petaflop machines.

  12. High-performance parallel approaches for three-dimensional light detection and ranging point clouds gridding

    Science.gov (United States)

    Rizki, Permata Nur Miftahur; Lee, Heezin; Lee, Minsu; Oh, Sangyoon

    2017-01-01

    With the rapid advance of remote sensing technology, the amount of three-dimensional point-cloud data has increased extraordinarily, requiring faster processing in the construction of digital elevation models. There have been several attempts to accelerate the computation using parallel methods; however, little attention has been given to investigating different approaches for selecting the most suited parallel programming model for a given computing environment. We present our findings and insights identified by implementing three popular high-performance parallel approaches (message passing interface, MapReduce, and GPGPU) on time demanding but accurate kriging interpolation. The performances of the approaches are compared by varying the size of the grid and input data. In our empirical experiment, we demonstrate the significant acceleration by all three approaches compared to a C-implemented sequential-processing method. In addition, we also discuss the pros and cons of each method in terms of usability, complexity infrastructure, and platform limitation to give readers a better understanding of utilizing those parallel approaches for gridding purposes.

  13. P3T+: A Performance Estimator for Distributed and Parallel Programs

    Directory of Open Access Journals (Sweden)

    T. Fahringer

    2000-01-01

    Full Text Available Developing distributed and parallel programs on today's multiprocessor architectures is still a challenging task. Particular distressing is the lack of effective performance tools that support the programmer in evaluating changes in code, problem and machine sizes, and target architectures. In this paper we introduce P3T+ which is a performance estimator for mostly regular HPF (High Performance Fortran programs but partially covers also message passing programs (MPI. P3T+ is unique by modeling programs, compiler code transformations, and parallel and distributed architectures. It computes at compile-time a variety of performance parameters including work distribution, number of transfers, amount of data transferred, transfer times, computation times, and number of cache misses. Several novel technologies are employed to compute these parameters: loop iteration spaces, array access patterns, and data distributions are modeled by employing highly effective symbolic analysis. Communication is estimated by simulating the behavior of a communication library used by the underlying compiler. Computation times are predicted through pre-measured kernels on every target architecture of interest. We carefully model most critical architecture specific factors such as cache lines sizes, number of cache lines available, startup times, message transfer time per byte, etc. P3T+ has been implemented and is closely integrated with the Vienna High Performance Compiler (VFC to support programmers develop parallel and distributed applications. Experimental results for realistic kernel codes taken from real-world applications are presented to demonstrate both accuracy and usefulness of P3T+.

  14. Computational performance of a smoothed particle hydrodynamics simulation for shared-memory parallel computing

    Science.gov (United States)

    Nishiura, Daisuke; Furuichi, Mikito; Sakaguchi, Hide

    2015-09-01

    The computational performance of a smoothed particle hydrodynamics (SPH) simulation is investigated for three types of current shared-memory parallel computer devices: many integrated core (MIC) processors, graphics processing units (GPUs), and multi-core CPUs. We are especially interested in efficient shared-memory allocation methods for each chipset, because the efficient data access patterns differ between compute unified device architecture (CUDA) programming for GPUs and OpenMP programming for MIC processors and multi-core CPUs. We first introduce several parallel implementation techniques for the SPH code, and then examine these on our target computer architectures to determine the most effective algorithms for each processor unit. In addition, we evaluate the effective computing performance and power efficiency of the SPH simulation on each architecture, as these are critical metrics for overall performance in a multi-device environment. In our benchmark test, the GPU is found to produce the best arithmetic performance as a standalone device unit, and gives the most efficient power consumption. The multi-core CPU obtains the most effective computing performance. The computational speed of the MIC processor on Xeon Phi approached that of two Xeon CPUs. This indicates that using MICs is an attractive choice for existing SPH codes on multi-core CPUs parallelized by OpenMP, as it gains computational acceleration without the need for significant changes to the source code.

  15. Effect of enzymes on anaerobic digestion of primary sludge and septic tank performance.

    Science.gov (United States)

    Diak, James; Örmeci, Banu; Kennedy, Kevin J

    2012-11-01

    Enzyme additives are believed to improve septic tank performance by increasing the hydrolysis and digestion rates and maintaining a healthy microbial population. Previous studies reported mixed results on the effectiveness of enzymes on mesophilic and thermophilic digestion, and it is not clear whether enzymes would be effective under septic tank conditions where there is no heating or mixing, quantities of enzymes added are small, and they can be washed out quickly. In this study, batch reactors and continuous-flow reactors designed and operated as septic tanks were used to evaluate whether enzymatic treatment would increase the hydrolysis and digestion rates in primary sludge. Total solids, volatile solids, total suspended solids, total and soluble chemical oxygen demand, concentrations of protein, carbohydrate, ammonia and volatile acids in sludge and effluent samples were measured to determine the differences in digestion rates in the presence and absence of enzymes. Overall, no significant improvement was observed in enzyme-treated reactors compared with the control reactors.

  16. 9th International Workshop on Parallel Tools for High Performance Computing

    CERN Document Server

    Hilbrich, Tobias; Niethammer, Christoph; Gracia, José; Nagel, Wolfgang; Resch, Michael

    2016-01-01

    High Performance Computing (HPC) remains a driver that offers huge potentials and benefits for science and society. However, a profound understanding of the computational matters and specialized software is needed to arrive at effective and efficient simulations. Dedicated software tools are important parts of the HPC software landscape, and support application developers. Even though a tool is by definition not a part of an application, but rather a supplemental piece of software, it can make a fundamental difference during the development of an application. Such tools aid application developers in the context of debugging, performance analysis, and code optimization, and therefore make a major contribution to the development of robust and efficient parallel software. This book introduces a selection of the tools presented and discussed at the 9th International Parallel Tools Workshop held in Dresden, Germany, September 2-3, 2015, which offered an established forum for discussing the latest advances in paral...

  17. Progress on H5Part: A Portable High Performance Parallel Data Interface for Electromagnetics Simulations

    International Nuclear Information System (INIS)

    Adelmann, Andreas; Gsell, Achim; Oswald, Benedikt; Schietinger, Thomas; Bethel, Wes; Shalf, John; Siegerist, Cristina; Stockinger, Kurt

    2007-01-01

    Significant problems facing all experimental and computational sciences arise from growing data size and complexity. Common to all these problems is the need to perform efficient data I/O on diverse computer architectures. In our scientific application, the largest parallel particle simulations generate vast quantities of six-dimensional data. Such a simulation run produces data for an aggregate data size up to several TB per run. Motivated by the need to address data I/O and access challenges, we have implemented H5Part, an open source data I/O API that simplifies the use of the Hierarchical Data Format v5 library (HDF5). HDF5 is an industry standard for high performance, cross-platform data storage and retrieval that runs on all contemporary architectures from large parallel supercomputers to laptops. H5Part, which is oriented to the needs of the particle physics and cosmology communities, provides support for parallel storage and retrieval of particles, structured and in the future unstructured meshes. In this paper, we describe recent work focusing on I/O support for particles and structured meshes and provide data showing performance on modern supercomputer architectures like the IBM POWER 5

  18. NETRA: A parallel architecture for integrated vision systems 2: Algorithms and performance evaluation

    Science.gov (United States)

    Choudhary, Alok N.; Patel, Janak H.; Ahuja, Narendra

    1989-01-01

    In part 1 architecture of NETRA is presented. A performance evaluation of NETRA using several common vision algorithms is also presented. Performance of algorithms when they are mapped on one cluster is described. It is shown that SIMD, MIMD, and systolic algorithms can be easily mapped onto processor clusters, and almost linear speedups are possible. For some algorithms, analytical performance results are compared with implementation performance results. It is observed that the analysis is very accurate. Performance analysis of parallel algorithms when mapped across clusters is presented. Mappings across clusters illustrate the importance and use of shared as well as distributed memory in achieving high performance. The parameters for evaluation are derived from the characteristics of the parallel algorithms, and these parameters are used to evaluate the alternative communication strategies in NETRA. Furthermore, the effect of communication interference from other processors in the system on the execution of an algorithm is studied. Using the analysis, performance of many algorithms with different characteristics is presented. It is observed that if communication speeds are matched with the computation speeds, good speedups are possible when algorithms are mapped across clusters.

  19. Effects of exogenous tannase enzyme on growth performance ...

    African Journals Online (AJOL)

    An experiment was conducted to study the effects of dietary addition of tannase to feed of chicks including grape pomace (GP) on growth performance, antioxidant status, immune response, blood parameters, gut morphology, intestinal microflora, liver function, and histopathological responses. The experimental diets were i) ...

  20. Visualizing Network Traffic to Understand the Performance of Massively Parallel Simulations

    KAUST Repository

    Landge, A. G.

    2012-12-01

    The performance of massively parallel applications is often heavily impacted by the cost of communication among compute nodes. However, determining how to best use the network is a formidable task, made challenging by the ever increasing size and complexity of modern supercomputers. This paper applies visualization techniques to aid parallel application developers in understanding the network activity by enabling a detailed exploration of the flow of packets through the hardware interconnect. In order to visualize this large and complex data, we employ two linked views of the hardware network. The first is a 2D view, that represents the network structure as one of several simplified planar projections. This view is designed to allow a user to easily identify trends and patterns in the network traffic. The second is a 3D view that augments the 2D view by preserving the physical network topology and providing a context that is familiar to the application developers. Using the massively parallel multi-physics code pF3D as a case study, we demonstrate that our tool provides valuable insight that we use to explain and optimize pF3D-s performance on an IBM Blue Gene/P system. © 1995-2012 IEEE.

  1. GROMACS: High performance molecular simulations through multi-level parallelism from laptops to supercomputers

    Directory of Open Access Journals (Sweden)

    Mark James Abraham

    2015-09-01

    Full Text Available GROMACS is one of the most widely used open-source and free software codes in chemistry, used primarily for dynamical simulations of biomolecules. It provides a rich set of calculation types, preparation and analysis tools. Several advanced techniques for free-energy calculations are supported. In version 5, it reaches new performance heights, through several new and enhanced parallelization algorithms. These work on every level; SIMD registers inside cores, multithreading, heterogeneous CPU–GPU acceleration, state-of-the-art 3D domain decomposition, and ensemble-level parallelization through built-in replica exchange and the separate Copernicus framework. The latest best-in-class compressed trajectory storage format is supported.

  2. Evaluating the performance of the particle finite element method in parallel architectures

    Science.gov (United States)

    Gimenez, Juan M.; Nigro, Norberto M.; Idelsohn, Sergio R.

    2014-05-01

    This paper presents a high performance implementation for the particle-mesh based method called particle finite element method two (PFEM-2). It consists of a material derivative based formulation of the equations with a hybrid spatial discretization which uses an Eulerian mesh and Lagrangian particles. The main aim of PFEM-2 is to solve transport equations as fast as possible keeping some level of accuracy. The method was found to be competitive with classical Eulerian alternatives for these targets, even in their range of optimal application. To evaluate the goodness of the method with large simulations, it is imperative to use of parallel environments. Parallel strategies for Finite Element Method have been widely studied and many libraries can be used to solve Eulerian stages of PFEM-2. However, Lagrangian stages, such as streamline integration, must be developed considering the parallel strategy selected. The main drawback of PFEM-2 is the large amount of memory needed, which limits its application to large problems with only one computer. Therefore, a distributed-memory implementation is urgently needed. Unlike a shared-memory approach, using domain decomposition the memory is automatically isolated, thus avoiding race conditions; however new issues appear due to data distribution over the processes. Thus, a domain decomposition strategy for both particle and mesh is adopted, which minimizes the communication between processes. Finally, performance analysis running over multicore and multinode architectures are presented. The Courant-Friedrichs-Lewy number used influences the efficiency of the parallelization and, in some cases, a weighted partitioning can be used to improve the speed-up. However the total cputime for cases presented is lower than that obtained when using classical Eulerian strategies.

  3. Enhancing Application Performance Using Mini-Apps: Comparison of Hybrid Parallel Programming Paradigms

    Science.gov (United States)

    Lawson, Gary; Sosonkina, Masha; Baurle, Robert; Hammond, Dana

    2017-01-01

    In many fields, real-world applications for High Performance Computing have already been developed. For these applications to stay up-to-date, new parallel strategies must be explored to yield the best performance; however, restructuring or modifying a real-world application may be daunting depending on the size of the code. In this case, a mini-app may be employed to quickly explore such options without modifying the entire code. In this work, several mini-apps have been created to enhance a real-world application performance, namely the VULCAN code for complex flow analysis developed at the NASA Langley Research Center. These mini-apps explore hybrid parallel programming paradigms with Message Passing Interface (MPI) for distributed memory access and either Shared MPI (SMPI) or OpenMP for shared memory accesses. Performance testing shows that MPI+SMPI yields the best execution performance, while requiring the largest number of code changes. A maximum speedup of 23 was measured for MPI+SMPI, but only 11 was measured for MPI+OpenMP.

  4. In-cylinder diesel spray combustion simulations using parallel computation: A performance benchmarking study

    International Nuclear Information System (INIS)

    Pang, Kar Mun; Ng, Hoon Kiat; Gan, Suyin

    2012-01-01

    Highlights: ► A performance benchmarking exercise is conducted for diesel combustion simulations. ► The reduced chemical mechanism shows its advantages over base and skeletal models. ► High efficiency and great reduction of CPU runtime are achieved through 4-node solver. ► Increasing ISAT memory from 0.1 to 2 GB reduces the CPU runtime by almost 35%. ► Combustion and soot processes are predicted well with minimal computational cost. - Abstract: In the present study, in-cylinder diesel combustion simulation was performed with parallel processing on an Intel Xeon Quad-Core platform to allow both fluid dynamics and chemical kinetics of the surrogate diesel fuel model to be solved simultaneously on multiple processors. Here, Cartesian Z-Coordinate was selected as the most appropriate partitioning algorithm since it computationally bisects the domain such that the dynamic load associated with fuel particle tracking was evenly distributed during parallel computations. Other variables examined included number of compute nodes, chemistry sizes and in situ adaptive tabulation (ISAT) parameters. Based on the performance benchmarking test conducted, parallel configuration of 4-compute node was found to reduce the computational runtime most efficiently whereby a parallel efficiency of up to 75.4% was achieved. The simulation results also indicated that accuracy level was insensitive to the number of partitions or the partitioning algorithms. The effect of reducing the number of species on computational runtime was observed to be more significant than reducing the number of reactions. Besides, the study showed that an increase in the ISAT maximum storage of up to 2 GB reduced the computational runtime by 50%. Also, the ISAT error tolerance of 10 −3 was chosen to strike a balance between results accuracy and computational runtime. The optimised parameters in parallel processing and ISAT, as well as the use of the in-house reduced chemistry model allowed accurate

  5. Modifications Caused by Enzyme-Retting and Their Effect on Composite Performance

    Directory of Open Access Journals (Sweden)

    Jonn A. Foulk

    2011-01-01

    Full Text Available Bethune seed flax was collected from Canada with seed removed using a stripper header and straw pulled and left in field for several weeks. Unretted straw was decorticated providing a coarse fiber bundle feedstock for enzyme treatments. Enzyme treatments using a bacterial pectinolytic enzyme with lyase activity were conducted in lab-scale reactors. Four fiber specimens were created: no retting, minimal retting, moderate retting, and full retting. Fiber characterization tests: strength, elongation, diameter, metal content, wax content, and pH were conducted with significant differences between fibers. Thermosetting vinyl ester resin was used to produce composite panels via vacuum-assisted infusion. Composite performance was evaluated using fiber bundle pull-out, tensile, impact, and interlaminar shear tests. Composite tests indicate that composite panels are largely unchanged among fiber samples. Variation in composite performance might not be realized due to poor interfacial bonding being of larger impact than the more subtle changes incurred by the enzyme treatment.

  6. Enzyme Biosensors for Biomedical Applications: Strategies for Safeguarding Analytical Performances in Biological Fluids

    Science.gov (United States)

    Rocchitta, Gaia; Spanu, Angela; Babudieri, Sergio; Latte, Gavinella; Madeddu, Giordano; Galleri, Grazia; Nuvoli, Susanna; Bagella, Paola; Demartis, Maria Ilaria; Fiore, Vito; Manetti, Roberto; Serra, Pier Andrea

    2016-01-01

    Enzyme-based chemical biosensors are based on biological recognition. In order to operate, the enzymes must be available to catalyze a specific biochemical reaction and be stable under the normal operating conditions of the biosensor. Design of biosensors is based on knowledge about the target analyte, as well as the complexity of the matrix in which the analyte has to be quantified. This article reviews the problems resulting from the interaction of enzyme-based amperometric biosensors with complex biological matrices containing the target analyte(s). One of the most challenging disadvantages of amperometric enzyme-based biosensor detection is signal reduction from fouling agents and interference from chemicals present in the sample matrix. This article, therefore, investigates the principles of functioning of enzymatic biosensors, their analytical performance over time and the strategies used to optimize their performance. Moreover, the composition of biological fluids as a function of their interaction with biosensing will be presented. PMID:27249001

  7. Design of parallel dual-energy X-ray beam and its performance for security radiography

    International Nuclear Information System (INIS)

    Kim, Kwang Hyun; Myoung, Sung Min; Chung, Yong Hyun

    2011-01-01

    A new concept of dual-energy X-ray beam generation and acquisition of dual-energy security radiography is proposed. Erbium (Er) and rhodium (Rh) with a copper filter were positioned in front of X-ray tube to generate low- and high-energy X-ray spectra. Low- and high-energy X-rays were guided to separately enter into two parallel detectors. Monte Carlo code of MCNPX was used to derive an optimum thickness of each filter for improved dual X-ray image quality. It was desired to provide separation ability between organic and inorganic matters for the condition of 140 kVp/0.8 mA as used in the security application. Acquired dual-energy X-ray beams were evaluated by the dual-energy Z-map yielding enhanced performance compared with a commercial dual-energy detector. A collimator for the parallel dual-energy X-ray beam was designed to minimize X-ray beam interference between low- and high-energy parallel beams for 500 mm source-to-detector distance.

  8. High performance parallel computing of flows in complex geometries: I. Methods

    International Nuclear Information System (INIS)

    Gourdain, N; Gicquel, L; Montagnac, M; Vermorel, O; Staffelbach, G; Garcia, M; Boussuge, J-F; Gazaix, M; Poinsot, T

    2009-01-01

    Efficient numerical tools coupled with high-performance computers, have become a key element of the design process in the fields of energy supply and transportation. However flow phenomena that occur in complex systems such as gas turbines and aircrafts are still not understood mainly because of the models that are needed. In fact, most computational fluid dynamics (CFD) predictions as found today in industry focus on a reduced or simplified version of the real system (such as a periodic sector) and are usually solved with a steady-state assumption. This paper shows how to overcome such barriers and how such a new challenge can be addressed by developing flow solvers running on high-end computing platforms, using thousands of computing cores. Parallel strategies used by modern flow solvers are discussed with particular emphases on mesh-partitioning, load balancing and communication. Two examples are used to illustrate these concepts: a multi-block structured code and an unstructured code. Parallel computing strategies used with both flow solvers are detailed and compared. This comparison indicates that mesh-partitioning and load balancing are more straightforward with unstructured grids than with multi-block structured meshes. However, the mesh-partitioning stage can be challenging for unstructured grids, mainly due to memory limitations of the newly developed massively parallel architectures. Finally, detailed investigations show that the impact of mesh-partitioning on the numerical CFD solutions, due to rounding errors and block splitting, may be of importance and should be accurately addressed before qualifying massively parallel CFD tools for a routine industrial use.

  9. Forced-convection boiling tests performed in parallel simulated LMR fuel assemblies

    International Nuclear Information System (INIS)

    Rose, S.D.; Carbajo, J.J.; Levin, A.E.; Lloyd, D.B.; Montgomery, B.H.; Wantland, J.L.

    1985-01-01

    Forced-convection tests have been carried out using parallel simulated Liquid Metal Reactor fuel assemblies in an engineering-scale sodium loop, the Thermal-Hydraulic Out-of-Reactor Safety facility. The tests, performed under single- and two-phase conditions, have shown that for low forced-convection flow there is significant flow augmentation by thermal convection, an important phenomenon under degraded shutdown heat removal conditions in an LMR. The power and flows required for boiling and dryout to occur are much higher than decay heat levels. The experimental evidence supports analytical results that heat removal from an LMR is possible with a degraded shutdown heat removal system

  10. Performance analysis of a refrigeration system with parallel control of evaporation pressure

    International Nuclear Information System (INIS)

    Lee, Jong Suk

    2008-01-01

    The conventional refrigeration system is composed of a compressor, condenser, receiver, expansion valve or capillary tube, and an evaporator. The refrigeration system used in this study has additional expansion valve and evaporator along with an Evaporation Pressure Regulator(EPR) at the exit side of the evaporator. The two evaporators can be operated at different temperatures according to the opening of the EPR. The experimental results obtained using the refrigeration system with parallel control of evaporation pressure are presented and the performance analysis of the refrigeration system with two evaporators is conducted

  11. Comparing the performance of different meta-heuristics for unweighted parallel machine scheduling

    Directory of Open Access Journals (Sweden)

    Adamu, Mumuni Osumah

    2015-08-01

    Full Text Available This article considers the due window scheduling problem to minimise the number of early and tardy jobs on identical parallel machines. This problem is known to be NP complete and thus finding an optimal solution is unlikely. Three meta-heuristics and their hybrids are proposed and extensive computational experiments are conducted. The purpose of this paper is to compare the performance of these meta-heuristics and their hybrids and to determine the best among them. Detailed comparative tests have also been conducted to analyse the different heuristics with the simulated annealing hybrid giving the best result.

  12. Performance of a parallel plate ionization chamber in beta radiation dosimetry

    Energy Technology Data Exchange (ETDEWEB)

    Antonio, Patricia L.; Caldas, Linda V.E., E-mail: patrilan@ipen.b, E-mail: lcaldas@ipen.b [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil)

    2011-07-01

    A homemade parallel plate ionization chamber with graphite collecting electrode, and developed for use in mammography beams, was tested in relation to its usefulness in beta radiation dosimetry at the Calibration Laboratory of IPEN. Characterization tests of this ionization chamber were performed, using the Sr-90 + Y-90, Kr-85 and Pm-147 sources of a beta secondary standard system. The results of saturation, leakage current, stabilization time, response stability, linearity, angular dependence, and calibration coefficients are within the recommended limits of international recommendations that indicate that this chamber may be used for beta radiation dosimetry. (author)

  13. Performance of a parallel plate ionization chamber in beta radiation dosimetry

    International Nuclear Information System (INIS)

    Antonio, Patricia L.; Caldas, Linda V.E.

    2011-01-01

    A homemade parallel plate ionization chamber with graphite collecting electrode, and developed for use in mammography beams, was tested in relation to its usefulness in beta radiation dosimetry at the Calibration Laboratory of IPEN. Characterization tests of this ionization chamber were performed, using the Sr-90 + Y-90, Kr-85 and Pm-147 sources of a beta secondary standard system. The results of saturation, leakage current, stabilization time, response stability, linearity, angular dependence, and calibration coefficients are within the recommended limits of international recommendations that indicate that this chamber may be used for beta radiation dosimetry. (author)

  14. Impact of interference on the performance of selection based parallel multiuser scheduling

    KAUST Repository

    Nam, Sungsik

    2012-02-01

    In conventional multiuser parallel scheduling schemes, every scheduled user is interfering with every other scheduled user, which limits the capacity and performance of multiuser systems, and the level of interference becomes substantial as the number of scheduled users increases. Based on the above observations, we investigate the trade-off between the system throughput and the number of scheduled users through the exact analysis of the total average sum rate capacity and the average spectral efficiency. Our analytical results can help the system designer to carefully select the appropriate number of scheduled users to maximize the overall throughput while maintaining an acceptable quality of service under certain channel conditions. © 2012 IEEE.

  15. Open | SpeedShop: An Open Source Infrastructure for Parallel Performance Analysis

    Directory of Open Access Journals (Sweden)

    Martin Schulz

    2008-01-01

    Full Text Available Over the last decades a large number of performance tools has been developed to analyze and optimize high performance applications. Their acceptance by end users, however, has been slow: each tool alone is often limited in scope and comes with widely varying interfaces and workflow constraints, requiring different changes in the often complex build and execution infrastructure of the target application. We started the Open | SpeedShop project about 3 years ago to overcome these limitations and provide efficient, easy to apply, and integrated performance analysis for parallel systems. Open | SpeedShop has two different faces: it provides an interoperable tool set covering the most common analysis steps as well as a comprehensive plugin infrastructure for building new tools. In both cases, the tools can be deployed to large scale parallel applications using DPCL/Dyninst for distributed binary instrumentation. Further, all tools developed within or on top of Open | SpeedShop are accessible through multiple fully equivalent interfaces including an easy-to-use GUI as well as an interactive command line interface reducing the usage threshold for those tools.

  16. Performance and scalability analysis of teraflop-scale parallel architectures using multidimensional wavefront applications

    International Nuclear Information System (INIS)

    Hoisie, A.; Lubeck, O.; Wasserman, H.

    1998-01-01

    The authors develop a model for the parallel performance of algorithms that consist of concurrent, two-dimensional wavefronts implemented in a message passing environment. The model, based on a LogGP machine parameterization, combines the separate contributions of computation and communication wavefronts. They validate the model on three important supercomputer systems, on up to 500 processors. They use data from a deterministic particle transport application taken from the ASCI workload, although the model is general to any wavefront algorithm implemented on a 2-D processor domain. They also use the validated model to make estimates of performance and scalability of wavefront algorithms on 100-TFLOPS computer systems expected to be in existence within the next decade as part of the ASCI program and elsewhere. In this context, they analyze two problem sizes. The model shows that on the largest such problem (1 billion cells), inter-processor communication performance is not the bottleneck. Single-node efficiency is the dominant factor

  17. High-performance parallel processors based on star-coupled wavelength division multiplexing optical interconnects

    Science.gov (United States)

    Deri, Robert J.; DeGroot, Anthony J.; Haigh, Ronald E.

    2002-01-01

    As the performance of individual elements within parallel processing systems increases, increased communication capability between distributed processor and memory elements is required. There is great interest in using fiber optics to improve interconnect communication beyond that attainable using electronic technology. Several groups have considered WDM, star-coupled optical interconnects. The invention uses a fiber optic transceiver to provide low latency, high bandwidth channels for such interconnects using a robust multimode fiber technology. Instruction-level simulation is used to quantify the bandwidth, latency, and concurrency required for such interconnects to scale to 256 nodes, each operating at 1 GFLOPS performance. Performance scales have been shown to .apprxeq.100 GFLOPS for scientific application kernels using a small number of wavelengths (8 to 32), only one wavelength received per node, and achievable optoelectronic bandwidth and latency.

  18. Effect of Different Levels of Extruded Soybean and Avizyme Enzyme on Broiler Performance

    Directory of Open Access Journals (Sweden)

    H Nasiri Mogadam

    2012-01-01

    Full Text Available An experiment was conducted to examine the effect of different levels of extruded soybean and enzyme on broiler performance. In a completely randomized design with 2×3 factorial arrangement, 480 one day-old, Ross broiler chickens were divided into 40 groups, 12 chicks per pen. Treatments were consisting of combination of four levels of extruded soybean (0.0, 5.0, 10.0 and 15.0 % and two levels of enzyme (0.0 and 500 g per ton. Different levels of extruded soybean and enzyme had no significant effect on blood factors such as cholesterol, triglyceride and the weight of liver and heart. The usage of extruded soybean and enzyme showed significantly higher weight gain and better feed conversion (p

  19. Homemade Buckeye-Pi: A Learning Many-Node Platform for High-Performance Parallel Computing

    Science.gov (United States)

    Amooie, M. A.; Moortgat, J.

    2017-12-01

    We report on the "Buckeye-Pi" cluster, the supercomputer developed in The Ohio State University School of Earth Sciences from 128 inexpensive Raspberry Pi (RPi) 3 Model B single-board computers. Each RPi is equipped with fast Quad Core 1.2GHz ARMv8 64bit processor, 1GB of RAM, and 32GB microSD card for local storage. Therefore, the cluster has a total RAM of 128GB that is distributed on the individual nodes and a flash capacity of 4TB with 512 processors, while it benefits from low power consumption, easy portability, and low total cost. The cluster uses the Message Passing Interface protocol to manage the communications between each node. These features render our platform the most powerful RPi supercomputer to date and suitable for educational applications in high-performance-computing (HPC) and handling of large datasets. In particular, we use the Buckeye-Pi to implement optimized parallel codes in our in-house simulator for subsurface media flows with the goal of achieving a massively-parallelized scalable code. We present benchmarking results for the computational performance across various number of RPi nodes. We believe our project could inspire scientists and students to consider the proposed unconventional cluster architecture as a mainstream and a feasible learning platform for challenging engineering and scientific problems.

  20. Performance of Polycrystalline Photovoltaic and Thermal Collector (PVT on Serpentine-Parallel Absorbers Design

    Directory of Open Access Journals (Sweden)

    Mustofa Mustofa

    2017-03-01

    Full Text Available This paper presents the performance of an unglazed polycrystalline photovoltaic-thermal PVT on 0.045 kg/s mass flow rate. PVT combine photovoltaic modules and solar thermal collectors, forming a single device that receive solar radiation and produces heat and electricity simultaneously. The collector figures out serpentine-parallel tubes that can prolong fluid heat conductivity from morning till afternoon. During testing, cell PV, inlet and outlet fluid temperaturs were recorded by thermocouple digital LM35 Arduino Mega 2560. Panel voltage and electric current were also noted in which they were connected to computer and presented each second data recorded. But, in this performance only shows in the certain significant time data. This because the electric current was only noted by multimeter device not the digital one. Based on these testing data, average cell efficieny was about 19%, while thermal efficiency of above 50% and correspondeng cell efficiency of 11%, respectively

  1. Performance of Polycrystalline Photovoltaic and Thermal Collector (PVT on Serpentine-Parallel Absor

    Directory of Open Access Journals (Sweden)

    Mustofa

    2015-10-01

    Full Text Available This paper presents the performance of an unglazed polycrystalline photovoltaic-thermal PVT on 0.045 kg/s mass flow rate. PVT combine photovoltaic modules and solar thermal collectors, forming a single device that receive solar radiation and produces heat and electricity simultaneously. The collector figures out serpentine-parallel tubes that can prolong fluid heat conductivity from morning till afternoon. During testing, cell PV, inlet and outlet fluid temperatures were recorded by thermocouple digital LM35 Arduino Mega 2560. Panel voltage and electric current were also noted in which they were connected to computer and presented each second data recorded. But, in this performance only shows in the certain significant time data. This because the electric current was only noted by multimeter device not the digital one. Based on these testing data, average cell efficiency was about 19%, while thermal efficiency of above 50% and correspondent cell efficiency of 11%, respectively.

  2. High-Performance Parallel and Stream Processing of X-ray Microdiffraction Data on Multicores

    International Nuclear Information System (INIS)

    Bauer, Michael A; McIntyre, Stewart; Xie Yuzhen; Biem, Alain; Tamura, Nobumichi

    2012-01-01

    We present the design and implementation of a high-performance system for processing synchrotron X-ray microdiffraction (XRD) data in IBM InfoSphere Streams on multicore processors. We report on the parallel and stream processing techniques that we use to harvest the power of clusters of multicores to analyze hundreds of gigabytes of synchrotron XRD data in order to reveal the microtexture of polycrystalline materials. The timing to process one XRD image using one pipeline is about ten times faster than the best C program at present. With the support of InfoSphere Streams platform, our software is able to be scaled up to operate on clusters of multi-cores for processing multiple images concurrently. This system provides a high-performance processing kernel to achieve near real-time data analysis of image data from synchrotron experiments.

  3. HPC-NMF: A High-Performance Parallel Algorithm for Nonnegative Matrix Factorization

    Energy Technology Data Exchange (ETDEWEB)

    2016-08-22

    NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient distributed algorithms to solve the problem for big data sets. We propose a high-performance distributed-memory parallel algorithm that computes the factorization by iteratively solving alternating non-negative least squares (NLS) subproblems for $\\WW$ and $\\HH$. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). As opposed to previous implementation, our algorithm is also flexible: It performs well for both dense and sparse matrices, and allows the user to choose any one of the multiple algorithms for solving the updates to low rank factors $\\WW$ and $\\HH$ within the alternating iterations.

  4. High performance parallel computing of flows in complex geometries: II. Applications

    International Nuclear Information System (INIS)

    Gourdain, N; Gicquel, L; Staffelbach, G; Vermorel, O; Duchaine, F; Boussuge, J-F; Poinsot, T

    2009-01-01

    Present regulations in terms of pollutant emissions, noise and economical constraints, require new approaches and designs in the fields of energy supply and transportation. It is now well established that the next breakthrough will come from a better understanding of unsteady flow effects and by considering the entire system and not only isolated components. However, these aspects are still not well taken into account by the numerical approaches or understood whatever the design stage considered. The main challenge is essentially due to the computational requirements inferred by such complex systems if it is to be simulated by use of supercomputers. This paper shows how new challenges can be addressed by using parallel computing platforms for distinct elements of a more complex systems as encountered in aeronautical applications. Based on numerical simulations performed with modern aerodynamic and reactive flow solvers, this work underlines the interest of high-performance computing for solving flow in complex industrial configurations such as aircrafts, combustion chambers and turbomachines. Performance indicators related to parallel computing efficiency are presented, showing that establishing fair criterions is a difficult task for complex industrial applications. Examples of numerical simulations performed in industrial systems are also described with a particular interest for the computational time and the potential design improvements obtained with high-fidelity and multi-physics computing methods. These simulations use either unsteady Reynolds-averaged Navier-Stokes methods or large eddy simulation and deal with turbulent unsteady flows, such as coupled flow phenomena (thermo-acoustic instabilities, buffet, etc). Some examples of the difficulties with grid generation and data analysis are also presented when dealing with these complex industrial applications.

  5. Performance modeling and analysis of parallel Gaussian elimination on multi-core computers

    Directory of Open Access Journals (Sweden)

    Fadi N. Sibai

    2014-01-01

    Full Text Available Gaussian elimination is used in many applications and in particular in the solution of systems of linear equations. This paper presents mathematical performance models and analysis of four parallel Gaussian Elimination methods (precisely the Original method and the new Meet in the Middle –MiM– algorithms and their variants with SIMD vectorization on multi-core systems. Analytical performance models of the four methods are formulated and presented followed by evaluations of these models with modern multi-core systems’ operation latencies. Our results reveal that the four methods generally exhibit good performance scaling with increasing matrix size and number of cores. SIMD vectorization only makes a large difference in performance for low number of cores. For a large matrix size (n ⩾ 16 K, the performance difference between the MiM and Original methods falls from 16× with four cores to 4× with 16 K cores. The efficiencies of all four methods are low with 1 K cores or more stressing a major problem of multi-core systems where the network-on-chip and memory latencies are too high in relation to basic arithmetic operations. Thus Gaussian Elimination can greatly benefit from the resources of multi-core systems, but higher performance gains can be achieved if multi-core systems can be designed with lower memory operation, synchronization, and interconnect communication latencies, requirements of utmost importance and challenge in the exascale computing age.

  6. Optimal design of a spherical parallel manipulator based on kinetostatic performance using evolutionary techniques

    Energy Technology Data Exchange (ETDEWEB)

    Daneshmand, Morteza [University of Tartu, Tartu (Estonia); Saadatzi, Mohammad Hossein [Colorado School of Mines, Golden (United States); Kaloorazi, Mohammad Hadi [École de Technologie Supérieur, Montréal (Canada); Masouleh, Mehdi Tale [University of Tehran, Tehran (Iran, Islamic Republic of); Anbarjafari, Gholamreza [Hasan Kalyoncu University, Gaziantep (Turkmenistan)

    2016-03-15

    This study aims to provide an optimal design for a Spherical parallel manipulator (SPM), namely, the Agile Eye. This aim is approached by investigating kinetostatic performance and workspace and searching for the most promising design. Previously recommended designs are examined to determine whether they provide acceptable kinetostatic performance and workspace. Optimal designs are provided according to different kinetostatic performance indices, especially kinematic sensitivity. The optimization process is launched based on the concept of the genetic algorithm. A single-objective process is implemented in accordance with the guidelines of an evolutionary algorithm called differential evolution. A multi-objective procedure is then provided following the reasoning of the nondominated sorting genetic algorithm-II. This process results in several sets of Pareto points for reconciliation between kinetostatic performance indices and workspace. The concept of numerous kinetostatic performance indices and the results of optimization algorithms are elaborated. The conclusions provide hints on the provided set of designs and their credibility to provide a well-conditioned workspace and acceptable kinetostatic performance for the SPM under study, which can be well extended to other types of SPMs.

  7. MulticoreBSP for C : A high-performance library for shared-memory parallel programming

    NARCIS (Netherlands)

    Yzelman, A. N.; Bisseling, R. H.; Roose, D.; Meerbergen, K.

    2014-01-01

    The bulk synchronous parallel (BSP) model, as well as parallel programming interfaces based on BSP, classically target distributed-memory parallel architectures. In earlier work, Yzelman and Bisseling designed a MulticoreBSP for Java library specifically for shared-memory architectures. In the

  8. H5Part A Portable High Performance Parallel Data Interface for Particle Simulations

    CERN Document Server

    Adelmann, Andreas; Shalf, John M; Siegerist, Cristina

    2005-01-01

    Largest parallel particle simulations, in six dimensional phase space generate wast amont of data. It is also desirable to share data and data analysis tools such as ParViT (Particle Visualization Toolkit) among other groups who are working on particle-based accelerator simulations. We define a very simple file schema built on top of HDF5 (Hierarchical Data Format version 5) as well as an API that simplifies the reading/writing of the data to the HDF5 file format. HDF5 offers a self-describing machine-independent binary file format that supports scalable parallel I/O performance for MPI codes on a variety of supercomputing systems and works equally well on laptop computers. The API is available for C, C++, and Fortran codes. The file format will enable disparate research groups with very different simulation implementations to share data transparently and share data analysis tools. For instance, the common file format will enable groups that depend on completely different simulation implementations to share c...

  9. Air-side performance of a parallel-flow parallel-fin (PF{sup 2}) heat exchanger in sequential frosting

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Ping [Zhejiang Vocational College of Commerce, Hangzhou, Binwen Road 470 (China); Department of Mechanical Science and Engineering, University of Illinois at Urbana-Champaign, 1206 West Green Street, Urbana, IL 61801 (United States); Hrnjak, P.S. [Department of Mechanical Science and Engineering, University of Illinois at Urbana-Champaign, 1206 West Green Street, Urbana, IL 61801 (United States)

    2010-09-15

    The thermal-hydraulic performance in periodic frosting conditions is experimentally studied for the parallel-flow parallel-fin heat exchanger, henceforth referred to as a PF{sup 2} heat exchanger, a new style of heat exchanger that uses louvered bent fins on flat tubes to enhance water drainage when the flat tubes are horizontal. Typically, it takes a few frosting/defrosting cycles to come to repeatable conditions. The criterion for the initiation of defrost and a sufficiently long defrost period are determined for the test PF{sup 2} heat exchanger and test condition. The effects of blower operation on the pressure drop, frost accumulation, water retention, and capacity in time are compared under the conditions of 15 sequential frosting cycles. Pressure drop across the heat exchanger and overall heat transfer coefficient are quantified under frost conditions as functions of the air humidity and air face velocity. The performances of two types of flat-tube heat exchangers, PF{sup 2} heat exchanger and conventional parallel-flow serpentine-fin (PFSF) heat exchanger, are compared and the results obtained are presented. (author)

  10. Effect of Alchornea cordifolia leaf meal inclusion and enzyme supplementation on performance and digestibility of rabbits

    Directory of Open Access Journals (Sweden)

    S.O. Ayodele

    2016-09-01

    Full Text Available A feeding trial was conducted to study the performance, digestibility and health status of weaner rabbits fed diets including Alchornea cordifolia leaf meal (ALM: 18% crude protein [CP] and 12.9% crude fibre and supplemented with a multi-enzyme additive (cellulase, xylanase, β-glucanase, α-amylase, protease, lipase. Six experimental diets were arranged factorially: 3 levels of ALM (0, 5 and 10% substituting palm kernel cake: 16.3% CP and 39.1% neutral detergent fibre combined with 2 levels of enzyme supplementation (0 and 0.35 g/kg. One hundred and eighty healthy, 5-wk-old weaner rabbits of cross-breeds were randomly allotted to 6 dietary treatments (30 rabbits/treatment, 3 rabbits/replicate. Growth rate was not affected (P>0.05 by the main factors (exogenous enzyme and ALM inclusion and their interactions (13.5 g/d on av.. Daily feed intake and feed conversion ratio decreased (P=0.01 with the ALM inclusion by 8%, but did not affect faecal digestibility. However, enzyme supplementation improved crude protein and crude fibre digestibility (P<0.001 by 6%. In conclusion, ALM inclusion and enzyme supplementation had no adverse effect on the performance and digestibility of rabbits.

  11. Parallel Backprojection: A Case Study in High-Performance Reconfigurable Computing

    Directory of Open Access Journals (Sweden)

    Cordes Ben

    2009-01-01

    Full Text Available High-performance reconfigurable computing (HPRC is a novel approach to provide large-scale computing power to modern scientific applications. Using both general-purpose processors and FPGAs allows application designers to exploit fine-grained and coarse-grained parallelism, achieving high degrees of speedup. One scientific application that benefits from this technique is backprojection, an image formation algorithm that can be used as part of a synthetic aperture radar (SAR processing system. We present an implementation of backprojection for SAR on an HPRC system. Using simulated data taken at a variety of ranges, our implementation runs over 200 times faster than a similar software program, with an overall application speedup better than 50x. The backprojection application is easily parallelizable, achieving near-linear speedup when run on multiple nodes of a clustered HPRC system. The results presented can be applied to other systems and other algorithms with similar characteristics.

  12. Parallel Backprojection: A Case Study in High-Performance Reconfigurable Computing

    Directory of Open Access Journals (Sweden)

    2009-03-01

    Full Text Available High-performance reconfigurable computing (HPRC is a novel approach to provide large-scale computing power to modern scientific applications. Using both general-purpose processors and FPGAs allows application designers to exploit fine-grained and coarse-grained parallelism, achieving high degrees of speedup. One scientific application that benefits from this technique is backprojection, an image formation algorithm that can be used as part of a synthetic aperture radar (SAR processing system. We present an implementation of backprojection for SAR on an HPRC system. Using simulated data taken at a variety of ranges, our implementation runs over 200 times faster than a similar software program, with an overall application speedup better than 50x. The backprojection application is easily parallelizable, achieving near-linear speedup when run on multiple nodes of a clustered HPRC system. The results presented can be applied to other systems and other algorithms with similar characteristics.

  13. Optimization Solutions for Improving the Performance of the Parallel Reduction Algorithm Using Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Ion LUNGU

    2012-01-01

    Full Text Available In this paper, we research, analyze and develop optimization solutions for the parallel reduction function using graphics processing units (GPUs that implement the Compute Unified Device Architecture (CUDA, a modern and novel approach for improving the software performance of data processing applications and algorithms. Many of these applications and algorithms make use of the reduction function in their computational steps. After having designed the function and its algorithmic steps in CUDA, we have progressively developed and implemented optimization solutions for the reduction function. In order to confirm, test and evaluate the solutions' efficiency, we have developed a custom tailored benchmark suite. We have analyzed the obtained experimental results regarding: the comparison of the execution time and bandwidth when using graphic processing units covering the main CUDA architectures (Tesla GT200, Fermi GF100, Kepler GK104 and a central processing unit; the data type influence; the binary operator's influence.

  14. The design and performance of the parallel multiprocessor nuclear physics data acquisition system, DAPHNE

    International Nuclear Information System (INIS)

    Welch, L.C.; Moog, T.H.; Daly, R.T.; Videbaek, F.

    1987-05-01

    The ever increasing complexity of nuclear physics experiments places severe demands on computerized data acquisition systems. A natural evolution of these systems, taking advantages of the independent nature of ''events,'' is to use identical parallel microcomputers in a front end to simultaneously analyze separate events. Such a system has been developed at Argonne to serve the needs of the experimental program of ATLAS, a new superconducting heavy-ion accelerator and other on-going research. Using microcomputers based on the National Semiconductor 32016 microprocessor housed in a Multibus I cage, CPU power equivalent to several VAXs is obtained at a fraction of the cost of one VAX. The front end interfacs to a VAX 11/750 on which an extensive user friendly command language based on DCL resides. The whole system, known as DAPHNE, also provides the means to reply data using the same command language. Design concepts, data structures, performance, and experience to data are discussed

  15. The design, creation, and performance of the parallel multiprocessor nuclear physics data acquisition system, DAPHNE

    International Nuclear Information System (INIS)

    Welch, L.C.; Moog, T.H.; Daly, R.T.; Videbaek, F.

    1986-01-01

    The ever increasing complexity of nuclear physics experiments places severe demands on computerized data acquisition systems. A natural evolution of these system, taking advantage of the independent nature of ''events'', is to use identical parallel microcomputers in a front end to simultaneously analyze separate events. Such a system has been developed at Argonne to serve the needs of the experimental program of ATLAS, a new superconducting heavy-ion accelerator and other on-going research. Using microcomputers based on the National Semiconductor 32016 microprocessor housed in a Multibus I cage, multi-VAX cpu power is obtained at a fraction of the cost of one VAX. The front end interfaces to a VAX 750 on which an extensive user friendly command language based on DCL resides. The whole system, known as DAPHNE, also provides the means to replay data using the same command language. Design concepts, data structures, performance, and experience to data are discussed. 5 refs., 2 figs

  16. Workspace optimization and kinematic performance evaluation of 2-DOF parallel mechanisms

    International Nuclear Information System (INIS)

    Nam, Yun Joo; Park, Myeong Kwan

    2006-01-01

    This paper presents the kinematics and workspace optimization of the two different 2-DOF (Degrees-of-Freedom) planar parallel mechanisms: one (called 2-RPR mechanism) with translational actuators and the other (called 2-RRR mechanism) with rotational ones. First of all, the inverse kinematics and Jacobian matrix for each mechanism are derived analytically. Then, the workspace including the output-space and the joint-space is systematically analyzed in order to determine the geometric parameters and the operating range of the actuators. Finally, the kinematic optimization of the mechanisms is performed in consideration of their dexterity and rigidity. It is expected that the optimization results can be effectively used as a basic material for the applications of the presented mechanisms to more industrial fields

  17. Performance Analysis of a Threshold-Based Parallel Multiple Beam Selection Scheme for WDM FSO Systems

    KAUST Repository

    Nam, Sung Sik

    2018-04-09

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme for a free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has occurred under independent identically distributed Gamma-Gamma fading conditions. To simplify the mathematical analysis, we additionally consider Gamma turbulence conditions, which are a good approximation of Gamma-Gamma distribution. Specifically, we statistically analyze the characteristics in operation under conventional detection schemes (i.e., heterodyne detection (HD) and intensity modulation/direct detection (IM/DD) techniques) for both adaptive modulation (AM) case in addition to non-AM case (i.e., coherent/non-coherent binary modulation). Then, based on the statistically derived results, we evaluate the outage probability of a selected beam, the average spectral efficiency (ASE), the average number of selected beams (ANSB) and the average bit error rate (BER). Selected results show that we can obtain higher spectral efficiency and simultaneously reduce the potential for increasing the complexity of implementation caused by applying the selection-based beam selection scheme without considerable performance loss. Especially for the AM case, the ASE can be increased further compared to the non- AM cases. Our derived results based on the Gamma distribution as an approximation of the Gamma-Gamma distribution can be used as approximated performance measure bounds, especially, they may lead to lower bounds on the approximated considered performance measures.

  18. Ultra-performance liquid chromatography-tandem mass spectrometry-based multiplex enzyme assay for six enzymes associated with hereditary hemolytic anemia.

    Science.gov (United States)

    Park, Chul Min; Lee, Kyunghoon; Jun, Sun-Hee; Song, Sang Hoon; Song, Junghan

    2017-08-15

    Deficiencies in erythrocyte metabolic enzymes are associated with hereditary hemolytic anemia. Here, we report the development of a novel multiplex enzyme assay for six major enzymes, namely glucose-6-phosphate dehydrogenase, pyruvate kinase, pyrimidine 5'-nucleotidase, hexokinase, triosephosphate isomerase, and adenosine deaminase, deficiencies in which are implicated in erythrocyte enzymopathies. To overcome the drawbacks of traditional spectrophotometric enzyme assays, the present assay was based on ultra-performance liquid chromatography-tandem mass spectrometry (UPLC-MS/MS). The products of the six enzymes were directly measured by using ion pairing UPLC-MS/MS, and the precision, linearity, ion suppression, optimal sample amounts, and incubation times were evaluated. Eighty-three normal individuals and 13 patients with suspected enzymopathy were analyzed. The UPLC running time was within 5min. No ion suppression was observed at the retention time for the products or internal standards. We selected an optimal dilution factor and incubation time for each enzyme system. The intra- and inter-assay imprecision values (CVs) were 2.5-12.1% and 2.9-14.3%, respectively. The linearity of each system was good, with R 2 values >0.97. Patient samples showed consistently lower enzyme activities than those from normal individuals. The present ion paring UPLC-MS/MS assay enables facile and reproducible multiplex evaluation of the activity of enzymes implicated in enzymopathy-associated hemolytic anemia. Copyright © 2017 Elsevier B.V. All rights reserved.

  19. PERFORMANCE ANALYSIS BETWEEN EXPLICIT SCHEDULING AND IMPLICIT SCHEDULING OF PARALLEL ARRAY-BASED DOMAIN DECOMPOSITION USING OPENMP

    Directory of Open Access Journals (Sweden)

    MOHAMMED FAIZ ABOALMAALY

    2014-10-01

    Full Text Available With the continuous revolution of multicore architecture, several parallel programming platforms have been introduced in order to pave the way for fast and efficient development of parallel algorithms. Back into its categories, parallel computing can be done through two forms: Data-Level Parallelism (DLP or Task-Level Parallelism (TLP. The former can be done by the distribution of data among the available processing elements while the latter is based on executing independent tasks concurrently. Most of the parallel programming platforms have built-in techniques to distribute the data among processors, these techniques are technically known as automatic distribution (scheduling. However, due to their wide range of purposes, variation of data types, amount of distributed data, possibility of extra computational overhead and other hardware-dependent factors, manual distribution could achieve better outcomes in terms of performance when compared to the automatic distribution. In this paper, this assumption is investigated by conducting a comparison between automatic and our newly proposed manual distribution of data among threads in parallel. Empirical results of matrix addition and matrix multiplication show a considerable performance gain when manual distribution is applied against automatic distribution.

  20. Parallelization of the AliRoot event reconstruction by performing a semi- automatic source-code transformation

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    side bus or processor interconnections. Parallelism can only result in performance gain, if the memory usage is optimized, memory locality improved and the communication between threads is minimized. But the domain of concurrent programming has become a field for highly skilled experts, as the implementation of multithreading is difficult, error prone and labor intensive. A full re-implementation for parallel execution of existing offline frameworks, like AliRoot in ALICE, is thus unaffordable. An alternative method, is to use a semi-automatic source-to-source transformation for getting a simple parallel design, with almost no interference between threads. This reduces the need of rewriting the develop...

  1. Parallel Application Performance on Two Generations of Intel Xeon HPC Platforms

    Energy Technology Data Exchange (ETDEWEB)

    Chang, Christopher H.; Long, Hai; Sides, Scott; Vaidhynathan, Deepthi; Jones, Wesley

    2015-10-15

    Two next-generation node configurations hosting the Haswell microarchitecture were tested with a suite of microbenchmarks and application examples, and compared with a current Ivy Bridge production node on NREL" tm s Peregrine high-performance computing cluster. A primary conclusion from this study is that the additional cores are of little value to individual task performance--limitations to application parallelism, or resource contention among concurrently running but independent tasks, limits effective utilization of these added cores. Hyperthreading generally impacts throughput negatively, but can improve performance in the absence of detailed attention to runtime workflow configuration. The observations offer some guidance to procurement of future HPC systems at NREL. First, raw core count must be balanced with available resources, particularly memory bandwidth. Balance-of-system will determine value more than processor capability alone. Second, hyperthreading continues to be largely irrelevant to the workloads that are commonly seen, and were tested here, at NREL. Finally, perhaps the most impactful enhancement to productivity might occur through enabling multiple concurrent jobs per node. Given the right type and size of workload, more may be achieved by doing many slow things at once, than fast things in order.

  2. Effects of protease and non-starch polysaccharide enzyme on performance, digestive function, activity and gene expression of endogenous enzyme of broilers.

    Directory of Open Access Journals (Sweden)

    Lin Yuan

    Full Text Available Three hundred one-day-old male broiler chickens (Ross-308 were fed corn-soybean basal diets containing non-starch polysaccharide (NSP enzyme and different levels of acid protease from 1 to 42 days of age to investigate the effects of exogenous enzymes on growth performance, digestive function, activity of endogenous digestive enzymes in the pancreas and mRNA expression of pancreatic digestive enzymes. For days 1-42, compared to the control chickens, average daily feed intake (ADFI and average daily gain (ADG were significantly enhanced by the addition of NSP enzyme in combination with protease supplementation at 40 or 80 mg/kg (p<0.05. Feed-to-gain ratio (FGR was significantly improved by supplementation with NSP enzymes or NSP enzyme combined with 40 or 80 mg/kg protease compared to the control diet (p<0.05. Apparent digestibility of crude protein (ADCP was significantly enhanced by the addition of NSP enzyme or NSP enzyme combined with 40 or 80 mg/kg protease (p<0.05. Cholecystokinin (CCK level in serum was reduced by 31.39% with NSP enzyme combined with protease supplementation at 160 mg/kg (p<0.05, but the CCK level in serum was increased by 26.51% with NSP enzyme supplementation alone. After 21 days, supplementation with NSP enzyme and NSP enzyme combined with 40 or 80 mg/kg protease increased the activity of pancreatic trypsin by 74.13%, 70.66% and 42.59% (p<0.05, respectively. After 42 days, supplementation with NSP enzyme and NSP enzyme combined with 40 mg/kg protease increased the activity of pancreatic trypsin by 32.45% and 27.41%, respectively (p<0.05. However, supplementation with NSP enzyme and 80 or 160 mg/kg protease decreased the activity of pancreatic trypsin by 10.75% and 25.88%, respectively (p<0.05. The activities of pancreatic lipase and amylase were significantly higher in treated animals than they were in the control group (p<0.05. Supplementation with NSP enzyme, NSP enzyme combined with 40 or 80 mg/kg protease increased

  3. Operational mesoscale atmospheric dispersion prediction using high performance parallel computing cluster for emergency response

    International Nuclear Information System (INIS)

    Srinivas, C.V.; Venkatesan, R.; Muralidharan, N.V.; Das, Someshwar; Dass, Hari; Eswara Kumar, P.

    2005-08-01

    An operational atmospheric dispersion prediction system is implemented on a cluster super computer for 'Online Emergency Response' for Kalpakkam nuclear site. The numerical system constitutes a parallel version of a nested grid meso-scale meteorological model MM5 coupled to a random walk particle dispersion model FLEXPART. The system provides 48 hour forecast of the local weather and radioactive plume dispersion due to hypothetical air borne releases in a range of 100 km around the site. The parallel code was implemented on different cluster configurations like distributed and shared memory systems. Results of MM5 run time performance for 1-day prediction are reported on all the machines available for testing. A reduction of 5 times in runtime is achieved using 9 dual Xeon nodes (18 physical/36 logical processors) compared to a single node sequential run. Based on the above run time results a cluster computer facility with 9-node Dual Xeon is commissioned at IGCAR for model operation. The run time of a triple nested domain MM5 is about 4 h for 24 h forecast. The system has been operated continuously for a few months and results were ported on the IMSc home page. Initial and periodic boundary condition data for MM5 are provided by NCMRWF, New Delhi. An alternative source is found to be NCEP, USA. These two sources provide the input data to the operational models at different spatial and temporal resolutions and using different assimilation methods. A comparative study on the results of forecast is presented using these two data sources for present operational use. Slight improvement is noticed in rainfall, winds, geopotential heights and the vertical atmospheric structure while using NCEP data probably because of its high spatial and temporal resolution. (author)

  4. A high performance data parallel tensor contraction framework: Application to coupled electro-mechanics

    Science.gov (United States)

    Poya, Roman; Gil, Antonio J.; Ortigosa, Rogelio

    2017-07-01

    The paper presents aspects of implementation of a new high performance tensor contraction framework for the numerical analysis of coupled and multi-physics problems on streaming architectures. In addition to explicit SIMD instructions and smart expression templates, the framework introduces domain specific constructs for the tensor cross product and its associated algebra recently rediscovered by Bonet et al. (2015, 2016) in the context of solid mechanics. The two key ingredients of the presented expression template engine are as follows. First, the capability to mathematically transform complex chains of operations to simpler equivalent expressions, while potentially avoiding routes with higher levels of computational complexity and, second, to perform a compile time depth-first or breadth-first search to find the optimal contraction indices of a large tensor network in order to minimise the number of floating point operations. For optimisations of tensor contraction such as loop transformation, loop fusion and data locality optimisations, the framework relies heavily on compile time technologies rather than source-to-source translation or JIT techniques. Every aspect of the framework is examined through relevant performance benchmarks, including the impact of data parallelism on the performance of isomorphic and nonisomorphic tensor products, the FLOP and memory I/O optimality in the evaluation of tensor networks, the compilation cost and memory footprint of the framework and the performance of tensor cross product kernels. The framework is then applied to finite element analysis of coupled electro-mechanical problems to assess the speed-ups achieved in kernel-based numerical integration of complex electroelastic energy functionals. In this context, domain-aware expression templates combined with SIMD instructions are shown to provide a significant speed-up over the classical low-level style programming techniques.

  5. The Effects of Enzyme Complex on Performance, Intestinal Health and Nutrient Digestibility of Weaned Pigs

    Directory of Open Access Journals (Sweden)

    J. Q. Yi

    2013-08-01

    Full Text Available Two experiments were conducted to evaluate the effect of supplementing a corn-soybean meal-based diet with an enzyme complex containing amylase, protease and xylanase on the performance, intestinal health, apparent ileal digestibility of amino acids and nutrient digestibility of weaned pigs. In Exp. 1, 108 piglets weaned at 28 d of age were fed one of three diets containing 0 (control, 100, or 150 ppm enzyme complex for 4 wks, based on a two-phase feeding program namely 1 to 7 d (phase 1 and 8 to 28 d (phase 2. At the end of the experiment, six pigs from the control group and the group supplemented with 150 ppm enzyme complex were chosen to collect digesta samples from intestine to measure viscosity and pH in the stomach, ileum, and cecum, as well as volatile fatty acid concentrations and composition of the microflora in the cecum and colon. There were linear increases (p<0.01 in weight gain, gain: feed ratio and digestibility of gross energy with the increasing dose rate of enzyme supplementation during the whole experiment. Supplementation with enzyme complex increased the digesta viscosity in the stomach (p<0.05 and significantly increased (p<0.01 the concentrations of acetic, propionic and butyric acid in the cecum and colon. Enzyme supplementation also significantly increased the population of Lactobacilli (p<0.01 in the cecum and decreased the population of E. coli (p<0.05 in the colon. In Exp. 2, six crossbred barrows (initial body weight: 18.26±1.21 kg, fitted with a simple T-cannula at the distal ileum, were assigned to three dietary treatments according to a replicated 3×3 Latin Square design. The experimental diets were the same as the diets used in phase 2 in Exp. 1. Apparent ileal digestibility of isoleucine (p<0.01, valine (p<0.05 and aspartic acid (p<0.05 linearly increased with the increasing dose rate of enzyme supplementation. In conclusion, supplementation of the diet with an enzyme complex containing amylase, protease and

  6. JANUS: A Compilation System for Balancing Parallelism and Performance in OpenVX

    Science.gov (United States)

    Omidian, Hossein; Lemieux, Guy G. F.

    2018-04-01

    Embedded systems typically do not have enough on-chip memory for entire an image buffer. Programming systems like OpenCV operate on entire image frames at each step, making them use excessive memory bandwidth and power. In contrast, the paradigm used by OpenVX is much more efficient; it uses image tiling, and the compilation system is allowed to analyze and optimize the operation sequence, specified as a compute graph, before doing any pixel processing. In this work, we are building a compilation system for OpenVX that can analyze and optimize the compute graph to take advantage of parallel resources in many-core systems or FPGAs. Using a database of prewritten OpenVX kernels, it automatically adjusts the image tile size as well as using kernel duplication and coalescing to meet a defined area (resource) target, or to meet a specified throughput target. This allows a single compute graph to target implementations with a wide range of performance needs or capabilities, e.g. from handheld to datacenter, that use minimal resources and power to reach the performance target.

  7. Ruminant Nutrition Symposium: Improving cell wall digestion and animal performance with fibrolytic enzymes.

    Science.gov (United States)

    Adesogan, A T; Ma, Z X; Romero, J J; Arriola, K G

    2014-04-01

    This paper aimed to summarize published responses to treatment of cattle diets with exogenous fibrolytic enzymes (EFE), to discuss reasons for variable EFE efficacy in animal trials, to recommend strategies for improving enzyme testing and EFE efficacy in ruminant diets, and to identify proteomic differences between effective and ineffective EFE. A meta-analysis of 20 dairy cow studies with 30 experiments revealed that only a few increased lactational performance and the response was inconsistent. This variability is attributable to several enzyme, feed, animal, and management factors that were discussed in this paper. The variability reflects our limited understanding of the synergistic and sequential interactions between exogenous glycosyl hydrolases, autochthonous ruminal microbes, and endogenous fibrolytic enzymes that are necessary to optimize ruminal fiber digestion. An added complication is that many of the standard methods of assaying EFE activities may over- or underestimate their potential effects because they are based on pure substrate saccharification and do not simulate ruminal conditions. Our recent evaluation of 18 commercial EFE showed that 78 and 83% of them exhibited optimal endoglucanase and xylanase activities, respectively, at 50 °C, and 77 and 61% had optimal activities at pH 4 to 5, respectively, indicating that most would likely act suboptimally in the rumen. Of the many fibrolytic activities that act synergistically to degrade forage fiber, the few usually assayed, typically endoglucanase and xylanase, cannot hydrolyze the recalcitrant phenolic acid-lignin linkages that are the main constraints to ruminal fiber degradation. These factors highlight the futility of random addition of EFE to diets. This paper discusses reasons for the variable animal responses to dietary addition of fibrolytic enzymes, advances explanations for the inconsistency, suggests a strategy to improve enzyme efficacy in ruminant diets, and describes differences

  8. Implementation of a Monte Carlo simulation environment for fully 3D PET on a high-performance parallel platform

    CERN Document Server

    Zaidi, H; Morel, Christian

    1998-01-01

    This paper describes the implementation of the Eidolon Monte Carlo program designed to simulate fully three-dimensional (3D) cylindrical positron tomographs on a MIMD parallel architecture. The original code was written in Objective-C and developed under the NeXTSTEP development environment. Different steps involved in porting the software on a parallel architecture based on PowerPC 604 processors running under AIX 4.1 are presented. Basic aspects and strategies of running Monte Carlo calculations on parallel computers are described. A linear decrease of the computing time was achieved with the number of computing nodes. The improved time performances resulting from parallelisation of the Monte Carlo calculations makes it an attractive tool for modelling photon transport in 3D positron tomography. The parallelisation paradigm used in this work is independent from the chosen parallel architecture

  9. Study on High Performance of MPI-Based Parallel FDTD from WorkStation to Super Computer Platform

    Directory of Open Access Journals (Sweden)

    Z. L. He

    2012-01-01

    Full Text Available Parallel FDTD method is applied to analyze the electromagnetic problems of the electrically large targets on super computer. It is well known that the more the number of processors the less computing time consumed. Nevertheless, with the same number of processors, computing efficiency is affected by the scheme of the MPI virtual topology. Then, the influence of different virtual topology schemes on parallel performance of parallel FDTD is studied in detail. The general rules are presented on how to obtain the highest efficiency of parallel FDTD algorithm by optimizing MPI virtual topology. To show the validity of the presented method, several numerical results are given in the later part. Various comparisons are made and some useful conclusions are summarized.

  10. Performance characteristics of bioassay, radioenzymatic assay, homogeneous enzyme immunoassay, and high-performance liquid chromatographic determination of serum gentamicin

    International Nuclear Information System (INIS)

    Delaney, C.J.; Opheim, K.E.; Smith, A.L.; Plorde, J.J.

    1982-01-01

    We compared the accuracy, precision, and between-method error of the microbiological assay, the radioenzymatic assay, the homogeneous enzyme immunoassay, and the high-performance liquid chromatographic assay for the quantitation of gentamicin in serum. Precision and accuracy were evaluated by reference samples prepared to contain 0.0 to 32.7 micrograms of gentamicin per ml. Correlations between the methods utilized patient sera with gentamicin concentrations ranging from 0.6 to 13.3 micrograms/ml. All methods were reliable within acceptable limits for routine clinical use; intermethod correlation coefficients exceeded 0.96. Relative to the microbiological assay, the alternative methods offer the advantage of rapid analysis. The elapsed times for acquiring data on a set of 10 specimens under routine operating conditions were 0.5 h by the enzyme immunoassay, 4 h by the radioenzymatic assay, 5 h by the high-performance liquid chromatographic assay, and 10 h by the microbiological assay

  11. A parallel calibration utility for WRF-Hydro on high performance computers

    Science.gov (United States)

    Wang, J.; Wang, C.; Kotamarthi, V. R.

    2017-12-01

    A successful modeling of complex hydrological processes comprises establishing an integrated hydrological model which simulates the hydrological processes in each water regime, calibrates and validates the model performance based on observation data, and estimates the uncertainties from different sources especially those associated with parameters. Such a model system requires large computing resources and often have to be run on High Performance Computers (HPC). The recently developed WRF-Hydro modeling system provides a significant advancement in the capability to simulate regional water cycles more completely. The WRF-Hydro model has a large range of parameters such as those in the input table files — GENPARM.TBL, SOILPARM.TBL and CHANPARM.TBL — and several distributed scaling factors such as OVROUGHRTFAC. These parameters affect the behavior and outputs of the model and thus may need to be calibrated against the observations in order to obtain a good modeling performance. Having a parameter calibration tool specifically for automate calibration and uncertainty estimates of WRF-Hydro model can provide significant convenience for the modeling community. In this study, we developed a customized tool using the parallel version of the model-independent parameter estimation and uncertainty analysis tool, PEST, to enabled it to run on HPC with PBS and SLURM workload manager and job scheduler. We also developed a series of PEST input file templates that are specifically for WRF-Hydro model calibration and uncertainty analysis. Here we will present a flood case study occurred in April 2013 over Midwest. The sensitivity and uncertainties are analyzed using the customized PEST tool we developed.

  12. Study Behaviors and USMLE Step 1 Performance: Implications of a Student Self-Directed Parallel Curriculum.

    Science.gov (United States)

    Burk-Rafel, Jesse; Santen, Sally A; Purkiss, Joel

    2017-11-01

    To determine medical students' study behaviors when preparing for the United States Medical Licensing Examination (USMLE) Step 1, and how these behaviors are associated with Step 1 scores when controlling for likely covariates. The authors distributed a study-behaviors survey in 2014 and 2015 at their institution to two cohorts of medical students who had recently taken Step 1. Demographic and academic data were linked to responses. Descriptive statistics, bivariate correlations, and multiple linear regression analyses were performed. Of 332 medical students, 274 (82.5%) participated. Most students (n = 211; 77.0%) began studying for Step 1 during their preclinical curriculum, increasing their intensity during a protected study period during which they averaged 11.0 hours studying per day (standard deviation [SD] 2.1) over a period of 35.3 days (SD 6.2). Students used numerous third-party resources, including reading an exam-specific 700-page review book on average 2.1 times (SD 0.8) and completing an average of 3,597 practice multiple-choice questions (SD 1,611). Initiating study prior to the designated study period, increased review book usage, and attempting more practice questions were all associated with higher Step 1 scores, even when controlling for Medical College Admission Test scores, preclinical exam performance, and self-identified score goal (adjusted R = 0.56, P < .001). Medical students at one public institution engaged in a self-directed, "parallel" Step 1 curriculum using third-party study resources. Several study behaviors were associated with improved USMLE Step 1 performance, informing both institutional- and student-directed preparation for this high-stakes exam.

  13. Cpl6: The New Extensible, High-Performance Parallel Coupler forthe Community Climate System Model

    Energy Technology Data Exchange (ETDEWEB)

    Craig, Anthony P.; Jacob, Robert L.; Kauffman, Brain; Bettge,Tom; Larson, Jay; Ong, Everest; Ding, Chris; He, Yun

    2005-03-24

    Coupled climate models are large, multiphysics applications designed to simulate the Earth's climate and predict the response of the climate to any changes in the forcing or boundary conditions. The Community Climate System Model (CCSM) is a widely used state-of-art climate model that has released several versions to the climate community over the past ten years. Like many climate models, CCSM employs a coupler, a functional unit that coordinates the exchange of data between parts of climate system such as the atmosphere and ocean. This paper describes the new coupler, cpl6, contained in the latest version of CCSM,CCSM3. Cpl6 introduces distributed-memory parallelism to the coupler, a class library for important coupler functions, and a standardized interface for component models. Cpl6 is implemented entirely in Fortran90 and uses Model Coupling Toolkit as the base for most of its classes. Cpl6 gives improved performance over previous versions and scales well on multiple platforms.

  14. Effect of Cereal Type and Enzyme Addition on Performance, Pancreatic Enzyme Activity, Intestinal Microflora and Gut Morphology of Broilers

    Directory of Open Access Journals (Sweden)

    Kalantar M

    2016-06-01

    Full Text Available The effects of grain and carbohydrase enzyme supplementation were investigated on digestive physiology of chickens. A total of 625 one-day-old chicks (Ross 308 were randomly assigned to five treatments in a completely randomized design. Treatments included two different types of grains (wheat, and barley with or without a multi-carbohydrase supplement. A corn-based diet was also considered to serve as a control. Feeding barley-based diet with multi-carbohydrase led to higher feed intake (P < 0.01 than those fed corn- and wheat-based diets. Birds fed on barley and wheat diets had lower weight gain despite a higher feed conversion ratio (P < 0.01. Total count and number of different type of bacteria including Gram-negative, E. coli, and Clostridia increased after feeding wheat and barley but the number of Lactobacilli and Bifidobacteria decreased (P < 0.01. Feeding barley and wheat diets reduced villus height in different parts of the small intestine when compared to those fed on a corn diet. However, enzyme supplementation of barley and wheat diets improved weight gain and feed conversion ratio and resulted in reduced number of E. coli and Clostridia and increased number of Lactobacilli and Bifidobacteria, and also restored the negative effects on intestinal villi height (P < 0.01. The activities of pancreatic α-amylase and lipase were (P < 0.01 increased in chickens fed wheat and barley diets when compared to the control fed on a corn diet. Enzyme supplementation reduced the activities of pancreatic α-amylase and lipase (P < 0.01. In conclusion, various dietary non-starch polysaccharides without enzyme supplementation have an adverse effect on digesta viscosity, ileal microflora, villi morphology, and pancreatic enzyme activity.

  15. Performance evaluations of advanced massively parallel platforms based on gyrokinetic toroidal five-dimensional Eulerian code GT5D

    International Nuclear Information System (INIS)

    Idomura, Yasuhiro; Jolliet, Sebastien

    2010-01-01

    A gyrokinetic toroidal five dimensional Eulerian code GT5D is ported on six advanced massively parallel platforms and comprehensive benchmark tests are performed. A parallelisation technique based on physical properties of the gyrokinetic equation is presented. By extending the parallelisation technique with a hybrid parallel model, the scalability of the code is improved on platforms with multi-core processors. In the benchmark tests, a good salability is confirmed up to several thousands cores on every platforms, and the maximum sustained performance of ∼18.6 Tflops is achieved using 16384 cores of BX900. (author)

  16. Performance of a parallel plate volume cell prototype for a fast iron/gas calorimeter

    International Nuclear Information System (INIS)

    Bizzeti, A.; Civinini, C.; D'alessandro, R.; Ferrando, A.

    1993-01-01

    We present the first test of the application of the parallel plate chamber principles for the design of a very fast and radiation-hard iron/gas sampling calorimeter, suitable for very forward regions in detectors for LBC; based on the use of thick iron plates as electrodes. We have built a one cell prototype consisting of three parallel thick iron plates (117 mn each). Results on efficiencies and mean collected charge for minimum ionizing particles with different gases are presented. (Author)

  17. Evaluation of thermal performance of all-GaN power module in parallel operation

    International Nuclear Information System (INIS)

    Chou, Po-Chien; Cheng, Stone; Chen, Szu-Hao

    2014-01-01

    This work presents an extensive thermal characterization of a single discrete GaN high-electron-mobility transistor (HEMT) device when operated in parallel at temperatures of 25 °C–175 °C. The maximum drain current (I D max ), on-resistance (R ON ), pinch-off voltage (V P ) and peak transconductance (g m ) at various chamber temperatures are measured and correlations among these parameters studied. Understanding the dependence of key transistor parameters on temperature is crucial to inhibiting the generation of hot spots and the equalization of currents in the parallel operation of HEMTs. A detailed analysis of the current imbalance between two parallel HEMT cells and its consequential effect on the junction temperature are also presented. The results from variations in the characteristics of the parallel-connected devices further verify that the thermal stability and switching behavior of these cells are balanced. Two parallel HEMT cells are operated at a safe working distance from thermal runaway to prevent destruction of the hottest cell. - Highlights: • This work reveals the sorting process of GaN devices for parallel operation. • The variations of I D max , R ON , V P , and g m with temperature are established. • The temperature-dependence parameters are crucial to prevent hot spots generation. • Safe working operation prevents thermal runaway and hottest cell destruction

  18. Effects of Maize Source and Complex Enzymes on Performance and Nutrient Utilization of Broilers

    Directory of Open Access Journals (Sweden)

    Defu Tang

    2014-12-01

    Full Text Available The objective of this study was to investigate the effect of maize source and complex enzymes containing amylase, xylanase and protease on performance and nutrient utilization of broilers. The experiment was a 4×3 factorial design with diets containing four source maize samples (M1, M2, M3, and M4 and without or with two kinds of complex enzyme A (Axtra XAP and B (Avizyme 1502. Nine hundred and sixty day old Arbor Acres broiler chicks were used in the trial (12 treatments with 8 replicate pens of 10 chicks. Birds fed M1 diet had better body weight gain (BWG and lower feed/gain ratio compared with those fed M3 diet and M4 diet (p0.05, respectively. The fresh feces output was significantly decreased by the addition of enzyme B (p<0.05. Maize source affects the nutrients digestibility and performance of broilers, and a combination of amylase, xylanase and protease is effective in improving the growth profiles of broilers fed maize-soybean-rapeseed-cotton mixed diets.

  19. Synergistic Enhancement of Enzyme Performance and Resilience via Orthogonal Peptide-Protein Chemistry Enabled Multilayer Construction.

    Science.gov (United States)

    Zhang, Xue-Jian; Wang, Xiao-Wei; Sun, Jiaxing; Su, Chao; Yang, Shuguang; Zhang, Wen-Bin

    2018-05-16

    Protein immobilization is critical to utilize their unique functions in diverse applications. Herein, we report that orthogonal peptide-protein chemistry enabled multilayer construction can facilitate the incorporation of various folded structural domains, including calmodulin in different states, affibody and dihydrofolate reductase (DHFR). An extended conformation is found to be the most advantageous for steady film growth. The resulting protein thin films exhibit sensitive and selective responsive behaviors to bio-signals (Ca2+, TFP, NADPH, etc.) and fully maintain the catalytic activity of DHFR. The approach is applicable to different substrates such as hydrophobic gold and hydrophilic silica microparticles. The DHFR enzyme can be immobilized onto silica microparticles with tunable amounts. The multi-layer set-up exhibits a synergistic enhancement of DHFR activity with increasing number of bilayers and also makes the embedded DHFR more resilient to lyophilization. Therefore, this is a convenient and versatile method for protein immobilization with potential benefits of synergistic enhancement in enzyme performance and resilience.

  20. Performance evaluation of parallel electric field tunnel field-effect transistor by a distributed-element circuit model

    Science.gov (United States)

    Morita, Yukinori; Mori, Takahiro; Migita, Shinji; Mizubayashi, Wataru; Tanabe, Akihito; Fukuda, Koichi; Matsukawa, Takashi; Endo, Kazuhiko; O'uchi, Shin-ichi; Liu, Yongxun; Masahara, Meishoku; Ota, Hiroyuki

    2014-12-01

    The performance of parallel electric field tunnel field-effect transistors (TFETs), in which band-to-band tunneling (BTBT) was initiated in-line to the gate electric field was evaluated. The TFET was fabricated by inserting an epitaxially-grown parallel-plate tunnel capacitor between heavily doped source wells and gate insulators. Analysis using a distributed-element circuit model indicated there should be a limit of the drain current caused by the self-voltage-drop effect in the ultrathin channel layer.

  1. Development of three-dimensional neoclassical transport simulation code with high performance Fortran on a vector-parallel computer

    International Nuclear Information System (INIS)

    Satake, Shinsuke; Okamoto, Masao; Nakajima, Noriyoshi; Takamaru, Hisanori

    2005-11-01

    A neoclassical transport simulation code (FORTEC-3D) applicable to three-dimensional configurations has been developed using High Performance Fortran (HPF). Adoption of computing techniques for parallelization and a hybrid simulation model to the δf Monte-Carlo method transport simulation, including non-local transport effects in three-dimensional configurations, makes it possible to simulate the dynamism of global, non-local transport phenomena with a self-consistent radial electric field within a reasonable computation time. In this paper, development of the transport code using HPF is reported. Optimization techniques in order to achieve both high vectorization and parallelization efficiency, adoption of a parallel random number generator, and also benchmark results, are shown. (author)

  2. Performance of a plasma fluid code on the Intel parallel computers

    International Nuclear Information System (INIS)

    Lynch, V.E.; Carreras, B.A.; Drake, J.B.; Leboeuf, J.N.; Liewer, P.

    1992-01-01

    One approach to improving the real-time efficiency of plasma turbulence calculations is to use a parallel algorithm. A parallel algorithm for plasma turbulence calculations was tested on the Intel iPSC/860 hypercube and the Touchtone Delta machine. Using the 128 processors of the Intel iPSC/860 hypercube, a factor of 5 improvement over a single-processor CRAY-2 is obtained. For the Touchtone Delta machine, the corresponding improvement factor is 16. For plasma edge turbulence calculations, an extrapolation of the present results to the Intel σ machine gives an improvement factor close to 64 over the single-processor CRAY-2

  3. Performance of a plasma fluid code on the Intel parallel computers

    International Nuclear Information System (INIS)

    Lynch, V.E.; Carreras, B.A.; Drake, J.B.; Leboeuf, J.N.; Liewer, P.

    1992-01-01

    One approach to improving the real-time efficiency of plasma turbulence calculations is to use a parallel algorithm. A parallel algorithm for plasma turbulence calculations was tested on the Intel iPSC/860 hypercube and the Touchtone Delta machine. Using the 128 processors of the Intel iPSC/860 hypercube, a factor of 5 improvement over a single-processor CRAY-2 is obtained. For the Touchtone Delta machine, the corresponding improvement factor is 16. For plasma edge turbulence calculations, an extrapolation of the present results to the Intel (sigma) machine gives an improvement factor close to 64 over the single-processor CRAY-2. 12 refs

  4. Performance of a plasma fluid code on the Intel parallel computers

    Science.gov (United States)

    Lynch, V. E.; Carreras, B. A.; Drake, J. B.; Leboeuf, J. N.; Liewer, P.

    1992-01-01

    One approach to improving the real-time efficiency of plasma turbulence calculations is to use a parallel algorithm. A parallel algorithm for plasma turbulence calculations was tested on the Intel iPSC/860 hypercube and the Touchtone Delta machine. Using the 128 processors of the Intel iPSC/860 hypercube, a factor of 5 improvement over a single-processor CRAY-2 is obtained. For the Touchtone Delta machine, the corresponding improvement factor is 16. For plasma edge turbulence calculations, an extrapolation of the present results to the Intel (sigma) machine gives an improvement factor close to 64 over the single-processor CRAY-2.

  5. Performance of a parallel plate volume cell prototype for a fast iron/gas calorimeter

    Energy Technology Data Exchange (ETDEWEB)

    Bizzeti, A.; Civinini, C.; D' Alessandro, R.; Ferrando, A.; Malinin, A.; Martinez-Laso, L.; Pojidaev, V.

    1993-07-01

    We present the first test of the application of the parallel plate chamber principles for the design of a very fast and radiation-hard iron/gas sampling calorimeter, suitable for very forward regions in detectors for LHC, based on the use of thick iron plates as electrodes. We have built a one cell prototype consisting of three parallel thick iron plates (17 mm each). Results on efficiencies and mean collected charge for minimum ionizing particles with different gases are presented. (Author) 7 refs.

  6. Performance of a parallel plate volume cell prototype for a fast iron/gas calorimeter

    International Nuclear Information System (INIS)

    Bizzeti, A.; Civinini, C.; D'Alessandro, R.; Ferrando, A.; Malinin, A.; Martinez-Laso, L.; Pojidaev, V.

    1993-01-01

    We present the first test of the application of the parallel plate chamber principles for the design of a very fast and radiation-hard iron/gas sampling calorimeter, suitable for very forward regions in detectors for LHC, based on the use of thick iron plates as electrodes. We have built a one cell prototype consisting of three parallel thick iron plates (17 mm each). Results on efficiencies and mean collected charge for minimum ionizing particles with different gases are presented. (Author) 7 refs

  7. EFFECTS OF EXOGENOUS ENZYMES ON NUTRIENTS DIGESTIBILITY AND GROWTH PERFORMANCE IN SHEEP AND GOATS

    Directory of Open Access Journals (Sweden)

    Abdel-Fattah Z.M. Salem

    2011-07-01

    Full Text Available Six crossbred sheep (32.00±0.603 kg BW and 6 Baladi goats (18.00±0.703 kg BW were used in 2×2 factorial design to evaluate the effect of exogenous enzymes of ZADO® (i.e., ENZ and on digestibility and growth performance. Animals were fed on wheat straw ad libitum and restricted amount of commercial concentrate with (+ENZ or without (-ENZ 10 g/animal/day of ZADO to cover 120% of their maintenance requirements. Nutrients digestibilities were increased (P

  8. Probe colorimeter for quantitating enzyme-linked immunosorbent assays and other colorimetric assays performed with microplates.

    Science.gov (United States)

    Ackerman, S B; Kelley, E A

    1983-03-01

    The performance of a fiberoptic probe colorimeter (model PC800; Brinkmann Instruments, Inc., Westbury, N.Y.) for quantitating enzymatic or colorimetric assays in 96-well microtiter plates was compared with the performances of a spectrophotometer (model 240; Gilford Instrument Laboratories, Inc., Oberlin, Ohio) and a commercially available enzyme immunoassay reader (model MR590; Dynatech Laboratories, Inc., Alexandria, Va.). Alkaline phosphatase-p-nitrophenyl phosphate in 3 M NaOH was used as the chromophore source. Six types of plates were evaluated for use with the probe colorimeter; they generated reproducibility values (100% coefficient of variation) ranging from 91 to 98% when one individual made 24 independent measurements on the same dilution of chromophore on each plate. Eleven individuals each performed 24 measurements with the colorimeter on either a visually light (absorbance of 0.10 at 420 nm) or a dark (absorbance of 0.80 at 420 nm) dilution of chromophore; reproducibilities averaged 87% for the light dilution and 97% for the dark dilution. When one individual measured the same chromophore sample at least 20 times in the colorimeter, in the spectrophotometer or in the enzyme immunoassay reader, reproducibility for each instrument was greater than 99%. Measurements of a dilution series of chromophore in a fixed volume indicated that the optical responses of each instrument were linear in a range of 0.05 to 1.10 absorbance units.

  9. The boat hull model : adapting the roofline model to enable performance prediction for parallel computing

    NARCIS (Netherlands)

    Nugteren, C.; Corporaal, H.

    2012-01-01

    Multi-core and many-core were already major trends for the past six years, and are expected to continue for the next decades. With these trends of parallel computing, it becomes increasingly difficult to decide on which architecture to run a given application. In this work, we use an algorithm

  10. Performance analysis of parallel identical machines with a generalized shortest queue arrival mechanism

    NARCIS (Netherlands)

    van Houtum, Geert-Jan; Adan, I.J.B.F.; Wessels, J.; Zijm, Willem H.M.

    In this paper we study a production system consisting of a group of parallel machines producing multiple job types. Each machine has its own queue and it can process a restricted set of job types only. On arrival a job joins the shortest queue among all queues capable of serving that job. Under the

  11. The boat hull model : enabling performance prediction for parallel computing prior to code development

    NARCIS (Netherlands)

    Nugteren, C.; Corporaal, H.

    2012-01-01

    Multi-core and many-core were already major trends for the past six years and are expected to continue for the next decade. With these trends of parallel computing, it becomes increasingly difficult to decide on which processor to run a given application, mainly because the programming of these

  12. Design and performance characterization of electronic structure calculations on massively parallel supercomputers

    DEFF Research Database (Denmark)

    Romero, N. A.; Glinsvad, Christian; Larsen, Ask Hjorth

    2013-01-01

    Density function theory (DFT) is the most widely employed electronic structure method because of its favorable scaling with system size and accuracy for a broad range of molecular and condensed-phase systems. The advent of massively parallel supercomputers has enhanced the scientific community...

  13. High Performance Parallel Processing Project: Industrial computing initiative. Progress reports for fiscal year 1995

    Energy Technology Data Exchange (ETDEWEB)

    Koniges, A.

    1996-02-09

    This project is a package of 11 individual CRADA`s plus hardware. This innovative project established a three-year multi-party collaboration that is significantly accelerating the availability of commercial massively parallel processing computing software technology to U.S. government, academic, and industrial end-users. This report contains individual presentations from nine principal investigators along with overall program information.

  14. Very Large-Scale Neighborhoods with Performance Guarantees for Minimizing Makespan on Parallel Machines

    NARCIS (Netherlands)

    Brueggemann, T.; Hurink, Johann L.; Vredeveld, T.; Woeginger, Gerhard

    2006-01-01

    We study the problem of minimizing the makespan on m parallel machines. We introduce a very large-scale neighborhood of exponential size (in the number of machines) that is based on a matching in a complete graph. The idea is to partition the jobs assigned to the same machine into two sets. This

  15. A concurrent visualization system for large-scale unsteady simulations. Parallel vector performance on an NEC SX-4

    International Nuclear Information System (INIS)

    Takei, Toshifumi; Doi, Shun; Matsumoto, Hideki; Muramatsu, Kazuhiro

    2000-01-01

    We have developed a concurrent visualization system RVSLIB (Real-time Visual Simulation Library). This paper shows the effectiveness of the system when it is applied to large-scale unsteady simulations, for which the conventional post-processing approach may no longer work, on high-performance parallel vector supercomputers. The system performs almost all of the visualization tasks on a computation server and uses compressed visualized image data for efficient communication between the server and the user terminal. We have introduced several techniques, including vectorization and parallelization, into the system to minimize the computational costs of the visualization tools. The performance of RVSLIB was evaluated by using an actual CFD code on an NEC SX-4. The computational time increase due to the concurrent visualization was at most 3% for a smaller (1.6 million) grid and less than 1% for a larger (6.2 million) one. (author)

  16. Fast magnetic resonance imaging of the knee using a parallel acquisition technique (mSENSE): a prospective performance evaluation

    International Nuclear Information System (INIS)

    Kreitner, K.F.; Romaneehsen, Bernd; Oberholzer, Katja; Dueber, Christoph; Krummenauer, Frank; Mueller, L.P.

    2006-01-01

    The performance of a magnetic resonance (MR) imaging strategy that uses multiple receiver coil elements and integrated parallel imaging techniques (iPAT) in traumatic and degenerative disorders of the knee and to compare this technique with a standard MR imaging protocol was evaluated. Ninety patients with suspected internal derangements of the knee joint prospectively underwent MR imaging at 1.5 T. For signal detection, a 6-channel array coil was used. All patients were investigated with a standard imaging protocol consisting of different turbo spin-echo sequences proton density (PD), T 2 -weighted turbo spin echo (TSE) with and without fat suppression in three imaging planes. All sequences were repeated with an integrated parallel acquisition technique (iPAT) using the modified sensitivity encoding (mSENSE) algorithm with an acceleration factor of 2. Two radiologists independently evaluated and scored all images with regard to overall image quality, artefacts and pathologic findings. Agreement of the parallel ratings between readers and imaging techniques, respectively, was evaluated by means of pairwise kappa coefficients that were stratified for the area of evaluation. Agreement between the parallel readers for both the iPAT imaging and the conventional technique, respectively, as well as between imaging techniques was found encouraging with inter-observer kappa values ranging between 0.78 and 0.98 for both imaging techniques, and the inter-method kappa values ranging between 0.88 and 1.00 for both clinical readers. All pathological findings (e.g. occult fractures, meniscal and cruciate ligament tears, torn and interpositioned Hoffa's cleft, cartilage damage) were detected by both techniques with comparable performance. The use of iPAT lead to a 48% reduction of acquisition time compared with standard technique. Parallel imaging using mSENSE proved to be an efficient and economic tool for fast musculoskeletal MR imaging of the knee joint with comparable

  17. Scalable High-Performance Parallel Design for Network Intrusion Detection Systems on Many-Core Processors

    OpenAIRE

    Jiang, Hayang; Xie, Gaogang; Salamatian, Kavé; Mathy, Laurent

    2013-01-01

    Network Intrusion Detection Systems (NIDSes) face significant challenges coming from the relentless network link speed growth and increasing complexity of threats. Both hardware accelerated and parallel software-based NIDS solutions, based on commodity multi-core and GPU processors, have been proposed to overcome these challenges. Network Intrusion Detection Systems (NIDSes) face significant challenges coming from the relentless network link speed growth and increasing complexity of threats. ...

  18. Parallel segmented outlet flow high performance liquid chromatography with multiplexed detection

    International Nuclear Information System (INIS)

    Camenzuli, Michelle; Terry, Jessica M.; Shalliker, R. Andrew; Conlan, Xavier A.; Barnett, Neil W.; Francis, Paul S.

    2013-01-01

    Graphical abstract: -- Highlights: •Multiplexed detection for liquid chromatography. •‘Parallel segmented outlet flow’ distributes inner and outer portions of the analyte zone. •Three detectors were used simultaneously for the determination of opiate alkaloids. -- Abstract: We describe a new approach to multiplex detection for HPLC, exploiting parallel segmented outlet flow – a new column technology that provides pressure-regulated control of eluate flow through multiple outlet channels, which minimises the additional dead volume associated with conventional post-column flow splitting. Using three detectors: one UV-absorbance and two chemiluminescence systems (tris(2,2′-bipyridine)ruthenium(III) and permanganate), we examine the relative responses for six opium poppy (Papaver somniferum) alkaloids under conventional and multiplexed conditions, where approximately 30% of the eluate was distributed to each detector and the remaining solution directed to a collection vessel. The parallel segmented outlet flow mode of operation offers advantages in terms of solvent consumption, waste generation, total analysis time and solute band volume when applying multiple detectors to HPLC, but the manner in which each detection system is influenced by changes in solute concentration and solution flow rates must be carefully considered

  19. Parallel segmented outlet flow high performance liquid chromatography with multiplexed detection

    Energy Technology Data Exchange (ETDEWEB)

    Camenzuli, Michelle [Australian Centre for Research on Separation Science (ACROSS), School of Science and Health, University of Western Sydney (Parramatta), Sydney, NSW (Australia); Terry, Jessica M. [Centre for Chemistry and Biotechnology, School of Life and Environmental Sciences, Deakin University, Geelong, Victoria 3216 (Australia); Shalliker, R. Andrew, E-mail: r.shalliker@uws.edu.au [Australian Centre for Research on Separation Science (ACROSS), School of Science and Health, University of Western Sydney (Parramatta), Sydney, NSW (Australia); Conlan, Xavier A.; Barnett, Neil W. [Centre for Chemistry and Biotechnology, School of Life and Environmental Sciences, Deakin University, Geelong, Victoria 3216 (Australia); Francis, Paul S., E-mail: paul.francis@deakin.edu.au [Centre for Chemistry and Biotechnology, School of Life and Environmental Sciences, Deakin University, Geelong, Victoria 3216 (Australia)

    2013-11-25

    Graphical abstract: -- Highlights: •Multiplexed detection for liquid chromatography. •‘Parallel segmented outlet flow’ distributes inner and outer portions of the analyte zone. •Three detectors were used simultaneously for the determination of opiate alkaloids. -- Abstract: We describe a new approach to multiplex detection for HPLC, exploiting parallel segmented outlet flow – a new column technology that provides pressure-regulated control of eluate flow through multiple outlet channels, which minimises the additional dead volume associated with conventional post-column flow splitting. Using three detectors: one UV-absorbance and two chemiluminescence systems (tris(2,2′-bipyridine)ruthenium(III) and permanganate), we examine the relative responses for six opium poppy (Papaver somniferum) alkaloids under conventional and multiplexed conditions, where approximately 30% of the eluate was distributed to each detector and the remaining solution directed to a collection vessel. The parallel segmented outlet flow mode of operation offers advantages in terms of solvent consumption, waste generation, total analysis time and solute band volume when applying multiple detectors to HPLC, but the manner in which each detection system is influenced by changes in solute concentration and solution flow rates must be carefully considered.

  20. New treatment strategy against advanced rectal cancer. Enzyme-targeting and radio-sensitization treatment under parallel use of TS-1

    International Nuclear Information System (INIS)

    Obata, Shiro; Yamanishi, Mikio; Katsumi, Shingo

    2015-01-01

    Preoperative chemoradiotherapy was applied to two cases of advanced rectal cancer. In addition, radiation sensitizers were injected to the lesion endoscopically at a pace of twice a week in order to enhance therapeutic effects (so-called enzyme-targeting and radio-sensitization treatment: KORTUC [Kochi Oxydol Radio-sensitization Treatment for Unresectable Carcinomas]). The flattening of the lesion shape was observed for both cases in a short period of time, then, Mile's and lateral lymphnode dissection was performed. The remnant of lesion was not pointed out in postoperative pathological specimens for both cases, and histological judgment after the treatment was ranked as Grade 3. In light of the better-than-expected results, this hospital is preparing for clinical trials, and planning to carefully accumulate the cases. As one of the curative treatment strategies against advanced rectal cancer, the authors are willing to make this KORTUC more objectively reliable as a safe and minimally invasive therapy. (A.O.)

  1. A double blind, randomised, parallel group study on the efficacy and safety of treating acute lateral ankle sprain with oral hydrolytic enzymes

    NARCIS (Netherlands)

    Kerkhoffs, G. M. M. J.; Struijs, P. A. A.; de Wit, C.; Rahlfs, V. W.; Zwipp, H.; van Dijk, C. N.

    2004-01-01

    Objective: To compare the effectiveness and safety of the triple combination Phlogenzym ( rutoside, bromelain, and trypsin) with double combinations, the single substances, and placebo. Design: Multinational, multicentre, double blind, randomised, parallel group design with eight groups structured

  2. High performance computing of density matrix renormalization group method for 2-dimensional model. Parallelization strategy toward peta computing

    International Nuclear Information System (INIS)

    Yamada, Susumu; Igarashi, Ryo; Machida, Masahiko; Imamura, Toshiyuki; Okumura, Masahiko; Onishi, Hiroaki

    2010-01-01

    We parallelize the density matrix renormalization group (DMRG) method, which is a ground-state solver for one-dimensional quantum lattice systems. The parallelization allows us to extend the applicable range of the DMRG to n-leg ladders i.e., quasi two-dimension cases. Such an extension is regarded to bring about several breakthroughs in e.g., quantum-physics, chemistry, and nano-engineering. However, the straightforward parallelization requires all-to-all communications between all processes which are unsuitable for multi-core systems, which is a mainstream of current parallel computers. Therefore, we optimize the all-to-all communications by the following two steps. The first one is the elimination of the communications between all processes by only rearranging data distribution with the communication data amount kept. The second one is the avoidance of the communication conflict by rescheduling the calculation and the communication. We evaluate the performance of the DMRG method on multi-core supercomputers and confirm that our two-steps tuning is quite effective. (author)

  3. Effects of treating sorghum wet distillers grains with solubles with fibrolytic enzymes on nutrient digestibility and performance in finishing beef steers

    Science.gov (United States)

    Two experiments were conducted to determine the effects of treating sorghum WDG with solubles (SWDG) with an enzyme, or enzyme-buffer combination on diet digestibility and feedlot performance. Experimental treatments are; 1) untreated SWDG (control), 2) addition of an enzyme complex to SWDG (enzyme...

  4. Performance of optical biosensor using alcohol oxidase enzyme for formaldehyde detection

    Science.gov (United States)

    Sari, A. P.; Rachim, A.; Nurlely, Fauzia, V.

    2017-07-01

    The recent issue in the world is the long exposure of formaldehyde which is can increase the risk of human health, therefore, that is very important to develop a device and method that can be optimized to detect the formaldehyde elements accurately, have a long lifetime and can be fabricated and produced in large quantities. A new and simple prepared optical biosensor for detection of formaldehyde in aqueous solutions using alcohol oxidase (AOX) enzyme was successfully fabricated. The poly-n-butyl acrylic-co-N-acryloxysuccinimide (nBA-NAS) membranes containing chromoionophore ETH5294 were used for immobilization of alcohol oxidase enzyme (AOX). Biosensor response was based on the colour change of chromoionophore as a result of enzymatic oxidation of formaldehyde and correlated with the detection concentration of formaldehyde. The performance of biosensor parameters were measured through the optical absorption value using UV-Vis spectrophotometer including the repeatability, reproducibility, selectivity and lifetime. The results showed that the prepared biosensor has good repeatability (RSD = 1.9 %) and good reproducibility (RSD = 2.1 %). The biosensor was selective formaldehyde with no disturbance by methanol, ethanol, and acetaldehyde, and also stable before 49 days and decrease by 41.77 % after 49 days.

  5. Cu₂O-Au nanocomposites for enzyme-free glucose sensing with enhanced performances.

    Science.gov (United States)

    Hu, Qiyan; Wang, Fenyun; Fang, Zhen; Liu, Xiaowang

    2012-06-15

    A facile method for the synthesis of Cu(2)O-Au nanocomposites has been reported by injecting Cu(2)O nanocubes into Au precursor directly with the assistance of ultrasound radiation at room temperature. The ultrasound radiation is not a necessary requirement but can make the distribution of Au nanoparticles more homogenous. The formation of Cu(2)O-Au nanocomposites is attributed to following two reasons. The first one is the difference in the reduction potential between Cu(2+)/Cu(2)O and AuCl(4)(-)/Au, which can also be considered as the driving force for the redox reaction. The other one is the low lattice mismatch between (200) planes of Cu(2)O and (200) facets of Au, which is favorable for the formation of heterostructure. The electrochemical investigation demonstrates that the performances of Cu(2)O nanocubes in enzyme-free glucose sensing have been improved significantly after the decoration of Au nanoparticles which may be derived from the polarization effect provided by Au nanoparticles. As-prepared Cu(2)O-Au nanocomposites have great potential in enzyme-free glucose sensing. Copyright © 2012 Elsevier B.V. All rights reserved.

  6. Effect of Enzyme Supplementation and Irradiation of Barley on Broiler Chicks Performance

    International Nuclear Information System (INIS)

    Farag, D.H.M.; Abd El-Hakeim, N.F.

    1999-01-01

    The experiments were conducted to study the influence of irradiation treatment at dose levels of 0.20 and 60 kGy on barley beta-glucan and the effect of enzyme supplementation and irradiation of barley on broiler chicks performance. The amount of total and water-soluble beta-glucan in raw barley was 36 kg -1 , respectively. The effect of irradiation treatment on total beta-glucan was insignificant while the level of soluble beta-glucan was increased with increasing the dose levels of irradiation. The effect of irradiation treatment and enzyme supplementation of barley diets on growth and conversion performance of broiler chicks indicated that birds fed raw barley diet had lower body weight, body weight gain and feed conversion than those fed control diet throughout the experimental period. Irradiation of barley at dose of 20 kGy did not affect the chick performance (feed consumption, weight gain feed-gain ratio) that received the B 20 diet from 7 to 21 days of age, but when bird maintained on B 20 diet from 7 28 days of age, only feed-gain ratio was improved by 14.4%. The results indicate that there was a significant effect of irradiation of barley at 60 kGy (B 60) on feed -gain ratio of chicks when were fed B 60 diet from 7 to 21 days of age. The corresponding improvement in feed-gain ratio was 16.4%. When birds were fed B 60 diet from 7-28 days of age, the improvement in body weight and feed-gain ratio was 25.5 and 19.6%, respectively

  7. High Performance Computation of a Jet in Crossflow by Lattice Boltzmann Based Parallel Direct Numerical Simulation

    Directory of Open Access Journals (Sweden)

    Jiang Lei

    2015-01-01

    Full Text Available Direct numerical simulation (DNS of a round jet in crossflow based on lattice Boltzmann method (LBM is carried out on multi-GPU cluster. Data parallel SIMT (single instruction multiple thread characteristic of GPU matches the parallelism of LBM well, which leads to the high efficiency of GPU on the LBM solver. With present GPU settings (6 Nvidia Tesla K20M, the present DNS simulation can be completed in several hours. A grid system of 1.5 × 108 is adopted and largest jet Reynolds number reaches 3000. The jet-to-free-stream velocity ratio is set as 3.3. The jet is orthogonal to the mainstream flow direction. The validated code shows good agreement with experiments. Vortical structures of CRVP, shear-layer vortices and horseshoe vortices, are presented and analyzed based on velocity fields and vorticity distributions. Turbulent statistical quantities of Reynolds stress are also displayed. Coherent structures are revealed in a very fine resolution based on the second invariant of the velocity gradients.

  8. Improving the performance of electrochemical microsensors based on enzymes entrapped in a redox hydrogel

    International Nuclear Information System (INIS)

    Mitala, J.J.; Michael, A.C.

    2006-01-01

    Microsensors based on carbon fiber microelectrodes coated with enzyme-entrapping redox hydrogels facilitate the in vivo detection of substances of interest within the central nervous system, including hydrogen peroxide, glucose, choline and glutamate. The hydrogel, formed by cross-linking a redox polymer, entraps the enzymes and mediates electron transfer between the enzymes and the electrode. It is important that the enzymes are entrapped in their enzymatically active state. Should entrapment cause enzyme denaturation, the sensitivity and the selectivity of the sensor may be compromised. Synthesis of the redox polymer according to published procedures may yield a product that precipitates when added to aqueous enzyme solutions. Casting hydrogels from solutions that contain the precipitate produces microsensors with low sensitivity and selectivity, suggesting that the precipitation disrupts the structure of the enzymes. Herein, we show that a surfactant, sodium dodecyl sulfate (SDS), can prevent the precipitation and improve the sensitivity and selectivity of the sensors

  9. High-Performance Control of Paralleled Three-Phase Inverters for Residential Microgrid Architectures Based on Online Uninterruptable Power Systems

    DEFF Research Database (Denmark)

    Zhang, Chi; Guerrero, Josep M.; Vasquez, Juan Carlos

    2015-01-01

    In this paper, a control strategy for the parallel operation of three-phase inverters forming an online uninterruptible power system (UPS) is presented. The UPS system consists of a cluster of paralleled inverters with LC filters directly connected to an AC critical bus and an AC/DC forming a DC...... bus. The proposed control scheme is performed on two layers: (i) a local layer that contains a “reactive power vs phase” in order to synchronize the phase angle of each inverter and a virtual resistance loop that guarantees equal power sharing among inverters; (ii) a central controller that guarantees...... synchronization with an external real/fictitious utility, and critical bus voltage restoration. Constant transient and steady-state frequency, active, reactive and harmonic power sharing, and global phase-locked loop resynchronization capability are achieved. Detailed system topology and control architecture...

  10. Diagnostic performances of serum liver enzymes and cytokines in non-alcoholic fatty liver disease

    Directory of Open Access Journals (Sweden)

    Hakan Turkon

    2015-03-01

    Full Text Available Objective:Non-alcoholic fatty liver disease (NAFLD is affecting people worldwide with increasing prevalence. Non-invasive tests are required for both diagnosis and staging of the disease. We aimed to evaluate diagnostic accuracy of routine liver enzymes and cytokines in NAFLD. Methods:A total of 88 cases, aged between 20 and 62 years, were included in the study. Serum ALT, AST, GGT, triglyceride, TNF-alpha, IL-6 and IL-8 were measured in 40 patients with NAFLD and in 48 healthy control patients with similar BMI and demographic characteristics. Diagnostic performances of serum biomarkers for diagnosis of NAFLD were evaluated with ROC analysis. Results:ALT and AST showed good diagnostic performance in predicting patients with NAFLD in the overall group (AUC=0.817; 95% CI[0.721-0.913], AUC=0.815;95% CI[0.718-0.911] respectively but in obese subjects ALT and AST showed poor performance (AUC=0.659;95% CI[0.478-0.841], AUC=0.680; 95% CI[0.498-0.861] respectively. Among cytokines TNF-alpha showed best performance in the diagnosis of NAFLD in both overall group and obese subjects (AUC=0.892; 95% CI[0.824- 0.959], AUC=0.858; 95% CI[0.739-0.977] respectively. The optimal cut off value for TNF-alpha was 10.65pg/ml with a sensitivity of 75% and a specificity of 93% in the overall group. IL-6 and IL-8 showed poor performance. Conclusion: TNF-alpha may be a good parameter for predicting patients with NAFLD. J Clin Exp Invest 2015;6 (1: 16-20

  11. Achieving high performance in numerical computations on RISC workstations and parallel systems

    Energy Technology Data Exchange (ETDEWEB)

    Goedecker, S. [Max-Planck Inst. for Solid State Research, Stuttgart (Germany); Hoisie, A. [Los Alamos National Lab., NM (United States)

    1997-08-20

    The nominal peak speeds of both serial and parallel computers is raising rapidly. At the same time however it is becoming increasingly difficult to get out a significant fraction of this high peak speed from modern computer architectures. In this tutorial the authors give the scientists and engineers involved in numerically demanding calculations and simulations the necessary basic knowledge to write reasonably efficient programs. The basic principles are rather simple and the possible rewards large. Writing a program by taking into account optimization techniques related to the computer architecture can significantly speedup your program, often by factors of 10--100. As such, optimizing a program can for instance be a much better solution than buying a faster computer. If a few basic optimization principles are applied during program development, the additional time needed for obtaining an efficient program is practically negligible. In-depth optimization is usually only needed for a few subroutines or kernels and the effort involved is therefore also acceptable.

  12. A 3-DOF parallel robot with spherical motion for the rehabilitation and evaluation of balance performance.

    Science.gov (United States)

    Patanè, Fabrizio; Cappa, Paolo

    2011-04-01

    In this paper a novel electrically actuated parallel robot with three degrees-of-freedom (3 DOF) for dynamic postural studies is presented. The design has been described, the solution to the inverse kinematics has been found, and a numerical solution for the direct kinematics has been proposed. The workspace of the implemented robot is characterized by an angular range of motion of about ±10° for roll and pitch when yaw is in the range ±15°. The robot was constructed and the orientation accuracy was tested by means of an optoelectronic system and by imposing a sinusoidal input, with a frequency of 1 Hz and amplitude of 10°, along the three axes, in sequence. The collected data indicated a phase delay of 1° and an amplitude error of 0.5%-1.5%; similar values were observed for cross-axis sensitivity errors. We also conducted a clinical application on a group of normal subjects, who were standing in equilibrium on the robot base with eyes open (EO) and eyes closed (EC), which was rotated with a tri-axial sinusoidal trajectory with a frequency of 0.5 Hz and amplitude 5° for roll and pitch and 10° for the yaw. The postural configuration of the subjects was recorded with an optoelectronic system. However, due to the mainly technical nature of this paper, only initial validation outcomes are reported here. The clinical application showed that only the tilt and displacement on the sagittal pane of head, trunk, and pelvis in the trials conducted with eyes closed were affected by drift and that the reduction of the yaw rotation and of the mediolateral translation was not a controlled parameter, as happened, instead, for the other anatomical directions.

  13. High-performance parallel computing in the classroom using the public goods game as an example

    Science.gov (United States)

    Perc, Matjaž

    2017-07-01

    The use of computers in statistical physics is common because the sheer number of equations that describe the behaviour of an entire system particle by particle often makes it impossible to solve them exactly. Monte Carlo methods form a particularly important class of numerical methods for solving problems in statistical physics. Although these methods are simple in principle, their proper use requires a good command of statistical mechanics, as well as considerable computational resources. The aim of this paper is to demonstrate how the usage of widely accessible graphics cards on personal computers can elevate the computing power in Monte Carlo simulations by orders of magnitude, thus allowing live classroom demonstration of phenomena that would otherwise be out of reach. As an example, we use the public goods game on a square lattice where two strategies compete for common resources in a social dilemma situation. We show that the second-order phase transition to an absorbing phase in the system belongs to the directed percolation universality class, and we compare the time needed to arrive at this result by means of the main processor and by means of a suitable graphics card. Parallel computing on graphics processing units has been developed actively during the last decade, to the point where today the learning curve for entry is anything but steep for those familiar with programming. The subject is thus ripe for inclusion in graduate and advanced undergraduate curricula, and we hope that this paper will facilitate this process in the realm of physics education. To that end, we provide a documented source code for an easy reproduction of presented results and for further development of Monte Carlo simulations of similar systems.

  14. The Computational Complexity, Parallel Scalability, and Performance of Atmospheric Data Assimilation Algorithms

    Science.gov (United States)

    Lyster, Peter M.; Guo, J.; Clune, T.; Larson, J. W.; Atlas, Robert (Technical Monitor)

    2001-01-01

    The computational complexity of algorithms for Four Dimensional Data Assimilation (4DDA) at NASA's Data Assimilation Office (DAO) is discussed. In 4DDA, observations are assimilated with the output of a dynamical model to generate best-estimates of the states of the system. It is thus a mapping problem, whereby scattered observations are converted into regular accurate maps of wind, temperature, moisture and other variables. The DAO is developing and using 4DDA algorithms that provide these datasets, or analyses, in support of Earth System Science research. Two large-scale algorithms are discussed. The first approach, the Goddard Earth Observing System Data Assimilation System (GEOS DAS), uses an atmospheric general circulation model (GCM) and an observation-space based analysis system, the Physical-space Statistical Analysis System (PSAS). GEOS DAS is very similar to global meteorological weather forecasting data assimilation systems, but is used at NASA for climate research. Systems of this size typically run at between 1 and 20 gigaflop/s. The second approach, the Kalman filter, uses a more consistent algorithm to determine the forecast error covariance matrix than does GEOS DAS. For atmospheric assimilation, the gridded dynamical fields typically have More than 10(exp 6) variables, therefore the full error covariance matrix may be in excess of a teraword. For the Kalman filter this problem can easily scale to petaflop/s proportions. We discuss the computational complexity of GEOS DAS and our implementation of the Kalman filter. We also discuss and quantify some of the technical issues and limitations in developing efficient, in terms of wall clock time, and scalable parallel implementations of the algorithms.

  15. Enzyme Informatics

    Science.gov (United States)

    Alderson, Rosanna G.; Ferrari, Luna De; Mavridis, Lazaros; McDonagh, James L.; Mitchell, John B. O.; Nath, Neetika

    2012-01-01

    Over the last 50 years, sequencing, structural biology and bioinformatics have completely revolutionised biomolecular science, with millions of sequences and tens of thousands of three dimensional structures becoming available. The bioinformatics of enzymes is well served by, mostly free, online databases. BRENDA describes the chemistry, substrate specificity, kinetics, preparation and biological sources of enzymes, while KEGG is valuable for understanding enzymes and metabolic pathways. EzCatDB, SFLD and MACiE are key repositories for data on the chemical mechanisms by which enzymes operate. At the current rate of genome sequencing and manual annotation, human curation will never finish the functional annotation of the ever-expanding list of known enzymes. Hence there is an increasing need for automated annotation, though it is not yet widespread for enzyme data. In contrast, functional ontologies such as the Gene Ontology already profit from automation. Despite our growing understanding of enzyme structure and dynamics, we are only beginning to be able to design novel enzymes. One can now begin to trace the functional evolution of enzymes using phylogenetics. The ability of enzymes to perform secondary functions, albeit relatively inefficiently, gives clues as to how enzyme function evolves. Substrate promiscuity in enzymes is one example of imperfect specificity in protein-ligand interactions. Similarly, most drugs bind to more than one protein target. This may sometimes result in helpful polypharmacology as a drug modulates plural targets, but also often leads to adverse side-effects. Many cheminformatics approaches can be used to model the interactions between druglike molecules and proteins in silico. We can even use quantum chemical techniques like DFT and QM/MM to compute the structural and energetic course of enzyme catalysed chemical reaction mechanisms, including a full description of bond making and breaking. PMID:23116471

  16. Development of high-performance functional materials for enzyme immobilization by the use of ionizing radiation

    International Nuclear Information System (INIS)

    SALIM, R.D.M.

    2013-01-01

    Isomerization of glucose to fructose was carried out using Glucose isomerase (GI) that immobilized by entrapment into Poly (acrylic acid) P (AA) and Poly (acrylic acid-co- 2-Acrylamido 2- methyl Propane sulfonic acid) P (AA-co-AMPS) polymer networks, the enzyme carriers were prepared by radiation induced co-polymerization in presence of (Methylene- bis acrylamide) (MBAA) as a crosslinking agent. Effects of immobilization conditions such as irradiation dose, methylene bis acrylamide concentration, comonomer composition, and amount of GI were investigated. The influence of reaction conditions on the activity of immobilized GI were studied, the optimum ph value of reaction solution is 7.5 and reaction temperature is 65 degree C. The immobilized GI into P (AA-co-AMPS) and P (AA) polymer networks retained 81% and 69%,respectively, of its initial activity after recycled for 15 times while it retained 87% and 71% ,respectively ,of its initial activity after stored at 4 degree C for 48 days , The Km values of free and immobilized GI onto P(AA-co-AMPS) and onto P(AA) matrices were found to be 34, 29.2 , 14.5 mg/ml respectively while the Vmax Values calculated to be 3.87 ,1.6,0.79 mg/ml.min, respectively, Therefore , the bio conversion of glucose to fructose can be successfully performed by GI entrapped into P (AA-co-AMPS) hydrogel .

  17. PERFORMANCE OF LAYER HEN FED FERMENTED Jatropha Curcas L. MEAL SUPPLEMENTED WITH CELLULASE AND PHYTASE ENZYME

    Directory of Open Access Journals (Sweden)

    S. Sumiati

    2014-10-01

    Full Text Available The objective of the experiment was to study the effect of feeding fermented Jatropha curcas L.meal (JCM supplemented with cellulase and phytase on the performances of ISA-Brown laying henaged 25-30 weeks. The Jatropha curcas meal was fermented using Rizhopus oligosporus. In this study200 laying hens were used and distributed to 5 treatments and 4 replications in Completely RandomizedDesign. The diet treatments were: R0 = control diet (without JCM, R1; diet contained fermented JCM7.5%, R2; diet contained fermented JCM 7.5% + celullase 200 g/ton, R3; diet contained fermented JCM7.5% + phytase 200 g/ton and R4; diet contained fermented JCM 7.5% + cellulase 200 g/ton + phytase200 g/ton. The parameters observed were feed consumption, hen day egg production, egg massproduction, egg weight and feed conversion ratio. The results showed that feeding fermented JCM 7.5%,both enzyme supplemented as well as unsupplemented significantly decreased (P<0.05 the feedconsumption, hen day egg and egg mass production. However, the treatments did not influence the eggweight. Supplementation of cellulase (R2 or phytase (R3 improved the feed conversion ratio with thevalue as same as the R0 diet.

  18. Parallel 50 ampere hour nickel cadmium battery performance in the Modular Power Subsystems (MPS)

    Science.gov (United States)

    Webb, D. A.

    1980-01-01

    The thermal performance of 50-ampere-hour, nickel cadmium batteries for use in a modular spacecraft is examined in near-Earth orbit simulation. Battery voltage and temperature profiles for temperature extreme cycles are given and discussed.

  19. Parallel Processing Performance Evaluation of Mixed T10/T100 Ethernet Topologies on Linux Pentium Systems

    National Research Council Canada - National Science Library

    Decato, Steven

    1997-01-01

    ... performed on relatively inexpensive off the shelf components. Alternative network topologies were implemented using 10 and 100 megabit-per-second Ethernet cards under the Linux operating system on Pentium based personal computer platforms...

  20. Practical parallel computing

    CERN Document Server

    Morse, H Stephen

    1994-01-01

    Practical Parallel Computing provides information pertinent to the fundamental aspects of high-performance parallel processing. This book discusses the development of parallel applications on a variety of equipment.Organized into three parts encompassing 12 chapters, this book begins with an overview of the technology trends that converge to favor massively parallel hardware over traditional mainframes and vector machines. This text then gives a tutorial introduction to parallel hardware architectures. Other chapters provide worked-out examples of programs using several parallel languages. Thi

  1. PetClaw: Parallelization and Performance Optimization of a Python-Based Nonlinear Wave Propagation Solver Using PETSc

    KAUST Repository

    Alghamdi, Amal Mohammed

    2012-04-01

    Clawpack, a conservation laws package implemented in Fortran, and its Python-based version, PyClaw, are existing tools providing nonlinear wave propagation solvers that use state of the art finite volume methods. Simulations using those tools can have extensive computational requirements to provide accurate results. Therefore, a number of tools, such as BearClaw and MPIClaw, have been developed based on Clawpack to achieve significant speedup by exploiting parallel architectures. However, none of them has been shown to scale on a large number of cores. Furthermore, these tools, implemented in Fortran, achieve parallelization by inserting parallelization logic and MPI standard routines throughout the serial code in a non modular manner. Our contribution in this thesis research is three-fold. First, we demonstrate an advantageous use case of Python in implementing easy-to-use modular extensible scalable scientific software tools by developing an implementation of a parallelization framework, PetClaw, for PyClaw using the well-known Portable Extensible Toolkit for Scientific Computation, PETSc, through its Python wrapper petsc4py. Second, we demonstrate the possibility of getting acceptable Python code performance when compared to Fortran performance after introducing a number of serial optimizations to the Python code including integrating Clawpack Fortran kernels into PyClaw for low-level computationally intensive parts of the code. As a result of those optimizations, the Python overhead in PetClaw for a shallow water application is only 12 percent when compared to the corresponding Fortran Clawpack application. Third, we provide a demonstration of PetClaw scalability on up to the entirety of Shaheen; a 16-rack Blue Gene/P IBM supercomputer that comprises 65,536 cores and located at King Abdullah University of Science and Technology (KAUST). The PetClaw solver achieved above 0.98 weak scaling efficiency for an Euler application on the whole machine excluding the

  2. The influence of non-magnetocaloric properties on the performance in parallel-plate AMRs

    DEFF Research Database (Denmark)

    Nielsen, Kaspar Kirstein; Bahl, Christian R.H.; Smith, Anders

    2014-01-01

    a strong dependence on the orientation of the applied field and the regenerator geometry. Finally, the flow maldistribution of non-uniform regenerator geometries is found to degrade the AMR performance even at minor deviations from perfectly homogeneous regenerator matrices. This paper reflects a summary......The performance of Active Magnetic Regenerators (AMR) does not depend solely on the magnetocaloric effect of their constituents. Rather, it depends on several additional parameters, including, magnetic field, geometry (hydraulic diameter, cross-sectional area, regenerator length etc.), thermal...... properties (conductivity, specific heat and mass density) and operating parameters (utilization, frequency, number of transfer units etc.). In this paper we focus on the influence of three parameters on regenerator performance: 1) Solid thermal conductivity, 2) magnetostatic demagnetization and 3) flow...

  3. Effects of Enzyme Treated Palm Kernel Expeller on Metabolizable Energy, Growth Performance, Villus Height and Digesta Viscosity in Broiler Chickens

    Directory of Open Access Journals (Sweden)

    P. Saenphoom

    2013-04-01

    Full Text Available This study examined whether pre-treating palm kernel expeller (PKE with exogenous enzyme would degrade its fiber content; thus improving its metabolizable energy (ME, growth performance, villus height and digesta viscosity in broiler chickens fed diets containing PKE. Our results showed that enzyme treatment decreased (p0.05 among treatment groups in the finisher period, ADG of chickens in the control (PKE-free diet was higher (p0.05 FCR. The intestinal villus height and crypt depth (duodenum, jejunum and ileum were not different (p>0.05 among treatments except for duodenal crypt depth. The villus height and crypt depth of birds in enzyme treated PKE diets were higher (p0.05 among treatments. Results of this study suggest that exogenous enzyme is effective in hydrolyzing the fiber (hemicellulose and cellulose component and improved the ME values of PKE, however, the above positive effects were not reflected in the growth performance in broiler chickens fed the enzyme treated PKE compared to those received raw PKE. The results suggest that PKE can be included up to 5% in the grower diet and 20% in the finisher diet without any significant negative effect on FCR in broiler chickens.

  4. INVESTIGATION OF FLIP-FLOP PERFORMANCE ON DIFFERENT TYPE AND ARCHITECTURE IN SHIFT REGISTER WITH PARALLEL LOAD APPLICATIONS

    Directory of Open Access Journals (Sweden)

    Dwi Purnomo

    2015-08-01

    Full Text Available Register is one of the computer components that have a key role in computer organisation. Every computer contains millions of registers that are manifested by flip-flop. This research focuses on the investigation of flip-flop performance based on its type (D, T, S-R, and J-K and architecture (structural, behavioural, and hybrid. Each type of flip-flop on each architecture would be tested in different bit of shift register with parallel load applications. The experiment criteria that will be assessed are power consumption, resources required, memory required, latency, and efficiency. Based on the experiment, it could be shown that D flip-flop and hybrid architecture showed the best performance in required memory, latency, power consumption, and efficiency. In addition, the experiment results showed that the greater the register number, the less efficient the system would be.

  5. Performance analysis of a threshold-based parallel multiple beam selection scheme for WDM-based systems for Gamma-Gamma distributions

    KAUST Repository

    Nam, Sung Sik; Yoon, Chang Seok; Alouini, Mohamed-Slim

    2017-01-01

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme (TPMBS) for Free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has

  6. Systematic Design Method and Experimental Validation of a 2-DOF Compliant Parallel Mechanism with Excellent Input and Output Decoupling Performances

    Directory of Open Access Journals (Sweden)

    Yao Jiang

    2017-06-01

    Full Text Available The output and input coupling characteristics of the compliant parallel mechanism (CPM bring difficulty in the motion control and challenge its high performance and operational safety. This paper presents a systematic design method for a 2-degrees-of-freedom (DOFs CPM with excellent decoupling performance. A symmetric kinematic structure can guarantee a CPM with a complete output decoupling characteristic; input coupling is reduced by resorting to a flexure-based decoupler. This work discusses the stiffness design requirement of the decoupler and proposes a compound flexure hinge as its basic structure. Analytical methods have been derived to assess the mechanical performances of the CPM in terms of input and output stiffness, motion stroke, input coupling degree, and natural frequency. The CPM’s geometric parameters were optimized to minimize the input coupling while ensuring key performance indicators at the same time. The optimized CPM’s performances were then evaluated by using a finite element analysis. Finally, a prototype was constructed and experimental validations were carried out to test the performance of the CPM and verify the effectiveness of the design method. The design procedure proposed in this paper is systematic and can be extended to design the CPMs with other types of motion.

  7. Gaming of performance measurement in health care: parallels with tax compliance.

    Science.gov (United States)

    Mears, Alex; Webley, Paul

    2010-10-01

    Gaming of performance measurement in health care distorts performance, making it appear better than it is. This can conceal potentially hazardous practice and endanger patients and staff. Research has investigated and categorized this behaviour but as yet has offered little in the way of potential solutions as the drivers are still not well understood. Studies of the psychology of tax behaviour, specifically tax avoidance and evasion, reveal some insights into what the underlying causes are. Looking at health care and tax compliance, seven similarities in response can be detected: negative view of those subject to it; not salient except for specific classes of people; general understanding of the need versus resentment of actuality; cognitive dissonance; moral versus legal grey areas; two levels of objective non-compliance; and four categories of subjective behaviour and attitude. We suggest a model of reaction and compliance for gaming and offer some suggestions as to how this phenomenon might be minimized.

  8. Performance Characteristics of a Refrigerator-Freezer with Parallel Evaporators using a Linear Compressor

    OpenAIRE

    Min, Byungchae; Song, Sangjin; Noh, Kiyoul; Kim, Geonwoo; Yoon, Teaseung; Na, Sangkyung; Song, Sanghoon; Yang, Jangsik; Choi, Gyungmin; Kim, Duckjool

    2016-01-01

    A linear compressor for a domestic refrigerator-freezer has energy saving potential compared with a reciprocating compressor because of a low friction loss and free piston system. A linear compressor can control the piston stroke since it does not have mechanical restriction of piston movement. Therefore, the energy consumption of a domestic refrigerator-freezer using a linear compressor can be reduced by changing the cooling capacity of the compressor. In order to investigate the performance...

  9. High Performance Molecular Visualization: In-Situ and Parallel Rendering with EGL

    Science.gov (United States)

    Stone, John E.; Messmer, Peter; Sisneros, Robert; Schulten, Klaus

    2016-01-01

    Large scale molecular dynamics simulations produce terabytes of data that is impractical to transfer to remote facilities. It is therefore necessary to perform visualization tasks in-situ as the data are generated, or by running interactive remote visualization sessions and batch analyses co-located with direct access to high performance storage systems. A significant challenge for deploying visualization software within clouds, clusters, and supercomputers involves the operating system software required to initialize and manage graphics acceleration hardware. Recently, it has become possible for applications to use the Embedded-system Graphics Library (EGL) to eliminate the requirement for windowing system software on compute nodes, thereby eliminating a significant obstacle to broader use of high performance visualization applications. We outline the potential benefits of this approach in the context of visualization applications used in the cloud, on commodity clusters, and supercomputers. We discuss the implementation of EGL support in VMD, a widely used molecular visualization application, and we outline benefits of the approach for molecular visualization tasks on petascale computers, clouds, and remote visualization servers. We then provide a brief evaluation of the use of EGL in VMD, with tests using developmental graphics drivers on conventional workstations and on Amazon EC2 G2 GPU-accelerated cloud instance types. We expect that the techniques described here will be of broad benefit to many other visualization applications. PMID:27747137

  10. Parallel implementation and performance optimization of the configuration-interaction method

    Energy Technology Data Exchange (ETDEWEB)

    Shan, H; Williams, S; Johnson, C; McElvain, K; Ormand, WE

    2015-11-20

    The configuration-interaction (CI) method, long a popular approach to describe quantum many-body systems, is cast as a very large sparse matrix eigenpair problem with matrices whose dimension can exceed one billion. Such formulations place high demands on memory capacity and memory bandwidth - - two quantities at a premium today. In this paper, we describe an efficient, scalable implementation, BIGSTICK, which, by factorizing both the basis and the interaction into two levels, can reconstruct the nonzero matrix elements on the fly, reduce the memory requirements by one or two orders of magnitude, and enable researchers to trade reduced resources for increased computational time. We optimize BIGSTICK on two leading HPC platforms - - the Cray XC30 and the IBM Blue Gene/Q. Specifically, we not only develop an empirically-driven load balancing strategy that can evenly distribute the matrix-vector multiplication across 256K threads, we also developed techniques that improve the performance of the Lanczos reorthogonalization. Combined, these optimizations improved performance by 1.3-8× depending on platform and configuration.

  11. Choosing processor array configuration by performance modeling for a highly parallel linear algebra algorithm

    International Nuclear Information System (INIS)

    Littlefield, R.J.; Maschhoff, K.J.

    1991-04-01

    Many linear algebra algorithms utilize an array of processors across which matrices are distributed. Given a particular matrix size and a maximum number of processors, what configuration of processors, i.e., what size and shape array, will execute the fastest? The answer to this question depends on tradeoffs between load balancing, communication startup and transfer costs, and computational overhead. In this paper we analyze in detail one algorithm: the blocked factored Jacobi method for solving dense eigensystems. A performance model is developed to predict execution time as a function of the processor array and matrix sizes, plus the basic computation and communication speeds of the underlying computer system. In experiments on a large hypercube (up to 512 processors), this model has been found to be highly accurate (mean error ∼ 2%) over a wide range of matrix sizes (10 x 10 through 200 x 200) and processor counts (1 to 512). The model reveals, and direct experiment confirms, that the tradeoffs mentioned above can be surprisingly complex and counterintuitive. We propose decision procedures based directly on the performance model to choose configurations for fastest execution. The model-based decision procedures are compared to a heuristic strategy and shown to be significantly better. 7 refs., 8 figs., 1 tab

  12. Parallel performances of three 3D reconstruction methods on MIMD computers: Feldkamp, block ART and SIRT algorithms

    International Nuclear Information System (INIS)

    Laurent, C.; Chassery, J.M.; Peyrin, F.; Girerd, C.

    1996-01-01

    This paper deals with the parallel implementations of reconstruction methods in 3D tomography. 3D tomography requires voluminous data and long computation times. Parallel computing, on MIMD computers, seems to be a good approach to manage this problem. In this study, we present the different steps of the parallelization on an abstract parallel computer. Depending on the method, we use two main approaches to parallelize the algorithms: the local approach and the global approach. Experimental results on MIMD computers are presented. Two 3D images reconstructed from realistic data are showed

  13. Influence of fungal morphology on the performance of industrial fermentation processes for enzyme production

    DEFF Research Database (Denmark)

    Quintanilla Hernandez, Daniela Alejandra

    Production of industrial enzymes is usually carried out as submerged aerobic fermentations. Filamentous microorganisms are widely used as hosts in these processes due to multiple advantages. Nevertheless, they also present major drawbacks, due to the unavoidable oxygen transfer limitations...... in this work, along with its correlation to viscosity and other process variables. Considerable research work has been conducted through the years to study fungal morphology and its relation to productivity. However, the work reported in the literature lacks relevant industrial data. In this work, a platform...... was developed which was able to produce high enzyme titers in comparison with what has been reported thus far in fed-batch fermentation using a soluble inducer (lactose). Different nitrogen sources were compared, and it was found that soy meal allowed for higher enzyme titers compared to what has been reported...

  14. High performance separation of xylose and glucose by enzyme assisted nanofiltration

    DEFF Research Database (Denmark)

    Morthensen, Sofie Thage; Luo, Jianquan; Meyer, Anne S.

    2015-01-01

    of the integrated system. Full conversion of glucose to gluconic acid assisted by glucose oxidase (GOD) could be achieved by coupling a parallel reaction catalyzed by catalase (CAT), where H2O2 (GOD-inhibitor formed in the first reaction) was decomposed to water and oxygen. GOD has a high oxygen...

  15. Using an Inducible Promoter of a Gene Encoding Penicillium verruculosum Glucoamylase for Production of Enzyme Preparations with Enhanced Cellulase Performance.

    Directory of Open Access Journals (Sweden)

    Alexander G Bulakhov

    Full Text Available Penicillium verruculosum is an efficient producer of highly active cellulase multienzyme system. One of the approaches for enhancing cellulase performance in hydrolysis of cellulosic substrates is to enrich the reaction system with β -glucosidase and/or accessory enzymes, such as lytic polysaccharide monooxygenases (LPMO displaying a synergism with cellulases.Genes bglI, encoding β-glucosidase from Aspergillus niger (AnBGL, and eglIV, encoding LPMO (formerly endoglucanase IV from Trichoderma reesei (TrLPMO, were cloned and expressed by P. verruculosum B1-537 strain under the control of the inducible gla1 gene promoter. Content of the heterologous AnBGL in the secreted multienzyme cocktails (hBGL1, hBGL2 and hBGL3 varied from 4 to 10% of the total protein, while the content of TrLPMO in the hLPMO sample was ~3%. The glucose yields in 48-h hydrolysis of Avicel and milled aspen wood by the hBGL1, hBGL2 and hBGL3 preparations increased by up to 99 and 80%, respectively, relative to control enzyme preparations without the heterologous AnBGL (at protein loading 5 mg/g substrate for all enzyme samples. The heterologous TrLPMO in the hLPMO preparation boosted the conversion of the lignocellulosic substrate by 10-43%; however, in hydrolysis of Avicel the hLPMO sample was less effective than the control preparations. The highest product yield in hydrolysis of aspen wood was obtained when the hBGL2 and hLPMO preparations were used at the ratio 1:1.The enzyme preparations produced by recombinant P. verruculosum strains, expressing the heterologous AnBGL or TrLPMO under the control of the gla1 gene promoter in a starch-containing medium, proved to be more effective in hydrolysis of a lignocellulosic substrate than control enzyme preparations without the heterologous enzymes. The enzyme composition containing both AnBGL and TrLPMO demonstrated the highest performance in lignocellulose hydrolysis, providing a background for developing a fungal strain capable

  16. Performance Characteristics of a PEM Fuel Cell with Parallel Flow Channels at Different Cathode Relative Humidity Levels

    Directory of Open Access Journals (Sweden)

    Sang Soon Hwang

    2009-11-01

    Full Text Available In fuel cells flow configuration and operating conditions such as cell temperature, humidity at each electrode and stoichiometric number are very crucial for improving performance. Too many flow channels could enhance the performance but result in high parasite loss. Therefore a trade-off between pressure drop and efficiency of a fuel cell should be considered for optimum design. This work focused on numerical simulation of the effects of operating conditions, especially cathode humidity, with simple micro parallel flow channels. It is known that the humidity at the cathode flow channel becomes very important for enhancing the ion conductivity of polymer membrane because fully humidified condition was normally set at anode. To investigate the effect of humidity on the performance of a fuel cell, in this study humidification was set to 100% at the anode flow channel and was changed by 0–100% at the cathode flow channel. Results showed that the maximum power density could be obtained under 60% humidified condition at the cathode where oxygen concentration was moderately high while maintaining high ion conductivity at a membrane.

  17. Structural Directed Growth of Ultrathin Parallel Birnessite on β-MnO2 for High-Performance Asymmetric Supercapacitors.

    Science.gov (United States)

    Zhu, Shijin; Li, Li; Liu, Jiabin; Wang, Hongtao; Wang, Tian; Zhang, Yuxin; Zhang, Lili; Ruoff, Rodney S; Dong, Fan

    2018-02-27

    Two-dimensional birnessite has attracted attention for electrochemical energy storage because of the presence of redox active Mn 4+ /Mn 3+ ions and spacious interlayer channels available for ions diffusion. However, current strategies are largely limited to enhancing the electrical conductivity of birnessite. One key limitation affecting the electrochemical properties of birnessite is the poor utilization of the MnO 6 unit. Here, we assemble β-MnO 2 /birnessite core-shell structure that exploits the exposed crystal face of β-MnO 2 as the core and ultrathin birnessite sheets that have the structure advantage to enhance the utilization efficiency of the Mn from the bulk. Our birnessite that has sheets parallel to each other is found to have unusual crystal structure with interlayer spacing, Mn(III)/Mn(IV) ratio and the content of the balancing cations differing from that of the common birnessite. The substrate directed growth mechanism is carefully investigated. The as-prepared core-shell nanostructures enhance the exposed surface area of birnessite and achieve high electrochemical performances (for example, 657 F g -1 in 1 M Na 2 SO 4 electrolyte based on the weight of parallel birnessite) and excellent rate capability over a potential window of up to 1.2 V. This strategy opens avenues for fundamental studies of birnessite and its properties and suggests the possibility of its use in energy storage and other applications. The potential window of an asymmetric supercapacitor that was assembled with this material can be enlarged to 2.2 V (in aqueous electrolyte) with a good cycling ability.

  18. Modular and efficient ozone systems based on massively parallel chemical processing in microchannel plasma arrays: performance and commercialization

    Science.gov (United States)

    Kim, M.-H.; Cho, J. H.; Park, S.-J.; Eden, J. G.

    2017-08-01

    Plasmachemical systems based on the production of a specific molecule (O3) in literally thousands of microchannel plasmas simultaneously have been demonstrated, developed and engineered over the past seven years, and commercialized. At the heart of this new plasma technology is the plasma chip, a flat aluminum strip fabricated by photolithographic and wet chemical processes and comprising 24-48 channels, micromachined into nanoporous aluminum oxide, with embedded electrodes. By integrating 4-6 chips into a module, the mass output of an ozone microplasma system is scaled linearly with the number of modules operating in parallel. A 115 g/hr (2.7 kg/day) ozone system, for example, is realized by the combined output of 18 modules comprising 72 chips and 1,800 microchannels. The implications of this plasma processing architecture for scaling ozone production capability, and reducing capital and service costs when introducing redundancy into the system, are profound. In contrast to conventional ozone generator technology, microplasma systems operate reliably (albeit with reduced output) in ambient air and humidity levels up to 90%, a characteristic attributable to the water adsorption/desorption properties and electrical breakdown strength of nanoporous alumina. Extensive testing has documented chip and system lifetimes (MTBF) beyond 5,000 hours, and efficiencies >130 g/kWh when oxygen is the feedstock gas. Furthermore, the weight and volume of microplasma systems are a factor of 3-10 lower than those for conventional ozone systems of comparable output. Massively-parallel plasmachemical processing offers functionality, performance, and commercial value beyond that afforded by conventional technology, and is currently in operation in more than 30 countries worldwide.

  19. Effect of using the Matrix Values for NSP-degrading enzymes on performance, water intake, litter moisture and jejunal digesta viscosity of broilers fed barley-based diet

    Directory of Open Access Journals (Sweden)

    Seyed Adel Moftakharzadeh

    2017-02-01

    Full Text Available In this study, we have evaluated the effect of three multi-enzymes nutrient matrix values and compared the results with that fed barley and the corn diets without enzyme. In entire period, addition of all enzymes to the barley-based diet significantly (p 0.05. Litter moisture and water to feed ratio at 15, 25, and 33 days of age significantly decreased by addition of all enzymes (p < 0.05. In conclusion, considering nutrient matrix values for all used enzymes improved performance of broilers and can be used in formulating diets commercial broiler diets based on barley.

  20. Improving the performance of dairy cattle with a xylanase-rich exogenous enzyme preparation.

    Science.gov (United States)

    Romero, J J; Macias, E G; Ma, Z X; Martins, R M; Staples, C R; Beauchemin, K A; Adesogan, A T

    2016-05-01

    The objective of this experiment was to examine effects of adding 2 exogenous fibrolytic enzymes (EFE) to the total mixed ration (TMR) on the performance of lactating dairy cows (experiment 1) and the kinetics of ruminal degradation of the diet (experiment 2). Twelve EFE had been screened in a series of in vitro assays that identified the most potent EFE and their optimal doses for increasing the digestibility of bermudagrass. In experiment 1, 66 Holstein cows (21±5 d in milk) were grouped by previous milk production and parity (45 multiparous and 21 primiparous) and assigned randomly to 1 of the following 3 treatments: (1) control (CON, untreated), (2) Xylanase Plus [2A, 1mL/kg of TMR dry matter (DM); Dyadic International, Jupiter, FL], and (3) a 75:25 (vol/vol) mixture of Cellulase Plus and Xylanase Plus EFE (3A, 3.4mL/kg of TMR DM; Dyadic International). The EFE were sprayed twice daily onto a TMR (10% bermudagrass silage, 35% corn silage, 5% alfalfa-orchardgrass hay mixture, and 50% concentrates; DM basis) and fed for a 14-d training and covariate period and a 70-d measurement period. Experiment 2 aimed to examine the in situ DM ruminal degradability and ruminal fermentation measurements of the diets fed in experiment 1. Three ruminally fistulated lactating Holstein cows were assigned to the diets. The experiment had a 3×3 Latin square design with 23-d periods. In experiment 1, application of 2A increased intakes (kg/d) of DM (23.5 vs. 22.6), organic matter (21.9 vs. 20.9), and crude protein (3.9 vs. 3.7) and tended to increase yields (kg/d) of fat-corrected milk (41.8 vs. 40.7) and milk fat (1.48 vs. 1.44). In particular, 2A increased milk yield (kg/d) during wk 3 (41.2 vs. 39.8, tendency), 6 (41.9 vs. 40.1), and 7 (42.1 vs. 40.4), whereas 3A increased milk yield (kg/d) during wk 6 (41.5 vs. 40.1, tendency), 8 (41.8 vs. 40.0), and 9 (40.9 vs. 39.5, tendency). In experiment 2, EFE treatment did not affect ruminal DM degradation kinetics or ruminal pH, ammonia

  1. A meta-analysis on the effect of dietary application of exogenous fibrolytic enzymes on the performance of dairy cows.

    Science.gov (United States)

    Arriola, Kathy G; Oliveira, Andre S; Ma, Zhengxin X; Lean, Ian J; Giurcanu, Mihai C; Adesogan, Adegbola T

    2017-06-01

    The aim of this study was to use meta-analytical methods to estimate effects of adding exogenous fibrolytic enzymes (EFE) to dairy cow diets on their performance and to determine which factors affect the response. Fifteen studies with 17 experiments and 36 observations met the study selection criteria for inclusion in the meta-analysis. The effects were compared by using random-effect models to examine the raw mean difference (RMD) and standardized mean difference between EFE and control treatments after both were weighted with the inverse of the study variances. Heterogeneity sources evaluated by meta-regression included experimental duration, EFE type and application rate, form (liquid or solid), and method (application to the forage, concentrate, or total mixed ration). Only the cellulase-xylanase (C-X) enzymes had a substantial number of observations (n = 13 studies). Application of EFE, overall, did not affect dry matter intake, feed efficiency but tended to increase total-tract dry matter digestibility and neutral detergent fiber digestibility (NDFD) by relatively small amounts (1.36 and 2.30%, respectively, or 50%) was detected for total-tract dry matter digestibility and NDFD. Milk production responses were higher for the C-X enzymes (RMD = 1.04 kg/d; 95% confidence interval: 0.33 to 1.74), but were still only moderate, about 0.35 standardized mean difference. A 24% numerical increase in the RMD resulting from examining only C-X enzymes instead of all enzymes (RMD = 1.04 vs. 0.83 kg/d) suggests that had more studies met the inclusion criteria, the C-X enzymes would have statistically increased the milk response relative to that for all enzymes. Increasing the EFE application rate had no effect on performance measures. Application of EFE to the total mixed ration improved only milk protein concentration, and application to the forage or concentrate had no effect. Applying EFE tended to increase dry matter digestibility and NDFD and increased milk yield by

  2. Effect of exogenous fibrolytic enzymes on performance and blood profile in early and mid-lactation Holstein cows

    Directory of Open Access Journals (Sweden)

    Anja Peters

    2015-09-01

    Full Text Available The supplementation of exogenous fibrolytic enzymes (EFE to dairy cows diets could be a strategy to improve fiber degradation in the rumen which is especially important for the early lactating cows characterized by a high milk energy output and an insufficient energy intake. The objective of this study was to examine the effects of a fibrolytic enzyme product (Roxazyme G2 Liquid, 3.8 and 3.9 mL/kg total mixed ration [TMR] DM supplemented to a TMR on production performance and blood parameters of dairy cows during early (trial 1 and mid-lactation (trail 2. In addition, rumination activity was measured in trial 2. The nutrient digestibility of the experimental TMR was obtained by using wethers. In the digestibility trial, EFE was supplemented at a rate of 4.4 mL/kg Roxazyme G2 Liquid TMR-DM. The TMR contained 60% forage and 40% concentrate (DM basis. Twenty eight 50 ± 16 days in milk (DIM and twenty six 136 ± 26 DIM Holstein cows were used in two 8-wk completely randomized trails, stratified by parity and milk yield level. One milliliter of the enzyme product contained primarily cellulase and xylanase activities (8,000 units endo-1,4-ß glucanase, 18,000 units endo-1,3(4-ß glucanase and 26,000 units 1,4-ß xylanase. No differences in digestibility of DM, OM, CP, NDF and ADF were observed (P > 0.05 between the control and the EFE supplemented TMR. Addition of EFE to the TMR fed to early (trial 1 and mid-lactation cows (trial 2 did not affect daily dry matter intake (DMI, milk yield, 4% fat-corrected milk, energy-corrected milk (ECM, concentration of milk fat, protein, fat-protein-quotients, somatic cell score, energy balance, and gross feed efficiency of early and mid-lactation cows (P > 0.05. Mid-lactation cows (trial 2 fed with TMR enzyme showed a tendency of a slightly higher ECM yield (P = 0.09. The tested blood parameters were not affected by treatment in trials 1 and 2 (P > 0.05. Exogenous fibrolytic enzymes supplementation did not alter

  3. Performance and Serum Hepatic Enzymes of Hy-Line W-36 Laying Hens Intoxicated with Dietary Carbon Tetrachloride

    Directory of Open Access Journals (Sweden)

    Hadavi A

    2015-12-01

    Full Text Available An experiment was conducted to study the effects of carbon tetrachloride (CCl4 on post-peak performance and serum enzymes of Hy-Line W-36 laying hens from 32-36 weeks of age. The experiment was carried out with a total of 192 laying hens in a completely randomized block design. During the experiment laying hens were allocated to 4 groups consisted of T1 no CCl4 as control diet, T2, T3 and T4 control diet supplemented with 1, 3 and 5 mL CCl4/100 g diet, respectively. Each experimental group was divided into 6 blocks of 8 hens each. Egg production, cracked egg percentage and feed intake were recorded weekly. Blood samples were taken from wing veins of hens at the middle and end of the experiment to measure serum hepatic enzymes of alkaline phosphatase, alanine aminotransferase and aspartate aminotransferase. Data showed that in comparison with the control group, the inclusion of CCl4 to the diets had no significant effect on performance parameters. However, by increasing the level of CCl4, egg production was linearly decreased and feed intake was linearly increased (P < 0.05. The effect of CCl4 on cracked eggs was significant and this effect was linearly increased (P < 0.05. Dietary supplementation of 3 and 5 mL CCl4 elevated the serum concentration of hepatic enzymes of alkaline phosphatase, aspartate aminotransferase and alanine aminotransferase, linearly (P < 0.0001. In conclusion, the dietary supplementation of CCl4 has the ability to decrease the performance and egg quality. CCl4 is also a potent hepatic toxicity inducer and may damage liver hepatocytes. Therefore, the level of 3 mL CCl4 was assigned as the one had the maximum negative effect on serum hepatic enzymes concentration (maximum liver damage alongside the minimum negative effect on laying hen performance for further studies.

  4. Achieving Performance Speed-up in FPGA Based Bit-Parallel Multipliers using Embedded Primitive and Macro support

    Directory of Open Access Journals (Sweden)

    Burhan Khurshid

    2015-05-01

    Full Text Available Modern Field Programmable Gate Arrays (FPGA are fast moving into the consumer market and their domain has expanded from prototype designing to low and medium volume productions. FPGAs are proving to be an attractive replacement for Application Specific Integrated Circuits (ASIC primarily because of the low Non-recurring Engineering (NRE costs associated with FPGA platforms. This has prompted FPGA vendors to improve the capacity and flexibility of the underlying primitive fabric and include specialized macro support and intellectual property (IP cores in their offerings. However, most of the work related to FPGA implementations does not take full advantage of these offerings. This is primarily because designers rely mainly on the technology-independent optimization to enhance the performance of the system and completely neglect the speed-up that is achievable using these embedded primitives and macro support. In this paper, we consider the technology-dependent optimization of fixed-point bit-parallel multipliers by carrying out their implementations using embedded primitives and macro support that are inherent in modern day FPGAs. Our implementation targets three different FPGA families viz. Spartan-6, Virtex-4 and Virtex-5. The implementation results indicate that a considerable speed up in performance is achievable using these embedded FPGA resources.

  5. Using the Eclipse Parallel Tools Platform to Assist Earth Science Model Development and Optimization on High Performance Computers

    Science.gov (United States)

    Alameda, J. C.

    2011-12-01

    Development and optimization of computational science models, particularly on high performance computers, and with the advent of ubiquitous multicore processor systems, practically on every system, has been accomplished with basic software tools, typically, command-line based compilers, debuggers, performance tools that have not changed substantially from the days of serial and early vector computers. However, model complexity, including the complexity added by modern message passing libraries such as MPI, and the need for hybrid code models (such as openMP and MPI) to be able to take full advantage of high performance computers with an increasing core count per shared memory node, has made development and optimization of such codes an increasingly arduous task. Additional architectural developments, such as many-core processors, only complicate the situation further. In this paper, we describe how our NSF-funded project, "SI2-SSI: A Productive and Accessible Development Workbench for HPC Applications Using the Eclipse Parallel Tools Platform" (WHPC) seeks to improve the Eclipse Parallel Tools Platform, an environment designed to support scientific code development targeted at a diverse set of high performance computing systems. Our WHPC project to improve Eclipse PTP takes an application-centric view to improve PTP. We are using a set of scientific applications, each with a variety of challenges, and using PTP to drive further improvements to both the scientific application, as well as to understand shortcomings in Eclipse PTP from an application developer perspective, to drive our list of improvements we seek to make. We are also partnering with performance tool providers, to drive higher quality performance tool integration. We have partnered with the Cactus group at Louisiana State University to improve Eclipse's ability to work with computational frameworks and extremely complex build systems, as well as to develop educational materials to incorporate into

  6. Optimizing Immobilized Enzyme Performance in Cell-Free Environments to Produce Liquid Fuels

    Energy Technology Data Exchange (ETDEWEB)

    Belfort, Georges [Rensselaer Polytechnic Inst., Troy, NY (United States). Dept. of Chemical and Biological Engineering; Grimaldi, Joseph J. [Rensselaer Polytechnic Inst., Troy, NY (United States). Dept. of Chemical and Biological Engineering

    2015-01-27

    Limitations on biofuel production using cell culture (Escherichia coli, Clostridium, Saccharomyces cerevisiae, brown microalgae, blue-green algae and others) include low product (alcohol) concentrations (≤0.2 vol%) due to feedback inhibition, instability of cells, and lack of economical product recovery processes. To overcome these challenges, an alternate simplified biofuel production scheme was tested based on a cell-free immobilized enzyme system. Using this cell free system, we were able to obtain about 2.6 times higher concentrations of iso-butanol using our non-optimized system as compared with live cell systems. This process involved two steps: (i) converts acid to aldehyde using keto-acid decarboxylase (KdcA), and (ii) produces alcohol from aldehyde using alcohol dehydrogenase (ADH) with a cofactor (NADH) conversion from inexpensive formate using a third enzyme, formate dehydrogenase (FDH). To increase stability and conversion efficiency with easy separations, the first two enzymes were immobilized onto methacrylate resin. Fusion proteins of labile KdcA (fKdcA) were expressed to stabilize the covalently immobilized KdcA. Covalently immobilized ADH exhibited long-term stability and efficient conversion of aldehyde to alcohol over multiple batch cycles without fusions. High conversion rates and low protein leaching were achieved by covalent immobilization of enzymes on methacrylate resin. The complete reaction scheme was demonstrated by immobilizing both ADH and fKdcA and using FDH free in solution. The new system without in situ removal of isobutanol achieved a 55% conversion of ketoisovaleric acid to isobutanol at a concentration of 0.5 % (v/v). Further increases in titer will require continuous removal of the isobutanol using our novel brush membrane system that exhibits a 1.5 fold increase in the separation factor of isobutanol from water versus that obtained for commercial silicone rubber membranes. These bio-inspired brush membranes are based on the

  7. Novel, Highly-Parallel Software for the Online Storage System of the ATLAS Experiment at CERN: Design and Performances

    CERN Document Server

    Colombo, T; The ATLAS collaboration

    2012-01-01

    The ATLAS experiment observes proton-proton collisions delivered by the LHC accelerator at CERN. The ATLAS Trigger and Data Acquisition (TDAQ) system selects interesting events on-line in a three-level trigger system in order to store them at a budgeted rate of several hundred Hz, for an average event size of ~1.5 MB. This paper focuses on the TDAQ data-logging system and in particular on the implementation and performance of a novel SW design, reporting on the effort of exploiting the full power of recently installed multi-core hardware. In this respect, the main challenge presented by the data-logging workload is the conflict between the largely parallel nature of the event processing, including the recently introduced on-line event-compression, and the constraint of sequential file writing and checksum evaluation. This is further complicated by the necessity of operating in a fully data-driven mode, to cope with continuously evolving trigger and detector configurations. In this paper we report on the desig...

  8. Novel, Highly-Parallel Software for the Online Storage System of the ATLAS Experiment at CERN: Design and Performances

    CERN Document Server

    Colombo, T; The ATLAS collaboration

    2012-01-01

    Abstract--- The ATLAS experiment observes proton-proton collisions delivered by the LHC accelerator at CERN. The ATLAS Trigger and Data Acquisition (TDAQ) system selects interesting events on-line in a three-level trigger system in order to store them at a budgeted rate of several hundred Hz, for an average event size of ~1.5 MB. This paper focuses on the TDAQ data-logging system and in particular on the implementation and performance of a novel software design, reporting on the effort of exploiting the full power of multi-core hardware. In this respect, the main challenge presented by the data-logging workload is the conflict between the largely parallel nature of the event processing, including the recently introduced on-line event-compression, and the constraint of sequential file writing and checksum evaluation. This is further complicated by the necessity of operating in a fully data-driven mode, to cope with continuously evolving trigger and detector configurations. In this paper we will briefly discuss...

  9. Map-Based Power-Split Strategy Design with Predictive Performance Optimization for Parallel Hybrid Electric Vehicles

    Directory of Open Access Journals (Sweden)

    Jixiang Fan

    2015-09-01

    Full Text Available In this paper, a map-based optimal energy management strategy is proposed to improve the consumption economy of a plug-in parallel hybrid electric vehicle. In the design of the maps, which provide both the torque split between engine and motor and the gear shift, not only the current vehicle speed and power demand, but also the optimality based on the predicted trajectory of vehicle dynamics are considered. To seek the optimality, the equivalent consumption, which trades off the fuel and electricity usages, is chosen as the cost function. Moreover, in order to decrease the model errors in the process of optimization conducted in the discrete time domain, the variational integrator is employed to calculate the evolution of the vehicle dynamics. To evaluate the proposed energy management strategy, the simulation results performed on a professional GT-Suit simulator are demonstrated and the comparison to a real-time optimization method is also given to show the advantage of the proposed off-line optimization approach.

  10. Fundamental understanding of liquid water effects on the performance of a PEMFC with serpentine-parallel channels

    International Nuclear Information System (INIS)

    Le, Anh Dinh; Zhou Biao

    2009-01-01

    A three-dimensional and unsteady proton exchange membrane fuel cell (PEMFC) model with serpentine-parallel channels has been incorporated to simulate not only the fluid flow, heat transfer, species transport, electrochemical reaction, and current density distribution but also the behaviors of liquid water in the gas-liquid flow of the channels and porous media. Using this general model, the behaviors of liquid water were investigated by performing the motion, deformation, coalescence and detachment of water droplets inside the channels and the penetration of liquid through the porous media at different time instants. The results showed that: tracking the interface of liquid water in a reacting gas-liquid flow in PEMFC can be fulfilled by using volume-of-fluid (VOF) algorithm combined with solving the conservation equations of continuity, momentum, energy, species transport and electrochemistry; the presence of liquid water in the channels has a significant impact on the flow fields, e.g., the gas flow became unevenly distributed due to the blockage of liquid water where the high pressure would be suddenly built up and the reactant gas transport in the channels and porous media would be hindered by liquid water occupation

  11. Coordinate changes in photosynthesis, sugar accumulation and antioxidative enzymes improve the performance of Jatropha curcas plants under drought stress

    International Nuclear Information System (INIS)

    Silva, Evandro N.; Ribeiro, Rafael V.; Ferreira-Silva, Sérgio L.; Vieira, Suyanne A.; Ponte, Luiz F.A.; Silveira, Joaquim A.G.

    2012-01-01

    The aim of this study was to assess the relationships between photosynthesis, sugars and photo-oxidative protection mechanisms in Jatropha curcas under drought stress. Leaf CO 2 assimilation rate (P N ) and instantaneous carboxylation efficiency decreased progressively as the water deficit increased. The sucrose and reducing sugar concentrations were negatively and highly correlated with photosynthesis indicating a modulation by negative feedback mechanism. The alternative electron sinks (ETR s '/P N ), relative excess of light energy (EXC) and non-photochemical quenching were strongly increased by drought, indicating effective mechanisms of energy excess dissipation. The photochemistry data indicate partial preservation of photosystem II integrity and function even under severe drought. EXC was positively correlated with superoxide dismutase (SOD) and ascorbate peroxidase (APX) activities evidencing an effective role of these enzymes in the oxidative protection against excess of reactive oxygen species in chloroplasts. Leaf H 2 O 2 content and lipid peroxidation were inversely and highly correlated with catalase (CAT) activity indicating that drought-induced inhibition of this enzyme might have allowed oxidative damage. Our data suggest that drought triggers a coordinate down-regulation in photosynthesis through sucrose and reducing sugar accumulation and an energy excess dissipation at PSII level by non-photochemical mechanisms associate with enhancement in photorespiration, restricting photo-damages. In parallel, drought up-regulates SOD and APX activities avoiding accumulation of reactive oxygen species, while CAT activity is not able to avoid H 2 O 2 accumulation in drought-stressed J. curcas leaves. -- Highlights: ► Drought triggers a down-regulation in photosynthesis by sucrose and reducing sugar. ► Drought induces energy dissipation at PSII level and increase in photorespiration. ► Drought up-regulates SOD and APX activities avoiding accumulation of

  12. Relationship between mathematical abstraction in learning parallel coordinates concept and performance in learning analytic geometry of pre-service mathematics teachers: an investigation

    Science.gov (United States)

    Nurhasanah, F.; Kusumah, Y. S.; Sabandar, J.; Suryadi, D.

    2018-05-01

    As one of the non-conventional mathematics concepts, Parallel Coordinates is potential to be learned by pre-service mathematics teachers in order to give them experiences in constructing richer schemes and doing abstraction process. Unfortunately, the study related to this issue is still limited. This study wants to answer a research question “to what extent the abstraction process of pre-service mathematics teachers in learning concept of Parallel Coordinates could indicate their performance in learning Analytic Geometry”. This is a case study that part of a larger study in examining mathematical abstraction of pre-service mathematics teachers in learning non-conventional mathematics concept. Descriptive statistics method is used in this study to analyze the scores from three different tests: Cartesian Coordinate, Parallel Coordinates, and Analytic Geometry. The participants in this study consist of 45 pre-service mathematics teachers. The result shows that there is a linear association between the score on Cartesian Coordinate and Parallel Coordinates. There also found that the higher levels of the abstraction process in learning Parallel Coordinates are linearly associated with higher student achievement in Analytic Geometry. The result of this study shows that the concept of Parallel Coordinates has a significant role for pre-service mathematics teachers in learning Analytic Geometry.

  13. Portable and Transparent Message Compression in MPI Libraries to Improve the Performance and Scalability of Parallel Applications

    Energy Technology Data Exchange (ETDEWEB)

    Albonesi, David; Burtscher, Martin

    2009-04-17

    The goal of this project has been to develop a lossless compression algorithm for message-passing libraries that can accelerate HPC systems by reducing the communication time. Because both compression and decompression have to be performed in software in real time, the algorithm has to be extremely fast while still delivering a good compression ratio. During the first half of this project, they designed a new compression algorithm called FPC for scientific double-precision data, made the source code available on the web, and published two papers describing its operation, the first in the proceedings of the Data Compression Conference and the second in the IEEE Transactions on Computers. At comparable average compression ratios, this algorithm compresses and decompresses 10 to 100 times faster than BZIP2, DFCM, FSD, GZIP, and PLMI on the three architectures tested. With prediction tables that fit into the CPU's L1 data acache, FPC delivers a guaranteed throughput of six gigabits per second on a 1.6 GHz Itanium 2 system. The C source code and documentation of FPC are posted on-line and have already been downloaded hundreds of times. To evaluate FPC, they gathered 13 real-world scientific datasets from around the globe, including satellite data, crash-simulation data, and messages from HPC systems. Based on the large number of requests they received, they also made these datasets available to the community (with permission of the original sources). While FPC represents a great step forward, it soon became clear that its throughput was too slow for the emerging 10 gigabits per second networks. Hence, no speedup can be gained by including this algorithm in an MPI library. They therefore changed the aim of the second half of the project. Instead of implementing FPC in an MPI library, they refocused their efforts to develop a parallel compression algorithm to further boost the throughput. After all, all modern high-end microprocessors contain multiple CPUs on a

  14. Performance evaluation of the HEP, ELXSI and CRAY X-MP parallel processors on hydrocode test problems

    International Nuclear Information System (INIS)

    Liebrock, L.M.; McGrath, J.F.; Hicks, D.L.

    1986-01-01

    Parallel programming promises improved processing speeds for hydrocodes, magnetohydrocodes, multiphase flow codes, thermal-hydraulics codes, wavecodes and other continuum dynamics codes. This paper presents the results of some investigations of parallel algorithms on three parallel processors: the CRAY X-MP, ELXSI and the HEP computers. Introduction and Background: We report the results of investigations of parallel algorithms for computational continuum dynamics. These programs (hydrocodes, wavecodes, etc.) produce simulations of the solutions to problems arising in the motion of continua: solid dynamics, liquid dynamics, gas dynamics, plasma dynamics, multiphase flow dynamics, thermal-hydraulic dynamics and multimaterial flow dynamics. This report restricts its scope to one-dimensional algorithms such as the von Neumann-Richtmyer (1950) scheme

  15. Parallel rendering

    Science.gov (United States)

    Crockett, Thomas W.

    1995-01-01

    This article provides a broad introduction to the subject of parallel rendering, encompassing both hardware and software systems. The focus is on the underlying concepts and the issues which arise in the design of parallel rendering algorithms and systems. We examine the different types of parallelism and how they can be applied in rendering applications. Concepts from parallel computing, such as data decomposition, task granularity, scalability, and load balancing, are considered in relation to the rendering problem. We also explore concepts from computer graphics, such as coherence and projection, which have a significant impact on the structure of parallel rendering algorithms. Our survey covers a number of practical considerations as well, including the choice of architectural platform, communication and memory requirements, and the problem of image assembly and display. We illustrate the discussion with numerous examples from the parallel rendering literature, representing most of the principal rendering methods currently used in computer graphics.

  16. An Overview of High-performance Parallel Big Data transfers over multiple network channels with Transport Layer Security (TLS) and TLS plus Perfect Forward Secrecy (PFS)

    Energy Technology Data Exchange (ETDEWEB)

    Fang, Chin [SLAC National Accelerator Lab., Menlo Park, CA (United States); Corttrell, R. A. [SLAC National Accelerator Lab., Menlo Park, CA (United States)

    2015-05-06

    This Technical Note provides an overview of high-performance parallel Big Data transfers with and without encryption for data in-transit over multiple network channels. It shows that with the parallel approach, it is feasible to carry out high-performance parallel "encrypted" Big Data transfers without serious impact to throughput. But other impacts, e.g. the energy-consumption part should be investigated. It also explains our rationales of using a statistics-based approach for gaining understanding from test results and for improving the system. The presentation is of high-level nature. Nevertheless, at the end we will pose some questions and identify potentially fruitful directions for future work.

  17. Performance analysis of a parallel Monte Carlo code for simulating solar radiative transfer in cloudy atmospheres using CUDA-enabled NVIDIA GPU

    Science.gov (United States)

    Russkova, Tatiana V.

    2017-11-01

    One tool to improve the performance of Monte Carlo methods for numerical simulation of light transport in the Earth's atmosphere is the parallel technology. A new algorithm oriented to parallel execution on the CUDA-enabled NVIDIA graphics processor is discussed. The efficiency of parallelization is analyzed on the basis of calculating the upward and downward fluxes of solar radiation in both a vertically homogeneous and inhomogeneous models of the atmosphere. The results of testing the new code under various atmospheric conditions including continuous singlelayered and multilayered clouds, and selective molecular absorption are presented. The results of testing the code using video cards with different compute capability are analyzed. It is shown that the changeover of computing from conventional PCs to the architecture of graphics processors gives more than a hundredfold increase in performance and fully reveals the capabilities of the technology used.

  18. Parallel computations

    CERN Document Server

    1982-01-01

    Parallel Computations focuses on parallel computation, with emphasis on algorithms used in a variety of numerical and physical applications and for many different types of parallel computers. Topics covered range from vectorization of fast Fourier transforms (FFTs) and of the incomplete Cholesky conjugate gradient (ICCG) algorithm on the Cray-1 to calculation of table lookups and piecewise functions. Single tridiagonal linear systems and vectorized computation of reactive flow are also discussed.Comprised of 13 chapters, this volume begins by classifying parallel computers and describing techn

  19. Performance of a fine-grained parallel model for multi-group nodal-transport calculations in three-dimensional pin-by-pin reactor geometry

    International Nuclear Information System (INIS)

    Masahiro, Tatsumi; Akio, Yamamoto

    2003-01-01

    A production code SCOPE2 was developed based on the fine-grained parallel algorithm by the red/black iterative method targeting parallel computing environments such as a PC-cluster. It can perform a depletion calculation in a few hours using a PC-cluster with the model based on a 9-group nodal-SP3 transport method in 3-dimensional pin-by-pin geometry for in-core fuel management of commercial PWRs. The present algorithm guarantees the identical convergence process as that in serial execution, which is very important from the viewpoint of quality management. The fine-mesh geometry is constructed by hierarchical decomposition with introduction of intermediate management layer as a block that is a quarter piece of a fuel assembly in radial direction. A combination of a mesh division scheme forcing even meshes on each edge and a latency-hidden communication algorithm provided simplicity and efficiency to message passing to enhance parallel performance. Inter-processor communication and parallel I/O access were realized using the MPI functions. Parallel performance was measured for depletion calculations by the 9-group nodal-SP3 transport method in 3-dimensional pin-by-pin geometry with 340 x 340 x 26 meshes for full core geometry and 170 x 170 x 26 for quarter core geometry. A PC cluster that consists of 24 Pentium-4 processors connected by the Fast Ethernet was used for the performance measurement. Calculations in full core geometry gave better speedups compared to those in quarter core geometry because of larger granularity. Fine-mesh sweep and feedback calculation parts gave almost perfect scalability since granularity is large enough, while 1-group coarse-mesh diffusion acceleration gave only around 80%. The speedup and parallel efficiency for total computation time were 22.6 and 94%, respectively, for the calculation in full core geometry with 24 processors. (authors)

  20. Parallel computing works

    Energy Technology Data Exchange (ETDEWEB)

    1991-10-23

    An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

  1. Effect of dietary genistein on growth performance, digestive enzyme activity, and body composition of Nile tilapia Oreochromis niloticus

    Science.gov (United States)

    Chen, Dong; Wang, Wei; Ru, Shaoguo

    2015-01-01

    An 8-week feeding experiment was performed to evaluate the effect of dietary genistein on growth performance, body composition, and digestive enzymes activity of juvenile Nile tilapia ( Oreochromis niloticus). Four isonitrogenous and isoenergetic diets were formulated containing four graded supplements of genistein: 0, 30, 300, and 3 000 μg/g. Each diet was randomly assigned in triplicate to tanks stocked with 15 juvenile tilapia (10.47±1.24 g). The results show that 30 and 300 μg/g dietary genistein had no significant effect on growth performance of Nile tilapia, but the higher level of genistein (3 000 μg/g) significantly depressed the final body weight and specific growth rate. There was no significant difference in survival rate, feed intake, feed efficiency ratio or whole body composition among all dietary treatments. An assay of digestive enzymes showed that the diet containing 3 000 μg/ggenistein decreased stomach and hepatopancreas protease activity, and amylase activity in the liver and intestine, while a dietary level of 300 μg/g genistein depressed stomach protease and intestine amylase activities. However, no significant difference in stomach amylase activity was found among dietary treatments. Overall, the results of the present study indicate that a high level of dietary genistein (3 000 μg/g, or above) would significantly reduce the growth of Nile tilapia, partly because of its inhibitory effect on the activity of major digestive enzymes. Accordingly, the detrimental effects of genistein, as found in soybean products, should not be ignored when applied as an alternative ingredient source in aquaculture.

  2. Improved performance of parallel surface/packed-bed discharge reactor for indoor VOCs decomposition: optimization of the reactor structure

    International Nuclear Information System (INIS)

    Jiang, Nan; Hui, Chun-Xue; Li, Jie; Lu, Na; Shang, Ke-Feng; Wu, Yan; Mizuno, Akira

    2015-01-01

    The purpose of this paper is to develop a high-efficiency air-cleaning system for volatile organic compounds (VOCs) existing in the workshop of a chemical factory. A novel parallel surface/packed-bed discharge (PSPBD) reactor, which utilized a combination of surface discharge (SD) plasma with packed-bed discharge (PBD) plasma, was designed and employed for VOCs removal in a closed vessel. In order to optimize the structure of the PSPBD reactor, the discharge characteristic, benzene removal efficiency, and energy yield were compared for different discharge lengths, quartz tube diameters, shapes of external high-voltage electrode, packed-bed discharge gaps, and packing pellet sizes, respectively. In the circulation test, 52.8% of benzene was removed and the energy yield achieved 0.79 mg kJ −1 after a 210 min discharge treatment in the PSPBD reactor, which was 10.3% and 0.18 mg kJ −1 higher, respectively, than in the SD reactor, 21.8% and 0.34 mg kJ −1 higher, respectively, than in the PBD reactor at 53 J l −1 . The improved performance in benzene removal and energy yield can be attributed to the plasma chemistry effect of the sequential processing in the PSPBD reactor. The VOCs mineralization and organic intermediates generated during discharge treatment were followed by CO x selectivity and FT-IR analyses. The experimental results indicate that the PSPBD plasma process is an effective and energy-efficient approach for VOCs removal in an indoor environment. (paper)

  3. Frontal cortex and hippocampus neurotransmitter receptor complex level parallels spatial memory performance in the radial arm maze.

    Science.gov (United States)

    Shanmugasundaram, Bharanidharan; Sase, Ajinkya; Miklosi, András G; Sialana, Fernando J; Subramaniyan, Saraswathi; Aher, Yogesh D; Gröger, Marion; Höger, Harald; Bennett, Keiryn L; Lubec, Gert

    2015-08-01

    Several neurotransmitter receptors have been proposed to be involved in memory formation. However, information on receptor complexes (RCs) in the radial arm maze (RAM) is missing. It was therefore the aim of this study to determine major neurotransmitter RCs levels that are modulated by RAM training because receptors are known to work in homo-or heteromeric assemblies. Immediate early gene Arc expression was determined by immunohistochemistry to show if prefrontal cortices (PFC) and hippocampi were activated following RAM training as these regions are known to be mainly implicated in spatial memory. Twelve rats per group, trained and untrained in the twelve arm RAM were used, frontal cortices and hippocampi were taken, RCs in membrane protein were quantified by blue-native PAGE immunoblotting. RCs components were characterised by co-immunoprecipitation followed by mass spectrometrical analysis and by the use of the proximity ligation assay. Arc expression was significantly higher in PFC of trained as compared to untrained rats whereas it was comparable in hippocampi. Frontal cortical levels of RCs containing AMPA receptors GluA1, GluA2, NMDA receptors GluN1 and GluN2A, dopamine receptor D1, acetylcholine nicotinic receptor alpha 7 (nAChR-α7) and hippocampal levels of RCs containing D1, GluN1, GluN2B and nAChR-α7 were increased in the trained group; phosphorylated dopamine transporter levels were decreased in the trained group. D1 and GluN1 receptors were shown to be in the same complex. Taken together, distinct RCs were paralleling performance in the RAM which is relevant for interpretation of previous and design of future work on RCs in memory studies. Copyright © 2015 Elsevier B.V. All rights reserved.

  4. Parallel discrete event simulation

    NARCIS (Netherlands)

    Overeinder, B.J.; Hertzberger, L.O.; Sloot, P.M.A.; Withagen, W.J.

    1991-01-01

    In simulating applications for execution on specific computing systems, the simulation performance figures must be known in a short period of time. One basic approach to the problem of reducing the required simulation time is the exploitation of parallelism. However, in parallelizing the simulation

  5. Parallel reservoir simulator computations

    International Nuclear Information System (INIS)

    Hemanth-Kumar, K.; Young, L.C.

    1995-01-01

    The adaptation of a reservoir simulator for parallel computations is described. The simulator was originally designed for vector processors. It performs approximately 99% of its calculations in vector/parallel mode and relative to scalar calculations it achieves speedups of 65 and 81 for black oil and EOS simulations, respectively on the CRAY C-90

  6. HPDC ´12 : proceedings of the 21st ACM symposium on high-performance parallel and distributed computing, June 18-22, 2012, Delft, The Netherlands

    NARCIS (Netherlands)

    Epema, D.H.J.; Kielmann, T.; Ripeanu, M.

    2012-01-01

    Welcome to ACM HPDC 2012! This is the twenty-first year of HPDC and we are pleased to report that our community continues to grow in size, quality and reputation. The program consists of three days packed with presentations on the latest developments in high-performance parallel and distributed

  7. Enzimas exógenas em dietas de frangos de corte: desempenho Exogenous enzymes in broilers fed diets: performance

    Directory of Open Access Journals (Sweden)

    Nei André Arruda Barbosa

    2012-08-01

    Full Text Available O experimento foi conduzido para avaliar a eficiência de enzimas exógenas em dietas sobre o desempenho de frangos de corte. Um total de 1440 pintos de corte machos (Cobb 500® foram distribuídos em um delineamento inteiramente casualizado em esquema fatorial 2x2 (duas dietas com diferentes densidades nutricionais; com e sem a adição enzimática com 8 repetições de 45 aves em cada unidade experimental. As dietas com diferentes densidades nutricionais foram: controle positivo (níveis normais e um controle negativo (com redução de 4,3 e 4,5% de energia metabolizável; 16,7 e 17,7% de cálcio e 35 e 42,7% de fósforo, nas fases inicial e de crescimento, respectivamente. A suplementação enzimática consistiu da combinação das enzimas fitase (100g t-1 e amilase, xilanase e protease (500g t-1. Foi avaliado o desempenho das aves nas fases de 1 a 21 e 1 a 42 dias de idade. Na fase total de criação, as aves alimentadas com a dieta controle negativo com adição enzimática tiveram consumo de ração (5,97%, peso médio (8,47%, ganho de peso (8,64% e conversão alimentar (2,92% melhores (P0,05 ao grupo alimentado com a dieta controle positivo com ou sem enzima. A adição de enzimas exógenas em dieta de frangos de corte com redução de energia metabolizável, cálcio e fósforo, proporciona um consumo de ração, peso vivo médio e ganho de peso similar a uma dieta com níveis nutricionais adequados.The experiment was conducted to evaluate the effect of enzyme efficiency in diets with and without nutrient reduction on the performance of broilers. Thousand and hundred forty male broiler chicks (Cobb 500® were distributed in a completely randomized design with 4 treatments in a factorial 2x2 (two diets with different nutrient densities, with and without enzyme with 8 replications of 45 birds. The diets were: positive control (normal and a negative control (with a reduction of 4.3% and 4.5% of metabolizable energy, 16.7% and 17.7% of

  8. Development of design technology on thermal-hydraulic performance in tight-lattice rod bundle. 4. Large paralleled simulation by the advanced two-fluid model code

    International Nuclear Information System (INIS)

    Misawa, Takeharu; Yoshida, Hiroyuki; Akimoto, Hajime

    2008-01-01

    In Japan Atomic Energy Agency (JAEA), the Innovative Water Reactor for Flexible Fuel Cycle (FLWR) has been developed. For thermal design of FLWR, it is necessary to develop analytical method to predict boiling transition of FLWR. Japan Atomic Energy Agency (JAEA) has been developing three-dimensional two-fluid model analysis code ACE-3D, which adopts boundary fitted coordinate system to simulate complex shape channel flow. In this paper, as a part of development of ACE-3D to apply to rod bundle analysis, introduction of parallelization to ACE-3D and assessments of ACE-3D are shown. In analysis of large-scale domain such as a rod bundle, even two-fluid model requires large number of computational cost, which exceeds upper limit of memory amount of 1 CPU. Therefore, parallelization was introduced to ACE-3D to divide data amount for analysis of large-scale domain among large number of CPUs, and it is confirmed that analysis of large-scale domain such as a rod bundle can be performed by parallel computation with keeping parallel computation performance even using large number of CPUs. ACE-3D adopts two-phase flow models, some of which are dependent upon channel geometry. Therefore, analyses in the domains, which simulate individual subchannel and 37 rod bundle, are performed, and compared with experiments. It is confirmed that the results obtained by both analyses using ACE-3D show agreement with past experimental result qualitatively. (author)

  9. Parallel algorithms

    CERN Document Server

    Casanova, Henri; Robert, Yves

    2008-01-01

    ""…The authors of the present book, who have extensive credentials in both research and instruction in the area of parallelism, present a sound, principled treatment of parallel algorithms. … This book is very well written and extremely well designed from an instructional point of view. … The authors have created an instructive and fascinating text. The book will serve researchers as well as instructors who need a solid, readable text for a course on parallelism in computing. Indeed, for anyone who wants an understandable text from which to acquire a current, rigorous, and broad vi

  10. Angiotensin I - Converting Enzyme (ACE) gene polymorphism in relation to physical performance, cognition and survival

    DEFF Research Database (Denmark)

    Frederiksen, Henrik; Gaist, David; Bathum, Lise

    2003-01-01

    Studies of younger individuals have suggested an association between ACE genotype and physical and cognitive performance. Using a longitudinal study of elderly twins we studied the association between ACE genotype and physical and cognitive functioning and survival in old age.......Studies of younger individuals have suggested an association between ACE genotype and physical and cognitive performance. Using a longitudinal study of elderly twins we studied the association between ACE genotype and physical and cognitive functioning and survival in old age....

  11. Effect of differences in gas-dynamic behaviour on the separation performance of ultracentrifuges connected in parallel

    International Nuclear Information System (INIS)

    Portoghese, C.C.P.; Buchmann, J.H.

    1996-01-01

    This paper is concerned with the degradation of separation factors occurred when groups of ultracentrifuges having different gas-dynamic behaviour are connected in parallel arrangements. Differences in the gas-dynamic behavior were traduced in terms of different tails pressures for the same operational conditions, that are feed flow rate, product pressure and cut number. A mathematical model describing the ratio of the tails flow rates as a function of the tails pressure ratios and the feed flow rate was developed using experimental data collected from a pair of different ultracentrifuges connected in parallel. The optimization of model parameters was made using Marquardt's algorithm. The model developed was used to simulate the separation factors degradation in some parallel arrangements containing more than two centrifuges. Te obtained results were compared with experimental data collected from different groups of ultracentrifuges. It was observed that the calculated results were in good agreement with experimental data. This mathematical model, which parameters were determined in a two-centrifuges parallel arrangement, is useful to simulate the effect of quantified gas-dynamic differences in the separation factors of groups containing any number of different ultracentrifuges and, consequently, to analyze cascade losses due to this kind of occurrence. (author)

  12. Multi-objective based on parallel vector evaluated particle swarm optimization for optimal steady-state performance of power systems

    DEFF Research Database (Denmark)

    Vlachogiannis, Ioannis (John); Lee, K Y

    2009-01-01

    In this paper the state-of-the-art extended particle swarm optimization (PSO) methods for solving multi-objective optimization problems are represented. We emphasize in those, the co-evolution technique of the parallel vector evaluated PSO (VEPSO), analysed and applied in a multi-objective problem...

  13. Impact of the Flipped Classroom on Student Performance and Retention: A Parallel Controlled Study in General Chemistry

    Science.gov (United States)

    Ryan, Michael D.; Reid, Scott A.

    2016-01-01

    Despite much recent interest in the flipped classroom, quantitative studies are slowly emerging, particularly in the sciences. We report a year-long parallel controlled study of the flipped classroom in a second-term general chemistry course. The flipped course was piloted in the off-semester course in Fall 2014, and the availability of the…

  14. High performance shallow water kernels for parallel overland flow simulations based on FullSWOF2D

    KAUST Repository

    Wittmann, Roland; Bungartz, Hans-Joachim; Neumann, Philipp

    2017-01-01

    -by-step transformation of the second order finite volume scheme in FullSWOF2D towards MPI parallelization. Second, the computational kernels are optimized by the use of templates and a portable vectorization approach. We discuss the load imbalance of the flux computation

  15. Influence of Guava by-Product, Enzyme Supplementation and Gamma Irradiation on Performance and Digestive Utilization of Fattening Rabbits

    International Nuclear Information System (INIS)

    Mekkawy, S.H.; El-Faramawy, A.A.; Zakaria, S. M.

    2000-01-01

    A total number of 32 New Zealand white rabbits weighing about 850 g were used to study the influence of guava by-product (GBP)on growth, feed consumption, feed efficiency, carcass dressing percentage, blood parameters and digestive efficiency. Four diets were formulated to provide about 17% crude fiber (CF). Control diet, diet with 16% (GBP), diet with the same percent of (GBP) supplemented with enzyme supplement and the last one also the same with the third diet in addition to treatment by gamma irradiation (3 kGy). The results indicated that there were no significant differences (P<0.05) between the experimental groups and control in growth, feed consumption, feed efficiency and carcass dressing percentage. Blood parameters (total protein, albumin, globulin, total lipids, and alkaline phosphatase) were within normal range through out the groups. Apparent digestibility coefficient of nutrients ( OM, CP, CF, EE and NFE) were significantly higher in rabbits fed the diet of (GBP) and enzyme supplementation. Our data indicate that (GBP) can replace 16% of alfalfa hay without decreasing growth performance of rabbits

  16. Performance evaluation of enzyme immunoassay for voriconazole therapeutic drug monitoring with automated clinical chemistry analyzers

    Directory of Open Access Journals (Sweden)

    Yongbum Jeon

    2017-08-01

    Full Text Available Objective: Voriconazole is a triazole antifungal developed for the treatment of fungal infectious disease, and the clinical utility of its therapeutic drug monitoring has been evaluated. Recently, a new assay for analyzing the serum voriconazole concentration with an automated clinical chemistry analyzer was developed. We evaluated the performance of the new assay based on standardized protocols. Methods: The analytical performance of the assay was evaluated according to its precision, trueness by recovery, limit of quantitation, linearity, and correlation with results from liquid chromatography-tandem mass spectrometry (LC-MS/MS. The evaluation was performed with the same protocol on two different routine chemistry analyzers. All evaluations were performed according to CLSI Guidelines EP15, EP17, EP6, and EP9 [1–4]. Results: Coefficients of variation for within-run and between-day imprecision were 3.2–5.1% and 1.5–3.0%, respectively, on the two different analyzers for pooled serum samples. The recovery rates were in the range of 95.4–102.2%. The limit of blank was 0.0049 μg/mL, and the limit of detection of the samples was 0.0266–0.0376 μg/mL. The percent recovery at three LoQ levels were 67.9–74.6% for 0.50 μg/mL, 75.5–80.2% for 0.60 μg/mL, and 89.9–96.6% for 0.70 μg/mL. A linear relationship was demonstrated between 0.5 μg/mL and 16.0 μg/mL (R2=0.9995–0.9998. The assay correlated well with LC-MS/MS results (R2=0.9739–0.9828. Conclusions: The assay showed acceptable precision, trueness, linearity, and limit of quantification, and correlated well with LC-MS/MS. Therefore, its analytical performance is satisfactory for monitoring the drug concentration of voriconazole. Keywords: Voriconazole, Antifungal agents, Therapeutic drug monitoring

  17. Performance evaluation of enzyme immunoassay for voriconazole therapeutic drug monitoring with automated clinical chemistry analyzers.

    Science.gov (United States)

    Jeon, Yongbum; Han, Minje; Han, Eun Young; Lee, Kyunghoon; Song, Junghan; Song, Sang Hoon

    2017-08-01

    Voriconazole is a triazole antifungal developed for the treatment of fungal infectious disease, and the clinical utility of its therapeutic drug monitoring has been evaluated. Recently, a new assay for analyzing the serum voriconazole concentration with an automated clinical chemistry analyzer was developed. We evaluated the performance of the new assay based on standardized protocols. The analytical performance of the assay was evaluated according to its precision, trueness by recovery, limit of quantitation, linearity, and correlation with results from liquid chromatography-tandem mass spectrometry (LC-MS/MS). The evaluation was performed with the same protocol on two different routine chemistry analyzers. All evaluations were performed according to CLSI Guidelines EP15, EP17, EP6, and EP9 [1-4]. Coefficients of variation for within-run and between-day imprecision were 3.2-5.1% and 1.5-3.0%, respectively, on the two different analyzers for pooled serum samples. The recovery rates were in the range of 95.4-102.2%. The limit of blank was 0.0049 μg/mL, and the limit of detection of the samples was 0.0266-0.0376 μg/mL. The percent recovery at three LoQ levels were 67.9-74.6% for 0.50 μg/mL, 75.5-80.2% for 0.60 μg/mL, and 89.9-96.6% for 0.70 μg/mL. A linear relationship was demonstrated between 0.5 μg/mL and 16.0 μg/mL ( R 2 =0.9995-0.9998). The assay correlated well with LC-MS/MS results ( R 2 =0.9739-0.9828). The assay showed acceptable precision, trueness, linearity, and limit of quantification, and correlated well with LC-MS/MS. Therefore, its analytical performance is satisfactory for monitoring the drug concentration of voriconazole.

  18. High performance shallow water kernels for parallel overland flow simulations based on FullSWOF2D

    KAUST Repository

    Wittmann, Roland

    2017-01-25

    We describe code optimization and parallelization procedures applied to the sequential overland flow solver FullSWOF2D. Major difficulties when simulating overland flows comprise dealing with high resolution datasets of large scale areas which either cannot be computed on a single node either due to limited amount of memory or due to too many (time step) iterations resulting from the CFL condition. We address these issues in terms of two major contributions. First, we demonstrate a generic step-by-step transformation of the second order finite volume scheme in FullSWOF2D towards MPI parallelization. Second, the computational kernels are optimized by the use of templates and a portable vectorization approach. We discuss the load imbalance of the flux computation due to dry and wet cells and propose a solution using an efficient cell counting approach. Finally, scalability results are shown for different test scenarios along with a flood simulation benchmark using the Shaheen II supercomputer.

  19. Massively parallel mathematical sieves

    Energy Technology Data Exchange (ETDEWEB)

    Montry, G.R.

    1989-01-01

    The Sieve of Eratosthenes is a well-known algorithm for finding all prime numbers in a given subset of integers. A parallel version of the Sieve is described that produces computational speedups over 800 on a hypercube with 1,024 processing elements for problems of fixed size. Computational speedups as high as 980 are achieved when the problem size per processor is fixed. The method of parallelization generalizes to other sieves and will be efficient on any ensemble architecture. We investigate two highly parallel sieves using scattered decomposition and compare their performance on a hypercube multiprocessor. A comparison of different parallelization techniques for the sieve illustrates the trade-offs necessary in the design and implementation of massively parallel algorithms for large ensemble computers.

  20. Design of a chemical batch plant : a study of dedicated parallel lines with intermediate storage and the plant performance

    OpenAIRE

    Verbiest, Floor; Cornelissens, Trijntje; Springael, Johan

    2016-01-01

    Abstract: Production plants worldwide face huge challenges in satisfying high service levels and outperforming competition. These challenges require appropriate strategic decisions on plant design and production strategies. In this paper, we focus on multiproduct chemical batch plants, which are typically equipped with multiple production lines and intermediate storage tanks. First we extend the existing MI(N) LP design models with the concept of parallel production lines, and optimise the as...

  1. Fast volume reconstruction in positron emission tomography: Implementation of four algorithms on a high-performance scalable parallel platform

    International Nuclear Information System (INIS)

    Egger, M.L.; Scheurer, A.H.; Joseph, C.

    1996-01-01

    The issue of long reconstruction times in PET has been addressed from several points of view, resulting in an affordable dedicated system capable of handling routine 3D reconstruction in a few minutes per frame: on the hardware side using fast processors and a parallel architecture, and on the software side, using efficient implementations of computationally less intensive algorithms. Execution times obtained for the PRT-1 data set on a parallel system of five hybrid nodes, each combining an Alpha processor for computation and a transputer for communication, are the following (256 sinograms of 96 views by 128 radial samples): Ramp algorithm 56 s, Favor 81 s and reprojection algorithm of Kinahan and Rogers 187 s. The implementation of fast rebinning algorithms has shown our hardware platform to become communications-limited; they execute faster on a conventional single-processor Alpha workstation: single-slice rebinning 7 s, Fourier rebinning 22 s, 2D filtered backprojection 5 s. The scalability of the system has been demonstrated, and a saturation effect at network sizes above ten nodes has become visible; new T9000-based products lifting most of the constraints on network topology and link throughput are expected to result in improved parallel efficiency and scalability properties

  2. A high performance image processing platform based on CPU-GPU heterogeneous cluster with parallel image reconstroctions for micro-CT

    International Nuclear Information System (INIS)

    Ding Yu; Qi Yujin; Zhang Xuezhu; Zhao Cuilan

    2011-01-01

    In this paper, we report the development of a high-performance image processing platform, which is based on CPU-GPU heterogeneous cluster. Currently, it consists of a Dell Precision T7500 and HP XW8600 workstations with parallel programming and runtime environment, using the message-passing interface (MPI) and CUDA (Compute Unified Device Architecture). We succeeded in developing parallel image processing techniques for 3D image reconstruction of X-ray micro-CT imaging. The results show that a GPU provides a computing efficiency of about 194 times faster than a single CPU, and the CPU-GPU clusters provides a computing efficiency of about 46 times faster than the CPU clusters. These meet the requirements of rapid 3D image reconstruction and real time image display. In conclusion, the use of CPU-GPU heterogeneous cluster is an effective way to build high-performance image processing platform. (authors)

  3. A hybrid, massively parallel implementation of a genetic algorithm for optimization of the impact performance of a metal/polymer composite plate

    KAUST Repository

    Narayanan, Kiran

    2012-07-17

    A hybrid parallelization method composed of a coarse-grained genetic algorithm (GA) and fine-grained objective function evaluations is implemented on a heterogeneous computational resource consisting of 16 IBM Blue Gene/P racks, a single x86 cluster node and a high-performance file system. The GA iterator is coupled with a finite-element (FE) analysis code developed in house to facilitate computational steering in order to calculate the optimal impact velocities of a projectile colliding with a polyurea/structural steel composite plate. The FE code is capable of capturing adiabatic shear bands and strain localization, which are typically observed in high-velocity impact applications, and it includes several constitutive models of plasticity, viscoelasticity and viscoplasticity for metals and soft materials, which allow simulation of ductile fracture by void growth. A strong scaling study of the FE code was conducted to determine the optimum number of processes run in parallel. The relative efficiency of the hybrid, multi-level parallelization method is studied in order to determine the parameters for the parallelization. Optimal impact velocities of the projectile calculated using the proposed approach, are reported. © The Author(s) 2012.

  4. Evaluation of biofilm performance as a protective barrier against biocorrosion using an enzyme electrode.

    Science.gov (United States)

    Soleimani, S; Ormeci, B; Isgor, O B; Papavinasam, S

    2011-01-01

    Sulfide is known to be an important factor in microbiologically influenced corrosion (MIC) of metals and concrete deterioration in wastewater treatment structures and sewer pipelines. A sulfide biosensor was used to determine the effectiveness of Escherichia coli DH5 alpha biofilm as a protective barrier against MIC. The biofilm was shown to be effective in protecting surfaces from sulfide and helping to reduce MIC using amperometric measurements. The results also indicated that the growth conditions of E. coli DH5 alpha may have an impact on the performance of the biofilm as a sulfide barrier. The simple method provided in this work enables the comparison of several microbial biofilms and selection of the ones with potential to prevent MIC in a relatively short time.

  5. Effects of Lipotropic Products on Productive Performance, Liver Lipid and Enzymes Activity in Broiler Chickens

    Directory of Open Access Journals (Sweden)

    Khosravinia H

    2015-12-01

    Full Text Available In a 42-d experiment, 576 one-day-old Vencobb 308 broiler chicks were used to investigate the effects of lecithin extract (0.5 g/kg, choline chloride 60% (1 g/kg and Bio choline (1 g/kg in diets of moderate and high energy in a 4 × 2 factorial arrangement on performance and certain physiological traits in broiler chickens. The inclusion of Bio choline and lecithin extract in the diet significantly increased average daily gain and improved feed conversion ratio  in overall (1 to 42 d period (P < 0.05. Performance efficiency index was improved in the birds fed with Bio choline compared to those fed control diet. Broilers fed diets containing Bio choline and lecithin extract had less abdominal fat percentage than those fed choline chloride or control diet. Regardless of dietary energy level, supplementation of diet with Bio choline, choline chloride and lecithin extract significantly decreased liver lipid concentration (P < 0.05. Aspartate aminotransferase activity increased in the serum of broilers fed high energy diets while it was decreased in the birds received diets containing choline chloride. Lipotropic compounds decreased serum aspartate aminotransferase activity in the birds fed on high energy diets. The addition of Bio choline and lecithin extract to diet significantly decreased serum γ–glutamyltransferase activity (P < 0.05. Results of the present study revealed that dietary supplementation of commercial lipotropic compounds could remove potential detrimental effects from high energy diets through reducing liver fat and maintaining liver health.

  6. Soybean hull and enzyme inclusion effects on diet digestibility and growth performance in beef steers consuming corn-based diets.

    Science.gov (United States)

    Russell, J R; Sexten, W J; Kerley, M S

    2016-06-01

    A beef feedlot study was conducted to determine the effects of increasing soybean hull (SH) inclusion and enzyme addition on diet digestibility and animal performance. The hypothesis was SH inclusion and enzyme addition would increase fiber digestibility with no negative effect on animal performance. Eight treatments (TRT) were arranged in a 4 × 2 factorial using four diets and two enzyme (ENZ) inclusion rates. The diets were composed primarily of whole shell corn (WSC) with 0%, 7%, 14%, or 28% SH replacing corn. The ENZ was a commercial proprietary mix of , and (Cattlemace, R&D Life Sciences, Menomonie, WI) included in the diets at 0% (S0, S7, S14, S28) or 0.045% DM basis (S0e, S7e, S14e, S28e). Eighty steers (287 ± 31 kg, SD) were stratified by weight and blocked into pens with 1 heavy and 1 light pen per TRT (2 pen/TRT, 5 steers/pen). Steers were fed for 70 d with titanium dioxide included in the diets for the final 15 d. Fecal samples were collected on d 70 to determine diet digestibility. Diets were balanced for AA and RDP requirement based on available ME. Individual DMI was measured using a GrowSafe system. Diet, ENZ, and diet × ENZ effects were analyzed using the MIXED procedure of SAS. Initial BW was applied as a covariate for final BW (FBW), and DMI was included as a covariate for all digestibility measures. The diet × ENZ interaction had no effect on FBW, ADG, DMI, or any digestibility measure ( ≥ 0.11). Steers fed ENZ tended to have greater FBW ( = 0.09) and had numerically greater ADG than steers not fed ENZ. Diet influenced DMI ( digestibility ( ≥ 0.2). Diet had an effect on NDF and ADF digestibility ( ≤ 0.04) which decreased as SH inclusion increased. The addition of ENZ tended to decrease NDF digestibility ( = 0.08) but had no effect on ADF digestibility ( = 0.8). Fiber digestibility in WSC diets did not improve with SH inclusion or ENZ addition but steers fed diets with 14% to 28% of WSC replaced by SH and the addition of 0.045% ENZ

  7. Effects of exogenous inosine monophosphate on growth performance, flavor compounds, enzyme activity, and gene expression of muscle tissues in chicken.

    Science.gov (United States)

    Yan, Junshu; Liu, Peifeng; Xu, Liangmei; Huan, Hailin; Zhou, Weiren; Xu, Xiaoming; Shi, Zhendan

    2018-04-01

    The goal of this experiment was to examine effects of diets supplemented with exogenous inosine monophosphate (IMP) on the growth performance, flavor compounds, enzyme activity and gene expression of chicken. A total of 1,500 healthy, 1-day-old male 3-yellow chickens were used for a 52-d experimental period. Individuals were randomly divided into 5 groups (group I, II, III, IV, V) with 6 replicates per group, and fed a basal diet supplemented with 0.0, 0.05, 0.1, 0.2, and 0.3% IMP, respectively. There was no significant response to the increasing dietary IMP level in average daily feed intake (ADFI), average daily gain (ADG), and feed:gain ratio (F/G) (P ≥ 0.05). IMP content of the breast and thigh muscle showed an exponential and linear response to the increasing dietary IMP level (P exogenous IMP was fed. There were significant effects of IMP level in diet on free amino acids (FAA) (exponential, linear and quadratic effect, P exogenous IMP was fed. Dietary IMP supplementation had a quadratic effect on 5΄-NT and the alkaline phosphatase (ALP) enzyme activity in the breast muscle (P exogenous IMP group had the highest (AMPD1) gene expression of the breast muscle and ATIC gene expression of the thigh muscle. These results indicate that dietary IMP did not affect the growth performance of chicken, the diet with 0.2 to 0.3% exogenous IMP is optimal to improve the meat flavor quality in chicken.

  8. Effects of dietary supplementation of resveratrol on performance, egg quality, yolk cholesterol and antioxidant enzyme activity of laying hens.

    Science.gov (United States)

    Feng, Z H; Gong, J G; Zhao, G X; Lin, X; Liu, Y C; Ma, K W

    2017-10-01

    1. This experiment was conducted to evaluate the effects of dietary supplementation of resveratrol on laying performance, egg quality, egg yolk cholesterol and antioxidant enzyme activities of laying hens. 2. A total of 360 Beijing PINK-1 laying hens (60 weeks old) were randomly distributed among five dietary treatments, each of which included 6 replicates of 12 hens. Dietary treatments were basal diet supplemented with 0 (control), 0.5, 1.0, 2.0 and 4.0 g/kg diet resveratrol. The study lasted for 9 weeks including 1 week of adaptation and 8 weeks of the main experimental period. 3. The results indicated that dietary resveratrol significantly improved feed conversion ratios during 5-8 weeks and 1-8 weeks of the trial. Increasing dietary concentrations of the resveratrol linearly improved Haugh unit and albumen height of eggs. 4. The content of total cholesterol (TC), total triglyceride (TG), low density lipoprotein cholesterol (LDL-C), very low density lipoprotein cholesterol (VLDL-C) in serum and cholesterol in yolk was significantly decreased by dietary resveratrol, and there were significant linear correlations between these indexes and resveratrol supplemental levels. 5. Dietary resveratrol supplementation significantly improved serum Glutathione peroxidase (GSH-Px) enzyme activity and decreased serum malondialdehyde (MDA) content in groups with 2.0 and 4.0 g/kg resveratrol as compared to the control, respectively. However, supplementation of resveratrol did not affect the activity of serum superoxide dismutase (SOD). 6. It is concluded that resveratrol supplementation has a positive effect on performance, lipid-related traits and antioxidant activity of laying hens.

  9. [Evaluation of the Performance of Two Kinds of Anti-TP Enzyme-Linked Immunosorbent Assay].

    Science.gov (United States)

    Gao, Nan; Huang, Li-Qin; Wang, Rui; Jia, Jun-Jie; Wu, Shuo; Zhang, Jing; Ge, Hong-Wei

    2018-06-01

    To evaluate the accuracy and precision of 2 kinds of anti-treponema pallidum (anti-TP) ELISA reagents in our laboratory for detecting the anti-TP in voluntary blood donors, so as to provide the data support for use of ELISA reagents after introduction of chemiluminescene immunoassay (CLIA). The route detection of anti-TP was performed by using 2 kinds of ELISA reagents, then 546 responsive positive samples detected by anti-TP ELISA were collected, and the infections status of samples confirmed by treponema pallidum particle agglutination (TPPA) test was identified. The confirmed results of responsive samples detected by 2 kinds of anti-TP ELISA reagents were compared, the accuracy of 2 kinds of anti-TP ELISA reagents was analyzed by drawing ROC and comparing area under curve (AUC), and precision of 2 kinds of anti-TP ELISA reagents was compared by statistical analysis of quality control data from 7.1 2016 to 6.30 2017. There were no statistical difference in confirmed positive rate of responsive samples and weak positive samples between 2 kinds of anti-TP ELISA reagents. The responsive samples detected by 2 kinds of anti-TP ELISA reagents accounted for 85.53%(467/546) of all responsive samples, the positive rate confirmed by TPPA test was 82.87%. 44 responsive samples detected by anti-TP ELISA reagent A and 35 responsive samples detected by anti-TP ELISA reagent B were confirmed to be negative by TPPA test. Comparison of AUC showed that the accuracy of 2 kinds of anti-TP ELISA reagents was more high, the difference between 2 reagents was not statistically significant. The coefficient of variation (CV) of anti-TP ELISA reagent A and B was 14.98% and 18.04% respectively, which met the precision requirement of ELISA test. The accuracy and precision of 2 kinds of anti-TP ELISA reagents used in our laboratory are similar, and using any one of anti-TP ELISA reagents all can satisfy the requirements of blood screening.

  10. Molecular simulation workflows as parallel algorithms: the execution engine of Copernicus, a distributed high-performance computing platform.

    Science.gov (United States)

    Pronk, Sander; Pouya, Iman; Lundborg, Magnus; Rotskoff, Grant; Wesén, Björn; Kasson, Peter M; Lindahl, Erik

    2015-06-09

    Computational chemistry and other simulation fields are critically dependent on computing resources, but few problems scale efficiently to the hundreds of thousands of processors available in current supercomputers-particularly for molecular dynamics. This has turned into a bottleneck as new hardware generations primarily provide more processing units rather than making individual units much faster, which simulation applications are addressing by increasingly focusing on sampling with algorithms such as free-energy perturbation, Markov state modeling, metadynamics, or milestoning. All these rely on combining results from multiple simulations into a single observation. They are potentially powerful approaches that aim to predict experimental observables directly, but this comes at the expense of added complexity in selecting sampling strategies and keeping track of dozens to thousands of simulations and their dependencies. Here, we describe how the distributed execution framework Copernicus allows the expression of such algorithms in generic workflows: dataflow programs. Because dataflow algorithms explicitly state dependencies of each constituent part, algorithms only need to be described on conceptual level, after which the execution is maximally parallel. The fully automated execution facilitates the optimization of these algorithms with adaptive sampling, where undersampled regions are automatically detected and targeted without user intervention. We show how several such algorithms can be formulated for computational chemistry problems, and how they are executed efficiently with many loosely coupled simulations using either distributed or parallel resources with Copernicus.

  11. Increased performance in the short-term water demand forecasting through the use of a parallel adaptive weighting strategy

    Science.gov (United States)

    Sardinha-Lourenço, A.; Andrade-Campos, A.; Antunes, A.; Oliveira, M. S.

    2018-03-01

    Recent research on water demand short-term forecasting has shown that models using univariate time series based on historical data are useful and can be combined with other prediction methods to reduce errors. The behavior of water demands in drinking water distribution networks focuses on their repetitive nature and, under meteorological conditions and similar consumers, allows the development of a heuristic forecast model that, in turn, combined with other autoregressive models, can provide reliable forecasts. In this study, a parallel adaptive weighting strategy of water consumption forecast for the next 24-48 h, using univariate time series of potable water consumption, is proposed. Two Portuguese potable water distribution networks are used as case studies where the only input data are the consumption of water and the national calendar. For the development of the strategy, the Autoregressive Integrated Moving Average (ARIMA) method and a short-term forecast heuristic algorithm are used. Simulations with the model showed that, when using a parallel adaptive weighting strategy, the prediction error can be reduced by 15.96% and the average error by 9.20%. This reduction is important in the control and management of water supply systems. The proposed methodology can be extended to other forecast methods, especially when it comes to the availability of multiple forecast models.

  12. EFFECT OF DIETARY SUPPLEMENTATION OF NON-STARCH POLYSACCHARIDE DEGRADING ENZYMES ON GROWTH PERFORMANCE OF BROILER CHICKS

    Directory of Open Access Journals (Sweden)

    M. A. Nadeem, M. I. Anjum, A. G. Khan and A. Azim

    2005-10-01

    Full Text Available An experiment was conducted to study the performance and carcass parameters of broilers chicks fed diets with and without supplementing non-starch polysaccharide degrading enzymes (NSPDE at the rate of 0.5 g/kg diet. A total of 300 day-old broiler chicks were randomly divided into 12 sets (replicates each comprising 25 chicks and three sets per treatment group, reared on deep litter from 1-42 days post-hatch. Group A was fed diets without NSPDE supplementation, while group B was fed diets supplemented with NSPDE (0.5 g/kg. Group C was fed diets containing 50 kcal/kg less metabolizable energy (ME without NSPDE and group D was fed diets having 50 kcal/kg less ME with NSPDE (0.5 g/kg supplementation. Feed and water were provided ad libitum. Feed intake and feed conversion ratio (FCR from 1-28 days and 1-42 days was significantly (p<0.05 improved in chicks fed NSPDE supplemented diets (groups B and D compared to non-supplemented diets (groups A and C. However, during 29-42 days of growing period enzymes supplementation did not influence feed intake and FCR. Body weight gain, dressing percentage and relative weights of heart, gizzard and shank at 42 days of age was found to be non-significantly different among all groups. However, liver weight reduced significantly (p<0.05 in NSPDE supplemented groups. The study suggested that NSPDE supplementation was beneficial in enhancing feed utilization during the starter phase, while its effects on weight gain, dressing percentage and weights of organs, except liver weight, were found to be non-significant.

  13. Parallel microscope-based fluorescence, absorbance and time-of-flight mass spectrometry detection for high performance liquid chromatography and determination of glucosamine in urine.

    Science.gov (United States)

    Xiong, Bo; Wang, Ling-Ling; Li, Qiong; Nie, Yu-Ting; Cheng, Shuang-Shuang; Zhang, Hui; Sun, Ren-Qiang; Wang, Yu-Jiao; Zhou, Hong-Bin

    2015-11-01

    A parallel microscope-based laser-induced fluorescence (LIF), ultraviolet-visible absorbance (UV) and time-of-flight mass spectrometry (TOF-MS) detection for high performance liquid chromatography (HPLC) was achieved and used to determine glucosamine in urines. First, a reliable and convenient LIF detection was developed based on an inverted microscope and corresponding modulations. Parallel HPLC-LIF/UV/TOF-MS detection was developed by the combination of preceding Microscope-based LIF detection and HPLC coupled with UV and TOF-MS. The proposed setup, due to its parallel scheme, was free of the influence from photo bleaching in LIF detection. Rhodamine B, glutamic acid and glucosamine have been determined to evaluate its performance. Moreover, the proposed strategy was used to determine the glucosamine in urines, and subsequent results suggested that glucosamine, which was widely used in the prevention of the bone arthritis, was metabolized to urines within 4h. Furthermore, its concentration in urines decreased to 5.4mM at 12h. Efficient glucosamine detection was achieved based on a sensitive quantification (LIF), a universal detection (UV) and structural characterizations (TOF-MS). This application indicated that the proposed strategy was sensitive, universal and versatile, and it was capable of improved analysis, especially for analytes with low concentrations in complex samples, compared with conventional HPLC-UV/TOF-MS. Copyright © 2015 Elsevier B.V. All rights reserved.

  14. Pancreatic Enzymes

    Science.gov (United States)

    ... Contact Us DONATE NOW GENERAL DONATION PURPLESTRIDE Pancreatic enzymes Home Facing Pancreatic Cancer Living with Pancreatic Cancer ... and see a registered dietitian. What are pancreatic enzymes? Pancreatic enzymes help break down fats, proteins and ...

  15. An investigation into the accuracy, stability and parallel performance of a highly stable explicit technique for stiff reaction-transport PDEs

    Energy Technology Data Exchange (ETDEWEB)

    Franz, A., LLNL

    1998-02-17

    The numerical simulation of chemically reacting flows is a topic, that has attracted a great deal of current research At the heart of numerical reactive flow simulations are large sets of coupled, nonlinear Partial Differential Equations (PDES). Due to the stiffness that is usually present, explicit time differencing schemes are not used despite their inherent simplicity and efficiency on parallel and vector machines, since these schemes require prohibitively small numerical stepsizes. Implicit time differencing schemes, although possessing good stability characteristics, introduce a great deal of computational overhead necessary to solve the simultaneous algebraic system at each timestep. This thesis examines an algorithm based on a preconditioned time differencing scheme. The algorithm is explicit and permits a large stable time step. An investigation of the algorithm`s accuracy, stability and performance on a parallel architecture is presented

  16. The Effect of Olive Cake, with or Without Enzymes Supplementation, on Growth Performance, Carcass Characteristics, Lymphoid Organs and Lipid Metabolism of Broiler Chickens

    Directory of Open Access Journals (Sweden)

    MA Al-Harthi

    Full Text Available ABSTRACT An experiment was carried out to investigate the effect of using olive cake (OC in broiler diets, when it is supplemented with multi-enzymes or phytase enzyme. The OC was included in isocaloric, isonitorgneous diets at 5 and 10% levels and fed to broilers from 1-28 days of age. Experimental diets were fed with or without either of the two enzymes: galzym or phytase. This resulted in 3 OC levels (0, 5, 10% × 3 enzyme supplementations (no enzyme, galzym enzyme, phytase enzyme. This included nine treatments, and each treatment was replicated eight times with seven broiler chickens each. Feed intake, feed conversion ratio, body weight gain, survival rate, dressing, inner and immune organ´s weights (compared to live body weight; and blood lipids constituents were investigated. According to the findings in this study, it could be concluded that OC is a valuable ingredient and might be included in the broiler diet up to 10% without galzym or phytase enzyme addition. Also, further studies should investigate the possibility of using higher ratios of it or mixed with another by-product in poultry diets; as a very cheap by-product. Moreover, these studies can be associated with suitable additives at different concentrations that might help to increase the utilization of olive cake or at least to keep performance equal to the control. On the other hand, it is worthwhile to follow the positive effect of phytase enzyme on cholesterol and very low density lipoprotein (VLDL concentrations, which may relate it´s use with chicken´s health.

  17. Enzyme Molecules in Solitary Confinement

    Directory of Open Access Journals (Sweden)

    Raphaela B. Liebherr

    2014-09-01

    Full Text Available Large arrays of homogeneous microwells each defining a femtoliter volume are a versatile platform for monitoring the substrate turnover of many individual enzyme molecules in parallel. The high degree of parallelization enables the analysis of a statistically representative enzyme population. Enclosing individual enzyme molecules in microwells does not require any surface immobilization step and enables the kinetic investigation of enzymes free in solution. This review describes various microwell array formats and explores their applications for the detection and investigation of single enzyme molecules. The development of new fabrication techniques and sensitive detection methods drives the field of single molecule enzymology. Here, we introduce recent progress in single enzyme molecule analysis in microwell arrays and discuss the challenges and opportunities.

  18. Parallel computation

    International Nuclear Information System (INIS)

    Jejcic, A.; Maillard, J.; Maurel, G.; Silva, J.; Wolff-Bacha, F.

    1997-01-01

    The work in the field of parallel processing has developed as research activities using several numerical Monte Carlo simulations related to basic or applied current problems of nuclear and particle physics. For the applications utilizing the GEANT code development or improvement works were done on parts simulating low energy physical phenomena like radiation, transport and interaction. The problem of actinide burning by means of accelerators was approached using a simulation with the GEANT code. A program of neutron tracking in the range of low energies up to the thermal region has been developed. It is coupled to the GEANT code and permits in a single pass the simulation of a hybrid reactor core receiving a proton burst. Other works in this field refers to simulations for nuclear medicine applications like, for instance, development of biological probes, evaluation and characterization of the gamma cameras (collimators, crystal thickness) as well as the method for dosimetric calculations. Particularly, these calculations are suited for a geometrical parallelization approach especially adapted to parallel machines of the TN310 type. Other works mentioned in the same field refer to simulation of the electron channelling in crystals and simulation of the beam-beam interaction effect in colliders. The GEANT code was also used to simulate the operation of germanium detectors designed for natural and artificial radioactivity monitoring of environment

  19. Performance characteristics of hybrid MPI/OpenMP implementations of NAS parallel benchmarks SP and BT on large-scale multicore supercomputers

    KAUST Repository

    Wu, Xingfu; Taylor, Valerie

    2011-01-01

    The NAS Parallel Benchmarks (NPB) are well-known applications with the fixed algorithms for evaluating parallel systems and tools. Multicore supercomputers provide a natural programming paradigm for hybrid programs, whereby OpenMP can be used with the data sharing with the multicores that comprise a node and MPI can be used with the communication between nodes. In this paper, we use SP and BT benchmarks of MPI NPB 3.3 as a basis for a comparative approach to implement hybrid MPI/OpenMP versions of SP and BT. In particular, we can compare the performance of the hybrid SP and BT with the MPI counterparts on large-scale multicore supercomputers. Our performance results indicate that the hybrid SP outperforms the MPI SP by up to 20.76%, and the hybrid BT outperforms the MPI BT by up to 8.58% on up to 10,000 cores on BlueGene/P at Argonne National Laboratory and Jaguar (Cray XT4/5) at Oak Ridge National Laboratory. We also use performance tools and MPI trace libraries available on these supercomputers to further investigate the performance characteristics of the hybrid SP and BT.

  20. Performance Characteristics of Hybrid MPI/OpenMP Implementations of NAS Parallel Benchmarks SP and BT on Large-Scale Multicore Clusters

    KAUST Repository

    Wu, X.; Taylor, V.

    2011-01-01

    The NAS Parallel Benchmarks (NPB) are well-known applications with fixed algorithms for evaluating parallel systems and tools. Multicore clusters provide a natural programming paradigm for hybrid programs, whereby OpenMP can be used with the data sharing with the multicores that comprise a node, and MPI can be used with the communication between nodes. In this paper, we use Scalar Pentadiagonal (SP) and Block Tridiagonal (BT) benchmarks of MPI NPB 3.3 as a basis for a comparative approach to implement hybrid MPI/OpenMP versions of SP and BT. In particular, we can compare the performance of the hybrid SP and BT with the MPI counterparts on large-scale multicore clusters, Intrepid (BlueGene/P) at Argonne National Laboratory and Jaguar (Cray XT4/5) at Oak Ridge National Laboratory. Our performance results indicate that the hybrid SP outperforms the MPI SP by up to 20.76 %, and the hybrid BT outperforms the MPI BT by up to 8.58 % on up to 10 000 cores on Intrepid and Jaguar. We also use performance tools and MPI trace libraries available on these clusters to further investigate the performance characteristics of the hybrid SP and BT. © 2011 The Author. Published by Oxford University Press on behalf of The British Computer Society. All rights reserved.

  1. Performance characteristics of hybrid MPI/OpenMP implementations of NAS parallel benchmarks SP and BT on large-scale multicore supercomputers

    KAUST Repository

    Wu, Xingfu

    2011-03-29

    The NAS Parallel Benchmarks (NPB) are well-known applications with the fixed algorithms for evaluating parallel systems and tools. Multicore supercomputers provide a natural programming paradigm for hybrid programs, whereby OpenMP can be used with the data sharing with the multicores that comprise a node and MPI can be used with the communication between nodes. In this paper, we use SP and BT benchmarks of MPI NPB 3.3 as a basis for a comparative approach to implement hybrid MPI/OpenMP versions of SP and BT. In particular, we can compare the performance of the hybrid SP and BT with the MPI counterparts on large-scale multicore supercomputers. Our performance results indicate that the hybrid SP outperforms the MPI SP by up to 20.76%, and the hybrid BT outperforms the MPI BT by up to 8.58% on up to 10,000 cores on BlueGene/P at Argonne National Laboratory and Jaguar (Cray XT4/5) at Oak Ridge National Laboratory. We also use performance tools and MPI trace libraries available on these supercomputers to further investigate the performance characteristics of the hybrid SP and BT.

  2. Performance Characteristics of Hybrid MPI/OpenMP Implementations of NAS Parallel Benchmarks SP and BT on Large-Scale Multicore Clusters

    KAUST Repository

    Wu, X.

    2011-07-18

    The NAS Parallel Benchmarks (NPB) are well-known applications with fixed algorithms for evaluating parallel systems and tools. Multicore clusters provide a natural programming paradigm for hybrid programs, whereby OpenMP can be used with the data sharing with the multicores that comprise a node, and MPI can be used with the communication between nodes. In this paper, we use Scalar Pentadiagonal (SP) and Block Tridiagonal (BT) benchmarks of MPI NPB 3.3 as a basis for a comparative approach to implement hybrid MPI/OpenMP versions of SP and BT. In particular, we can compare the performance of the hybrid SP and BT with the MPI counterparts on large-scale multicore clusters, Intrepid (BlueGene/P) at Argonne National Laboratory and Jaguar (Cray XT4/5) at Oak Ridge National Laboratory. Our performance results indicate that the hybrid SP outperforms the MPI SP by up to 20.76 %, and the hybrid BT outperforms the MPI BT by up to 8.58 % on up to 10 000 cores on Intrepid and Jaguar. We also use performance tools and MPI trace libraries available on these clusters to further investigate the performance characteristics of the hybrid SP and BT. © 2011 The Author. Published by Oxford University Press on behalf of The British Computer Society. All rights reserved.

  3. Digestibility, productive performance, and egg quality of laying hens as affected by dried cassava pulp replacement with corn and enzyme supplementation.

    Science.gov (United States)

    Khempaka, Sutisa; Maliwan, Prapot; Okrathok, Supattra; Molee, Wittawat

    2018-02-24

    Two experiments were conducted to investigate the potential use of dried cassava pulp (DCP) supplemented with enzymes as an alternative feed ingredient in laying hen diets. In experiment 1, 45 laying hens (Isa Brown) aged 45 weeks were placed in individual cages to measure nutrient digestibility for 10 days. Nine dietary treatments were control and DCP as a replacement for corn at 20, 25, 30, and 35% supplemented with mixed enzymes (cellulase, glucanase, and xylanase) at 0.10 and 0.15%. Results showed that the use of DCP at 20-35% added with mixed enzymes had no negative effects on dry matter digestibility, while organic matter digestibility and nitrogen retention decreased with increased DCP up to 30-35% in diets. Both enzyme levels (0.10 and 0.15%) showed similar results on nutrient digestibility and retention. In experiment 2, a total of 336 laying hens aged 32 weeks were randomly allocated to seven dietary treatments (control and DCP-substituted diets at 20, 25, and 30%) supplemented with mixed enzymes (0.10 and 0.15%). Diets incorporated with 20-30% of DCP and supplemented with mixed enzymes at both levels had no significant effects on egg production, egg weight, feed intake, egg mass, feed conversion ratio, protein efficiency ratio, or egg quality, except for egg yolk color being decreased with an increase of DCP in diets (P digestibility, productive performance, or egg quality.

  4. The influence of internal current loop on transient response performance of I-V droop controlled paralleled DC-DC converters

    DEFF Research Database (Denmark)

    Wang, Haojie; Han, Minxiao; Guerrero, Josep M.

    2017-01-01

    The external droop control loop of I-V droop control is designed as a voltage loop with embedded virtual impedance, so the internal current loop plays a major role in the system bandwidth. Thus, in this paper, the influence of internal current loop on transient response performance of I-V droop...... controlled paralleled dc-dc converters is analyzed, which is guided and significant for its industry application. The model which is used for dynamic analysis is built, and the root locus method is used based on the model to analyze the dynamic response of the system by shifting different control parameters...

  5. Comparison of performance between a parallel and a series solar-heat pump system; Solar heat pump system ni okeru heiretsu setsuzoku no seino hikaku

    Energy Technology Data Exchange (ETDEWEB)

    Kanayama, K; Zhao, J; Baba, H; Endo, N [Kitami Institute of Technology, Hokkaido (Japan)

    1997-11-25

    In a solar heat pump system, a single-tank system was fabricated, in which a heat pump is installed in series between a heat collecting tank and a heat storage tank. At the same time, a double-tank system was also fabricated, in which two tanks are assembled into one to which a solar system and a heat pump are connected in parallel. Performance of both systems was analyzed by using measured values and estimated values. Heat collecting efficiency in the double-tank system is higher by about 13 points than in the single-tank system. Nevertheless, the coefficient of performance for the single-tank system is 1.03 to 1.51 times greater than that of the double-tank system. Dependency of the single-tank system on natural energy is higher by 0.3 to 3 points than the double-tank system. Putting the above facts together, it may be said that the single-tank system connecting the solar system and the heat pump in parallel is superior in performance to the double-tank system of the series connection. 3 refs., 5 figs., 2 tabs.

  6. Parallel R

    CERN Document Server

    McCallum, Ethan

    2011-01-01

    It's tough to argue with R as a high-quality, cross-platform, open source statistical software product-unless you're in the business of crunching Big Data. This concise book introduces you to several strategies for using R to analyze large datasets. You'll learn the basics of Snow, Multicore, Parallel, and some Hadoop-related tools, including how to find them, how to use them, when they work well, and when they don't. With these packages, you can overcome R's single-threaded nature by spreading work across multiple CPUs, or offloading work to multiple machines to address R's memory barrier.

  7. Effects of Enzyme Complex Supplementation to a Paddy-based Diet on Performance and Nutrient Digestibility of Meat-type Ducks

    Directory of Open Access Journals (Sweden)

    P. Kang

    2013-02-01

    Full Text Available Paddy rice is rarely used as a feed because of its high fiber content. In this study, two experiments were conducted to study the effects of supplementing an enzyme complex consisting of xylanase, beta-glucanase and cellulase, to paddy-based diets on the performance and nutrient digestibility in meat-type ducks. In the both experiments, meat-type ducks (Cherry Valley were randomly assigned to four treatments. Treatment 1 was a basal diet of corn-soybean; treatment 2 was a basal diet of corn-paddy-soybean; treatment 3, had enzyme complex added to the corn-paddy-soybean basal diet at levels of 0.5 g/kg diet; and treatment 4, had enzyme complex added to the corn-paddy-soybean diet at levels of 1.0 g/kg diet. The results showed that the enzyme complex increased the ADG, and decreased the ADFI and F/G significantly (p0.05. The outcome of this research indicates that the application of enzyme complex made up of xylanase, beta-glucanase, and cellulase, in the corn-paddy-soybean diet, can improve performance and nutrition digestibility in meat-type ducks.

  8. Mesophilic and thermophilic conditions select for unique but highly parallel microbial communities to perform carboxylate platform biomass conversion.

    Directory of Open Access Journals (Sweden)

    Emily B Hollister

    Full Text Available The carboxylate platform is a flexible, cost-effective means of converting lignocellulosic materials into chemicals and liquid fuels. Although the platform's chemistry and engineering are well studied, relatively little is known about the mixed microbial communities underlying its conversion processes. In this study, we examined the metagenomes of two actively fermenting platform communities incubated under contrasting temperature conditions (mesophilic 40°C; thermophilic 55 °C, but utilizing the same inoculum and lignocellulosic feedstock. Community composition segregated by temperature. The thermophilic community harbored genes affiliated with Clostridia, Bacilli, and a Thermoanaerobacterium sp, whereas the mesophilic community metagenome was composed of genes affiliated with other Clostridia and Bacilli, Bacteriodia, γ-Proteobacteria, and Actinobacteria. Although both communities were able to metabolize cellulosic materials and shared many core functions, significant differences were detected with respect to the abundances of multiple Pfams, COGs, and enzyme families. The mesophilic metagenome was enriched in genes related to the degradation of arabinose and other hemicellulose-derived oligosaccharides, and the production of valerate and caproate. In contrast, the thermophilic community was enriched in genes related to the uptake of cellobiose and the transfer of genetic material. Functions assigned to taxonomic bins indicated that multiple community members at either temperature had the potential to degrade cellulose, cellobiose, or xylose and produce acetate, ethanol, and propionate. The results of this study suggest that both metabolic flexibility and functional redundancy contribute to the platform's ability to process lignocellulosic substrates and are likely to provide a degree of stability to the platform's fermentation processes.

  9. Effect of corn replacement with graded levels of wheat screening and enzyme supplementation on performance, blood lipids, viscosity and jejunal histomorphology of finisher broilers

    Energy Technology Data Exchange (ETDEWEB)

    Mazhari, M.; Golian, A.; Kermanshahi, H.

    2015-07-01

    An experiment was carried out to study the effect of corn replacement with five levels of wheat screening (0, 150, 300, 450 and 600 g/kg of diet) with (0.5 g/kg of diet) or without xylanase-glucanase enzyme on performance, blood lipids, viscosity and jejunal histomorphology of finisher broilers (25-42 days of age). Five hundred day-old Ross-308 male broiler chicks were fed by a standard commercial diet up to 24 days of age, then randomly assigned to 10 diets. Each diet was fed to five groups of ten chicks each. There was not significant differences in body weight gain (BWG), feed intake, and feed conversion ratio of birds fed with different levels of wheat screening (WS), whereas enzyme increased (p<0.05) BWG. Different levels of WS and enzyme did not have a significant effect on relative weights of carcass, breast, thigh, and abdominal fat of broilers. Relative weights of gizzard, pancreas, small and large intestine, and relative length of jejunum and jejunal and ileal viscosity were increased (p<0.05) by WS, while were decreased (p<0.05) by enzyme. The serum cholesterol level decreased (p<0.05) by increasing levels of WS. Jejunal histomorphological observations showed (p<0.05) shorter and thicker villus and lower crypt depth by increasing levels of WS, while addition of enzyme to the diets, affected (p<0.05) reversely to these parameters. The results showed that the addition of wheat screening up to an inclusion level of 600 g/kg of diet had no adverse effect on broiler performance in the finisher (25-42 d) phases whereas decreased serum cholesterol levels, increased viscosity and villus atrophy. The dietary administration of exogenous enzyme improved performance parameters and decreased viscosity and villus atrophy of broiler jejunum. (Author)

  10. Seeing or moving in parallel

    DEFF Research Database (Denmark)

    Christensen, Mark Schram; Ehrsson, H Henrik; Nielsen, Jens Bo

    2013-01-01

    a different network, involving bilateral dorsal premotor cortex (PMd), primary motor cortex, and SMA, was more active when subjects viewed parallel movements while performing either symmetrical or parallel movements. Correlations between behavioral instability and brain activity were present in right lateral...... adduction-abduction movements symmetrically or in parallel with real-time congruent or incongruent visual feedback of the movements. One network, consisting of bilateral superior and middle frontal gyrus and supplementary motor area (SMA), was more active when subjects performed parallel movements, whereas...

  11. Parallel Lines

    Directory of Open Access Journals (Sweden)

    James G. Worner

    2017-05-01

    Full Text Available James Worner is an Australian-based writer and scholar currently pursuing a PhD at the University of Technology Sydney. His research seeks to expose masculinities lost in the shadow of Australia’s Anzac hegemony while exploring new opportunities for contemporary historiography. He is the recipient of the Doctoral Scholarship in Historical Consciousness at the university’s Australian Centre of Public History and will be hosted by the University of Bologna during 2017 on a doctoral research writing scholarship.   ‘Parallel Lines’ is one of a collection of stories, The Shapes of Us, exploring liminal spaces of modern life: class, gender, sexuality, race, religion and education. It looks at lives, like lines, that do not meet but which travel in proximity, simultaneously attracted and repelled. James’ short stories have been published in various journals and anthologies.

  12. Modeling of Radiotherapy Linac Source Terms Using ARCHER Monte Carlo Code: Performance Comparison for GPU and MIC Parallel Computing Devices

    Science.gov (United States)

    Lin, Hui; Liu, Tianyu; Su, Lin; Bednarz, Bryan; Caracappa, Peter; Xu, X. George

    2017-09-01

    Monte Carlo (MC) simulation is well recognized as the most accurate method for radiation dose calculations. For radiotherapy applications, accurate modelling of the source term, i.e. the clinical linear accelerator is critical to the simulation. The purpose of this paper is to perform source modelling and examine the accuracy and performance of the models on Intel Many Integrated Core coprocessors (aka Xeon Phi) and Nvidia GPU using ARCHER and explore the potential optimization methods. Phase Space-based source modelling for has been implemented. Good agreements were found in a tomotherapy prostate patient case and a TrueBeam breast case. From the aspect of performance, the whole simulation for prostate plan and breast plan cost about 173s and 73s with 1% statistical error.

  13. Modeling of Radiotherapy Linac Source Terms Using ARCHER Monte Carlo Code: Performance Comparison for GPU and MIC Parallel Computing Devices

    Directory of Open Access Journals (Sweden)

    Lin Hui

    2017-01-01

    Full Text Available Monte Carlo (MC simulation is well recognized as the most accurate method for radiation dose calculations. For radiotherapy applications, accurate modelling of the source term, i.e. the clinical linear accelerator is critical to the simulation. The purpose of this paper is to perform source modelling and examine the accuracy and performance of the models on Intel Many Integrated Core coprocessors (aka Xeon Phi and Nvidia GPU using ARCHER and explore the potential optimization methods. Phase Space-based source modelling for has been implemented. Good agreements were found in a tomotherapy prostate patient case and a TrueBeam breast case. From the aspect of performance, the whole simulation for prostate plan and breast plan cost about 173s and 73s with 1% statistical error.

  14. Effects of carbohydrase enzyme supplementation on performance, eggshell quality, and bone parameters of laying hens fed on maize- and wheat-based diets.

    Science.gov (United States)

    Olgun, Osman; Altay, Y; Yildiz, Alp O

    2018-04-01

    1. This study was conducted to determine the effects of enzyme supplementation of maize/wheat-based diets on the performance, egg quality, and serum and bone parameters of laying hens. 2. During the 12-week experimental period, a total of 72 laying hens aged 52 weeks were randomly distributed among 6 experimental groups. Each experimental group contained 4 replicates, each with three birds. The experiment was a randomised design consisting of a 3 × 2 factorial arrangement, with three levels of wheat substitution and two levels of enzyme (xylanase: 1500.00 U/kg, β-glucanase: 100 000 U/kg, cellulase: 1 000 000 U/kg, α-amylase: 160 000 U/kg) inclusion in the diet. Wheat replaced 0, 50, or 100% of maize with or without 1.0 g/kg enzyme supplementation in iso-nitrogenous and iso-caloric experimental diets. 3. Body weight, egg production, egg weight, egg mass, eggshell thickness, and the feed conversion ratio were adversely affected by the wheat-based diet. The eggshell quality parameters decreased with enzyme supplementation to the diet. 4. Wheat-based diets adversely affected calcium and phosphorus concentrations in the tibia, but the addition of the enzymes to the wheat-based diet prevented the negative effects of wheat-based diets on tibia mineralisation in laying hens. The wheat-based diets tended to reduce plasma mineral contents, and the addition of enzymes tended to affect plasma minerals and biomechanical properties of the tibia positively in laying hens. 5. These results indicate that wheat-based diets in aged laying hens adversely affected the mineral metabolism compared with maize-based diets, and the negative effects of wheat on bone mineralisation can be prevented by enzyme supplementation to the diets in laying hens.

  15. Description and operating performance of a parallel-rail electric-arc system with helium driver gas for the Langley 6-inch expansion tube

    Science.gov (United States)

    Moore, J. A.

    1976-01-01

    A parallel-rail arc-discharge system to heat and pressurize the initial helium driver gas of the Langley 6-inch expansion tube is described. This system was designed for a 2.44-m-long driver vessel rated at 138 MPa, with a distance between rails of 20.3 cm. Electric energy was obtained from a capacitor storage system rated at 12,000 V with a maximum energy of 5 MJ. Tests were performed over a range of energy from 1.74 MJ to the maximum value. The operating experience and system performance are discussed, along with results from a limited number of expansion-tube tests with air and carbon dioxide as test gases.

  16. A task-based parallelism and vectorized approach to 3D Method of Characteristics (MOC) reactor simulation for high performance computing architectures

    Science.gov (United States)

    Tramm, John R.; Gunow, Geoffrey; He, Tim; Smith, Kord S.; Forget, Benoit; Siegel, Andrew R.

    2016-05-01

    In this study we present and analyze a formulation of the 3D Method of Characteristics (MOC) technique applied to the simulation of full core nuclear reactors. Key features of the algorithm include a task-based parallelism model that allows independent MOC tracks to be assigned to threads dynamically, ensuring load balancing, and a wide vectorizable inner loop that takes advantage of modern SIMD computer architectures. The algorithm is implemented in a set of highly optimized proxy applications in order to investigate its performance characteristics on CPU, GPU, and Intel Xeon Phi architectures. Speed, power, and hardware cost efficiencies are compared. Additionally, performance bottlenecks are identified for each architecture in order to determine the prospects for continued scalability of the algorithm on next generation HPC architectures.

  17. Standardized performance tests of collectors of solar thermal energy - A flat-plate copper collector with parallel mylar striping

    Science.gov (United States)

    Johnson, S. M.

    1976-01-01

    Basic test results are reported for a flat plate solar collector whose performance was determined in a solar simulator. The collector was tested over ranges of inlet temperatures, fluxes and one coolant flow rate. Collector efficiency is correlated in terms of inlet temperature and flux level.

  18. SPINning parallel systems software

    International Nuclear Information System (INIS)

    Matlin, O.S.; Lusk, E.; McCune, W.

    2002-01-01

    We describe our experiences in using Spin to verify parts of the Multi Purpose Daemon (MPD) parallel process management system. MPD is a distributed collection of processes connected by Unix network sockets. MPD is dynamic processes and connections among them are created and destroyed as MPD is initialized, runs user processes, recovers from faults, and terminates. This dynamic nature is easily expressible in the Spin/Promela framework but poses performance and scalability challenges. We present here the results of expressing some of the parallel algorithms of MPD and executing both simulation and verification runs with Spin

  19. Selective, Embedded, Just-In-Time Specialization (SEJITS): Portable Parallel Performance from Sequential, Productive, Embedded Domain-Specific Languages

    Science.gov (United States)

    2012-12-01

    approaches to speaker diarization become fast enough to obviate further research in offline approaches. (We have not yet investigated what components of the...Evans, Corinne Fredouille, G Friedland, and O Vinyals. Speaker diarization : A review of recent research. Accepted for publication in ”IEEE Transactions...implementation from [32]. (b) Diarizer application performance as a multiple of real time; “100×” means that 1 second of audio can be processed in 1/100

  20. Parallel Monte Carlo reactor neutronics

    International Nuclear Information System (INIS)

    Blomquist, R.N.; Brown, F.B.

    1994-01-01

    The issues affecting implementation of parallel algorithms for large-scale engineering Monte Carlo neutron transport simulations are discussed. For nuclear reactor calculations, these include load balancing, recoding effort, reproducibility, domain decomposition techniques, I/O minimization, and strategies for different parallel architectures. Two codes were parallelized and tested for performance. The architectures employed include SIMD, MIMD-distributed memory, and workstation network with uneven interactive load. Speedups linear with the number of nodes were achieved

  1. Efficient immobilization of AGE and NAL enzymes onto functional amino resin as recyclable and high-performance biocatalyst.

    Science.gov (United States)

    Cheng, Jian; Zhuang, Wei; Tang, Chenglun; Chen, Yong; Wu, Jinglan; Guo, Ting; Ying, Hanjie

    2017-03-01

    N-Acetylglucosamine-2-epimerase (AGE) and N-acetylneuraminic acid lyase (NAL) were immobilized for synthesis of N-acetylneuraminic acid (Neu5Ac) on three resins: Amberzyme oxirane resin (AOR), poly (styrene-co-DVB)-Br resin (PBR) and amino resin (AR). The loading capacity and immobilized enzyme activity showed that AR was the best carrier. Three methods of glutaraldehyde cross-linking were tested and simultaneous cross-linking and immobilization was demonstrated to be the best method. The functional properties of immobilized AGE and NAL were studied and compared to those of the free enzyme. The highest enzyme activities of free and immobilized AGE were obtained in 0.1 M potassium phosphate buffer at pH 7.5 and a temperature of 37 °C. Comparatively, the highest NAL activities were at pH 8.5. Meanwhile, an increase in K m (from 1.14 to 1.31 mg·mL -1 for AGE and from 1.05 to 1.25 mg·mL -1 for NAL) and a decrease in V max (from 177.53 to 106.37 µg·min -1 mL -1 for AGE and from 126.41 to 95.96 µg·min -1 mL -1 for NAL) were recorded after immobilization. The AR-glutaraldehyde-enzyme system exhibited better thermal stability than the free enzyme, and retained 72% of its initial activity even after eight repeated runs. The apparent activation energy (E a ) of the free and immobilized AGE (NAL) was 117.14 kJ·mol -1 (124.21 kJ·mol -1 ) and 78.45 kJ·mol -1 (66.64 kJ·mol -1 ), respectively, implying that the catalytic efficiency of the immobilized enzyme was restricted by mass-transfer rather than kinetic limit. Subsequently, Neu5Ac production from GlcNAc using immobilized enzymes in one reactor was carried out resulting 101.45 g·L -1 of Neu5Ac and the highest conversion ratio of 82%. This method of enzyme immobilization may have a promising future for Neu5Ac production in industry.

  2. Investigation and performance tests of a new parallel plate ionization chamber with double sensitive volume for measuring diagnostic X-rays

    Energy Technology Data Exchange (ETDEWEB)

    Sharifi, B., E-mail: babak_sharifi88@yahoo.com [Graduate University of Advanced Technology, Kerman (Iran, Islamic Republic of); Zamani Zeinali, H. [Application of Radiation Research School, Nuclear Science and Technology Research Institute, AEOI, Karaj (Iran, Islamic Republic of); Soltani, J.; Negarestani, A. [Graduate University of Advanced Technology, Kerman (Iran, Islamic Republic of); Shahvar, A. [Application of Radiation Research School, Nuclear Science and Technology Research Institute, AEOI, Karaj (Iran, Islamic Republic of)

    2015-01-11

    Medical diagnostic equipment, like diagnostic radiology and mammography require a dosimeter with high accuracy for dosimetry of the diagnostic X-ray beam. Ionization chambers are suitable instruments for dosimetry of diagnostic-range X-ray beams because of their appropriate response and high reliability. This work introduces the design and fabrication of a new parallel plate ionization chamber with a PMMA body, graphite-coated PMMA windows (0.5 mm thick) and a graphite-foil central electrode (0.1 mm thick, 0.7 g/cm{sup 3} dense). This design improves upon the response characteristics of existing designs through the specific choice of materials as well as the appropriate size and arrangement of the ionization chamber components. The results of performance tests conducted at the Secondary Standard Dosimetry laboratory in Karaj-Iran demonstrated the short and long-term stability, the low leakage current, the low directional dependence, and the high ion collection efficiency of the design. Furthermore, the FLUKA Monte Carlo simulations confirmed the low effect of central electrode on this new ionization chamber response. The response characteristics of the parallel plate ionization chamber presented in this work makes the instrument suitable for use as a standard dosimeter in laboratories.

  3. Effects of thai foot massage on balance performance in diabetic patients with peripheral neuropathy: a randomized parallel-controlled trial.

    Science.gov (United States)

    Chatchawan, Uraiwan; Eungpinichpong, Wichai; Plandee, Piyawan; Yamauchi, Junichiro

    2015-04-20

    BACKGROUND Peripheral neuropathy is the most common complications of diabetic patients and leads to loss of plantar cutaneous sensation, movement perception, and body balance. Thai foot massage is an alternative therapy to improve balance. Therefore, the purpose of this study was to investigate the effects of Thai foot massage on balance performance in diabetic patients with peripheral neuropathy. MATERIAL AND METHODS Sixty patients with type-2 diabetes were recruited and randomly assigned into either the Thai foot massage or control groups. The Thai foot massage group received a modified Thai traditional foot massage for 30 min, 3 days per week for 2 weeks. We measured timed up and go (TUG), one leg stance: OLS), the range of motion (ROM) of the foot, and foot sensation (SWMT) before treatment, after the first single session, and after the 2-week treatment. RESULTS After the single treatment session, only the Thai foot massage group showed a significant improvement in TUG. After the 2-week treatment, both Thai foot massage and control groups showed a significant improvement of TUG and OLS (Pfoot massage group showed better improvement in TUG than the control group (pfoot massage group also showed significant improvements in ROM and SWMT after the 2-week treatment. CONCLUSIONS The results of this study suggest that Thai foot massage is a viable alternative treatment for balance performance, ROM of the foot, and the foot sensation in diabetic patients with peripheral neuropathy.

  4. Effect of wheat cultivar and enzyme addition to broiler chicken diets on nutrient digestibility, performance, and apparent metabolizable energy content.

    NARCIS (Netherlands)

    Gutierrez del Alamo Oms, A.; Verstegen, M.W.A.; Hartog, den L.A.; Villamide, M.J.

    2008-01-01

    A total of 5,000 one-day-old male broiler chickens were assigned to 8 different treatments in a 4 x 2 factorial design. Four wheat cultivars (Amiro, Guadalupe, Isengrain, and Horzal) and 2 levels (0 or 1 kg/t of feed) of an enzyme cocktail (Avizyme 1300, xylanase, 2,500 U/kg and protease, 800 U/kg)

  5. Performance and advantages of a soft-core based parallel architecture for energy peak detection in the calorimeter Level 0 trigger for the NA62 experiment at CERN

    International Nuclear Information System (INIS)

    Ammendola, R.; Barbanera, M.; Bizzarri, M.; Bonaiuto, V.; Ceccucci, A.; Simone, N. De; Fantechi, R.; Fucci, A.; Lupi, M.; Ryjov, V.; Checcucci, B.; Papi, A.; Piccini, M.; Federici, L.; Paoluzzi, G.; Salamon, A.; Salina, G.; Sargeni, F.; Venditti, S.

    2017-01-01

    The NA62 experiment at CERN SPS has started its data-taking. Its aim is to measure the branching ratio of the ultra-rare decay K +  → π + ν ν̅ . In this context, rejecting the background is a crucial topic. One of the main background to the measurement is represented by the K +  → π + π 0  decay. In the 1-8.5 mrad decay region this background is rejected by the calorimetric trigger processor (Cal-L0). In this work we present the performance of a soft-core based parallel architecture built on FPGAs for the energy peak reconstruction as an alternative to an implementation completely founded on VHDL language.

  6. Performance and advantages of a soft-core based parallel architecture for energy peak detection in the calorimeter Level 0 trigger for the NA62 experiment at CERN

    Science.gov (United States)

    Ammendola, R.; Barbanera, M.; Bizzarri, M.; Bonaiuto, V.; Ceccucci, A.; Checcucci, B.; De Simone, N.; Fantechi, R.; Federici, L.; Fucci, A.; Lupi, M.; Paoluzzi, G.; Papi, A.; Piccini, M.; Ryjov, V.; Salamon, A.; Salina, G.; Sargeni, F.; Venditti, S.

    2017-03-01

    The NA62 experiment at CERN SPS has started its data-taking. Its aim is to measure the branching ratio of the ultra-rare decay K+ → π+ν ν̅ . In this context, rejecting the background is a crucial topic. One of the main background to the measurement is represented by the K+ → π+π0 decay. In the 1-8.5 mrad decay region this background is rejected by the calorimetric trigger processor (Cal-L0). In this work we present the performance of a soft-core based parallel architecture built on FPGAs for the energy peak reconstruction as an alternative to an implementation completely founded on VHDL language.

  7. Parallel Framework for Cooperative Processes

    Directory of Open Access Journals (Sweden)

    Mitică Craus

    2005-01-01

    Full Text Available This paper describes the work of an object oriented framework designed to be used in the parallelization of a set of related algorithms. The idea behind the system we are describing is to have a re-usable framework for running several sequential algorithms in a parallel environment. The algorithms that the framework can be used with have several things in common: they have to run in cycles and the work should be possible to be split between several "processing units". The parallel framework uses the message-passing communication paradigm and is organized as a master-slave system. Two applications are presented: an Ant Colony Optimization (ACO parallel algorithm for the Travelling Salesman Problem (TSP and an Image Processing (IP parallel algorithm for the Symmetrical Neighborhood Filter (SNF. The implementations of these applications by means of the parallel framework prove to have good performances: approximatively linear speedup and low communication cost.

  8. Integrating enzyme immobilization and protein engineering: An alternative path for the development of novel and improved industrial biocatalysts.

    Science.gov (United States)

    Bernal, Claudia; Rodríguez, Karen; Martínez, Ronny

    2018-06-09

    Enzyme immobilization often achieves reusable biocatalysts with improved operational stability and solvent resistance. However, these modifications are generally associated with a decrease in activity or detrimental modifications in catalytic properties. On the other hand, protein engineering aims to generate enzymes with increased performance at specific conditions by means of genetic manipulation, directed evolution and rational design. However, the achieved biocatalysts are generally generated as soluble enzymes, -thus not reusable- and their performance under real operational conditions is uncertain. Combined protein engineering and enzyme immobilization approaches have been employed as parallel or consecutive strategies for improving an enzyme of interest. Recent reports show efforts on simultaneously improving both enzymatic and immobilization components through genetic modification of enzymes and optimizing binding chemistry for site-specific and oriented immobilization. Nonetheless, enzyme engineering and immobilization are usually performed as separate workflows to achieve improved biocatalysts. In this review, we summarize and discuss recent research aiming to integrate enzyme immobilization and protein engineering and propose strategies to further converge protein engineering and enzyme immobilization efforts into a novel "immobilized biocatalyst engineering" research field. We believe that through the integration of both enzyme engineering and enzyme immobilization strategies, novel biocatalysts can be obtained, not only as the sum of independently improved intrinsic and operational properties of enzymes, but ultimately tailored specifically for increased performance as immobilized biocatalysts, potentially paving the way for a qualitative jump in the development of efficient, stable biocatalysts with greater real-world potential in challenging bioprocess applications. Copyright © 2018. Published by Elsevier Inc.

  9. The effect of dietary faba bean and non-starch polysaccharide degrading enzymes on the growth performance and gut physiology of young turkeys.

    Science.gov (United States)

    Mikulski, D; Juskiewicz, J; Przybylska-Gornowicz, B; Sosnowska, E; Slominski, B A; Jankowski, J; Zdunczyk, Z

    2017-12-01

    The aim of this study was to investigate the effect of dietary replacement of soya bean meal (SBM) with faba bean (FB) and a blend of non-starch polysaccharide (NSP) degrading enzymes on the gastrointestinal function, growth performance and welfare of young turkeys (1 to 56 days of age). An experiment with a 2×2 factorial design was performed to compare the efficacy of four diets: a SBM-based diet and a diet containing FB, with and without enzyme supplementation (C, FB, CE and FBE, respectively). In comparison with groups C, higher dry matter content and lower viscosity of the small intestinal digesta were noted in groups FB. The content of short-chain fatty acids (SCFAs) in the small intestinal digesta was higher in groups FB, but SCFA concentrations in the caecal digesta were comparable in groups C and FB. In comparison with control groups, similar BW gains, higher feed conversion ratio (FCR), higher dry matter content of excreta and milder symptoms of footpad dermatitis (FPD) were noted in groups FB. Enzyme supplementation increased the concentrations of acetate, butyrate and total SCFAs, but it did not increase the SCFA pool in the caecal digesta. The enzymatic preparation significantly improved FCR, reduced excreta hydration and the severity of FPD in turkeys. It can be concluded that in comparison with the SBM-based diet, the diet containing 30% of FB enables to achieve comparable BW gains accompanied by lower feed efficiency during the first 8 weeks of rearing. Non-starch polysaccharide-degrading enzymes can be used to improve the nutritional value of diets for young turkeys, but more desirable results of enzyme supplementation were noted in the SBM-based diet than in the FB-based diet.

  10. Influence of a direct-fed microbial and xylanase enzyme on the dietary energy uptake efficiency and performance of broiler chickens.

    Science.gov (United States)

    Murugesan, Ganapathi Raj; Persia, Michael E

    2015-09-01

    Efficacy of a multi-strain direct-fed microbial product (PoultryStar(®) ME; PS) and a xylanase enzyme product on the dietary energy utilization efficiency and resulting performance in broiler chickens was evaluated. Apart from performance parameters, cecal and serum metabolites and activities of hepatic enzymes involved in energy metabolism were also determined. Ross 308 chicks were fed one of four experimental diets [control (CON), CON + PS, CON + xylanase and CON + PS + xylanase] using a 2 × 2 factorial arrangement from 1-21 days of age. Cecal proportions of propionate and butyrate, as well as total short-chain fatty acid concentration were increased (P energy uptake and hepatic energy retention. The combination additively increased the FCR, suggesting involvement of synergistic modes of actions. © 2014 Society of Chemical Industry.

  11. Significance of membrane bioreactor design on the biocatalytic performance of glucose oxidase and catalase: Free vs. immobilized enzyme systems

    DEFF Research Database (Denmark)

    Morthensen, Sofie Thage; Meyer, Anne S.; Jørgensen, Henning

    2017-01-01

    Membrane separation of xylose and glucose can be accomplished via oxidation of glucose to gluconic acid by enzymatic glucose oxidase catalysis. Oxygen for this reaction can be supplied via decomposition of hydrogen peroxide by enzymatic catalase catalysis. In order to maximize the biocatalytic...... productivity of glucose oxidase and catalase (gluconic acid yield per total amount of enzyme) the following system set-ups were compared: immobilization of glucose oxidase alone; co-immobilization of glucose oxidase and catalase; glucose oxidase and catalase free in the membrane bioreactor. Fouling......-induced enzyme immobilization in the porous support of an ultrafiltration membrane was used as strategy for entrapment of glucose oxidase and catalase. The biocatalytic productivity of the membrane reactor was found to be highly related to the oxygen availability, which in turn depended on the reactor...

  12. Identification of Performance Problems in a Commercial Human Immunodeficiency Virus Type 1 Enzyme Immunoassay by Multiuser External Quality Control Monitoring and Real-Time Data Analysis▿ †

    OpenAIRE

    Kim, J.; Swantee, C.; Lee, B.; Gunning, H.; Chow, A.; Sidaway, F.; Sherlock, C.; Garceau, R.; Dimech, W.; Malloch, L.

    2009-01-01

    In June 2005, a pilot program was implemented in Canadian laboratories to monitor the performance of the Abbott human immunodeficiency virus types 1 and 2 (HIV-1/2) gO enzyme immunoassay (EIA). Two different external quality control (QC) reagents and a “real-time” software analysis program were evaluated. In November 2005, higher-than-expected calibrator rate values in these kits were first reported at the Ontario Ministry of Health (Etobicoke), followed by the Alberta Provincial Public Healt...

  13. Parallel hierarchical radiosity rendering

    Energy Technology Data Exchange (ETDEWEB)

    Carter, Michael [Iowa State Univ., Ames, IA (United States)

    1993-07-01

    In this dissertation, the step-by-step development of a scalable parallel hierarchical radiosity renderer is documented. First, a new look is taken at the traditional radiosity equation, and a new form is presented in which the matrix of linear system coefficients is transformed into a symmetric matrix, thereby simplifying the problem and enabling a new solution technique to be applied. Next, the state-of-the-art hierarchical radiosity methods are examined for their suitability to parallel implementation, and scalability. Significant enhancements are also discovered which both improve their theoretical foundations and improve the images they generate. The resultant hierarchical radiosity algorithm is then examined for sources of parallelism, and for an architectural mapping. Several architectural mappings are discussed. A few key algorithmic changes are suggested during the process of making the algorithm parallel. Next, the performance, efficiency, and scalability of the algorithm are analyzed. The dissertation closes with a discussion of several ideas which have the potential to further enhance the hierarchical radiosity method, or provide an entirely new forum for the application of hierarchical methods.

  14. Parallel k-means++

    Energy Technology Data Exchange (ETDEWEB)

    2017-04-04

    A parallelization of the k-means++ seed selection algorithm on three distinct hardware platforms: GPU, multicore CPU, and multithreaded architecture. K-means++ was developed by David Arthur and Sergei Vassilvitskii in 2007 as an extension of the k-means data clustering technique. These algorithms allow people to cluster multidimensional data, by attempting to minimize the mean distance of data points within a cluster. K-means++ improved upon traditional k-means by using a more intelligent approach to selecting the initial seeds for the clustering process. While k-means++ has become a popular alternative to traditional k-means clustering, little work has been done to parallelize this technique. We have developed original C++ code for parallelizing the algorithm on three unique hardware architectures: GPU using NVidia's CUDA/Thrust framework, multicore CPU using OpenMP, and the Cray XMT multithreaded architecture. By parallelizing the process for these platforms, we are able to perform k-means++ clustering much more quickly than it could be done before.

  15. Assessment of enzyme supplementation on growth performance and apparent nutrient digestibility in diets containing undecorticated sunflower seed meal in layer chicks.

    Science.gov (United States)

    Fafiolu, A O; Oduguwa, O O; Jegede, A V; Tukura, C C; Olarotimi, I D; Teniola, A A; Alabi, J O

    2015-08-01

    Six hundred and forty one-day-old layer chicks were used to investigate the effect of replacing soybean meal with undecorticated sunflower seed meal protein for protein at 0, 25, 50, and 75% levels. Diets were without enzyme supplementation or with enzyme supplementation with four replications of twenty birds. Growth performance and nutrient utilization were determined. Proximate composition of the undecorticated sunflower seed meal used revealed that undecorticated sunflower seed meal contained 925.9, 204.5, 336.2, 215.1, 52.0 and 192.2g/kg dry matter, crude protein, ether extract, crude fibre, ash and soluble carbohydrates, respectively. Results showed that the final weight of 484.4 g/bird was obtained for birds on 75% undecorticated sunflower seed meal diet, while the lowest value of 472.2g/bird was obtained for birds on 25% undecorticated sunflower seed meal diet. Weight gain per bird per day was not significantly (P > 0.05) affected as the level of undecorticated sunflower seed meal increased in the diets. Feed intake per bird per day increased (P meal inclusion in the diet. However, enzyme supplementation of the diets showed marked (P meal inclusion in the diet while crude protein digestibility progressively reduced (P meal increased in the diet. Ash digestibility values were, however, increased (P meal increased in the diets. Birds on enzyme-supplemented diets consistently showed superior (P meal in the diets of layer chicks showed a similar body weight gain/bird/day with the control. Undecorticated sunflower seed meal used in this study is a good source of crude protein, ether extract, and amino acids and had the potential to serve as feeding stuffs as replacement for soybeans. The nutritive value of undecorticated sunflower seed meal was improved for layer chicks with exogenous enzyme supplementation. © 2015 Poultry Science Association Inc.

  16. Performance analysis of a threshold-based parallel multiple beam selection scheme for WDM-based systems for Gamma-Gamma distributions

    KAUST Repository

    Nam, Sung Sik

    2017-03-02

    In this paper, we statistically analyze the performance of a threshold-based parallel multiple beam selection scheme (TPMBS) for Free-space optical (FSO) based system with wavelength division multiplexing (WDM) in cases where a pointing error has occurred for practical consideration over independent identically distributed (i.i.d.) Gamma-Gamma fading conditions. Specifically, we statistically analyze the characteristics in operation under conventional heterodyne detection (HD) scheme for both adaptive modulation (AM) case in addition to non-AM case (i.e., coherentnon-coherent binary modulation). Then, based on the statistically derived results, we evaluate the outage probability (CDF) of a selected beam, the average spectral efficiency (ASE), the average number of selected beams (ANSB), and the average bit error rate (BER). Some selected results shows that we can obtain the higher spectral efficiency and simultaneously reduce the potential increasing of the complexity of implementation caused by applying the selection based beam selection scheme without a considerable performance loss.

  17. Distributing coil elements in three dimensions enhances parallel transmission multiband RF performance: A simulation study in the human brain at 7 Tesla.

    Science.gov (United States)

    Wu, Xiaoping; Tian, Jinfeng; Schmitter, Sebastian; Vaughan, J Tommy; Uğurbil, Kâmil; Van de Moortele, Pierre-François

    2016-06-01

    We explore the advantages of using a double-ring radiofrequency (RF) array and slice orientation to design parallel transmission (pTx) multiband (MB) pulses for simultaneous multislice (SMS) imaging with whole-brain coverage at 7 Tesla (T). A double-ring head array with 16 elements split evenly in two rings stacked in the z-direction was modeled and compared with two single-ring arrays consisting of 8 or 16 elements. The array performance was evaluated by designing band-specific pTx MB pulses with local specific absorption rate (SAR) control. The impact of slice orientations was also investigated. The double-ring array consistently and significantly outperformed the other two single-ring arrays, with peak local SAR reduced by up to 40% at a fixed excitation error of 0.024. For all three arrays, exciting sagittal or coronal slices yielded better RF performance than exciting axial or oblique slices. A double-ring RF array can be used to drastically improve SAR versus excitation fidelity tradeoff for pTx MB pulse design for brain imaging at 7 T; therefore, it is preferable against single-ring RF array designs when pursuing various biomedical applications of pTx SMS imaging. In comparing the stripline arrays, coronal and sagittal slices are more advantageous than axial and oblique slices for pTx MB pulses. Magn Reson Med 75:2464-2472, 2016. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.

  18. Systematic approach for deriving feasible mappings of parallel algorithms to parallel computing platforms

    NARCIS (Netherlands)

    Arkin, Ethem; Tekinerdogan, Bedir; Imre, Kayhan M.

    2017-01-01

    The need for high-performance computing together with the increasing trend from single processor to parallel computer architectures has leveraged the adoption of parallel computing. To benefit from parallel computing power, usually parallel algorithms are defined that can be mapped and executed

  19. Dietary inclusion of raw faba bean instead of soybean meal and enzyme supplementation in laying hens: Effect on performance and egg quality

    Directory of Open Access Journals (Sweden)

    M.E. Abd El-Hack

    2017-02-01

    Full Text Available An experiment was conducted with 160 Hisex Brown laying hens to evaluate the effect of different inclusion levels of faba bean (FB and enzyme supplementation on productive performance and egg quality parameters. The experimental diets consisted of five levels of FB: 0% (control, 25%, 50%, 75% and 100%, substituting soybean meal (SBM, and two levels of enzyme supplementation (0 or 250 mg/kg. Each dietary treatment was assigned to four replicate groups and the experiment lasted 22 weeks. A positive relationship (P  0.05. The main effect of FB levels replacing for SBM affected (P < 0.05 yolk and shell percentages, yolk index, yolk to albumen ratio, shell thickness and egg shape index. It can be concluded that FB and enzyme supplementation could be included in hens diet at less than 50% instead of SBM to support egg productive performance, however higher raw FB levels negatively affected egg production indices and quality.

  20. Magnetically responsive enzyme powders

    Energy Technology Data Exchange (ETDEWEB)

    Pospiskova, Kristyna, E-mail: kristyna.pospiskova@upol.cz [Regional Centre of Advanced Technologies and Materials, Palacky University, Slechtitelu 11, 783 71 Olomouc (Czech Republic); Safarik, Ivo, E-mail: ivosaf@yahoo.com [Regional Centre of Advanced Technologies and Materials, Palacky University, Slechtitelu 11, 783 71 Olomouc (Czech Republic); Department of Nanobiotechnology, Institute of Nanobiology and Structural Biology of GCRC, Na Sadkach 7, 370 05 Ceske Budejovice (Czech Republic)

    2015-04-15

    Powdered enzymes were transformed into their insoluble magnetic derivatives retaining their catalytic activity. Enzyme powders (e.g., trypsin and lipase) were suspended in various liquid media not allowing their solubilization (e.g., saturated ammonium sulfate and highly concentrated polyethylene glycol solutions, ethanol, methanol, 2-propanol) and subsequently cross-linked with glutaraldehyde. Magnetic modification was successfully performed at low temperature in a freezer (−20 °C) using magnetic iron oxides nano- and microparticles prepared by microwave-assisted synthesis from ferrous sulfate. Magnetized cross-linked enzyme powders were stable at least for two months in water suspension without leakage of fixed magnetic particles. Operational stability of magnetically responsive enzymes during eight repeated reaction cycles was generally without loss of enzyme activity. Separation of magnetically modified cross-linked powdered enzymes from reaction mixtures was significantly simplified due to their magnetic properties. - Highlights: • Cross-linked enzyme powders were prepared in various liquid media. • Insoluble enzymes were magnetized using iron oxides particles. • Magnetic iron oxides particles were prepared by microwave-assisted synthesis. • Magnetic modification was performed under low (freezing) temperature. • Cross-linked powdered trypsin and lipase can be used repeatedly for reaction.

  1. Effect of feeding alfalfa hay or Tifton 85 bermudagrass haylage with or without a cellulase enzyme on performance of Holstein cows.

    Science.gov (United States)

    Bernard, J K; Castro, J J; Mullis, N A; Adesogan, A T; West, J W; Morantes, G

    2010-11-01

    Forty-four lactating Holstein cows (173±30 DIM, 42.5±6.8 kg of milk, 4.03±0.69% fat, 674±78 kg of body weight) were used in an 8-wk, completely randomized trial with a 2 × 2 factorial arrangement of treatments to determine the effect of forage source and supplemental cellulase enzyme on performance. Treatments included 2 forage combinations (corn silage plus 12.2% dry matter, DM, from either alfalfa hay or Tifton 85 bermudagrass haylage) with or without a commercial cellulase enzyme applied to the total mixed ration at the rate of 4 g/head per day (Promote N.E.T.-L, Cargill Animal Nutrition, Minneapolis, MN). Experimental diets were formulated to provide similar concentrations of protein (16.5% of DM), energy (1.63 Mcal of net energy for lactation/kg of DM), and neutral detergent fiber (41.7% of DM) and were fed once daily as a total mixed ration behind Calan doors for ad libitum intake. The cellulase enzyme provided 1,200 cellulase units of activity/g of product and was applied to the total mixed ration and allowed to mix for 5min before feeding. Before beginning the trial, all cows were trained to use Calan (American Calan, Northwood, NH) doors and then fed the alfalfa hay-based diet for 2 wk. Data collected during wk 2 were used as a covariate in the statistical analysis. At the beginning of the 6-wk experimental period, cows were assigned randomly to 1 of the 4 experimental diets. No interactions were observed between forage and enzyme for any measures. Daily DM intake; milk yield; concentrations of milk fat, true protein, lactose, and solids not fat; energy-corrected milk yield; and dairy efficiency were not different among alfalfa or Tifton 85 bermudagrass rations with or without cellulase enzyme supplementation. The results of this trial indicate that Tifton 85 bermudagrass haylage can replace alfalfa hay in diets fed to high-producing, lactating dairy cows without depressing DM intake or milk yield when rations are balanced for NDF. Although

  2. Parallel integer sorting with medium and fine-scale parallelism

    Science.gov (United States)

    Dagum, Leonardo

    1993-01-01

    Two new parallel integer sorting algorithms, queue-sort and barrel-sort, are presented and analyzed in detail. These algorithms do not have optimal parallel complexity, yet they show very good performance in practice. Queue-sort designed for fine-scale parallel architectures which allow the queueing of multiple messages to the same destination. Barrel-sort is designed for medium-scale parallel architectures with a high message passing overhead. The performance results from the implementation of queue-sort on a Connection Machine CM-2 and barrel-sort on a 128 processor iPSC/860 are given. The two implementations are found to be comparable in performance but not as good as a fully vectorized bucket sort on the Cray YMP.

  3. Parallel high-performance grid computing: capabilities and opportunities of a novel demanding service and business class allowing highest resource efficiency

    NARCIS (Netherlands)

    F.N. Kepper (Nick); R. Ettig (Ramona); F. Dickmann (Frank); R. Stehr (Rene); F.G. Grosveld (Frank); G. Wedemann (Gero); T.A. Knoch (Tobias)

    2010-01-01

    textabstractThe hardware and software requirements for parallel applications depend on the problem size, type and the number particles / parameters, the degree of parallelization possible, the load balancing over different processors / memory, the calculation type and the input / output and

  4. Effects of diets containing vegetable protein concentrates on performance and activity of digestive enzymes in silver catfish (Rhamdia quelen

    Directory of Open Access Journals (Sweden)

    Naglezi de Menezes Lovatto

    2014-02-01

    Full Text Available The purpose of study was to evaluate the effect of using protein concentrates crambe and sunflower meal in the diet of silver catfish juveniles, as substitute for animal protein source. A total of 300 silver catfish had been separate in 15 experimental units of 280 L, totaling five treatments with three replications. We evaluated two levels (25% and 50% replacement of the meat and bone meal by protein concentrates of crambe and sunflower meals. Evaluated growth parameters, biological index and digestive enzymes in fish. There was no statistical difference for mass (g and standard length (cm, but the fish diet CPFCr-25% had greater total length (cm. No difference in dry matter, crude protein and total protein deposited (calculated. However, there was a higher concentration of ash in the carcass of the animals fed the control diet and CPFCr-50% in relation to diet CPFG- 50%, in addition, higher levels of lipids in fish fed diet CPFG-50%. No significant differences for hepatosomatic index, digestive somatic index and intestinal quotient of animals subjected to different treatments. The activity of digestive enzymes trypsin and chymotrypsin did not change. There was increased activity of acid protease. The quantitative and qualitative increase in protein concentration from this fraction allows the use of bran protein concentrates crambe and sunflower as substitutes for animal protein source.

  5. The Galley Parallel File System

    Science.gov (United States)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    Most current multiprocessor file systems are designed to use multiple disks in parallel, using the high aggregate bandwidth to meet the growing I/0 requirements of parallel scientific applications. Many multiprocessor file systems provide applications with a conventional Unix-like interface, allowing the application to access multiple disks transparently. This interface conceals the parallelism within the file system, increasing the ease of programmability, but making it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. In addition to providing an insufficient interface, most current multiprocessor file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic scientific multiprocessor workloads. We discuss Galley's file structure and application interface, as well as the performance advantages offered by that interface.

  6. Ultrascalable petaflop parallel supercomputer

    Science.gov (United States)

    Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton On Hudson, NY; Chiu, George [Cross River, NY; Cipolla, Thomas M [Katonah, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Hall, Shawn [Pleasantville, NY; Haring, Rudolf A [Cortlandt Manor, NY; Heidelberger, Philip [Cortlandt Manor, NY; Kopcsay, Gerard V [Yorktown Heights, NY; Ohmacht, Martin [Yorktown Heights, NY; Salapura, Valentina [Chappaqua, NY; Sugavanam, Krishnan [Mahopac, NY; Takken, Todd [Brewster, NY

    2010-07-20

    A massively parallel supercomputer of petaOPS-scale includes node architectures based upon System-On-a-Chip technology, where each processing node comprises a single Application Specific Integrated Circuit (ASIC) having up to four processing elements. The ASIC nodes are interconnected by multiple independent networks that optimally maximize the throughput of packet communications between nodes with minimal latency. The multiple networks may include three high-speed networks for parallel algorithm message passing including a Torus, collective network, and a Global Asynchronous network that provides global barrier and notification functions. These multiple independent networks may be collaboratively or independently utilized according to the needs or phases of an algorithm for optimizing algorithm processing performance. The use of a DMA engine is provided to facilitate message passing among the nodes without the expenditure of processing resources at the node.

  7. Effects of Enzyme Supplementation on Productive Performance and Egg Quality of Laying Hens fed Diets Containing Graded Levels of Whole Date Waste

    Directory of Open Access Journals (Sweden)

    Torki M

    2014-12-01

    Full Text Available The objective of present study was to determine the effects of β-mannanase-based enzyme (Hemicell® on productive performance and egg quality in diets containing graded levels of Whole date waste (WDW fed to laying hens. A total of 336 Hy-line leghorn hens after production peak were randomly divided into 56 cages. Eight iso-energetic and iso-nitrogenous experimental diets in a 4 × 2 factorial arrangement including four levels of WDW (0, 10, 20 and 30% and 2 concentrations of supplemental β-mannanase (0 or 0.06 % were prepared. Each dietary treatment was fed to 7 cages (6 birds/cage from 32 to 38 wk of age. During the experiment, daily egg production, egg weight and feed intake were measured. At the 6th wk, egg quality traits were also recorded. The results showed that there was no interaction between WDW inclusion and enzyme supplementation on performance and egg traits. Dietary supplementation of WDW more than 10% significantly decreased egg production and egg mass compared to no WDW recipient hens (control diet during the entire experiment (P. Inclusion of 30% WDW to the diet, significantly increased overall feed conversion ratio compared to the control group (P. The treatment with 20 and 30% WDW also resulted in lower eggshell thickness as compared to 10% WDW (P. The dietary inclusion of 10% WDW also increased yolk index as compared to the control and 30% WDW groups (P. Enzyme supplementation had no significant effect on productive performance as well as egg quality characteristics. Based on the results of this experiment, it can be concluded that WDW could be included to laying hens diets up to 10% with no deleterious effects on performance and egg quality characteristics.

  8. Artificial Enzymes, "Chemzymes"

    DEFF Research Database (Denmark)

    Bjerre, Jeannette; Rousseau, Cyril Andre Raphaël; Pedersen, Lavinia Georgeta M

    2008-01-01

    Enzymes have fascinated scientists since their discovery and, over some decades, one aim in organic chemistry has been the creation of molecules that mimic the active sites of enzymes and promote catalysis. Nevertheless, even today, there are relatively few examples of enzyme models that successf......Enzymes have fascinated scientists since their discovery and, over some decades, one aim in organic chemistry has been the creation of molecules that mimic the active sites of enzymes and promote catalysis. Nevertheless, even today, there are relatively few examples of enzyme models...... that successfully perform Michaelis-Menten catalysis under enzymatic conditions (i.e., aqueous medium, neutral pH, ambient temperature) and for those that do, very high rate accelerations are seldomly seen. This review will provide a brief summary of the recent developments in artificial enzymes, so called...... "Chemzymes", based on cyclodextrins and other molecules. Only the chemzymes that have shown enzyme-like activity that has been quantified by different methods will be mentioned. This review will summarize the work done in the field of artificial glycosidases, oxidases, epoxidases, and esterases, as well...

  9. Aspects of computation on asynchronous parallel processors

    International Nuclear Information System (INIS)

    Wright, M.

    1989-01-01

    The increasing availability of asynchronous parallel processors has provided opportunities for original and useful work in scientific computing. However, the field of parallel computing is still in a highly volatile state, and researchers display a wide range of opinion about many fundamental questions such as models of parallelism, approaches for detecting and analyzing parallelism of algorithms, and tools that allow software developers and users to make effective use of diverse forms of complex hardware. This volume collects the work of researchers specializing in different aspects of parallel computing, who met to discuss the framework and the mechanics of numerical computing. The far-reaching impact of high-performance asynchronous systems is reflected in the wide variety of topics, which include scientific applications (e.g. linear algebra, lattice gauge simulation, ordinary and partial differential equations), models of parallelism, parallel language features, task scheduling, automatic parallelization techniques, tools for algorithm development in parallel environments, and system design issues

  10. More insight into the interplay of response selection and visual attention in dual-tasks: masked visual search and response selection are performed in parallel.

    Science.gov (United States)

    Reimer, Christina B; Schubert, Torsten

    2017-09-15

    Both response selection and visual attention are limited in capacity. According to the central bottleneck model, the response selection processes of two tasks in a dual-task situation are performed sequentially. In conjunction search, visual attention is required to select the items and to bind their features (e.g., color and form), which results in a serial search process. Search time increases as items are added to the search display (i.e., set size effect). When the search display is masked, visual attention deployment is restricted to a brief period of time and target detection decreases as a function of set size. Here, we investigated whether response selection and visual attention (i.e., feature binding) rely on a common or on distinct capacity limitations. In four dual-task experiments, participants completed an auditory Task 1 and a conjunction search Task 2 that were presented with an experimentally modulated temporal interval between them (Stimulus Onset Asynchrony, SOA). In Experiment 1, Task 1 was a two-choice discrimination task and the conjunction search display was not masked. In Experiment 2, the response selection difficulty in Task 1 was increased to a four-choice discrimination and the search task was the same as in Experiment 1. We applied the locus-of-slack method in both experiments to analyze conjunction search time, that is, we compared the set size effects across SOAs. Similar set size effects across SOAs (i.e., additive effects of SOA and set size) would indicate sequential processing of response selection and visual attention. However, a significantly smaller set size effect at short SOA compared to long SOA (i.e., underadditive interaction of SOA and set size) would indicate parallel processing of response selection and visual attention. In both experiments, we found underadditive interactions of SOA and set size. In Experiments 3 and 4, the conjunction search display in Task 2 was masked. Task 1 was the same as in Experiments 1 and 2

  11. Effects of Immune Stress on Performance Parameters, Intestinal Enzyme Activity and mRNA Expression of Intestinal Transporters in Broiler Chickens

    Directory of Open Access Journals (Sweden)

    Y. Feng

    2012-05-01

    Full Text Available Immune stress is the loss of immune homeostasis caused by external forces. The purpose of this experiment was to investigate the effects of immune stress on the growth performance, small intestinal enzymes and peristalsis rate, and mRNA expression of nutrient transporters in broiler chickens. Four hundred and thirty-two 1-d-old broilers (Cobb500 were randomly assigned to four groups for treatment; each group included nine cages with 12 birds per cage. Group 1 = no vaccine (NV; Group 2 = conventional vaccine (CV; group 3 = lipopolysaccharide (LPS+conventional vaccine (LPS; group 4 = cyclophosphamide (CYP+conventional vaccine (CYP. The results demonstrated that immune stress by LPS and CYP reduced body weight gain (BWG, feed intake (FI, small intestine peristalsis rate and sIgA content in small intestinal digesta (p<0.05. However, the feed conversion ratio (FCR remained unchanged during the feeding period. LPS and CYP increased intestinal enzyme activity, relative expression of SGLT-1, CaBP-D28k and L-FABP mRNAs (p<0.05. LPS and CYP injection had a negative effect on the growth performance of healthy broiler chickens. The present study demonstrated that NV and CV could improve growth performance while enzyme activity in small intestine and relative expression of nutrient transporter mRNA of NV and CV were decreased in the conditions of a controlled rational feeding environment. It is generally recommended that broilers only need to be vaccinated for the diseases to which they might be exposed.

  12. Parallel Computing Using Web Servers and "Servlets".

    Science.gov (United States)

    Lo, Alfred; Bloor, Chris; Choi, Y. K.

    2000-01-01

    Describes parallel computing and presents inexpensive ways to implement a virtual parallel computer with multiple Web servers. Highlights include performance measurement of parallel systems; models for using Java and intranet technology including single server, multiple clients and multiple servers, single client; and a comparison of CGI (common…

  13. Automatic Loop Parallelization via Compiler Guided Refactoring

    DEFF Research Database (Denmark)

    Larsen, Per; Ladelsky, Razya; Lidman, Jacob

    For many parallel applications, performance relies not on instruction-level parallelism, but on loop-level parallelism. Unfortunately, many modern applications are written in ways that obstruct automatic loop parallelization. Since we cannot identify sufficient parallelization opportunities...... for these codes in a static, off-line compiler, we developed an interactive compilation feedback system that guides the programmer in iteratively modifying application source, thereby improving the compiler’s ability to generate loop-parallel code. We use this compilation system to modify two sequential...... benchmarks, finding that the code parallelized in this way runs up to 8.3 times faster on an octo-core Intel Xeon 5570 system and up to 12.5 times faster on a quad-core IBM POWER6 system. Benchmark performance varies significantly between the systems. This suggests that semi-automatic parallelization should...

  14. Parallel Programming with Intel Parallel Studio XE

    CERN Document Server

    Blair-Chappell , Stephen

    2012-01-01

    Optimize code for multi-core processors with Intel's Parallel Studio Parallel programming is rapidly becoming a "must-know" skill for developers. Yet, where to start? This teach-yourself tutorial is an ideal starting point for developers who already know Windows C and C++ and are eager to add parallelism to their code. With a focus on applying tools, techniques, and language extensions to implement parallelism, this essential resource teaches you how to write programs for multicore and leverage the power of multicore in your programs. Sharing hands-on case studies and real-world examples, the

  15. Effect of different levels dietary vitamin C on growth performance, muscle composition, antioxidant and enzyme activity of freshwater prawn, Macrobrachium malcolmsonii

    Directory of Open Access Journals (Sweden)

    Annamalai Asaikkutti

    2016-05-01

    Full Text Available In the present study was conducted to examine the effects and interactions of dietary vitamin C levels on the growth performance, antioxidant ability, muscle composition and enzyme activity in freshwater prawn Macrobrachium malcolmsonii (M. malcolmsonii. Additional, the vitamins C was dietary supplemented for freshwater prawn M. malcolmsonii. The experimental basal diets were supplemented with M. malcolmsonii at the rates of 0 (control, 25, 50, 100, 200 and 400 mg/kg dry feed weight. The as-supplemented vitamin C was fed in M. malcolmsonii for a period of 90 days. In the present investigation revealed that prawns fed with diet supplemented with 25–100 mg/kg of vitamins C shows enhanced (P  0.05 in feed conversion ratio (FCR were observed in prawn fed different diets. Addition, prawns fed with 25–100 mg/kg of vitamins C supplemented diets achieved significant (P  0.05 alterations in prawns fed with 25–100 mg/kg of vitamin C supplemented diets. Therefore, the present study proposed that 100 mg/kg of vitamin C could be supplemented for flexible enhanced survival; growth, antioxidant defense system and production of M. malcolmsonii. Keywords: Macrobrachium malcolmsonii, Vitamin, Growth performance, Biochemical compositions, Antioxidant enzyme

  16. Effect of potential probiotic Rhodotorula benthica D30 on the growth performance, digestive enzyme activity and immunity in juvenile sea cucumber Apostichopus japonicus.

    Science.gov (United States)

    Wang, Ji-hui; Zhao, Liu-qun; Liu, Jin-feng; Wang, Han; Xiao, Shan

    2015-04-01

    The effects of dietary addition of yeast Rhodotorula benthica (R. benthica) D30 which isolated from local sea mud at levels of 0 (control), 10(5), 10(6) and 10(7) CFU/g feed on the growth performance, digestive enzyme activity, immunity and disease resistance of juvenile sea cucumber Apostichopus japonicus were investigated. It was shown that dietary addition of R. benthica D30 significantly increased the growth rates of sea cucumbers (p  0.05). It was observed that adding R. benthica D30 could significantly decrease the cumulative mortality of sea cucumbers. The present study demonstrated that dietary addition of R. benthica D30 could increase growth performance and some digestive enzyme activities, improve immunity and disease resistance of A. japonicus. And the medium (10(6) CFU) and high (10(7) CFU) additional levels showed better effects. It suggests that yeast R. benthica D30 could be a good probiotic for aquaculture. Copyright © 2015 Elsevier Ltd. All rights reserved.

  17. Template based parallel checkpointing in a massively parallel computer system

    Science.gov (United States)

    Archer, Charles Jens [Rochester, MN; Inglett, Todd Alan [Rochester, MN

    2009-01-13

    A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.

  18. Effects of thermo-resistant non-starch polysaccharide degrading multi-enzyme on growth performance, meat quality, relative weights of body organs and blood profile in broiler chickens.

    Science.gov (United States)

    Mohammadi Gheisar, M; Hosseindoust, A; Kim, I H

    2016-06-01

    This research was conducted to study the performance and carcass parameters of broiler chickens fed diets supplemented with heat-treated non-starch polysaccharide degrading enzyme. A total of 432 one-day old Ross 308 broiler chickens were allocated to five treatments: (i) CON (basal diet), (ii) E1: CON + 0.05% multi-enzyme, (iii) E2: CON + 0.1% multi-enzyme, (iv) E3: CON + 0.05% thermo-resistant multi-enzyme and (v) E4: CON + 0.1% thermo-resistant multi-enzyme, each treatment consisted of six replications and 12 chickens in each replication. The chickens were housed in three floor battery cages during 28-day experimental period. On days 1-7, gain in body weight (BWG) improved by feeding the diets supplemented with thermo-resistant multi-enzyme. On days 7-21 and 1-28, chickens fed the diets containing thermo-resistant multi-enzyme showed improved (p thermo-resistant multi-enzyme affected the percentage of drip loss on d 1 (p thermo-resistant multi-enzyme did not affect the relative weights of organs but compared to CON group, relative weight of breast muscle increased and abdominal fat decreased (p thermo-resistant multi-enzyme showed higher (p thermo-resistant multi-enzyme improved performance of broiler chickens. Journal of Animal Physiology and Animal Nutrition © 2015 Blackwell Verlag GmbH.

  19. Parallelization of 2-D lattice Boltzmann codes

    International Nuclear Information System (INIS)

    Suzuki, Soichiro; Kaburaki, Hideo; Yokokawa, Mitsuo.

    1996-03-01

    Lattice Boltzmann (LB) codes to simulate two dimensional fluid flow are developed on vector parallel computer Fujitsu VPP500 and scalar parallel computer Intel Paragon XP/S. While a 2-D domain decomposition method is used for the scalar parallel LB code, a 1-D domain decomposition method is used for the vector parallel LB code to be vectorized along with the axis perpendicular to the direction of the decomposition. High parallel efficiency of 95.1% by the vector parallel calculation on 16 processors with 1152x1152 grid and 88.6% by the scalar parallel calculation on 100 processors with 800x800 grid are obtained. The performance models are developed to analyze the performance of the LB codes. It is shown by our performance models that the execution speed of the vector parallel code is about one hundred times faster than that of the scalar parallel code with the same number of processors up to 100 processors. We also analyze the scalability in keeping the available memory size of one processor element at maximum. Our performance model predicts that the execution time of the vector parallel code increases about 3% on 500 processors. Although the 1-D domain decomposition method has in general a drawback in the interprocessor communication, the vector parallel LB code is still suitable for the large scale and/or high resolution simulations. (author)

  20. Parallelization of 2-D lattice Boltzmann codes

    Energy Technology Data Exchange (ETDEWEB)

    Suzuki, Soichiro; Kaburaki, Hideo; Yokokawa, Mitsuo

    1996-03-01

    Lattice Boltzmann (LB) codes to simulate two dimensional fluid flow are developed on vector parallel computer Fujitsu VPP500 and scalar parallel computer Intel Paragon XP/S. While a 2-D domain decomposition method is used for the scalar parallel LB code, a 1-D domain decomposition method is used for the vector parallel LB code to be vectorized along with the axis perpendicular to the direction of the decomposition. High parallel efficiency of 95.1% by the vector parallel calculation on 16 processors with 1152x1152 grid and 88.6% by the scalar parallel calculation on 100 processors with 800x800 grid are obtained. The performance models are developed to analyze the performance of the LB codes. It is shown by our performance models that the execution speed of the vector parallel code is about one hundred times faster than that of the scalar parallel code with the same number of processors up to 100 processors. We also analyze the scalability in keeping the available memory size of one processor element at maximum. Our performance model predicts that the execution time of the vector parallel code increases about 3% on 500 processors. Although the 1-D domain decomposition method has in general a drawback in the interprocessor communication, the vector parallel LB code is still suitable for the large scale and/or high resolution simulations. (author).

  1. Compiling Scientific Programs for Scalable Parallel Systems

    National Research Council Canada - National Science Library

    Kennedy, Ken

    2001-01-01

    ...). The research performed in this project included new techniques for recognizing implicit parallelism in sequential programs, a powerful and precise set-based framework for analysis and transformation...

  2. Parallel kinematics type, kinematics, and optimal design

    CERN Document Server

    Liu, Xin-Jun

    2014-01-01

    Parallel Kinematics- Type, Kinematics, and Optimal Design presents the results of 15 year's research on parallel mechanisms and parallel kinematics machines. This book covers the systematic classification of parallel mechanisms (PMs) as well as providing a large number of mechanical architectures of PMs available for use in practical applications. It focuses on the kinematic design of parallel robots. One successful application of parallel mechanisms in the field of machine tools, which is also called parallel kinematics machines, has been the emerging trend in advanced machine tools. The book describes not only the main aspects and important topics in parallel kinematics, but also references novel concepts and approaches, i.e. type synthesis based on evolution, performance evaluation and optimization based on screw theory, singularity model taking into account motion and force transmissibility, and others.   This book is intended for researchers, scientists, engineers and postgraduates or above with interes...

  3. DNA polymerase hybrids derived from the family-B enzymes of Pyrococcus furiosus and Thermococcus kodakarensis: improving performance in the polymerase chain reaction.

    Science.gov (United States)

    Elshawadfy, Ashraf M; Keith, Brian J; Ee Ooi, H'Ng; Kinsman, Thomas; Heslop, Pauline; Connolly, Bernard A

    2014-01-01

    The polymerase chain reaction (PCR) is widely applied across the biosciences, with archaeal Family-B DNA polymerases being preferred, due to their high thermostability and fidelity. The enzyme from Pyrococcus furiosus (Pfu-Pol) is more frequently used than the similar protein from Thermococcus kodakarensis (Tkod-Pol), despite the latter having better PCR performance. Here the two polymerases have been comprehensively compared, confirming that Tkod-Pol: (1) extends primer-templates more rapidly; (2) has higher processivity; (3) demonstrates superior performance in normal and real time PCR. However, Tkod-Pol is less thermostable than Pfu-Pol and both enzymes have equal fidelities. To understand the favorable properties of Tkod-Pol, hybrid proteins have been prepared. Single, double and triple mutations were used to site arginines, present at the "forked-point" (the junction of the exonuclease and polymerase channels) of Tkod-Pol, at the corresponding locations in Pfu-Pol, slightly improving PCR performance. The Pfu-Pol thumb domain, responsible for double-stranded DNA binding, has been entirely replaced with that from Tkod-Pol, again giving better PCR properties. Combining the "forked-point" and thumb swap mutations resulted in a marked increase in PCR capability, maintenance of high fidelity and retention of the superior thermostability associated with Pfu-Pol. However, even the arginine/thumb swap mutant falls short of Tkod-Pol in PCR, suggesting further improvement within the Pfu-Pol framework is attainable. The significance of this work is the observation that improvements in PCR performance are easily attainable by blending elements from closely related archaeal polymerases, an approach that may, in future, be extended by using more polymerases from these organisms.

  4. Effects of small peptides, probiotics, prebiotics, and synbiotics on growth performance, digestive enzymes, and oxidative stress in orange-spotted grouper, Epinephelus coioides, juveniles reared in artificial seawater

    Science.gov (United States)

    Wang, Tao; Cheng, Yongzhou; Chen, Xiaoyan; Liu, Zhaopu; Long, Xiaohua

    2017-01-01

    Aquaculture production efficiency may increase by using feed additives. This study investigated the effects of different dietary additives [w/w: 2% small peptides, 0.01% probiotics ( Bacillus licheniformis) and 0.2% prebiotics (inulin)] on growth performance, digestive enzyme activities, and oxidative stress in juvenile Epinephelus coioides reared in artificial seawater of two salt concentrations (13.5 vs. 28.5). Weight gain rate was significantly higher in fish fed the diet supplemented with small peptides, B. licheniformis, inulin, or synbiotics than that in fish fed the basal diet; the greatest weight gain rate was found in fish fed the small peptide treatment [56.0% higher than basal diet]. Higher feed efficiency was detected in fish fed the diet supplemented with small peptides than that of fish in the other dietary treatments. Total protease activity in the stomach and intestines was highest in fish fed the small peptide-treated diet, whereas lipase activity was highest in those fed synbiotics (combination of Bacillus licheniformis and inulin) than that in fish fed the other treatments. Antioxidant enzyme (total superoxide dismutase and catalase) activities and hepatic malondialdehyde content were higher in fish receiving the dietary supplements and maintained in artificial seawater containing 13.5 salinity compared with those in the control (28.5). Hepatic catalase activity in grouper fed the diets with small peptides or synbiotics decreased significantly compared with that in control fish. Overall, the three types of additives improved growth rate of juvenile grouper and digestive enzymes activities to varying degrees but did not effectively improve antioxidant capacity under low-salinity stress conditions.

  5. Effect of Potential Probiotic Lactococcus lactis Subsp. lactis on Growth Performance, Intestinal Microbiota, Digestive Enzyme Activities, and Disease Resistance of Litopenaeus vannamei.

    Science.gov (United States)

    Adel, Milad; El-Sayed, Abdel-Fattah M; Yeganeh, Sakineh; Dadar, Maryam; Giri, Sib Sankar

    2017-06-01

    The aims of this study were to evaluate the effects of Lactococcus lactis subsp. lactis on the growth, intestinal microbiota, digestive enzyme activity, and disease resistance of Litopenaeus vannamei. Diets containing four different concentrations of L. lactis (0 [basal diet], 10 6 , 10 7 , and 10 8  CFU g -1 ) were fed to white shrimps L. vannamei (average weight 5.89 ± 0.36 g) for 8 weeks. At the end of the feeding trial, shrimps were immersed in Caspian Seawater (10.8 ppt) contaminated with 10 6  CFU ml -1 pathogenic V. anguillarum for 2 h. Results revealed that growth rate, survival, and body protein level were increased with dietary supplementation of L. lactis. The activities of digestive enzymes (cellulose, lipase, amylase, and protease) were significantly higher in the groups fed with diets containing 10 7 or 10 8  CFU g -1 L. lactis than those in the control. The Lactobacillus and Bacillus counts were higher (P lactis-supplemented diets. In addition, higher level of L. lactis supplementation decreased the Vibrio counts. Moreover, L. vannamei fed diet supplemented with 10 8  CFU g -1 of L. lactis exhibited significantly the highest hematocyte count and post-challenge survival rate (79.2 %). Collectively, these results suggest that dietary supplementation of L. lactis subsp. lactis at 10 8  CFU g -1 can promote growth performance, digestive enzyme activity, and disease resistance of L. vannamei.

  6. Massively Parallel Finite Element Programming

    KAUST Repository

    Heister, Timo

    2010-01-01

    Today\\'s large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  7. Massively Parallel Finite Element Programming

    KAUST Repository

    Heister, Timo; Kronbichler, Martin; Bangerth, Wolfgang

    2010-01-01

    Today's large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  8. The STAPL Parallel Graph Library

    KAUST Repository

    Harshvardhan,

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable distributed graph container and a collection of commonly used parallel graph algorithms. The library introduces pGraph pViews that separate algorithm design from the container implementation. It supports three graph processing algorithmic paradigms, level-synchronous, asynchronous and coarse-grained, and provides common graph algorithms based on them. Experimental results demonstrate improved scalability in performance and data size over existing graph libraries on more than 16,000 cores and on internet-scale graphs containing over 16 billion vertices and 250 billion edges. © Springer-Verlag Berlin Heidelberg 2013.

  9. Parallel sorting algorithms

    CERN Document Server

    Akl, Selim G

    1985-01-01

    Parallel Sorting Algorithms explains how to use parallel algorithms to sort a sequence of items on a variety of parallel computers. The book reviews the sorting problem, the parallel models of computation, parallel algorithms, and the lower bounds on the parallel sorting problems. The text also presents twenty different algorithms, such as linear arrays, mesh-connected computers, cube-connected computers. Another example where algorithm can be applied is on the shared-memory SIMD (single instruction stream multiple data stream) computers in which the whole sequence to be sorted can fit in the

  10. Comparison among performances of a ligase chain reaction-based assay and two enzyme immunoassays in detecting Chlamydia trachomatis in urine specimens from men with nongonococcal urethritis.

    Science.gov (United States)

    Deguchi, T; Yasuda, M; Uno, M; Tada, K; Iwata, H; Komeda, H; Maeda, S; Latila, V; Saito, I; Kawada, Y

    1996-01-01

    We evaluated the performances of a ligase chain reaction (LCR)-based assay and two enzyme immunoassays (Chlamydiazyme and IDEIA) in the detection of Chlamydia trachomatis in urine specimens. We compared the results of testing urine specimens by these assays with those of urethral swab culture by examining samples from 131 men with nongonococcal urethritis. Discrepant results were analyzed by testing urethral swab specimens for C. trachomatis by a PCR-based assay. After the resolution of discrepant results, the sensitivity of urethral swab culture was 85.3%, whereas those of the LCR assay, Chlamydiazyme, and IDEIA with urine specimens were 94.1, 82.4, and 94.1%, respectively. The LCR assay and IDEIA were more sensitive than was urethral swab culture. In addition, the LCR assay, with a sensitivity equal to that of IDEIA, was more specific. Overall, the LCR assay proved to be superior to the enzyme immunoassays in detecting C. trachomatis in urine specimens. Testing urine specimens by LCR assay should be a helpful alternative method for diagnosing C. trachomatis urethral infection in men with nongonococcal urethritis. PMID:8784574

  11. Persona and the Performance of Identity : Parallel Developments in the Biographical Historiography of Science and Gender, and the Related Uses of Self Narrative

    NARCIS (Netherlands)

    Bosch, Mineke

    2013-01-01

    In this article Bosch explores the parallel development in scientific and gender biography to shed light on the relation between the individual and the collective, the self and society. In the history of science the relational/collective scientific self and the concept of the scientific persona (or

  12. A hybrid, massively parallel implementation of a genetic algorithm for optimization of the impact performance of a metal/polymer composite plate

    KAUST Repository

    Narayanan, Kiran; Mora Cordova, Angel; Allsopp, Nicholas; El Sayed, Tamer S.

    2012-01-01

    A hybrid parallelization method composed of a coarse-grained genetic algorithm (GA) and fine-grained objective function evaluations is implemented on a heterogeneous computational resource consisting of 16 IBM Blue Gene/P racks, a single x86 cluster

  13. Performance of hemicellulolytic enzymes in culture supernatants from a wide range of fungi on insoluble wheat straw and corn fiber fractions

    NARCIS (Netherlands)

    Gool, van M.P.; Toth, K.; Schols, H.A.; Szakacs, G.; Gruppen, H.

    2012-01-01

    Filamentous fungi are a good source of hemicellulolytic enzymes for biomass degradation. Enzyme preparations were obtained as culture supernatants from 78 fungal isolates grown on wheat straw as carbon source. These enzyme preparations were utilized in the hydrolysis of insoluble wheat straw and

  14. PDDP, A Data Parallel Programming Model

    Directory of Open Access Journals (Sweden)

    Karen H. Warren

    1996-01-01

    Full Text Available PDDP, the parallel data distribution preprocessor, is a data parallel programming model for distributed memory parallel computers. PDDP implements high-performance Fortran-compatible data distribution directives and parallelism expressed by the use of Fortran 90 array syntax, the FORALL statement, and the WHERE construct. Distributed data objects belong to a global name space; other data objects are treated as local and replicated on each processor. PDDP allows the user to program in a shared memory style and generates codes that are portable to a variety of parallel machines. For interprocessor communication, PDDP uses the fastest communication primitives on each platform.

  15. Parallelization of quantum molecular dynamics simulation code

    International Nuclear Information System (INIS)

    Kato, Kaori; Kunugi, Tomoaki; Shibahara, Masahiko; Kotake, Susumu

    1998-02-01

    A quantum molecular dynamics simulation code has been developed for the analysis of the thermalization of photon energies in the molecule or materials in Kansai Research Establishment. The simulation code is parallelized for both Scalar massively parallel computer (Intel Paragon XP/S75) and Vector parallel computer (Fujitsu VPP300/12). Scalable speed-up has been obtained with a distribution to processor units by division of particle group in both parallel computers. As a result of distribution to processor units not only by particle group but also by the particles calculation that is constructed with fine calculations, highly parallelization performance is achieved in Intel Paragon XP/S75. (author)

  16. Parallel Polarization State Generation.

    Science.gov (United States)

    She, Alan; Capasso, Federico

    2016-05-17

    The control of polarization, an essential property of light, is of wide scientific and technological interest. The general problem of generating arbitrary time-varying states of polarization (SOP) has always been mathematically formulated by a series of linear transformations, i.e. a product of matrices, imposing a serial architecture. Here we show a parallel architecture described by a sum of matrices. The theory is experimentally demonstrated by modulating spatially-separated polarization components of a laser using a digital micromirror device that are subsequently beam combined. This method greatly expands the parameter space for engineering devices that control polarization. Consequently, performance characteristics, such as speed, stability, and spectral range, are entirely dictated by the technologies of optical intensity modulation, including absorption, reflection, emission, and scattering. This opens up important prospects for polarization state generation (PSG) with unique performance characteristics with applications in spectroscopic ellipsometry, spectropolarimetry, communications, imaging, and security.

  17. Heat pump using dual heat sources of air and water. Performance with heat sources arranged in parallel; Mizu kuki ryonetsugen heat pump no kenkyu. Netsugen heiretsu unten ni okeru seino

    Energy Technology Data Exchange (ETDEWEB)

    Ito, S; Miura, N; Sato, S [Kanagawa Institute of Technology, Kanagawa (Japan); Uchikawa, Y; Hamada, K [Kubota Corp., Osaka (Japan)

    1996-10-27

    A heat pump system using water and air as heat sources was built and evaluated for its performance. In this system, evaporators may be operated singly or as connected in parallel or series, and, for each case, the quantity of heat acquired may be measured and system performance may be quantitatively evaluated. The findings follow. When the two heat sources are equal in temperature in the single-evaporator operation, the evaporation temperature is about 7{degree}C higher on the water side than on the air side, and the performance coefficient is about 0.7 higher. When the air heat source temperature is 25{degree}C in the parallel operation, like quantities of heat are obtained from both heat sources, and collection of heat from the water increases with a decrease in the air heat source temperature but, with an increase, collection from the air increases. When the air heat source temperature decreases, the evaporation temperature decreases in the single-evaporator working on the air and in the parallel operation but it levels off in the single-evaporator working on the water alone. When the water heat source temperature decreases, evaporation temperature drop is sharper in the single-evaporator working on the water than in the parallel operation, which suggests the transfer from the parallel operation to the single-evaporator working on the air. In the single-evaporator operation on the water heat source, the evaporation temperature linearly decreases with an increase in superheating. 1 ref., 10 figs.

  18. Introduction to parallel programming

    CERN Document Server

    Brawer, Steven

    1989-01-01

    Introduction to Parallel Programming focuses on the techniques, processes, methodologies, and approaches involved in parallel programming. The book first offers information on Fortran, hardware and operating system models, and processes, shared memory, and simple parallel programs. Discussions focus on processes and processors, joining processes, shared memory, time-sharing with multiple processors, hardware, loops, passing arguments in function/subroutine calls, program structure, and arithmetic expressions. The text then elaborates on basic parallel programming techniques, barriers and race

  19. Parallel computing works!

    CERN Document Server

    Fox, Geoffrey C; Messina, Guiseppe C

    2014-01-01

    A clear illustration of how parallel computers can be successfully appliedto large-scale scientific computations. This book demonstrates how avariety of applications in physics, biology, mathematics and other scienceswere implemented on real parallel computers to produce new scientificresults. It investigates issues of fine-grained parallelism relevant forfuture supercomputers with particular emphasis on hypercube architecture. The authors describe how they used an experimental approach to configuredifferent massively parallel machines, design and implement basic systemsoftware, and develop

  20. Growth performance, digestive enzyme activity and immune response of Macrobrachium rosenbergii fed with probiotic Clostridium butyricum incorporated diets

    Directory of Open Access Journals (Sweden)

    Mohammad Saifuddin Sumon

    2018-01-01

    Full Text Available To determine antagonistic effect of Clostridium butyricum against Vibrio harveyi and its probiotic effect on growth performance, digestibility and immune response of fresh water prawn, Macrobrachium rosenbergii juveniles were examined following feeding with C. butyricum incorporated feed for 60 days. Significant reduction of V. harveyi growth was found at 8 hr and onward in in-vitro and at 10 days and onward in in-vivo challenge test. After rearing prawn with the bacteria in feed treatment for 60 days, body weight and growth rate of prawns was significantly higher (p  0.05 compared to control group. This study revealed that probiotic, C. butyricum incorporated diets were found to be beneficial for M. rosenbergii culture in terms of hindering the growth of pathogenic bacteria and increasing the growth, protease and amylase activities of prawn. Results from this study will be helpful to improve fresh water prawn farming.

  1. A high-performance liquid chromatography-based radiometric assay for sucrose-phosphate synthase and other UDP-glucose requiring enzymes

    International Nuclear Information System (INIS)

    Salvucci, M.E.; Crafts-Brandner, S.J.

    1991-01-01

    A method for product analysis that eliminates a problematic step in the radiometric sucrose-phosphate synthase assay is described. The method uses chromatography on a boronate-derivatized high-performance liquid chromatography column to separate the labeled product, [14C]sucrose phosphate, from unreacted uridine 5'-diphosphate-[14C]glucose (UDP-Glc). Direct separation of these compounds eliminates the need for treatment of the reaction mixtures with alkaline phosphatase, thereby avoiding the problem of high background caused by contaminating phosphodiesterase activity in alkaline phosphatase preparations. The method presented in this paper can be applied to many UDP-Glc requiring enzymes; here the authors show its use for determining the activities of sucrose-phosphate synthase, sucrose synthase, and uridine diphosphate-glucose pyrophosphorylase in plant extracts

  2. Effect of an Enzyme Blend on the Performance, Diet Metabolizability, Phosphorous Retention, and Bone Mineralization of Broilers Fed Diets Containing Defatted Rice Bran

    Directory of Open Access Journals (Sweden)

    ML Moraes

    2015-06-01

    Full Text Available An experiment was conducted to evaluate the effect of an enzyme blend (EB on the performance, diet metabolizability, phosphorus (P retention, and bone mineralization of broilers fed diets containing 10% defatted rice bran (DRB. In total, 432 one- to 38-d-old male Cobb broilers were evaluated according to a completely randomized experimental design in 3 x 2 factorial arrangement. Three diets were tested with two nutrient reductions (NR in the matrix (standard diet; NR I of 75 kcal/kg ME, 0.1% Ca and 0.1% available P; and NR II of 100 kcal/kg ME, 0.1% Ca and 0.1% available P with or without the addition of an EB (200 g/t. The coefficients of total tract apparent retention (CTTAR of the diets and P retention were determined by collecting excreta during two periods (14 to 17 and 28 to 31 d. As expected, birds fed the standard diet had higher BW, BW gain, and G:F compared to birds on the NR diets. The EB did not show any positive effects on CTTAR or on performance; however, birds fed the EB retained 6.58% more P from d 14 to 17 (p ≤ 0.07 and 8.55% from d 28 to 31 (p < 0.05. Tibiotarsus ash percentage also increased by 2.45% (p ≤ 0.06 on d 38. In diets containing 10% DRB, the enzyme blend showed biological activity improving P retention and tibiotarsus mineralization.

  3. Effects of Bacillus subtilis on the growth performance, digestive enzymes, immune gene expression and disease resistance of white shrimp, Litopenaeus vannamei.

    Science.gov (United States)

    Zokaeifar, Hadi; Balcázar, José Luis; Saad, Che Roos; Kamarudin, Mohd Salleh; Sijam, Kamaruzaman; Arshad, Aziz; Nejat, Naghmeh

    2012-10-01

    We studied the effect of two probiotic Bacillus subtilis strains on the growth performance, digestive enzyme activity, immune gene expression and disease resistance of juvenile white shrimp (Litopenaeus vannamei). A mixture of two probiotic strains, L10 and G1 in equal proportions, was administered at two different doses 10(5) (BM5) and 10(8) (BM8) CFU g(-1) feed to shrimp for eight weeks. In comparison to untreated control group, final weight, weight gain and digestive enzyme activity were significantly greater in shrimp fed BM5 and BM8 diets. Significant differences for specific growth rate (SGR) and survival were recorded in shrimp fed BM8 diet as compared with the control; however, no significant differences were recorded for food conversion ratio (FCR) among all the experimental groups. Eight weeks after the start of the feeding period, shrimp were challenged with Vibrio harveyi. Statistical analysis revealed significant differences in shrimp survival between probiotic and control groups. Cumulative mortality of the control group was 63.3%, whereas cumulative mortality of the shrimp that had been given probiotics was 20.0% with BM8 and 33.3% with BM5. Subsequently, real-time PCR was employed to determine the mRNA levels of prophenoloxidase (proPO), peroxinectin (PE), lipopolysaccharide- and β-1,3-glucan-binding protein (LGBP) and serine protein (SP). The expression of all immune-related genes studied was significantly up-regulated (P growth performance and disease resistance through an enhanced immune response in shrimp. Copyright © 2012 Elsevier Ltd. All rights reserved.

  4. Massively Parallel QCD

    International Nuclear Information System (INIS)

    Soltz, R; Vranas, P; Blumrich, M; Chen, D; Gara, A; Giampap, M; Heidelberger, P; Salapura, V; Sexton, J; Bhanot, G

    2007-01-01

    The theory of the strong nuclear force, Quantum Chromodynamics (QCD), can be numerically simulated from first principles on massively-parallel supercomputers using the method of Lattice Gauge Theory. We describe the special programming requirements of lattice QCD (LQCD) as well as the optimal supercomputer hardware architectures that it suggests. We demonstrate these methods on the BlueGene massively-parallel supercomputer and argue that LQCD and the BlueGene architecture are a natural match. This can be traced to the simple fact that LQCD is a regular lattice discretization of space into lattice sites while the BlueGene supercomputer is a discretization of space into compute nodes, and that both are constrained by requirements of locality. This simple relation is both technologically important and theoretically intriguing. The main result of this paper is the speedup of LQCD using up to 131,072 CPUs on the largest BlueGene/L supercomputer. The speedup is perfect with sustained performance of about 20% of peak. This corresponds to a maximum of 70.5 sustained TFlop/s. At these speeds LQCD and BlueGene are poised to produce the next generation of strong interaction physics theoretical results

  5. Experiences in Data-Parallel Programming

    Directory of Open Access Journals (Sweden)

    Terry W. Clark

    1997-01-01

    Full Text Available To efficiently parallelize a scientific application with a data-parallel compiler requires certain structural properties in the source program, and conversely, the absence of others. A recent parallelization effort of ours reinforced this observation and motivated this correspondence. Specifically, we have transformed a Fortran 77 version of GROMOS, a popular dusty-deck program for molecular dynamics, into Fortran D, a data-parallel dialect of Fortran. During this transformation we have encountered a number of difficulties that probably are neither limited to this particular application nor do they seem likely to be addressed by improved compiler technology in the near future. Our experience with GROMOS suggests a number of points to keep in mind when developing software that may at some time in its life cycle be parallelized with a data-parallel compiler. This note presents some guidelines for engineering data-parallel applications that are compatible with Fortran D or High Performance Fortran compilers.

  6. The Use of Angiotensin-I Converting Enzyme I/D Genetic Polymorphism as a Biomarker of Athletic Performance in Humans

    Directory of Open Access Journals (Sweden)

    Maria Fernanda De Mello Costa

    2012-10-01

    Full Text Available Angiotensin II is a key regulator of blood pressure and cardiovascular function in mammals. The conversion of angiotensin into its active form is carried out by Angiotensin I-Converting Enzyme (ACE. The measurement of ACE concentration in plasma or serum, its enzymatic activity, and the correlation between an insertion/deletion (I/D genetic polymorphism of the ACE gene have been investigated as possible indicators of superior athletic performance in humans. In this context, other indicators of superior adaptation to exercise resulting in better athletic performance (such as ventricular hypertrophy, VO2 max, and competition results were mostly used to study the association between ACE I/D polymorphism and improved performance. Despite the fact that the existing literature presents little consensus, there is sufficient scientific evidence to warrant further investigation on the usage of ACE activity and the I/D ACE gene polymorphism as biomarkers of superior athletic performance in humans of specific ethnicities or in athletes involved in certain sports. In this sense, a biomarker would be a substance or genetic component that could be measured to provide a degree of certainty, or an indication, of the presence of a certain trait or characteristic that would be beneficial to the athlete’s performance. Difficulties in interpreting and comparing the results of scientific research on the topic arise from dissimilar protocols and variation in study design. This review aims to investigate the current literature on the use of ACE I/D polymorphism as a biomarker of performance in humans through the comparison of scientific publications.

  7. Evaluation of Imbrasia belina meal as a fishmeal substitute in Oreochromis mossambicus diets: Growth performance, histological analysis and enzyme activity

    Directory of Open Access Journals (Sweden)

    Mmaditshaba M. Rapatsa

    2017-02-01

    Full Text Available The main objective of this study was to investigate mopane worm (Imbrasia belina as a protein source in the diet of Oreochromis mossambicus. One thousand five hundred O. mossambicus fingerlings (40 ± 2.5 g were fed five isonitrogenous, isolipidic and isoenergetic diets formulated to contain 30% crude protein and 20 MJ/kggross energy (dry matter basis for 51 days. Fifteen indoor rectangular concrete tanks (1.5 m3 connected to a recirculating system were used. Water temperature ranged between 27 and 29 °C. The diets were prepared by replacing fishmeal with mopane worm meal at 10%, 20%, 40% and 60%. The diets were coded D2, D3, D4 and D5 respectively. A control diet with no mopane worm meal was coded D1. The diets were fed to triplicate groups of O. mossambicus twice a day. Specific growth rate (SGR, Thermal-unit growth coefficient (TGC, protein efficiency ratio (PER and apparent digestibility coefficient (ADC increased with higher inclusion levels of mopane worm meal. Feed conversion ratio (FCR also improved with higher inclusion levels. However, the highest growth performance (SGR: 3.49%; FCR: 1.29 was recorded in fish fed the fishmeal based control diet. Protease, amylase and lipase activities were determined in the intestines. Both protease and amylase activity were significantly higher (P < 0.05 at high mopane worm inclusion levels. It is suggested that the high protein levels of the mopane worm diet elicited high protease activity. The health status of the fish was evaluated by examining the liver and intestine histology. There were no evident histological alterations of either liver or intestine as mopane worm meal inclusion levels increased. This showed that mopane worm meal may be a good candidate for the replacement of fishmeal in O. mossambicus diets. The highest profit index (1.67 was recorded in the 60% mopane worm inclusion level. The lowest profit index was in the control. More studies on mopane worm meal as a substitute

  8. Programming massively parallel processors a hands-on approach

    CERN Document Server

    Kirk, David B

    2010-01-01

    Programming Massively Parallel Processors discusses basic concepts about parallel programming and GPU architecture. ""Massively parallel"" refers to the use of a large number of processors to perform a set of computations in a coordinated parallel way. The book details various techniques for constructing parallel programs. It also discusses the development process, performance level, floating-point format, parallel patterns, and dynamic parallelism. The book serves as a teaching guide where parallel programming is the main topic of the course. It builds on the basics of C programming for CUDA, a parallel programming environment that is supported on NVI- DIA GPUs. Composed of 12 chapters, the book begins with basic information about the GPU as a parallel computer source. It also explains the main concepts of CUDA, data parallelism, and the importance of memory access efficiency using CUDA. The target audience of the book is graduate and undergraduate students from all science and engineering disciplines who ...

  9. Parallel plasma fluid turbulence calculations

    International Nuclear Information System (INIS)

    Leboeuf, J.N.; Carreras, B.A.; Charlton, L.A.; Drake, J.B.; Lynch, V.E.; Newman, D.E.; Sidikman, K.L.; Spong, D.A.

    1994-01-01

    The study of plasma turbulence and transport is a complex problem of critical importance for fusion-relevant plasmas. To this day, the fluid treatment of plasma dynamics is the best approach to realistic physics at the high resolution required for certain experimentally relevant calculations. Core and edge turbulence in a magnetic fusion device have been modeled using state-of-the-art, nonlinear, three-dimensional, initial-value fluid and gyrofluid codes. Parallel implementation of these models on diverse platforms--vector parallel (National Energy Research Supercomputer Center's CRAY Y-MP C90), massively parallel (Intel Paragon XP/S 35), and serial parallel (clusters of high-performance workstations using the Parallel Virtual Machine protocol)--offers a variety of paths to high resolution and significant improvements in real-time efficiency, each with its own advantages. The largest and most efficient calculations have been performed at the 200 Mword memory limit on the C90 in dedicated mode, where an overlap of 12 to 13 out of a maximum of 16 processors has been achieved with a gyrofluid model of core fluctuations. The richness of the physics captured by these calculations is commensurate with the increased resolution and efficiency and is limited only by the ingenuity brought to the analysis of the massive amounts of data generated

  10. Enzyme-linked immunosorbent assay and polymerase chain reaction performance using Mexican and Guatemalan discrete typing unit I strains of Trypanosoma cruzi.

    Science.gov (United States)

    Ballinas-Verdugo, Martha; Reyes, Pedro Antonio; Mejia-Dominguez, Ana; López, Ruth; Matta, Vivian; Monteón, Victor M

    2011-12-01

    Thirteen Trypanosoma cruzi isolates from different geographic regions of Mexico and Guatemala belonging to discrete typing unit (DTU) I and a reference CL-Brener (DTU VI) strain were used to perform enzyme-linked immunosorbent assay (ELISA) and polymerase chain reaction (PCR). A panel of 57 Mexican serum samples of patients with chronic chagasic cardiopathy and asymptomatic infected subjects (blood bank donors) were used in this study. DNA from the above 14 T. cruzi strains were extracted and analyzed by PCR using different sets of primers designed from minicircle and satellite T. cruzi DNA. The chronic chagasic cardiopathy serum samples were easily recognized with ELISA regardless of the source of antigenic extract used, even with the CL-Brener TcVI, but positive serum samples from blood bank donors in some cases were not recognized by some Mexican antigenic extracts. On the other hand, PCR showed an excellent performance despite the set of primers used, since all Mexican and Guatemalan T. cruzi strains were correctly amplified. In general terms, Mexican, Guatemalan, and CL-Brener T. cruzi strains are equally good sources of antigen when using the ELISA test to detect Mexican serum samples. However, there are some strains with poor performance. The DTU I strains are easily detected using either kinetoplast or satellite DNA target designed from DTU VI strains.

  11. Parallel Atomistic Simulations

    Energy Technology Data Exchange (ETDEWEB)

    HEFFELFINGER,GRANT S.

    2000-01-18

    Algorithms developed to enable the use of atomistic molecular simulation methods with parallel computers are reviewed. Methods appropriate for bonded as well as non-bonded (and charged) interactions are included. While strategies for obtaining parallel molecular simulations have been developed for the full variety of atomistic simulation methods, molecular dynamics and Monte Carlo have received the most attention. Three main types of parallel molecular dynamics simulations have been developed, the replicated data decomposition, the spatial decomposition, and the force decomposition. For Monte Carlo simulations, parallel algorithms have been developed which can be divided into two categories, those which require a modified Markov chain and those which do not. Parallel algorithms developed for other simulation methods such as Gibbs ensemble Monte Carlo, grand canonical molecular dynamics, and Monte Carlo methods for protein structure determination are also reviewed and issues such as how to measure parallel efficiency, especially in the case of parallel Monte Carlo algorithms with modified Markov chains are discussed.

  12. Pasture, multi-enzymes, benzoic acid and essential oils positively influence performance, intestinal organ weight and egg quality in free-range laying hens.

    Science.gov (United States)

    Iqbal, Z; Roberts, J; Perez-Maldonado, R A; Goodarzi Boroojeni, F; Swick, R A; Ruhnke, I

    2018-04-01

    1. The objective of this study was to investigate the effect of range type, multi-enzyme applications, and a combination of benzoic acid (BA) and essential oils (EO) on the productive performance, organ weight and egg quality of free-range laying hens. 2. Three hundred laying hens were evaluated for the short-term (6 weeks) and long-term (12 weeks) effects of range type (G = no pasture, P = pasture) and feed additives (T1 = control; T2 = betaglucanase/pectinase/protease; T3 = BA/EO). Body weight, feed intake (FI), feed conversion ratio (FCR), egg production (EP), digestive organ weight, and egg quality (EQ) were evaluated. Data were analysed using SPSS 2.2 in a 2×2×3 factorial arrangement. 3. Hens that ranged on pasture were significantly heavier (2043 g vs. 1996 g; p ranged on gravel. Hens fed T2 were significantly heavier (2050 g) compared to hens fed T1 (2005 g) or T3 (2008 g). Organ weights (gizzard, liver and pancreas) were significantly heavier in hens ranged on pasture (16.8 g/kg BW, 22.3 g/kg BW and 1.89 g/kg BW, respectively) compared to hens ranged on gravel (14.2 g/kg BW, 21.7 g/kg BW and 1.83 g/kg BW, respectively). Over time, body weight (1970-2070 g; p < 0.001) and egg weight (59.5-62.8 g; p < 0.001) increased, FI (123-120 g; p = 0.024) was reduced and FCR (2.36-2.10; p = 0.002) improved 4. In conclusion, hens housed on pasture and fed multi-enzyme supplemented diets had significantly heavier body weight and produced heavier eggs with darker yolk colour. Pasture intake and enzyme supplementation increased digestive organ weight significantly.

  13. Effects of dietary amylose/amylopectin ratio on growth performance, feed utilization, digestive enzymes, and postprandial metabolic responses in juvenile obscure puffer Takifugu obscurus.

    Science.gov (United States)

    Liu, Xiang-he; Ye, Chao-xia; Ye, Ji-dan; Shen, Bi-duan; Wang, Chun-yan; Wang, An-li

    2014-10-01

    other diets. Plasma glucose and triglyceride concentrations showed a significant difference at 2 and 4 h after a meal and varied between dietary treatments. According to regression analysis of weight gain against dietary AM/AP ratio, the optimum dietary AM/AP ratio for maximum growth of obscure puffer was 0.25. The present result indicates that dietary AM/AP ratio could affect growth performance and feed utilization, some plasma parameters, digestive enzyme as well as hepatic glucose metabolic enzyme activities in juvenile obscure puffer.

  14. Growth performance and antioxidant enzyme activities in rainbow trout (Oncorhynchus mykiss) juveniles fed diets supplemented with sage, mint and thyme oils.

    Science.gov (United States)

    Sönmez, Adem Yavuz; Bilen, Soner; Alak, Gonca; Hisar, Olcay; Yanık, Talat; Biswas, Gouranga

    2015-02-01

    This study evaluated effects of dietary supplementation of sage (Salvia officinalis), mint (Mentha spicata) and thyme (Thymus vulgaris) oils on growth performance, lipid peroxidation level (melondialdehyde, MDA) and liver antioxidant enzyme activities (superoxide dismutase, SOD; catalase, CAT; glucose-6-phosphate dehydrogenase, G6PD; glutathione reductase, GR; glutathione-S-transferase, GST and glutathione peroxidase, GPx) in rainbow trout (Oncorhynchus mykiss) juveniles. For this purpose, triplicate groups of rainbow trout were fed daily ad libitum with diets containing sage, mint and thyme oils at 500, 1,000 and 1,500 mg kg(-1) for 60 days. While weight gain percentage of fish fed the diets containing sage and thyme oils was significantly higher than the control group, that of fish fed mint oil was the lowest. Similarly, specific growth rate was found to be the highest in all groups of the sage and thyme oil feeding and the lowest in the mint groups. Moreover, feed conversion ratio was significantly higher in the mint oil administered groups. Survival rate was also significantly reduced in the fish fed the diet containing mint oil. It was observed that SOD, G6PD and GPx activities were significantly increased in liver tissues of all the treated fish groups compared to that of control diet-fed group. However, CAT, GST and GR activities were significantly decreased in experimental diet-fed fish groups at the end of the experiment. On the other hand, a significant reduction was found in MDA levels in the fish fed the diets with sage and thyme oils compared to control and mint diets on the 30th and 60th days of experiment. Overall, dietary inclusion of sage and thyme oils is effective in enhancing rainbow trout growth, reduction in MDA and least changing antioxidant enzyme activities at a low level of 500 mg kg(-1) diet, and they can be used as important feed supplements for rainbow trout production.

  15. Parallel processing for fluid dynamics applications

    International Nuclear Information System (INIS)

    Johnson, G.M.

    1989-01-01

    The impact of parallel processing on computational science and, in particular, on computational fluid dynamics is growing rapidly. In this paper, particular emphasis is given to developments which have occurred within the past two years. Parallel processing is defined and the reasons for its importance in high-performance computing are reviewed. Parallel computer architectures are classified according to the number and power of their processing units, their memory, and the nature of their connection scheme. Architectures which show promise for fluid dynamics applications are emphasized. Fluid dynamics problems are examined for parallelism inherent at the physical level. CFD algorithms and their mappings onto parallel architectures are discussed. Several example are presented to document the performance of fluid dynamics applications on present-generation parallel processing devices

  16. Multiple-enzyme supplementation on digestive traits, carcass characteristics, blood lipid parameters and growth performance of broilers fed a wheat-based diet

    Directory of Open Access Journals (Sweden)

    Hamid Reza Taheri

    2017-09-01

    Full Text Available Objective A trial was conducted from 11 to 42 d post-hatch to investigate the effectiveness of the supplementation of a multiple-enzyme preparation (Natuzyme Plus in a wheat-based diet on digesta viscosity, pH and microbial population, villus morphology, feed passage time, nutrient retention, carcass characteristics, blood lipid parameters and growth performance of broiler chickens. Methods Three hundreds 10-d-old male Ross 308 chicks were allocated to three diets with five replicates of 20 birds per replicate. Dietary treatments were i a wheat-based diet (W, ii W+Natuzyme Plus (WN; 500 mg/kg of the diet, and iii a corn-based diet (C. Results Birds fed on the C diet had higher average daily gain (ADG, p0.05 difference compared to those of the C diet. Compared to those of the W diet, the WN diet showed the higher count of Lactobacilli and lower count of coliforms (p<0.01 and digesta viscosity (p<0.01. Conclusion In general, the results of this study showed that Natuzyme Plus supplementation in a wheat-based diet can be appropriate to achieve a comparable growth performance in broiler chickens to those given the C diet probably through improving digesta viscosity, VH, ET, TTAR of NT and EE, AMEn, count of Lactobacilli and coliforms.

  17. Synchronization Techniques in Parallel Discrete Event Simulation

    OpenAIRE

    Lindén, Jonatan

    2018-01-01

    Discrete event simulation is an important tool for evaluating system models in many fields of science and engineering. To improve the performance of large-scale discrete event simulations, several techniques to parallelize discrete event simulation have been developed. In parallel discrete event simulation, the work of a single discrete event simulation is distributed over multiple processing elements. A key challenge in parallel discrete event simulation is to ensure that causally dependent ...

  18. A survey of parallel multigrid algorithms

    Science.gov (United States)

    Chan, Tony F.; Tuminaro, Ray S.

    1987-01-01

    A typical multigrid algorithm applied to well-behaved linear-elliptic partial-differential equations (PDEs) is described. Criteria for designing and evaluating parallel algorithms are presented. Before evaluating the performance of some parallel multigrid algorithms, consideration is given to some theoretical complexity results for solving PDEs in parallel and for executing the multigrid algorithm. The effect of mapping and load imbalance on the partial efficiency of the algorithm is studied.

  19. Evaluating parallel optimization on transputers

    Directory of Open Access Journals (Sweden)

    A.G. Chalmers

    2003-12-01

    Full Text Available The faster processing power of modern computers and the development of efficient algorithms have made it possible for operations researchers to tackle a much wider range of problems than ever before. Further improvements in processing speed can be achieved utilising relatively inexpensive transputers to process components of an algorithm in parallel. The Davidon-Fletcher-Powell method is one of the most successful and widely used optimisation algorithms for unconstrained problems. This paper examines the algorithm and identifies the components that can be processed in parallel. The results of some experiments with these components are presented which indicates under what conditions parallel processing with an inexpensive configuration is likely to be faster than the traditional sequential implementations. The performance of the whole algorithm with its parallel components is then compared with the original sequential algorithm. The implementation serves to illustrate the practicalities of speeding up typical OR algorithms in terms of difficulty, effort and cost. The results give an indication of the savings in time a given parallel implementation can be expected to yield.

  20. Pattern-Driven Automatic Parallelization

    Directory of Open Access Journals (Sweden)

    Christoph W. Kessler

    1996-01-01

    Full Text Available This article describes a knowledge-based system for automatic parallelization of a wide class of sequential numerical codes operating on vectors and dense matrices, and for execution on distributed memory message-passing multiprocessors. Its main feature is a fast and powerful pattern recognition tool that locally identifies frequently occurring computations and programming concepts in the source code. This tool also works for dusty deck codes that have been "encrypted" by former machine-specific code transformations. Successful pattern recognition guides sophisticated code transformations including local algorithm replacement such that the parallelized code need not emerge from the sequential program structure by just parallelizing the loops. It allows access to an expert's knowledge on useful parallel algorithms, available machine-specific library routines, and powerful program transformations. The partially restored program semantics also supports local array alignment, distribution, and redistribution, and allows for faster and more exact prediction of the performance of the parallelized target code than is usually possible.

  1. Parallelization in Modern C++

    CERN Multimedia

    CERN. Geneva

    2016-01-01

    The traditionally used and well established parallel programming models OpenMP and MPI are both targeting lower level parallelism and are meant to be as language agnostic as possible. For a long time, those models were the only widely available portable options for developing parallel C++ applications beyond using plain threads. This has strongly limited the optimization capabilities of compilers, has inhibited extensibility and genericity, and has restricted the use of those models together with other, modern higher level abstractions introduced by the C++11 and C++14 standards. The recent revival of interest in the industry and wider community for the C++ language has also spurred a remarkable amount of standardization proposals and technical specifications being developed. Those efforts however have so far failed to build a vision on how to seamlessly integrate various types of parallelism, such as iterative parallel execution, task-based parallelism, asynchronous many-task execution flows, continuation s...

  2. Parallelism in matrix computations

    CERN Document Server

    Gallopoulos, Efstratios; Sameh, Ahmed H

    2016-01-01

    This book is primarily intended as a research monograph that could also be used in graduate courses for the design of parallel algorithms in matrix computations. It assumes general but not extensive knowledge of numerical linear algebra, parallel architectures, and parallel programming paradigms. The book consists of four parts: (I) Basics; (II) Dense and Special Matrix Computations; (III) Sparse Matrix Computations; and (IV) Matrix functions and characteristics. Part I deals with parallel programming paradigms and fundamental kernels, including reordering schemes for sparse matrices. Part II is devoted to dense matrix computations such as parallel algorithms for solving linear systems, linear least squares, the symmetric algebraic eigenvalue problem, and the singular-value decomposition. It also deals with the development of parallel algorithms for special linear systems such as banded ,Vandermonde ,Toeplitz ,and block Toeplitz systems. Part III addresses sparse matrix computations: (a) the development of pa...

  3. A parallel buffer tree

    DEFF Research Database (Denmark)

    Sitchinava, Nodar; Zeh, Norbert

    2012-01-01

    We present the parallel buffer tree, a parallel external memory (PEM) data structure for batched search problems. This data structure is a non-trivial extension of Arge's sequential buffer tree to a private-cache multiprocessor environment and reduces the number of I/O operations by the number of...... in the optimal OhOf(psortN + K/PB) parallel I/O complexity, where K is the size of the output reported in the process and psortN is the parallel I/O complexity of sorting N elements using P processors....

  4. Parallel MR imaging.

    Science.gov (United States)

    Deshmane, Anagha; Gulani, Vikas; Griswold, Mark A; Seiberlich, Nicole

    2012-07-01

    Parallel imaging is a robust method for accelerating the acquisition of magnetic resonance imaging (MRI) data, and has made possible many new applications of MR imaging. Parallel imaging works by acquiring a reduced amount of k-space data with an array of receiver coils. These undersampled data can be acquired more quickly, but the undersampling leads to aliased images. One of several parallel imaging algorithms can then be used to reconstruct artifact-free images from either the aliased images (SENSE-type reconstruction) or from the undersampled data (GRAPPA-type reconstruction). The advantages of parallel imaging in a clinical setting include faster image acquisition, which can be used, for instance, to shorten breath-hold times resulting in fewer motion-corrupted examinations. In this article the basic concepts behind parallel imaging are introduced. The relationship between undersampling and aliasing is discussed and two commonly used parallel imaging methods, SENSE and GRAPPA, are explained in detail. Examples of artifacts arising from parallel imaging are shown and ways to detect and mitigate these artifacts are described. Finally, several current applications of parallel imaging are presented and recent advancements and promising research in parallel imaging are briefly reviewed. Copyright © 2012 Wiley Periodicals, Inc.

  5. Parallel Algorithms and Patterns

    Energy Technology Data Exchange (ETDEWEB)

    Robey, Robert W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-06-16

    This is a powerpoint presentation on parallel algorithms and patterns. A parallel algorithm is a well-defined, step-by-step computational procedure that emphasizes concurrency to solve a problem. Examples of problems include: Sorting, searching, optimization, matrix operations. A parallel pattern is a computational step in a sequence of independent, potentially concurrent operations that occurs in diverse scenarios with some frequency. Examples are: Reductions, prefix scans, ghost cell updates. We only touch on parallel patterns in this presentation. It really deserves its own detailed discussion which Gabe Rockefeller would like to develop.

  6. Application Portable Parallel Library

    Science.gov (United States)

    Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott

    1995-01-01

    Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.

  7. Plant polyketide synthases: a chalcone synthase-type enzyme which performs a condensation reaction with methylmalonyl-CoA in the biosynthesis of C-methylated chalcones.

    Science.gov (United States)

    Schröder, J; Raiber, S; Berger, T; Schmidt, A; Schmidt, J; Soares-Sello, A M; Bardshiri, E; Strack, D; Simpson, T J; Veit, M; Schröder, G

    1998-06-09

    Heterologous screening of a cDNA library from Pinusstrobus seedlings identified clones for two chalcone synthase (CHS) related proteins (PStrCHS1 and PStrCHS2, 87.6% identity). Heterologous expression in Escherichia coli showed that PStrCHS1 performed the typical CHS reaction, that it used starter CoA-esters from the phenylpropanoid pathway, and that it performed three condensation reactions with malonyl-CoA, followed by the ring closure to the chalcone. PstrCHS2 was completely inactive with these starters and also with linear CoA-esters. Activity was detected only with a diketide derivative (N-acetylcysteamine thioester of 3-oxo-5-phenylpent-4-enoic acid) that corresponded to the CHS reaction intermediate postulated after the first condensation reaction. PstrCHS2 performed only one condensation, with 6-styryl-4-hydroxy-2-pyrone derivatives as release products. The enzyme preferred methylmalonyl-CoA against malonyl-CoA, if only methylmalonyl-CoA was available. These properties and a comparison with the CHS from Pinus sylvestris suggested for PstrCHS2 a special function in the biosynthesis of secondary products. In contrast to P. sylvestris, P. strobus contains C-methylated chalcone derivatives, and the methyl group is at the position predicted from a chain extension with methylmalonyl-CoA in the second condensation of the biosynthetic reaction sequence. We propose that PstrCHS2 specifically contributes the condensing reaction with methylmalonyl-CoA to yield a methylated triketide intermediate. We discuss a model that the biosynthesis of C-methylated chalcones represents the simplest example of a modular polyketide synthase.

  8. Implantable enzyme amperometric biosensors.

    Science.gov (United States)

    Kotanen, Christian N; Moussy, Francis Gabriel; Carrara, Sandro; Guiseppi-Elie, Anthony

    2012-05-15

    The implantable enzyme amperometric biosensor continues as the dominant in vivo format for the detection, monitoring and reporting of biochemical analytes related to a wide range of pathologies. Widely used in animal studies, there is increasing emphasis on their use in diabetes care and management, the management of trauma-associated hemorrhage and in critical care monitoring by intensivists in the ICU. These frontier opportunities demand continuous indwelling performance for up to several years, well in excess of the currently approved seven days. This review outlines the many challenges to successful deployment of chronically implantable amperometric enzyme biosensors and emphasizes the emerging technological approaches in their continued development. The foreign body response plays a prominent role in implantable biotransducer failure. Topics considering the approaches to mitigate the inflammatory response, use of biomimetic chemistries, nanostructured topographies, drug eluting constructs, and tissue-to-device interface modulus matching are reviewed. Similarly, factors that influence biotransducer performance such as enzyme stability, substrate interference, mediator selection and calibration are reviewed. For the biosensor system, the opportunities and challenges of integration, guided by footprint requirements, the limitations of mixed signal electronics, and power requirements, has produced three systems approaches. The potential is great. However, integration along the multiple length scales needed to address fundamental issues and integration across the diverse disciplines needed to achieve success of these highly integrated systems, continues to be a challenge in the development and deployment of implantable amperometric enzyme biosensor systems. Copyright © 2012 Elsevier B.V. All rights reserved.

  9. Computer-Aided Parallelizer and Optimizer

    Science.gov (United States)

    Jin, Haoqiang

    2011-01-01

    The Computer-Aided Parallelizer and Optimizer (CAPO) automates the insertion of compiler directives (see figure) to facilitate parallel processing on Shared Memory Parallel (SMP) machines. While CAPO currently is integrated seamlessly into CAPTools (developed at the University of Greenwich, now marketed as ParaWise), CAPO was independently developed at Ames Research Center as one of the components for the Legacy Code Modernization (LCM) project. The current version takes serial FORTRAN programs, performs interprocedural data dependence analysis, and generates OpenMP directives. Due to the widely supported OpenMP standard, the generated OpenMP codes have the potential to run on a wide range of SMP machines. CAPO relies on accurate interprocedural data dependence information currently provided by CAPTools. Compiler directives are generated through identification of parallel loops in the outermost level, construction of parallel regions around parallel loops and optimization of parallel regions, and insertion of directives with automatic identification of private, reduction, induction, and shared variables. Attempts also have been made to identify potential pipeline parallelism (implemented with point-to-point synchronization). Although directives are generated automatically, user interaction with the tool is still important for producing good parallel codes. A comprehensive graphical user interface is included for users to interact with the parallelization process.

  10. Electro-ultrafiltration of industrial enzyme solutions

    DEFF Research Database (Denmark)

    Enevoldsen, Ann Dorrit; Hansen, Erik Børresen; Jonsson, Gunnar Eigil

    2007-01-01

    To reduce the problems with fouling and concentration polarization during crossflow ultrafiltration of industrial enzyme solutions an electric field is applied across the membrane. The filtration performance during electro-ultrafiltration (EUF) has been tested with several enzymes. Results show...

  11. Advanced parallel processing with supercomputer architectures

    International Nuclear Information System (INIS)

    Hwang, K.

    1987-01-01

    This paper investigates advanced parallel processing techniques and innovative hardware/software architectures that can be applied to boost the performance of supercomputers. Critical issues on architectural choices, parallel languages, compiling techniques, resource management, concurrency control, programming environment, parallel algorithms, and performance enhancement methods are examined and the best answers are presented. The authors cover advanced processing techniques suitable for supercomputers, high-end mainframes, minisupers, and array processors. The coverage emphasizes vectorization, multitasking, multiprocessing, and distributed computing. In order to achieve these operation modes, parallel languages, smart compilers, synchronization mechanisms, load balancing methods, mapping parallel algorithms, operating system functions, application library, and multidiscipline interactions are investigated to ensure high performance. At the end, they assess the potentials of optical and neural technologies for developing future supercomputers

  12. Productive Parallel Programming: The PCN Approach

    Directory of Open Access Journals (Sweden)

    Ian Foster

    1992-01-01

    Full Text Available We describe the PCN programming system, focusing on those features designed to improve the productivity of scientists and engineers using parallel supercomputers. These features include a simple notation for the concise specification of concurrent algorithms, the ability to incorporate existing Fortran and C code into parallel applications, facilities for reusing parallel program components, a portable toolkit that allows applications to be developed on a workstation or small parallel computer and run unchanged on supercomputers, and integrated debugging and performance analysis tools. We survey representative scientific applications and identify problem classes for which PCN has proved particularly useful.

  13. Parallel thermal radiation transport in two dimensions

    International Nuclear Information System (INIS)

    Smedley-Stevenson, R.P.; Ball, S.R.

    2003-01-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  14. Parallel Algorithms for the Exascale Era

    Energy Technology Data Exchange (ETDEWEB)

    Robey, Robert W. [Los Alamos National Laboratory

    2016-10-19

    New parallel algorithms are needed to reach the Exascale level of parallelism with millions of cores. We look at some of the research developed by students in projects at LANL. The research blends ideas from the early days of computing while weaving in the fresh approach brought by students new to the field of high performance computing. We look at reproducibility of global sums and why it is important to parallel computing. Next we look at how the concept of hashing has led to the development of more scalable algorithms suitable for next-generation parallel computers. Nearly all of this work has been done by undergraduates and published in leading scientific journals.

  15. Parallel thermal radiation transport in two dimensions

    Energy Technology Data Exchange (ETDEWEB)

    Smedley-Stevenson, R.P.; Ball, S.R. [AWE Aldermaston (United Kingdom)

    2003-07-01

    This paper describes the distributed memory parallel implementation of a deterministic thermal radiation transport algorithm in a 2-dimensional ALE hydrodynamics code. The parallel algorithm consists of a variety of components which are combined in order to produce a state of the art computational capability, capable of solving large thermal radiation transport problems using Blue-Oak, the 3 Tera-Flop MPP (massive parallel processors) computing facility at AWE (United Kingdom). Particular aspects of the parallel algorithm are described together with examples of the performance on some challenging applications. (author)

  16. Effect of different levels of alpha tocopherol on performance traits, serum antioxidant enzymes, and trace elements in Japanese quail ( Coturnix coturnix japonica under low ambient temperature

    Directory of Open Access Journals (Sweden)

    Assar Ali Shah

    Full Text Available ABSTRACT This study was designed to find the effect of vitamin E supplementation on growth, serum antioxidant enzymes, and some trace elements in Japanese quail (Coturnix coturnix japonica under low ambient temperature. A total of 180 day-old Japanese quails were randomly divided into four groups and provided with 0 (group A, 50 (group B, 100 (group C, and 150 IU/kg (group D vitamin E (dl-α-tocopherol acetate under an average temperature of 9±0.5 °C for an experimental period of 42 days. The result showed that feed intake per day, body weight, weight gain per day, and feed conversion ratio did not differ significantly between the groups. Serum concentrations of superoxide and glutathione peroxidase were significantly high in birds supplemented with 150 mg/kg of vitamin E. The concentration of aspartate aminotransferase was not significantly affected between the control and treated groups; however, alanine transaminase concentration significantly reduced in group D. Zinc concentration in the blood increased significantly in group D, with no significant effect on copper and manganese between the control and treated groups. Vitamin E at the level of 150 IU/kg of feed improves the blood antioxidant status and zinc concentration, with no effect on the performance traits of quail reared under low ambient temperature.

  17. Application of 3D Printing Technology in Increasing the Diagnostic Performance of Enzyme-Linked Immunosorbent Assay (ELISA for Infectious Diseases

    Directory of Open Access Journals (Sweden)

    Harpal Singh

    2015-07-01

    Full Text Available Enzyme-linked Immunosorbent Assay (ELISA-based diagnosis is the mainstay for measuring antibody response in infectious diseases and to support pathogen identification of potential use in infectious disease outbreaks and clinical care of individual patients. The development of laboratory diagnostics using readily available 3D printing technologies provides a timely opportunity for further expansion of this technology into immunodetection systems. Utilizing available 3D printing platforms, a ‘3D well’ was designed and developed to have an increased surface area compared to those of 96-well plates. The ease and rapidity of the development of the 3D well prototype provided an opportunity for its rapid validation through the diagnostic performance of ELISA in infectious disease without modifying current laboratory practices for ELISA. The improved sensitivity of the 3D well of up to 2.25-fold higher compared to the 96-well ELISA provides a potential for the expansion of this technology towards miniaturization and Lab-On-a-Chip platforms to reduce time, volume of reagents and samples needed for such assays in the laboratory diagnosis of infectious and other diseases including applications in other disciplines.

  18. Potential Therapeutic Applications of Mucuna pruriens Peptide Fractions Purified by High-Performance Liquid Chromatography as Angiotensin-Converting Enzyme Inhibitors, Antioxidants, Antithrombotic and Hypocholesterolemic Agents.

    Science.gov (United States)

    Herrera-Chalé, Francisco; Ruiz-Ruiz, Jorge Carlos; Betancur-Ancona, David; Segura-Campos, Maira Rubi

    2016-02-01

    A Mucuna pruriens protein concentrate was hydrolyzed with a digestive (pepsin-pancreatin) enzymatic system. The soluble portion of the hydrolysate was fractionated by ultrafiltration and the ultrafiltered peptide fraction (PF) with lower molecular weight was purified by reversed-phase high-performance liquid chromatography. The PF obtained were evaluated by testing the biological activity in vitro. Fractions showed that the ability to inhibit the angiotensin-converting enzyme had IC50 values that ranged from 2.7 to 6.2 μg/mL. Trolox equivalent antioxidant capacity values ranged from 132.20 to 507.43 mM/mg. The inhibition of human platelet aggregation ranged from 1.59% to 11.11%, and the inhibition of cholesterol micellar solubility ranged from 0.24% to 0.47%. Hydrophobicity, size, and amino acid sequence could be factors in determining the biological activity of peptides contained in fractions. This is the first report that M. pruriens peptides act as antihypertensives, antioxidants, and inhibitors for human platelet aggregation and cholesterol micellar solubility in vitro.

  19. Application of 3D Printing Technology in Increasing the Diagnostic Performance of Enzyme-Linked Immunosorbent Assay (ELISA) for Infectious Diseases

    Science.gov (United States)

    Singh, Harpal; Shimojima, Masayuki; Shiratori, Tomomi; An, Le Van; Sugamata, Masami; Yang, Ming

    2015-01-01

    Enzyme-linked Immunosorbent Assay (ELISA)-based diagnosis is the mainstay for measuring antibody response in infectious diseases and to support pathogen identification of potential use in infectious disease outbreaks and clinical care of individual patients. The development of laboratory diagnostics using readily available 3D printing technologies provides a timely opportunity for further expansion of this technology into immunodetection systems. Utilizing available 3D printing platforms, a ‘3D well’ was designed and developed to have an increased surface area compared to those of 96-well plates. The ease and rapidity of the development of the 3D well prototype provided an opportunity for its rapid validation through the diagnostic performance of ELISA in infectious disease without modifying current laboratory practices for ELISA. The improved sensitivity of the 3D well of up to 2.25-fold higher compared to the 96-well ELISA provides a potential for the expansion of this technology towards miniaturization and Lab-On-a-Chip platforms to reduce time, volume of reagents and samples needed for such assays in the laboratory diagnosis of infectious and other diseases including applications in other disciplines. PMID:26184194

  20. Application of 3D Printing Technology in Increasing the Diagnostic Performance of Enzyme-Linked Immunosorbent Assay (ELISA) for Infectious Diseases.

    Science.gov (United States)

    Singh, Harpal; Shimojima, Masayuki; Shiratori, Tomomi; An, Le Van; Sugamata, Masami; Yang, Ming

    2015-07-08

    Enzyme-linked Immunosorbent Assay (ELISA)-based diagnosis is the mainstay for measuring antibody response in infectious diseases and to support pathogen identification of potential use in infectious disease outbreaks and clinical care of individual patients. The development of laboratory diagnostics using readily available 3D printing technologies provides a timely opportunity for further expansion of this technology into immunodetection systems. Utilizing available 3D printing platforms, a '3D well' was designed and developed to have an increased surface area compared to those of 96-well plates. The ease and rapidity of the development of the 3D well prototype provided an opportunity for its rapid validation through the diagnostic performance of ELISA in infectious disease without modifying current laboratory practices for ELISA. The improved sensitivity of the 3D well of up to 2.25-fold higher compared to the 96-well ELISA provides a potential for the expansion of this technology towards miniaturization and Lab-On-a-Chip platforms to reduce time, volume of reagents and samples needed for such assays in the laboratory diagnosis of infectious and other diseases including applications in other disciplines.

  1. Customizable Memory Schemes for Data Parallel Architectures

    NARCIS (Netherlands)

    Gou, C.

    2011-01-01

    Memory system efficiency is crucial for any processor to achieve high performance, especially in the case of data parallel machines. Processing capabilities of parallel lanes will be wasted, when data requests are not accomplished in a sustainable and timely manner. Irregular vector memory accesses

  2. Design considerations for parallel graphics libraries

    Science.gov (United States)

    Crockett, Thomas W.

    1994-01-01

    Applications which run on parallel supercomputers are often characterized by massive datasets. Converting these vast collections of numbers to visual form has proven to be a powerful aid to comprehension. For a variety of reasons, it may be desirable to provide this visual feedback at runtime. One way to accomplish this is to exploit the available parallelism to perform graphics operations in place. In order to do this, we need appropriate parallel rendering algorithms and library interfaces. This paper provides a tutorial introduction to some of the issues which arise in designing parallel graphics libraries and their underlying rendering algorithms. The focus is on polygon rendering for distributed memory message-passing systems. We illustrate our discussion with examples from PGL, a parallel graphics library which has been developed on the Intel family of parallel systems.

  3. Totally parallel multilevel algorithms

    Science.gov (United States)

    Frederickson, Paul O.

    1988-01-01

    Four totally parallel algorithms for the solution of a sparse linear system have common characteristics which become quite apparent when they are implemented on a highly parallel hypercube such as the CM2. These four algorithms are Parallel Superconvergent Multigrid (PSMG) of Frederickson and McBryan, Robust Multigrid (RMG) of Hackbusch, the FFT based Spectral Algorithm, and Parallel Cyclic Reduction. In fact, all four can be formulated as particular cases of the same totally parallel multilevel algorithm, which are referred to as TPMA. In certain cases the spectral radius of TPMA is zero, and it is recognized to be a direct algorithm. In many other cases the spectral radius, although not zero, is small enough that a single iteration per timestep keeps the local error within the required tolerance.

  4. Influence of dietary inclusion of Bacillus licheniformis on laying performance, egg quality, antioxidant enzyme activities, and intestinal barrier function of laying hens.

    Science.gov (United States)

    Lei, K; Li, Y L; Yu, D Y; Rajput, I R; Li, W F

    2013-09-01

    This experiment was conducted to evaluate the effects of dietary inclusion of Bacillus licheniformis on laying performance, egg quality, antioxidant enzyme activities, and intestinal barrier function of laying hens. Hy-Line Variety W-36 hens (n = 540; 28 wk of age) were randomized into 6 groups, each group with 6 replications (n = 15). The control group received the basal diet formulated with maize and soybean meal. The treatment groups received the same basal diets supplemented with 0.01, 0.02, 0.03, 0.06, and 0.09% Bacillus licheniformis powder (2 × 10(10) cfu/g) for an 8-wk trial. The results showed that dietary supplementation with 0.01 and 0.03% B. licheniformis significantly increased egg production and egg mass. However, no significant differences were observed in egg weight, feed consumption, and feed conversion efficiency among the 6 groups. Supplementation with different levels of B. licheniformis was found to be effective in improvement of egg quality by increasing egg shell thickness and strength. Compared with control, d-lactate content, diamine oxidase activity, and adrenocorticotropic hormone level in serum decreased significantly, and the level of estradiol and follicle-stimulating hormone increased significantly in plasma of all the experimental groups. Dietary supplementation with B. licheniformis increased the intestinal villus height and reduced the crypt depth. In conclusion, dietary inclusion of B. licheniformis could improve laying performance and egg quality significantly in a dose-dependent manner by decreasing the stress response, upregulating the growth hormone, and improving intestinal health.

  5. Comparison of clinical performance of antigen based-enzyme immunoassay (EIA) and major outer membrane protein (MOMP)-PCR for detection of genital Chlamydia trachomatis infection.

    Science.gov (United States)

    Nateghi Rostami, Mahmoud; Hossein Rashidi, Batool; Aghsaghloo, Fatemeh; Nazari, Razieh

    2016-06-01

    Chlamydia trachomatis is the most common sexually transmitted bacterial pathogen worldwide. Early detection and treatment of C.trachomatis genital infection prevent serious reproductive complications. Performances of enzyme immunoassay (EIA) and major outer membrane protein (MOMP)-polymerase chain reaction (PCR) for diagnosis of genital C.trachomatis infection in women were compared. In this cross sectional study a total of 518 women volunteers were included (33.67±8.3 yrs) who had been referred to Gynecology clinics of Qom province, Iran, were included. Endocervical swab specimens were collected to detect lipopolysaccharide (LPS) antigen in EIA and to amplify MOMP gene of C.trachomatis in PCR. Results were confirmed using ompI nested-PCR. Sensitivity, specificity, positive (PPV) and negative predictive values (NPV) were calculated for performance of the tests. Odds ratios were determined using binary logistic regression analysis. In total, 37 (7.14%) cases were positive by EIA and/or MOMP-PCR. All discrepant results were confirmed by nested-PCR. Sensitivity, specificity, PPV and NPV values of EIA were 59.46%, 100%, 100% and 96.98%, and those of MOMP-PCR were 97.30%, 100%, 100%, 99.79%, respectively. Reproductive complications including 2.7% ectopic pregnancy, 5.4% stillbirth, 5.4% infertility, and 10.8% PROM were recorded. The risk of developing chlamydiosis was increased 4.8-fold in volunteers with cervicitis (p<0.05; OR 4.80; 95% CI 1.25-18.48). C.trachomatis infection should be regarded in women of reproductive ages especially those with cervicitis. Primary screening of women by using the low cost antigen-EIA is recommended; however, due to the low sensitivity of Ag-EIA, verification of the negative results by a DNA amplification method is needed.

  6. Modern industrial simulation tools: Kernel-level integration of high performance parallel processing, object-oriented numerics, and adaptive finite element analysis. Final report, July 16, 1993--September 30, 1997

    Energy Technology Data Exchange (ETDEWEB)

    Deb, M.K.; Kennon, S.R.

    1998-04-01

    A cooperative R&D effort between industry and the US government, this project, under the HPPP (High Performance Parallel Processing) initiative of the Dept. of Energy, started the investigations into parallel object-oriented (OO) numerics. The basic goal was to research and utilize the emerging technologies to create a physics-independent computational kernel for applications using adaptive finite element method. The industrial team included Computational Mechanics Co., Inc. (COMCO) of Austin, TX (as the primary contractor), Scientific Computing Associates, Inc. (SCA) of New Haven, CT, Texaco and CONVEX. Sandia National Laboratory (Albq., NM) was the technology partner from the government side. COMCO had the responsibility of the main kernel design and development, SCA had the lead in parallel solver technology and guidance on OO technologies was Sandia`s main expertise in this venture. CONVEX and Texaco supported the partnership by hardware resource and application knowledge, respectively. As such, a minimum of fifty-percent cost-sharing was provided by the industry partnership during this project. This report describes the R&D activities and provides some details about the prototype kernel and example applications.

  7. Computational enzyme design: transitioning from catalytic proteins to enzymes.

    Science.gov (United States)

    Mak, Wai Shun; Siegel, Justin B

    2014-08-01

    The widespread interest in enzymes stem from their ability to catalyze chemical reactions under mild and ecologically friendly conditions with unparalleled catalytic proficiencies. While thousands of naturally occurring enzymes have been identified and characterized, there are still numerous important applications for which there are no biological catalysts capable of performing the desired chemical transformation. In order to engineer enzymes for which there is no natural starting point, efforts using a combination of quantum chemistry and force-field based protein molecular modeling have led to the design of novel proteins capable of catalyzing chemical reactions not catalyzed by naturally occurring enzymes. Here we discuss the current status and potential avenues to pursue as the field of computational enzyme design moves forward. Published by Elsevier Ltd.

  8. Parallel R-matrix computation

    International Nuclear Information System (INIS)

    Heggarty, J.W.

    1999-06-01

    For almost thirty years, sequential R-matrix computation has been used by atomic physics research groups, from around the world, to model collision phenomena involving the scattering of electrons or positrons with atomic or molecular targets. As considerable progress has been made in the understanding of fundamental scattering processes, new data, obtained from more complex calculations, is of current interest to experimentalists. Performing such calculations, however, places considerable demands on the computational resources to be provided by the target machine, in terms of both processor speed and memory requirement. Indeed, in some instances the computational requirements are so great that the proposed R-matrix calculations are intractable, even when utilising contemporary classic supercomputers. Historically, increases in the computational requirements of R-matrix computation were accommodated by porting the problem codes to a more powerful classic supercomputer. Although this approach has been successful in the past, it is no longer considered to be a satisfactory solution due to the limitations of current (and future) Von Neumann machines. As a consequence, there has been considerable interest in the high performance multicomputers, that have emerged over the last decade which appear to offer the computational resources required by contemporary R-matrix research. Unfortunately, developing codes for these machines is not as simple a task as it was to develop codes for successive classic supercomputers. The difficulty arises from the considerable differences in the computing models that exist between the two types of machine and results in the programming of multicomputers to be widely acknowledged as a difficult, time consuming and error-prone task. Nevertheless, unless parallel R-matrix computation is realised, important theoretical and experimental atomic physics research will continue to be hindered. This thesis describes work that was undertaken in

  9. Parallel imaging with phase scrambling.

    Science.gov (United States)

    Zaitsev, Maxim; Schultz, Gerrit; Hennig, Juergen; Gruetter, Rolf; Gallichan, Daniel

    2015-04-01

    Most existing methods for accelerated parallel imaging in MRI require additional data, which are used to derive information about the sensitivity profile of each radiofrequency (RF) channel. In this work, a method is presented to avoid the acquisition of separate coil calibration data for accelerated Cartesian trajectories. Quadratic phase is imparted to the image to spread the signals in k-space (aka phase scrambling). By rewriting the Fourier transform as a convolution operation, a window can be introduced to the convolved chirp function, allowing a low-resolution image to be reconstructed from phase-scrambled data without prominent aliasing. This image (for each RF channel) can be used to derive coil sensitivities to drive existing parallel imaging techniques. As a proof of concept, the quadratic phase was applied by introducing an offset to the x(2) - y(2) shim and the data were reconstructed using adapted versions of the image space-based sensitivity encoding and GeneRalized Autocalibrating Partially Parallel Acquisitions algorithms. The method is demonstrated in a phantom (1 × 2, 1 × 3, and 2 × 2 acceleration) and in vivo (2 × 2 acceleration) using a 3D gradient echo acquisition. Phase scrambling can be used to perform parallel imaging acceleration without acquisition of separate coil calibration data, demonstrated here for a 3D-Cartesian trajectory. Further research is required to prove the applicability to other 2D and 3D sampling schemes. © 2014 Wiley Periodicals, Inc.

  10. Wakefield calculations on parallel computers

    International Nuclear Information System (INIS)

    Schoessow, P.

    1990-01-01

    The use of parallelism in the solution of wakefield problems is illustrated for two different computer architectures (SIMD and MIMD). Results are given for finite difference codes which have been implemented on a Connection Machine and an Alliant FX/8 and which are used to compute wakefields in dielectric loaded structures. Benchmarks on code performance are presented for both cases. 4 refs., 3 figs., 2 tabs

  11. Enzyme detection by microfluidics

    DEFF Research Database (Denmark)

    2013-01-01

    Microfluidic-implemented methods of detecting an enzyme, in particular a DNA-modifying enzyme, are provided, as well as methods for detecting a cell, or a microorganism expressing said enzyme. The enzyme is detected by providing a nucleic acid substrate, which is specifically targeted...... by that enzyme...

  12. Algorithms for parallel computers

    International Nuclear Information System (INIS)

    Churchhouse, R.F.

    1985-01-01

    Until relatively recently almost all the algorithms for use on computers had been designed on the (usually unstated) assumption that they were to be run on single processor, serial machines. With the introduction of vector processors, array processors and interconnected systems of mainframes, minis and micros, however, various forms of parallelism have become available. The advantage of parallelism is that it offers increased overall processing speed but it also raises some fundamental questions, including: (i) which, if any, of the existing 'serial' algorithms can be adapted for use in the parallel mode. (ii) How close to optimal can such adapted algorithms be and, where relevant, what are the convergence criteria. (iii) How can we design new algorithms specifically for parallel systems. (iv) For multi-processor systems how can we handle the software aspects of the interprocessor communications. Aspects of these questions illustrated by examples are considered in these lectures. (orig.)

  13. Parallelism and array processing

    International Nuclear Information System (INIS)

    Zacharov, V.

    1983-01-01

    Modern computing, as well as the historical development of computing, has been dominated by sequential monoprocessing. Yet there is the alternative of parallelism, where several processes may be in concurrent execution. This alternative is discussed in a series of lectures, in which the main developments involving parallelism are considered, both from the standpoint of computing systems and that of applications that can exploit such systems. The lectures seek to discuss parallelism in a historical context, and to identify all the main aspects of concurrency in computation right up to the present time. Included will be consideration of the important question as to what use parallelism might be in the field of data processing. (orig.)

  14. Automatic Parallelization Tool: Classification of Program Code for Parallel Computing

    Directory of Open Access Journals (Sweden)

    Mustafa Basthikodi

    2016-04-01

    Full Text Available Performance growth of single-core processors has come to a halt in the past decade, but was re-enabled by the introduction of parallelism in processors. Multicore frameworks along with Graphical Processing Units empowered to enhance parallelism broadly. Couples of compilers are updated to developing challenges forsynchronization and threading issues. Appropriate program and algorithm classifications will have advantage to a great extent to the group of software engineers to get opportunities for effective parallelization. In present work we investigated current species for classification of algorithms, in that related work on classification is discussed along with the comparison of issues that challenges the classification. The set of algorithms are chosen which matches the structure with different issues and perform given task. We have tested these algorithms utilizing existing automatic species extraction toolsalong with Bones compiler. We have added functionalities to existing tool, providing a more detailed characterization. The contributions of our work include support for pointer arithmetic, conditional and incremental statements, user defined types, constants and mathematical functions. With this, we can retain significant data which is not captured by original speciesof algorithms. We executed new theories into the device, empowering automatic characterization of program code.

  15. Parallel magnetic resonance imaging

    International Nuclear Information System (INIS)

    Larkman, David J; Nunes, Rita G

    2007-01-01

    Parallel imaging has been the single biggest innovation in magnetic resonance imaging in the last decade. The use of multiple receiver coils to augment the time consuming Fourier encoding has reduced acquisition times significantly. This increase in speed comes at a time when other approaches to acquisition time reduction were reaching engineering and human limits. A brief summary of spatial encoding in MRI is followed by an introduction to the problem parallel imaging is designed to solve. There are a large number of parallel reconstruction algorithms; this article reviews a cross-section, SENSE, SMASH, g-SMASH and GRAPPA, selected to demonstrate the different approaches. Theoretical (the g-factor) and practical (coil design) limits to acquisition speed are reviewed. The practical implementation of parallel imaging is also discussed, in particular coil calibration. How to recognize potential failure modes and their associated artefacts are shown. Well-established applications including angiography, cardiac imaging and applications using echo planar imaging are reviewed and we discuss what makes a good application for parallel imaging. Finally, active research areas where parallel imaging is being used to improve data quality by repairing artefacted images are also reviewed. (invited topical review)

  16. Effects of juniper essential oil on growth performance, some rumen protozoa, rumen fermentation and antioxidant blood enzyme parameters of growing Saanen kids.

    Science.gov (United States)

    Yesilbag, D; Biricik, H; Cetin, I; Kara, C; Meral, Y; Cengiz, S S; Orman, A; Udum, D

    2017-10-01

    This study aimed to evaluate the effects of juniper essential oil on the growth performance, rumen fermentation parameters, rumen protozoa population, blood antioxidant enzyme parameters and faecal content in growing Saanen kids. Thirty-six male Saanen kids (36 ± 14 days of age) were used in the study. Each group consisted of 9 kids. The control group (G1) was fed with a diet that consisted of the above concentrated feed and oat hay, whereas the experimental groups consumed the same diet but with the concentrated feed uniformly sprayed with juniper essential oil 0.4 ml/kg (G2), 0.8 ml/kg (G3) or 2 ml/kg (G4). There were no differences (p > 0.05) in live weight, live weight gain or feed consumption between the control and experimental groups. There was a significant improvement (p rumen pH, rumen volatile fatty acid (VFA) profile or faecal pH of the control and experimental groups. The rumen NH 3 N values were similar at the middle and end of the experiment, but at the start of the experiment, the rumen NH 3 N values differed between the control and experimental groups (p < 0.05). The faecal score value was significantly (p < 0.05) decreased in the experimental groups. The addition of juniper essential oil supplementation to the rations caused significant effects on the kids' antioxidant blood parameters. Although the superoxide dismutase (SOD) activity, total antioxidant capacity (TAC) and catalase values were significantly (p < 0.05) increased in the experimental groups (G2, G3 and G4), especially group G4, the blood glutathione peroxidase (GPX) value significantly decreased in the experimental groups. The results of this study suggest that supplementation of juniper oil is more effective on antioxidant parameters than on performance parameters and may be used as a natural antioxidant product. Journal of Animal Physiology and Animal Nutrition © 2016 Blackwell Verlag GmbH.

  17. Effect of γ-Aminobutyric Acid-producing Strain on Laying Performance, Egg Quality and Serum Enzyme Activity in Hy-Line Brown Hens under Heat Stress

    Directory of Open Access Journals (Sweden)

    Y. Z. Zhu

    2015-07-01

    Full Text Available Heat-stress remains a costly issue for animal production, especially for poultry as they lack sweat glands, and alleviating heat-stress is necessary for ensuring animal production in hot environment. A high γ-aminobutyric acid (GABA-producer Lactobacillus strain was used to investigate the effect of dietary GABA-producer on laying performance and egg quality in heat-stressed Hy-line brown hens. Hy-Line brown hens (n = 1,164 at 280 days of age were randomly divided into 4 groups based on the amount of freeze-dried GABA-producer added to the basal diet as follows: i 0 mg/kg, ii 25 mg/kg, iii 50 mg/kg, and iv 100 mg/kg. All hens were subjected to heat-stress treatment through maintaining the temperature and the relative humidity at 28.83±3.85°C and 37% to 53.9%, respectively. During the experiment, laying rate, egg weight and feed intake of hens were recorded daily. At the 30th and 60th day after the start of the experiment, biochemical parameters, enzyme activity and immune activity in serum were measured. Egg production, average egg weight, average daily feed intake, feed conversion ratio and percentage of speckled egg, soft shell egg and misshaped egg were significantly improved (p<0.05 by the increasing supplementation of the dietary GABA-producer. Shape index, eggshell thickness, strength and weight were increased linearly with increasing GABA-producer supplementation. The level of calcium, phosphorus, glucose, total protein and albumin in serum of the hens fed GABA-producing strain supplemented diet was significantly higher (p<0.05 than that of the hens fed the basal diet, whereas cholesterol level was decreased. Compared with the basal diet, GABA-producer strain supplementation increased serum level of glutathione peroxidase (p = 0.009 and superoxide dismutase. In conclusion, GABA-producer played an important role in alleviating heat-stress, the isolated GABA-producer strain might be a potential natural and safe probiotic to use to

  18. Effect of γ-Aminobutyric Acid-producing Lactobacillus Strain on Laying Performance, Egg Quality and Serum Enzyme Activity in Hy-Line Brown Hens under Heat Stress.

    Science.gov (United States)

    Zhu, Y Z; Cheng, J L; Ren, M; Yin, L; Piao, X S

    2015-07-01

    Heat-stress remains a costly issue for animal production, especially for poultry as they lack sweat glands, and alleviating heat-stress is necessary for ensuring animal production in hot environment. A high γ-aminobutyric acid (GABA)-producer Lactobacillus strain was used to investigate the effect of dietary GABA-producer on laying performance and egg quality in heat-stressed Hy-line brown hens. Hy-Line brown hens (n = 1,164) at 280 days of age were randomly divided into 4 groups based on the amount of freeze-dried GABA-producer added to the basal diet as follows: i) 0 mg/kg, ii) 25 mg/kg, iii) 50 mg/kg, and iv) 100 mg/kg. All hens were subjected to heat-stress treatment through maintaining the temperature and the relative humidity at 28.83±3.85°C and 37% to 53.9%, respectively. During the experiment, laying rate, egg weight and feed intake of hens were recorded daily. At the 30th and 60th day after the start of the experiment, biochemical parameters, enzyme activity and immune activity in serum were measured. Egg production, average egg weight, average daily feed intake, feed conversion ratio and percentage of speckled egg, soft shell egg and misshaped egg were significantly improved (pGABA-producer. Shape index, eggshell thickness, strength and weight were increased linearly with increasing GABA-producer supplementation. The level of calcium, phosphorus, glucose, total protein and albumin in serum of the hens fed GABA-producing strain supplemented diet was significantly higher (plevel was decreased. Compared with the basal diet, GABA-producer strain supplementation increased serum level of glutathione peroxidase (p = 0.009) and superoxide dismutase. In conclusion, GABA-producer played an important role in alleviating heat-stress, the isolated GABA-producer strain might be a potential natural and safe probiotic to use to improve laying performance and egg quality in heat-stressed hens.

  19. Parallel Architectures and Parallel Algorithms for Integrated Vision Systems. Ph.D. Thesis

    Science.gov (United States)

    Choudhary, Alok Nidhi

    1989-01-01

    Computer vision is regarded as one of the most complex and computationally intensive problems. An integrated vision system (IVS) is a system that uses vision algorithms from all levels of processing to perform for a high level application (e.g., object recognition). An IVS normally involves algorithms from low level, intermediate level, and high level vision. Designing parallel architectures for vision systems is of tremendous interest to researchers. Several issues are addressed in parallel architectures and parallel algorithms for integrated vision systems.

  20. Re-targeting the Graze performance debugging tool for Java threads and analyzing the re-targeting to automatically parallelized (FORTRAN) code

    OpenAIRE

    Tsai, Pedro T. H.

    2000-01-01

    Approved for public release; distribution is unlimited This research focuses on the design of a language-independent concept, Glimpse, for performance debugging of multi-threaded programs. This research extends previous work on Graze, a tool designed and implemented for performance debugging of C++ programs. Not only is Glimpse easily portable among different programming languages, (i) it is useful in many different paradigms ranging from few long-lived threads to many short-lived...

  1. Coupled enzyme reactions performed in heterogeneous reaction media: experiments and modeling for glucose oxidase and horseradish peroxidase in a PEG/citrate aqueous two-phase system.

    Science.gov (United States)

    Aumiller, William M; Davis, Bradley W; Hashemian, Negar; Maranas, Costas; Armaou, Antonios; Keating, Christine D

    2014-03-06

    The intracellular environment in which biological reactions occur is crowded with macromolecules and subdivided into microenvironments that differ in both physical properties and chemical composition. The work described here combines experimental and computational model systems to help understand the consequences of this heterogeneous reaction media on the outcome of coupled enzyme reactions. Our experimental model system for solution heterogeneity is a biphasic polyethylene glycol (PEG)/sodium citrate aqueous mixture that provides coexisting PEG-rich and citrate-rich phases. Reaction kinetics for the coupled enzyme reaction between glucose oxidase (GOX) and horseradish peroxidase (HRP) were measured in the PEG/citrate aqueous two-phase system (ATPS). Enzyme kinetics differed between the two phases, particularly for the HRP. Both enzymes, as well as the substrates glucose and H2O2, partitioned to the citrate-rich phase; however, the Amplex Red substrate necessary to complete the sequential reaction partitioned strongly to the PEG-rich phase. Reactions in ATPS were quantitatively described by a mathematical model that incorporated measured partitioning and kinetic parameters. The model was then extended to new reaction conditions, i.e., higher enzyme concentration. Both experimental and computational results suggest mass transfer across the interface is vital to maintain the observed rate of product formation, which may be a means of metabolic regulation in vivo. Although outcomes for a specific system will depend on the particulars of the enzyme reactions and the microenvironments, this work demonstrates how coupled enzymatic reactions in complex, heterogeneous media can be understood in terms of a mathematical model.

  2. Elevated Liver Enzymes

    Science.gov (United States)

    Symptoms Elevated liver enzymes By Mayo Clinic Staff Elevated liver enzymes may indicate inflammation or damage to cells in the liver. Inflamed or ... than normal amounts of certain chemicals, including liver enzymes, into the bloodstream, which can result in elevated ...

  3. Differences Between Distributed and Parallel Systems

    Energy Technology Data Exchange (ETDEWEB)

    Brightwell, R.; Maccabe, A.B.; Rissen, R.

    1998-10-01

    Distributed systems have been studied for twenty years and are now coming into wider use as fast networks and powerful workstations become more readily available. In many respects a massively parallel computer resembles a network of workstations and it is tempting to port a distributed operating system to such a machine. However, there are significant differences between these two environments and a parallel operating system is needed to get the best performance out of a massively parallel system. This report characterizes the differences between distributed systems, networks of workstations, and massively parallel systems and analyzes the impact of these differences on operating system design. In the second part of the report, we introduce Puma, an operating system specifically developed for massively parallel systems. We describe Puma portals, the basic building blocks for message passing paradigms implemented on top of Puma, and show how the differences observed in the first part of the report have influenced the design and implementation of Puma.

  4. Parallel-In-Time For Moving Meshes

    Energy Technology Data Exchange (ETDEWEB)

    Falgout, R. D. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Manteuffel, T. A. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Southworth, B. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Schroder, J. B. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2016-02-04

    With steadily growing computational resources available, scientists must develop e ective ways to utilize the increased resources. High performance, highly parallel software has be- come a standard. However until recent years parallelism has focused primarily on the spatial domain. When solving a space-time partial di erential equation (PDE), this leads to a sequential bottleneck in the temporal dimension, particularly when taking a large number of time steps. The XBraid parallel-in-time library was developed as a practical way to add temporal parallelism to existing se- quential codes with only minor modi cations. In this work, a rezoning-type moving mesh is applied to a di usion problem and formulated in a parallel-in-time framework. Tests and scaling studies are run using XBraid and demonstrate excellent results for the simple model problem considered herein.

  5. 6th International Parallel Tools Workshop

    CERN Document Server

    Brinkmann, Steffen; Gracia, José; Resch, Michael; Nagel, Wolfgang

    2013-01-01

    The latest advances in the High Performance Computing hardware have significantly raised the level of available compute performance. At the same time, the growing hardware capabilities of modern supercomputing architectures have caused an increasing complexity of the parallel application development. Despite numerous efforts to improve and simplify parallel programming, there is still a lot of manual debugging and  tuning work required. This process  is supported by special software tools, facilitating debugging, performance analysis, and optimization and thus  making a major contribution to the development of  robust and efficient parallel software. This book introduces a selection of the tools, which were presented and discussed at the 6th International Parallel Tools Workshop, held in Stuttgart, Germany, 25-26 September 2012.

  6. The STAPL Parallel Graph Library

    KAUST Repository

    Harshvardhan,; Fidel, Adam; Amato, Nancy M.; Rauchwerger, Lawrence

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable

  7. Efeito do uso de enzimas sobre o desempenho e metabolismo de poedeiras Effect of the use of enzymes on the performance and metabolism of laying hens

    Directory of Open Access Journals (Sweden)

    Maurício Tárcio dos Santos Viana

    2009-06-01

    Full Text Available Objetivou-se verificar o efeito da adição do complexo enzimático (CE sobre o desempenho e o metabolismo de poedeiras. Foram utilizadas 216 poedeiras da linhagem Bovans Goldline, distribuídas em delineamento experimental inteiramente casualizado, com 12 repetições de seis aves por unidade experimental e 3 tratamentos: controle positivo; controle negativo; e controle negativo + Rovábio® Max (100 g/t. A dieta controle positivo foi formulada segundo recomendações das tabelas brasileiras e a dieta controle negativo foi calculada reduzindo os nutrientes presentes na matriz nutricional da enzima de acordo com a empresa produtora. Avaliaram-se o consumo de ração, a produção de ovos, o peso dos ovos, a massa de ovo, a conversão por massa de ovo, a conversão alimentar por dúzia (CAD e os componentes do ovo no período de 24 a 36 semanas de idade. No ensaio de metabolismo, calcularam-se os valores do coeficiente de digestibilidade da matéria seca, da energia metabolizável aparente (EMA e da energia metabolizável corrigida (EMAn, bem como seus coeficientes de metabolizabilidade e a retenção de nitrogênio e do balanço de P e Ca das dietas. Nas aves mantidas com a dieta controle negativo, a adição de complexo enzimático na dieta melhorou o percentual de postura e a conversão alimentar por dúzia. A redução dos níveis nutricionais das dietas resultou em menores valores de EMA e EMAn. A suplementação de complexo enzimático às dietas com menores níveis nutricionais melhorou os valores de EMA, resultando em valores similares aos obtidos nas aves alimentadas com a ração controle positivo. Poedeiras alimentadas com a ração controle positivo apresentam maior ingestão, excreção e retenção de fósforo (mg/ave/dia, entretanto, quando a ração é suplementada com complexo enzimático, ocorre maior retenção de fósforo pelas aves.The effect of the addition of multienzymatic complex on performance and metabolism of laying

  8. Parallel algorithms and cluster computing

    CERN Document Server

    Hoffmann, Karl Heinz

    2007-01-01

    This book presents major advances in high performance computing as well as major advances due to high performance computing. It contains a collection of papers in which results achieved in the collaboration of scientists from computer science, mathematics, physics, and mechanical engineering are presented. From the science problems to the mathematical algorithms and on to the effective implementation of these algorithms on massively parallel and cluster computers we present state-of-the-art methods and technology as well as exemplary results in these fields. This book shows that problems which seem superficially distinct become intimately connected on a computational level.

  9. Simulation of neutron transport equation using parallel Monte Carlo for deep penetration problems

    International Nuclear Information System (INIS)

    Bekar, K. K.; Tombakoglu, M.; Soekmen, C. N.

    2001-01-01

    Neutron transport equation is simulated using parallel Monte Carlo method for deep penetration neutron transport problem. Monte Carlo simulation is parallelized by using three different techniques; direct parallelization, domain decomposition and domain decomposition with load balancing, which are used with PVM (Parallel Virtual Machine) software on LAN (Local Area Network). The results of parallel simulation are given for various model problems. The performances of the parallelization techniques are compared with each other. Moreover, the effects of variance reduction techniques on parallelization are discussed

  10. Parallel Computing Strategies for Irregular Algorithms

    Science.gov (United States)

    Biswas, Rupak; Oliker, Leonid; Shan, Hongzhang; Biegel, Bryan (Technical Monitor)

    2002-01-01

    Parallel computing promises several orders of magnitude increase in our ability to solve realistic computationally-intensive problems, but relies on their efficient mapping and execution on large-scale multiprocessor architectures. Unfortunately, many important applications are irregular and dynamic in nature, making their effective parallel implementation a daunting task. Moreover, with the proliferation of parallel architectures and programming paradigms, the typical scientist is faced with a plethora of questions that must be answered in order to obtain an acceptable parallel implementation of the solution algorithm. In this paper, we consider three representative irregular applications: unstructured remeshing, sparse matrix computations, and N-body problems, and parallelize them using various popular programming paradigms on a wide spectrum of computer platforms ranging from state-of-the-art supercomputers to PC clusters. We present the underlying problems, the solution algorithms, and the parallel implementation strategies. Smart load-balancing, partitioning, and ordering techniques are used to enhance parallel performance. Overall results demonstrate the complexity of efficiently parallelizing irregular algorithms.

  11. Shared Variable Oriented Parallel Precompiler for SPMD Model

    Institute of Scientific and Technical Information of China (English)

    1995-01-01

    For the moment,commercial parallel computer systems with distributed memory architecture are usually provided with parallel FORTRAN or parallel C compliers,which are just traditional sequential FORTRAN or C compilers expanded with communication statements.Programmers suffer from writing parallel programs with communication statements. The Shared Variable Oriented Parallel Precompiler (SVOPP) proposed in this paper can automatically generate appropriate communication statements based on shared variables for SPMD(Single Program Multiple Data) computation model and greatly ease the parallel programming with high communication efficiency.The core function of parallel C precompiler has been successfully verified on a transputer-based parallel computer.Its prominent performance shows that SVOPP is probably a break-through in parallel programming technique.

  12. Parallel performance and accuracy of lattice Boltzmann and traditional finite difference methods for solving the unsteady two-dimensional Burger's equation

    Science.gov (United States)

    Velivelli, A. C.; Bryden, K. M.

    2006-03-01

    Lattice Boltzmann methods are gaining recognition in the field of computational fluid dynamics due to their computational efficiency. In order to quantify the computational efficiency and accuracy of the lattice Boltzmann method, it is compared with efficient traditional finite difference methods such as the alternating direction implicit scheme. The lattice Boltzmann algorithm implemented in previous studies does not approach peak performance for simulations where the data involved in computation per time step is more than the cache size. Due to this, data is obtained from the main memory and this access is much slower than access to cache memory. Using a cache-optimized lattice Boltzmann algorithm, this paper takes into account the full computational strength of the lattice Boltzmann method. The com parison is performed on both a single processor and multiple processors.

  13. Fast and parallel determination of PCB 77 and PCB 180 in plasma using ultra performance liquid chromatography with diode array detection: A pharmacokinetic study in Swiss albino mouse.

    Science.gov (United States)

    Ramanujam, N; Sivaselvakumar, M; Ramalingam, S

    2017-11-01

    A simple, sensitive and reproducible ultra-performance liquid chromatography (UPLC) method has been developed and validated for simultaneous estimation of polychlorinated biphenyl (PCB) 77 and PCB 180 in mouse plasma. The sample preparation was performed by simple liquid-liquid extraction technique. The analytes were chromatographed on a Waters Acquity H class UPLC system using isocratic mobile phase conditions at a flow rate of 0.3 mL/min and Acquity UPLC BEH shield RP 18 column maintained at 35°C. Quantification was performed on a photodiode array detector set at 215 nm and PCB 101 was used as internal standard (IS). PCB 77, PCB 180, and IS retention times were 2.6, 4.7 and 2.8 min, respectively, and the total run time was 6 min. The method was validated for specificity, selectivity, recovery, linearity, accuracy, precision and sample stability. The calibration curve was linear over the concentration range 10-3000 ng/mL for PCB 77 and PCB 180. Intra- and inter-day precisions for PCBs 77 and 180 were found to be good with CV <4.64%, and the accuracy ranged from 98.90 to 102.33% in mouse plasma. The validated UPLC method was successfully applied to the pharmacokinetic study of PCBs 77 and 180 in mouse plasma. Copyright © 2017 John Wiley & Sons, Ltd.

  14. Massively parallel multicanonical simulations

    Science.gov (United States)

    Gross, Jonathan; Zierenberg, Johannes; Weigel, Martin; Janke, Wolfhard

    2018-03-01

    Generalized-ensemble Monte Carlo simulations such as the multicanonical method and similar techniques are among the most efficient approaches for simulations of systems undergoing discontinuous phase transitions or with rugged free-energy landscapes. As Markov chain methods, they are inherently serial computationally. It was demonstrated recently, however, that a combination of independent simulations that communicate weight updates at variable intervals allows for the efficient utilization of parallel computational resources for multicanonical simulations. Implementing this approach for the many-thread architecture provided by current generations of graphics processing units (GPUs), we show how it can be efficiently employed with of the order of 104 parallel walkers and beyond, thus constituting a versatile tool for Monte Carlo simulations in the era of massively parallel computing. We provide the fully documented source code for the approach applied to the paradigmatic example of the two-dimensional Ising model as starting point and reference for practitioners in the field.

  15. Castor Oil Transesterification Catalysed by Liquid Enzymes

    DEFF Research Database (Denmark)

    Andrade, Thalles; Errico, Massimiliano; Christensen, Knud Villy

    2017-01-01

    In the present work, biodiesel production by reaction of non-edible castor oil with methanol under enzymatic catalysis is investigated. Two liquid enzymes were tested: Eversa Transform and Resinase HT. Reactions were performed at 35 °C and with a molar ratio of methanol to oil of 6:1. The reaction...... time was 8 hours. Stepwise addition of methanol was necessary to avoid enzyme inhibition by methanol. In order to minimize the enzyme costs, the influence of enzyme activity loss during reuse of both enzymes was evaluated under two distinct conditions. In the former, the enzymes were recovered...... and fully reused; in the latter, a mixture of 50 % reused and 50 % fresh enzymes was tested. In the case of total reuse after three cycles, both enzymes achieved only low conversions. The biodiesel content in the oil-phase using Eversa Transform was 94.21 % for the first cycle, 68.39 % in the second, and 33...

  16. Development of parallel/serial program analyzing tool

    International Nuclear Information System (INIS)

    Watanabe, Hiroshi; Nagao, Saichi; Takigawa, Yoshio; Kumakura, Toshimasa

    1999-03-01

    Japan Atomic Energy Research Institute has been developing 'KMtool', a parallel/serial program analyzing tool, in order to promote the parallelization of the science and engineering computation program. KMtool analyzes the performance of program written by FORTRAN77 and MPI, and it reduces the effort for parallelization. This paper describes development purpose, design, utilization and evaluation of KMtool. (author)

  17. Automatic Management of Parallel and Distributed System Resources

    Science.gov (United States)

    Yan, Jerry; Ngai, Tin Fook; Lundstrom, Stephen F.

    1990-01-01

    Viewgraphs on automatic management of parallel and distributed system resources are presented. Topics covered include: parallel applications; intelligent management of multiprocessing systems; performance evaluation of parallel architecture; dynamic concurrent programs; compiler-directed system approach; lattice gaseous cellular automata; and sparse matrix Cholesky factorization.

  18. Parallel Application Development Using Architecture View Driven Model Transformations

    NARCIS (Netherlands)

    Arkin, E.; Tekinerdogan, B.

    2015-01-01

    o realize the increased need for computing performance the current trend is towards applying parallel computing in which the tasks are run in parallel on multiple nodes. On its turn we can observe the rapid increase of the scale of parallel computing platforms. This situation has led to a complexity

  19. Parallel programming with Python

    CERN Document Server

    Palach, Jan

    2014-01-01

    A fast, easy-to-follow and clear tutorial to help you develop Parallel computing systems using Python. Along with explaining the fundamentals, the book will also introduce you to slightly advanced concepts and will help you in implementing these techniques in the real world. If you are an experienced Python programmer and are willing to utilize the available computing resources by parallelizing applications in a simple way, then this book is for you. You are required to have a basic knowledge of Python development to get the most of this book.

  20. Development and validation of a method for the determination of regulated fragrance allergens by High-Performance Liquid Chromatography and Parallel Factor Analysis 2.

    Science.gov (United States)

    Pérez-Outeiral, Jessica; Elcoroaristizabal, Saioa; Amigo, Jose Manuel; Vidal, Maider

    2017-12-01

    This work presents the development and validation of a multivariate method for quantitation of 6 potentially allergenic substances (PAS) related to fragrances by ultrasound-assisted emulsification microextraction coupled with HPLC-DAD and PARAFAC2 in the presence of other 18 PAS. The objective is the extension of a previously proposed univariate method to be able to determine the 24 PAS currently considered as allergens. The suitability of the multivariate approach for the qualitative and quantitative analysis of the analytes is discussed through datasets of increasing complexity, comprising the assessment and validation of the method performance. PARAFAC2 showed to adequately model the data facing up different instrumental and chemical issues, such as co-elution profiles, overlapping spectra, unknown interfering compounds, retention time shifts and baseline drifts. Satisfactory quality parameters of the model performance were obtained (R 2 ≥0.94), as well as meaningful chromatographic and spectral profiles (r≥0.97). Moreover, low errors of prediction in external validation standards (below 15% in most cases) as well as acceptable quantification errors in real spiked samples (recoveries from 82 to 119%) confirmed the suitability of PARAFAC2 for resolution and quantification of the PAS. The combination of the previously proposed univariate approach, for the well-resolved peaks, with the developed multivariate method allows the determination of the 24 regulated PAS. Copyright © 2017 Elsevier B.V. All rights reserved.

  1. Parallel processing of genomics data

    Science.gov (United States)

    Agapito, Giuseppe; Guzzi, Pietro Hiram; Cannataro, Mario

    2016-10-01

    The availability of high-throughput experimental platforms for the analysis of biological samples, such as mass spectrometry, microarrays and Next Generation Sequencing, have made possible to analyze a whole genome in a single experiment. Such platforms produce an enormous volume of data per single experiment, thus the analysis of this enormous flow of data poses several challenges in term of data storage, preprocessing, and analysis. To face those issues, efficient, possibly parallel, bioinformatics software needs to be used to preprocess and analyze data, for instance to highlight genetic variation associated with complex diseases. In this paper we present a parallel algorithm for the parallel preprocessing and statistical analysis of genomics data, able to face high dimension of data and resulting in good response time. The proposed system is able to find statistically significant biological markers able to discriminate classes of patients that respond to drugs in different ways. Experiments performed on real and synthetic genomic datasets show good speed-up and scalability.

  2. Data parallel sorting for particle simulation

    Science.gov (United States)

    Dagum, Leonardo

    1992-01-01

    Sorting on a parallel architecture is a communications intensive event which can incur a high penalty in applications where it is required. In the case of particle simulation, only integer sorting is necessary, and sequential implementations easily attain the minimum performance bound of O (N) for N particles. Parallel implementations, however, have to cope with the parallel sorting problem which, in addition to incurring a heavy communications cost, can make the minimun performance bound difficult to attain. This paper demonstrates how the sorting problem in a particle simulation can be reduced to a merging problem, and describes an efficient data parallel algorithm to solve this merging problem in a particle simulation. The new algorithm is shown to be optimal under conditions usual for particle simulation, and its fieldwise implementation on the Connection Machine is analyzed in detail. The new algorithm is about four times faster than a fieldwise implementation of radix sort on the Connection Machine.

  3. System performances of optical space code-division multiple-access-based fiber-optic two-dimensional parallel data link.

    Science.gov (United States)

    Nakamura, M; Kitayama, K

    1998-05-10

    Optical space code-division multiple access is a scheme to multiplex and link data between two-dimensional processors such as smart pixels and spatial light modulators or arrays of optical sources like vertical-cavity surface-emitting lasers. We examine the multiplexing characteristics of optical space code-division multiple access by using optical orthogonal signature patterns. The probability density function of interference noise in interfering optical orthogonal signature patterns is calculated. The bit-error rate is derived from the result and plotted as a function of receiver threshold, code length, code weight, and number of users. Furthermore, we propose a prethresholding method to suppress the interference noise, and we experimentally verify that the method works effectively in improving system performance.

  4. High performance mass spectrometry based proteomics reveals enzyme and signaling pathway regulation in neutrophils during the early stage of surgical trauma

    DEFF Research Database (Denmark)

    Arshid, Samina; Tahir, Muhammad; Fontes, Belchor

    2017-01-01

    and surgical trauma rats in this study. Extracted proteins were analyzed using nano liquid chromatography coupled to tandem mass spectrometry. A total of 2924 rat neutrophil proteins were identified in our analysis, of which 393 were found differentially regulated between control and trauma groups. By using...... functional pathways analysis of the 190 proteins up-regulated in surgical trauma we found proteins related to transcription initiation and protein biosynthesis. On the other hand, among the 203 proteins down-regulated in surgical trauma we found enrichment for proteins of the immune response, proteasome...... degradation and actin cytoskeleton. Overall, enzyme prediction analysis revealed that regulated enzymes are directly involved in neutrophil apoptosis, directional migration and chemotaxis. Our observations were then confirmed by in silico protein-protein interaction analysis. Collectively, our results reveal...

  5. Expressing Parallelism with ROOT

    Energy Technology Data Exchange (ETDEWEB)

    Piparo, D. [CERN; Tejedor, E. [CERN; Guiraud, E. [CERN; Ganis, G. [CERN; Mato, P. [CERN; Moneta, L. [CERN; Valls Pla, X. [CERN; Canal, P. [Fermilab

    2017-11-22

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  6. Expressing Parallelism with ROOT

    Science.gov (United States)

    Piparo, D.; Tejedor, E.; Guiraud, E.; Ganis, G.; Mato, P.; Moneta, L.; Valls Pla, X.; Canal, P.

    2017-10-01

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  7. Parallel Fast Legendre Transform

    NARCIS (Netherlands)

    Alves de Inda, M.; Bisseling, R.H.; Maslen, D.K.

    1998-01-01

    We discuss a parallel implementation of a fast algorithm for the discrete polynomial Legendre transform We give an introduction to the DriscollHealy algorithm using polynomial arithmetic and present experimental results on the eciency and accuracy of our implementation The algorithms were

  8. Practical parallel programming

    CERN Document Server

    Bauer, Barr E

    2014-01-01

    This is the book that will teach programmers to write faster, more efficient code for parallel processors. The reader is introduced to a vast array of procedures and paradigms on which actual coding may be based. Examples and real-life simulations using these devices are presented in C and FORTRAN.

  9. Parallel universes beguile science

    CERN Multimedia

    2007-01-01

    A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too. We may not be able -- as least not yet -- to prove they exist, many serious scientists say, but there are plenty of reasons to think that parallel dimensions are more than figments of eggheaded imagination.

  10. Parallel plate detectors

    International Nuclear Information System (INIS)

    Gardes, D.; Volkov, P.

    1981-01-01

    A 5x3cm 2 (timing only) and a 15x5cm 2 (timing and position) parallel plate avalanche counters (PPAC) are considered. The theory of operation and timing resolution is given. The measurement set-up and the curves of experimental results illustrate the possibilities of the two counters [fr

  11. Parallel hierarchical global illumination

    Energy Technology Data Exchange (ETDEWEB)

    Snell, Quinn O. [Iowa State Univ., Ames, IA (United States)

    1997-10-08

    Solving the global illumination problem is equivalent to determining the intensity of every wavelength of light in all directions at every point in a given scene. The complexity of the problem has led researchers to use approximation methods for solving the problem on serial computers. Rather than using an approximation method, such as backward ray tracing or radiosity, the authors have chosen to solve the Rendering Equation by direct simulation of light transport from the light sources. This paper presents an algorithm that solves the Rendering Equation to any desired accuracy, and can be run in parallel on distributed memory or shared memory computer systems with excellent scaling properties. It appears superior in both speed and physical correctness to recent published methods involving bidirectional ray tracing or hybrid treatments of diffuse and specular surfaces. Like progressive radiosity methods, it dynamically refines the geometry decomposition where required, but does so without the excessive storage requirements for ray histories. The algorithm, called Photon, produces a scene which converges to the global illumination solution. This amounts to a huge task for a 1997-vintage serial computer, but using the power of a parallel supercomputer significantly reduces the time required to generate a solution. Currently, Photon can be run on most parallel environments from a shared memory multiprocessor to a parallel supercomputer, as well as on clusters of heterogeneous workstations.

  12. PIXE analysis of Zn enzymes

    International Nuclear Information System (INIS)

    Solis, C.; Oliver, A.; Andrade, E.; Ruvalcaba-Sil, J.L.; Romero, I.; Celis, H.

    1999-01-01

    Zinc is a necessary component in the action and structural stability of many enzymes. Some of them are well characterized, but in others, Zn stoichiometry and its association is not known. PIXE has been proven to be a suitable technique for analyzing metallic proteins embedded in electrophoresis gels. In this study, PIXE has been used to investigate the Zn content of enzymes that are known to carry Zn atoms. These include the carbonic anhydrase, an enzyme well characterized by other methods and the cytoplasmic pyrophosphatase of Rhodospirillum rubrum that is known to require Zn to be stable but not how many metal ions are involved or how they are bound to the enzyme. Native proteins have been purified by polyacrylamide gel electrophoresis and direct identification and quantification of Zn in the gel bands was performed with an external proton beam of 3.7 MeV energy

  13. Chromatographic background drift correction coupled with parallel factor analysis to resolve coelution problems in three-dimensional chromatographic data: quantification of eleven antibiotics in tap water samples by high-performance liquid chromatography coupled with a diode array detector.

    Science.gov (United States)

    Yu, Yong-Jie; Wu, Hai-Long; Fu, Hai-Yan; Zhao, Juan; Li, Yuan-Na; Li, Shu-Fang; Kang, Chao; Yu, Ru-Qin

    2013-08-09

    Chromatographic background drift correction has been an important field of research in chromatographic analysis. In the present work, orthogonal spectral space projection for background drift correction of three-dimensional chromatographic data was described in detail and combined with parallel factor analysis (PARAFAC) to resolve overlapped chromatographic peaks and obtain the second-order advantage. This strategy was verified by simulated chromatographic data and afforded significant improvement in quantitative results. Finally, this strategy was successfully utilized to quantify eleven antibiotics in tap water samples. Compared with the traditional methodology of introducing excessive factors for the PARAFAC model to eliminate the effect of background drift, clear improvement in the quantitative performance of PARAFAC was observed after background drift correction by orthogonal spectral space projection. Copyright © 2013 Elsevier B.V. All rights reserved.

  14. Parallel changes in the onset of blood lactate accumulation (OBLA) and threshold of psychomotor performance deterioration during incremental exercise after training in athletes.

    Science.gov (United States)

    Chmura, Jan; Nazar, Krystyna

    2010-03-01

    During aerobic exercise with increasing intensities choice reaction time (CRT) progressively shortens up to 60-80% of maximal workload, and then it rapidly increases. The aim of this study was to determine whether workload associated with the shortest CRT operationally called "the psychomotor fatigue threshold" is related to the metabolic response to exercise. Thirteen male soccer players (aged 23.3 + or - 1.0 yrs) participated in this study. Before and after 6 weeks of training in the pre-competition period they underwent treadmill test at 0 grade with running speed increasing every 3 min by 2 km/h starting from 6 km/h until exhaustion. At each stage of exercise CRT, heart rate, respiratory gas exchange and blood lactate [LA] were measured and the workload corresponding to [LA] of 4 mmol/l (OBLA) was recorded. After training, CRT was significantly shortened at rest (from m + or - SEM = 345 + or - 12 to 317 + or - 12 ms) and during exercise (from 304 + or - 10 to 285 + or - 11 ms at the psychomotor fatigue threshold and from 359 + or - 13 to 331 + or - 13 ms, pchanges in OBLA occurring during training and those in psychomotor fatigue threshold (r = 0.88). It is concluded that endurance training not only increases exercise tolerance due to its influence on metabolism but also facilitates psychomotor performance during heavy exercise. Copyright 2010 Elsevier B.V. All rights reserved.

  15. Parallel processing of structural integrity analysis codes

    International Nuclear Information System (INIS)

    Swami Prasad, P.; Dutta, B.K.; Kushwaha, H.S.

    1996-01-01

    Structural integrity analysis forms an important role in assessing and demonstrating the safety of nuclear reactor components. This analysis is performed using analytical tools such as Finite Element Method (FEM) with the help of digital computers. The complexity of the problems involved in nuclear engineering demands high speed computation facilities to obtain solutions in reasonable amount of time. Parallel processing systems such as ANUPAM provide an efficient platform for realising the high speed computation. The development and implementation of software on parallel processing systems is an interesting and challenging task. The data and algorithm structure of the codes plays an important role in exploiting the parallel processing system capabilities. Structural analysis codes based on FEM can be divided into two categories with respect to their implementation on parallel processing systems. The first category codes such as those used for harmonic analysis, mechanistic fuel performance codes need not require the parallelisation of individual modules of the codes. The second category of codes such as conventional FEM codes require parallelisation of individual modules. In this category, parallelisation of equation solution module poses major difficulties. Different solution schemes such as domain decomposition method (DDM), parallel active column solver and substructuring method are currently used on parallel processing systems. Two codes, FAIR and TABS belonging to each of these categories have been implemented on ANUPAM. The implementation details of these codes and the performance of different equation solvers are highlighted. (author). 5 refs., 12 figs., 1 tab

  16. [Advances on enzymes and enzyme inhibitors research based on microfluidic devices].

    Science.gov (United States)

    Hou, Feng-Hua; Ye, Jian-Qing; Chen, Zuan-Guang; Cheng, Zhi-Yi

    2010-06-01

    With the continuous development in microfluidic fabrication technology, microfluidic analysis has evolved from a concept to one of research frontiers in last twenty years. The research of enzymes and enzyme inhibitors based on microfluidic devices has also made great progress. Microfluidic technology improved greatly the analytical performance of the research of enzymes and enzyme inhibitors by reducing the consumption of reagents, decreasing the analysis time, and developing automation. This review focuses on the development and classification of enzymes and enzyme inhibitors research based on microfluidic devices.

  17. Parallel computation of rotating flows

    DEFF Research Database (Denmark)

    Lundin, Lars Kristian; Barker, Vincent A.; Sørensen, Jens Nørkær

    1999-01-01

    This paper deals with the simulation of 3‐D rotating flows based on the velocity‐vorticity formulation of the Navier‐Stokes equations in cylindrical coordinates. The governing equations are discretized by a finite difference method. The solution is advanced to a new time level by a two‐step process...... is that of solving a singular, large, sparse, over‐determined linear system of equations, and the iterative method CGLS is applied for this purpose. We discuss some of the mathematical and numerical aspects of this procedure and report on the performance of our software on a wide range of parallel computers. Darbe...

  18. Substrate mediated enzyme prodrug therapy

    DEFF Research Database (Denmark)

    Fejerskov, Betina; Jarlstad Olesen, Morten T; Zelikin, Alexander N

    2017-01-01

    Substrate mediated enzyme prodrug therapy (SMEPT) is a biomedical platform developed to perform a localized synthesis of drugs mediated by implantable biomaterials. This approach combines the benefits and at the same time offers to overcome the drawbacks for traditional pill-based drug administra......Substrate mediated enzyme prodrug therapy (SMEPT) is a biomedical platform developed to perform a localized synthesis of drugs mediated by implantable biomaterials. This approach combines the benefits and at the same time offers to overcome the drawbacks for traditional pill-based drug...

  19. Parallel grid population

    Science.gov (United States)

    Wald, Ingo; Ize, Santiago

    2015-07-28

    Parallel population of a grid with a plurality of objects using a plurality of processors. One example embodiment is a method for parallel population of a grid with a plurality of objects using a plurality of processors. The method includes a first act of dividing a grid into n distinct grid portions, where n is the number of processors available for populating the grid. The method also includes acts of dividing a plurality of objects into n distinct sets of objects, assigning a distinct set of objects to each processor such that each processor determines by which distinct grid portion(s) each object in its distinct set of objects is at least partially bounded, and assigning a distinct grid portion to each processor such that each processor populates its distinct grid portion with any objects that were previously determined to be at least partially bounded by its distinct grid portion.

  20. More parallel please

    DEFF Research Database (Denmark)

    Gregersen, Frans; Josephson, Olle; Kristoffersen, Gjert

    of departure that English may be used in parallel with the various local, in this case Nordic, languages. As such, the book integrates the challenge of internationalization faced by any university with the wish to improve quality in research, education and administration based on the local language......Abstract [en] More parallel, please is the result of the work of an Inter-Nordic group of experts on language policy financed by the Nordic Council of Ministers 2014-17. The book presents all that is needed to plan, practice and revise a university language policy which takes as its point......(s). There are three layers in the text: First, you may read the extremely brief version of the in total 11 recommendations for best practice. Second, you may acquaint yourself with the extended version of the recommendations and finally, you may study the reasoning behind each of them. At the end of the text, we give...

  1. PARALLEL MOVING MECHANICAL SYSTEMS

    Directory of Open Access Journals (Sweden)

    Florian Ion Tiberius Petrescu

    2014-09-01

    Full Text Available Normal 0 false false false EN-US X-NONE X-NONE MicrosoftInternetExplorer4 Moving mechanical systems parallel structures are solid, fast, and accurate. Between parallel systems it is to be noticed Stewart platforms, as the oldest systems, fast, solid and precise. The work outlines a few main elements of Stewart platforms. Begin with the geometry platform, kinematic elements of it, and presented then and a few items of dynamics. Dynamic primary element on it means the determination mechanism kinetic energy of the entire Stewart platforms. It is then in a record tail cinematic mobile by a method dot matrix of rotation. If a structural mottoelement consists of two moving elements which translates relative, drive train and especially dynamic it is more convenient to represent the mottoelement as a single moving components. We have thus seven moving parts (the six motoelements or feet to which is added mobile platform 7 and one fixed.

  2. Xyce parallel electronic simulator.

    Energy Technology Data Exchange (ETDEWEB)

    Keiter, Eric R; Mei, Ting; Russo, Thomas V.; Rankin, Eric Lamont; Schiek, Richard Louis; Thornquist, Heidi K.; Fixel, Deborah A.; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R.

    2010-05-01

    This document is a reference guide to the Xyce Parallel Electronic Simulator, and is a companion document to the Xyce Users Guide. The focus of this document is (to the extent possible) exhaustively list device parameters, solver options, parser options, and other usage details of Xyce. This document is not intended to be a tutorial. Users who are new to circuit simulation are better served by the Xyce Users Guide.

  3. Stability of parallel flows

    CERN Document Server

    Betchov, R

    2012-01-01

    Stability of Parallel Flows provides information pertinent to hydrodynamical stability. This book explores the stability problems that occur in various fields, including electronics, mechanics, oceanography, administration, economics, as well as naval and aeronautical engineering. Organized into two parts encompassing 10 chapters, this book starts with an overview of the general equations of a two-dimensional incompressible flow. This text then explores the stability of a laminar boundary layer and presents the equation of the inviscid approximation. Other chapters present the general equation

  4. Algorithmically specialized parallel computers

    CERN Document Server

    Snyder, Lawrence; Gannon, Dennis B

    1985-01-01

    Algorithmically Specialized Parallel Computers focuses on the concept and characteristics of an algorithmically specialized computer.This book discusses the algorithmically specialized computers, algorithmic specialization using VLSI, and innovative architectures. The architectures and algorithms for digital signal, speech, and image processing and specialized architectures for numerical computations are also elaborated. Other topics include the model for analyzing generalized inter-processor, pipelined architecture for search tree maintenance, and specialized computer organization for raster

  5. Adapting algorithms to massively parallel hardware

    CERN Document Server

    Sioulas, Panagiotis

    2016-01-01

    In the recent years, the trend in computing has shifted from delivering processors with faster clock speeds to increasing the number of cores per processor. This marks a paradigm shift towards parallel programming in which applications are programmed to exploit the power provided by multi-cores. Usually there is gain in terms of the time-to-solution and the memory footprint. Specifically, this trend has sparked an interest towards massively parallel systems that can provide a large number of processors, and possibly computing nodes, as in the GPUs and MPPAs (Massively Parallel Processor Arrays). In this project, the focus was on two distinct computing problems: k-d tree searches and track seeding cellular automata. The goal was to adapt the algorithms to parallel systems and evaluate their performance in different cases.

  6. Implementing Shared Memory Parallelism in MCBEND

    Directory of Open Access Journals (Sweden)

    Bird Adam

    2017-01-01

    Full Text Available MCBEND is a general purpose radiation transport Monte Carlo code from AMEC Foster Wheelers’s ANSWERS® Software Service. MCBEND is well established in the UK shielding community for radiation shielding and dosimetry assessments. The existing MCBEND parallel capability effectively involves running the same calculation on many processors. This works very well except when the memory requirements of a model restrict the number of instances of a calculation that will fit on a machine. To more effectively utilise parallel hardware OpenMP has been used to implement shared memory parallelism in MCBEND. This paper describes the reasoning behind the choice of OpenMP, notes some of the challenges of multi-threading an established code such as MCBEND and assesses the performance of the parallel method implemented in MCBEND.

  7. The effect of aerobic exercise training on growth performance, digestive enzyme activities and postprandial metabolic response in juvenile qingbo (Spinibarbus sinensis).

    Science.gov (United States)

    Li, Xiu-Ming; Yu, Li-Juan; Wang, Chuan; Zeng, Ling-Qing; Cao, Zhen-Dong; Fu, Shi-Jian; Zhang, Yao-Guang

    2013-09-01

    Continual swimming exercise usually promotes growth in fish at a moderate water velocity. We hypothesized that the improvement in growth in exercise-trained fish may be accompanied by increases in digestive enzyme activity, respiratory capacity and, hence, postprandial metabolism. Juvenile qingbo fish (Spinibarbus sinensis) were subjected to aerobic training for 8weeks at a water velocity of control (3cms(-1)), 1, 2 and 4 body length (bl)s(-1) at a constant temperature of 25°C. The feed intake (FI), food conversion rate (FCR), specific growth rate (SGR), whole-body composition, trypsin and lipase activities, maximal oxygen consumption (M˙O2max) and postprandial M˙O2 response were measured at the end of the training period. Aerobic exercise training induced a significant increase in FI compared with the control group, while the FCR of the 4bls(-1) group was significantly lower than for the other three groups (PFI after long-term training; (3) and aerobic exercise training boosted the activity of digestive enzymes and maximum digestive metabolism, which could favor fast digestion and growth in juvenile S. sinensis. Copyright © 2013 Elsevier Inc. All rights reserved.

  8. Parallelization of an existing high energy physics event reconstruction software package

    International Nuclear Information System (INIS)

    Schiefer, R.; Francis, D.

    1996-01-01

    Software parallelization allows an efficient use of available computing power to increase the performance of applications. In a case study the authors have investigated the parallelization of high energy physics event reconstruction software in terms of costs (effort, computing resource requirements), benefits (performance increase) and the feasibility of a systematic parallelization approach. Guidelines facilitating a parallel implementation are proposed for future software development

  9. Advances in randomized parallel computing

    CERN Document Server

    Rajasekaran, Sanguthevar

    1999-01-01

    The technique of randomization has been employed to solve numerous prob­ lems of computing both sequentially and in parallel. Examples of randomized algorithms that are asymptotically better than their deterministic counterparts in solving various fundamental problems abound. Randomized algorithms have the advantages of simplicity and better performance both in theory and often in practice. This book is a collection of articles written by renowned experts in the area of randomized parallel computing. A brief introduction to randomized algorithms In the aflalysis of algorithms, at least three different measures of performance can be used: the best case, the worst case, and the average case. Often, the average case run time of an algorithm is much smaller than the worst case. 2 For instance, the worst case run time of Hoare's quicksort is O(n ), whereas its average case run time is only O( n log n). The average case analysis is conducted with an assumption on the input space. The assumption made to arrive at t...

  10. Improvement of Parallel Algorithm for MATRA Code

    International Nuclear Information System (INIS)

    Kim, Seong-Jin; Seo, Kyong-Won; Kwon, Hyouk; Hwang, Dae-Hyun

    2014-01-01

    The feasibility study to parallelize the MATRA code was conducted in KAERI early this year. As a result, a parallel algorithm for the MATRA code has been developed to decrease a considerably required computing time to solve a bigsize problem such as a whole core pin-by-pin problem of a general PWR reactor and to improve an overall performance of the multi-physics coupling calculations. It was shown that the performance of the MATRA code was greatly improved by implementing the parallel algorithm using MPI communication. For problems of a 1/8 core and whole core for SMART reactor, a speedup was evaluated as about 10 when the numbers of used processor were 25. However, it was also shown that the performance deteriorated as the axial node number increased. In this paper, the procedure of a communication between processors is optimized to improve the previous parallel algorithm.. To improve the performance deterioration of the parallelized MATRA code, the communication algorithm between processors was newly presented. It was shown that the speedup was improved and stable regardless of the axial node number

  11. 21 CFR 864.9400 - Stabilized enzyme solution.

    Science.gov (United States)

    2010-04-01

    ... 21 Food and Drugs 8 2010-04-01 2010-04-01 false Stabilized enzyme solution. 864.9400 Section 864... and Blood Products § 864.9400 Stabilized enzyme solution. (a) Identification. A stabilized enzyme... enzyme solutions include papain, bromelin, ficin, and trypsin. (b) Classification. Class II (performance...

  12. Towards a streaming model for nested data parallelism

    DEFF Research Database (Denmark)

    Madsen, Frederik Meisner; Filinski, Andrzej

    2013-01-01

    The language-integrated cost semantics for nested data parallelism pioneered by NESL provides an intuitive, high-level model for predicting performance and scalability of parallel algorithms with reasonable accuracy. However, this predictability, obtained through a uniform, parallelism-flattening......The language-integrated cost semantics for nested data parallelism pioneered by NESL provides an intuitive, high-level model for predicting performance and scalability of parallel algorithms with reasonable accuracy. However, this predictability, obtained through a uniform, parallelism......-processable in a streaming fashion. This semantics is directly compatible with previously proposed piecewise execution models for nested data parallelism, but allows the expected space usage to be reasoned about directly at the source-language level. The language definition and implementation are still very much work...

  13. An Algorithm for Parallel Sn Sweeps on Unstructured Meshes

    International Nuclear Information System (INIS)

    Pautz, Shawn D.

    2002-01-01

    A new algorithm for performing parallel S n sweeps on unstructured meshes is developed. The algorithm uses a low-complexity list ordering heuristic to determine a sweep ordering on any partitioned mesh. For typical problems and with 'normal' mesh partitionings, nearly linear speedups on up to 126 processors are observed. This is an important and desirable result, since although analyses of structured meshes indicate that parallel sweeps will not scale with normal partitioning approaches, no severe asymptotic degradation in the parallel efficiency is observed with modest (≤100) levels of parallelism. This result is a fundamental step in the development of efficient parallel S n methods

  14. Embodied and Distributed Parallel DJing.

    Science.gov (United States)

    Cappelen, Birgitta; Andersson, Anders-Petter

    2016-01-01

    Everyone has a right to take part in cultural events and activities, such as music performances and music making. Enforcing that right, within Universal Design, is often limited to a focus on physical access to public areas, hearing aids etc., or groups of persons with special needs performing in traditional ways. The latter might be people with disabilities, being musicians playing traditional instruments, or actors playing theatre. In this paper we focus on the innovative potential of including people with special needs, when creating new cultural activities. In our project RHYME our goal was to create health promoting activities for children with severe disabilities, by developing new musical and multimedia technologies. Because of the users' extreme demands and rich contribution, we ended up creating both a new genre of musical instruments and a new art form. We call this new art form Embodied and Distributed Parallel DJing, and the new genre of instruments for Empowering Multi-Sensorial Things.

  15. Evaluation of the chiral recognition properties as well as the column performance of four chiral stationary phases based on cellulose (3,5-dimethylphenylcarbamate) by parallel HPLC and SFC.

    Science.gov (United States)

    Nelander, Hanna; Andersson, Shalini; Ohlén, Kristina

    2011-12-30

    The performance of four commercially available cellulose tris(3,5-dimethylphenylcarbamate) based chiral stationary phases (CSPs) was evaluated with parallel high performance liquid chromatography (HPLC) and super critical fluid chromatography (SFC). Retention, enantioselectivity, resolution and efficiency were compared for a set of neutral, basic and acidic compounds having different physico-chemical properties by using different mobile phase conditions. Although the chiral selector is the same in all the four CSPs, a large difference in the ability to retain and resolve enantiomers was observed under the same chromatographic conditions. We believe that this is mainly due to differences in the silica matrix and immobilization techniques used by the different vendors. An extended study of metoprolol and structure analogues gave a deeper understanding of the accessibility of the chiral discriminating interactions and its impact on the resolution of the racemic compounds on the four CSPs studied. Also, a clear difference in enantioselectivity is observed between SFC and LC mode, hydrogen bonding was found to play an important role in the differential binding of the enantiomers to the CSPs. Copyright © 2011 Elsevier B.V. All rights reserved.

  16. Enzyme inhibition by iminosugars

    DEFF Research Database (Denmark)

    López, Óscar; Qing, Feng-Ling; Pedersen, Christian Marcus

    2013-01-01

    Imino- and azasugar glycosidase inhibitors display pH dependant inhibition reflecting that both the inhibitor and the enzyme active site have groups that change protonation state with pH. With the enzyme having two acidic groups and the inhibitor one basic group, enzyme-inhibitor complexes...

  17. Adsorption and enzyme activity of asparaginase at lipid Langmuir and Langmuir-Blodgett films

    International Nuclear Information System (INIS)

    Rocha Junior, Carlos da; Caseli, Luciano

    2017-01-01

    In this present work, the surface activity of the enzyme asparaginase was investigated at the air-water interface, presenting surface activity in high ionic strengths. Asparaginase was incorporated in Langmuir monolayers of the phospholipid dipalmitoylphosphatidylcholine (DPPC), forming a mixed film, which was characterized with surface pressure-area isotherms, surface potential-area isotherms, polarization-modulated infrared reflection-absorption spectroscopy (PM-IRRAS), and Brewster angle microscopy (BAM). The adsorption of the enzyme at the air-water interface condensed the lipid monolayer and increased the film compressibility at high surface pressures. Amide bands in the PM-IRRAS spectra were identified, with the C−N and C =O dipole moments lying parallel to monolayer plane, revealing the structuring of the enzyme into α-helices and β-sheets. The floating monolayers were transferred to solid supports as Langmuir-Blodgett (LB) films and characterized with fluorescence spectroscopy and atomic force microscopy. Catalytic activities of the films were measured and compared to the homogenous medium. The enzyme accommodated in the LB films preserved more than 78% of the enzyme activity after 30 days, in contrast for the homogeneous medium, which preserved less than 13%. The method presented in this work not only allows for an enhanced catalytic activity, but also can help explain why certain film architectures exhibit better performance. - Highlights: • Biomembranes are mimicked with Langmuir monolayers. • Asparaginase is incorporated into the lipid monolayer. • Enzyme adsorption is confirmed with tensiometry and infrared spectroscopy. • Langmuir-Blodgett films of the enzyme present enzyme activity.

  18. Adsorption and enzyme activity of asparaginase at lipid Langmuir and Langmuir-Blodgett films

    Energy Technology Data Exchange (ETDEWEB)

    Rocha Junior, Carlos da; Caseli, Luciano, E-mail: lcaseli@unifesp.br

    2017-04-01

    In this present work, the surface activity of the enzyme asparaginase was investigated at the air-water interface, presenting surface activity in high ionic strengths. Asparaginase was incorporated in Langmuir monolayers of the phospholipid dipalmitoylphosphatidylcholine (DPPC), forming a mixed film, which was characterized with surface pressure-area isotherms, surface potential-area isotherms, polarization-modulated infrared reflection-absorption spectroscopy (PM-IRRAS), and Brewster angle microscopy (BAM). The adsorption of the enzyme at the air-water interface condensed the lipid monolayer and increased the film compressibility at high surface pressures. Amide bands in the PM-IRRAS spectra were identified, with the C−N and C =O dipole moments lying parallel to monolayer plane, revealing the structuring of the enzyme into α-helices and β-sheets. The floating monolayers were transferred to solid supports as Langmuir-Blodgett (LB) films and characterized with fluorescence spectroscopy and atomic force microscopy. Catalytic activities of the films were measured and compared to the homogenous medium. The enzyme accommodated in the LB films preserved more than 78% of the enzyme activity after 30 days, in contrast for the homogeneous medium, which preserved less than 13%. The method presented in this work not only allows for an enhanced catalytic activity, but also can help explain why certain film architectures exhibit better performance. - Highlights: • Biomembranes are mimicked with Langmuir monolayers. • Asparaginase is incorporated into the lipid monolayer. • Enzyme adsorption is confirmed with tensiometry and infrared spectroscopy. • Langmuir-Blodgett films of the enzyme present enzyme activity.

  19. Resistor Combinations for Parallel Circuits.

    Science.gov (United States)

    McTernan, James P.

    1978-01-01

    To help simplify both teaching and learning of parallel circuits, a high school electricity/electronics teacher presents and illustrates the use of tables of values for parallel resistive circuits in which total resistances are whole numbers. (MF)

  20. SOFTWARE FOR DESIGNING PARALLEL APPLICATIONS

    Directory of Open Access Journals (Sweden)

    M. K. Bouza

    2017-01-01

    Full Text Available The object of research is the tools to support the development of parallel programs in C/C ++. The methods and software which automates the process of designing parallel applications are proposed.

  1. Parallelization of Subchannel Analysis Code MATRA

    International Nuclear Information System (INIS)

    Kim, Seongjin; Hwang, Daehyun; Kwon, Hyouk

    2014-01-01

    A stand-alone calculation of MATRA code used up pertinent computing time for the thermal margin calculations while a relatively considerable time is needed to solve the whole core pin-by-pin problems. In addition, it is strongly required to improve the computation speed of the MATRA code to satisfy the overall performance of the multi-physics coupling calculations. Therefore, a parallel approach to improve and optimize the computability of the MATRA code is proposed and verified in this study. The parallel algorithm is embodied in the MATRA code using the MPI communication method and the modification of the previous code structure was minimized. An improvement is confirmed by comparing the results between the single and multiple processor algorithms. The speedup and efficiency are also evaluated when increasing the number of processors. The parallel algorithm was implemented to the subchannel code MATRA using the MPI. The performance of the parallel algorithm was verified by comparing the results with those from the MATRA with the single processor. It is also noticed that the performance of the MATRA code was greatly improved by implementing the parallel algorithm for the 1/8 core and whole core problems

  2. Parallel External Memory Graph Algorithms

    DEFF Research Database (Denmark)

    Arge, Lars Allan; Goodrich, Michael T.; Sitchinava, Nodari

    2010-01-01

    In this paper, we study parallel I/O efficient graph algorithms in the Parallel External Memory (PEM) model, one o f the private-cache chip multiprocessor (CMP) models. We study the fundamental problem of list ranking which leads to efficient solutions to problems on trees, such as computing lowest...... an optimal speedup of ¿(P) in parallel I/O complexity and parallel computation time, compared to the single-processor external memory counterparts....

  3. Performance of four different indirect enzyme-linked immunosorbent assays (ELISAs) to detect specific IgG, IgA, and IgM in Legionnaires' disease

    DEFF Research Database (Denmark)

    Bangsborg, Jette Marie; Shand, G H; Hansen, K

    1994-01-01

    Currently recommended methods in Legionnaires' disease serology are based upon crude whole-cell antigenic preparations. To investigate whether purified antigens would perform better in a given diagnostic test for antibodies against Legionella pneumophila, we compared the performance of three...

  4. Parallel inter channel interaction mechanisms

    International Nuclear Information System (INIS)

    Jovic, V.; Afgan, N.; Jovic, L.

    1995-01-01

    Parallel channels interactions are examined. For experimental researches of nonstationary regimes flow in three parallel vertical channels results of phenomenon analysis and mechanisms of parallel channel interaction for adiabatic condition of one-phase fluid and two-phase mixture flow are shown. (author)

  5. Parallel pic plasma simulation through particle decomposition techniques

    International Nuclear Information System (INIS)

    Briguglio, S.; Vlad, G.; Di Martino, B.; Naples, Univ. 'Federico II'

    1998-02-01

    Particle-in-cell (PIC) codes are among the major candidates to yield a satisfactory description of the detail of kinetic effects, such as the resonant wave-particle interaction, relevant in determining the transport mechanism in magnetically confined plasmas. A significant improvement of the simulation performance of such codes con be expected from parallelization, e.g., by distributing the particle population among several parallel processors. Parallelization of a hybrid magnetohydrodynamic-gyrokinetic code has been accomplished within the High Performance Fortran (HPF) framework, and tested on the IBM SP2 parallel system, using a 'particle decomposition' technique. The adopted technique requires a moderate effort in porting the code in parallel form and results in intrinsic load balancing and modest inter processor communication. The performance tests obtained confirm the hypothesis of high effectiveness of the strategy, if targeted towards moderately parallel architectures. Optimal use of resources is also discussed with reference to a specific physics problem [it

  6. A Parallel Butterfly Algorithm

    KAUST Repository

    Poulson, Jack; Demanet, Laurent; Maxwell, Nicholas; Ying, Lexing

    2014-01-01

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  7. A Parallel Butterfly Algorithm

    KAUST Repository

    Poulson, Jack

    2014-02-04

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  8. Fast parallel event reconstruction

    CERN Multimedia

    CERN. Geneva

    2010-01-01

    On-line processing of large data volumes produced in modern HEP experiments requires using maximum capabilities of modern and future many-core CPU and GPU architectures.One of such powerful feature is a SIMD instruction set, which allows packing several data items in one register and to operate on all of them, thus achievingmore operations per clock cycle. Motivated by the idea of using the SIMD unit ofmodern processors, the KF based track fit has been adapted for parallelism, including memory optimization, numerical analysis, vectorization with inline operator overloading, and optimization using SDKs. The speed of the algorithm has been increased in 120000 times with 0.1 ms/track, running in parallel on 16 SPEs of a Cell Blade computer.  Running on a Nehalem CPU with 8 cores it shows the processing speed of 52 ns/track using the Intel Threading Building Blocks. The same KF algorithm running on an Nvidia GTX 280 in the CUDA frameworkprovi...

  9. Parallel computing in enterprise modeling.

    Energy Technology Data Exchange (ETDEWEB)

    Goldsby, Michael E.; Armstrong, Robert C.; Shneider, Max S.; Vanderveen, Keith; Ray, Jaideep; Heath, Zach; Allan, Benjamin A.

    2008-08-01

    This report presents the results of our efforts to apply high-performance computing to entity-based simulations with a multi-use plugin for parallel computing. We use the term 'Entity-based simulation' to describe a class of simulation which includes both discrete event simulation and agent based simulation. What simulations of this class share, and what differs from more traditional models, is that the result sought is emergent from a large number of contributing entities. Logistic, economic and social simulations are members of this class where things or people are organized or self-organize to produce a solution. Entity-based problems never have an a priori ergodic principle that will greatly simplify calculations. Because the results of entity-based simulations can only be realized at scale, scalable computing is de rigueur for large problems. Having said that, the absence of a spatial organizing principal makes the decomposition of the problem onto processors problematic. In addition, practitioners in this domain commonly use the Java programming language which presents its own problems in a high-performance setting. The plugin we have developed, called the Parallel Particle Data Model, overcomes both of these obstacles and is now being used by two Sandia frameworks: the Decision Analysis Center, and the Seldon social simulation facility. While the ability to engage U.S.-sized problems is now available to the Decision Analysis Center, this plugin is central to the success of Seldon. Because Seldon relies on computationally intensive cognitive sub-models, this work is necessary to achieve the scale necessary for realistic results. With the recent upheavals in the financial markets, and the inscrutability of terrorist activity, this simulation domain will likely need a capability with ever greater fidelity. High-performance computing will play an important part in enabling that greater fidelity.

  10. Parallel visualization on leadership computing resources

    Energy Technology Data Exchange (ETDEWEB)

    Peterka, T; Ross, R B [Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, IL 60439 (United States); Shen, H-W [Department of Computer Science and Engineering, Ohio State University, Columbus, OH 43210 (United States); Ma, K-L [Department of Computer Science, University of California at Davis, Davis, CA 95616 (United States); Kendall, W [Department of Electrical Engineering and Computer Science, University of Tennessee at Knoxville, Knoxville, TN 37996 (United States); Yu, H, E-mail: tpeterka@mcs.anl.go [Sandia National Laboratories, California, Livermore, CA 94551 (United States)

    2009-07-01

    Changes are needed in the way that visualization is performed, if we expect the analysis of scientific data to be effective at the petascale and beyond. By using similar techniques as those used to parallelize simulations, such as parallel I/O, load balancing, and effective use of interprocess communication, the supercomputers that compute these datasets can also serve as analysis and visualization engines for them. Our team is assessing the feasibility of performing parallel scientific visualization on some of the most powerful computational resources of the U.S. Department of Energy's National Laboratories in order to pave the way for analyzing the next generation of computational results. This paper highlights some of the conclusions of that research.

  11. Java parallel secure stream for grid computing

    International Nuclear Information System (INIS)

    Chen, J.; Akers, W.; Chen, Y.; Watson, W.

    2001-01-01

    The emergence of high speed wide area networks makes grid computing a reality. However grid applications that need reliable data transfer still have difficulties to achieve optimal TCP performance due to network tuning of TCP window size to improve the bandwidth and to reduce latency on a high speed wide area network. The authors present a pure Java package called JPARSS (Java Parallel Secure Stream) that divides data into partitions that are sent over several parallel Java streams simultaneously and allows Java or Web applications to achieve optimal TCP performance in a gird environment without the necessity of tuning the TCP window size. Several experimental results are provided to show that using parallel stream is more effective than tuning TCP window size. In addition X.509 certificate based single sign-on mechanism and SSL based connection establishment are integrated into this package. Finally a few applications using this package will be discussed

  12. Abstract Level Parallelization of Finite Difference Methods

    Directory of Open Access Journals (Sweden)

    Edwin Vollebregt

    1997-01-01

    Full Text Available A formalism is proposed for describing finite difference calculations in an abstract way. The formalism consists of index sets and stencils, for characterizing the structure of sets of data items and interactions between data items (“neighbouring relations”. The formalism provides a means for lifting programming to a more abstract level. This simplifies the tasks of performance analysis and verification of correctness, and opens the way for automaticcode generation. The notation is particularly useful in parallelization, for the systematic construction of parallel programs in a process/channel programming paradigm (e.g., message passing. This is important because message passing, unfortunately, still is the only approach that leads to acceptable performance for many more unstructured or irregular problems on parallel computers that have non-uniform memory access times. It will be shown that the use of index sets and stencils greatly simplifies the determination of which data must be exchanged between different computing processes.

  13. Parallel visualization on leadership computing resources

    International Nuclear Information System (INIS)

    Peterka, T; Ross, R B; Shen, H-W; Ma, K-L; Kendall, W; Yu, H

    2009-01-01

    Changes are needed in the way that visualization is performed, if we expect the analysis of scientific data to be effective at the petascale and beyond. By using similar techniques as those used to parallelize simulations, such as parallel I/O, load balancing, and effective use of interprocess communication, the supercomputers that compute these datasets can also serve as analysis and visualization engines for them. Our team is assessing the feasibility of performing parallel scientific visualization on some of the most powerful computational resources of the U.S. Department of Energy's National Laboratories in order to pave the way for analyzing the next generation of computational results. This paper highlights some of the conclusions of that research.

  14. Limiting Concentrate during Growing Period Affect Performance and Gene Expression of Hepatic Gluconeogenic Enzymes and Visfatin in Korean Native Beef Calves

    Directory of Open Access Journals (Sweden)

    S. S. Chang

    2013-02-01

    Full Text Available This study elucidated the effects of limited concentrate feeding on growth, plasma profile, and gene expression of gluconeogenic enzymes and visfatin in the liver of Hanwoo beef calves. The purpose of this study was to test that reducing the amount of concentrate would partially be compensated by increasing the intake of forage and by altering the metabolic status. The study utilized 20 Korean native beef calves (Hanwoo; 60 to 70 d of age divided into two groups of 10 calves each for 158 d. Control group calves received the amount of concentrate as per the established Korean feeding standards for Hanwoo, whereas calves in the restricted group only received half the amount of concentrate as per standard requirements. Good quality forage (Timothy hay was available for ad libitum consumption to both groups. Since calves were with their dam until 4 months of age in breeding pens before weaning, the intake of milk before weaning was not recorded, however, the concentrate and forage intakes were recorded daily. Body weights (BW were recorded at start and on 10 d interval. Blood samples were collected at start and at 50 d interval. On the final day of the experiment, liver biopsies were collected from all animals in each group. The BW was not different between the groups at all times, but tended to be higher (p = 0.061 only at final BW in control than restricted group. Total BW gain in the control group was 116.2 kg as opposed to 84.1 kg in restricted group that led to average BW gain of 736 g/d and 532 g/d in respective groups, and the differences were significant (p<0.01. As planned, the calves in the control group had higher concentrate and lower forage intake than the restricted group. The plasma variables like total protein and urea were higher (p<0.05 in control than restricted group. The mRNA expressions for the gluconeogenic enzymes such as cytosolic phosphoenol pyruvate carboxykinase (EC 4.1.1.32 and pyruvate carboxylase (EC 6.4.1.1, and

  15. Limiting Concentrate during Growing Period Affect Performance and Gene Expression of Hepatic Gluconeogenic Enzymes and Visfatin in Korean Native Beef Calves.

    Science.gov (United States)

    Chang, S S; Lohakare, J D; Singh, N K; Kwon, E G; Nejad, J G; Sung, K I; Hong, S K

    2013-02-01

    This study elucidated the effects of limited concentrate feeding on growth, plasma profile, and gene expression of gluconeogenic enzymes and visfatin in the liver of Hanwoo beef calves. The purpose of this study was to test that reducing the amount of concentrate would partially be compensated by increasing the intake of forage and by altering the metabolic status. The study utilized 20 Korean native beef calves (Hanwoo; 60 to 70 d of age) divided into two groups of 10 calves each for 158 d. Control group calves received the amount of concentrate as per the established Korean feeding standards for Hanwoo, whereas calves in the restricted group only received half the amount of concentrate as per standard requirements. Good quality forage (Timothy hay) was available for ad libitum consumption to both groups. Since calves were with their dam until 4 months of age in breeding pens before weaning, the intake of milk before weaning was not recorded, however, the concentrate and forage intakes were recorded daily. Body weights (BW) were recorded at start and on 10 d interval. Blood samples were collected at start and at 50 d interval. On the final day of the experiment, liver biopsies were collected from all animals in each group. The BW was not different between the groups at all times, but tended to be higher (p = 0.061) only at final BW in control than restricted group. Total BW gain in the control group was 116.2 kg as opposed to 84.1 kg in restricted group that led to average BW gain of 736 g/d and 532 g/d in respective groups, and the differences were significant (pforage intake than the restricted group. The plasma variables like total protein and urea were higher (p<0.05) in control than restricted group. The mRNA expressions for the gluconeogenic enzymes such as cytosolic phosphoenol pyruvate carboxykinase (EC 4.1.1.32) and pyruvate carboxylase (EC 6.4.1.1), and visfatin measured by quantitative real-time PCR in liver biopsies showed higher expression (p<0.05) in

  16. Design strategies for irregularly adapting parallel applications

    International Nuclear Information System (INIS)

    Oliker, Leonid; Biswas, Rupak; Shan, Hongzhang; Sing, Jaswinder Pal

    2000-01-01

    Achieving scalable performance for dynamic irregular applications is eminently challenging. Traditional message-passing approaches have been making steady progress towards this goal; however, they suffer from complex implementation requirements. The use of a global address space greatly simplifies the programming task, but can degrade the performance of dynamically adapting computations. In this work, we examine two major classes of adaptive applications, under five competing programming methodologies and four leading parallel architectures. Results indicate that it is possible to achieve message-passing performance using shared-memory programming techniques by carefully following the same high level strategies. Adaptive applications have computational work loads and communication patterns which change unpredictably at runtime, requiring dynamic load balancing to achieve scalable performance on parallel machines. Efficient parallel implementations of such adaptive applications are therefore a challenging task. This work examines the implementation of two typical adaptive applications, Dynamic Remeshing and N-Body, across various programming paradigms and architectural platforms. We compare several critical factors of the parallel code development, including performance, programmability, scalability, algorithmic development, and portability

  17. Step by step parallel programming method for molecular dynamics code

    International Nuclear Information System (INIS)

    Orii, Shigeo; Ohta, Toshio

    1996-07-01

    Parallel programming for a numerical simulation program of molecular dynamics is carried out with a step-by-step programming technique using the two phase method. As a result, within the range of a certain computing parameters, it is found to obtain parallel performance by using the level of parallel programming which decomposes the calculation according to indices of do-loops into each processor on the vector parallel computer VPP500 and the scalar parallel computer Paragon. It is also found that VPP500 shows parallel performance in wider range computing parameters. The reason is that the time cost of the program parts, which can not be reduced by the do-loop level of the parallel programming, can be reduced to the negligible level by the vectorization. After that, the time consuming parts of the program are concentrated on less parts that can be accelerated by the do-loop level of the parallel programming. This report shows the step-by-step parallel programming method and the parallel performance of the molecular dynamics code on VPP500 and Paragon. (author)

  18. Parallel Computing in SCALE

    International Nuclear Information System (INIS)

    DeHart, Mark D.; Williams, Mark L.; Bowman, Stephen M.

    2010-01-01

    The SCALE computational architecture has remained basically the same since its inception 30 years ago, although constituent modules and capabilities have changed significantly. This SCALE concept was intended to provide a framework whereby independent codes can be linked to provide a more comprehensive capability than possible with the individual programs - allowing flexibility to address a wide variety of applications. However, the current system was designed originally for mainframe computers with a single CPU and with significantly less memory than today's personal computers. It has been recognized that the present SCALE computation system could be restructured to take advantage of modern hardware and software capabilities, while retaining many of the modular features of the present system. Preliminary work is being done to define specifications and capabilities for a more advanced computational architecture. This paper describes the state of current SCALE development activities and plans for future development. With the release of SCALE 6.1 in 2010, a new phase of evolutionary development will be available to SCALE users within the TRITON and NEWT modules. The SCALE (Standardized Computer Analyses for Licensing Evaluation) code system developed by Oak Ridge National Laboratory (ORNL) provides a comprehensive and integrated package of codes and nuclear data for a wide range of applications in criticality safety, reactor physics, shielding, isotopic depletion and decay, and sensitivity/uncertainty (S/U) analysis. Over the last three years, since the release of version 5.1 in 2006, several important new codes have been introduced within SCALE, and significant advances applied to existing codes. Many of these new features became available with the release of SCALE 6.0 in early 2009. However, beginning with SCALE 6.1, a first generation of parallel computing is being introduced. In addition to near-term improvements, a plan for longer term SCALE enhancement

  19. The use of enzymes for beer brewing

    NARCIS (Netherlands)

    Donkelaar, van Laura H.G.; Mostert, Joost; Zisopoulos, Filippos K.; Boom, Remko M.; Goot, van der Atze Jan

    2016-01-01

    The exergetic performance of beer produced by the conventional malting and brewing process is compared with that of beer produced using an enzyme-assisted process. The aim is to estimate if the use of an exogenous enzyme formulation reduces the environmental impact of the overall brewing process.

  20. Parallel imaging microfluidic cytometer.

    Science.gov (United States)

    Ehrlich, Daniel J; McKenna, Brian K; Evans, James G; Belkina, Anna C; Denis, Gerald V; Sherr, David H; Cheung, Man Ching

    2011-01-01

    By adding an additional degree of freedom from multichannel flow, the parallel microfluidic cytometer (PMC) combines some of the best features of fluorescence-activated flow cytometry (FCM) and microscope-based high-content screening (HCS). The PMC (i) lends itself to fast processing of large numbers of samples, (ii) adds a 1D imaging capability for intracellular localization assays (HCS), (iii) has a high rare-cell sensitivity, and (iv) has an unusual capability for time-synchronized sampling. An inability to practically handle large sample numbers has restricted applications of conventional flow cytometers and microscopes in combinatorial cell assays, network biology, and drug discovery. The PMC promises to relieve a bottleneck in these previously constrained applications. The PMC may also be a powerful tool for finding rare primary cells in the clinic. The multichannel architecture of current PMC prototypes allows 384 unique samples for a cell-based screen to be read out in ∼6-10 min, about 30 times the speed of most current FCM systems. In 1D intracellular imaging, the PMC can obtain protein localization using HCS marker strategies at many times for the sample throughput of charge-coupled device (CCD)-based microscopes or CCD-based single-channel flow cytometers. The PMC also permits the signal integration time to be varied over a larger range than is practical in conventional flow cytometers. The signal-to-noise advantages are useful, for example, in counting rare positive cells in the most difficult early stages of genome-wide screening. We review the status of parallel microfluidic cytometry and discuss some of the directions the new technology may take. Copyright © 2011 Elsevier Inc. All rights reserved.