WorldWideScience

Sample records for core parallels climate

  1. On the effective parallel programming of multi-core processors

    NARCIS (Netherlands)

    Varbanescu, A.L.

    2010-01-01

    Multi-core processors are considered now the only feasible alternative to the large single-core processors which have become limited by technological aspects such as power consumption and heat dissipation. However, due to their inherent parallel structure and their diversity, multi-cores are

  2. Adaptive query parallelization in multi-core column stores

    NARCIS (Netherlands)

    M.M. Gawade (Mrunal); M.L. Kersten (Martin); M.M. Gawade (Mrunal); M.L. Kersten (Martin)

    2016-01-01

    htmlabstractWith the rise of multi-core CPU platforms, their optimal utilization for in-memory OLAP workloads using column store databases has become one of the biggest challenges. Some of the inherent limi- tations in the achievable query parallelism are due to the degree of parallelism

  3. Climate models on massively parallel computers

    International Nuclear Information System (INIS)

    Vitart, F.; Rouvillois, P.

    1993-01-01

    First results got on massively parallel computers (Multiple Instruction Multiple Data and Simple Instruction Multiple Data) allow to consider building of coupled models with high resolutions. This would make possible simulation of thermoaline circulation and other interaction phenomena between atmosphere and ocean. The increasing of computers powers, and then the improvement of resolution will go us to revise our approximations. Then hydrostatic approximation (in ocean circulation) will not be valid when the grid mesh will be of a dimension lower than a few kilometers: We shall have to find other models. The expert appraisement got in numerical analysis at the Center of Limeil-Valenton (CEL-V) will be used again to imagine global models taking in account atmosphere, ocean, ice floe and biosphere, allowing climate simulation until a regional scale

  4. Hydraulic Profiling of a Parallel Channel Type Reactor Core

    International Nuclear Information System (INIS)

    Seo, Kyong-Won; Hwang, Dae-Hyun; Lee, Chung-Chan

    2006-01-01

    An advanced reactor core which consisted of closed multiple parallel channels was optimized to maximize the thermal margin of the core. The closed multiple parallel channel configurations have different characteristics to the open channels of conventional PWRs. The channels, usually assemblies, are isolated hydraulically from each other and there is no cross flow between channels. The distribution of inlet flow rate between channels is a very important design parameter in the core because distribution of inlet flow is directly proportional to a margin for a certain hydraulic parameter. The thermal hydraulic parameter may be the boiling margin, maximum fuel temperature, and critical heat flux. The inlet flow distribution of the core was optimized for the boiling margins by grouping the inlet orifices by several hydraulic regions. The procedure is called a hydraulic profiling

  5. Multi-core parallelism in a column-store

    NARCIS (Netherlands)

    Gawade, M.M.

    2017-01-01

    The research reported in this thesis addresses several challenges of improving the efficiency and effectiveness of parallel processing of analytical database queries on modern multi- and many-core systems, using an open-source column-oriented analytical database management system, MonetDB, for

  6. Highly parallel line-based image coding for many cores.

    Science.gov (United States)

    Peng, Xiulian; Xu, Jizheng; Zhou, You; Wu, Feng

    2012-01-01

    Computers are developing along with a new trend from the dual-core and quad-core processors to ones with tens or even hundreds of cores. Multimedia, as one of the most important applications in computers, has an urgent need to design parallel coding algorithms for compression. Taking intraframe/image coding as a start point, this paper proposes a pure line-by-line coding scheme (LBLC) to meet the need. In LBLC, an input image is processed line by line sequentially, and each line is divided into small fixed-length segments. The compression of all segments from prediction to entropy coding is completely independent and concurrent at many cores. Results on a general-purpose computer show that our scheme can get a 13.9 times speedup with 15 cores at the encoder and a 10.3 times speedup at the decoder. Ideally, such near-linear speeding relation with the number of cores can be kept for more than 100 cores. In addition to the high parallelism, the proposed scheme can perform comparatively or even better than the H.264 high profile above middle bit rates. At near-lossless coding, it outperforms H.264 more than 10 dB. At lossless coding, up to 14% bit-rate reduction is observed compared with H.264 lossless coding at the high 4:4:4 profile.

  7. Climate Ocean Modeling on Parallel Computers

    Science.gov (United States)

    Wang, P.; Cheng, B. N.; Chao, Y.

    1998-01-01

    Ocean modeling plays an important role in both understanding the current climatic conditions and predicting future climate change. However, modeling the ocean circulation at various spatial and temporal scales is a very challenging computational task.

  8. Peformance Tuning and Evaluation of a Parallel Community Climate Model

    Energy Technology Data Exchange (ETDEWEB)

    Drake, J.B.; Worley, P.H.; Hammond, S.

    1999-11-13

    The Parallel Community Climate Model (PCCM) is a message-passing parallelization of version 2.1 of the Community Climate Model (CCM) developed by researchers at Argonne and Oak Ridge National Laboratories and at the National Center for Atmospheric Research in the early to mid 1990s. In preparation for use in the Department of Energy's Parallel Climate Model (PCM), PCCM has recently been updated with new physics routines from version 3.2 of the CCM, improvements to the parallel implementation, and ports to the SGIKray Research T3E and Origin 2000. We describe our experience in porting and tuning PCCM on these new platforms, evaluating the performance of different parallel algorithm options and comparing performance between the T3E and Origin 2000.

  9. Implementation of a parallel version of a regional climate model

    Energy Technology Data Exchange (ETDEWEB)

    Gerstengarbe, F.W. [ed.; Kuecken, M. [Potsdam-Institut fuer Klimafolgenforschung (PIK), Potsdam (Germany); Schaettler, U. [Deutscher Wetterdienst, Offenbach am Main (Germany). Geschaeftsbereich Forschung und Entwicklung

    1997-10-01

    A regional climate model developed by the Max Planck Institute for Meterology and the German Climate Computing Centre in Hamburg based on the `Europa` and `Deutschland` models of the German Weather Service has been parallelized and implemented on the IBM RS/6000 SP computer system of the Potsdam Institute for Climate Impact Research including parallel input/output processing, the explicit Eulerian time-step, the semi-implicit corrections, the normal-mode initialization and the physical parameterizations of the German Weather Service. The implementation utilizes Fortran 90 and the Message Passing Interface. The parallelization strategy used is a 2D domain decomposition. This report describes the parallelization strategy, the parallel I/O organization, the influence of different domain decomposition approaches for static and dynamic load imbalances and first numerical results. (orig.)

  10. Parallel Access of Out-Of-Core Dense Extendible Arrays

    Energy Technology Data Exchange (ETDEWEB)

    Otoo, Ekow J; Rotem, Doron

    2007-07-26

    Datasets used in scientific and engineering applications are often modeled as dense multi-dimensional arrays. For very large datasets, the corresponding array models are typically stored out-of-core as array files. The array elements are mapped onto linear consecutive locations that correspond to the linear ordering of the multi-dimensional indices. Two conventional mappings used are the row-major order and the column-major order of multi-dimensional arrays. Such conventional mappings of dense array files highly limit the performance of applications and the extendibility of the dataset. Firstly, an array file that is organized in say row-major order causes applications that subsequently access the data in column-major order, to have abysmal performance. Secondly, any subsequent expansion of the array file is limited to only one dimension. Expansions of such out-of-core conventional arrays along arbitrary dimensions, require storage reorganization that can be very expensive. Wepresent a solution for storing out-of-core dense extendible arrays that resolve the two limitations. The method uses a mapping function F*(), together with information maintained in axial vectors, to compute the linear address of an extendible array element when passed its k-dimensional index. We also give the inverse function, F-1*() for deriving the k-dimensional index when given the linear address. We show how the mapping function, in combination with MPI-IO and a parallel file system, allows for the growth of the extendible array without reorganization and no significant performance degradation of applications accessing elements in any desired order. We give methods for reading and writing sub-arrays into and out of parallel applications that run on a cluster of workstations. The axial-vectors are replicated and maintained in each node that accesses sub-array elements.

  11. Climate change from air in ice cores

    International Nuclear Information System (INIS)

    Riedel, K.

    2013-01-01

    How sensitive is our climate to greenhouse gas concentrations? What feedbacks will trigger further emissions in a warming world and at which thresholds? Over the last 200 years human activity has increased greenhouse gases to well beyond the natural range for the last 800,000 years. In order to mitigate changes - or adapt to them - we need a better understanding of greenhouse gas sources and sinks in the recent past. Ice cores with occluded ancient air hold the key to understanding the linkages between climate change and greenhouse gas variations. (author). 22 refs., 1 tab.

  12. Parallel community climate model: Description and user`s guide

    Energy Technology Data Exchange (ETDEWEB)

    Drake, J.B.; Flanery, R.E.; Semeraro, B.D.; Worley, P.H. [and others

    1996-07-15

    This report gives an overview of a parallel version of the NCAR Community Climate Model, CCM2, implemented for MIMD massively parallel computers using a message-passing programming paradigm. The parallel implementation was developed on an Intel iPSC/860 with 128 processors and on the Intel Delta with 512 processors, and the initial target platform for the production version of the code is the Intel Paragon with 2048 processors. Because the implementation uses a standard, portable message-passing libraries, the code has been easily ported to other multiprocessors supporting a message-passing programming paradigm. The parallelization strategy used is to decompose the problem domain into geographical patches and assign each processor the computation associated with a distinct subset of the patches. With this decomposition, the physics calculations involve only grid points and data local to a processor and are performed in parallel. Using parallel algorithms developed for the semi-Lagrangian transport, the fast Fourier transform and the Legendre transform, both physics and dynamics are computed in parallel with minimal data movement and modest change to the original CCM2 source code. Sequential or parallel history tapes are written and input files (in history tape format) are read sequentially by the parallel code to promote compatibility with production use of the model on other computer systems. A validation exercise has been performed with the parallel code and is detailed along with some performance numbers on the Intel Paragon and the IBM SP2. A discussion of reproducibility of results is included. A user`s guide for the PCCM2 version 2.1 on the various parallel machines completes the report. Procedures for compilation, setup and execution are given. A discussion of code internals is included for those who may wish to modify and use the program in their own research.

  13. The Glasgow Parallel Reduction Machine: Programming Shared-memory Many-core Systems using Parallel Task Composition

    Directory of Open Access Journals (Sweden)

    Ashkan Tousimojarad

    2013-12-01

    Full Text Available We present the Glasgow Parallel Reduction Machine (GPRM, a novel, flexible framework for parallel task-composition based many-core programming. We allow the programmer to structure programs into task code, written as C++ classes, and communication code, written in a restricted subset of C++ with functional semantics and parallel evaluation. In this paper we discuss the GPRM, the virtual machine framework that enables the parallel task composition approach. We focus the discussion on GPIR, the functional language used as the intermediate representation of the bytecode running on the GPRM. Using examples in this language we show the flexibility and power of our task composition framework. We demonstrate the potential using an implementation of a merge sort algorithm on a 64-core Tilera processor, as well as on a conventional Intel quad-core processor and an AMD 48-core processor system. We also compare our framework with OpenMP tasks in a parallel pointer chasing algorithm running on the Tilera processor. Our results show that the GPRM programs outperform the corresponding OpenMP codes on all test platforms, and can greatly facilitate writing of parallel programs, in particular non-data parallel algorithms such as reductions.

  14. Synthesis of parallel and antiparallel core-shell triangular nanoparticles

    Science.gov (United States)

    Bhattacharjee, Gourab; Satpati, Biswarup

    2018-04-01

    Core-shell triangular nanoparticles were synthesized by seed mediated growth. Using triangular gold (Au) nanoparticle as template, we have grown silver (Ag) shellto get core-shell nanoparticle. Here by changing the chemistry we have grown two types of core-shell structures where core and shell is having same symmetry and also having opposite symmetry. Both core and core-shell nanoparticles were characterized using transmission electron microscopy (TEM) and energy dispersive X-ray spectroscopy (EDX) to know the crystal structure and composition of these synthesized core-shell nanoparticles. From diffraction pattern analysis and energy filtered TEM (EFTEM) we have confirmed the crystal facet in core is responsible for such two dimensional growth of core-shell nanostructures.

  15. Scalable Parallelization of Skyline Computation for Multi-core Processors

    DEFF Research Database (Denmark)

    Chester, Sean; Sidlauskas, Darius; Assent, Ira

    2015-01-01

    The skyline is an important query operator for multi-criteria decision making. It reduces a dataset to only those points that offer optimal trade-offs of dimensions. In general, it is very expensive to compute. Recently, multi-core CPU algorithms have been proposed to accelerate the computation...... of the skyline. However, they do not sufficiently minimize dominance tests and so are not competitive with state-of-the-art sequential algorithms. In this paper, we introduce a novel multi-core skyline algorithm, Hybrid, which processes points in blocks. It maintains a shared, global skyline among all threads...

  16. Mathematical Methods and Algorithms of Mobile Parallel Computing on the Base of Multi-core Processors

    Directory of Open Access Journals (Sweden)

    Alexander B. Bakulev

    2012-11-01

    Full Text Available This article deals with mathematical models and algorithms, providing mobility of sequential programs parallel representation on the high-level language, presents formal model of operation environment processes management, based on the proposed model of programs parallel representation, presenting computation process on the base of multi-core processors.

  17. First results from core-edge parallel composition in the FACETS project.

    Energy Technology Data Exchange (ETDEWEB)

    Cary, J. R.; Candy, J.; Cohen, R. H.; Krasheninnikov, S.; McCune, D. C.; Estep, D. J.; Larson, J.; Malony, A. D.; Pankin, A.; Worley, P. H.; Carlsson, J. A.; Hakim, A. H.; Hamill, P.; Kruger, S.; Miah, M.; Muzsala, S.; Pletzer, A.; Shasharina, S.; Wade-Stein, D.; Wang, N.; Balay, S.; McInnes, L.; Zhang, H.; Casper, T.; Diachin, L. (Mathematics and Computer Science); (Tech-X Corp.); (General Atomics); (LLNL); (Univ. of California at San Diego); (Princeton Plasma Physics Lab.); (Colorado State Univ.); (ParaTools Inc.); (Lehigh Univ.); (ORNL)

    2008-01-01

    FACETS (Framework Application for Core-Edge Transport Simulations), now in its second year, has achieved its first coupled core-edge transport simulations. In the process, a number of accompanying accomplishments were achieved. These include a new parallel core component, a new wall component, improvements in edge and source components, and the framework for coupling all of this together. These accomplishments were a result of an interdisciplinary collaboration among computational physics, computer scientists, and applied mathematicians on the team.

  18. First results from core-edge parallel composition in the FACETS project

    Energy Technology Data Exchange (ETDEWEB)

    Cary, J R; Carlsson, J A; Hakim, A H; Hamill, P; Kruger, S; Miah, M; Muzsala, S; Pletzer, A; Shasharina, S; Wade-Stein, D; Wang, N [Tech-X Corporation, Boulder, CO 80303 (United States); Candy, J [General Atomics, San Diego, CA 92186 (United States); Cohen, R H [Lawrence Livermore National Laboratory, Livermore, CA 94550 (United States); Krasheninnikov, S [University of California at San Diego, San Diego, CA 92093 (United States); McCune, D C [Princeton Plasma Physics Laboratory, Princeton, NJ 08543 (United States); Estep, D J [Colorado State University, Fort Collins, CO 80523 (United States); Larson, J [Argonne National Laboratory, Argonne, IL 60439 (United States); Malony, A D [ParaTools, Inc., Eugene, OR 97405 (United States); Pankin, A [Lehigh University, Bethlehem, PA 18015 (United States); Worley, P H [Oak Ridge National Laboratory, Oak Ridge, TN 37831 (United States)], E-mail: cary@txcorp.com (and others)

    2008-07-15

    FACETS (Framework Application for Core-Edge Transport Simulations), now in its second year, has achieved its first coupled core-edge transport simulations. In the process, a number of accompanying accomplishments were achieved. These include a new parallel core component, a new wall component, improvements in edge and source components, and the framework for coupling all of this together. These accomplishments were a result of an interdisciplinary collaboration among computational physics, computer scientists, and applied mathematicians on the team.

  19. First results from core-edge parallel composition in the FACETS project

    Energy Technology Data Exchange (ETDEWEB)

    Cary, John R. [Tech-X Corporation; Candy, Jeff [General Atomics; Cohen, Ronald H. [Lawrence Livermore National Laboratory (LLNL); Krasheninnikov, Sergei [University of California, San Diego; McCune, Douglas [Princeton Plasma Physics Laboratory (PPPL); Estep, Donald J [Colorado State University, Fort Collins; Larson, Jay [Argonne National Laboratory (ANL); Malony, Allen [University of Oregon; Pankin, A. [Lehigh University, Bethlehem, PA; Worley, Patrick H [ORNL; Carlsson, Johann [Tech-X Corporation; Hakim, A H [Tech-X Corporation; Hamill, P [Tech-X Corporation; Kruger, Scott [Tech-X Corporation; Miah, Mahmood [Tech-X Corporation; Muzsala, S [Tech-X Corporation; Pletzer, Alexander [Tech-X Corporation; Shasharina, Svetlana [Tech-X Corporation; Wade-Stein, D [Tech-X Corporation; Wang, N [Tech-X Corporation; Balay, Satish [Argonne National Laboratory (ANL); McInnes, Lois [Argonne National Laboratory (ANL); Zhang, Hong [Argonne National Laboratory (ANL); Casper, T. A. [Lawrence Livermore National Laboratory (LLNL); Diachin, Lori [Lawrence Livermore National Laboratory (LLNL); Epperly, Thomas [Lawrence Livermore National Laboratory (LLNL); Rognlien, T. D. [Lawrence Livermore National Laboratory (LLNL); Fahey, Mark R [ORNL; Cobb, John W [ORNL; Morris, A [University of Oregon; Shende, Sameer [University of Oregon; Hammett, Greg [Princeton Plasma Physics Laboratory (PPPL); Indireshkumar, K [Tech-X Corporation; Stotler, D. [Princeton Plasma Physics Laboratory (PPPL); Pigarov, A [University of California, San Diego

    2008-01-01

    FACETS (Framework Application for Core-Edge Transport Simulations), now in its second year, has achieved its first coupled core-edge transport simulations. In the process, a number of accompanying accomplishments were achieved. These include a new parallel core component, a new wall component, improvements in edge and source components, and the framework for coupling all of this together. These accomplishments were a result of an interdisciplinary collaboration among computational physics, computer scientists, and applied mathematicians on the team.

  20. Parallelization of a three-dimensional whole core transport code DeCART

    Energy Technology Data Exchange (ETDEWEB)

    Jin Young, Cho; Han Gyu, Joo; Ha Yong, Kim; Moon-Hee, Chang [Korea Atomic Energy Research Institute, Yuseong-gu, Daejon (Korea, Republic of)

    2003-07-01

    Parallelization of the DeCART (deterministic core analysis based on ray tracing) code is presented that reduces the computational burden of the tremendous computing time and memory required in three-dimensional whole core transport calculations. The parallelization employs the concept of MPI grouping and the MPI/OpenMP mixed scheme as well. Since most of the computing time and memory are used in MOC (method of characteristics) and the multi-group CMFD (coarse mesh finite difference) calculation in DeCART, variables and subroutines related to these two modules are the primary targets for parallelization. Specifically, the ray tracing module was parallelized using a planar domain decomposition scheme and an angular domain decomposition scheme. The parallel performance of the DeCART code is evaluated by solving a rodded variation of the C5G7MOX three dimensional benchmark problem and a simplified three-dimensional SMART PWR core problem. In C5G7MOX problem with 24 CPUs, a speedup of maximum 21 is obtained on an IBM Regatta machine and 22 on a LINUX Cluster in the MOC kernel, which indicates good parallel performance of the DeCART code. In the simplified SMART problem, the memory requirement of about 11 GBytes in the single processor cases reduces to 940 Mbytes with 24 processors, which means that the DeCART code can now solve large core problems with affordable LINUX clusters. (authors)

  1. Ice core melt features in relation to Antarctic coastal climate

    NARCIS (Netherlands)

    Kaczmarska, M.; Isaksson, E.; Karlöf, L.; Brandt, O.; Winther, J.G.; van de Wal, R.S.W.; van den Broeke, M.R.; Johnsen, S.J.

    2006-01-01

    Measurement of light intensity transmission was carried out on an ice core S100 from coastal Dronning Maud Land (DML). Ice lenses were observed in digital pictures of the core and recorded as peaks in the light transmittance record. The frequency of ice layer occurrence was compared with climate

  2. Parallelization characteristics of a three-dimensional whole-core code DeCART

    International Nuclear Information System (INIS)

    Cho, J. Y.; Joo, H.K.; Kim, H. Y.; Lee, J. C.; Jang, M. H.

    2003-01-01

    Neutron transport calculation for three-dimensional amount of computing time but also huge memory. Therefore, whole-core codes such as DeCART need both also parallel computation and distributed memory capabilities. This paper is to implement such parallel capabilities based on MPI grouping and memory distribution on the DeCART code, and then to evaluate the performance by solving the C5G7 three-dimensional benchmark and a simplified three-dimensional SMART core problem. In C5G7 problem with 24 CPUs, a speedup of maximum 22 is obtained on IBM regatta machine and 21 on a LINUX cluster for the MOC kernel, which indicates good parallel performance of the DeCART code. The simplified SMART problem which need about 11 GBytes memory with one processors requires about 940 MBytes, which means that the DeCART code can now solve large core problems on affordable LINUX clusters

  3. OS and Runtime Support for Efficiently Managing Cores in Parallel Applications

    OpenAIRE

    Klues, Kevin Alan

    2015-01-01

    Parallel applications can benefit from the ability to explicitly control their thread scheduling policies in user-space. However, modern operating systems lack the interfaces necessary to make this type of “user-level” scheduling efficient. The key component missing is the ability for applications to gain direct access to cores and keep control of those cores even when making I/O operations that traditionally block in the kernel. A number of former systems provided limited support for these c...

  4. Optimization of multi-phase compressible lattice Boltzmann codes on massively parallel multi-core systems

    NARCIS (Netherlands)

    Biferale, L.; Mantovani, F.; Pivanti, M.; Pozzati, F.; Sbragaglia, M.; Schifano, S.F.; Toschi, F.; Tripiccione, R.

    2011-01-01

    We develop a Lattice Boltzmann code for computational fluid-dynamics and optimize it for massively parallel systems based on multi-core processors. Our code describes 2D multi-phase compressible flows. We analyze the performance bottlenecks that we find as we gradually expose a larger fraction of

  5. GTfold: Enabling parallel RNA secondary structure prediction on multi-core desktops

    DEFF Research Database (Denmark)

    Swenson, M Shel; Anderson, Joshua; Ash, Andrew

    2012-01-01

    achieved significant improvements in runtime, but their implementations were not portable from niche high-performance computers or easily accessible to most RNA researchers. With the increasing prevalence of multi-core desktop machines, a new parallel prediction program is needed to take full advantage...

  6. Performance modeling and analysis of parallel Gaussian elimination on multi-core computers

    Directory of Open Access Journals (Sweden)

    Fadi N. Sibai

    2014-01-01

    Full Text Available Gaussian elimination is used in many applications and in particular in the solution of systems of linear equations. This paper presents mathematical performance models and analysis of four parallel Gaussian Elimination methods (precisely the Original method and the new Meet in the Middle –MiM– algorithms and their variants with SIMD vectorization on multi-core systems. Analytical performance models of the four methods are formulated and presented followed by evaluations of these models with modern multi-core systems’ operation latencies. Our results reveal that the four methods generally exhibit good performance scaling with increasing matrix size and number of cores. SIMD vectorization only makes a large difference in performance for low number of cores. For a large matrix size (n ⩾ 16 K, the performance difference between the MiM and Original methods falls from 16× with four cores to 4× with 16 K cores. The efficiencies of all four methods are low with 1 K cores or more stressing a major problem of multi-core systems where the network-on-chip and memory latencies are too high in relation to basic arithmetic operations. Thus Gaussian Elimination can greatly benefit from the resources of multi-core systems, but higher performance gains can be achieved if multi-core systems can be designed with lower memory operation, synchronization, and interconnect communication latencies, requirements of utmost importance and challenge in the exascale computing age.

  7. Climatic Changes on Tibetan Plateau Based on Ice Core Records

    Science.gov (United States)

    Yao, T.

    2008-12-01

    Climatic changes have been reconstructed for the Tibetan Plateau based on ice core records. The Guliya ice core on the Tibetan Plateau presents climatic changes in the past 100,000 years, thus is comparative with that from Vostok ice core in Antarctica and GISP2 record in Arctic. These three records share an important common feature, i.e., our climate is not stable. It is also evident that the major patterns of climatic changes are similar on the earth. Why does climatic change over the earth follow a same pattern? It might be attributed to solar radiation. We found that the cold periods correspond to low insolation periods, and warm periods to high insolation periods. We found abrupt climatic change in the ice core climatic records, which presented dramatic temperature variation of as much as 10 °C in 50 or 60 years. Our major challenge in the study of both climate and environment is that greenhouse gases such as CO2, CH4 are possibly amplifying global warming, though at what degree remains unclear. One of the ways to understand the role of greenhouse gases is to reconstruct the past greenhouse gases recorded in ice. In 1997, we drilled an ice core from 7100 m a.s.l. in the Himalayas to reconstruct methane record. Based on the record, we found seasonal cycles in methane variation. In particular, the methane concentration is high in summer, suggestiing active methane emission from wet land in summer. Based on the seasonal cycle, we can reconstruct the methane fluctuation history in the past 500 years. The most prominent feature of the methane record in the Himalayan ice core is the abrupt increase since 1850 A.D.. This is closely related to the industrial revolution worldwide. We can also observe sudden decrease in methane concentration during the World War I and World War II. It implies that the industrial revolution has dominated the atmospheric greenhouse gas emission for about 100 years. Besides, the average methane concentration in the Himalayan ice core is

  8. Integral manifolding structure for fuel cell core having parallel gas flow

    Science.gov (United States)

    Herceg, Joseph E.

    1984-01-01

    Disclosed herein are manifolding means for directing the fuel and oxidant gases to parallel flow passageways in a fuel cell core. Each core passageway is defined by electrolyte and interconnect walls. Each electrolyte and interconnect wall consists respectively of anode and cathode materials layered on the opposite sides of electrolyte material, or on the opposite sides of interconnect material. A core wall projects beyond the open ends of the defined core passageways and is disposed approximately midway between and parallel to the adjacent overlaying and underlying interconnect walls to define manifold chambers therebetween on opposite sides of the wall. Each electrolyte wall defining the flow passageways is shaped to blend into and be connected to this wall in order to redirect the corresponding fuel and oxidant passageways to the respective manifold chambers either above or below this intermediate wall. Inlet and outlet connections are made to these separate manifold chambers respectively, for carrying the fuel and oxidant gases to the core, and for carrying their reaction products away from the core.

  9. Massively Parallel Sort-Merge Joins in Main Memory Multi-Core Database Systems

    OpenAIRE

    Albutiu, Martina-Cezara; Kemper, Alfons; Neumann, Thomas

    2012-01-01

    Two emerging hardware trends will dominate the database system technology in the near future: increasing main memory capacities of several TB per server and massively parallel multi-core processing. Many algorithmic and control techniques in current database technology were devised for disk-based systems where I/O dominated the performance. In this work we take a new look at the well-known sort-merge join which, so far, has not been in the focus of research in scalable massively parallel mult...

  10. Massively Parallel Sort-Merge Joins in Main Memory Multi-Core Database Systems

    OpenAIRE

    Martina-Cezara Albutiu, Alfons Kemper, Thomas Neumann

    2012-01-01

    Two emerging hardware trends will dominate the database system technology in the near future: increasing main memory capacities of several TB per server and massively parallel multi-core processing. Many algorithmic and control techniques in current database technology were devised for disk-based systems where I/O dominated the performance. In this work we take a new look at the well-known sort-merge join which, so far, has not been in the focus of research ...

  11. Scalable High-Performance Parallel Design for Network Intrusion Detection Systems on Many-Core Processors

    OpenAIRE

    Jiang, Hayang; Xie, Gaogang; Salamatian, Kavé; Mathy, Laurent

    2013-01-01

    Network Intrusion Detection Systems (NIDSes) face significant challenges coming from the relentless network link speed growth and increasing complexity of threats. Both hardware accelerated and parallel software-based NIDS solutions, based on commodity multi-core and GPU processors, have been proposed to overcome these challenges. Network Intrusion Detection Systems (NIDSes) face significant challenges coming from the relentless network link speed growth and increasing complexity of threats. ...

  12. Cache Locality-Centric Parallel String Matching on Many-Core Accelerator Chips

    OpenAIRE

    Tran, Nhat-Phuong; Lee, Myungho; Choi, Dong Hoon

    2015-01-01

    Aho-Corasick (AC) algorithm is a multiple patterns string matching algorithm commonly used in computer and network security and bioinformatics, among many others. In order to meet the highly demanding computational requirements imposed on these applications, achieving high performance for the AC algorithm is crucial. In this paper, we present a high performance parallelization of the AC on the many-core accelerator chips such as the Graphic Processing Unit (GPU) from Nvidia and...

  13. Parallel processing architecture for H.264 deblocking filter on multi-core platforms

    Science.gov (United States)

    Prasad, Durga P.; Sonachalam, Sekar; Kunchamwar, Mangesh K.; Gunupudi, Nageswara Rao

    2012-03-01

    Massively parallel computing (multi-core) chips offer outstanding new solutions that satisfy the increasing demand for high resolution and high quality video compression technologies such as H.264. Such solutions not only provide exceptional quality but also efficiency, low power, and low latency, previously unattainable in software based designs. While custom hardware and Application Specific Integrated Circuit (ASIC) technologies may achieve lowlatency, low power, and real-time performance in some consumer devices, many applications require a flexible and scalable software-defined solution. The deblocking filter in H.264 encoder/decoder poses difficult implementation challenges because of heavy data dependencies and the conditional nature of the computations. Deblocking filter implementations tend to be fixed and difficult to reconfigure for different needs. The ability to scale up for higher quality requirements such as 10-bit pixel depth or a 4:2:2 chroma format often reduces the throughput of a parallel architecture designed for lower feature set. A scalable architecture for deblocking filtering, created with a massively parallel processor based solution, means that the same encoder or decoder will be deployed in a variety of applications, at different video resolutions, for different power requirements, and at higher bit-depths and better color sub sampling patterns like YUV, 4:2:2, or 4:4:4 formats. Low power, software-defined encoders/decoders may be implemented using a massively parallel processor array, like that found in HyperX technology, with 100 or more cores and distributed memory. The large number of processor elements allows the silicon device to operate more efficiently than conventional DSP or CPU technology. This software programing model for massively parallel processors offers a flexible implementation and a power efficiency close to that of ASIC solutions. This work describes a scalable parallel architecture for an H.264 compliant deblocking

  14. Earth's Climate History from Glaciers and Ice Cores

    Science.gov (United States)

    Thompson, Lonnie

    2013-03-01

    Glaciers serve both as recorders and early indicators of climate change. Over the past 35 years our research team has recovered climatic and environmental histories from ice cores drilled in both Polar Regions and from low to mid-latitude, high-elevation ice fields. Those ice core -derived proxy records extending back 25,000 years have made it possible to compare glacial stage conditions in the Tropics with those in the Polar Regions. High-resolution records of δ18O (in part a temperature proxy) demonstrate that the current warming at high elevations in the mid- to lower latitudes is unprecedented for the last two millennia, although at many sites the early Holocene was warmer than today. Remarkable similarities between changes in the highland and coastal cultures of Peru and regional climate variability, especially precipitation, imply a strong connection between prehistoric human activities and regional climate. Ice cores retrieved from shrinking glaciers around the world confirm their continuous existence for periods ranging from hundreds to thousands of years, suggesting that current climatological conditions in those regions today are different from those under which these ice fields originated and have been sustained. The ongoing widespread melting of high-elevation glaciers and ice caps, particularly in low to middle latitudes, provides strong evidence that a large-scale, pervasive and, in some cases, rapid change in Earth's climate system is underway. Observations of glacier shrinkage during the 20th and 21st century girdle the globe from the South American Andes, the Himalayas, Kilimanjaro (Tanzania, Africa) and glaciers near Puncak Jaya, Indonesia (New Guinea). The history and fate of these ice caps, told through the adventure, beauty and the scientific evidence from some of world's most remote mountain tops, provide a global perspective for contemporary climate. NSF Paleoclimate Program

  15. Efficient parallel and out of core algorithms for constructing large bi-directed de Bruijn graphs

    Directory of Open Access Journals (Sweden)

    Vaughn Matthew

    2010-11-01

    Full Text Available Abstract Background Assembling genomic sequences from a set of overlapping reads is one of the most fundamental problems in computational biology. Algorithms addressing the assembly problem fall into two broad categories - based on the data structures which they employ. The first class uses an overlap/string graph and the second type uses a de Bruijn graph. However with the recent advances in short read sequencing technology, de Bruijn graph based algorithms seem to play a vital role in practice. Efficient algorithms for building these massive de Bruijn graphs are very essential in large sequencing projects based on short reads. In an earlier work, an O(n/p time parallel algorithm has been given for this problem. Here n is the size of the input and p is the number of processors. This algorithm enumerates all possible bi-directed edges which can overlap with a node and ends up generating Θ(nΣ messages (Σ being the size of the alphabet. Results In this paper we present a Θ(n/p time parallel algorithm with a communication complexity that is equal to that of parallel sorting and is not sensitive to Σ. The generality of our algorithm makes it very easy to extend it even to the out-of-core model and in this case it has an optimal I/O complexity of Θ(nlog(n/BBlog(M/B (M being the main memory size and B being the size of the disk block. We demonstrate the scalability of our parallel algorithm on a SGI/Altix computer. A comparison of our algorithm with the previous approaches reveals that our algorithm is faster - both asymptotically and practically. We demonstrate the scalability of our sequential out-of-core algorithm by comparing it with the algorithm used by VELVET to build the bi-directed de Bruijn graph. Our experiments reveal that our algorithm can build the graph with a constant amount of memory, which clearly outperforms VELVET. We also provide efficient algorithms for the bi-directed chain compaction problem. Conclusions The bi

  16. Efficient parallel and out of core algorithms for constructing large bi-directed de Bruijn graphs.

    Science.gov (United States)

    Kundeti, Vamsi K; Rajasekaran, Sanguthevar; Dinh, Hieu; Vaughn, Matthew; Thapar, Vishal

    2010-11-15

    Assembling genomic sequences from a set of overlapping reads is one of the most fundamental problems in computational biology. Algorithms addressing the assembly problem fall into two broad categories - based on the data structures which they employ. The first class uses an overlap/string graph and the second type uses a de Bruijn graph. However with the recent advances in short read sequencing technology, de Bruijn graph based algorithms seem to play a vital role in practice. Efficient algorithms for building these massive de Bruijn graphs are very essential in large sequencing projects based on short reads. In an earlier work, an O(n/p) time parallel algorithm has been given for this problem. Here n is the size of the input and p is the number of processors. This algorithm enumerates all possible bi-directed edges which can overlap with a node and ends up generating Θ(nΣ) messages (Σ being the size of the alphabet). In this paper we present a Θ(n/p) time parallel algorithm with a communication complexity that is equal to that of parallel sorting and is not sensitive to Σ. The generality of our algorithm makes it very easy to extend it even to the out-of-core model and in this case it has an optimal I/O complexity of Θ(nlog(n/B)Blog(M/B)) (M being the main memory size and B being the size of the disk block). We demonstrate the scalability of our parallel algorithm on a SGI/Altix computer. A comparison of our algorithm with the previous approaches reveals that our algorithm is faster--both asymptotically and practically. We demonstrate the scalability of our sequential out-of-core algorithm by comparing it with the algorithm used by VELVET to build the bi-directed de Bruijn graph. Our experiments reveal that our algorithm can build the graph with a constant amount of memory, which clearly outperforms VELVET. We also provide efficient algorithms for the bi-directed chain compaction problem. The bi-directed de Bruijn graph is a fundamental data structure for

  17. Coarse-grained parallel genetic algorithm applied to a nuclear reactor core design optimization problem

    International Nuclear Information System (INIS)

    Pereira, Claudio M.N.A.; Lapa, Celso M.F.

    2003-01-01

    This work extends the research related to generic algorithms (GA) in core design optimization problems, which basic investigations were presented in previous work. Here we explore the use of the Island Genetic Algorithm (IGA), a coarse-grained parallel GA model, comparing its performance to that obtained by the application of a traditional non-parallel GA. The optimization problem consists on adjusting several reactor cell parameters, such as dimensions, enrichment and materials, in order to minimize the average peak-factor in a 3-enrichment zone reactor, considering restrictions on the average thermal flux, criticality and sub-moderation. Our IGA implementation runs as a distributed application on a conventional local area network (LAN), avoiding the use of expensive parallel computers or architectures. After exhaustive experiments, taking more than 1500 h in 550 MHz personal computers, we have observed that the IGA provided gains not only in terms of computational time, but also in the optimization outcome. Besides, we have also realized that, for such kind of problem, which fitness evaluation is itself time consuming, the time overhead in the IGA, due to the communication in LANs, is practically imperceptible, leading to the conclusion that the use of expensive parallel computers or architecture can be avoided

  18. A global database with parallel measurements to study non-climatic changes

    Science.gov (United States)

    Venema, Victor; Auchmann, Renate; Aguilar, Enric; Auer, Ingeborg; Azorin-Molina, Cesar; Brandsma, Theo; Brunetti, Michele; Dienst, Manuel; Domonkos, Peter; Gilabert, Alba; Lindén, Jenny; Milewska, Ewa; Nordli, Øyvind; Prohom, Marc; Rennie, Jared; Stepanek, Petr; Trewin, Blair; Vincent, Lucie; Willett, Kate; Wolff, Mareile

    2016-04-01

    In this work we introduce the rationale behind the ongoing compilation of a parallel measurements database, in the framework of the International Surface Temperatures Initiative (ISTI) and with the support of the World Meteorological Organization. We intend this database to become instrumental for a better understanding of inhomogeneities affecting the evaluation of long-term changes in daily climate data. Long instrumental climate records are usually affected by non-climatic changes, due to, e.g., (i) station relocations, (ii) instrument height changes, (iii) instrumentation changes, (iv) observing environment changes, (v) different sampling intervals or data collection procedures, among others. These so-called inhomogeneities distort the climate signal and can hamper the assessment of long-term trends and variability of climate. Thus to study climatic changes we need to accurately distinguish non-climatic and climatic signals. The most direct way to study the influence of non-climatic changes on the distribution and to understand the reasons for these biases is the analysis of parallel measurements representing the old and new situation (in terms of e.g. instruments, location, different radiation shields, etc.). According to the limited number of available studies and our understanding of the causes of inhomogeneity, we expect that they will have a strong impact on the tails of the distribution of air temperatures and most likely of other climate elements. Our abilities to statistically homogenize daily data will be increased by systematically studying different causes of inhomogeneity replicated through parallel measurements. Current studies of non-climatic changes using parallel data are limited to local and regional case studies. However, the effect of specific transitions depends on the local climate and the most interesting climatic questions are about the systematic large-scale biases produced by transitions that occurred in many regions. Important

  19. Development of a parallel genetic algorithm using MPI and its application in a nuclear reactor core. Design optimization

    International Nuclear Information System (INIS)

    Waintraub, Marcel; Pereira, Claudio M.N.A.; Baptista, Rafael P.

    2005-01-01

    This work presents the development of a distributed parallel genetic algorithm applied to a nuclear reactor core design optimization. In the implementation of the parallelism, a 'Message Passing Interface' (MPI) library, standard for parallel computation in distributed memory platforms, has been used. Another important characteristic of MPI is its portability for various architectures. The main objectives of this paper are: validation of the results obtained by the application of this algorithm in a nuclear reactor core optimization problem, through comparisons with previous results presented by Pereira et al.; and performance test of the Brazilian Nuclear Engineering Institute (IEN) cluster in reactors physics optimization problems. The experiments demonstrated that the developed parallel genetic algorithm using the MPI library presented significant gains in the obtained results and an accentuated reduction of the processing time. Such results ratify the use of the parallel genetic algorithms for the solution of nuclear reactor core optimization problems. (author)

  20. Multilevel parallel strategy on Monte Carlo particle transport for the large-scale full-core pin-by-pin simulations

    International Nuclear Information System (INIS)

    Zhang, B.; Li, G.; Wang, W.; Shangguan, D.; Deng, L.

    2015-01-01

    This paper introduces the Strategy of multilevel hybrid parallelism of JCOGIN Infrastructure on Monte Carlo Particle Transport for the large-scale full-core pin-by-pin simulations. The particle parallelism, domain decomposition parallelism and MPI/OpenMP parallelism are designed and implemented. By the testing, JMCT presents the parallel scalability of JCOGIN, which reaches the parallel efficiency 80% on 120,000 cores for the pin-by-pin computation of the BEAVRS benchmark. (author)

  1. Possible origin and significance of extension-parallel drainages in Arizona's metamophic core complexes

    Science.gov (United States)

    Spencer, J.E.

    2000-01-01

    The corrugated form of the Harcuvar, South Mountains, and Catalina metamorphic core complexes in Arizona reflects the shape of the middle Tertiary extensional detachment fault that projects over each complex. Corrugation axes are approximately parallel to the fault-displacement direction and to the footwall mylonitic lineation. The core complexes are locally incised by enigmatic, linear drainages that parallel corrugation axes and the inferred extension direction and are especially conspicuous on the crests of antiformal corrugations. These drainages have been attributed to erosional incision on a freshly denuded, planar, inclined fault ramp followed by folding that elevated and preserved some drainages on the crests of rising antiforms. According to this hypothesis, corrugations were produced by folding after subacrial exposure of detachment-fault foot-walls. An alternative hypothesis, proposed here, is as follows. In a setting where preexisting drainages cross an active normal fault, each fault-slip event will cut each drainage into two segments separated by a freshly denuded fault ramp. The upper and lower drainage segments will remain hydraulically linked after each fault-slip event if the drainage in the hanging-wall block is incised, even if the stream is on the flank of an antiformal corrugation and there is a large component of strike-slip fault movement. Maintenance of hydraulic linkage during sequential fault-slip events will guide the lengthening stream down the fault ramp as the ramp is uncovered, and stream incision will form a progressively lengthening, extension-parallel, linear drainage segment. This mechanism for linear drainage genesis is compatible with corrugations as original irregularities of the detachment fault, and does not require folding after early to middle Miocene footwall exhumations. This is desirable because many drainages are incised into nonmylonitic crystalline footwall rocks that were probably not folded under low

  2. Parallelized Kalman-Filter-Based Reconstruction of Particle Tracks on Many-Core Architectures

    Energy Technology Data Exchange (ETDEWEB)

    Cerati, Giuseppe [Fermilab; Elmer, Peter [Princeton U.; Krutelyov, Slava [UC, San Diego; Lantz, Steven [Cornell U., Phys. Dept.; Lefebvre, Matthieu [Princeton U.; Masciovecchio, Mario [UC, San Diego; McDermott, Kevin [Cornell U., Phys. Dept.; Riley, Daniel [Cornell U., Phys. Dept.; Tadel, Matevž [UC, San Diego; Wittich, Peter [Cornell U., Phys. Dept.; Würthwein, Frank [UC, San Diego; Yagil, Avi [UC, San Diego

    2017-11-16

    Faced with physical and energy density limitations on clock speed, contemporary microprocessor designers have increasingly turned to on-chip parallelism for performance gains. Examples include the Intel Xeon Phi, GPGPUs, and similar technologies. Algorithms should accordingly be designed with ample amounts of fine-grained parallelism if they are to realize the full performance of the hardware. This requirement can be challenging for algorithms that are naturally expressed as a sequence of small-matrix operations, such as the Kalman filter methods widely in use in high-energy physics experiments. In the High-Luminosity Large Hadron Collider (HL-LHC), for example, one of the dominant computational problems is expected to be finding and fitting charged-particle tracks during event reconstruction; today, the most common track-finding methods are those based on the Kalman filter. Experience at the LHC, both in the trigger and offline, has shown that these methods are robust and provide high physics performance. Previously we reported the significant parallel speedups that resulted from our efforts to adapt Kalman-filter-based tracking to many-core architectures such as Intel Xeon Phi. Here we report on how effectively those techniques can be applied to more realistic detector configurations and event complexity.

  3. Parallel structures for disaster risk reduction and climate change adaptation in Southern Africa

    Directory of Open Access Journals (Sweden)

    Per Becker

    2013-01-01

    Full Text Available During the last decade, the interest of the international community in the concepts of disaster risk reduction and climate change adaptation has been growing immensely. Even though an increasing number of scholars seem to view these concepts as two sides of the same coin (at least when not considering the potentially positive effects of climate change, in practice the two concepts have developed in parallel rather than in an integrated manner when it comes to policy, rhetoric and funding opportunities amongst international organisations and donors. This study investigates the extent of the creation of parallel structures for disaster risk reduction and climate change adaptation in the Southern African Development Community (SADC region. The chosen methodology for the study is a comparative case study and the data are collected through focus groups and content analysis of documentary sources, as well as interviews with key informants. The results indicate that parallel structures for disaster risk reduction and climate change adaptation have been established in all but one of the studied countries. The qualitative interviews performed in some of the countries indicate that stakeholders in disaster risk reduction view this duplication of structures as unfortunate, inefficient and a fertile setup for conflict over resources for the implementation of similar activities. Additional research is called for in order to study the concrete effects of having these parallel structures as a foundation for advocacy for more efficient future disaster risk reduction and climate change adaptation.

  4. Parallel analysis tools and new visualization techniques for ultra-large climate data set

    Energy Technology Data Exchange (ETDEWEB)

    Middleton, Don [National Center for Atmospheric Research, Boulder, CO (United States); Haley, Mary [National Center for Atmospheric Research, Boulder, CO (United States)

    2014-12-10

    ParVis was a project funded under LAB 10-05: “Earth System Modeling: Advanced Scientific Visualization of Ultra-Large Climate Data Sets”. Argonne was the lead lab with partners at PNNL, SNL, NCAR and UC-Davis. This report covers progress from January 1st, 2013 through Dec 1st, 2014. Two previous reports covered the period from Summer, 2010, through September 2011 and October 2011 through December 2012, respectively. While the project was originally planned to end on April 30, 2013, personnel and priority changes allowed many of the institutions to continue work through FY14 using existing funds. A primary focus of ParVis was introducing parallelism to climate model analysis to greatly reduce the time-to-visualization for ultra-large climate data sets. Work in the first two years was conducted on two tracks with different time horizons: one track to provide immediate help to climate scientists already struggling to apply their analysis to existing large data sets and another focused on building a new data-parallel library and tool for climate analysis and visualization that will give the field a platform for performing analysis and visualization on ultra-large datasets for the foreseeable future. In the final 2 years of the project, we focused mostly on the new data-parallel library and associated tools for climate analysis and visualization.

  5. Climate change and daily press : Italy vs Usa parallel analysis

    International Nuclear Information System (INIS)

    Borrelli, G.; Mazzotta, V.; Falconi, C.; Grossi, R.; Farabollini, F.

    1996-06-01

    Among ENEA (Italian National Agency for New Technologies, Energy, and the Environment) activities, one deals with analysis and strategies of environmental information. A survey of four daily newspaper coverage, on an issue (Global Climate Change) belonging to this area, has been realized. The involved newspapers are: two Italian ones, namely 'La Repubblica' and 'Il Corriere della Sera', two North-American ones, namely 'New York Times' and 'Washington Post'. Purpose of the work was that of detecting the qualitative and quantitative level of consciousness of the Italian press via a comparison with the North-American press, notoriously sensible and careful on environmental issues. The number of articled analyzed is partitioned in the following numerical data: 319 for the 'New York Times', 309 for the 'Washington Post', 146 for the 'Corriere della Sera', 81 articles for 'La Repubblica'. The time period covered for the analysis spans from 1989, initiatic year for the organization of the 1992 Rio Conference, to December 1994, deadline date for the submission of national

  6. Development of whole core thermal-hydraulic analysis program ACT. 4. Simplified fuel assembly model and parallelization by MPI

    International Nuclear Information System (INIS)

    Ohshima, Hiroyuki

    2001-10-01

    A whole core thermal-hydraulic analysis program ACT is being developed for the purpose of evaluating detailed in-core thermal hydraulic phenomena of fast reactors including the effect of the flow between wrapper-tube walls (inter-wrapper flow) under various reactor operation conditions. As appropriate boundary conditions in addition to a detailed modeling of the core are essential for accurate simulations of in-core thermal hydraulics, ACT consists of not only fuel assembly and inter-wrapper flow analysis modules but also a heat transport system analysis module that gives response of the plant dynamics to the core model. This report describes incorporation of a simplified model to the fuel assembly analysis module and program parallelization by a message passing method toward large-scale simulations. ACT has a fuel assembly analysis module which can simulate a whole fuel pin bundle in each fuel assembly of the core and, however, it may take much CPU time for a large-scale core simulation. Therefore, a simplified fuel assembly model that is thermal-hydraulically equivalent to the detailed one has been incorporated in order to save the simulation time and resources. This simplified model is applied to several parts of fuel assemblies in a core where the detailed simulation results are not required. With regard to the program parallelization, the calculation load and the data flow of ACT were analyzed and the optimum parallelization has been done including the improvement of the numerical simulation algorithm of ACT. Message Passing Interface (MPI) is applied to data communication between processes and synchronization in parallel calculations. Parallelized ACT was verified through a comparison simulation with the original one. In addition to the above works, input manuals of the core analysis module and the heat transport system analysis module have been prepared. (author)

  7. Par@Graph - a parallel toolbox for the construction and analysis of large complex climate networks

    NARCIS (Netherlands)

    Tantet, A.J.J.

    2015-01-01

    In this paper, we present Par@Graph, a software toolbox to reconstruct and analyze complex climate networks having a large number of nodes (up to at least 106) and edges (up to at least 1012). The key innovation is an efficient set of parallel software tools designed to leverage the inherited hybrid

  8. A theoretical concept for a thermal-hydraulic 3D parallel channel core model

    International Nuclear Information System (INIS)

    Hoeld, A.

    2004-01-01

    A detailed description of the theoretical concept of the 3D thermal-hydraulic single- and two-phase flow phenomena is presented. The theoretical concept is based on important development lines such as separate treatment of the mass and energy from the momentum balance eqs. The other line is the establishment of a procedure for the calculation of the mass flow distributions into different parallel channels based on the fact that the sum of pressure decrease terms over a closed loop must stay, despite of un-symmetric perturbations, zero. The concept is realized in the experimental code HERO-X3D, concentrating in a first step on an artificial BWR or PWR core which may consist of a central channel, four quadrants, and a bypass channel. (authors)

  9. Parallelized computation for computer simulation of electrocardiograms using personal computers with multi-core CPU and general-purpose GPU.

    Science.gov (United States)

    Shen, Wenfeng; Wei, Daming; Xu, Weimin; Zhu, Xin; Yuan, Shizhong

    2010-10-01

    Biological computations like electrocardiological modelling and simulation usually require high-performance computing environments. This paper introduces an implementation of parallel computation for computer simulation of electrocardiograms (ECGs) in a personal computer environment with an Intel CPU of Core (TM) 2 Quad Q6600 and a GPU of Geforce 8800GT, with software support by OpenMP and CUDA. It was tested in three parallelization device setups: (a) a four-core CPU without a general-purpose GPU, (b) a general-purpose GPU plus 1 core of CPU, and (c) a four-core CPU plus a general-purpose GPU. To effectively take advantage of a multi-core CPU and a general-purpose GPU, an algorithm based on load-prediction dynamic scheduling was developed and applied to setting (c). In the simulation with 1600 time steps, the speedup of the parallel computation as compared to the serial computation was 3.9 in setting (a), 16.8 in setting (b), and 20.0 in setting (c). This study demonstrates that a current PC with a multi-core CPU and a general-purpose GPU provides a good environment for parallel computations in biological modelling and simulation studies. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.

  10. Parallelized Kalman-Filter-Based Reconstruction of Particle Tracks on Many-Core Processors and GPUs

    Science.gov (United States)

    Cerati, Giuseppe; Elmer, Peter; Krutelyov, Slava; Lantz, Steven; Lefebvre, Matthieu; Masciovecchio, Mario; McDermott, Kevin; Riley, Daniel; Tadel, Matevž; Wittich, Peter; Würthwein, Frank; Yagil, Avi

    2017-08-01

    For over a decade now, physical and energy constraints have limited clock speed improvements in commodity microprocessors. Instead, chipmakers have been pushed into producing lower-power, multi-core processors such as Graphical Processing Units (GPU), ARM CPUs, and Intel MICs. Broad-based efforts from manufacturers and developers have been devoted to making these processors user-friendly enough to perform general computations. However, extracting performance from a larger number of cores, as well as specialized vector or SIMD units, requires special care in algorithm design and code optimization. One of the most computationally challenging problems in high-energy particle experiments is finding and fitting the charged-particle tracks during event reconstruction. This is expected to become by far the dominant problem at the High-Luminosity Large Hadron Collider (HL-LHC), for example. Today the most common track finding methods are those based on the Kalman filter. Experience with Kalman techniques on real tracking detector systems has shown that they are robust and provide high physics performance. This is why they are currently in use at the LHC, both in the trigger and offine. Previously we reported on the significant parallel speedups that resulted from our investigations to adapt Kalman filters to track fitting and track building on Intel Xeon and Xeon Phi. Here, we discuss our progresses toward the understanding of these processors and the new developments to port the Kalman filter to NVIDIA GPUs.

  11. Parallelized Kalman-Filter-Based Reconstruction of Particle Tracks on Many-Core Processors and GPUs

    Directory of Open Access Journals (Sweden)

    Cerati Giuseppe

    2017-01-01

    Full Text Available For over a decade now, physical and energy constraints have limited clock speed improvements in commodity microprocessors. Instead, chipmakers have been pushed into producing lower-power, multi-core processors such as Graphical Processing Units (GPU, ARM CPUs, and Intel MICs. Broad-based efforts from manufacturers and developers have been devoted to making these processors user-friendly enough to perform general computations. However, extracting performance from a larger number of cores, as well as specialized vector or SIMD units, requires special care in algorithm design and code optimization. One of the most computationally challenging problems in high-energy particle experiments is finding and fitting the charged-particle tracks during event reconstruction. This is expected to become by far the dominant problem at the High-Luminosity Large Hadron Collider (HL-LHC, for example. Today the most common track finding methods are those based on the Kalman filter. Experience with Kalman techniques on real tracking detector systems has shown that they are robust and provide high physics performance. This is why they are currently in use at the LHC, both in the trigger and offine. Previously we reported on the significant parallel speedups that resulted from our investigations to adapt Kalman filters to track fitting and track building on Intel Xeon and Xeon Phi. Here, we discuss our progresses toward the understanding of these processors and the new developments to port the Kalman filter to NVIDIA GPUs.

  12. Parallelized Kalman-Filter-Based Reconstruction of Particle Tracks on Many-Core Processors and GPUs

    Energy Technology Data Exchange (ETDEWEB)

    Cerati, Giuseppe [Fermilab; Elmer, Peter [Princeton U.; Krutelyov, Slava [UC, San Diego; Lantz, Steven [Cornell U.; Lefebvre, Matthieu [Princeton U.; Masciovecchio, Mario [UC, San Diego; McDermott, Kevin [Cornell U.; Riley, Daniel [Cornell U., LNS; Tadel, Matevž [UC, San Diego; Wittich, Peter [Cornell U.; Würthwein, Frank [UC, San Diego; Yagil, Avi [UC, San Diego

    2017-01-01

    For over a decade now, physical and energy constraints have limited clock speed improvements in commodity microprocessors. Instead, chipmakers have been pushed into producing lower-power, multi-core processors such as Graphical Processing Units (GPU), ARM CPUs, and Intel MICs. Broad-based efforts from manufacturers and developers have been devoted to making these processors user-friendly enough to perform general computations. However, extracting performance from a larger number of cores, as well as specialized vector or SIMD units, requires special care in algorithm design and code optimization. One of the most computationally challenging problems in high-energy particle experiments is finding and fitting the charged-particle tracks during event reconstruction. This is expected to become by far the dominant problem at the High-Luminosity Large Hadron Collider (HL-LHC), for example. Today the most common track finding methods are those based on the Kalman filter. Experience with Kalman techniques on real tracking detector systems has shown that they are robust and provide high physics performance. This is why they are currently in use at the LHC, both in the trigger and offine. Previously we reported on the significant parallel speedups that resulted from our investigations to adapt Kalman filters to track fitting and track building on Intel Xeon and Xeon Phi. Here, we discuss our progresses toward the understanding of these processors and the new developments to port the Kalman filter to NVIDIA GPUs.

  13. Optimization and Openmp Parallelization of a Discrete Element Code for Convex Polyhedra on Multi-Core Machines

    Science.gov (United States)

    Chen, Jian; Matuttis, Hans-Georg

    2013-02-01

    We report our experiences with the optimization and parallelization of a discrete element code for convex polyhedra on multi-core machines and introduce a novel variant of the sort-and-sweep neighborhood algorithm. While in theory the whole code in itself parallelizes ideally, in practice the results on different architectures with different compilers and performance measurement tools depend very much on the particle number and optimization of the code. After difficulties with the interpretation of the data for speedup and efficiency are overcome, respectable parallelization speedups could be obtained.

  14. Cpl6: The New Extensible, High-Performance Parallel Coupler forthe Community Climate System Model

    Energy Technology Data Exchange (ETDEWEB)

    Craig, Anthony P.; Jacob, Robert L.; Kauffman, Brain; Bettge,Tom; Larson, Jay; Ong, Everest; Ding, Chris; He, Yun

    2005-03-24

    Coupled climate models are large, multiphysics applications designed to simulate the Earth's climate and predict the response of the climate to any changes in the forcing or boundary conditions. The Community Climate System Model (CCSM) is a widely used state-of-art climate model that has released several versions to the climate community over the past ten years. Like many climate models, CCSM employs a coupler, a functional unit that coordinates the exchange of data between parts of climate system such as the atmosphere and ocean. This paper describes the new coupler, cpl6, contained in the latest version of CCSM,CCSM3. Cpl6 introduces distributed-memory parallelism to the coupler, a class library for important coupler functions, and a standardized interface for component models. Cpl6 is implemented entirely in Fortran90 and uses Model Coupling Toolkit as the base for most of its classes. Cpl6 gives improved performance over previous versions and scales well on multiple platforms.

  15. Spatial data analytics on heterogeneous multi- and many-core parallel architectures using python

    Science.gov (United States)

    Laura, Jason R.; Rey, Sergio J.

    2017-01-01

    Parallel vector spatial analysis concerns the application of parallel computational methods to facilitate vector-based spatial analysis. The history of parallel computation in spatial analysis is reviewed, and this work is placed into the broader context of high-performance computing (HPC) and parallelization research. The rise of cyber infrastructure and its manifestation in spatial analysis as CyberGIScience is seen as a main driver of renewed interest in parallel computation in the spatial sciences. Key problems in spatial analysis that have been the focus of parallel computing are covered. Chief among these are spatial optimization problems, computational geometric problems including polygonization and spatial contiguity detection, the use of Monte Carlo Markov chain simulation in spatial statistics, and parallel implementations of spatial econometric methods. Future directions for research on parallelization in computational spatial analysis are outlined.

  16. Climate prosperity, parallel paths: Canada-U.S. climate policy choices

    International Nuclear Information System (INIS)

    2011-01-01

    The National Round Table on the Environment and the Economy (NRTEE) has conducted a study on the economic risks and opportunities of climate policies for Canada in the Canada-United States relationship background. This research aims to inform future policy decisions and provide ideas on means to serve Canadian interests in the context of this climate changing world. The NRTEE's research presented in this report focuses on the economic and environmental implications for Canada of harmonizing with the U.S. on climate policy. The document follows two main objectives: evaluate the consequences of U.S. climate policy choices for Canada and identify the policy options that would lead to a long-term reduction of emissions, taking into account the economic risks associated to policy choices in the U.S. and in Canada. According to the NRTEE, the government of Canada should take under consideration the benefits of the implementation of its own strategy for harmonization. This document provides a path leading to the achievement of a climate policy harmonization with the United States. 52 refs., 18 tabs., 24 figs.

  17. Efficient Out of Core Sorting Algorithms for the Parallel Disks Model.

    Science.gov (United States)

    Kundeti, Vamsi; Rajasekaran, Sanguthevar

    2011-11-01

    In this paper we present efficient algorithms for sorting on the Parallel Disks Model (PDM). Numerous asymptotically optimal algorithms have been proposed in the literature. However many of these merge based algorithms have large underlying constants in the time bounds, because they suffer from the lack of read parallelism on PDM. The irregular consumption of the runs during the merge affects the read parallelism and contributes to the increased sorting time. In this paper we first introduce a novel idea called the dirty sequence accumulation that improves the read parallelism. Secondly, we show analytically that this idea can reduce the number of parallel I/O's required to sort the input close to the lower bound of [Formula: see text]. We experimentally verify our dirty sequence idea with the standard R-Way merge and show that our idea can reduce the number of parallel I/Os to sort on PDM significantly.

  18. Evidence for general instability of past climate from a 250-KYR ice-core record

    DEFF Research Database (Denmark)

    Johnsen, Sigfus Johann; Clausen, Henrik Brink; Dahl-Jensen, Dorthe

    1993-01-01

    decades. Here we present a detailed stable-isotope record for the full length of the Greenland Ice-core Project Summit ice core, extending over the past 250 kyr according to a calculated timescale. We find that climate instability was not confined to the last glaciation, but appears also to have been...... results1,2 from two ice cores drilled in central Greenland have revealed large, abrupt climate changes of at least regional extent during the late stages of the last glaciation, suggesting that climate in the North Atlantic region is able to reorganize itself rapidly, perhaps even within a few...

  19. The design of multi-core DSP parallel model based on message passing and multi-level pipeline

    Science.gov (United States)

    Niu, Jingyu; Hu, Jian; He, Wenjing; Meng, Fanrong; Li, Chuanrong

    2017-10-01

    Currently, the design of embedded signal processing system is often based on a specific application, but this idea is not conducive to the rapid development of signal processing technology. In this paper, a parallel processing model architecture based on multi-core DSP platform is designed, and it is mainly suitable for the complex algorithms which are composed of different modules. This model combines the ideas of multi-level pipeline parallelism and message passing, and summarizes the advantages of the mainstream model of multi-core DSP (the Master-Slave model and the Data Flow model), so that it has better performance. This paper uses three-dimensional image generation algorithm to validate the efficiency of the proposed model by comparing with the effectiveness of the Master-Slave and the Data Flow model.

  20. A parallel solution-adaptive scheme for predicting multi-phase core flows in solid propellant rocket motors

    International Nuclear Information System (INIS)

    Sachdev, J.S.; Groth, C.P.T.; Gottlieb, J.J.

    2003-01-01

    The development of a parallel adaptive mesh refinement (AMR) scheme is described for solving the governing equations for multi-phase (gas-particle) core flows in solid propellant rocket motors (SRM). An Eulerian formulation is used to described the coupled motion between the gas and particle phases. A cell-centred upwind finite-volume discretization and the use of limited solution reconstruction, Riemann solver based flux functions for the gas and particle phases, and explicit multi-stage time-stepping allows for high solution accuracy and computational robustness. A Riemann problem is formulated for prescribing boundary data at the burning surface. Efficient and scalable parallel implementations are achieved with domain decomposition on distributed memory multiprocessor architectures. Numerical results are described to demonstrate the capabilities of the approach for predicting SRM core flows. (author)

  1. Reactor core T-H characteristics determination in case of parallel operation of different fuel assembly types

    International Nuclear Information System (INIS)

    Hermansky, J.; Petenyi, V.; Zavodsky, M.

    2009-01-01

    The WWER-440 nuclear fuel vendor permanently improve the assortment of produced nuclear fuel assemblies for achieving better fuel cycle economy and reactor operation safety. Therefore it is necessary to have the skilled methodology and computing code for analyzing factors which affecting the accuracy of flow redistributed determination through reactor on flows through separate parts of reactor core in case of parallel operation different assembly types. Whereas the geometric parameters of new manufactured assemblies were changed recently, the calculated flows through the fuel parts of different type of assemblies are depended also on their real position in reactor core. Therefore the computing code CORFLO was developed in VUJE Trnava for carrying out stationary analyses of T-H characteristics of reactor core within 60 deg symmetry. The CORFLO code deals the area of the active core which consists of 312 fuel assemblies and 37 control assemblies. Regarding the rotational 60 deg symmetry of reactor core only 1/6 of reactor core with 59 fuel assemblies is calculated. Computing code is verified and validated at this time. Paper presents the short description of computing code CORFLO with some calculated results. (Authors)

  2. GRAPES: a software for parallel searching on biological graphs targeting multi-core architectures.

    Directory of Open Access Journals (Sweden)

    Rosalba Giugno

    Full Text Available Biological applications, from genomics to ecology, deal with graphs that represents the structure of interactions. Analyzing such data requires searching for subgraphs in collections of graphs. This task is computationally expensive. Even though multicore architectures, from commodity computers to more advanced symmetric multiprocessing (SMP, offer scalable computing power, currently published software implementations for indexing and graph matching are fundamentally sequential. As a consequence, such software implementations (i do not fully exploit available parallel computing power and (ii they do not scale with respect to the size of graphs in the database. We present GRAPES, software for parallel searching on databases of large biological graphs. GRAPES implements a parallel version of well-established graph searching algorithms, and introduces new strategies which naturally lead to a faster parallel searching system especially for large graphs. GRAPES decomposes graphs into subcomponents that can be efficiently searched in parallel. We show the performance of GRAPES on representative biological datasets containing antiviral chemical compounds, DNA, RNA, proteins, protein contact maps and protein interactions networks.

  3. Paleoclimate from ice cores : abrupt climate change and the prolonged Holocene

    International Nuclear Information System (INIS)

    White, J.W.C.

    2001-01-01

    Ice cores provide valuable information about the Earth's past climates and past environments. They can also help in predicting future climates and the nature of climate change. Recent findings in ice cores have shown large and abrupt climate changes in the past. This paper addressed abrupt climate changes and the peculiar nature of the Holocene. An abrupt climate change is a shift of 5 degrees C in mean annual temperature in less than 50 years. This is considered to be the most threatening aspect of potential future climate change since it leaves very little time for adaptation by humans or any other part of the Earth's ecosystem. This paper also discussed the arrival of the next glacial period. In the past 50 years, scientists have recognized the importance of the Earth's orbit around the sun in pacing the occurrence of large ice sheets. The timing of orbital forcing suggests that the Earth is overdue for the next major glaciation. The reason for this anomaly was discussed. Abrupt climate shifts seem to be caused by mode changes in sensitive points in the climate system, such as the North Atlantic Deep Water Formation and its impact on sea ice cover in the North Atlantic. These changes have been observed in ice cores in Greenland but they are not restricted to Greenland. Evidence from Antarctic ice cores suggest that abrupt climate change may also occur in the Southern Hemisphere. The Vostok ice core in Antarctica indicates that the 11,000 year long interglacial period that we are in right now is longer than the previous four interglacial periods. The Holocene epoch is unique because both methane and carbon dioxide rise in the last 6,000 years, an atypical response from these greenhouse gases during an interglacial period. It was suggested that the rise in methane can be attributed to human activities. 13 refs., 2 figs

  4. Operating system design of parallel computer for on-line management of nuclear pressurised water reactor cores

    International Nuclear Information System (INIS)

    Gougam, F.

    1991-04-01

    This study is part of the PHAETON project which aims at increasing the knowledge of safety parameters of PWR core and reducing operating margins during the reactor cycle. The on-line system associates a simulator process to compute the three dimensional flux distribution and an acquisition process of reactor core parameters from the central instrumentation. The 3D flux calculation is the most time consuming. So, for cost and safety reasons, the PHAETON project proposes an approach which is to parallelize the 3D diffusion calculation and to use a computer based on parallel processor architecture. This paper presents the design of the operating system on which the application is executed. The routine interface proposed, includes the main operations necessary for programming a real time and parallel application. The primitives include: task management, data transfer, synchronisation by event signalling and by using the rendez-vous mechanisms. The primitives which are proposed use standard softwares like real-time kernel and UNIX operating system [fr

  5. How to use MPI communication in highly parallel climate simulations more easily and more efficiently.

    Science.gov (United States)

    Behrens, Jörg; Hanke, Moritz; Jahns, Thomas

    2014-05-01

    In this talk we present a way to facilitate efficient use of MPI communication for developers of climate models. Exploitation of the performance potential of today's highly parallel supercomputers with real world simulations is a complex task. This is partly caused by the low level nature of the MPI communication library which is the dominant communication tool at least for inter-node communication. In order to manage the complexity of the task, climate simulations with non-trivial communication patterns often use an internal abstraction layer above MPI without exploiting the benefits of communication aggregation or MPI-datatypes. The solution for the complexity and performance problem we propose is the communication library YAXT. This library is built on top of MPI and takes high level descriptions of arbitrary domain decompositions and automatically derives an efficient collective data exchange. Several exchanges can be aggregated in order to reduce latency costs. Examples are given which demonstrate the simplicity and the performance gains for selected climate applications.

  6. High performance parallelism pearls 2 multicore and many-core programming approaches

    CERN Document Server

    Jeffers, Jim

    2015-01-01

    High Performance Parallelism Pearls Volume 2 offers another set of examples that demonstrate how to leverage parallelism. Similar to Volume 1, the techniques included here explain how to use processors and coprocessors with the same programming - illustrating the most effective ways to combine Xeon Phi coprocessors with Xeon and other multicore processors. The book includes examples of successful programming efforts, drawn from across industries and domains such as biomed, genetics, finance, manufacturing, imaging, and more. Each chapter in this edited work includes detailed explanations of t

  7. An Efficient Parallel SAT Solver Exploiting Multi-Core Environments, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — The hundreds of stream cores in the latest graphics processors (GPUs), and the possibility to execute non-graphics computations on them, open unprecedented levels of...

  8. Parallel DC3 Algorithm for Suffix Array Construction on Many-Core Accelerators

    KAUST Repository

    Liao, Gang; Ma, Longfei; Zang, Guangming; Tang, Lin

    2015-01-01

    In bioinformatics applications, suffix arrays are widely used to DNA sequence alignments in the initial exact match phase of heuristic algorithms. With the exponential growth and availability of data, using many-core accelerators, like GPUs, to optimize existing algorithms is very common. We present a new implementation of suffix array on GPU. As a result, suffix array construction on GPU achieves around 10x speedup on standard large data sets, which contain more than 100 million characters. The idea is simple, fast and scalable that can be easily scale to multi-core processors and even heterogeneous architectures. © 2015 IEEE.

  9. Parallel DC3 Algorithm for Suffix Array Construction on Many-Core Accelerators

    KAUST Repository

    Liao, Gang

    2015-05-01

    In bioinformatics applications, suffix arrays are widely used to DNA sequence alignments in the initial exact match phase of heuristic algorithms. With the exponential growth and availability of data, using many-core accelerators, like GPUs, to optimize existing algorithms is very common. We present a new implementation of suffix array on GPU. As a result, suffix array construction on GPU achieves around 10x speedup on standard large data sets, which contain more than 100 million characters. The idea is simple, fast and scalable that can be easily scale to multi-core processors and even heterogeneous architectures. © 2015 IEEE.

  10. Out of core, out of mind: Practical parallel I/O

    Energy Technology Data Exchange (ETDEWEB)

    Womble, D.E.; Greenberg, D.S.; Riesen, R.E.; Wheat, S.R.

    1993-11-01

    Parallel computers are becoming more powerful and more complex in response to the demand for computing power by scientists and engineers. Inevitably, new and more complex I/O systems will be developed for these systems. In particular we believe that the I/O system must provide the programmer with the ability to explcitly manage storage (despite the trend toward complex parallel file systems and caching schemes). One method of doing so is to have a partitioned secondary storage in which each processor owns a logical disk. Along with operating system enhancements which allow overheads such as buffer copying to be avoided and libraries to support optimal remapping of data, this sort of I/O system meets the needs of high performance computing.

  11. ISP: an optimal out-of-core image-set processing streaming architecture for parallel heterogeneous systems.

    Science.gov (United States)

    Ha, Linh Khanh; Krüger, Jens; Dihl Comba, João Luiz; Silva, Cláudio T; Joshi, Sarang

    2012-06-01

    Image population analysis is the class of statistical methods that plays a central role in understanding the development, evolution, and disease of a population. However, these techniques often require excessive computational power and memory that are compounded with a large number of volumetric inputs. Restricted access to supercomputing power limits its influence in general research and practical applications. In this paper we introduce ISP, an Image-Set Processing streaming framework that harnesses the processing power of commodity heterogeneous CPU/GPU systems and attempts to solve this computational problem. In ISP, we introduce specially designed streaming algorithms and data structures that provide an optimal solution for out-of-core multiimage processing problems both in terms of memory usage and computational efficiency. ISP makes use of the asynchronous execution mechanism supported by parallel heterogeneous systems to efficiently hide the inherent latency of the processing pipeline of out-of-core approaches. Consequently, with computationally intensive problems, the ISP out-of-core solution can achieve the same performance as the in-core solution. We demonstrate the efficiency of the ISP framework on synthetic and real datasets.

  12. The Crystal Structures of the N-terminal Photosensory Core Module of Agrobacterium Phytochrome Agp1 as Parallel and Anti-parallel Dimers.

    Science.gov (United States)

    Nagano, Soshichiro; Scheerer, Patrick; Zubow, Kristina; Michael, Norbert; Inomata, Katsuhiko; Lamparter, Tilman; Krauß, Norbert

    2016-09-23

    Agp1 is a canonical biliverdin-binding bacteriophytochrome from the soil bacterium Agrobacterium fabrum that acts as a light-regulated histidine kinase. Crystal structures of the photosensory core modules (PCMs) of homologous phytochromes have provided a consistent picture of the structural changes that these proteins undergo during photoconversion between the parent red light-absorbing state (Pr) and the far-red light-absorbing state (Pfr). These changes include secondary structure rearrangements in the so-called tongue of the phytochrome-specific (PHY) domain and structural rearrangements within the long α-helix that connects the cGMP-specific phosphodiesterase, adenylyl cyclase, and FhlA (GAF) and the PHY domains. We present the crystal structures of the PCM of Agp1 at 2.70 Å resolution and of a surface-engineered mutant of this PCM at 1.85 Å resolution in the dark-adapted Pr states. Whereas in the mutant structure the dimer subunits are in anti-parallel orientation, the wild-type structure contains parallel subunits. The relative orientations between the PAS-GAF bidomain and the PHY domain are different in the two structures, due to movement involving two hinge regions in the GAF-PHY connecting α-helix and the tongue, indicating pronounced structural flexibility that may give rise to a dynamic Pr state. The resolution of the mutant structure enabled us to detect a sterically strained conformation of the chromophore at ring A that we attribute to the tight interaction with Pro-461 of the conserved PRXSF motif in the tongue. Based on this observation and on data from mutants where residues in the tongue region were replaced by alanine, we discuss the crucial roles of those residues in Pr-to-Pfr photoconversion. © 2016 by The American Society for Biochemistry and Molecular Biology, Inc.

  13. The Crystal Structures of the N-terminal Photosensory Core Module of Agrobacterium Phytochrome Agp1 as Parallel and Anti-parallel Dimers*

    Science.gov (United States)

    Nagano, Soshichiro; Scheerer, Patrick; Zubow, Kristina; Michael, Norbert; Inomata, Katsuhiko; Lamparter, Tilman; Krauß, Norbert

    2016-01-01

    Agp1 is a canonical biliverdin-binding bacteriophytochrome from the soil bacterium Agrobacterium fabrum that acts as a light-regulated histidine kinase. Crystal structures of the photosensory core modules (PCMs) of homologous phytochromes have provided a consistent picture of the structural changes that these proteins undergo during photoconversion between the parent red light-absorbing state (Pr) and the far-red light-absorbing state (Pfr). These changes include secondary structure rearrangements in the so-called tongue of the phytochrome-specific (PHY) domain and structural rearrangements within the long α-helix that connects the cGMP-specific phosphodiesterase, adenylyl cyclase, and FhlA (GAF) and the PHY domains. We present the crystal structures of the PCM of Agp1 at 2.70 Å resolution and of a surface-engineered mutant of this PCM at 1.85 Å resolution in the dark-adapted Pr states. Whereas in the mutant structure the dimer subunits are in anti-parallel orientation, the wild-type structure contains parallel subunits. The relative orientations between the PAS-GAF bidomain and the PHY domain are different in the two structures, due to movement involving two hinge regions in the GAF-PHY connecting α-helix and the tongue, indicating pronounced structural flexibility that may give rise to a dynamic Pr state. The resolution of the mutant structure enabled us to detect a sterically strained conformation of the chromophore at ring A that we attribute to the tight interaction with Pro-461 of the conserved PRXSF motif in the tongue. Based on this observation and on data from mutants where residues in the tongue region were replaced by alanine, we discuss the crucial roles of those residues in Pr-to-Pfr photoconversion. PMID:27466363

  14. Direct north-south synchronization of abrupt climate change record in ice cores using Beryllium 10

    Directory of Open Access Journals (Sweden)

    G. M. Raisbeck

    2007-09-01

    Full Text Available A new, decadally resolved record of the 10Be peak at 41 kyr from the EPICA Dome C ice core (Antarctica is used to match it with the same peak in the GRIP ice core (Greenland. This permits a direct synchronisation of the climatic variations around this time period, independent of uncertainties related to the ice age-gas age difference in ice cores. Dansgaard-Oeschger event 10 is in the period of best synchronisation and is found to be coeval with an Antarctic temperature maximum. Simulations using a thermal bipolar seesaw model agree reasonably well with the observed relative climate chronology in these two cores. They also reproduce three Antarctic warming events observed between A1 and A2.

  15. Early Holocene climate oscillations recorded in three Greenland ice cores

    DEFF Research Database (Denmark)

    Rasmussen, Sune Olander; Vinther, Bo Møllesøe; Clausen, Henrik Brink

    2007-01-01

    around 9.3 ka before present, and the Preboreal Oscillation during the first centuries of the Holocene. For each of these sections, we present a d18O anomaly curve and a common accumulation signal that represents regional changes in the accumulation rate over the Greenland ice cap....... and accumulation anomalies that are common to the three cores in the Early Holocene (7.9–11.7 ka before present). Three time periods with significant and synchronous anomalies in the d18O and accumulation signals stand out: the well-known 8.2 ka event, an event of shorter duration but of almost similar amplitude...

  16. A MULTI-CORE PARALLEL MOSAIC ALORITHM FOR MULTI-VIEW UAV IMAGES

    Directory of Open Access Journals (Sweden)

    X. Pan

    2017-09-01

    Full Text Available As the spread of the error and accumulation often lead to distortion or failure of image mosaic during the multi-view UAV (Unmanned Aerial Vehicle images stitching. In this paper, to solve the problem we propose a mosaic strategy to construct a mosaic ring and multi-level grouping parallel acceleration as an auxiliary. First, the input images will be divided into several groups, each group in the ring way to stitch. Then, use SIFT for matching, RANSAC to remove the wrong matching points. And then, calculate the perspective transformation matrix. Finally weaken the error by using the adjustment equation. All these steps run between different groups at the same time. By using real UAV images, the experiment results show that this method can effectively reduce the influence of accumulative error, improve the precision of mosaic and reduce the mosaic time by 60 %. The proposed method can be used as one of the effective ways to minimize the accumulative error.

  17. The Principalship: Essential Core Competencies for Instructional Leadership and Its Impact on School Climate

    Science.gov (United States)

    Ross, Dorrell J.; Cozzens, Jeffry A.

    2016-01-01

    The purpose of this quantitative study was to investigate teachers' perceptions of principals' leadership behaviors influencing the schools' climate according to Green's (2010) ideologies of the 13 core competencies within the four dimensions of principal leadership. Data from the "Leadership Behavior Inventory" (Green, 2014) suggest 314…

  18. Long memory effect of past climate change in Vostok ice core records

    International Nuclear Information System (INIS)

    Yamamoto, Yuuki; Kitahara, Naoki; Kano, Makoto

    2012-01-01

    Time series analysis of Vostok ice core data has been done for understanding of palaeoclimate change from a stochastic perspective. The Vostok ice core is one of the proxy data for palaeoclimate in which local temperature and precipitation rate, moisture source conditions, wind strength and aerosol fluxes of marine, volcanic, terrestrial, cosmogenic and anthropogenic origin are indirectly stored. Palaeoclimate data has a periodic feature and a stochastic feature. For the proxy data, spectrum analysis and detrended fluctuation analysis (DFA) were conducted to characterize periodicity and scaling property (long memory effect) in the climate change. The result of spectrum analysis indicates there exist periodicities corresponding to the Milankovitch cycle in past climate change occurred. DFA clarified time variability of scaling exponents (Hurst exponent) is associated with abrupt warming in past climate.

  19. Hominin Sites and Paleolakes Drilling Project. Chew Bahir, southern Ethiopia: How to get from three tonnes of sediment core to > 500 ka of continuous climate history?

    Science.gov (United States)

    Foerster, Verena; Asrat, Asfawossen; Cohen, Andrew S.; Gromig, Raphael; Günter, Christina; Junginger, Annett; Lamb, Henry F.; Schaebitz, Frank; Trauth, Martin H.

    2016-04-01

    In search of the environmental context of the evolution and dispersal of Homo sapiens and our close relatives within and beyond the African continent, the ICDP-funded Hominin Sites and Paleolakes Drilling Project (HSPDP) has recently cored five fluvio-lacustrine archives of climate change in East Africa. The sediment cores collected in Ethiopia and Kenya are expected to provide valuable insights into East African environmental variability during the last ~3.5 Ma. The tectonically-bound Chew Bahir basin in the southern Ethiopian rift is one of the five sites within HSPDP, located in close proximity to the Lower Omo River valley, the site of the oldest known fossils of anatomically modern humans. In late 2014, the two cores (279 and 266 m long respectively, HSPDP-CHB14-2A and 2B) were recovered, summing up to nearly three tonnes of mostly calcareous clays and silts. Deciphering an environmental record from multiple records, from the source region of modern humans could eventually allow us to reconstruct the pronounced variations of moisture availability during the transition into Middle Stone Age, and its implications for the origin and dispersal of Homo sapiens. Here we present the first results of our analysis of the Chew Bahir cores. Following the HSPDP protocols, the two parallel Chew Bahir sediment cores have been merged into one single, 280 m long and nearly continuous (>90%) composite core on the basis of a high resolution MSCL data set (e.g., magnetic susceptibility, gamma ray density, color intensity transects, core photographs). Based on the obvious cyclicities in the MSCL, correlated with orbital cycles, the time interval covered by our sediment archive of climate change is inferred to span the last 500-600 kyrs. Combining our first results from the long cores with the results from the accomplished pre-study of short cores taken in 2009/10 along a NW-SE transect across the basin (Foerster et al., 2012, Trauth et al., 2015), we have developed a hypothesis

  20. Past temperature reconstructions from deep ice cores: relevance for future climate change

    Directory of Open Access Journals (Sweden)

    V. Masson-Delmotte

    2006-01-01

    Full Text Available Ice cores provide unique archives of past climate and environmental changes based only on physical processes. Quantitative temperature reconstructions are essential for the comparison between ice core records and climate models. We give an overview of the methods that have been developed to reconstruct past local temperatures from deep ice cores and highlight several points that are relevant for future climate change. We first analyse the long term fluctuations of temperature as depicted in the long Antarctic record from EPICA Dome C. The long term imprint of obliquity changes in the EPICA Dome C record is highlighted and compared to simulations conducted with the ECBILT-CLIO intermediate complexity climate model. We discuss the comparison between the current interglacial period and the long interglacial corresponding to marine isotopic stage 11, ~400 kyr BP. Previous studies had focused on the role of precession and the thresholds required to induce glacial inceptions. We suggest that, due to the low eccentricity configuration of MIS 11 and the Holocene, the effect of precession on the incoming solar radiation is damped and that changes in obliquity must be taken into account. The EPICA Dome C alignment of terminations I and VI published in 2004 corresponds to a phasing of the obliquity signals. A conjunction of low obliquity and minimum northern hemisphere summer insolation is not found in the next tens of thousand years, supporting the idea of an unusually long interglacial ahead. As a second point relevant for future climate change, we discuss the magnitude and rate of change of past temperatures reconstructed from Greenland (NorthGRIP and Antarctic (Dome C ice cores. Past episodes of temperatures above the present-day values by up to 5°C are recorded at both locations during the penultimate interglacial period. The rate of polar warming simulated by coupled climate models forced by a CO2 increase of 1% per year is compared to ice-core

  1. The ice-core record - Climate sensitivity and future greenhouse warming

    Science.gov (United States)

    Lorius, C.; Raynaud, D.; Jouzel, J.; Hansen, J.; Le Treut, H.

    1990-01-01

    The prediction of future greenhouse-gas-warming depends critically on the sensitivity of earth's climate to increasing atmospheric concentrations of these gases. Data from cores drilled in polar ice sheets show a remarkable correlation between past glacial-interglacial temperature changes and the inferred atmospheric concentration of gases such as carbon dioxide and methane. These and other palaeoclimate data are used to assess the role of greenhouse gases in explaining past global climate change, and the validity of models predicting the effect of increasing concentrations of such gases in the atmosphere.

  2. Parallel computing of a climate model on the dawn 1000 by domain decomposition method

    Science.gov (United States)

    Bi, Xunqiang

    1997-12-01

    In this paper the parallel computing of a grid-point nine-level atmospheric general circulation model on the Dawn 1000 is introduced. The model was developed by the Institute of Atmospheric Physics (IAP), Chinese Academy of Sciences (CAS). The Dawn 1000 is a MIMD massive parallel computer made by National Research Center for Intelligent Computer (NCIC), CAS. A two-dimensional domain decomposition method is adopted to perform the parallel computing. The potential ways to increase the speed-up ratio and exploit more resources of future massively parallel supercomputation are also discussed.

  3. Annually resolved ice core records of tropical climate variability over the past ~1800 years.

    Science.gov (United States)

    Thompson, L G; Mosley-Thompson, E; Davis, M E; Zagorodnov, V S; Howat, I M; Mikhalenko, V N; Lin, P-N

    2013-05-24

    Ice cores from low latitudes can provide a wealth of unique information about past climate in the tropics, but they are difficult to recover and few exist. Here, we report annually resolved ice core records from the Quelccaya ice cap (5670 meters above sea level) in Peru that extend back ~1800 years and provide a high-resolution record of climate variability there. Oxygen isotopic ratios (δ(18)O) are linked to sea surface temperatures in the tropical eastern Pacific, whereas concentrations of ammonium and nitrate document the dominant role played by the migration of the Intertropical Convergence Zone in the region of the tropical Andes. Quelccaya continues to retreat and thin. Radiocarbon dates on wetland plants exposed along its retreating margins indicate that it has not been smaller for at least six millennia.

  4. RICE ice core: Black Carbon reflects climate variability at Roosevelt Island, West Antarctica

    Science.gov (United States)

    Ellis, Aja; Edwards, Ross; Bertler, Nancy; Winton, Holly; Goodwin, Ian; Neff, Peter; Tuohy, Andrea; Proemse, Bernadette; Hogan, Chad; Feiteng, Wang

    2015-04-01

    The Roosevelt Island Climate Evolution (RICE) project successfully drilled a deep ice core from Roosevelt Island during the 2011/2012 and 2012/2013 seasons. Located in the Ross Ice Shelf in West Antarctica, the site is an ideal location for investigating climate variability and the past stability of the Ross Ice Shelf. Black carbon (BC) aerosols are emitted by both biomass burning and fossil fuels, and BC particles emitted in the southern hemisphere are transported in the atmosphere and preserved in Antarctic ice. The past record of BC is expected to be sensitive to climate variability, as it is modulated by both emissions and transport. To investigate BC variability over the past 200 years, we developed a BC record from two overlapping ice cores (~1850-2012) and a high-resolution snow pit spanning 2010-2012 (cal. yr). Consistent results are found between the snow pit profiles and ice core records. Distinct decadal trends are found with respect to BC particle size, and the record indicates a steady rise in BC particle size over the last 100 years. Differences in emission sources and conditions may be a possible explanation for changes in BC size. These records also show a significant increase in BC concentration over the past decade with concentrations rising over 1.5 ppb (1.5*10^-9 ng/g), suggesting a fundamental shift in BC deposition to the site.

  5. 12-core x 3-mode Dense Space Division Multiplexed Transmission over 40 km Employing Multi-carrier Signals with Parallel MIMO Equalization

    DEFF Research Database (Denmark)

    Mizuno, T.; Kobayashi, T.; Takara, H.

    2014-01-01

    We demonstrate dense SDM transmission of 20-WDM multi-carrier PDM-32QAM signals over a 40-km 12-core x 3-mode fiber with 247.9-b/s/Hz spectral efficiency. Parallel MIMO equalization enables 21-ns DMD compensation with 61 TDE taps per subcarrier....

  6. Marine sediment cores database for the Mediterranean Basin: a tool for past climatic and environmental studies

    Science.gov (United States)

    Alberico, I.; Giliberti, I.; Insinga, D. D.; Petrosino, P.; Vallefuoco, M.; Lirer, F.; Bonomo, S.; Cascella, A.; Anzalone, E.; Barra, R.; Marsella, E.; Ferraro, L.

    2017-06-01

    Paleoclimatic data are essential for fingerprinting the climate of the earth before the advent of modern recording instruments. They enable us to recognize past climatic events and predict future trends. Within this framework, a conceptual and logical model was drawn to physically implement a paleoclimatic database named WDB-Paleo that includes the paleoclimatic proxies data of marine sediment cores of the Mediterranean Basin. Twenty entities were defined to record four main categories of data: a) the features of oceanographic cruises and cores (metadata); b) the presence/absence of paleoclimatic proxies pulled from about 200 scientific papers; c) the quantitative analysis of planktonic and benthonic foraminifera, pollen, calcareous nannoplankton, magnetic susceptibility, stable isotopes, radionuclides values of about 14 cores recovered by Institute for Coastal Marine Environment (IAMC) of Italian National Research Council (CNR) in the framework of several past research projects; d) specific entities recording quantitative data on δ18O, AMS 14C (Accelerator Mass Spectrometry) and tephra layers available in scientific papers. Published data concerning paleoclimatic proxies in the Mediterranean Basin are recorded only for 400 out of 6000 cores retrieved in the area and they show a very irregular geographical distribution. Moreover, the data availability decreases when a constrained time interval is investigated or more than one proxy is required. We present three applications of WDB-Paleo for the Younger Dryas (YD) paleoclimatic event at Mediterranean scale and point out the potentiality of this tool for integrated stratigraphy studies.

  7. Impact of climate fluctuations on deposition of DDT and hexachlorocyclohexane in mountain glaciers: Evidence from ice core records

    International Nuclear Information System (INIS)

    Wang Xiaoping; Gong Ping; Zhang, Qianggong; Yao Tandong

    2010-01-01

    How do climate fluctuations affect DDT and hexachlorocyclohexane (HCH) distribution in the global scale? In this study, the interactions between climate variations and depositions of DDT and HCH in ice cores from Mt. Everest (the Tibetan Plateau), Mt. Muztagata (the eastern Pamirs) and the Rocky Mountains were investigated. All data regarding DDT/HCH deposition were obtained from the published results. Concentrations of DDT and HCH in an ice core from Mt. Everest were associated with the El Nino-Southern Oscillation. Concentrations of DDT in an ice core from Mt. Muztagata were significantly correlated with the Siberia High pattern. Concentrations of HCH in an ice core from Snow Dome of the Rocky Mountains responded to the North Atlantic Oscillation. These associations suggested that there are some linkages between climate variations and the global distribution of persistent organic pollutants. - Our study approves the potential contribution of ice core records of POPs to transport mechanisms of POPs.

  8. Three-dimensional gyrokinetic particle-in-cell simulation of plasmas on a massively parallel computer: Final report on LDRD Core Competency Project, FY 1991--FY 1993

    International Nuclear Information System (INIS)

    Byers, J.A.; Williams, T.J.; Cohen, B.I.; Dimits, A.M.

    1994-01-01

    One of the programs of the Magnetic fusion Energy (MFE) Theory and computations Program is studying the anomalous transport of thermal energy across the field lines in the core of a tokamak. We use the method of gyrokinetic particle-in-cell simulation in this study. For this LDRD project we employed massively parallel processing, new algorithms, and new algorithms, and new formal techniques to improve this research. Specifically, we sought to take steps toward: researching experimentally-relevant parameters in our simulations, learning parallel computing to have as a resource for our group, and achieving a 100 x speedup over our starting-point Cray2 simulation code's performance

  9. High performance statistical computing with parallel R: applications to biology and climate modelling

    International Nuclear Information System (INIS)

    Samatova, Nagiza F; Branstetter, Marcia; Ganguly, Auroop R; Hettich, Robert; Khan, Shiraj; Kora, Guruprasad; Li, Jiangtian; Ma, Xiaosong; Pan, Chongle; Shoshani, Arie; Yoginath, Srikanth

    2006-01-01

    Ultrascale computing and high-throughput experimental technologies have enabled the production of scientific data about complex natural phenomena. With this opportunity, comes a new problem - the massive quantities of data so produced. Answers to fundamental questions about the nature of those phenomena remain largely hidden in the produced data. The goal of this work is to provide a scalable high performance statistical data analysis framework to help scientists perform interactive analyses of these raw data to extract knowledge. Towards this goal we have been developing an open source parallel statistical analysis package, called Parallel R, that lets scientists employ a wide range of statistical analysis routines on high performance shared and distributed memory architectures without having to deal with the intricacies of parallelizing these routines

  10. Climatic changes on orbital and sub-orbital time scale recorded by the Guliya ice core in Tibetan Plateau

    Institute of Scientific and Technical Information of China (English)

    姚檀栋; 徐柏青; 蒲健辰

    2001-01-01

    Based on ice core records in the Tibetan Plateau and Greenland, the features and possible causes of climatic changes on orbital and sub-orbital time scale were discussed. Orbital time scale climatic change recorded in ice core from the Tibetan Plateau is typically ahead of that from polar regions, which indicates that climatic change in the Tibetan Plateau might be earlier than polar regions. The solar radiation change is a major factor that dominates the climatic change on orbital time scale. However, climatic events on sub-orbital time scale occurred later in the Tibetan Plateau than in the Arctic Region, indicating a different mechanism. For example, the Younger Dryas and Heinrich events took place earlier in Greenland ice core record than in Guliya ice core record. It is reasonable to propose the hypothesis that these climatic events were affected possibly by the Laurentide Ice Sheet. Therefore, ice sheet is critically important to climatic change on sub-orbital time scale in some ice ages.

  11. Dead Sea deep cores: A window into past climate and seismicity

    Science.gov (United States)

    Stein, Mordechai; Ben-Avraham, Zvi; Goldstein, Steven L.

    2011-12-01

    The area surrounding the Dead Sea was the locus of humankind's migration out of Africa and thus has been the home of peoples since the Stone Age. For this reason, understanding the climate and tectonic history of the region provides valuable insight into archaeology and studies of human history and helps to gain a better picture of future climate and tectonic scenarios. The deposits at the bottom of the Dead Sea are a geological archive of the environmental conditions (e.g., rains, floods, dust storms, droughts) during ice ages and warm ages, as well as of seismic activity in this key region. An International Continental Scientific Drilling Program (ICDP) deep drilling project was performed in the Dead Sea between November 2010 and March 2011. The project was funded by the ICDP and agencies in Israel, Germany, Japan, Norway, Switzerland, and the United States. Drilling was conducted using the new Large Lake Drilling Facility (Figure 1), a barge with a drilling rig run by DOSECC, Inc. (Drilling, Observation and Sampling of the Earth's Continental Crust), a nonprofit corporation dedicated to advancing scientific drilling worldwide. The main purpose of the project was to recover a long, continuous core to provide a high resolution record of the paleoclimate, paleoenvironment, paleoseismicity, and paleomagnetism of the Dead Sea Basin. With this, scientists are beginning to piece together a record of the climate and seismic history of the Middle East during the past several hundred thousand years in millennial to decadal to annual time resolution.

  12. Parallel responses of bees to Pleistocene climate change in three isolated archipelagos of the southwestern Pacific.

    Science.gov (United States)

    Groom, Scott V C; Stevens, Mark I; Schwarz, Michael P

    2014-06-22

    The impacts of glacial cycles on the geographical distribution and size of populations have been explored for numerous terrestrial and marine taxa. However, most studies have focused on high latitudes, with only a few focused on the response of biota to the last glacial maximum (LGM) in equatorial regions. Here, we examine how population sizes of key bee fauna in the southwest Pacific archipelagos of Fiji, Vanuatu and Samoa have fluctuated over the Quaternary. We show that all three island faunas suffered massive population declines, roughly corresponding in time to the LGM, followed by rapid expansion post-LGM. Our data therefore suggest that Pleistocene climate change has had major impacts across a very broad tropical region. While other studies indicate widespread Holarctic effects of the LGM, our data suggest a much wider range of latitudes, extending to the tropics, where these climate change repercussions were important. As key pollinators, the inferred changes in these bee faunas may have been critical in the development of the diverse Pacific island flora. The magnitude of these responses indicates future climate change scenarios may have alarming consequences for Pacific island systems involving pollinator-dependent plant communities and agricultural crops.

  13. 1500 Years of Annual Climate and Environmental Variability as Recorded in Bona-Churchill (Alaska) Ice Cores

    Science.gov (United States)

    Thompson, L. G.; Mosley-Thompson, E. S.; Zagorodnov, V.; Davis, M. E.; Mashiotta, T. A.; Lin, P.

    2004-12-01

    In 2003, six ice cores measuring 10.5, 11.5, 11.8, 12.4, 114 and 460 meters were recovered from the col between Mount Bona and Mount Churchill (61° 24'N; 141° 42'W; 4420 m asl). These cores have been analyzed for stable isotopic ratios, insoluble dust content and concentrations of major chemical species. Total Beta radioactivity was measured in the upper sections. The 460-meter core, extending to bedrock, captured the entire depositional record at this site where ice temperatures ranged from -24° C at 10 meters to -19.8° C at the ice/bedrock contact. The shallow cores allow assessment of surface processes under modern meteorological conditions while the deep core offers a ˜1500-year climate and environmental perspective. The average annual net balance is ˜~1000 mm of water equivalent and distinct annual signals in dust and calcium concentrations along with δ 18O allow annual resolution over most of the core. The excess sulfate record reflects many known large volcanic eruptions such as Katmai, Krakatau, Tambora, and Laki which allow validation of the time scale in the upper part of the core. The lower part of the core yields a history of earlier volcanic events. The 460-m Bona-Churchill ice core provides a detailed history of the `Little Ice Age' and medieval warm periods for southeastern Alaska. The source of the White River Ash will be discussed in light of the evidence from this core. The 460-m core also provides a long-term history of the dust fall that originates in north-central China. The annual ice core-derived climate records from southeastern Alaska will facilitate an investigation of the likelihood that the high resolution 1500-year record from the tropical Quelccaya Ice Cap (Peru) preserves a history of the variability of both the PDO and the Aleutian Low.

  14. Hybrid MPI/OpenMP parallelization of the explicit Volterra integral equation solver for multi-core computer architectures

    KAUST Repository

    Al Jarro, Ahmed

    2011-08-01

    A hybrid MPI/OpenMP scheme for efficiently parallelizing the explicit marching-on-in-time (MOT)-based solution of the time-domain volume (Volterra) integral equation (TD-VIE) is presented. The proposed scheme equally distributes tested field values and operations pertinent to the computation of tested fields among the nodes using the MPI standard; while the source field values are stored in all nodes. Within each node, OpenMP standard is used to further accelerate the computation of the tested fields. Numerical results demonstrate that the proposed parallelization scheme scales well for problems involving three million or more spatial discretization elements. © 2011 IEEE.

  15. Finding Tropical Cyclones on a Cloud Computing Cluster: Using Parallel Virtualization for Large-Scale Climate Simulation Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Hasenkamp, Daren; Sim, Alexander; Wehner, Michael; Wu, Kesheng

    2010-09-30

    Extensive computing power has been used to tackle issues such as climate changes, fusion energy, and other pressing scientific challenges. These computations produce a tremendous amount of data; however, many of the data analysis programs currently only run a single processor. In this work, we explore the possibility of using the emerging cloud computing platform to parallelize such sequential data analysis tasks. As a proof of concept, we wrap a program for analyzing trends of tropical cyclones in a set of virtual machines (VMs). This approach allows the user to keep their familiar data analysis environment in the VMs, while we provide the coordination and data transfer services to ensure the necessary input and output are directed to the desired locations. This work extensively exercises the networking capability of the cloud computing systems and has revealed a number of weaknesses in the current cloud system software. In our tests, we are able to scale the parallel data analysis job to a modest number of VMs and achieve a speedup that is comparable to running the same analysis task using MPI. However, compared to MPI based parallelization, the cloud-based approach has a number of advantages. The cloud-based approach is more flexible because the VMs can capture arbitrary software dependencies without requiring the user to rewrite their programs. The cloud-based approach is also more resilient to failure; as long as a single VM is running, it can make progress while as soon as one MPI node fails the whole analysis job fails. In short, this initial work demonstrates that a cloud computing system is a viable platform for distributed scientific data analyses traditionally conducted on dedicated supercomputing systems.

  16. Finding Tropical Cyclones on a Cloud Computing Cluster: Using Parallel Virtualization for Large-Scale Climate Simulation Analysis

    International Nuclear Information System (INIS)

    Hasenkamp, Daren; Sim, Alexander; Wehner, Michael; Wu, Kesheng

    2010-01-01

    Extensive computing power has been used to tackle issues such as climate changes, fusion energy, and other pressing scientific challenges. These computations produce a tremendous amount of data; however, many of the data analysis programs currently only run a single processor. In this work, we explore the possibility of using the emerging cloud computing platform to parallelize such sequential data analysis tasks. As a proof of concept, we wrap a program for analyzing trends of tropical cyclones in a set of virtual machines (VMs). This approach allows the user to keep their familiar data analysis environment in the VMs, while we provide the coordination and data transfer services to ensure the necessary input and output are directed to the desired locations. This work extensively exercises the networking capability of the cloud computing systems and has revealed a number of weaknesses in the current cloud system software. In our tests, we are able to scale the parallel data analysis job to a modest number of VMs and achieve a speedup that is comparable to running the same analysis task using MPI. However, compared to MPI based parallelization, the cloud-based approach has a number of advantages. The cloud-based approach is more flexible because the VMs can capture arbitrary software dependencies without requiring the user to rewrite their programs. The cloud-based approach is also more resilient to failure; as long as a single VM is running, it can make progress while as soon as one MPI node fails the whole analysis job fails. In short, this initial work demonstrates that a cloud computing system is a viable platform for distributed scientific data analyses traditionally conducted on dedicated supercomputing systems.

  17. Hybrid MPI/OpenMP parallelization of the explicit Volterra integral equation solver for multi-core computer architectures

    KAUST Repository

    Al Jarro, Ahmed; Bagci, Hakan

    2011-01-01

    A hybrid MPI/OpenMP scheme for efficiently parallelizing the explicit marching-on-in-time (MOT)-based solution of the time-domain volume (Volterra) integral equation (TD-VIE) is presented. The proposed scheme equally distributes tested field values

  18. Experiences Using Hybrid MPI/OpenMP in the Real World: Parallelization of a 3D CFD Solver for Multi-Core Node Clusters

    Directory of Open Access Journals (Sweden)

    Gabriele Jost

    2010-01-01

    Full Text Available Today most systems in high-performance computing (HPC feature a hierarchical hardware design: shared-memory nodes with several multi-core CPUs are connected via a network infrastructure. When parallelizing an application for these architectures it seems natural to employ a hierarchical programming model such as combining MPI and OpenMP. Nevertheless, there is the general lore that pure MPI outperforms the hybrid MPI/OpenMP approach. In this paper, we describe the hybrid MPI/OpenMP parallelization of IR3D (Incompressible Realistic 3-D code, a full-scale real-world application, which simulates the environmental effects on the evolution of vortices trailing behind control surfaces of underwater vehicles. We discuss performance, scalability and limitations of the pure MPI version of the code on a variety of hardware platforms and show how the hybrid approach can help to overcome certain limitations.

  19. Past climate changes derived from isotope measurements in polar ice cores

    International Nuclear Information System (INIS)

    Beer, J.; Muscheler, R.; Wagner, G.; Kubik, P.K.

    2002-01-01

    Measurements of stable and radioactive isotopes in polar ice cores provide a wealth of information on the climate conditions of the past. Stable isotopes (δ 18 O, δD) reflect mainly the temperature, whereas δ 18 O of oxygen in air bubbles reveals predominantly the global ice volume and the biospheric activity. Cosmic ray produced radioisotopes (cosmogenic nuclides) such as 10 Be and 36 Cl record information on the solar variability and possibly also on the solar irradiance. If the flux of a cosmogenic nuclide into the ice is known the accumulation rate can be derived from the measured concentration. The comparison of 10 Be from ice with 14 C from tree rings allows deciding whether observed 14 C variations are caused by production or system effects. Finally, isotope measurements are very useful for establishing and improving time scales. The 10 Be/ 36 Cl ratio changes with an apparent half-life of 376,000 years and is therefore well suited to date old ice. Significant abrupt changes in the records of 10 Be, 36 Cl from ice and of δ 18 O from atmospheric oxygen representing global signals can be used to synchronize ice and sediment cores. (author)

  20. Characterization of rapid climate changes through isotope analyses of ice and entrapped air in the NEEM ice core

    DEFF Research Database (Denmark)

    Guillevic, Myriam

    Greenland ice core have revealed the occurrence of rapid climatic instabilities during the last glacial period, known as Dansgaard-Oeschger (DO) events, while marine cores from the North Atlantic have evidenced layers of ice rafted debris deposited by icebergs melt, caused by the collapse...... mechanisms at play. Recent analytical developments have made possible to measure new paleoclimate proxies in Greenland ice cores. In this thesis we first contribute to these analytical developments by measuring the new innovative parameter 17O-excess at LSCE (Laboratoire des Sciences du Climatet de l......'Environnement, France). At the Centre for Ice and Climate (CIC, Denmark) we contribute to the development of a protocol for absolute referencing of methane gas isotopes, and making full air standard with known concentration and isotopic composition of methane. Then, air (δ15N) and water stable isotope measurements from...

  1. Optimization and parallelization of B-spline based orbital evaluations in QMC on multi/many-core shared memory processors

    OpenAIRE

    Mathuriya, Amrita; Luo, Ye; Benali, Anouar; Shulenburger, Luke; Kim, Jeongnim

    2016-01-01

    B-spline based orbital representations are widely used in Quantum Monte Carlo (QMC) simulations of solids, historically taking as much as 50% of the total run time. Random accesses to a large four-dimensional array make it challenging to efficiently utilize caches and wide vector units of modern CPUs. We present node-level optimizations of B-spline evaluations on multi/many-core shared memory processors. To increase SIMD efficiency and bandwidth utilization, we first apply data layout transfo...

  2. Trends in historical mercury deposition inferred from lake sediment cores across a climate gradient in the Canadian High Arctic.

    Science.gov (United States)

    Korosi, Jennifer B; Griffiths, Katherine; Smol, John P; Blais, Jules M

    2018-06-02

    Recent climate change may be enhancing mercury fluxes to Arctic lake sediments, confounding the use of sediment cores to reconstruct histories of atmospheric deposition. Assessing the independent effects of climate warming on mercury sequestration is challenging due to temporal overlap between warming temperatures and increased long-range transport of atmospheric mercury following the Industrial Revolution. We address this challenge by examining mercury trends in short cores (the last several hundred years) from eight lakes centered on Cape Herschel (Canadian High Arctic) that span a gradient in microclimates, including two lakes that have not yet been significantly altered by climate warming due to continued ice cover. Previous research on subfossil diatoms and inferred primary production indicated the timing of limnological responses to climate warming, which, due to prevailing ice cover conditions, varied from ∼1850 to ∼1990 for lakes that have undergone changes. We show that climate warming may have enhanced mercury deposition to lake sediments in one lake (Moraine Pond), while another (West Lake) showed a strong signal of post-industrial mercury enrichment without any corresponding limnological changes associated with warming. Our results provide insights into the role of climate warming and organic carbon cycling as drivers of mercury deposition to Arctic lake sediments. Copyright © 2018 Elsevier Ltd. All rights reserved.

  3. Application of the distributed genetic algorithm for in-core fuel optimization problems under parallel computational environment

    International Nuclear Information System (INIS)

    Yamamoto, Akio; Hashimoto, Hiroshi

    2002-01-01

    The distributed genetic algorithm (DGA) is applied for loading pattern optimization problems of the pressurized water reactors. A basic concept of DGA follows that of the conventional genetic algorithm (GA). However, DGA equally distributes candidates of solutions (i.e. loading patterns) to several independent ''islands'' and evolves them in each island. Communications between islands, i.e. migrations of some candidates between islands are performed with a certain period. Since candidates of solutions independently evolve in each island while accepting different genes of migrants, premature convergence in the conventional GA can be prevented. Because many candidate loading patterns should be evaluated in GA or DGA, the parallelization is efficient to reduce turn around time. Parallel efficiency of DGA was measured using our optimization code and good efficiency was attained even in a heterogeneous cluster environment due to dynamic distribution of the calculation load. The optimization code is based on the client/server architecture with the TCP/IP native socket and a client (optimization) module and calculation server modules communicate the objects of loading patterns each other. Throughout the sensitivity study on optimization parameters of DGA, a suitable set of the parameters for a test problem was identified. Finally, optimization capability of DGA and the conventional GA was compared in the test problem and DGA provided better optimization results than the conventional GA. (author)

  4. Diatom Stratigraphy of FA-1 Core, Qarun Lake, Records of Holocene Environmental and Climatic Change in Faiyum Oasis, Egypt

    Directory of Open Access Journals (Sweden)

    Zalat Abdelfattah A.

    2017-06-01

    Full Text Available This study evaluates changes in the environmental and climatic conditions in the Faiyum Oasis during the Holocene based on diatom analyses of the sediment FA-1 core from the southern seashore of the Qarun Lake. The studied FA-1 core was 26 m long and covered the time span ca. 9.000 cal. yrs BP. Diatom taxa were abundant and moderately to well-preserved throughout the core sediments. Planktonic taxa were most abundant than the benthic and epiphytic forms, which were very rare and sparsely distributed. The most dominant planktonic genera were Aulacoseira and Stephanodiscus followed by frequently distribution of Cyclostephanos and Cyclotella species. The stratigraphic distribution patterns of the recorded diatoms through the Holocene sediments explained five ecological diatom groups. These groups represent distinctive environmental conditions, which were mainly related to climatic changes through the early and middle Holocene, in addition to anthropogenic activity during the late Holocene. Comparison of diatom assemblages in the studied sediment core suggests that considerable changes occurred in water level as well as salinity. There were several high stands of the freshwater lake level during humid, warmer-wet climatic phases marked by dominance of planktonic, oligohalobous and alkaliphilous diatoms alternated with lowering of the lake level and slight increases in salinity and alkalinity during warm arid conditions evident by prevalence of brackish water diatoms.

  5. Millennial and sub-millennial scale climatic variations recorded in polar ice cores over the last glacial period

    DEFF Research Database (Denmark)

    Capron, E.; Landais, A.; Chappellaz, J.

    2010-01-01

    Since its discovery in Greenland ice cores, the millennial scale climatic variability of the last glacial period has been increasingly documented at all latitudes with studies focusing mainly on Marine Isotopic Stage 3 (MIS 3; 28–60 thousand of years before present, hereafter ka) and characterized...... that when ice sheets are extensive, Antarctica does not necessarily warm during the whole GS as the thermal bipolar seesaw model would predict, questioning the Greenland ice core temperature records as a proxy for AMOC changes throughout the glacial period....

  6. Parallel transport studies of high-Z impurities in the core of Alcator C-Mod plasmas

    Energy Technology Data Exchange (ETDEWEB)

    Reinke, M. L.; Hutchinson, I. H.; Rice, J. E.; Greenwald, M.; Howard, N. T.; Hubbard, A.; Hughes, J. W.; Terry, J. L.; Wolfe, S. M. [MIT-Plasma Science and Fusion Center Cambridge, Massachusetts 02139 (United States)

    2013-05-15

    Measurements of poloidal variation, ñ{sub z}/, in high-Z impurity density have been made using photodiode arrays sensitive to vacuum ultraviolet and soft x-ray emission in Alcator C-Mod plasmas. In/out asymmetries in the range of −0.2<0.3 are observed for r/a<0.8, and accumulation on both the high-field side, n{sub z,cos}<0, and low-field side, n{sub z,cos}>0, of a flux surface is found to be well described by a combination of centrifugal, poloidal electric field, and ion-impurity friction effects. Up/down asymmetries, −0.05<0.10, are observed over 0.50 corresponding to accumulation opposite the ion ∇B drift direction. Measurements of the up/down asymmetry of molybdenum are found to disagree with predictions from recent neoclassical theory in the trace limit, n{sub z}Z{sup 2}/n{sub i}≪1. Non-trace levels of impurities are expected to modify the main-ion poloidal flow and thus change friction-driven impurity density asymmetries and impurity poloidal rotation, v{sub θ,z}. Artificially modifying main-ion flow in parallel transport simulations is shown to impact both ñ{sub z}/ and v{sub θ,z}, but simultaneous agreement between measured and predicted up/down and in/out asymmetry as well as impurity poloidal rotation is not possible for these C-Mod data. This link between poloidal flow and poloidal impurity density variation outlines a more stringent test for parallel neoclassical transport theory than has previously been performed. Measurement and computational techniques specific to the study of poloidal impurity asymmetry physics are discussed as well.

  7. Performance and advantages of a soft-core based parallel architecture for energy peak detection in the calorimeter Level 0 trigger for the NA62 experiment at CERN

    International Nuclear Information System (INIS)

    Ammendola, R.; Barbanera, M.; Bizzarri, M.; Bonaiuto, V.; Ceccucci, A.; Simone, N. De; Fantechi, R.; Fucci, A.; Lupi, M.; Ryjov, V.; Checcucci, B.; Papi, A.; Piccini, M.; Federici, L.; Paoluzzi, G.; Salamon, A.; Salina, G.; Sargeni, F.; Venditti, S.

    2017-01-01

    The NA62 experiment at CERN SPS has started its data-taking. Its aim is to measure the branching ratio of the ultra-rare decay K +  → π + ν ν̅ . In this context, rejecting the background is a crucial topic. One of the main background to the measurement is represented by the K +  → π + π 0  decay. In the 1-8.5 mrad decay region this background is rejected by the calorimetric trigger processor (Cal-L0). In this work we present the performance of a soft-core based parallel architecture built on FPGAs for the energy peak reconstruction as an alternative to an implementation completely founded on VHDL language.

  8. Performance and advantages of a soft-core based parallel architecture for energy peak detection in the calorimeter Level 0 trigger for the NA62 experiment at CERN

    Science.gov (United States)

    Ammendola, R.; Barbanera, M.; Bizzarri, M.; Bonaiuto, V.; Ceccucci, A.; Checcucci, B.; De Simone, N.; Fantechi, R.; Federici, L.; Fucci, A.; Lupi, M.; Paoluzzi, G.; Papi, A.; Piccini, M.; Ryjov, V.; Salamon, A.; Salina, G.; Sargeni, F.; Venditti, S.

    2017-03-01

    The NA62 experiment at CERN SPS has started its data-taking. Its aim is to measure the branching ratio of the ultra-rare decay K+ → π+ν ν̅ . In this context, rejecting the background is a crucial topic. One of the main background to the measurement is represented by the K+ → π+π0 decay. In the 1-8.5 mrad decay region this background is rejected by the calorimetric trigger processor (Cal-L0). In this work we present the performance of a soft-core based parallel architecture built on FPGAs for the energy peak reconstruction as an alternative to an implementation completely founded on VHDL language.

  9. McCall Glacier record of Arctic climate change: Interpreting a northern Alaska ice core with regional water isotopes

    Science.gov (United States)

    Klein, E. S.; Nolan, M.; McConnell, J.; Sigl, M.; Cherry, J.; Young, J.; Welker, J. M.

    2016-01-01

    We explored modern precipitation and ice core isotope ratios to better understand both modern and paleo climate in the Arctic. Paleoclimate reconstructions require an understanding of how modern synoptic climate influences proxies used in those reconstructions, such as water isotopes. Therefore we measured periodic precipitation samples at Toolik Lake Field Station (Toolik) in the northern foothills of the Brooks Range in the Alaskan Arctic to determine δ18O and δ2H. We applied this multi-decadal local precipitation δ18O/temperature regression to ∼65 years of McCall Glacier (also in the Brooks Range) ice core isotope measurements and found an increase in reconstructed temperatures over the late-20th and early-21st centuries. We also show that the McCall Glacier δ18O isotope record is negatively correlated with the winter bidecadal North Pacific Index (NPI) climate oscillation. McCall Glacier deuterium excess (d-excess, δ2H - 8*δ18O) values display a bidecadal periodicity coherent with the NPI and suggest shifts from more southwestern Bering Sea moisture sources with less sea ice (lower d-excess values) to more northern Arctic Ocean moisture sources with more sea ice (higher d-excess values). Northern ice covered Arctic Ocean McCall Glacier moisture sources are associated with weak Aleutian Low (AL) circulation patterns and the southern moisture sources with strong AL patterns. Ice core d-excess values significantly decrease over the record, coincident with warmer temperatures and a significant reduction in Alaska sea ice concentration, which suggests that ice free northern ocean waters are increasingly serving as terrestrial precipitation moisture sources; a concept recently proposed by modeling studies and also present in Greenland ice core d-excess values during previous transitions to warm periods. This study also shows the efficacy and importance of using ice cores from Arctic valley glaciers in paleoclimate reconstructions.

  10. Ice core records of climate variability on the Third Pole with emphasis on the Guliya ice cap, western Kunlun Mountains

    Science.gov (United States)

    Thompson, Lonnie G.; Yao, Tandong; Davis, Mary E.; Mosley-Thompson, Ellen; Wu, Guangjian; Porter, Stacy E.; Xu, Baiqing; Lin, Ping-Nan; Wang, Ninglian; Beaudon, Emilie; Duan, Keqin; Sierra-Hernández, M. Roxana; Kenny, Donald V.

    2018-05-01

    Records of recent climate from ice cores drilled in 2015 on the Guliya ice cap in the western Kunlun Mountains of the Tibetan Plateau, which with the Himalaya comprises the Third Pole (TP), demonstrate that this region has become warmer and moister since at least the middle of the 19th century. Decadal-scale linkages are suggested between ice core temperature and snowfall proxies, North Atlantic oceanic and atmospheric processes, Arctic temperatures, and Indian summer monsoon intensity. Correlations between annual-scale oxygen isotopic ratios (δ18O) and tropical western Pacific and Indian Ocean sea surface temperatures are also demonstrated. Comparisons of climate records during the last millennium from ice cores acquired throughout the TP illustrate centennial-scale differences between monsoon and westerlies dominated regions. Among these records, Guliya shows the highest rate of warming since the end of the Little Ice Age, but δ18O data over the last millennium from TP ice cores support findings that elevation-dependent warming is most pronounced in the Himalaya. This, along with the decreasing precipitation rates in the Himalaya region, is having detrimental effects on the cryosphere. Although satellite monitoring of glaciers on the TP indicates changes in surface area, only a few have been directly monitored for mass balance and ablation from the surface. This type of ground-based study is essential to obtain a better understanding of the rate of ice shrinkage on the TP.

  11. Peeking Below the Snow Surface to Explore Amundsen Sea Climate Variability and Locate Optimal Ice-Core Sites

    Science.gov (United States)

    Neff, P. D.; Fudge, T. J.; Medley, B.

    2016-12-01

    Observations over recent decades reveal rapid changes in ice shelves and fast-flowing grounded ice along the Amundsen Sea coast of the West Antarctic Ice Sheet (WAIS). Long-term perspectives on this ongoing ice loss are needed to address a central question of Antarctic research: how much and how fast will Antarctic ice-loss raise sea level? Ice cores can provide insight into past variability of the atmospheric (wind) forcing of regional ocean dynamics affecting ice loss. Interannual variability of snow accumulation on coastal ice domes grounded near or within ice shelves reflects local to regional atmospheric circulation near the ice-ocean interface. Records of snow accumulation inferred from shallow ice cores strongly correlate with reanalysis precipitation and pressure fields, but ice cores have not yet been retrieved along the Amundsen Sea coast. High-frequency airborne radar data (NASA Operation IceBridge), however, have been collected over this region and we demonstrate that these data accurately reflect annual stratigraphy in shallow snow and firn (1 to 2 decades of accumulation). This further validates the agreement between radar snow accumulation records and climate reanalysis products. We then explore regional climate controls on local snow accumulation through comparison with gridded reanalysis products, providing a preview of what information longer coastal ice core records may provide with respect to past atmospheric forcing of ocean circulation and WAIS ice loss.

  12. Multiproxy records of Holocene climate and glacier variability from sediment cores in the Cordillera Vilcabamba of southern Peru

    Science.gov (United States)

    Schweinsberg, A. D.; Licciardi, J. M.; Rodbell, D. T.; Stansell, N.; Tapia, P. M.

    2012-12-01

    Sediments contained in glacier-fed lakes and bogs provide continuous high-resolution records of glacial activity, and preserve multiproxy evidence of Holocene climate change. Tropical glacier fluctuations offer critical insight on regional paleoclimatic trends and controls, however, continuous sediment records of past tropical climates are limited. Recent cosmogenic 10Be surface exposure ages of moraine sequences in the Cordillera Vilcabamba of southern Peru (13°20'S latitude) reveal a glacial culmination during the early Holocene and a less extensive glaciation coincident with the Little Ice Age of the Northern Hemisphere. Here we supplement the existing 10Be moraine chronology with the first continuous records of multiproxy climate data in this mountain range from sediment cores recovered from bogs in direct stratigraphic contact with 10Be-dated moraines. Radiocarbon-dated sedimentological changes in a 2-meter long bog core reveal that the Holocene is characterized by alternating inorganic and organic-rich laminae, suggesting high-frequency climatic variability. Carbon measurements, bulk density, and bulk sedimentation rates are used to derive a record of clastic sediment flux that serves as a proxy indicator of former glacier activity. Preliminary analyses of the bog core reveal approximately 70 diatom taxa that indicate both rheophilic and lentic environments. Initial results show a general decrease in magnetic susceptibility and clastic flux throughout the early to mid-Holocene, which suggests an interval of deglaciation. An episode of high clastic flux from 3.8 to 2.0 ka may reflect a late Holocene glacial readvance. Volcanic glass fragments and an anomalous peak in magnetic susceptibility may correspond to the historical 1600 AD eruption of Huaynaputina. Ten new bog and lake sediment cores were collected during the 2012 field expedition and analytical measurements are underway. Ongoing efforts are focused on analyzing diatom assemblage data, developing

  13. Archive of Geosample Data and Information from the Ohio State University Byrd Polar and Climate Research Center (BPCRC) Sediment Core Repository

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The Byrd Polar and Climate Research Center (BPCRC) Sediment Core Repository operated by the Ohio State University is a partner in the Index to Marine and Lacustrine...

  14. Insights Into Deglacial Through Holocene Climate Variability At The Peru-Chile Margin From Very High Sedimentation Rate Marine Cores

    Science.gov (United States)

    Chazen, C.; Dejong, H.; Altabet, M.; Herbert, T.

    2007-12-01

    The Peru-Chile upwelling system is situated at the epicenter of the modern ENSO System. The high settling flux of organic materials and poor ventilation of subsurface waters makes the Peru upwelling system one of the world's three major oxygen minimum/denitrification zones (Codispoti and Christensen, 1985). Extremely high sedimentation rates and permanent hypoxic/anoxic subsurface waters create excellent conditions for the preservation of organic matter. Despite the significance of this region in regards to paleoceanography and paleoclimatology, relatively little work has been done to characterize past Peruvian climate because carbonate dissolution hinders the use of conventional paleoclimate methods and hiatuses frequently interrupt the record. However, using nitrogen isotopes and alkenone paleothermometry on multiple sediment cores from the Margin we have managed to overcome many of these challenges to create a nearly continuous SST (Uk`37), productivity (C37total), biogenic opal and denitrification (δN15) record from the LGM through the late Holocene. Remarkably, recent work has revealed an annually laminated core, which spans from 1.4-8.0ka uninterrupted, providing a unique window into Holocene climate variability. Modern-day upwelling induced climate at the Peru-Chile margin is characterized by cold temperatures (21.5°C) high productivity and strong denitrification, which has persisted since the mid Holocene (4ka). The mid Holocene also marks the beginning of a dramatic increase in seasonality and ENSO variability consistent with other tropical climate indicators. Climate variability in the Mid-early Holocene shows a distinctively different pattern from that of the late Holocene; unproductive warm temperatures persist through the early Holocene in what can be described as a permanent El Niño-like state. Early tropical warming occurred near 17ka along with an unprecedented increase in denitrification, which is decoupled from local productivity. Early onset

  15. Millennial and sub-millennial scale climatic variations recorded in polar ice cores over the last glacial period

    Directory of Open Access Journals (Sweden)

    E. Capron

    2010-06-01

    Full Text Available Since its discovery in Greenland ice cores, the millennial scale climatic variability of the last glacial period has been increasingly documented at all latitudes with studies focusing mainly on Marine Isotopic Stage 3 (MIS 3; 28–60 thousand of years before present, hereafter ka and characterized by short Dansgaard-Oeschger (DO events. Recent and new results obtained on the EPICA and NorthGRIP ice cores now precisely describe the rapid variations of Antarctic and Greenland temperature during MIS 5 (73.5–123 ka, a time period corresponding to relatively high sea level. The results display a succession of abrupt events associated with long Greenland InterStadial phases (GIS enabling us to highlight a sub-millennial scale climatic variability depicted by (i short-lived and abrupt warming events preceding some GIS (precursor-type events and (ii abrupt warming events at the end of some GIS (rebound-type events. The occurrence of these sub-millennial scale events is suggested to be driven by the insolation at high northern latitudes together with the internal forcing of ice sheets. Thanks to a recent NorthGRIP-EPICA Dronning Maud Land (EDML common timescale over MIS 5, the bipolar sequence of climatic events can be established at millennial to sub-millennial timescale. This shows that for extraordinary long stadial durations the accompanying Antarctic warming amplitude cannot be described by a simple linear relationship between the two as expected from the bipolar seesaw concept. We also show that when ice sheets are extensive, Antarctica does not necessarily warm during the whole GS as the thermal bipolar seesaw model would predict, questioning the Greenland ice core temperature records as a proxy for AMOC changes throughout the glacial period.

  16. Parallel Programming with Intel Parallel Studio XE

    CERN Document Server

    Blair-Chappell , Stephen

    2012-01-01

    Optimize code for multi-core processors with Intel's Parallel Studio Parallel programming is rapidly becoming a "must-know" skill for developers. Yet, where to start? This teach-yourself tutorial is an ideal starting point for developers who already know Windows C and C++ and are eager to add parallelism to their code. With a focus on applying tools, techniques, and language extensions to implement parallelism, this essential resource teaches you how to write programs for multicore and leverage the power of multicore in your programs. Sharing hands-on case studies and real-world examples, the

  17. 10Be and δ2H in polar ice cores as a probe of the solar variability's influence on climate

    International Nuclear Information System (INIS)

    Raisbeck, G.M.; Yiou, F.; Jouzel, J.; Domaine Univ., 38 - St-Martin-d'Heres; Petit, J.R.

    1990-01-01

    By using the technique of accelerator mass spectrometry, it is now possible to measure detailed profiles of cosmogenic (cosmic ray produced) 10 Be in polar ice cores. Recent work has demonstrated that these profiles contain information on solar activity, via its influence on the intensity of galactic cosmic rays arriving in the Earth's atmosphere. It has been known for some time that, as a result of temperature-dependent fractionation effects, the stable isotope profiles δ 2 O and δ 2 H in polar ice cores contain palaeoclimate information. Thus by comparing the 10 Be and stable isotope profiles in the same ice core, one can test the influence of solar variability on climate, and this independent of possible uncertainties in the absolute chronology of the records. We present here the results of such a comparison for two Antarctic ice cores; one from the South Pole, covering the past ca. 1000 years, and one from Dome C, covering the past ca. 3000 years. (author)

  18. Late holocene ice core records of climate and environment from the Tropical Andes, Peru

    Directory of Open Access Journals (Sweden)

    1995-01-01

    relacionado con el “Optimum Medieval”. El registro de polvo del Huascarán, de nivel constante durante los 3 000 años, fue interrumpido por un evento de alta concentración de polvo entre 2 000 y 1 800 años BP (0-200 A.D. centrado en los años 1 900s BP (100 A.D.. Picos menos marcados se observan de 1 400 a 1 600 años BP (400 a 600 A.D. y de 1300 a 1030 años BP (700 a 960 A.D. El análisis del polvo asociado a este evento indica un material acumulado por los vientos de misma composición que la roca que constituyen la Cordillera Blanca (granodiórita. Los más recientes picos son parcialmente sincrónicos con el mayor evento de 400 a 620 A.D. encontrado en Quelccaya, lo que sugiere que este evento fue muy extenso. El más reciente evento de alta concentración de polvo registrado en Quelccaya y fechado de 830 a 960 A.D. permite observar bajas concentraciones en el Huascarán éste podría ser un argumento más para relacionar el principio de la emisión de polvo en Quelccaya con las actividades agrícolas en la cuenca del Titicaca (Thompson et al., 1988. A 1 500-year history of climatic and environmental variations from the Quelccaya ice cap (13°56’S, 70°50’W, 5 670 m a.s.l. is compared to a similar 3 000-year ice core record from the col of Huascarán (9°06’S, 77°36’W, 6 048 m a.s.l.. The parameters which are presented are oxygen isotopic ratios ( 18O, considered to be indicative of temperature, insoluble dust, and (for Huascarán only nitrate concentrations (NO3- which is indicative of vegetation fluctuations in the Amazon rainforest. The Huascarán  18O and NO3- profiles for the most recent 3 000 years show that there has been a general decrease in temperature along with a decrease in biological activity in the Amazon Basin, with the lowest values occurring during the “Little Ice Age” (LIA, 200 to 500 yrs BP. This was followed by an abrupt change in  18O, which increased to the levels of 3 000 years ago. This abrupt warming has

  19. Past climate change on Sky Islands drives novelty in a core developmental gene network and its phenotype.

    Science.gov (United States)

    Favé, Marie-Julie; Johnson, Robert A; Cover, Stefan; Handschuh, Stephan; Metscher, Brian D; Müller, Gerd B; Gopalan, Shyamalika; Abouheif, Ehab

    2015-09-04

    A fundamental and enduring problem in evolutionary biology is to understand how populations differentiate in the wild, yet little is known about what role organismal development plays in this process. Organismal development integrates environmental inputs with the action of gene regulatory networks to generate the phenotype. Core developmental gene networks have been highly conserved for millions of years across all animals, and therefore, organismal development may bias variation available for selection to work on. Biased variation may facilitate repeatable phenotypic responses when exposed to similar environmental inputs and ecological changes. To gain a more complete understanding of population differentiation in the wild, we integrated evolutionary developmental biology with population genetics, morphology, paleoecology and ecology. This integration was made possible by studying how populations of the ant species Monomorium emersoni respond to climatic and ecological changes across five 'Sky Islands' in Arizona, which are mountain ranges separated by vast 'seas' of desert. Sky Islands represent a replicated natural experiment allowing us to determine how repeatable is the response of M. emersoni populations to climate and ecological changes at the phenotypic, developmental, and gene network levels. We show that a core developmental gene network and its phenotype has kept pace with ecological and climate change on each Sky Island over the last ~90,000 years before present (BP). This response has produced two types of evolutionary change within an ant species: one type is unpredictable and contingent on the pattern of isolation of Sky lsland populations by climate warming, resulting in slight changes in gene expression, organ growth, and morphology. The other type is predictable and deterministic, resulting in the repeated evolution of a novel wingless queen phenotype and its underlying gene network in response to habitat changes induced by climate warming. Our

  20. Climatic changes inferred fron analyses of lake-sediment cores, Walker Lake, Nevada

    International Nuclear Information System (INIS)

    Yang, In Che.

    1989-01-01

    Organic and inorganic fractions of sediment collected from the bottom of Walker Lake, Nevada, have been dated by carbon-14 techniques. Sedimentation rates and the organic-carbon content of the sediment were correlated with climatic change. The cold climate between 25,000 and 21,000 years ago caused little runoff, snow accumulation on the mountains, and rapid substantial glacial advances; this period of cold climate resulted in a slow sedimentation rate (0.20 millimeter per year) and in a small organic-carbon content in the sediment. Also, organic-carbon accumulation rates in the lake during this period were slow. The most recent period of slow sedimentation rate and small organic-carbon content occurred between 10,000 and 5500 years ago, indicative of low lake stage and dry climatic conditions. This period of dry climate also was evidenced by dry conditions for Lake Lahontan in Nevada and Searles Lake in California, as cited in the literature. Walker Lake filled rapidly with water between 5500 and 4500 years ago. The data published in this report was not produced under an approved Site Investigation Plan (SIP) or Study Plan (SP) and will not be used in the licensing process. 10 refs., 3 figs., 2 tabs

  1. Parallel computation

    International Nuclear Information System (INIS)

    Jejcic, A.; Maillard, J.; Maurel, G.; Silva, J.; Wolff-Bacha, F.

    1997-01-01

    The work in the field of parallel processing has developed as research activities using several numerical Monte Carlo simulations related to basic or applied current problems of nuclear and particle physics. For the applications utilizing the GEANT code development or improvement works were done on parts simulating low energy physical phenomena like radiation, transport and interaction. The problem of actinide burning by means of accelerators was approached using a simulation with the GEANT code. A program of neutron tracking in the range of low energies up to the thermal region has been developed. It is coupled to the GEANT code and permits in a single pass the simulation of a hybrid reactor core receiving a proton burst. Other works in this field refers to simulations for nuclear medicine applications like, for instance, development of biological probes, evaluation and characterization of the gamma cameras (collimators, crystal thickness) as well as the method for dosimetric calculations. Particularly, these calculations are suited for a geometrical parallelization approach especially adapted to parallel machines of the TN310 type. Other works mentioned in the same field refer to simulation of the electron channelling in crystals and simulation of the beam-beam interaction effect in colliders. The GEANT code was also used to simulate the operation of germanium detectors designed for natural and artificial radioactivity monitoring of environment

  2. High-resolution Greenland Ice Core data show abrupt climate change happens in few years

    DEFF Research Database (Denmark)

    Steffensen, Jørgen Peder; Andersen, Katrine Krogh; Bigler, Matthias

    2008-01-01

    The last two abrupt warmings at the onset of our present warm interglacial period, interrupted by the Younger Dryas cooling event, were investigated at high temporal resolution from the North Greenland Ice Core Project ice core. The deuterium excess, a proxy of Greenland precipitation moisture...... source, switched mode within 1 to 3 years over these transitions and initiated a more gradual change (over 50 years) of the Greenland air temperature, as recorded by stable water isotopes. The onsets of both abrupt Greenland warmings were slightly preceded by decreasing Greenland dust deposition...

  3. Optimization of High-Resolution Continuous Flow Analysis for Transient Climate Signals in Ice Cores

    DEFF Research Database (Denmark)

    Bigler, Matthias; Svensson, Anders; Kettner, Ernesto

    2011-01-01

    Over the past two decades, continuous flow analysis (CFA) systems have been refined and widely used to measure aerosol constituents in polar and alpine ice cores in very high-depth resolution. Here we present a newly designed system consisting of sodium, ammonium, dust particles, and electrolytic...... meltwater conductivity detection modules. The system is optimized for high- resolution determination of transient signals in thin layers of deep polar ice cores. Based on standard measurements and by comparing sections of early Holocene and glacial ice from Greenland, we find that the new system features...

  4. A Multiproxy Approach to Unraveling Climate and Human Demography in the Peruvian Altiplano from a 5000 year Lake Sediment Core

    Science.gov (United States)

    Vaught-Mijares, R. M.; Hillman, A. L.; Abbott, M. B.; Werne, J. P.; Arkush, E.

    2017-12-01

    Drought and flood events are thought to have shaped the ways in which Andean societies have adapted to life in the Titicaca Basin region, particularly with regard to land use practices and settlement patterns. This study examines a small lake in the region, Laguna Orurillo. Water isotopes suggest that the lake primarily loses water through evaporation, making it hydrologically sensitive. In 2015, a 3.4 m overlapping sediment record was collected and inspected for evidence of shallow water facies and erosional unconformities to reconstruct paleohydrology. Sediment core chronology was established using 7 AMS radiocarbon dates and 210Pb dating and indicates that the core spans 5000 years. Additional sediment core measurements include magnetic susceptibility, bulk density, organic/carbonate content, and XRD. Results show a pronounced change in sediment composition from brittle, angular salt deposits to massive calcareous silt and clay around 5000 years BP. Multiple transitions from clay to sand show potential lake level depressions at 1540, 2090, and 2230, yr BP that are supported by a drastic increase in carbonate composition from 2760-1600 yr BP. Additional shallow-water periods may be reflected in the presence of rip-up clasts from 4000 to 3000 yr BP. These early interpretations align well with existing hydrologic records from Lake Titicaca. In order to develop a more detailed climate and land use record, isotope analyses of authigenic carbonate minerals using δ13C and δ18O and leaf waxes using δD are being developed. Ultimately, this record will be linked with records from nearby Lagunas Arapa and Umayo. Additional proxies for human population such as fecal 5β-stanols and proximal anthropologic surveys will be synthesized to contribute to a regional understanding of Holocene climate variability and human demography in the Peruvian Altiplano.

  5. High resolution climate reconstructions of recent warming using instrumental and ice core records from coastal Antarctica

    Digital Repository Service at National Institute of Oceanography (India)

    Thamban, M.; Naik, S.S.; Laluraj, C.M.; Ravindra, R.

    for the past 4 years (Thamban et al., 2006). Chronological control for the IND-25/B5 ice core was based on multiple and complimentary methods: (i) atomic bomb (Tritium) markers; (ii) annual layer counting using stable isotope records; and (iii) volcanic...

  6. Deuterium excess record in a southern Tibetan ice core and its potential climatic implications

    Energy Technology Data Exchange (ETDEWEB)

    Zhao, Huabiao; Xu, Baiqing; Yao, Tandong; Wu, Guangjian; Lin, Shubiao; Gao, Jing; Wang, Mo [Chinese Academy of Sciences, Key Laboratory of Tibetan Environment Changes and Land Surface Processes, Institute of Tibetan Plateau Research, Beijing (China)

    2012-05-15

    A 55-m long ice core, drilled close to bedrock from Mt. Noijin Kangsang on the southern Tibetan Plateau in summer 2007, was annually dated covering the period of 1864-2006 AD. The stable isotope ratios ({delta} {sup 18}O and {delta}D) of the ice core were measured and thereby the deuterium excess (d) was calculated by d = {delta}D - 8*{delta} {sup 18}O for the individual ice samples. Results show that the d values of the ice samples were predominantly controlled by the moisture sources. The significant increasing trend of annual mean d values along the ice core is mainly related to the rapid warming of the tropical Indian Ocean, although the tendency is subjected to the modulation by the western-derived moisture. The decreasing Indian monsoon precipitation on the southern Tibetan Plateau, physically linked with the increasing tropical Indian Ocean SST, reduced the share of monsoon precipitation in the annual total accumulation, making an additional contribution to the significant increase of annual mean d in the Noijin Kangsang ice core with high values during the past 143 years. (orig.)

  7. Climatic Cycles and Gradients of the El Niño Core Region in North Peru

    Directory of Open Access Journals (Sweden)

    Rütger Rollenbeck

    2015-01-01

    Full Text Available Climatic processes in northern Peru are evaluated on surface observation independent of modelling studies. The region is characterized by regular oscillations, but episodic El Niño-events introduce strong disturbances. Conceptual models based on observations, remote sensing data, and output of regional climate models are compared with data from a new station network. The results show regular oscillations of all climate variables on the annual and daily time scale. The daily cycle is probably associated with thermotidal forcings, causing gravity waves to emanate from the Andes Cordillera. Main factors are the interaction of large scale pressure systems like the Southeast Pacific High and the intertropical convergence zone (ITCZ. Also, there are regional factors: an extended sea-breeze system, the barrier-effect of the Andes, additional energy input by elevated radiation absorption at the mountain slopes, local wind systems, and the variations of the sea surface temperature. At the coast, a low-level jet works as a thermodynamic energy sink, suppressing deep convection and supporting the aridity. Those patterns are found in most of the station data and the processes of this climate can generally be confirmed. The overturning of this stable system with the onset of El Niño-conditions is possibly caused by disruptions of the regional circulation.

  8. Records of climatic changes and volcanic events in an ice core from ...

    Indian Academy of Sciences (India)

    R. Narasimhan (Krishtel eMaging) 1461 1996 Oct 15 13:05:22

    the volcanic event that occurred in 1815 AD, has been identified based on electrical conductance ... tions and accumulation rates of ice, climatic and ..... The peak saturated values of currents (µ amp) at about 5 and 30m depths identify the past volcanic episodes Augung ..... in promoting the scientific activities by allowing us.

  9. Patterns of volcanism, weathering, and climate history from high-resolution geochemistry of the BINGO core, Mono Lake, California, USA

    Science.gov (United States)

    Zimmerman, S. R.; Starratt, S.; Hemming, S. R.

    2012-12-01

    Mono Lake, California is a closed-basin lake on the east side of the Sierra Nevada, and inflow from snowmelt dominates the modern hydrology. Changes in wetness during the last glacial period (>12,000 years ago) and over the last 2,000 years have been extensively described, but are poorly known for the intervening period. We have recovered a 6.25 m-long core from ~3 m of water in the western embayment of Mono Lake, which is shown by initial radiocarbon dates to cover at least the last 10,000 years. The sediments of the core are variable, ranging from black to gray silts near the base, laminated olive-green silt through the center, to layers of peach-colored carbonate nodules interbedded with gray and olive silts and pea-green organic ooze. Volcanic tephras from Bodie and Adobe Hills to the north, east, and south. The rhyolitic tephras of the Mono-Inyo Craters are much lower in TiO2 than the bedrock (10,000 calibrated years before present (cal yr BP) higher in the core, and significant disruption of the fine layers, this interval likely indicates a relatively deep lake persisting into the early Holocene, after the initial dramatic regression from late Pleistocene levels. The finely laminated olive-green silt of the period ~10,700 to ~7500 cal yr BP is very homogenous chemically, probably indicating a stable, stratified lake and a relatively wet climate. This section merits mm-scale scanning and petrographic examination in the future. The upper boundary of the laminated section shows rising Ca/K and decreasing Ti and Si/K, marking the appearance of authigenic carbonate layers. After ~7500 cal yr BP, the sediment in BINGO becomes highly variable, with increased occurrence of tephra layers and carbonate, indicating a lower and more variable lake level. A short interval of olive-green, laminated fine sand/silt just above a radiocarbon date of 3870 ± 360 cal yr BP may record the Dechambeau Ranch highstand of Stine (1990; PPP v. 78 pp 333-381), and is marked by a distinct

  10. Ice Cores

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Records of past temperature, precipitation, atmospheric trace gases, and other aspects of climate and environment derived from ice cores drilled on glaciers and ice...

  11. Paleo-Climate and Glaciological Reconstruction in Central Asia through the Collection and Analysis of Ice Cores and Instrumental Data from the Tien Shan

    International Nuclear Information System (INIS)

    Vladimir Aizen; Donald Bren; Karl Kreutz; Cameron Wake

    2001-01-01

    While the majority of ice core investigations have been undertaken in the polar regions, a few ice cores recovered from carefully selected high altitude/mid-to-low latitude glaciers have also provided valuable records of climate variability in these regions. A regional array of high resolution, multi-parameter ice core records developed from temperate and tropical regions of the globe can be used to document regional climate and environmental change in the latitudes which are home to the vase majority of the Earth's human population. In addition, these records can be directly compared with ice core records available from the polar regions and can therefore expand our understanding of inter-hemispheric dynamics of past climate changes. The main objectives of our paleoclimate research in the Tien Shan mountains of middle Asia combine the development of detailed paleoenvironmental records via the physical and chemical analysis of ice cores with the analysis of modern meteorological and hydrological data. The first step in this research was the collection of ice cores from the accumulation zone of the Inylchek Glacier and the collection of meteorological data from a variety of stations throughout the Tien Shan. The research effort described in this report was part of a collaborative effort with the United State Geological Survey's (USGS) Global Environmental Research Program which began studying radionuclide deposition in mid-latitude glaciers in 1995

  12. Paleo-Climate and Glaciological Reconstruction in Central Asia through the Collection and Analysis of Ice Cores and Instrumental Data from the Tien Shan

    Energy Technology Data Exchange (ETDEWEB)

    Vladimir Aizen; Donald Bren; Karl Kreutz; Cameron Wake

    2001-05-30

    While the majority of ice core investigations have been undertaken in the polar regions, a few ice cores recovered from carefully selected high altitude/mid-to-low latitude glaciers have also provided valuable records of climate variability in these regions. A regional array of high resolution, multi-parameter ice core records developed from temperate and tropical regions of the globe can be used to document regional climate and environmental change in the latitudes which are home to the vase majority of the Earth's human population. In addition, these records can be directly compared with ice core records available from the polar regions and can therefore expand our understanding of inter-hemispheric dynamics of past climate changes. The main objectives of our paleoclimate research in the Tien Shan mountains of middle Asia combine the development of detailed paleoenvironmental records via the physical and chemical analysis of ice cores with the analysis of modern meteorological and hydrological data. The first step in this research was the collection of ice cores from the accumulation zone of the Inylchek Glacier and the collection of meteorological data from a variety of stations throughout the Tien Shan. The research effort described in this report was part of a collaborative effort with the United State Geological Survey's (USGS) Global Environmental Research Program which began studying radionuclide deposition in mid-latitude glaciers in 1995.

  13. Two Extreme Climate Events of the Last 1000 Years Recorded in Himalayan and Andean Ice Cores: Impacts on Humans

    Science.gov (United States)

    Thompson, L. G.; Mosley-Thompson, E. S.; Davis, M. E.; Kenny, D. V.; Lin, P.

    2013-12-01

    In the last few decades numerous studies have linked pandemic influenza, cholera, malaria, and viral pneumonia, as well as droughts, famines and global crises, to the El Niño-Southern Oscillation (ENSO). Two annually resolved ice core records, one from Dasuopu Glacier in the Himalaya and one from the Quelccaya Ice Cap in the tropical Peruvian Andes provide an opportunity to investigate these relationships on opposite sides of the Pacific Basin for the last 1000 years. The Dasuopu record provides an annual history from 1440 to 1997 CE and a decadally resolved record from 1000 to 1440 CE while the Quelccaya ice core provides annual resolution over the last 1000 years. Major ENSO events are often recorded in the oxygen isotope, insoluble dust, and chemical records from these cores. Here we investigate outbreaks of diseases, famines and global crises during two of the largest events recorded in the chemistry of these cores, particularly large peaks in the concentrations of chloride (Cl-) and fluoride (Fl-). One event is centered on 1789 to 1800 CE and the second begins abruptly in 1345 and tapers off after 1360 CE. These Cl- and F- peaks represent major droughts and reflect the abundance of continental atmospheric dust, derived in part from dried lake beds in drought stricken regions upwind of the core sites. For Dasuopu the likely sources are in India while for Quelccaya the sources would be the Andean Altiplano. Both regions are subject to drought conditions during the El Niño phase of the ENSO cycle. These two events persist longer (10 to 15 years) than today's typical ENSO events in the Pacific Ocean Basin. The 1789 to 1800 CE event was associated with a very strong El Niño event and was coincidental with the Boji Bara famine resulting from extended droughts that led to over 600,000 deaths in central India by 1792. Similarly extensive droughts are documented in Central and South America. Likewise, the 1345 to 1360 CE event, although poorly documented

  14. Stable water isotopes of precipitation and firn cores from the northern Antarctic Peninsula region as a proxy for climate reconstruction

    Directory of Open Access Journals (Sweden)

    F. Fernandoy

    2012-03-01

    Full Text Available In order to investigate the climate variability in the northern Antarctic Peninsula region, this paper focuses on the relationship between stable isotope content of precipitation and firn, and main meteorological variables (air temperature, relative humidity, sea surface temperature, and sea ice extent. Between 2008 and 2010, we collected precipitation samples and retrieved firn cores from several key sites in this region. We conclude that the deuterium excess oscillation represents a robust indicator of the meteorological variability on a seasonal to sub-seasonal scale. Low absolute deuterium excess values and the synchronous variation of both deuterium excess and air temperature imply that the evaporation of moisture occurs in the adjacent Southern Ocean. The δ18O-air temperature relationship is complicated and significant only at a (multiseasonal scale. Backward trajectory calculations show that air-parcels arriving at the region during precipitation events predominantly originate at the South Pacific Ocean and Bellingshausen Sea. These investigations will be used as a calibration for ongoing and future research in the area, suggesting that appropriate locations for future ice core research are located above 600 m a.s.l. We selected the Plateau Laclavere, Antarctic Peninsula as the most promising site for a deeper drilling campaign.

  15. Resolving climate change in the period 15-23 ka in Greenland ice cores: A new application of spectral trend analysis

    NARCIS (Netherlands)

    de Jong, M.G.G.; Nio, D.S.; Böhm, A.R.; Seijmonsbergen, H.C.; de Graaff, L.W.S.

    2009-01-01

    Northern Hemisphere climate history through and following the Last Glacial Maximum is recorded in detail in ice cores from Greenland. However, the period between Greenland Interstadials 1 and 2 (15-23 ka), i.e. the period of deglaciation following the last major glaciation, has been difficult to

  16. Modelling the regional climate and isotopic composition of Svalbard precipitation using REMOiso: a comparison with available GNIP and ice core data

    NARCIS (Netherlands)

    Divine, D.V.; Sjolte, J.; Isaksson, E.; Meijer, H.A.J.; van de Wal, R.S.W.; Martma, T.; Pohjola, V.; Sturm, C.; Godtliebsen, F.

    2011-01-01

    Simulations of a regional (approx. 50 km resolution) circulation model REMOiso with embedded stable water isotope module covering the period 1958-2001 are compared with the two instrumental climate and four isotope series (δ18O) from western Svalbard. We examine the data from ice cores drilled on

  17. Modelling the regional climate and isotopic composition of Svalbard precipitation using REMOiso : a comparison with available GNIP and ice core data

    NARCIS (Netherlands)

    Divine, D. V.; Sjolte, J.; Isaksson, E.; Meijer, H. A. J.; van de Wal, R. S. W.; Martma, T.; Pohjola, V.; Sturm, C.; Godtliebsen, F.

    2011-01-01

    Simulations of a regional (approx. 50 km resolution) circulation model REMOiso with embedded stable water isotope module covering the period 1958-2001 are compared with the two instrumental climate and four isotope series (d18O) from western Svalbard. We examine the data from ice cores drilled on

  18. Unveiling exceptional Baltic bog ecohydrology, autogenic succession and climate change during the last 2000 years in CE Europe using replicate cores, multi-proxy data and functional traits of testate amoebae

    Science.gov (United States)

    Gałka, Mariusz; Tobolski, Kazimierz; Lamentowicz, Łukasz; Ersek, Vasile; Jassey, Vincent E. J.; van der Knaap, Willem O.; Lamentowicz, Mariusz

    2017-01-01

    We present the results of high-resolution, multi-proxy palaeoecological investigations of two parallel peat cores from the Baltic raised bog Mechacz Wielki in NE Poland. We aim to evaluate the role of regional climate and autogenic processes of the raised bog itself in driving the vegetation and hydrology dynamics. Based on partly synchronous changes in Sphagnum communities in the two study cores we suggest that extrinsic factors (climate) played an important role as a driver in mire development during the bog stage (500-2012 CE). Using a testate amoebae transfer function, we found exceptionally stable hydrological conditions during the last 2000 years with a relatively high water table and lack of local fire events that allowed for rapid peat accumulation (2.75 mm/year) in the bog. Further, the strong correlation between pH and community-weighted mean of testate amoeba traits suggests that other variables than water-table depth play a role in driving microbial properties under stable hydrological conditions. There is a difference in hydrological dynamics in bogs between NW and NE Poland until ca 1500 CE, after which the water table reconstructions show more similarities. Our results illustrate how various functional traits relate to different environmental variables in a range of trophic and hydrological scenarios on long time scales. Moreover, our data suggest a common regional climatic forcing in Mechacz Wielki, Gązwa and Kontolanrahka. Though it may still be too early to attempt a regional summary of wetness change in the southern Baltic region, this study is a next step to better understand the long-term peatland palaeohydrology in NE Europe.

  19. Coring of Karakel’ Lake sediments (Teberda River valley and prospects for reconstruction of glaciation and Holocene climate history in the Caucasus

    Directory of Open Access Journals (Sweden)

    O. N. Solomina

    2013-01-01

    Full Text Available Lacustrine sediments represent an important data source for glacial and palaeoclimatic reconstructions. Having a number of certain advantages, they can be successfully used as a means of specification of glacier situation and age of moraine deposits, as well as a basis for detailed climatic models of the Holocene. The article focuses on the coring of sediments of Lake Kakakel (Western Caucasus that has its goal to clarify the Holocene climatic history for the region, providing the sampling methods, lithologic description of the sediment core, obtained radiocarbon dating and the element composition of the sediments. The primary outlook over the results of coring of the sediments of the Lake Karakyol helped to reconsider the conventional opinion on the glacial fluctuations in the valley of Teberda and to assume the future possibility for high-definition palaeoclimatic reconstruction for Western Caucasus.

  20. Roosevelt Island Climate Evolution Project (RICE): A 65 Kyr ice core record of black carbon aerosol deposition to the Ross Ice Shelf, West Antarctica.

    Science.gov (United States)

    Edwards, Ross; Bertler, Nancy; Tuohy, Andrea; Neff, Peter; Proemse, Bernedette; Feiteng, Wang; Goodwin, Ian; Hogan, Chad

    2015-04-01

    Emitted by fires, black carbon aerosols (rBC) perturb the atmosphere's physical and chemical properties and are climatically active. Sedimentary charcoal and other paleo-fire records suggest that rBC emissions have varied significantly in the past due to human activity and climate variability. However, few paleo rBC records exist to constrain reconstructions of the past rBC atmospheric distribution and its climate interaction. As part of the international Roosevelt Island Climate Evolution (RICE) project, we have developed an Antarctic rBC ice core record spanning the past ~65 Kyr. The RICE deep ice core was drilled from the Roosevelt Island ice dome in West Antarctica from 2011 to 2013. The high depth resolution (~ 1 cm) record was developed using a single particle intracavity laser-induced incandescence soot photometer (SP2) coupled to an ice core melter system. The rBC record displays sub-annual variability consistent with both austral dry-season and summer biomass burning. The record exhibits significant decadal to millennial-scale variability consistent with known changes in climate. Glacial rBC concentrations were much lower than Holocene concentrations with the exception of several periods of abrupt increases in rBC. The transition from glacial to interglacial rBC concentrations occurred over a much longer time relative to other ice core climate proxies such as water isotopes and suggests . The protracted increase in rBC during the transition may reflected Southern hemisphere ecosystem / fire regime changes in response to hydroclimate and human activity.

  1. Atmospheric CO2 variations over the last three glacial-interglacial climatic cycles deduced from the Dome Fuji deep ice core, Antarctica using a wet extraction technique

    International Nuclear Information System (INIS)

    Kawamura, Kenji; Nakazawa, Takakiyo; Aoki, Shuji

    2003-01-01

    A deep ice core drilled at Dome Fuji, East Antarctica was analyzed for the CO 2 concentration using a wet extraction method in order to reconstruct its atmospheric variations over the past 320 kyr, which includes three full glacial-interglacial climatic cycles, with a mean time resolution of about 1.1 kyr. The CO 2 concentration values derived for the past 65 kyr are very close to those obtained from other Antarctic ice cores using dry extraction methods, although the wet extraction method is generally thought to be inappropriate for the determination of the CO 2 concentration. The comparison between the CO 2 and Ca 2+ concentrations deduced from the Dome Fuji core suggests that calcium carbonate emitted from lands was mostly neutralized in the atmosphere before reaching the central part of Antarctica, or that only a small part of calcium carbonate was involved in CO 2 production during the wet extraction process. The CO 2 concentration for the past 320 kyr deduced from the Dome Fuji core varies between 190 and 300 ppmv, showing clear glacial-interglacial variations similar to the result of the Vostok ice core. However, for some periods, the concentration values of the Dome Fuji core are higher by up to 20 ppmv than those of the Vostok core. There is no clear indication that such differences are related to variations of chemical components of Ca 2+ , microparticle and acidity of the Dome Fuji core

  2. The core of Ure2p prion fibrils is formed by the N-terminal segment in a parallel cross-β structure: evidence from solid-state NMR.

    Science.gov (United States)

    Kryndushkin, Dmitry S; Wickner, Reed B; Tycko, Robert

    2011-06-03

    Intracellular fibril formation by Ure2p produces the non-Mendelian genetic element [URE3] in Saccharomyces cerevisiae, making Ure2p a prion protein. We show that solid-state NMR spectra of full-length Ure2p fibrils, seeded with infectious prions from a specific [URE3] strain and labeled with uniformly (15)N-(13)C-enriched Ile, include strong, sharp signals from Ile residues in the globular C-terminal domain (CTD) with both helical and nonhelical (13)C chemical shifts. Treatment with proteinase K eliminates these CTD signals, leaving only nonhelical signals from the Gln-rich and Asn-rich N-terminal segment, which are also observed in the solid-state NMR spectra of Ile-labeled fibrils formed by residues 1-89 of Ure2p. Thus, the N-terminal segment, or "prion domain" (PD), forms the fibril core, while CTD units are located outside the core. We additionally show that, after proteinase K treatment, Ile-labeled Ure2p fibrils formed without prion seeding exhibit a broader set of solid-state NMR signals than do prion-seeded fibrils, consistent with the idea that structural variations within the PD core account for prion strains. Measurements of (13)C-(13)C magnetic dipole-dipole couplings among (13)C-labeled Ile carbonyl sites in full-length Ure2p fibrils support an in-register parallel β-sheet structure for the PD core of Ure2p fibrils. Finally, we show that a model in which CTD units are attached rigidly to the parallel β-sheet core is consistent with steric constraints. Published by Elsevier Ltd.

  3. Climate change and daily press : Italy vs Usa parallel analysis; Stampa e cambiamento climatico : un confronto internazionale

    Energy Technology Data Exchange (ETDEWEB)

    Borrelli, G; Mazzotta, V [ENEA, Centro Ricerche Casaccia, Rome (Italy). Dip. Ambiente; Falconi, C; Grossi, R; Farabollini, F

    1996-06-01

    Among ENEA (Italian National Agency for New Technologies, Energy, and the Environment) activities, one deals with analysis and strategies of environmental information. A survey of four daily newspaper coverage, on an issue (Global Climate Change) belonging to this area, has been realized. The involved newspapers are: two Italian ones, namely `La Repubblica` and `Il Corriere della Sera`, two North-American ones, namely `New York Times` and `Washington Post`. Purpose of the work was that of detecting the qualitative and quantitative level of consciousness of the Italian press via a comparison with the North-American press, notoriously sensible and careful on environmental issues. The number of articled analyzed is partitioned in the following numerical data: 319 for the `New York Times`, 309 for the `Washington Post`, 146 for the `Corriere della Sera`, 81 articles for `La Repubblica`. The time period covered for the analysis spans from 1989, initiatic year for the organization of the 1992 Rio Conference, to December 1994, deadline date for the submission of national

  4. Badlands: A parallel basin and landscape dynamics model

    Directory of Open Access Journals (Sweden)

    T. Salles

    2016-01-01

    Full Text Available Over more than three decades, a number of numerical landscape evolution models (LEMs have been developed to study the combined effects of climate, sea-level, tectonics and sediments on Earth surface dynamics. Most of them are written in efficient programming languages, but often cannot be used on parallel architectures. Here, I present a LEM which ports a common core of accepted physical principles governing landscape evolution into a distributed memory parallel environment. Badlands (acronym for BAsin anD LANdscape DynamicS is an open-source, flexible, TIN-based landscape evolution model, built to simulate topography development at various space and time scales.

  5. The carbon dioxide content in ice cores - climatic curves of carbon dioxide. Zu den CO sub 2 -Klimakurven aus Eisbohrkernen

    Energy Technology Data Exchange (ETDEWEB)

    Heyke, H.E.

    1992-05-01

    The 'greenhouse effect', which implies a temperature of 15 deg C as against -18 deg C, owes its effect to 80% from water (clouds and gaseous phase) and to 10% from carbon dioxide, besides other components. Whereas water is largely unaccounted for, carbon dioxide has been postulated as the main cause of anticipated climatic catastrophe. The carbon dioxide concentration in the atmosphere has risen presently to such levels that all previous figures seem to have been left far behind. The reference point is the concentration of carbon dioxide in the air bubbles trapped in ice cores of Antartic and Greenland ice dated 160 000 years ago, which show much lower values than at present. A review of the most relevant publications indicates that many basic laws of chemistry seem to have been left largely unconsidered and experimental errors have made the results rather doubtful. Appropriate arguments have been presented. The investigations considered should be repeated under improved and more careful conditions. (orig.).

  6. 基于OpenMP的电磁场FDTD多核并行程序设计%Design of electromagnetic field FDTD multi-core parallel program based on OpenMP

    Institute of Scientific and Technical Information of China (English)

    吕忠亭; 张玉强; 崔巍

    2013-01-01

    探讨了基于OpenMP的电磁场FDTD多核并行程序设计的方法,以期实现该方法在更复杂的算法中应用具有更理想的性能提升。针对一个一维电磁场FDTD算法问题,对其计算方法与过程做了简单描述。在Fortran语言环境中,采用OpenMP+细粒度并行的方式实现了并行化,即只对循环部分进行并行计算,并将该并行方法在一个三维瞬态场电偶极子辐射FDTD程序中进行了验证。该并行算法取得了较其他并行FDTD算法更快的加速比和更高的效率。结果表明基于OpenMP的电磁场FDTD并行算法具有非常好的加速比和效率。%The method of the electromagnetic field FDTD multi-core parallel programm design based on OpenMP is dis-cussed,in order to implement ideal performance improvement of this method in the application of more sophisticated algorithms. Aiming at a problem existing in one-dimensional electromagnetic FDTD algorithm , its calculation method and process are described briefly. In Fortran language environment,the parallelism is achieved with OpenMP technology and fine-grained parallel way,that is,the parallel computation is performed only for the cycle part. The parallel method was verified in a three-dimensional transient electromagnetic field FDTD program for dipole radiation. The parallel algorithm has achieved faster speedup and higher efficiency than other parallel FDTD algoritms. The results indicate that the electromagnetic field FDTD parallel algorithm based on OpenMP has a good speedup and efficiency.

  7. The Chew Bahir Drilling Project (HSPDP). Deciphering climate information from the Chew Bahir sediment cores: Towards a continuous half-million year climate record near the Omo - Turkana key palaeonanthropological Site

    Science.gov (United States)

    Foerster, Verena E.; Asrat, Asfawossen; Chapot, Melissa S.; Cohen, Andrew S.; Dean, Jonathan R.; Deino, Alan; Günter, Christina; Junginger, Annett; Lamb, Henry F.; Leng, Melanie J.; Roberts, Helen M.; Schaebitz, Frank; Trauth, Martin H.

    2017-04-01

    As a contribution towards an enhanced understanding of human-climate interactions, the Hominin Sites and Paleolakes Drilling Project (HSPDP) has successfully completed coring five dominantly lacustrine archives of climate change during the last 3.5 Ma in East Africa. All five sites in Ethiopia and Kenya are adjacent to key paleoanthropological research areas encompassing diverse milestones in human evolution, dispersal episodes, and technological innovation. The 280 m-long Chew Bahir sediment records, recovered from a tectonically-bound basin in the southern Ethiopian rift in late 2014, cover the past 550 ka of environmental history, a time period that includes the transition to the Middle Stone Age, and the origin and dispersal of modern Homo sapiens. Deciphering climate information from lake sediments is challenging, due to the complex relationship between climate parameters and sediment composition. We will present the first results in our efforts to develop a reliable climate-proxy tool box for Chew Bahir by deconvolving the relationship between sedimentological and geochemical sediment composition and strongly climate-controlled processes in the basin, such as incongruent weathering, transportation and authigenic mineral alteration. Combining our first results from the long cores with those from a pilot study of short cores taken in 2009/10 along a NW-SE transect of the basin, we have developed a hypothesis linking climate forcing and paleoenvironmental signal-formation processes in the basin. X-ray diffraction analysis of the first sample sets from the long Chew Bahir record reveals similar processes that have been recognized for the uppermost 20 m during the pilot-study of the project: the diagenetic illitization of smectites during episodes of higher alkalinity and salinity in the closed-basin lake induced by a drier climate. The precise time resolution, largely continuous record and (eventually) a detailed understanding of site specific proxy formation

  8. Climate

    International Nuclear Information System (INIS)

    Fellous, J.L.

    2005-02-01

    This book starts with a series of about 20 preconceived ideas about climate and climatic change and analyses each of them in the light of the present day knowledge. Using this approach, it makes a status of the reality of the climatic change, of its causes and of the measures to be implemented to limit its impacts and reduce its most harmful consequences. (J.S.)

  9. Parallel rendering

    Science.gov (United States)

    Crockett, Thomas W.

    1995-01-01

    This article provides a broad introduction to the subject of parallel rendering, encompassing both hardware and software systems. The focus is on the underlying concepts and the issues which arise in the design of parallel rendering algorithms and systems. We examine the different types of parallelism and how they can be applied in rendering applications. Concepts from parallel computing, such as data decomposition, task granularity, scalability, and load balancing, are considered in relation to the rendering problem. We also explore concepts from computer graphics, such as coherence and projection, which have a significant impact on the structure of parallel rendering algorithms. Our survey covers a number of practical considerations as well, including the choice of architectural platform, communication and memory requirements, and the problem of image assembly and display. We illustrate the discussion with numerous examples from the parallel rendering literature, representing most of the principal rendering methods currently used in computer graphics.

  10. Parallel computations

    CERN Document Server

    1982-01-01

    Parallel Computations focuses on parallel computation, with emphasis on algorithms used in a variety of numerical and physical applications and for many different types of parallel computers. Topics covered range from vectorization of fast Fourier transforms (FFTs) and of the incomplete Cholesky conjugate gradient (ICCG) algorithm on the Cray-1 to calculation of table lookups and piecewise functions. Single tridiagonal linear systems and vectorized computation of reactive flow are also discussed.Comprised of 13 chapters, this volume begins by classifying parallel computers and describing techn

  11. Reconstruction of Antarctic climate change using ice core proxy records from the coastal Dronning Maud Land, East Antarctica

    Digital Repository Service at National Institute of Oceanography (India)

    Thamban, M.; Laluraj, C.M.; Naik, S.S.; Chaturvedi, A.

    the austral summer of 2003. The retrieved ice core samples were labelled, packed in good quality LDPE containers and subsequently shipped in -20ºC deep freezer facilities. These cores were archived in frozen conditions in custom-made expanded polypropylene...: Glaciochemistry, Stable isotope, Ice core, Solar activity, Dronning Maud Land, Antarctica. regions offer continuous and highly resolved long-term records of reliable information on major atmospheric parameters like temperature, composition and trace gases. Among...

  12. Parallel algorithms

    CERN Document Server

    Casanova, Henri; Robert, Yves

    2008-01-01

    ""…The authors of the present book, who have extensive credentials in both research and instruction in the area of parallelism, present a sound, principled treatment of parallel algorithms. … This book is very well written and extremely well designed from an instructional point of view. … The authors have created an instructive and fascinating text. The book will serve researchers as well as instructors who need a solid, readable text for a course on parallelism in computing. Indeed, for anyone who wants an understandable text from which to acquire a current, rigorous, and broad vi

  13. Climate Changes Documented in Ice Core Records from Third Pole Glaciers, with Emphasis on the Guliya Ice Cap in the Western Kunlun Mountains over the Last 100 Years

    Science.gov (United States)

    Thompson, L. G.; Yao, T.; Beaudon, E.; Mosley-Thompson, E.; Davis, M. E.; Kenny, D. V.; Lin, P. N.

    2016-12-01

    The Third Pole (TP) is a rapidly warming region containing 100,000 km2 of ice cover that collectively holds one of Earth's largest stores of freshwater that feeds Asia's largest rivers and helps sustain 1.5 billion people. Information on the accelerating warming in the region, its impact on the glaciers and subsequently on future water resources is urgently needed to guide mitigation and adaptation policies. Ice core histories collected over the last three decades across the TP demonstrate its climatic complexity and diversity. Here we present preliminary results from the flagship project of the Third Pole Environment Program, the 2015 Sino-American cooperative ice core drilling of the Guliya ice cap in the Kunlun Mountains in the western TP near the northern limit of the region influenced by the southwest monsoon. Three ice cores, each 51 meters in length, were recovered from the summit ( 6700 masl) while two deeper cores, one to bedrock ( 310 meters), were recovered from the plateau ( 6200 masl). Across the ice cap the net balance (accumulation) has increased annually by 2.3 cm of water equivalent from 1963-1992 to 1992-2015, and average oxygen isotopic ratios (δ18O) have enriched by 2‰. This contrasts with the recent ablation on the Naimona'nyi glacier located 540 km south of Guliya in the western Himalaya. Borehole temperatures in 2015 on the Guliya plateau have warmed substantially in the upper 30 meters of the ice compared to temperatures in 1992, when the first deep-drilling of the Guliya plateau was conducted. Compared with glaciers in the northern and western TP, the Himalayan ice fields are more sensitive to both fluctuations in the South Asian Monsoon and rising temperatures in the region. We examine the climatic changes of the last century preserved in ice core records from sites throughout the TP and compare them with those reconstructed for earlier warm epochs, such as the Medieval Climate Anomaly ( 950-1250 AD), the early Holocene "Hypsithermal

  14. Large-scale drivers of Caucasus climate variability in meteorological records and Mt El'brus ice cores

    Science.gov (United States)

    Kozachek, Anna; Mikhalenko, Vladimir; Masson-Delmotte, Valérie; Ekaykin, Alexey; Ginot, Patrick; Kutuzov, Stanislav; Legrand, Michel; Lipenkov, Vladimir; Preunkert, Susanne

    2017-05-01

    A 181.8 m ice core was recovered from a borehole drilled into bedrock on the western plateau of Mt El'brus (43°20'53.9'' N, 42°25'36.0'' E; 5115 m a.s.l.) in the Caucasus, Russia, in 2009 (Mikhalenko et al., 2015). Here, we report on the results of the water stable isotope composition from this ice core with additional data from the shallow cores. The distinct seasonal cycle of the isotopic composition allows dating by annual layer counting. Dating has been performed for the upper 126 m of the deep core combined with 20 m from the shallow cores. The whole record covers 100 years, from 2013 back to 1914. Due to the high accumulation rate (1380 mm w.e. year-1) and limited melting, we obtained isotopic composition and accumulation rate records with seasonal resolution. These values were compared with available meteorological data from 13 weather stations in the region and also with atmosphere circulation indices, back-trajectory calculations, and Global Network of Isotopes in Precipitation (GNIP) data in order to decipher the drivers of accumulation and ice core isotopic composition in the Caucasus region. In the warm season (May-October) the isotopic composition depends on local temperatures, but the correlation is not persistent over time, while in the cold season (November-April), atmospheric circulation is the predominant driver of the ice core's isotopic composition. The snow accumulation rate correlates well with the precipitation rate in the region all year round, which made it possible to reconstruct and expand the precipitation record at the Caucasus highlands from 1914 until 1966, when reliable meteorological observations of precipitation at high elevation began.

  15. Long-term Records of Pacific Salmon Abundance From Sediment Core Analysis: Relationships to Past Climatic Change, and Implications for the Future

    Science.gov (United States)

    Finney, B.

    2002-12-01

    The response of Pacific salmon to future climatic change is uncertain, but will have large impacts on the economy, culture and ecology of the North Pacific Rim. Relationships between sockeye salmon populations and climatic change can be determined by analyzing sediment cores from lakes where sockeye return to spawn. Sockeye salmon return to their natal lake system to spawn and subsequently die following 2 - 3 years of feeding in the North Pacific Ocean. Sockeye salmon abundance can be reconstructed from stable nitrogen isotope analysis of lake sediment cores as returning sockeye transport significant quantities of N, relatively enriched in N-15, from the ocean to freshwater systems. Temporal changes in the input of salmon-derived N, and hence salmon abundance, can be quantified through downcore analysis of N isotopes. Reconstructions of sockeye salmon abundance from lakes in several regions of Alaska show similar temporal patterns, with variability occurring on decadal to millennial timescales. Over the past 2000 years, shifts in sockeye salmon abundance far exceed the historical decadal-scale variability. A decline occurred from about 100 BC - 800 AD, but salmon were consistently more abundant 1200 - 1900 AD. Declines since 1900 AD coincide with the period of extensive commercial fishing. Correspondence between these records and paleoclimatic data suggest that changes in salmon abundance are related to large scale climatic changes over the North Pacific. For example, the increase in salmon abundance c.a. 1200 AD corresponds to a period of glacial advance in southern Alaska, and a shift to drier conditions in western North America. Although the regionally coherent patterns in reconstructed salmon abundance are consistent with the hypothesis that climate is an important driver, the relationships do not always follow patterns observed in the 20th century. A main feature of recorded climate variability in this region is the alternation between multi-decade periods of

  16. Causal Chains Arising from Climate Change in Mountain Regions: the Core Program of the Mountain Research Initiative

    Science.gov (United States)

    Greenwood, G. B.

    2014-12-01

    Mountains are a widespread terrestrial feature, covering from 12 to 24 percent of the world's terrestrial surface, depending of the definition. Topographic relief is central to the definition of mountains, to the benefits and costs accruing to society and to the cascade of changes expected from climate change. Mountains capture and store water, particularly important in arid regions and in all areas for energy production. In temperate and boreal regions, mountains have a great range in population densities, from empty to urban, while tropical mountains are often densely settled and farmed. Mountain regions contain a wide range of habitats, important for biodiversity, and for primary, secondary and tertiary sectors of the economy. Climate change interacts with this relief and consequent diversity. Elevation itself may accentuate warming (elevationi dependent warming) in some mountain regions. Even average warming starts complex chains of causality that reverberate through the diverse social ecological mountain systems affecting both the highlands and adjacent lowlands. A single feature of climate change such as higher snow lines affect the climate through albedo, the water cycle through changes in timing of release , water quality through the weathering of newly exposed material, geomorphology through enhanced erosion, plant communities through changes in climatic water balance, and animal and human communities through changes in habitat conditions and resource availabilities. Understanding these causal changes presents a particular interdisciplinary challenge to researchers, from assessing the existence and magnitude of elevation dependent warming and monitoring the full suite of changes within the social ecological system to climate change, to understanding how social ecological systems respond through individual and institutional behavior with repercussions on the long-term sustainability of these systems.

  17. Automatic Loop Parallelization via Compiler Guided Refactoring

    DEFF Research Database (Denmark)

    Larsen, Per; Ladelsky, Razya; Lidman, Jacob

    For many parallel applications, performance relies not on instruction-level parallelism, but on loop-level parallelism. Unfortunately, many modern applications are written in ways that obstruct automatic loop parallelization. Since we cannot identify sufficient parallelization opportunities...... for these codes in a static, off-line compiler, we developed an interactive compilation feedback system that guides the programmer in iteratively modifying application source, thereby improving the compiler’s ability to generate loop-parallel code. We use this compilation system to modify two sequential...... benchmarks, finding that the code parallelized in this way runs up to 8.3 times faster on an octo-core Intel Xeon 5570 system and up to 12.5 times faster on a quad-core IBM POWER6 system. Benchmark performance varies significantly between the systems. This suggests that semi-automatic parallelization should...

  18. Massively Parallel Finite Element Programming

    KAUST Repository

    Heister, Timo

    2010-01-01

    Today\\'s large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  19. Massively Parallel Finite Element Programming

    KAUST Repository

    Heister, Timo; Kronbichler, Martin; Bangerth, Wolfgang

    2010-01-01

    Today's large finite element simulations require parallel algorithms to scale on clusters with thousands or tens of thousands of processor cores. We present data structures and algorithms to take advantage of the power of high performance computers in generic finite element codes. Existing generic finite element libraries often restrict the parallelization to parallel linear algebra routines. This is a limiting factor when solving on more than a few hundreds of cores. We describe routines for distributed storage of all major components coupled with efficient, scalable algorithms. We give an overview of our effort to enable the modern and generic finite element library deal.II to take advantage of the power of large clusters. In particular, we describe the construction of a distributed mesh and develop algorithms to fully parallelize the finite element calculation. Numerical results demonstrate good scalability. © 2010 Springer-Verlag.

  20. Collaborative Observation and Research (CORE) Watersheds: new strategies for tracking the regional effects of climate change on complex systems

    Science.gov (United States)

    Murdoch, P. S.

    2007-12-01

    The past 30 years of environmental research have shown that our world is not made up of discrete components acting independently, but rather of a mosaic of complex relations among air, land, water, living resources, and human activities. Recent warming of the climate is having a significant effect on the functioning of those systems. A national imperative is developing to quickly establish local, regional, and national systems for anticipating environmental degradation from a changing climate and developing cost-effective adaptation or mitigation strategies. In these circumstances, the debate over research versus monitoring becomes moot--there is a clear need for the integrated application of both across a range of temporal and spatial scales. A national framework that effectively addresses the multiple scales and complex multi-disciplinary processes of climate change is being assembled largely from existing programs through collaboration among Federal, State, local, and NGO organizations. The result will be an observation and research network capable of interpreting complex environmental changes at a range of spatial and temporal scales, but at less cost than if the network were funded as an independent initiative. A pilot implementation of the collaborative framework in the Delaware River Basin yielded multi-scale assessments of carbon storage and flux, and the effects of forest fragmentation and soil calcium depletion on ecosystem function. A prototype of a national climate-effects observation and research network linking research watersheds, regional surveys, remote sensing, and ecosystem modeling is being initiated in the Yukon River Basin where carbon flux associated with permafrost thaw could accelerate global warming.

  1. Benchmark Comparison of Dual- and Quad-Core Processor Linux Clusters with Two Global Climate Modeling Workloads

    Science.gov (United States)

    McGalliard, James

    2008-01-01

    This viewgraph presentation details the science and systems environments that NASA High End computing program serves. Included is a discussion of the workload that is involved in the processing for the Global Climate Modeling. The Goddard Earth Observing System Model, Version 5 (GEOS-5) is a system of models integrated using the Earth System Modeling Framework (ESMF). The GEOS-5 system was used for the Benchmark tests, and the results of the tests are shown and discussed. Tests were also run for the Cubed Sphere system, results for these test are also shown.

  2. Climate and hydrology of the last interglaciation (MIS 5) in Owens Basin, California: Isotopic and geochemical evidence from core OL-92

    Science.gov (United States)

    Li, H.-C.; Bischoff, J.L.; Ku, T.-L.; Zhu, Z.-Y.

    2004-01-01

    ??18O, ??13C, total organic carbon, total inorganic carbon, and acid-leachable Li, Mg and Sr concentrations on 443 samples from 32 to 83 m depth in Owens Lake core OL-92 were analyzed to study the climatic and hydrological conditions between 60 and 155 ka with a resolution of ???200 a. The multi-proxy data show that Owens Lake overflowed during wet/cold conditions of marine isotope stages (MIS) 4, 5b and 6, and was closed during the dry/warm conditions of MIS 5a, c and e. The lake partially overflowed during MIS 5d. Our age model places the MIS 4/5 boundary at ca 72.5 ka and the MIS 5/6 boundary (Termination II) at ca 140 ka, agreeing with the Devils Hole chronology. The diametrical precipitation intensities between the Great Basin (cold/wet) and eastern China (cold/dry) on Milankovitch time scales imply a climatic teleconnection across the Pacific. It also probably reflects the effect of high-latitude ice sheets on the southward shifts of both the summer monsoon frontal zone in eastern Asia and the polar jet stream in western North America during glacial periods. ?? 2003 Elsevier Ltd. All rights reserved.

  3. Parallel R

    CERN Document Server

    McCallum, Ethan

    2011-01-01

    It's tough to argue with R as a high-quality, cross-platform, open source statistical software product-unless you're in the business of crunching Big Data. This concise book introduces you to several strategies for using R to analyze large datasets. You'll learn the basics of Snow, Multicore, Parallel, and some Hadoop-related tools, including how to find them, how to use them, when they work well, and when they don't. With these packages, you can overcome R's single-threaded nature by spreading work across multiple CPUs, or offloading work to multiple machines to address R's memory barrier.

  4. Late Quaternary vegetational and climate dynamics in northeastern Brazil, inferences from marine core GeoB 3104-1

    Science.gov (United States)

    Behling, Hermann; W. Arz, Helge; Pätzold, Jürgen; Wefer, Gerold

    2000-06-01

    Late Quaternary paleoenvironments from northeastern (NE) Brazil have been studied by pollen analysis of marine sediment. The studied core GeoB 3104-1 (3°40' S, 37°43' W, 767 m b.s.l.) from the upper continental slope off NE Brazil is 517 cm long and >42,000 14C yr BP old. Chronological control was obtained by 12 radiocarbon (AMS) dates from individuals of the foraminiferal species Globigerinoides sacculifer. Modern pollen analogs were received from 15 river, lake and forest soil surface samples from NE Brazil. Marine pollen dates indicate the predominance of semi-arid caatinga vegetation in NE Brazil during the recorded period between >42,000 and 8500 14C yr BP. The increased fluvial input of terrigenous material, with high concentrations of pollen and specially fern spores, into the marine deposits, about 40,000, 33,000 and 24,000 14C yr BP and between 15,500 and 11,800 14C yr BP, indicate short-term periods of strong rainfall on the NE Brazilian continent. The expansion of mountain, floodplain and gallery forests characterize the interval between 15,500 and 11,800 14C yr BP as the wettest recorded period in NE Brazil, which allowed floristic exchanges between Atlantic rain forest and Amazonian rain forest, and vice versa. The paleodata from core GeoB 3104-1 confirm the, in general, dry pre-Last Glacial Maximum (LGM) and LGM conditions and the change to wet Lateglacial environments in tropical South America. The annual movement of the intertropical convergence zone over NE Brazil, the strong influence of the Antarctic cold fronts and changes of the high-pressure cell over the southern Atlantic, may explain the very wet Lateglacial period in NE Brazil. The documented NE Brazilian short-term signals correlate with the documented Dansgaard-Oeschger cycles and Heinrich events from the northern Hemisphere and suggest strong teleconnections.

  5. Long-term vegetation, climate and ocean dynamics inferred from a 73,500 years old marine sediment core (GeoB2107-3) off southern Brazil

    Science.gov (United States)

    Gu, Fang; Zonneveld, Karin A. F.; Chiessi, Cristiano M.; Arz, Helge W.; Pätzold, Jürgen; Behling, Hermann

    2017-09-01

    Long-term changes in vegetation and climate of southern Brazil, as well as ocean dynamics of the adjacent South Atlantic, were studied by analyses of pollen, spores and organic-walled dinoflagellate cysts (dinocysts) in marine sediment core GeoB2107-3 collected offshore southern Brazil covering the last 73.5 cal kyr BP. The pollen record indicates that grasslands were much more frequent in the landscapes of southern Brazil during the last glacial period if compared to the late Holocene, reflecting relatively colder and/or less humid climatic conditions. Patches of forest occurred in the lowlands and probably also on the exposed continental shelf that was mainly covered by salt marshes. Interestingly, drought-susceptible Araucaria trees were frequent in the highlands (with a similar abundance as during the late Holocene) until 65 cal kyr BP, but were rare during the following glacial period. Atlantic rainforest was present in the northern lowlands of southern Brazil during the recorded last glacial period, but was strongly reduced from 38.5 until 13.0 cal kyr BP. The reduction was probably controlled by colder and/or less humid climatic conditions. Atlantic rainforest expanded to the south since the Lateglacial period, while Araucaria forests advanced in the highlands only during the late Holocene. Dinocysts data indicate that the Brazil Current (BC) with its warm, salty and nutrient-poor waters influenced the study area throughout the investigated period. However, variations in the proportion of dinocyst taxa indicating an eutrophic environment reflect the input of nutrients transported mainly by the Brazilian Coastal Current (BCC) and partly discharged by the Rio Itajaí (the major river closest to the core site). This was strongly related to changes in sea level. A stronger influence of the BCC with nutrient rich waters occurred during Marine Isotope Stage (MIS) 4 and in particular during the late MIS 3 and MIS 2 under low sea level. Evidence of Nothofagus pollen

  6. The paradox of cooling streams in a warming world: Regional climate trends do not parallel variable local trends in stream temperature in the Pacific continental United States

    Science.gov (United States)

    Arismendi, Ivan; Johnson, Sherri; Dunham, Jason B.; Haggerty, Roy; Hockman-Wert, David

    2012-01-01

    Temperature is a fundamentally important driver of ecosystem processes in streams. Recent warming of terrestrial climates around the globe has motivated concern about consequent increases in stream temperature. More specifically, observed trends of increasing air temperature and declining stream flow are widely believed to result in corresponding increases in stream temperature. Here, we examined the evidence for this using long-term stream temperature data from minimally and highly human-impacted sites located across the Pacific continental United States. Based on hypothesized climate impacts, we predicted that we should find warming trends in the maximum, mean and minimum temperatures, as well as increasing variability over time. These predictions were not fully realized. Warming trends were most prevalent in a small subset of locations with longer time series beginning in the 1950s. More recent series of observations (1987-2009) exhibited fewer warming trends and more cooling trends in both minimally and highly human-influenced systems. Trends in variability were much less evident, regardless of the length of time series. Based on these findings, we conclude that our perspective of climate impacts on stream temperatures is clouded considerably by a lack of long-termdata on minimally impacted streams, and biased spatio-temporal representation of existing time series. Overall our results highlight the need to develop more mechanistic, process-based understanding of linkages between climate change, other human impacts and stream temperature, and to deploy sensor networks that will provide better information on trends in stream temperatures in the future.

  7. The paradox of cooling streams in a warming world: regional climate trends do not parallel variable local trends in stream temperature in the Pacific continental United States

    Science.gov (United States)

    Ivan Arismendi; Sherri L. Johnson; Jason B. Dunham; Roy Haggerty

    2012-01-01

    Temperature is a fundamentally important driver of ecosystem processes in streams. Recent warming of terrestrial climates around the globe has motivated concern about consequent increases in stream temperature. More specifically, observed trends of increasing air temperature and declining stream flow are widely believed to result in corresponding increases in stream...

  8. Performing a local reduction operation on a parallel computer

    Science.gov (United States)

    Blocksome, Michael A.; Faraj, Daniel A.

    2012-12-11

    A parallel computer including compute nodes, each including two reduction processing cores, a network write processing core, and a network read processing core, each processing core assigned an input buffer. Copying, in interleaved chunks by the reduction processing cores, contents of the reduction processing cores' input buffers to an interleaved buffer in shared memory; copying, by one of the reduction processing cores, contents of the network write processing core's input buffer to shared memory; copying, by another of the reduction processing cores, contents of the network read processing core's input buffer to shared memory; and locally reducing in parallel by the reduction processing cores: the contents of the reduction processing core's input buffer; every other interleaved chunk of the interleaved buffer; the copied contents of the network write processing core's input buffer; and the copied contents of the network read processing core's input buffer.

  9. Accelerating Climate Simulations Through Hybrid Computing

    Science.gov (United States)

    Zhou, Shujia; Sinno, Scott; Cruz, Carlos; Purcell, Mark

    2009-01-01

    Unconventional multi-core processors (e.g., IBM Cell B/E and NYIDIDA GPU) have emerged as accelerators in climate simulation. However, climate models typically run on parallel computers with conventional processors (e.g., Intel and AMD) using MPI. Connecting accelerators to this architecture efficiently and easily becomes a critical issue. When using MPI for connection, we identified two challenges: (1) identical MPI implementation is required in both systems, and; (2) existing MPI code must be modified to accommodate the accelerators. In response, we have extended and deployed IBM Dynamic Application Virtualization (DAV) in a hybrid computing prototype system (one blade with two Intel quad-core processors, two IBM QS22 Cell blades, connected with Infiniband), allowing for seamlessly offloading compute-intensive functions to remote, heterogeneous accelerators in a scalable, load-balanced manner. Currently, a climate solar radiation model running with multiple MPI processes has been offloaded to multiple Cell blades with approx.10% network overhead.

  10. Parallel Lines

    Directory of Open Access Journals (Sweden)

    James G. Worner

    2017-05-01

    Full Text Available James Worner is an Australian-based writer and scholar currently pursuing a PhD at the University of Technology Sydney. His research seeks to expose masculinities lost in the shadow of Australia’s Anzac hegemony while exploring new opportunities for contemporary historiography. He is the recipient of the Doctoral Scholarship in Historical Consciousness at the university’s Australian Centre of Public History and will be hosted by the University of Bologna during 2017 on a doctoral research writing scholarship.   ‘Parallel Lines’ is one of a collection of stories, The Shapes of Us, exploring liminal spaces of modern life: class, gender, sexuality, race, religion and education. It looks at lives, like lines, that do not meet but which travel in proximity, simultaneously attracted and repelled. James’ short stories have been published in various journals and anthologies.

  11. A Parallel Butterfly Algorithm

    KAUST Repository

    Poulson, Jack; Demanet, Laurent; Maxwell, Nicholas; Ying, Lexing

    2014-01-01

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  12. A Parallel Butterfly Algorithm

    KAUST Repository

    Poulson, Jack

    2014-02-04

    The butterfly algorithm is a fast algorithm which approximately evaluates a discrete analogue of the integral transform (Equation Presented.) at large numbers of target points when the kernel, K(x, y), is approximately low-rank when restricted to subdomains satisfying a certain simple geometric condition. In d dimensions with O(Nd) quasi-uniformly distributed source and target points, when each appropriate submatrix of K is approximately rank-r, the running time of the algorithm is at most O(r2Nd logN). A parallelization of the butterfly algorithm is introduced which, assuming a message latency of α and per-process inverse bandwidth of β, executes in at most (Equation Presented.) time using p processes. This parallel algorithm was then instantiated in the form of the open-source DistButterfly library for the special case where K(x, y) = exp(iΦ(x, y)), where Φ(x, y) is a black-box, sufficiently smooth, real-valued phase function. Experiments on Blue Gene/Q demonstrate impressive strong-scaling results for important classes of phase functions. Using quasi-uniform sources, hyperbolic Radon transforms, and an analogue of a three-dimensional generalized Radon transform were, respectively, observed to strong-scale from 1-node/16-cores up to 1024-nodes/16,384-cores with greater than 90% and 82% efficiency, respectively. © 2014 Society for Industrial and Applied Mathematics.

  13. Fast parallel event reconstruction

    CERN Multimedia

    CERN. Geneva

    2010-01-01

    On-line processing of large data volumes produced in modern HEP experiments requires using maximum capabilities of modern and future many-core CPU and GPU architectures.One of such powerful feature is a SIMD instruction set, which allows packing several data items in one register and to operate on all of them, thus achievingmore operations per clock cycle. Motivated by the idea of using the SIMD unit ofmodern processors, the KF based track fit has been adapted for parallelism, including memory optimization, numerical analysis, vectorization with inline operator overloading, and optimization using SDKs. The speed of the algorithm has been increased in 120000 times with 0.1 ms/track, running in parallel on 16 SPEs of a Cell Blade computer.  Running on a Nehalem CPU with 8 cores it shows the processing speed of 52 ns/track using the Intel Threading Building Blocks. The same KF algorithm running on an Nvidia GTX 280 in the CUDA frameworkprovi...

  14. The STAPL Parallel Graph Library

    KAUST Repository

    Harshvardhan,

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable distributed graph container and a collection of commonly used parallel graph algorithms. The library introduces pGraph pViews that separate algorithm design from the container implementation. It supports three graph processing algorithmic paradigms, level-synchronous, asynchronous and coarse-grained, and provides common graph algorithms based on them. Experimental results demonstrate improved scalability in performance and data size over existing graph libraries on more than 16,000 cores and on internet-scale graphs containing over 16 billion vertices and 250 billion edges. © Springer-Verlag Berlin Heidelberg 2013.

  15. From School of Rock to Building Core Knowledge: Teaching about Cenozoic climate change with data and case studies from the primary literature

    Science.gov (United States)

    Leckie, R. M.; St John, K. K.; Jones, M. H.; Pound, K. S.; Krissek, L. A.; Peart, L. W.

    2011-12-01

    The School of Rock (SoR) began in 2005 as a pilot geoscience professional development program for K-12 teachers and informal educators aboard the JOIDES Resolution (JR). Since then, the highly successful SoR program, sponsored by the Consortium for Ocean Leadership's Deep Earth Academy, has conducted on-shore professional development at the Integrated Ocean Drilling Program (IODP) core repository in College Station, TX, and on the JR. The success of the SoR program stems from the natural synergy that develops between research scientists and educators when their combined pedagogical skills and scientific knowledge are used to uncover a wealth of scientific ocean drilling discoveries and research findings. Educators are challenged with authentic inquiry based on sediment archives; these lessons from the past are then made transferable to the general public and to classrooms through the creation of age-appropriate student-active learning materials (http://www.oceanleadership.org/education/deep-earth-academy/educators/classroom-activities/). This science made accessible approach was the basis for a successful NSF Course Curriculum and Laboratory Improvement (CCLI) proposal to develop teaching materials for use at the college level. Our Building Core Knowledge project resulted in a series of 14 linked, yet independent, inquiry-based exercise modules around the theme of Reconstructing Earth's Climate History. All of the exercises build upon authentic data from peer reviewed scientific publications. These multiple part modules cover fundamental paleoclimate principles, tools and proxies, and Cenozoic case studies. It is important to teach students how we know what we know. For example, paleoclimate records must be systematically described, ages must be determined, and indirect evidence (i.e., proxies) of past climate must be analyzed. Much like the work of a detective, geoscientists and paleoclimatologists reconstruct what happened in the past, and when and how it

  16. Distributed Memory Parallel Computing with SEAWAT

    Science.gov (United States)

    Verkaik, J.; Huizer, S.; van Engelen, J.; Oude Essink, G.; Ram, R.; Vuik, K.

    2017-12-01

    Fresh groundwater reserves in coastal aquifers are threatened by sea-level rise, extreme weather conditions, increasing urbanization and associated groundwater extraction rates. To counteract these threats, accurate high-resolution numerical models are required to optimize the management of these precious reserves. The major model drawbacks are long run times and large memory requirements, limiting the predictive power of these models. Distributed memory parallel computing is an efficient technique for reducing run times and memory requirements, where the problem is divided over multiple processor cores. A new Parallel Krylov Solver (PKS) for SEAWAT is presented. PKS has recently been applied to MODFLOW and includes Conjugate Gradient (CG) and Biconjugate Gradient Stabilized (BiCGSTAB) linear accelerators. Both accelerators are preconditioned by an overlapping additive Schwarz preconditioner in a way that: a) subdomains are partitioned using Recursive Coordinate Bisection (RCB) load balancing, b) each subdomain uses local memory only and communicates with other subdomains by Message Passing Interface (MPI) within the linear accelerator, c) it is fully integrated in SEAWAT. Within SEAWAT, the PKS-CG solver replaces the Preconditioned Conjugate Gradient (PCG) solver for solving the variable-density groundwater flow equation and the PKS-BiCGSTAB solver replaces the Generalized Conjugate Gradient (GCG) solver for solving the advection-diffusion equation. PKS supports the third-order Total Variation Diminishing (TVD) scheme for computing advection. Benchmarks were performed on the Dutch national supercomputer (https://userinfo.surfsara.nl/systems/cartesius) using up to 128 cores, for a synthetic 3D Henry model (100 million cells) and the real-life Sand Engine model ( 10 million cells). The Sand Engine model was used to investigate the potential effect of the long-term morphological evolution of a large sand replenishment and climate change on fresh groundwater resources

  17. Dense SDM (12-Core × 3-Mode) Transmission Over 527 km With 33.2-ns Mode-Dispersion Employing Low-Complexity Parallel MIMO Frequency-Domain Equalization

    DEFF Research Database (Denmark)

    Shibahara, Kohki; Lee, Doohwan; Kobayashi, Takayuki

    2016-01-01

    We propose long-haul space-division-multiplexing (SDM) transmission systems employing parallel multiple-input multiple-output (MIMO) frequency-domain equalization (FDE) and transmission fiber with low differential mode delay (DMD). We first discuss the advantages of parallel MIMO FDE technique in...

  18. Synchronizing Greenland ice-core records and the Meerfelder maar sediment record via the global cosmogenic radionuclide signature and insights on climate around 11,230 years BP

    Science.gov (United States)

    Mekhaldi, F.; Czymzik, M.; Brauer, A.; Martin-Puertas, C.; Aldahan, A.; Possnert, G.; Muscheler, R.

    2017-12-01

    The causal investigation of multiple paleoclimate records relies on the accuracy of their respective chronostratigraphy. To achieve relative synchronization, cosmogenic radionuclides are an excellent tool because their common signature is global and can be retrieved and measured in different paleoclimate archives. For instance, 10Be can be measured in both ice cores and lake sediments (Berggren et al., 2013; Czymzik et al., 2016) which allows for both archives to be anchored onto radiocarbon timescales by synchronizing 10Be with 14C. We investigate the period 11,500-11,000 years BP when a short cold climate spell is known, from ice-core proxy records, to have occurred in Greenland shortly after the onset of the Holocene - the Preboreal Oscillation (PBO). This period also coincides with one of the largest and longest-lived increase in 14C production rate during the Holocene, which most likely corresponds to a grand solar minimum (around 11,230-11,000 years BP). In consequence, this period ideally illustrates the potential of using a known and clear signal in the production rate of cosmogenic radionuclides as a synchronizing tool, such as caused by large variations in solar activity. Here we measure 10Be in Meerfelder Maar (a well-dated and widely used sediment record from Germany) around 11,230 years BP which allows us to align the 10Be signal in both the Meerfelder Maar (MFM) sediment record and the GRIP ice core to IntCal13. Doing so, we report that i) the structure of the grand solar minimum is well-preserved in the 10Be signal of MFM sediments, ii) the PBO in Greenland occurs during high levels of solar activity and is not clearly observed in MFM, and iii) the PBO in Greenland ends precisely at the onset of the grand solar minimum at 11,230 years BP which also corresponds to a depositional change in MFM sediments (Martin-Puertas et al., 2017). These results thus suggest that changes in solar activity could have been a forcing at play eventually resulting in the

  19. Parallel plasma fluid turbulence calculations

    International Nuclear Information System (INIS)

    Leboeuf, J.N.; Carreras, B.A.; Charlton, L.A.; Drake, J.B.; Lynch, V.E.; Newman, D.E.; Sidikman, K.L.; Spong, D.A.

    1994-01-01

    The study of plasma turbulence and transport is a complex problem of critical importance for fusion-relevant plasmas. To this day, the fluid treatment of plasma dynamics is the best approach to realistic physics at the high resolution required for certain experimentally relevant calculations. Core and edge turbulence in a magnetic fusion device have been modeled using state-of-the-art, nonlinear, three-dimensional, initial-value fluid and gyrofluid codes. Parallel implementation of these models on diverse platforms--vector parallel (National Energy Research Supercomputer Center's CRAY Y-MP C90), massively parallel (Intel Paragon XP/S 35), and serial parallel (clusters of high-performance workstations using the Parallel Virtual Machine protocol)--offers a variety of paths to high resolution and significant improvements in real-time efficiency, each with its own advantages. The largest and most efficient calculations have been performed at the 200 Mword memory limit on the C90 in dedicated mode, where an overlap of 12 to 13 out of a maximum of 16 processors has been achieved with a gyrofluid model of core fluctuations. The richness of the physics captured by these calculations is commensurate with the increased resolution and efficiency and is limited only by the ingenuity brought to the analysis of the massive amounts of data generated

  20. FAMOUS, faster: using parallel computing techniques to accelerate the FAMOUS/HadCM3 climate model with a focus on the radiative transfer algorithm

    Directory of Open Access Journals (Sweden)

    P. Hanappe

    2011-09-01

    Full Text Available We have optimised the atmospheric radiation algorithm of the FAMOUS climate model on several hardware platforms. The optimisation involved translating the Fortran code to C and restructuring the algorithm around the computation of a single air column. Instead of the existing MPI-based domain decomposition, we used a task queue and a thread pool to schedule the computation of individual columns on the available processors. Finally, four air columns are packed together in a single data structure and computed simultaneously using Single Instruction Multiple Data operations.

    The modified algorithm runs more than 50 times faster on the CELL's Synergistic Processing Element than on its main PowerPC processing element. On Intel-compatible processors, the new radiation code runs 4 times faster. On the tested graphics processor, using OpenCL, we find a speed-up of more than 2.5 times as compared to the original code on the main CPU. Because the radiation code takes more than 60 % of the total CPU time, FAMOUS executes more than twice as fast. Our version of the algorithm returns bit-wise identical results, which demonstrates the robustness of our approach. We estimate that this project required around two and a half man-years of work.

  1. Pollen stratigraphy, vegetation and climate history of the last 215 ka in the Azzano Decimo core (plain of Friuli, north-eastern Italy)

    Science.gov (United States)

    Pini, R.; Ravazzi, C.; Donegana, M.

    2009-06-01

    The pollen record of the long succession of marine and continental deposits filling the subsident north-Adriatic foredeep basin (NE Italy) documents the history of vegetation, the landscape evolution and the climate forcing during the last 215 ka at the south-eastern Alpine foreland. The chronology relies on several 14C determinations as well as on estimated ages of pollen-stratigraphical and sea-level event tie-points derived from comparison with high-resolution marine records, speleothemes and ice cores. Mixed temperate rainforests persisted throughout MIS 7a-7c, being replaced by conifer forests after the local glacioeustatic regression during early MIS 6. The Alpine piedmont facing the Adriatic foredeeep was glaciated at the culmination of the penultimate glaciation, as directly testified by in situ fluvioglacial aggradation related to the building of a large morainic amphitheatre. The pollen record allows correlation with other European records and with the IRD from N-Atlantic and off Iberia, thus the duration of the penultimate glacial culmination at the southalpine fringe is estimated less than 13 ka between 148 ± 1 and >135 ka. The site was not reached by the Last Interglacial maximum sea transgression and enregistered a typical, though incomplete, Eemian forest record, lacking Mediterranean evergreen trees. A complex sequence of stadial-interstadial episodes is reconstructed during the Early and Middle Würm: major xerophyte peaks match IRD maxima occurred during Heinrich events in deep-sea cores offshore Iberia and in the N-Atlantic and allows to frame lumps of interstadial phases, marked by Picea peaks, each one including several DO warm events. Broad-leaved thermophilous forests disappeared from the north-eastern plain of Italy at the end of the Early Würm, whereas reduced populations of Abies and Fagus probably sheltered even during the Last Glacial Maximum. A renewed fluvioglacial in situ deposition between 30.4 ± 0.4 and 21.6 ± 0.5 ka cal BP sets

  2. Dense SDM (12-core × 3-mode) transmission over 527 km with 33.2-ns mode-dispersion employing low-complexity parallel MIMO frequency-domain equalization

    DEFF Research Database (Denmark)

    Shibahara, K.; Mizuno, T.; Takara, H.

    We demonstrate 12-core × 3-mode dense SDM transmission over 527 km graded-index multi-core few-mode fiber without mode-dispersion management. Employing low baud rate multi-carrier signal and frequency-domain equalization enables 33.2-ns DMD compensation with low computational complexity. © 2015 OSA...

  3. Parallel Algorithms for the Exascale Era

    Energy Technology Data Exchange (ETDEWEB)

    Robey, Robert W. [Los Alamos National Laboratory

    2016-10-19

    New parallel algorithms are needed to reach the Exascale level of parallelism with millions of cores. We look at some of the research developed by students in projects at LANL. The research blends ideas from the early days of computing while weaving in the fresh approach brought by students new to the field of high performance computing. We look at reproducibility of global sums and why it is important to parallel computing. Next we look at how the concept of hashing has led to the development of more scalable algorithms suitable for next-generation parallel computers. Nearly all of this work has been done by undergraduates and published in leading scientific journals.

  4. Ice cores and palaeoclimate

    International Nuclear Information System (INIS)

    Krogh Andersen, K.; Ditlevsen, P.; Steffensen, J.P.

    2001-01-01

    Ice cores from Greenland give testimony of a highly variable climate during the last glacial period. Dramatic climate warmings of 15 to 25 deg. C for the annual average temperature in less than a human lifetime have been documented. Several questions arise: Why is the Holocene so stable? Is climatic instability only a property of glacial periods? What is the mechanism behind the sudden climate changes? Are the increased temperatures in the past century man-made? And what happens in the future? The ice core community tries to attack some of these problems. The NGRIP ice core currently being drilled is analysed in very high detail, allowing for a very precise dating of climate events. It will be possible to study some of the fast changes on a year by year basis and from this we expect to find clues to the sequence of events during rapid changes. New techniques are hoped to allow for detection of annual layers as far back as 100,000 years and thus a much improved time scale over past climate changes. It is also hoped to find ice from the Eemian period. If the Eemian layers confirm the GRIP sequence, the Eemian was actually climatically unstable just as the glacial period. This would mean that the stability of the Holocene is unique. It would also mean, that if human made global warming indeed occurs, we could jeopardize the Holocene stability and create an unstable 'Eemian situation' which ultimately could start an ice age. Currenlty mankind is changing the composition of the atmosphere. Ice cores document significant increases in greenhouse gases, and due to increased emissions of sulfuric and nitric acid from fossil fuel burning, combustion engines and agriculture, modern Greenland snow is 3 - 5 times more acidic than pre-industrial snow (Mayewski et al., 1986). However, the magnitude and abruptness of the temperature changes of the past century do not exceed the magnitude of natural variability. It is from the ice core perspective thus not possible to attribute the

  5. Climatic risks

    International Nuclear Information System (INIS)

    Lamarre, D.; Favier, R.; Bourg, D.; Marchand, J.P.

    2005-04-01

    The climatic risks are analyzed in this book under the cross-vision of specialists of different domains: philosophy, sociology, economic history, law, geography, climatology and hydrology. The prevention of risks and the precautionary principle are presented first. Then, the relations between climatic risk and geography are analyzed using the notion of territoriality. The territory aspect is in the core of the present day debates about the geography of risks, in particular when the links between climate change and public health are considered. Then the main climatic risks are presented. Droughts and floods are the most damaging ones and the difficulties of prevention-indemnification coupling remain important. (J.S.)

  6. New perspectives for European climate services: HORIZON2020

    Science.gov (United States)

    Bruning, Claus; Tilche, Andrea

    2014-05-01

    The developing of new end-to-end climate services was one of the core priorities of 7th Framework for Research and Technological Development of the European Commission and will become one of the key strategic priorities of Societal Challenge 5 of HORIZON2020 (the new EU Framework Programme for Research and Innovation 2014-2020). Results should increase the competitiveness of European businesses, and the ability of regional and national authorities to make effective decisions in climate-sensitive sectors. In parallel, the production of new tailored climate information should strengthen the resilience of the European society to climate change. In this perspective the strategy to support and foster the underpinning science for climate services in HORIZON2020 will be presented.

  7. Core Hunter 3: flexible core subset selection.

    Science.gov (United States)

    De Beukelaer, Herman; Davenport, Guy F; Fack, Veerle

    2018-05-31

    Core collections provide genebank curators and plant breeders a way to reduce size of their collections and populations, while minimizing impact on genetic diversity and allele frequency. Many methods have been proposed to generate core collections, often using distance metrics to quantify the similarity of two accessions, based on genetic marker data or phenotypic traits. Core Hunter is a multi-purpose core subset selection tool that uses local search algorithms to generate subsets relying on one or more metrics, including several distance metrics and allelic richness. In version 3 of Core Hunter (CH3) we have incorporated two new, improved methods for summarizing distances to quantify diversity or representativeness of the core collection. A comparison of CH3 and Core Hunter 2 (CH2) showed that these new metrics can be effectively optimized with less complex algorithms, as compared to those used in CH2. CH3 is more effective at maximizing the improved diversity metric than CH2, still ensures a high average and minimum distance, and is faster for large datasets. Using CH3, a simple stochastic hill-climber is able to find highly diverse core collections, and the more advanced parallel tempering algorithm further increases the quality of the core and further reduces variability across independent samples. We also evaluate the ability of CH3 to simultaneously maximize diversity, and either representativeness or allelic richness, and compare the results with those of the GDOpt and SimEli methods. CH3 can sample equally representative cores as GDOpt, which was specifically designed for this purpose, and is able to construct cores that are simultaneously more diverse, and either are more representative or have higher allelic richness, than those obtained by SimEli. In version 3, Core Hunter has been updated to include two new core subset selection metrics that construct cores for representativeness or diversity, with improved performance. It combines and outperforms the

  8. Waves in the core and mechanical core-mantle interactions

    DEFF Research Database (Denmark)

    Jault, D.; Finlay, Chris

    2015-01-01

    This Chapter focuses on time-dependent uid motions in the core interior, which can beconstrained by observations of the Earth's magnetic eld, on timescales which are shortcompared to the magnetic diusion time. This dynamics is strongly inuenced by the Earth's rapid rotation, which rigidies...... the motions in the direction parallel to the Earth'srotation axis. This property accounts for the signicance of the core-mantle topography.In addition, the stiening of the uid in the direction parallel to the rotation axis gives riseto a magnetic diusion layer attached to the core-mantle boundary, which would...... otherwisebe dispersed by Alfven waves. This Chapter complements the descriptions of large-scaleow in the core (8.04), of turbulence in the core (8.06) and of core-mantle interactions(8.12), which can all be found in this volume. We rely on basic magnetohydrodynamictheory, including the derivation...

  9. Adapting algorithms to massively parallel hardware

    CERN Document Server

    Sioulas, Panagiotis

    2016-01-01

    In the recent years, the trend in computing has shifted from delivering processors with faster clock speeds to increasing the number of cores per processor. This marks a paradigm shift towards parallel programming in which applications are programmed to exploit the power provided by multi-cores. Usually there is gain in terms of the time-to-solution and the memory footprint. Specifically, this trend has sparked an interest towards massively parallel systems that can provide a large number of processors, and possibly computing nodes, as in the GPUs and MPPAs (Massively Parallel Processor Arrays). In this project, the focus was on two distinct computing problems: k-d tree searches and track seeding cellular automata. The goal was to adapt the algorithms to parallel systems and evaluate their performance in different cases.

  10. Proteomics Core

    Data.gov (United States)

    Federal Laboratory Consortium — Proteomics Core is the central resource for mass spectrometry based proteomics within the NHLBI. The Core staff help collaborators design proteomics experiments in a...

  11. Parallelism and Scalability in an Image Processing Application

    DEFF Research Database (Denmark)

    Rasmussen, Morten Sleth; Stuart, Matthias Bo; Karlsson, Sven

    2008-01-01

    parallel programs. This paper investigates parallelism and scalability of an embedded image processing application. The major challenges faced when parallelizing the application were to extract enough parallelism from the application and to reduce load imbalance. The application has limited immediately......The recent trends in processor architecture show that parallel processing is moving into new areas of computing in the form of many-core desktop processors and multi-processor system-on-chip. This means that parallel processing is required in application areas that traditionally have not used...

  12. Parallelism and Scalability in an Image Processing Application

    DEFF Research Database (Denmark)

    Rasmussen, Morten Sleth; Stuart, Matthias Bo; Karlsson, Sven

    2009-01-01

    parallel programs. This paper investigates parallelism and scalability of an embedded image processing application. The major challenges faced when parallelizing the application were to extract enough parallelism from the application and to reduce load imbalance. The application has limited immediately......The recent trends in processor architecture show that parallel processing is moving into new areas of computing in the form of many-core desktop processors and multi-processor system-on-chips. This means that parallel processing is required in application areas that traditionally have not used...

  13. ENSO and interdecadal climate variability over the last century documented by geochemical records of two coral cores from the South West Pacific

    Directory of Open Access Journals (Sweden)

    T. Ourbak

    2006-01-01

    Full Text Available The south west Pacific is affected by climatic phenomena such as ENSO (El Niño Southern Oscillation or the PDO (Pacific Decadal Oscillation. Near-monthly resolution calibrations of Sr/Ca, U/Ca and δ18Oc were made on corals taken from New Caledonia and Wallis Island. These geochemical variations could be linked to SST (sea surface temperature and SSS (sea surface salinity variations over the last two decades, itselves dependent on ENSO occurrences. On the other hand, near-half-yearly resolution over the last century smoothes seasonal and interannual climate signals, but emphasizes low frequency climate variability.

  14. Practical parallel computing

    CERN Document Server

    Morse, H Stephen

    1994-01-01

    Practical Parallel Computing provides information pertinent to the fundamental aspects of high-performance parallel processing. This book discusses the development of parallel applications on a variety of equipment.Organized into three parts encompassing 12 chapters, this book begins with an overview of the technology trends that converge to favor massively parallel hardware over traditional mainframes and vector machines. This text then gives a tutorial introduction to parallel hardware architectures. Other chapters provide worked-out examples of programs using several parallel languages. Thi

  15. Parallel sorting algorithms

    CERN Document Server

    Akl, Selim G

    1985-01-01

    Parallel Sorting Algorithms explains how to use parallel algorithms to sort a sequence of items on a variety of parallel computers. The book reviews the sorting problem, the parallel models of computation, parallel algorithms, and the lower bounds on the parallel sorting problems. The text also presents twenty different algorithms, such as linear arrays, mesh-connected computers, cube-connected computers. Another example where algorithm can be applied is on the shared-memory SIMD (single instruction stream multiple data stream) computers in which the whole sequence to be sorted can fit in the

  16. Parallel phase model : a programming model for high-end parallel machines with manycores.

    Energy Technology Data Exchange (ETDEWEB)

    Wu, Junfeng (Syracuse University, Syracuse, NY); Wen, Zhaofang; Heroux, Michael Allen; Brightwell, Ronald Brian

    2009-04-01

    This paper presents a parallel programming model, Parallel Phase Model (PPM), for next-generation high-end parallel machines based on a distributed memory architecture consisting of a networked cluster of nodes with a large number of cores on each node. PPM has a unified high-level programming abstraction that facilitates the design and implementation of parallel algorithms to exploit both the parallelism of the many cores and the parallelism at the cluster level. The programming abstraction will be suitable for expressing both fine-grained and coarse-grained parallelism. It includes a few high-level parallel programming language constructs that can be added as an extension to an existing (sequential or parallel) programming language such as C; and the implementation of PPM also includes a light-weight runtime library that runs on top of an existing network communication software layer (e.g. MPI). Design philosophy of PPM and details of the programming abstraction are also presented. Several unstructured applications that inherently require high-volume random fine-grained data accesses have been implemented in PPM with very promising results.

  17. Reassessment of the Upper Fremont Glacier ice-core chronologies by synchronizing of ice-core-water isotopes to a nearby tree-ring chronology

    Science.gov (United States)

    Chellman, Nathan J.; McConnell, Joseph R.; Arienzo, Monica; Pederson, Gregory T.; Aarons, Sarah; Csank, Adam

    2017-01-01

    The Upper Fremont Glacier (UFG), Wyoming, is one of the few continental glaciers in the contiguous United States known to preserve environmental and climate records spanning recent centuries. A pair of ice cores taken from UFG have been studied extensively to document changes in climate and industrial pollution (most notably, mid-19th century increases in mercury pollution). Fundamental to these studies is the chronology used to map ice-core depth to age. Here, we present a revised chronology for the UFG ice cores based on new measurements and using a novel dating approach of synchronizing continuous water isotope measurements to a nearby tree-ring chronology. While consistent with the few unambiguous age controls underpinning the previous UFG chronologies, the new interpretation suggests a very different time scale for the UFG cores with changes of up to 80 years. Mercury increases previously associated with the mid-19th century Gold Rush now coincide with early-20th century industrial emissions, aligning the UFG record with other North American mercury records from ice and lake sediment cores. Additionally, new UFG records of industrial pollutants parallel changes documented in ice cores from southern Greenland, further validating the new UFG chronologies while documenting the extent of late 19th and early 20th century pollution in remote North America.

  18. Shared Variable Oriented Parallel Precompiler for SPMD Model

    Institute of Scientific and Technical Information of China (English)

    1995-01-01

    For the moment,commercial parallel computer systems with distributed memory architecture are usually provided with parallel FORTRAN or parallel C compliers,which are just traditional sequential FORTRAN or C compilers expanded with communication statements.Programmers suffer from writing parallel programs with communication statements. The Shared Variable Oriented Parallel Precompiler (SVOPP) proposed in this paper can automatically generate appropriate communication statements based on shared variables for SPMD(Single Program Multiple Data) computation model and greatly ease the parallel programming with high communication efficiency.The core function of parallel C precompiler has been successfully verified on a transputer-based parallel computer.Its prominent performance shows that SVOPP is probably a break-through in parallel programming technique.

  19. A Tutorial on Parallel and Concurrent Programming in Haskell

    Science.gov (United States)

    Peyton Jones, Simon; Singh, Satnam

    This practical tutorial introduces the features available in Haskell for writing parallel and concurrent programs. We first describe how to write semi-explicit parallel programs by using annotations to express opportunities for parallelism and to help control the granularity of parallelism for effective execution on modern operating systems and processors. We then describe the mechanisms provided by Haskell for writing explicitly parallel programs with a focus on the use of software transactional memory to help share information between threads. Finally, we show how nested data parallelism can be used to write deterministically parallel programs which allows programmers to use rich data types in data parallel programs which are automatically transformed into flat data parallel versions for efficient execution on multi-core processors.

  20. Introduction to parallel programming

    CERN Document Server

    Brawer, Steven

    1989-01-01

    Introduction to Parallel Programming focuses on the techniques, processes, methodologies, and approaches involved in parallel programming. The book first offers information on Fortran, hardware and operating system models, and processes, shared memory, and simple parallel programs. Discussions focus on processes and processors, joining processes, shared memory, time-sharing with multiple processors, hardware, loops, passing arguments in function/subroutine calls, program structure, and arithmetic expressions. The text then elaborates on basic parallel programming techniques, barriers and race

  1. Parallel computing works!

    CERN Document Server

    Fox, Geoffrey C; Messina, Guiseppe C

    2014-01-01

    A clear illustration of how parallel computers can be successfully appliedto large-scale scientific computations. This book demonstrates how avariety of applications in physics, biology, mathematics and other scienceswere implemented on real parallel computers to produce new scientificresults. It investigates issues of fine-grained parallelism relevant forfuture supercomputers with particular emphasis on hypercube architecture. The authors describe how they used an experimental approach to configuredifferent massively parallel machines, design and implement basic systemsoftware, and develop

  2. Benchmarking NWP Kernels on Multi- and Many-core Processors

    Science.gov (United States)

    Michalakes, J.; Vachharajani, M.

    2008-12-01

    Increased computing power for weather, climate, and atmospheric science has provided direct benefits for defense, agriculture, the economy, the environment, and public welfare and convenience. Today, very large clusters with many thousands of processors are allowing scientists to move forward with simulations of unprecedented size. But time-critical applications such as real-time forecasting or climate prediction need strong scaling: faster nodes and processors, not more of them. Moreover, the need for good cost- performance has never been greater, both in terms of performance per watt and per dollar. For these reasons, the new generations of multi- and many-core processors being mass produced for commercial IT and "graphical computing" (video games) are being scrutinized for their ability to exploit the abundant fine- grain parallelism in atmospheric models. We present results of our work to date identifying key computational kernels within the dynamics and physics of a large community NWP model, the Weather Research and Forecast (WRF) model. We benchmark and optimize these kernels on several different multi- and many-core processors. The goals are to (1) characterize and model performance of the kernels in terms of computational intensity, data parallelism, memory bandwidth pressure, memory footprint, etc. (2) enumerate and classify effective strategies for coding and optimizing for these new processors, (3) assess difficulties and opportunities for tool or higher-level language support, and (4) establish a continuing set of kernel benchmarks that can be used to measure and compare effectiveness of current and future designs of multi- and many-core processors for weather and climate applications.

  3. Parallelization of Subchannel Analysis Code MATRA

    International Nuclear Information System (INIS)

    Kim, Seongjin; Hwang, Daehyun; Kwon, Hyouk

    2014-01-01

    A stand-alone calculation of MATRA code used up pertinent computing time for the thermal margin calculations while a relatively considerable time is needed to solve the whole core pin-by-pin problems. In addition, it is strongly required to improve the computation speed of the MATRA code to satisfy the overall performance of the multi-physics coupling calculations. Therefore, a parallel approach to improve and optimize the computability of the MATRA code is proposed and verified in this study. The parallel algorithm is embodied in the MATRA code using the MPI communication method and the modification of the previous code structure was minimized. An improvement is confirmed by comparing the results between the single and multiple processor algorithms. The speedup and efficiency are also evaluated when increasing the number of processors. The parallel algorithm was implemented to the subchannel code MATRA using the MPI. The performance of the parallel algorithm was verified by comparing the results with those from the MATRA with the single processor. It is also noticed that the performance of the MATRA code was greatly improved by implementing the parallel algorithm for the 1/8 core and whole core problems

  4. Improvement of Parallel Algorithm for MATRA Code

    International Nuclear Information System (INIS)

    Kim, Seong-Jin; Seo, Kyong-Won; Kwon, Hyouk; Hwang, Dae-Hyun

    2014-01-01

    The feasibility study to parallelize the MATRA code was conducted in KAERI early this year. As a result, a parallel algorithm for the MATRA code has been developed to decrease a considerably required computing time to solve a bigsize problem such as a whole core pin-by-pin problem of a general PWR reactor and to improve an overall performance of the multi-physics coupling calculations. It was shown that the performance of the MATRA code was greatly improved by implementing the parallel algorithm using MPI communication. For problems of a 1/8 core and whole core for SMART reactor, a speedup was evaluated as about 10 when the numbers of used processor were 25. However, it was also shown that the performance deteriorated as the axial node number increased. In this paper, the procedure of a communication between processors is optimized to improve the previous parallel algorithm.. To improve the performance deterioration of the parallelized MATRA code, the communication algorithm between processors was newly presented. It was shown that the speedup was improved and stable regardless of the axial node number

  5. A solution for automatic parallelization of sequential assembly code

    Directory of Open Access Journals (Sweden)

    Kovačević Đorđe

    2013-01-01

    Full Text Available Since modern multicore processors can execute existing sequential programs only on a single core, there is a strong need for automatic parallelization of program code. Relying on existing algorithms, this paper describes one new software solution tool for parallelization of sequential assembly code. The main goal of this paper is to develop the parallelizator which reads sequential assembler code and at the output provides parallelized code for MIPS processor with multiple cores. The idea is the following: the parser translates assembler input file to program objects suitable for further processing. After that the static single assignment is done. Based on the data flow graph, the parallelization algorithm separates instructions on different cores. Once sequential code is parallelized by the parallelization algorithm, registers are allocated with the algorithm for linear allocation, and the result at the end of the program is distributed assembler code on each of the cores. In the paper we evaluate the speedup of the matrix multiplication example, which was processed by the parallelizator of assembly code. The result is almost linear speedup of code execution, which increases with the number of cores. The speed up on the two cores is 1.99, while on 16 cores the speed up is 13.88.

  6. Parallel Atomistic Simulations

    Energy Technology Data Exchange (ETDEWEB)

    HEFFELFINGER,GRANT S.

    2000-01-18

    Algorithms developed to enable the use of atomistic molecular simulation methods with parallel computers are reviewed. Methods appropriate for bonded as well as non-bonded (and charged) interactions are included. While strategies for obtaining parallel molecular simulations have been developed for the full variety of atomistic simulation methods, molecular dynamics and Monte Carlo have received the most attention. Three main types of parallel molecular dynamics simulations have been developed, the replicated data decomposition, the spatial decomposition, and the force decomposition. For Monte Carlo simulations, parallel algorithms have been developed which can be divided into two categories, those which require a modified Markov chain and those which do not. Parallel algorithms developed for other simulation methods such as Gibbs ensemble Monte Carlo, grand canonical molecular dynamics, and Monte Carlo methods for protein structure determination are also reviewed and issues such as how to measure parallel efficiency, especially in the case of parallel Monte Carlo algorithms with modified Markov chains are discussed.

  7. Automatic Parallelization Tool: Classification of Program Code for Parallel Computing

    Directory of Open Access Journals (Sweden)

    Mustafa Basthikodi

    2016-04-01

    Full Text Available Performance growth of single-core processors has come to a halt in the past decade, but was re-enabled by the introduction of parallelism in processors. Multicore frameworks along with Graphical Processing Units empowered to enhance parallelism broadly. Couples of compilers are updated to developing challenges forsynchronization and threading issues. Appropriate program and algorithm classifications will have advantage to a great extent to the group of software engineers to get opportunities for effective parallelization. In present work we investigated current species for classification of algorithms, in that related work on classification is discussed along with the comparison of issues that challenges the classification. The set of algorithms are chosen which matches the structure with different issues and perform given task. We have tested these algorithms utilizing existing automatic species extraction toolsalong with Bones compiler. We have added functionalities to existing tool, providing a more detailed characterization. The contributions of our work include support for pointer arithmetic, conditional and incremental statements, user defined types, constants and mathematical functions. With this, we can retain significant data which is not captured by original speciesof algorithms. We executed new theories into the device, empowering automatic characterization of program code.

  8. Should the moral core of climate issues be emphasized or downplayed in public discourse? Three ways to successfully manage the double-edged sword of moral communication

    NARCIS (Netherlands)

    Täuber, Susanne; van Zomeren, Martijn; Kutlaca, Maja

    The main objective of this paper is to identify a serious problem for communicators regarding the framing of climate issues in public discourse, namely that moralizing such an issue can motivate individuals while at the same time defensively lead them to avoid solving the problem. We review recent

  9. Ice Sheets & Ice Cores

    DEFF Research Database (Denmark)

    Mikkelsen, Troels Bøgeholm

    Since the discovery of the Ice Ages it has been evident that Earth’s climate is liable to undergo dramatic changes. The previous climatic period known as the Last Glacial saw large oscillations in the extent of ice sheets covering the Northern hemisphere. Understanding these oscillations known....... The first part concerns time series analysis of ice core data obtained from the Greenland Ice Sheet. We analyze parts of the time series where DO-events occur using the so-called transfer operator and compare the results with time series from a simple model capable of switching by either undergoing...

  10. Parallelization in Modern C++

    CERN Multimedia

    CERN. Geneva

    2016-01-01

    The traditionally used and well established parallel programming models OpenMP and MPI are both targeting lower level parallelism and are meant to be as language agnostic as possible. For a long time, those models were the only widely available portable options for developing parallel C++ applications beyond using plain threads. This has strongly limited the optimization capabilities of compilers, has inhibited extensibility and genericity, and has restricted the use of those models together with other, modern higher level abstractions introduced by the C++11 and C++14 standards. The recent revival of interest in the industry and wider community for the C++ language has also spurred a remarkable amount of standardization proposals and technical specifications being developed. Those efforts however have so far failed to build a vision on how to seamlessly integrate various types of parallelism, such as iterative parallel execution, task-based parallelism, asynchronous many-task execution flows, continuation s...

  11. Parallelism in matrix computations

    CERN Document Server

    Gallopoulos, Efstratios; Sameh, Ahmed H

    2016-01-01

    This book is primarily intended as a research monograph that could also be used in graduate courses for the design of parallel algorithms in matrix computations. It assumes general but not extensive knowledge of numerical linear algebra, parallel architectures, and parallel programming paradigms. The book consists of four parts: (I) Basics; (II) Dense and Special Matrix Computations; (III) Sparse Matrix Computations; and (IV) Matrix functions and characteristics. Part I deals with parallel programming paradigms and fundamental kernels, including reordering schemes for sparse matrices. Part II is devoted to dense matrix computations such as parallel algorithms for solving linear systems, linear least squares, the symmetric algebraic eigenvalue problem, and the singular-value decomposition. It also deals with the development of parallel algorithms for special linear systems such as banded ,Vandermonde ,Toeplitz ,and block Toeplitz systems. Part III addresses sparse matrix computations: (a) the development of pa...

  12. Tectonic/climatic control on sediment provenance in the Cape Roberts Project core record (southern Victoria Land, Antarctica): A pulsing late Oligocene/early Miocene signal from south revealed by detrital thermochronology

    Science.gov (United States)

    Olivetti, V.; Balestrieri, M. L.; Rossetti, F.; Talarico, F. M.

    2012-04-01

    The Mesozoic-Cenozoic West Antarctic Rift System (WARS) is one of the largest intracontinental rift on Earth. The Transantarctic Mountains (TAM) form its western shoulder, marking the boundary between the East and West Antarctica. The rifting evolution is commonly considered polyphase and involves an Early Cretaceous phase linked to the Gondwana break-up followed by a major Cenozoic one, starting at c. 50-40 Ma. This Cenozoic episode corresponds to the major uplift/denudation phase of the TAM, which occurred concurrently with transition from orthogonal to oblique rifting. The Cenozoic rift reorganization occurred concurrently with a major change in the global climate system and a global reorganization of plate motions. This area thus provide an outstanding natural laboratory for studying a range of geological problems that involve feedback relationships between tectonics and climate. A key to address the tectonic/climate feedback relations is to look on apparent synchronicity in erosion signal between different segments, and to compare these with well-dated regional and global climatic events. However, due to the paucity of Cenozoic rock sequences exposed along the TAM front, a few information is available about the neotectonics of the rift and rift-flank uplift system. The direct physical record of the tectonic/climate history of the WARS recovered by core drillings along the western margin of the Ross sea (DSDP, CIROS, Cape Roberts and ANDRILL projects) provides an invaluable tool to address this issue. Twenty-three samples distributed throughout the entire composite drill-cored stratigraphic succession of Cape Roberts were analyzed. Age probability plots of eighteen detrital samples with depositional ages between 34 Ma and the Pliocene were decomposed into statistically significant age populations or peaks using binomial peak-fitting. Moreover, three granitic pebbles, one dolerite clast and one sample of Beacon sandstones have been dated. From detrital samples

  13. A parallel buffer tree

    DEFF Research Database (Denmark)

    Sitchinava, Nodar; Zeh, Norbert

    2012-01-01

    We present the parallel buffer tree, a parallel external memory (PEM) data structure for batched search problems. This data structure is a non-trivial extension of Arge's sequential buffer tree to a private-cache multiprocessor environment and reduces the number of I/O operations by the number of...... in the optimal OhOf(psortN + K/PB) parallel I/O complexity, where K is the size of the output reported in the process and psortN is the parallel I/O complexity of sorting N elements using P processors....

  14. Parallel MR imaging.

    Science.gov (United States)

    Deshmane, Anagha; Gulani, Vikas; Griswold, Mark A; Seiberlich, Nicole

    2012-07-01

    Parallel imaging is a robust method for accelerating the acquisition of magnetic resonance imaging (MRI) data, and has made possible many new applications of MR imaging. Parallel imaging works by acquiring a reduced amount of k-space data with an array of receiver coils. These undersampled data can be acquired more quickly, but the undersampling leads to aliased images. One of several parallel imaging algorithms can then be used to reconstruct artifact-free images from either the aliased images (SENSE-type reconstruction) or from the undersampled data (GRAPPA-type reconstruction). The advantages of parallel imaging in a clinical setting include faster image acquisition, which can be used, for instance, to shorten breath-hold times resulting in fewer motion-corrupted examinations. In this article the basic concepts behind parallel imaging are introduced. The relationship between undersampling and aliasing is discussed and two commonly used parallel imaging methods, SENSE and GRAPPA, are explained in detail. Examples of artifacts arising from parallel imaging are shown and ways to detect and mitigate these artifacts are described. Finally, several current applications of parallel imaging are presented and recent advancements and promising research in parallel imaging are briefly reviewed. Copyright © 2012 Wiley Periodicals, Inc.

  15. Parallel Algorithms and Patterns

    Energy Technology Data Exchange (ETDEWEB)

    Robey, Robert W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-06-16

    This is a powerpoint presentation on parallel algorithms and patterns. A parallel algorithm is a well-defined, step-by-step computational procedure that emphasizes concurrency to solve a problem. Examples of problems include: Sorting, searching, optimization, matrix operations. A parallel pattern is a computational step in a sequence of independent, potentially concurrent operations that occurs in diverse scenarios with some frequency. Examples are: Reductions, prefix scans, ghost cell updates. We only touch on parallel patterns in this presentation. It really deserves its own detailed discussion which Gabe Rockefeller would like to develop.

  16. Application Portable Parallel Library

    Science.gov (United States)

    Cole, Gary L.; Blech, Richard A.; Quealy, Angela; Townsend, Scott

    1995-01-01

    Application Portable Parallel Library (APPL) computer program is subroutine-based message-passing software library intended to provide consistent interface to variety of multiprocessor computers on market today. Minimizes effort needed to move application program from one computer to another. User develops application program once and then easily moves application program from parallel computer on which created to another parallel computer. ("Parallel computer" also include heterogeneous collection of networked computers). Written in C language with one FORTRAN 77 subroutine for UNIX-based computers and callable from application programs written in C language or FORTRAN 77.

  17. Combined ice core and climate-model evidence for the collapse of the West Antarctic Ice Sheet during Marine Isotope Stage 5e.

    Science.gov (United States)

    Steig, Eric J.; Huybers, Kathleen; Singh, Hansi A.; Steiger, Nathan J.; Frierson, Dargan M. W.; Popp, Trevor; White, James W. C.

    2015-04-01

    It has been speculated that collapse of the West Antarctic Ice Sheet explains the very high eustatic sea level rise during the last interglacial period, marine isotope stage (MIS) 5e, but the evidence remains equivocal. Changes in atmospheric circulation resulting from a collapse of the West Antarctic Ice Sheet (WAIS) would have significant regional impacts that should be detectable in ice core records. We conducted simulations using general circulation models (GCMs) at varying levels of complexity: a gray-radiation aquaplanet moist GCM (GRaM), the slab ocean version of GFDL-AM2 (also as an aquaplanet), and the fully-coupled version of NCAR's CESM with realistic topography. In all the experiments, decreased elevation from the removal of the WAIS leads to greater cyclonic circulation over the West Antarctic region. This creates increased advection of relatively warm marine air from the Amundsen-Bellingshausen Seas towards the South Pole, and increased cold-air advection from the East Antarctic plateau towards the Ross Sea and coastal Marie Byrd Land. The result is anomalous warming in some areas of the East Antarctic interior, and significant cooling in Marie Byrd Land. Comparison of ice core records shows good agreement with the model predictions. In particular, isotope-paleotemperature records from ice cores in East Antarctica warmed more between the previous glacial period (MIS 6) and MIS 5e than coastal Marie Byrd Land. These results add substantial support to other evidence for WAIS collapse during the last interglacial period.

  18. Fiscal 2000 achievement report on the venture business assisting type regional consortium - Core industry creation type. Research and development of high-accuracy fabrication technology of aspheric optical component creation technology using parallel grinding method; 2000 nendo chiiki consortium kenkyu kaihatsu jigyo seika hokokusho. Parallel kensaku hoshiki ni yoru koseido hikyumen kogaku soshi sosei gijutsu no kenkyu kaihatsu

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2001-03-01

    Research and development was carried out for the establishment and commercialization of the technical seeds of a newly proposed 'parallel grinding type aspheric surface grinding method,' based on the results of studies of ultra-accuracy grinding and aspheric surface machining technologies conducted at Tohoku University. The new method is characterized in that a work point on the work piece faces the grinding point on the grindstone cross section on the one-to-one basis because the grinding point travels on the grindstone cross section. Accordingly, the method is quite suitable for corrective grinding for eliminating only the error in case a geometrical error occurs on the wrought surface. For this purpose, measurement of the machine surface on board the machine is indispensable. Hence: a study of corrective grinding methods under an on-board measuring system. As the result, the initially intended geometrical accuracy level was attained in the machining of axis-symmetric aspheric glass lenses and aspheric metal molds, and coarseness on the finished surface in the grinding of glass lenses was reduced to 0.3{mu}m using a diamond grindstone. In conclusion, machining accuracy and finished surface coarseness were remarkably improved, and this enables machining without the need of polishing and demonstrates that there are possibilities of commercialization. (NEDO)

  19. Parallel discrete event simulation

    NARCIS (Netherlands)

    Overeinder, B.J.; Hertzberger, L.O.; Sloot, P.M.A.; Withagen, W.J.

    1991-01-01

    In simulating applications for execution on specific computing systems, the simulation performance figures must be known in a short period of time. One basic approach to the problem of reducing the required simulation time is the exploitation of parallelism. However, in parallelizing the simulation

  20. Parallel reservoir simulator computations

    International Nuclear Information System (INIS)

    Hemanth-Kumar, K.; Young, L.C.

    1995-01-01

    The adaptation of a reservoir simulator for parallel computations is described. The simulator was originally designed for vector processors. It performs approximately 99% of its calculations in vector/parallel mode and relative to scalar calculations it achieves speedups of 65 and 81 for black oil and EOS simulations, respectively on the CRAY C-90

  1. Rubus: A compiler for seamless and extensible parallelism.

    Directory of Open Access Journals (Sweden)

    Muhammad Adnan

    Full Text Available Nowadays, a typical processor may have multiple processing cores on a single chip. Furthermore, a special purpose processing unit called Graphic Processing Unit (GPU, originally designed for 2D/3D games, is now available for general purpose use in computers and mobile devices. However, the traditional programming languages which were designed to work with machines having single core CPUs, cannot utilize the parallelism available on multi-core processors efficiently. Therefore, to exploit the extraordinary processing power of multi-core processors, researchers are working on new tools and techniques to facilitate parallel programming. To this end, languages like CUDA and OpenCL have been introduced, which can be used to write code with parallelism. The main shortcoming of these languages is that programmer needs to specify all the complex details manually in order to parallelize the code across multiple cores. Therefore, the code written in these languages is difficult to understand, debug and maintain. Furthermore, to parallelize legacy code can require rewriting a significant portion of code in CUDA or OpenCL, which can consume significant time and resources. Thus, the amount of parallelism achieved is proportional to the skills of the programmer and the time spent in code optimizations. This paper proposes a new open source compiler, Rubus, to achieve seamless parallelism. The Rubus compiler relieves the programmer from manually specifying the low-level details. It analyses and transforms a sequential program into a parallel program automatically, without any user intervention. This achieves massive speedup and better utilization of the underlying hardware without a programmer's expertise in parallel programming. For five different benchmarks, on average a speedup of 34.54 times has been achieved by Rubus as compared to Java on a basic GPU having only 96 cores. Whereas, for a matrix multiplication benchmark the average execution speedup of 84

  2. Totally parallel multilevel algorithms

    Science.gov (United States)

    Frederickson, Paul O.

    1988-01-01

    Four totally parallel algorithms for the solution of a sparse linear system have common characteristics which become quite apparent when they are implemented on a highly parallel hypercube such as the CM2. These four algorithms are Parallel Superconvergent Multigrid (PSMG) of Frederickson and McBryan, Robust Multigrid (RMG) of Hackbusch, the FFT based Spectral Algorithm, and Parallel Cyclic Reduction. In fact, all four can be formulated as particular cases of the same totally parallel multilevel algorithm, which are referred to as TPMA. In certain cases the spectral radius of TPMA is zero, and it is recognized to be a direct algorithm. In many other cases the spectral radius, although not zero, is small enough that a single iteration per timestep keeps the local error within the required tolerance.

  3. Parallel computing works

    Energy Technology Data Exchange (ETDEWEB)

    1991-10-23

    An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

  4. Massively parallel mathematical sieves

    Energy Technology Data Exchange (ETDEWEB)

    Montry, G.R.

    1989-01-01

    The Sieve of Eratosthenes is a well-known algorithm for finding all prime numbers in a given subset of integers. A parallel version of the Sieve is described that produces computational speedups over 800 on a hypercube with 1,024 processing elements for problems of fixed size. Computational speedups as high as 980 are achieved when the problem size per processor is fixed. The method of parallelization generalizes to other sieves and will be efficient on any ensemble architecture. We investigate two highly parallel sieves using scattered decomposition and compare their performance on a hypercube multiprocessor. A comparison of different parallelization techniques for the sieve illustrates the trade-offs necessary in the design and implementation of massively parallel algorithms for large ensemble computers.

  5. 40Ar/ 39Ar, K-Ar and 230Th- 238U dating of the Laschamp excursion: A radioisotopic tie-point for ice core and climate chronologies

    Science.gov (United States)

    Singer, Brad S.; Guillou, Hervé; Jicha, Brian R.; Laj, Carlo; Kissel, Catherine; Beard, Brian L.; Johnson, Clark M.

    2009-08-01

    A brief period of enhanced 10Be flux that straddles the interstadial warm period known as Dansgaard-Oeschger event 10 in Greenland and its counterpart in Antarctica, the Antarctic Isotope Maximum 10 is but one consequence of the weakening of Earth's magnetic field associated with the Laschamp excursion. This 10Be peak measured in the GRIP ice core is dated at 41,250 y b2k (= before year 2000 AD) in the most recent GICC05 age model obtained from the NorthGRIP core via multi-parameter counting of annual layers. Uncertainty in the age of the 10Be peak is, however, no better than ± 1630 y at the 95% confidence level, reflecting accumulated error in identifying annual layers. The age of the Laschamp excursion [Guillou, H., Singer, B.S., Laj, C., Kissel, C., Scaillet, S., Jicha, B., 2004. On the age of the Laschamp geomagnetic excursion. Earth Planet. Sci. Lett. 227, 331-343.] is revised on the basis of new 40Ar/ 39Ar, unspiked K-Ar and 238U- 230Th data from three lava flows in the Massif Central, France, together with the 40Ar/ 39Ar age of a transitionally magnetized lava flow at Auckland, New Zealand. Combined, these data yield an age of 40,700 ± 950 y b2k, where the uncertainty includes both analytical and systematic ( 40K and 230Th decay constant) errors. Taking the radioisotopic age as a calibration tie point suggests that the layer-counting chronologies for the NorthGRIP and GISP2 ice cores are more accurate and precise than previously thought at depths corresponding to the Laschamp excursion.

  6. Parallelization of the Physical-Space Statistical Analysis System (PSAS)

    Science.gov (United States)

    Larson, J. W.; Guo, J.; Lyster, P. M.

    1999-01-01

    Atmospheric data assimilation is a method of combining observations with model forecasts to produce a more accurate description of the atmosphere than the observations or forecast alone can provide. Data assimilation plays an increasingly important role in the study of climate and atmospheric chemistry. The NASA Data Assimilation Office (DAO) has developed the Goddard Earth Observing System Data Assimilation System (GEOS DAS) to create assimilated datasets. The core computational components of the GEOS DAS include the GEOS General Circulation Model (GCM) and the Physical-space Statistical Analysis System (PSAS). The need for timely validation of scientific enhancements to the data assimilation system poses computational demands that are best met by distributed parallel software. PSAS is implemented in Fortran 90 using object-based design principles. The analysis portions of the code solve two equations. The first of these is the "innovation" equation, which is solved on the unstructured observation grid using a preconditioned conjugate gradient (CG) method. The "analysis" equation is a transformation from the observation grid back to a structured grid, and is solved by a direct matrix-vector multiplication. Use of a factored-operator formulation reduces the computational complexity of both the CG solver and the matrix-vector multiplication, rendering the matrix-vector multiplications as a successive product of operators on a vector. Sparsity is introduced to these operators by partitioning the observations using an icosahedral decomposition scheme. PSAS builds a large (approx. 128MB) run-time database of parameters used in the calculation of these operators. Implementing a message passing parallel computing paradigm into an existing yet developing computational system as complex as PSAS is nontrivial. One of the technical challenges is balancing the requirements for computational reproducibility with the need for high performance. The problem of computational

  7. Parallel transposition of sparse data structures

    DEFF Research Database (Denmark)

    Wang, Hao; Liu, Weifeng; Hou, Kaixi

    2016-01-01

    Many applications in computational sciences and social sciences exploit sparsity and connectivity of acquired data. Even though many parallel sparse primitives such as sparse matrix-vector (SpMV) multiplication have been extensively studied, some other important building blocks, e.g., parallel tr...... transposition in the latest vendor-supplied library on an Intel multicore CPU platform, and the MergeTrans approach achieves on average of 3.4-fold (up to 11.7-fold) speedup on an Intel Xeon Phi many-core processor....

  8. High-resolution paleoclimatology of the Santa Barbara Basin during the Medieval Climate Anomaly and early Little Ice Age based on diatom and silicoflagellate assemblages in Kasten core SPR0901-02KC

    Science.gov (United States)

    Barron, John A.; Bukry, David B.; Hendy, Ingrid L.

    2015-01-01

    Diatom and silicoflagellate assemblages documented in a high-resolution time series spanning 800 to 1600 AD in varved sediment recovered in Kasten core SPR0901-02KC (34°16.845’ N, 120°02.332’ W, water depth 588 m) from the Santa Barbara Basin (SBB) reveal that SBB surface water conditions during the Medieval Climate Anomaly (MCA) and the early part of the Little Ice Age (LIA) were not extreme by modern standards, mostly falling within one standard deviation of mean conditions during the pre anthropogenic interval of 1748 to 1900. No clear differences between the character of MCA and the early LIA conditions are apparent. During intervals of extreme droughts identified by terrigenous proxy scanning XRF analyses, diatom and silicoflagellate proxies for coastal upwelling typically exceed one standard deviation above mean values for 1748-1900, supporting the hypothesis that droughts in southern California are associated with cooler (or La Niña-like) sea surface temperatures (SSTs). Increased percentages of diatoms transported downslope generally coincide with intervals of increased siliciclastic flux to the SBB identified by scanning XRF analyses. Diatom assemblages suggest only two intervals of the MCA (at ~897 to 922 and ~1151 to 1167) when proxy SSTs exceeded one standard deviation above mean values for 1748 to 1900. Conversely, silicoflagellates imply extreme warm water events only at ~830 to 860 (early MCA) and ~1360 to 1370 (early LIA) that are not supported by the diatom data. Silicoflagellates appear to be more suitable for characterizing average climate during the 5 to 11 year-long sample intervals studied in the SPR0901-02KC core than diatoms, probably because diatom relative abundances may be dominated by seasonal blooms of a particular year.

  9. Transformer core

    NARCIS (Netherlands)

    Mehendale, A.; Hagedoorn, Wouter; Lötters, Joost Conrad

    2008-01-01

    A transformer core includes a stack of a plurality of planar core plates of a magnetically permeable material, which plates each consist of a first and a second sub-part that together enclose at least one opening. The sub-parts can be fitted together via contact faces that are located on either side

  10. Transformer core

    NARCIS (Netherlands)

    Mehendale, A.; Hagedoorn, Wouter; Lötters, Joost Conrad

    2010-01-01

    A transformer core includes a stack of a plurality of planar core plates of a magnetically permeable material, which plates each consist of a first and a second sub-part that together enclose at least one opening. The sub-parts can be fitted together via contact faces that are located on either side

  11. Ground Penetrating Radar, Magnetic and Compositional Analysis of Sediment Cores and Surface Samples: The Relationships Between Lacustrine Sediments and Holocene Lake- Level and Climate Change at Deming Lake, Minnesota, USA

    Science.gov (United States)

    Murray, R.; Lascu, I.; Plank, C.

    2007-12-01

    Deming Lake is a small (Deming. Cores were sampled continuously at a 1-2 cm resolution and sediment composition (in terms of percent organic matter, carbonate material, and minerogenic residue) was determined via loss on ignition (LOI). Isothermal remanent magnetization (IRM) and anhysteretic remanent magnetization (ARM) were used as proxies of magnetic mineral concentration and grain size. Four lithostratigraphic units were identified and correlated between cores based on these analyses. Changes in GPR facies corroborate the correlation between the two shallow cores. In order to inform our interpretation of down-core variations in magnetic properties and LOI values in terms of variations in lake depth, a suite of over 70 modern sediment samples were collected from the basin and analyzed. LOI compositional variability across the basin was high, with no clear trends related to depth or distance from shore. A sharp decrease in minerogenic content was observed at depths consistent with a predicted wave-base of 0.5 m, but aside from this trend it appears the steep slopes of much of the basin promote gravity driven slumping and mixing of sediments at depth. In the profundal sediments IRM values are routinely 5% higher than in the slope and littoral environments, while ARM/IRM ratios indicate an increase in magnetic grain size with water depth. We infer that an increase in coarse organic material in the shallow-water cores of Deming records a period of aridity (associated with a decrease lake-level less than 2m based on GPR profiles) and/or increased water clarity during the regionally expansive mid-Holocene dry period. We do not see clear evidence of late-Holocene lake level change of a significant magnitude (i.e. >1m). While remanence measurements (especially IRM) often correlate with the LOI residue, interference in the IRM resulting from the dissolution of magnetic minerals casts uncertainty into the reliability of our magnetic measurements as a signal of climate

  12. Options for Parallelizing a Planning and Scheduling Algorithm

    Science.gov (United States)

    Clement, Bradley J.; Estlin, Tara A.; Bornstein, Benjamin D.

    2011-01-01

    Space missions have a growing interest in putting multi-core processors onboard spacecraft. For many missions processing power significantly slows operations. We investigate how continual planning and scheduling algorithms can exploit multi-core processing and outline different potential design decisions for a parallelized planning architecture. This organization of choices and challenges helps us with an initial design for parallelizing the CASPER planning system for a mesh multi-core processor. This work extends that presented at another workshop with some preliminary results.

  13. SCORPIO - VVER core surveillance system

    International Nuclear Information System (INIS)

    Zalesky, K.; Svarny, J.; Novak, L.; Rosol, J.; Horanes, A.

    1997-01-01

    The Halden Project has developed the core surveillance system SCORPIO which has two parallel modes of operation: the Core Follow Mode and the Predictive Mode. The main motivation behind the development of SCORPIO is to make a practical tool for reactor operators which can increase the quality and quantity of information presented on core status and dynamic behavior. This can first of all improve plant safety as undesired core conditions are detected and prevented. Secondly, more flexible and efficient plant operation is made possible. So far the system has only been implemented on western PWRs but the basic concept is applicable to a wide range of reactor including WWERs. The main differences between WWERs and typical western PWRs with respect to core surveillance requirements are outlined. The development of a WWER version of SCORPIO was initiated in cooperation with the Nuclear Research Institute at Rez and industry partners in the Czech Republic. The first system will be installed at the Dukovany NPP. (author)

  14. Parallelizing the spectral transform method: A comparison of alternative parallel algorithms

    International Nuclear Information System (INIS)

    Foster, I.; Worley, P.H.

    1993-01-01

    The spectral transform method is a standard numerical technique for solving partial differential equations on the sphere and is widely used in global climate modeling. In this paper, we outline different approaches to parallelizing the method and describe experiments that we are conducting to evaluate the efficiency of these approaches on parallel computers. The experiments are conducted using a testbed code that solves the nonlinear shallow water equations on a sphere, but are designed to permit evaluation in the context of a global model. They allow us to evaluate the relative merits of the approaches as a function of problem size and number of processors. The results of this study are guiding ongoing work on PCCM2, a parallel implementation of the Community Climate Model developed at the National Center for Atmospheric Research

  15. Current distribution characteristics of superconducting parallel circuits

    International Nuclear Information System (INIS)

    Mori, K.; Suzuki, Y.; Hara, N.; Kitamura, M.; Tominaka, T.

    1994-01-01

    In order to increase the current carrying capacity of the current path of the superconducting magnet system, the portion of parallel circuits such as insulated multi-strand cables or parallel persistent current switches (PCS) are made. In superconducting parallel circuits of an insulated multi-strand cable or a parallel persistent current switch (PCS), the current distribution during the current sweep, the persistent mode, and the quench process were investigated. In order to measure the current distribution, two methods were used. (1) Each strand was surrounded with a pure iron core with the air gap. In the air gap, a Hall probe was located. The accuracy of this method was deteriorated by the magnetic hysteresis of iron. (2) The Rogowski coil without iron was used for the current measurement of each path in a 4-parallel PCS. As a result, it was shown that the current distribution characteristics of a parallel PCS is very similar to that of an insulated multi-strand cable for the quench process

  16. Parallelized Seeded Region Growing Using CUDA

    Directory of Open Access Journals (Sweden)

    Seongjin Park

    2014-01-01

    Full Text Available This paper presents a novel method for parallelizing the seeded region growing (SRG algorithm using Compute Unified Device Architecture (CUDA technology, with intention to overcome the theoretical weakness of SRG algorithm of its computation time being directly proportional to the size of a segmented region. The segmentation performance of the proposed CUDA-based SRG is compared with SRG implementations on single-core CPUs, quad-core CPUs, and shader language programming, using synthetic datasets and 20 body CT scans. Based on the experimental results, the CUDA-based SRG outperforms the other three implementations, advocating that it can substantially assist the segmentation during massive CT screening tests.

  17. Massive Asynchronous Parallelization of Sparse Matrix Factorizations

    Energy Technology Data Exchange (ETDEWEB)

    Chow, Edmond [Georgia Inst. of Technology, Atlanta, GA (United States)

    2018-01-08

    Solving sparse problems is at the core of many DOE computational science applications. We focus on the challenge of developing sparse algorithms that can fully exploit the parallelism in extreme-scale computing systems, in particular systems with massive numbers of cores per node. Our approach is to express a sparse matrix factorization as a large number of bilinear constraint equations, and then solving these equations via an asynchronous iterative method. The unknowns in these equations are the matrix entries of the factorization that is desired.

  18. Parallel execution of chemical software on EGEE Grid

    CERN Document Server

    Sterzel, Mariusz

    2008-01-01

    Constant interest among chemical community to study larger and larger molecules forces the parallelization of existing computational methods in chemistry and development of new ones. These are main reasons of frequent port updates and requests from the community for the Grid ports of new packages to satisfy their computational demands. Unfortunately some parallelization schemes used by chemical packages cannot be directly used in Grid environment. Here we present a solution for Gaussian package. The current state of development of Grid middleware allows easy parallel execution in case of software using any of MPI flavour. Unfortunately many chemical packages do not use MPI for parallelization therefore special treatment is needed. Gaussian can be executed in parallel on SMP architecture or via Linda. These require reservation of certain number of processors/cores on a given WN and the equal number of processors/cores on each WN, respectively. The current implementation of EGEE middleware does not offer such f...

  19. A hybrid algorithm for parallel molecular dynamics simulations

    Science.gov (United States)

    Mangiardi, Chris M.; Meyer, R.

    2017-10-01

    This article describes algorithms for the hybrid parallelization and SIMD vectorization of molecular dynamics simulations with short-range forces. The parallelization method combines domain decomposition with a thread-based parallelization approach. The goal of the work is to enable efficient simulations of very large (tens of millions of atoms) and inhomogeneous systems on many-core processors with hundreds or thousands of cores and SIMD units with large vector sizes. In order to test the efficiency of the method, simulations of a variety of configurations with up to 74 million atoms have been performed. Results are shown that were obtained on multi-core systems with Sandy Bridge and Haswell processors as well as systems with Xeon Phi many-core processors.

  20. Core lifter

    Energy Technology Data Exchange (ETDEWEB)

    Pavlov, N G; Edel' man, Ya A

    1981-02-15

    A core lifter is suggested which contains a housing, core-clamping elements installed in the housing depressions in the form of semirings with projections on the outer surface restricting the rotation of the semirings in the housing depressions. In order to improve the strength and reliability of the core lifter, the semirings have a variable transverse section formed from the outside by the surface of the rotation body of the inner arc of the semiring aroung the rotation axis and from the inner a cylindrical surface which is concentric to the outer arc of the semiring. The core-clamping elements made in this manner have the possibility of freely rotating in the housing depressions under their own weight and from contact with the core sample. These semirings do not have weakened sections, have sufficient strength, are inserted into the limited ring section of the housing of the core lifter without reduction in its through opening and this improve the reliability of the core lifter in operation.

  1. Algorithms for parallel computers

    International Nuclear Information System (INIS)

    Churchhouse, R.F.

    1985-01-01

    Until relatively recently almost all the algorithms for use on computers had been designed on the (usually unstated) assumption that they were to be run on single processor, serial machines. With the introduction of vector processors, array processors and interconnected systems of mainframes, minis and micros, however, various forms of parallelism have become available. The advantage of parallelism is that it offers increased overall processing speed but it also raises some fundamental questions, including: (i) which, if any, of the existing 'serial' algorithms can be adapted for use in the parallel mode. (ii) How close to optimal can such adapted algorithms be and, where relevant, what are the convergence criteria. (iii) How can we design new algorithms specifically for parallel systems. (iv) For multi-processor systems how can we handle the software aspects of the interprocessor communications. Aspects of these questions illustrated by examples are considered in these lectures. (orig.)

  2. Parallelism and array processing

    International Nuclear Information System (INIS)

    Zacharov, V.

    1983-01-01

    Modern computing, as well as the historical development of computing, has been dominated by sequential monoprocessing. Yet there is the alternative of parallelism, where several processes may be in concurrent execution. This alternative is discussed in a series of lectures, in which the main developments involving parallelism are considered, both from the standpoint of computing systems and that of applications that can exploit such systems. The lectures seek to discuss parallelism in a historical context, and to identify all the main aspects of concurrency in computation right up to the present time. Included will be consideration of the important question as to what use parallelism might be in the field of data processing. (orig.)

  3. The new landscape of parallel computer architecture

    Energy Technology Data Exchange (ETDEWEB)

    Shalf, John [NERSC Division, Lawrence Berkeley National Laboratory 1 Cyclotron Road, Berkeley California, 94720 (United States)

    2007-07-15

    The past few years has seen a sea change in computer architecture that will impact every facet of our society as every electronic device from cell phone to supercomputer will need to confront parallelism of unprecedented scale. Whereas the conventional multicore approach (2, 4, and even 8 cores) adopted by the computing industry will eventually hit a performance plateau, the highest performance per watt and per chip area is achieved using manycore technology (hundreds or even thousands of cores). However, fully unleashing the potential of the manycore approach to ensure future advances in sustained computational performance will require fundamental advances in computer architecture and programming models that are nothing short of reinventing computing. In this paper we examine the reasons behind the movement to exponentially increasing parallelism, and its ramifications for system design, applications and programming models.

  4. The new landscape of parallel computer architecture

    International Nuclear Information System (INIS)

    Shalf, John

    2007-01-01

    The past few years has seen a sea change in computer architecture that will impact every facet of our society as every electronic device from cell phone to supercomputer will need to confront parallelism of unprecedented scale. Whereas the conventional multicore approach (2, 4, and even 8 cores) adopted by the computing industry will eventually hit a performance plateau, the highest performance per watt and per chip area is achieved using manycore technology (hundreds or even thousands of cores). However, fully unleashing the potential of the manycore approach to ensure future advances in sustained computational performance will require fundamental advances in computer architecture and programming models that are nothing short of reinventing computing. In this paper we examine the reasons behind the movement to exponentially increasing parallelism, and its ramifications for system design, applications and programming models

  5. Parallel channel effects under BWR LOCA conditions

    International Nuclear Information System (INIS)

    Suzuki, H.; Hatamiya, S.; Murase, M.

    1988-01-01

    Due to parallel channel effects, different flow patterns such as liquid down-flow and gas up-flow appear simultaneously in fuel bundles of a BWR core during postulated LOCAs. Applying the parallel channel effects to the fuel bundle, water drain tubes with a restricted bottom end have been developed in order to mitigate counter-current flow limiting and to increase the falling water flow rate at the upper tie plate. The upper tie plate with water drain tubes is an especially effective means of increasing the safety margin of a reactor with narrow gaps between fuel rods and high steam velocity at the upper tie plate. The characteristics of the water drain tubes have been experimentally investigated using a small-scaled steam-water system simulating a BWR core. Then, their effect on the fuel cladding temperature was evaluated using the LOCA analysis program SAFER. (orig.)

  6. Toward an ultra-high resolution community climate system model for the BlueGene platform

    International Nuclear Information System (INIS)

    Dennis, John M; Jacob, Robert; Vertenstein, Mariana; Craig, Tony; Loy, Raymond

    2007-01-01

    Global climate models need to simulate several small, regional-scale processes which affect the global circulation in order to accurately simulate the climate. This is particularly important in the ocean where small scale features such as oceanic eddies are currently represented with adhoc parameterizations. There is also a need for higher resolution to provide climate predictions at small, regional scales. New high-performance computing platforms such as the IBM BlueGene can provide the necessary computational power to perform ultra-high resolution climate model integrations. We have begun to investigate the scaling of the individual components of the Community Climate System Model to prepare it for integrations on BlueGene and similar platforms. Our investigations show that it is possible to successfully utilize O(32K) processors. We describe the scalability of five models: the Parallel Ocean Program (POP), the Community Ice CodE (CICE), the Community Land Model (CLM), and the new CCSM sequential coupler (CPL7) which are components of the next generation Community Climate System Model (CCSM); as well as the High-Order Method Modeling Environment (HOMME) which is a dynamical core currently being evaluated within the Community Atmospheric Model. For our studies we concentrate on 1/10 0 resolution for CICE, POP, and CLM models and 1/4 0 resolution for HOMME. The ability to simulate high resolutions on the massively parallel petascale systems that will dominate high-performance computing for the foreseeable future is essential to the advancement of climate science

  7. Emergency core cooling device

    International Nuclear Information System (INIS)

    Suzaki, Kiyoshi; Inoue, Akihiro.

    1979-01-01

    Purpose: To improve core cooling effect by making the operation region for a plurality of water injection pumps more broader. Constitution: An emergency reactor core cooling device actuated upon failure of recycling pipe ways is adapted to be fed with cooling water through a thermal sleeve by way of a plurality of water injection pump from pool water in a condensate storage tank and a pressure suppression chamber as water feed source. Exhaust pipes and suction pipes of each of the pumps are connected by way of switching valves and the valves are switched so that the pumps are set to a series operation if the pressure in the pressure vessel is high and the pumps are set to a parallel operation if the pressure in the pressure vessel is low. (Furukawa, Y.)

  8. Parallel magnetic resonance imaging

    International Nuclear Information System (INIS)

    Larkman, David J; Nunes, Rita G

    2007-01-01

    Parallel imaging has been the single biggest innovation in magnetic resonance imaging in the last decade. The use of multiple receiver coils to augment the time consuming Fourier encoding has reduced acquisition times significantly. This increase in speed comes at a time when other approaches to acquisition time reduction were reaching engineering and human limits. A brief summary of spatial encoding in MRI is followed by an introduction to the problem parallel imaging is designed to solve. There are a large number of parallel reconstruction algorithms; this article reviews a cross-section, SENSE, SMASH, g-SMASH and GRAPPA, selected to demonstrate the different approaches. Theoretical (the g-factor) and practical (coil design) limits to acquisition speed are reviewed. The practical implementation of parallel imaging is also discussed, in particular coil calibration. How to recognize potential failure modes and their associated artefacts are shown. Well-established applications including angiography, cardiac imaging and applications using echo planar imaging are reviewed and we discuss what makes a good application for parallel imaging. Finally, active research areas where parallel imaging is being used to improve data quality by repairing artefacted images are also reviewed. (invited topical review)

  9. Reactor core

    International Nuclear Information System (INIS)

    Azekura, Kazuo; Kurihara, Kunitoshi.

    1992-01-01

    In a BWR type reactor, a great number of pipes (spectral shift pipes) are disposed in the reactor core. Moderators having a small moderating cross section (heavy water) are circulated in the spectral shift pipes to suppress the excess reactivity while increasing the conversion ratio at an initial stage of the operation cycle. After the intermediate stage of the operation cycle in which the reactor core reactivity is lowered, reactivity is increased by circulating moderators having a great moderating cross section (light water) to extend the taken up burnup degree. Further, neutron absorbers such as boron are mixed to the moderator in the spectral shift pipe to control the concentration thereof. With such a constitution, control rods and driving mechanisms are no more necessary, to simplify the structure of the reactor core. This can increase the fuel conversion ratio and control great excess reactivity. Accordingly, a nuclear reactor core of high conversion and high burnup degree can be attained. (I.N.)

  10. Hierarchical approach to optimization of parallel matrix multiplication on large-scale platforms

    KAUST Repository

    Hasanov, Khalid; Quintin, Jean-Noë l; Lastovetsky, Alexey

    2014-01-01

    -scale parallelism in mind. Indeed, while in 1990s a system with few hundred cores was considered a powerful supercomputer, modern top supercomputers have millions of cores. In this paper, we present a hierarchical approach to optimization of message-passing parallel

  11. Introducing 'bones' : a parallelizing source-to-source compiler based on algorithmic skeletons.

    NARCIS (Netherlands)

    Nugteren, C.; Corporaal, H.

    2012-01-01

    Recent advances in multi-core and many-core processors requires programmers to exploit an increasing amount of parallelism from their applications. Data parallel languages such as CUDA and OpenCL make it possible to take advantage of such processors, but still require a large amount of effort from

  12. Kalman Filter Tracking on Parallel Architectures

    International Nuclear Information System (INIS)

    Cerati, Giuseppe; Elmer, Peter; Krutelyov, Slava; Lantz, Steven; Lefebvre, Matthieu; McDermott, Kevin; Riley, Daniel; Tadel, Matevž; Wittich, Peter; Würthwein, Frank; Yagil, Avi

    2016-01-01

    Power density constraints are limiting the performance improvements of modern CPUs. To address this we have seen the introduction of lower-power, multi-core processors such as GPGPU, ARM and Intel MIC. In order to achieve the theoretical performance gains of these processors, it will be necessary to parallelize algorithms to exploit larger numbers of lightweight cores and specialized functions like large vector units. Track finding and fitting is one of the most computationally challenging problems for event reconstruction in particle physics. At the High-Luminosity Large Hadron Collider (HL-LHC), for example, this will be by far the dominant problem. The need for greater parallelism has driven investigations of very different track finding techniques such as Cellular Automata or Hough Transforms. The most common track finding techniques in use today, however, are those based on a Kalman filter approach. Significant experience has been accumulated with these techniques on real tracking detector systems, both in the trigger and offline. They are known to provide high physics performance, are robust, and are in use today at the LHC. Given the utility of the Kalman filter in track finding, we have begun to port these algorithms to parallel architectures, namely Intel Xeon and Xeon Phi. We report here on our progress towards an end-to-end track reconstruction algorithm fully exploiting vectorization and parallelization techniques in a simplified experimental environment

  13. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2004-01-01

    Ice cores are the most direct, continuous, and high resolution archive for Late Quaternary paleoclimate reconstruction. Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate migration strategies for New Zealand. (author). 23 refs., 15 figs., 1 tab

  14. The STAPL Parallel Graph Library

    KAUST Repository

    Harshvardhan,; Fidel, Adam; Amato, Nancy M.; Rauchwerger, Lawrence

    2013-01-01

    This paper describes the stapl Parallel Graph Library, a high-level framework that abstracts the user from data-distribution and parallelism details and allows them to concentrate on parallel graph algorithm development. It includes a customizable

  15. Incorporating Parallel Computing into the Goddard Earth Observing System Data Assimilation System (GEOS DAS)

    Science.gov (United States)

    Larson, Jay W.

    1998-01-01

    Atmospheric data assimilation is a method of combining actual observations with model forecasts to produce a more accurate description of the earth system than the observations or forecast alone can provide. The output of data assimilation, sometimes called the analysis, are regular, gridded datasets of observed and unobserved variables. Analysis plays a key role in numerical weather prediction and is becoming increasingly important for climate research. These applications, and the need for timely validation of scientific enhancements to the data assimilation system pose computational demands that are best met by distributed parallel software. The mission of the NASA Data Assimilation Office (DAO) is to provide datasets for climate research and to support NASA satellite and aircraft missions. The system used to create these datasets is the Goddard Earth Observing System Data Assimilation System (GEOS DAS). The core components of the the GEOS DAS are: the GEOS General Circulation Model (GCM), the Physical-space Statistical Analysis System (PSAS), the Observer, the on-line Quality Control (QC) system, the Coupler (which feeds analysis increments back to the GCM), and an I/O package for processing the large amounts of data the system produces (which will be described in another presentation in this session). The discussion will center on the following issues: the computational complexity for the whole GEOS DAS, assessment of the performance of the individual elements of GEOS DAS, and parallelization strategy for some of the components of the system.

  16. Climatic Oscillations 10,000-155,000 yr B.P. at Owens Lake, California Reflected in Glacial Rock Flour Abundance and Lake Salinity in Core OL-92

    Science.gov (United States)

    Bischoff, J.L.; Menking, K.M.; Fitts, J.P.; Fitzpatrick, J.A.

    1997-01-01

    Chemical analyses of the acid-soluble and clay-size fractions of sediment samples (1500-yr resolution) reveal oscillations of lake salinity and of glacial advances in core OL-92 back to 155,000 yr B.P. Relatively saline conditions are indicated by the abundance of carbonate and smectite (both pedogenic and authigenic), reflected by Ca, Sr, and Mg in the acid-soluble suite, and by Cs2O, excess MgO, and LOI (loss on ignition) in the clay-size fraction. Rock flour produced during glacial advances is represented by the abundance of detrital plagioclase and biotite in the clay-size fraction, the ratio of which remains essentially constant over the entire time span. These phases are quantitatively represented by Na2O, TiO2, Ba, and Mn in the clay fraction. The rock-flour record indicates two major ice-advances during the penultimate glacial cycle corresponding to marine isotope stage (MIS) 6, no major advances during the last interglaciation (entire MIS 5), and three major advances during the last glacial cycle (MIS 2, 3, and 4). The ages of the latter three correspond rather well to 36Cl dates reported for Sierra Nevada moraines. The onset of the last interglaciation is shown by abrupt increases in authigenic CaCO3 and an abrupt decrease in rock flour, at about 118,000 yr B.P. according to our time scale. In contrast, the boundary appears to be gradual in the ??18O record in which the change from light to heavy values begins at about 140,000 yrs B.P. The exact position of the termination, therefore, may be proxy-dependent. Conditions of high carbonate and low rock flour prevailed during the entire period from 118,000 yr B.P. until the glacial advance at 53,000 yr B.P. signaled the end of this long interglaciation. ?? 1997 University of Washington.

  17. Climatic Oscillations 10,000-155,000 yr B.P. at Owens Lake, California Reflected in Glacial Rock Flour Abundance and Lake Salinity in Core OL-92

    Science.gov (United States)

    Bischoff, James L.; Menking, Kirsten M.; Fitts, Jeffrey P.; Fitzpatrick, John A.

    1997-11-01

    Chemical analyses of the acid-soluble and clay-size fractions of sediment samples (1500-yr resolution) reveal oscillations of lake salinity and of glacial advances in core OL-92 back to 155,000 yr B.P. Relatively saline conditions are indicated by the abundance of carbonate and smectite (both pedogenic and authigenic), reflected by Ca, Sr, and Mg in the acid-soluble suite, and by Cs 2O, excess MgO, and LOI (loss on ignition) in the clay-size fraction. Rock flour produced during glacial advances is represented by the abundance of detrital plagioclase and biotite in the clay-size fraction, the ratio of which remains essentially constant over the entire time span. These phases are quantitatively represented by Na 2O, TiO 2, Ba, and Mn in the clay fraction. The rock-flour record indicates two major ice-advances during the penultimate glacial cycle corresponding to marine isotope stage (MIS) 6, no major advances during the last interglaciation (entire MIS 5), and three major advances during the last glacial cycle (MIS 2, 3, and 4). The ages of the latter three correspond rather well to 36Cl dates reported for Sierra Nevada moraines. The onset of the last interglaciation is shown by abrupt increases in authigenic CaCO 3and an abrupt decrease in rock flour, at about 118,000 yr B.P. according to our time scale. In contrast, the boundary appears to be gradual in the δ 18O record in which the change from light to heavy values begins at about 140,000 yrs B.P. The exact position of the termination, therefore, may be proxy-dependent. Conditions of high carbonate and low rock flour prevailed during the entire period from 118,000 yr B.P. until the glacial advance at 53,000 yr B.P. signaled the end of this long interglaciation.

  18. Effect of core body temperature, time of day, and climate conditions on behavioral patterns of lactating dairy cows experiencing mild to moderate heat stress.

    Science.gov (United States)

    Allen, J D; Hall, L W; Collier, R J; Smith, J F

    2015-01-01

    Cattle show several responses to heat load, including spending more time standing. Little is known about what benefit this may provide for the animals. Data from 3 separate cooling management trials were analyzed to investigate the relationship between behavioral patterns in lactating dairy cows experiencing mild to moderate heat stress and their body temperature. Cows (n=157) were each fitted with a leg data logger that measured position and an intravaginal data logger that measures core body temperature (CBT). Ambient conditions were also collected. All data were standardized to 5-min intervals, and information was divided into several categories: when standing and lying bouts were initiated and the continuance of each bout (7,963 lying and 6,276 standing bouts). In one location, cows were continuously subjected to heat-stress levels according to temperature-humidity index (THI) range (THI≥72). The THI range for the other 2 locations was below and above a heat-stress threshold of 72 THI. Overall and regardless of period of day, cows stood up at greater CBT compared with continuing to stand or switching to a lying position. In contrast, cows lay down at lower CBT compared with continuing to lie or switching to a standing position, and lying bouts lasted longer when cows had lower CBT. Standing bouts also lasted longer when cattle had greater CBT, and they were less likely to lie down (less than 50% of lying bouts initiated) when their body temperature was over 38.8°C. Also, cow standing behavior was affected once THI reached 68. Increasing CBT decreased lying duration and increased standing duration. A CBT of 38.93°C marked a 50% likelihood a cow would be standing. This is the first physiological evidence that standing may help cool cows and provides insight into a communally observed behavioral response to heat. Copyright © 2015 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.

  19. The parallel algorithm for the 2D discrete wavelet transform

    Science.gov (United States)

    Barina, David; Najman, Pavel; Kleparnik, Petr; Kula, Michal; Zemcik, Pavel

    2018-04-01

    The discrete wavelet transform can be found at the heart of many image-processing algorithms. Until now, the transform on general-purpose processors (CPUs) was mostly computed using a separable lifting scheme. As the lifting scheme consists of a small number of operations, it is preferred for processing using single-core CPUs. However, considering a parallel processing using multi-core processors, this scheme is inappropriate due to a large number of steps. On such architectures, the number of steps corresponds to the number of points that represent the exchange of data. Consequently, these points often form a performance bottleneck. Our approach appropriately rearranges calculations inside the transform, and thereby reduces the number of steps. In other words, we propose a new scheme that is friendly to parallel environments. When evaluating on multi-core CPUs, we consistently overcome the original lifting scheme. The evaluation was performed on 61-core Intel Xeon Phi and 8-core Intel Xeon processors.

  20. I/O Parallelization for the Goddard Earth Observing System Data Assimilation System (GEOS DAS)

    Science.gov (United States)

    Lucchesi, Rob; Sawyer, W.; Takacs, L. L.; Lyster, P.; Zero, J.

    1998-01-01

    The National Aeronautics and Space Administration (NASA) Data Assimilation Office (DAO) at the Goddard Space Flight Center (GSFC) has developed the GEOS DAS, a data assimilation system that provides production support for NASA missions and will support NASA's Earth Observing System (EOS) in the coming years. The GEOS DAS will be used to provide background fields of meteorological quantities to EOS satellite instrument teams for use in their data algorithms as well as providing assimilated data sets for climate studies on decadal time scales. The DAO has been involved in prototyping parallel implementations of the GEOS DAS for a number of years and is now embarking on an effort to convert the production version from shared-memory parallelism to distributed-memory parallelism using the portable Message-Passing Interface (MPI). The GEOS DAS consists of two main components, an atmospheric General Circulation Model (GCM) and a Physical-space Statistical Analysis System (PSAS). The GCM operates on data that are stored on a regular grid while PSAS works with observational data that are scattered irregularly throughout the atmosphere. As a result, the two components have different data decompositions. The GCM is decomposed horizontally as a checkerboard with all vertical levels of each box existing on the same processing element(PE). The dynamical core of the GCM can also operate on a rotated grid, which requires communication-intensive grid transformations during GCM integration. PSAS groups observations on PEs in a more irregular and dynamic fashion.

  1. Massively parallel multicanonical simulations

    Science.gov (United States)

    Gross, Jonathan; Zierenberg, Johannes; Weigel, Martin; Janke, Wolfhard

    2018-03-01

    Generalized-ensemble Monte Carlo simulations such as the multicanonical method and similar techniques are among the most efficient approaches for simulations of systems undergoing discontinuous phase transitions or with rugged free-energy landscapes. As Markov chain methods, they are inherently serial computationally. It was demonstrated recently, however, that a combination of independent simulations that communicate weight updates at variable intervals allows for the efficient utilization of parallel computational resources for multicanonical simulations. Implementing this approach for the many-thread architecture provided by current generations of graphics processing units (GPUs), we show how it can be efficiently employed with of the order of 104 parallel walkers and beyond, thus constituting a versatile tool for Monte Carlo simulations in the era of massively parallel computing. We provide the fully documented source code for the approach applied to the paradigmatic example of the two-dimensional Ising model as starting point and reference for practitioners in the field.

  2. On the application and interpretation of Keeling plots in paleo climate research – deciphering δ13C of atmospheric CO2 measured in ice cores

    Directory of Open Access Journals (Sweden)

    P. Köhler

    2006-01-01

    Full Text Available The Keeling plot analysis is an interpretation method widely used in terrestrial carbon cycle research to quantify exchange processes of carbon between terrestrial reservoirs and the atmosphere. Here, we analyse measured data sets and artificial time series of the partial pressure of atmospheric carbon dioxide (pCO2 and of δ13C of CO2 over industrial and glacial/interglacial time scales and investigate to what extent the Keeling plot methodology can be applied to longer time scales. The artificial time series are simulation results of the global carbon cycle box model BICYCLE. The signals recorded in ice cores caused by abrupt terrestrial carbon uptake or release loose information due to air mixing in the firn before bubble enclosure and limited sampling frequency. Carbon uptake by the ocean cannot longer be neglected for less abrupt changes as occurring during glacial cycles. We introduce an equation for the calculation of long-term changes in the isotopic signature of atmospheric CO2 caused by an injection of terrestrial carbon to the atmosphere, in which the ocean is introduced as third reservoir. This is a paleo extension of the two reservoir mass balance equations of the Keeling plot approach. It gives an explanation for the bias between the isotopic signature of the terrestrial release and the signature deduced with the Keeling plot approach for long-term processes, in which the oceanic reservoir cannot be neglected. These deduced isotopic signatures are similar (−8.6‰ for steady state analyses of long-term changes in the terrestrial and marine biosphere which both perturb the atmospheric carbon reservoir. They are more positive than the δ13C signals of the sources, e.g. the terrestrial carbon pools themselves (−25‰. A distinction of specific processes acting on the global carbon cycle from the Keeling plot approach is not straightforward. In general, processes related to biogenic fixation or release of carbon have lower y

  3. Climate certificates

    International Nuclear Information System (INIS)

    1998-10-01

    Reduced emissions of climate gases at the lowest cost require international cooperation in order to ensure that the most cost-efficient measures are taken. A market for emission rights is one way of achieving this. However, creating the right conditions for such a market to operate requires an unambiguous definition of the product to be traded. In this PM, the Swedish Power Association sketches out how such a product could be defined, and how a market for the resulting unambiguously defined product could be operated internationally, in parallel with other markets for energy products. Trade in climate certificates could become a joint EU approach to achieving common results within the field of climate policy. The main features of the proposal are as follows: Electricity producers would be allowed to issue climate certificates for electricity produced without climate-affecting emissions, e.g. in wind power plants. 1 kWh of electricity produced without emissions would entitle the utility to issue a climate certificate for 1 kWh. Electricity from power stations having low emissions, e.g. modern natural gas-fired plants, would entitle the utility to issue certificates in proportion to how much lower their emissions were in comparison with those from conventional coal-fired power stations. The number of certificates would be reduced by an individual coefficient, related directly to the quantity of climate-affecting emissions from the plant concerned. They would be traded and noted on markets in the various member countries. The certificates would not be nationally restricted, but could be traded across borders. Exchanges would be authorised by national authorities, in accordance with overall EU directives. These authorised exchanges would act as certification bodies, checking that certificates had been properly issued in accordance with a corresponding volume of electricity production. Electricity and certificates could be purchased from different suppliers. The

  4. SPINning parallel systems software

    International Nuclear Information System (INIS)

    Matlin, O.S.; Lusk, E.; McCune, W.

    2002-01-01

    We describe our experiences in using Spin to verify parts of the Multi Purpose Daemon (MPD) parallel process management system. MPD is a distributed collection of processes connected by Unix network sockets. MPD is dynamic processes and connections among them are created and destroyed as MPD is initialized, runs user processes, recovers from faults, and terminates. This dynamic nature is easily expressible in the Spin/Promela framework but poses performance and scalability challenges. We present here the results of expressing some of the parallel algorithms of MPD and executing both simulation and verification runs with Spin

  5. Parallel programming with Python

    CERN Document Server

    Palach, Jan

    2014-01-01

    A fast, easy-to-follow and clear tutorial to help you develop Parallel computing systems using Python. Along with explaining the fundamentals, the book will also introduce you to slightly advanced concepts and will help you in implementing these techniques in the real world. If you are an experienced Python programmer and are willing to utilize the available computing resources by parallelizing applications in a simple way, then this book is for you. You are required to have a basic knowledge of Python development to get the most of this book.

  6. Professional Parallel Programming with C# Master Parallel Extensions with NET 4

    CERN Document Server

    Hillar, Gastón

    2010-01-01

    Expert guidance for those programming today's dual-core processors PCs As PC processors explode from one or two to now eight processors, there is an urgent need for programmers to master concurrent programming. This book dives deep into the latest technologies available to programmers for creating professional parallel applications using C#, .NET 4, and Visual Studio 2010. The book covers task-based programming, coordination data structures, PLINQ, thread pools, asynchronous programming model, and more. It also teaches other parallel programming techniques, such as SIMD and vectorization.Teach

  7. Effects of adjustable and stationary fans with misters on core body temperature and lying behavior of lactating dairy cows in a semiarid climate.

    Science.gov (United States)

    Anderson, S D; Bradford, B J; Harner, J P; Tucker, C B; Choi, C Y; Allen, J D; Hall, L W; Rungruang, S; Collier, R J; Smith, J F

    2013-07-01

    Cows readily seek shade to reduce solar heat load during periods of high ambient temperature. Typically, auxiliary cooling systems are oriented to maximize cooling for shaded cows. However, when a shade structure is oriented north-south, stationary fan and mister cooling systems are unable to track shade as the sun's angle shifts throughout the day, and thus can become ineffective. The FlipFan Dairy Cooling System (Schaefer Ventilation Equipment, Sauk Rapids, MN) employs fans and misters that follow shade and compensate for wind speed by rotating on a horizontal axis. Multiparous, lactating Holstein cows (n=144) on a commercial dairy in Arizona were cooled by a fixed system comprised of stationary fans and misters acting as control or the adjustable FlipFan operated for 16.5 h/d (0830 to 0100 h). Core body temperatures (CBT) of 64 cows (4 pens/treatment; 8 cows/pen; 6d) and lying behavior of 144 cows (4 pens/treatment; 18 cows/pen; 5d) were collected by intravaginal and leg data loggers, respectively. Cows were balanced by milk production, blocked by days in milk, and randomly assigned to pen within block. Pen was the experimental unit. In a second experiment, isothermal maps were developed using a fixed system of thermal data loggers arranged in the shaded areas of the pens at different times of day and were analyzed for differences in the temperature-humidity index (THI) achieved by each cooling treatment. Ambient conditions consisted of a mean temperature of 33.0°C, mean relative humidity of 40.3%, and mean THI of 80.2. Mean 24-h CBT for FlipFan was lower than control (38.9 vs. 39.1±0.04°C). A treatment × time interaction was observed in which CBT of FlipFan was 0.4°C lower than control from 0600 to 0800h and 1500 to 1600h. Cows cooled by FlipFan spent more time lying down compared with those cooled by control (9.5 vs. 8.6 h/d). Cows under FlipFan had more frequent lying bouts than did those under control (12.8 vs. 10.7 bouts/d). Lower CBT and decreased

  8. Expressing Parallelism with ROOT

    Energy Technology Data Exchange (ETDEWEB)

    Piparo, D. [CERN; Tejedor, E. [CERN; Guiraud, E. [CERN; Ganis, G. [CERN; Mato, P. [CERN; Moneta, L. [CERN; Valls Pla, X. [CERN; Canal, P. [Fermilab

    2017-11-22

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  9. Expressing Parallelism with ROOT

    Science.gov (United States)

    Piparo, D.; Tejedor, E.; Guiraud, E.; Ganis, G.; Mato, P.; Moneta, L.; Valls Pla, X.; Canal, P.

    2017-10-01

    The need for processing the ever-increasing amount of data generated by the LHC experiments in a more efficient way has motivated ROOT to further develop its support for parallelism. Such support is being tackled both for shared-memory and distributed-memory environments. The incarnations of the aforementioned parallelism are multi-threading, multi-processing and cluster-wide executions. In the area of multi-threading, we discuss the new implicit parallelism and related interfaces, as well as the new building blocks to safely operate with ROOT objects in a multi-threaded environment. Regarding multi-processing, we review the new MultiProc framework, comparing it with similar tools (e.g. multiprocessing module in Python). Finally, as an alternative to PROOF for cluster-wide executions, we introduce the efforts on integrating ROOT with state-of-the-art distributed data processing technologies like Spark, both in terms of programming model and runtime design (with EOS as one of the main components). For all the levels of parallelism, we discuss, based on real-life examples and measurements, how our proposals can increase the productivity of scientists.

  10. Parallel Fast Legendre Transform

    NARCIS (Netherlands)

    Alves de Inda, M.; Bisseling, R.H.; Maslen, D.K.

    1998-01-01

    We discuss a parallel implementation of a fast algorithm for the discrete polynomial Legendre transform We give an introduction to the DriscollHealy algorithm using polynomial arithmetic and present experimental results on the eciency and accuracy of our implementation The algorithms were

  11. Practical parallel programming

    CERN Document Server

    Bauer, Barr E

    2014-01-01

    This is the book that will teach programmers to write faster, more efficient code for parallel processors. The reader is introduced to a vast array of procedures and paradigms on which actual coding may be based. Examples and real-life simulations using these devices are presented in C and FORTRAN.

  12. Parallel hierarchical radiosity rendering

    Energy Technology Data Exchange (ETDEWEB)

    Carter, Michael [Iowa State Univ., Ames, IA (United States)

    1993-07-01

    In this dissertation, the step-by-step development of a scalable parallel hierarchical radiosity renderer is documented. First, a new look is taken at the traditional radiosity equation, and a new form is presented in which the matrix of linear system coefficients is transformed into a symmetric matrix, thereby simplifying the problem and enabling a new solution technique to be applied. Next, the state-of-the-art hierarchical radiosity methods are examined for their suitability to parallel implementation, and scalability. Significant enhancements are also discovered which both improve their theoretical foundations and improve the images they generate. The resultant hierarchical radiosity algorithm is then examined for sources of parallelism, and for an architectural mapping. Several architectural mappings are discussed. A few key algorithmic changes are suggested during the process of making the algorithm parallel. Next, the performance, efficiency, and scalability of the algorithm are analyzed. The dissertation closes with a discussion of several ideas which have the potential to further enhance the hierarchical radiosity method, or provide an entirely new forum for the application of hierarchical methods.

  13. Parallel universes beguile science

    CERN Multimedia

    2007-01-01

    A staple of mind-bending science fiction, the possibility of multiple universes has long intrigued hard-nosed physicists, mathematicians and cosmologists too. We may not be able -- as least not yet -- to prove they exist, many serious scientists say, but there are plenty of reasons to think that parallel dimensions are more than figments of eggheaded imagination.

  14. Parallel k-means++

    Energy Technology Data Exchange (ETDEWEB)

    2017-04-04

    A parallelization of the k-means++ seed selection algorithm on three distinct hardware platforms: GPU, multicore CPU, and multithreaded architecture. K-means++ was developed by David Arthur and Sergei Vassilvitskii in 2007 as an extension of the k-means data clustering technique. These algorithms allow people to cluster multidimensional data, by attempting to minimize the mean distance of data points within a cluster. K-means++ improved upon traditional k-means by using a more intelligent approach to selecting the initial seeds for the clustering process. While k-means++ has become a popular alternative to traditional k-means clustering, little work has been done to parallelize this technique. We have developed original C++ code for parallelizing the algorithm on three unique hardware architectures: GPU using NVidia's CUDA/Thrust framework, multicore CPU using OpenMP, and the Cray XMT multithreaded architecture. By parallelizing the process for these platforms, we are able to perform k-means++ clustering much more quickly than it could be done before.

  15. Parallel plate detectors

    International Nuclear Information System (INIS)

    Gardes, D.; Volkov, P.

    1981-01-01

    A 5x3cm 2 (timing only) and a 15x5cm 2 (timing and position) parallel plate avalanche counters (PPAC) are considered. The theory of operation and timing resolution is given. The measurement set-up and the curves of experimental results illustrate the possibilities of the two counters [fr

  16. Parallel hierarchical global illumination

    Energy Technology Data Exchange (ETDEWEB)

    Snell, Quinn O. [Iowa State Univ., Ames, IA (United States)

    1997-10-08

    Solving the global illumination problem is equivalent to determining the intensity of every wavelength of light in all directions at every point in a given scene. The complexity of the problem has led researchers to use approximation methods for solving the problem on serial computers. Rather than using an approximation method, such as backward ray tracing or radiosity, the authors have chosen to solve the Rendering Equation by direct simulation of light transport from the light sources. This paper presents an algorithm that solves the Rendering Equation to any desired accuracy, and can be run in parallel on distributed memory or shared memory computer systems with excellent scaling properties. It appears superior in both speed and physical correctness to recent published methods involving bidirectional ray tracing or hybrid treatments of diffuse and specular surfaces. Like progressive radiosity methods, it dynamically refines the geometry decomposition where required, but does so without the excessive storage requirements for ray histories. The algorithm, called Photon, produces a scene which converges to the global illumination solution. This amounts to a huge task for a 1997-vintage serial computer, but using the power of a parallel supercomputer significantly reduces the time required to generate a solution. Currently, Photon can be run on most parallel environments from a shared memory multiprocessor to a parallel supercomputer, as well as on clusters of heterogeneous workstations.

  17. A task parallel implementation of fast multipole methods

    KAUST Repository

    Taura, Kenjiro

    2012-11-01

    This paper describes a task parallel implementation of ExaFMM, an open source implementation of fast multipole methods (FMM), using a lightweight task parallel library MassiveThreads. Although there have been many attempts on parallelizing FMM, experiences have almost exclusively been limited to formulation based on flat homogeneous parallel loops. FMM in fact contains operations that cannot be readily expressed in such conventional but restrictive models. We show that task parallelism, or parallel recursions in particular, allows us to parallelize all operations of FMM naturally and scalably. Moreover it allows us to parallelize a \\'\\'mutual interaction\\'\\' for force/potential evaluation, which is roughly twice as efficient as a more conventional, unidirectional force/potential evaluation. The net result is an open source FMM that is clearly among the fastest single node implementations, including those on GPUs; with a million particles on a 32 cores Sandy Bridge 2.20GHz node, it completes a single time step including tree construction and force/potential evaluation in 65 milliseconds. The study clearly showcases both programmability and performance benefits of flexible parallel constructs over more monolithic parallel loops. © 2012 IEEE.

  18. Reactor core

    International Nuclear Information System (INIS)

    Matsuura, Tetsuaki; Nomura, Teiji; Tokunaga, Kensuke; Okuda, Shin-ichi

    1990-01-01

    Fuel assemblies in the portions where the gradient of fast neutron fluxes between two opposing faces of a channel box is great are kept loaded at the outermost peripheral position of the reactor core also in the second operation cycle in the order to prevent interference between a control rod and the channel box due to bending deformation of the channel box. Further, the fuel assemblies in the second row from the outer most periphery in the first operation cycle are also kept loaded at the second row in the second operation cycle. Since the gradient of the fast neutrons in the reactor core is especially great at the outer circumference of the reactor core, the channel box at the outer circumference is bent such that the surface facing to the center of the reactor core is convexed and the channel box in the second row is also bent to the identical direction, the insertion of the control rod is not interfered. Further, if the positions for the fuels at the outermost periphery and the fuels in the second row are not altered in the second operation cycle, the gaps are not reduced to prevent the interference between the control rod and the channel box. (N.H.)

  19. A Parallel Saturation Algorithm on Shared Memory Architectures

    Science.gov (United States)

    Ezekiel, Jonathan; Siminiceanu

    2007-01-01

    Symbolic state-space generators are notoriously hard to parallelize. However, the Saturation algorithm implemented in the SMART verification tool differs from other sequential symbolic state-space generators in that it exploits the locality of ring events in asynchronous system models. This paper explores whether event locality can be utilized to efficiently parallelize Saturation on shared-memory architectures. Conceptually, we propose to parallelize the ring of events within a decision diagram node, which is technically realized via a thread pool. We discuss the challenges involved in our parallel design and conduct experimental studies on its prototypical implementation. On a dual-processor dual core PC, our studies show speed-ups for several example models, e.g., of up to 50% for a Kanban model, when compared to running our algorithm only on a single core.

  20. Core Values

    Science.gov (United States)

    Martin, Tim

    2016-01-01

    In this article, two lessons are introduced in which students examine Arctic lake sediments from Lake El'gygytgyn in Russia and discover a climate signal in a lake or pond near their own school. The lessons allow students to experience fieldwork, understand lab procedure, practice basic measurement and observation skills, and learn how to…

  1. Core baffle for nuclear reactors

    International Nuclear Information System (INIS)

    Machado, O.J.; Berringer, R.T.

    1977-01-01

    The invention concerns the design of the core of a LWR with a large number of fuel assemblies formed by fuel rods and kept in position by spacer grids. According to the invention, at the level of the spacer grids match plates are mounted with openings so the flow of coolant directed upwards will not be obstructed and a parallel bypass will be obtained in the space between the core barrel and the baffle plates. In case of an accident, this configuration reduces or avoids damage from overpressure reactions. (HP) [de

  2. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.A.N.

    2015-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Ice core records provide an annual-scale, 'instrumental-quality' baseline of atmospheric temperature and circulation changes back many thousands of years. (author).

  3. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.A.N.

    2014-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Ice core records provide an annual-scale, 'instrumental-quality' baseline of atmospheric temperature and circulation changes back many thousands of years. (author)

  4. A Parallel Algebraic Multigrid Solver on Graphics Processing Units

    KAUST Repository

    Haase, Gundolf

    2010-01-01

    The paper presents a multi-GPU implementation of the preconditioned conjugate gradient algorithm with an algebraic multigrid preconditioner (PCG-AMG) for an elliptic model problem on a 3D unstructured grid. An efficient parallel sparse matrix-vector multiplication scheme underlying the PCG-AMG algorithm is presented for the many-core GPU architecture. A performance comparison of the parallel solver shows that a singe Nvidia Tesla C1060 GPU board delivers the performance of a sixteen node Infiniband cluster and a multi-GPU configuration with eight GPUs is about 100 times faster than a typical server CPU core. © 2010 Springer-Verlag.

  5. Parallelization of the model-based iterative reconstruction algorithm DIRA

    International Nuclear Information System (INIS)

    Oertenberg, A.; Sandborg, M.; Alm Carlsson, G.; Malusek, A.; Magnusson, M.

    2016-01-01

    New paradigms for parallel programming have been devised to simplify software development on multi-core processors and many-core graphical processing units (GPU). Despite their obvious benefits, the parallelization of existing computer programs is not an easy task. In this work, the use of the Open Multiprocessing (OpenMP) and Open Computing Language (OpenCL) frameworks is considered for the parallelization of the model-based iterative reconstruction algorithm DIRA with the aim to significantly shorten the code's execution time. Selected routines were parallelized using OpenMP and OpenCL libraries; some routines were converted from MATLAB to C and optimised. Parallelization of the code with the OpenMP was easy and resulted in an overall speedup of 15 on a 16-core computer. Parallelization with OpenCL was more difficult owing to differences between the central processing unit and GPU architectures. The resulting speedup was substantially lower than the theoretical peak performance of the GPU; the cause was explained. (authors)

  6. Parallel grid population

    Science.gov (United States)

    Wald, Ingo; Ize, Santiago

    2015-07-28

    Parallel population of a grid with a plurality of objects using a plurality of processors. One example embodiment is a method for parallel population of a grid with a plurality of objects using a plurality of processors. The method includes a first act of dividing a grid into n distinct grid portions, where n is the number of processors available for populating the grid. The method also includes acts of dividing a plurality of objects into n distinct sets of objects, assigning a distinct set of objects to each processor such that each processor determines by which distinct grid portion(s) each object in its distinct set of objects is at least partially bounded, and assigning a distinct grid portion to each processor such that each processor populates its distinct grid portion with any objects that were previously determined to be at least partially bounded by its distinct grid portion.

  7. Ultrascalable petaflop parallel supercomputer

    Science.gov (United States)

    Blumrich, Matthias A [Ridgefield, CT; Chen, Dong [Croton On Hudson, NY; Chiu, George [Cross River, NY; Cipolla, Thomas M [Katonah, NY; Coteus, Paul W [Yorktown Heights, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Hall, Shawn [Pleasantville, NY; Haring, Rudolf A [Cortlandt Manor, NY; Heidelberger, Philip [Cortlandt Manor, NY; Kopcsay, Gerard V [Yorktown Heights, NY; Ohmacht, Martin [Yorktown Heights, NY; Salapura, Valentina [Chappaqua, NY; Sugavanam, Krishnan [Mahopac, NY; Takken, Todd [Brewster, NY

    2010-07-20

    A massively parallel supercomputer of petaOPS-scale includes node architectures based upon System-On-a-Chip technology, where each processing node comprises a single Application Specific Integrated Circuit (ASIC) having up to four processing elements. The ASIC nodes are interconnected by multiple independent networks that optimally maximize the throughput of packet communications between nodes with minimal latency. The multiple networks may include three high-speed networks for parallel algorithm message passing including a Torus, collective network, and a Global Asynchronous network that provides global barrier and notification functions. These multiple independent networks may be collaboratively or independently utilized according to the needs or phases of an algorithm for optimizing algorithm processing performance. The use of a DMA engine is provided to facilitate message passing among the nodes without the expenditure of processing resources at the node.

  8. More parallel please

    DEFF Research Database (Denmark)

    Gregersen, Frans; Josephson, Olle; Kristoffersen, Gjert

    of departure that English may be used in parallel with the various local, in this case Nordic, languages. As such, the book integrates the challenge of internationalization faced by any university with the wish to improve quality in research, education and administration based on the local language......Abstract [en] More parallel, please is the result of the work of an Inter-Nordic group of experts on language policy financed by the Nordic Council of Ministers 2014-17. The book presents all that is needed to plan, practice and revise a university language policy which takes as its point......(s). There are three layers in the text: First, you may read the extremely brief version of the in total 11 recommendations for best practice. Second, you may acquaint yourself with the extended version of the recommendations and finally, you may study the reasoning behind each of them. At the end of the text, we give...

  9. PARALLEL MOVING MECHANICAL SYSTEMS

    Directory of Open Access Journals (Sweden)

    Florian Ion Tiberius Petrescu

    2014-09-01

    Full Text Available Normal 0 false false false EN-US X-NONE X-NONE MicrosoftInternetExplorer4 Moving mechanical systems parallel structures are solid, fast, and accurate. Between parallel systems it is to be noticed Stewart platforms, as the oldest systems, fast, solid and precise. The work outlines a few main elements of Stewart platforms. Begin with the geometry platform, kinematic elements of it, and presented then and a few items of dynamics. Dynamic primary element on it means the determination mechanism kinetic energy of the entire Stewart platforms. It is then in a record tail cinematic mobile by a method dot matrix of rotation. If a structural mottoelement consists of two moving elements which translates relative, drive train and especially dynamic it is more convenient to represent the mottoelement as a single moving components. We have thus seven moving parts (the six motoelements or feet to which is added mobile platform 7 and one fixed.

  10. Xyce parallel electronic simulator.

    Energy Technology Data Exchange (ETDEWEB)

    Keiter, Eric R; Mei, Ting; Russo, Thomas V.; Rankin, Eric Lamont; Schiek, Richard Louis; Thornquist, Heidi K.; Fixel, Deborah A.; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R.

    2010-05-01

    This document is a reference guide to the Xyce Parallel Electronic Simulator, and is a companion document to the Xyce Users Guide. The focus of this document is (to the extent possible) exhaustively list device parameters, solver options, parser options, and other usage details of Xyce. This document is not intended to be a tutorial. Users who are new to circuit simulation are better served by the Xyce Users Guide.

  11. Stability of parallel flows

    CERN Document Server

    Betchov, R

    2012-01-01

    Stability of Parallel Flows provides information pertinent to hydrodynamical stability. This book explores the stability problems that occur in various fields, including electronics, mechanics, oceanography, administration, economics, as well as naval and aeronautical engineering. Organized into two parts encompassing 10 chapters, this book starts with an overview of the general equations of a two-dimensional incompressible flow. This text then explores the stability of a laminar boundary layer and presents the equation of the inviscid approximation. Other chapters present the general equation

  12. Algorithmically specialized parallel computers

    CERN Document Server

    Snyder, Lawrence; Gannon, Dennis B

    1985-01-01

    Algorithmically Specialized Parallel Computers focuses on the concept and characteristics of an algorithmically specialized computer.This book discusses the algorithmically specialized computers, algorithmic specialization using VLSI, and innovative architectures. The architectures and algorithms for digital signal, speech, and image processing and specialized architectures for numerical computations are also elaborated. Other topics include the model for analyzing generalized inter-processor, pipelined architecture for search tree maintenance, and specialized computer organization for raster

  13. Scariest thing about climate change: climate flips

    International Nuclear Information System (INIS)

    Beaulieu, P.

    1997-01-01

    The idea that an increase in greenhouse gases will cause the global average temperature to rise slowly over the next decades was discussed. Studies of ice core from Greenland have shown that in the past climate shifts seem to have happened quickly. Some scientists fear that increasingly frequent extreme weather events could be a sign that the climate system is nearing its threshold and a rapid climate flip may be just ahead. In the case of global climatic system, the danger is that stresses from greenhouse gas effects are pushing the present system over the threshold where it must flip into a new warmer system that will be stable, but different from the climate on which our agriculture, economy, settlements and lives depend. 4 refs

  14. Parallel Execution of Multi Set Constraint Rewrite Rules

    DEFF Research Database (Denmark)

    Sulzmann, Martin; Lam, Edmund Soon Lee

    2008-01-01

    that the underlying constraint rewrite implementation executes rewrite steps in parallel on increasingly popular becoming multi-core architectures. We design and implement efficient algorithms which allow for the parallel execution of multi-set constraint rewrite rules. Our experiments show that we obtain some......Multi-set constraint rewriting allows for a highly parallel computational model and has been used in a multitude of application domains such as constraint solving, agent specification etc. Rewriting steps can be applied simultaneously as long as they do not interfere with each other.We wish...

  15. Abrupt climate change:Debate or action

    Institute of Scientific and Technical Information of China (English)

    CHENG Hai

    2004-01-01

    Global abrupt climate changes have been documented by various climate records, including ice cores,ocean sediment cores, lake sediment cores, cave deposits,loess deposits and pollen records. The climate system prefers to be in one of two stable states, i.e. interstadial or stadial conditions, but not in between. The transition between two states has an abrupt character. Abrupt climate changes are,in general, synchronous in the northern hemisphere and tropical regions. The timescale for abrupt climate changes can be as short as a decade. As the impacts may be potentially serious, we need to take actions such as reducing CO2emissions to the atmosphere.

  16. Design Patterns: establishing a discipline of parallel software engineering

    CERN Multimedia

    CERN. Geneva

    2010-01-01

    Many core processors present us with a software challenge. We must turn our serial code into parallel code. To accomplish this wholesale transformation of our software ecosystem, we must define established practice is in parallel programming and then develop tools to support that practice. This leads to design patterns supported by frameworks optimized at runtime with advanced autotuning compilers. In this talk I provide an update of my ongoing research with the ParLab at UC Berkeley to realize this vision. In particular, I will describe our draft parallel pattern language, our early experiments with software frameworks, and the associated runtime optimization tools.About the speakerTim Mattson is a parallel programmer (Ph.D. Chemistry, UCSC, 1985). He does linear algebra, finds oil, shakes molecules, solves differential equations, and models electrons in simple atomic systems. He has spent his career working with computer scientists to make sure the needs of parallel applications programmers are met.Tim has ...

  17. Multiscale Architectures and Parallel Algorithms for Video Object Tracking

    Science.gov (United States)

    2011-10-01

    larger number of cores using the IBM QS22 Blade for handling higher video processing workloads (but at higher cost per core), low power consumption and...Cell/B.E. Blade processors which have a lot more main memory but also higher power consumption . More detailed performance figures for HD and SD video...Parallelism in Algorithms and Architectures, pages 289–298, 2007. [3] S. Ali and M. Shah. COCOA - Tracking in aerial imagery. In Daniel J. Henry

  18. An Automatic Instruction-Level Parallelization of Machine Code

    Directory of Open Access Journals (Sweden)

    MARINKOVIC, V.

    2018-02-01

    Full Text Available Prevailing multicores and novel manycores have made a great challenge of modern day - parallelization of embedded software that is still written as sequential. In this paper, automatic code parallelization is considered, focusing on developing a parallelization tool at the binary level as well as on the validation of this approach. The novel instruction-level parallelization algorithm for assembly code which uses the register names after SSA to find independent blocks of code and then to schedule independent blocks using METIS to achieve good load balance is developed. The sequential consistency is verified and the validation is done by measuring the program execution time on the target architecture. Great speedup, taken as the performance measure in the validation process, and optimal load balancing are achieved for multicore RISC processors with 2 to 16 cores (e.g. MIPS, MicroBlaze, etc.. In particular, for 16 cores, the average speedup is 7.92x, while in some cases it reaches 14x. An approach to automatic parallelization provided by this paper is useful to researchers and developers in the area of parallelization as the basis for further optimizations, as the back-end of a compiler, or as the code parallelization tool for an embedded system.

  19. Electromagnetic Physics Models for Parallel Computing Architectures

    International Nuclear Information System (INIS)

    Amadio, G; Bianchini, C; Iope, R; Ananya, A; Apostolakis, J; Aurora, A; Bandieramonte, M; Brun, R; Carminati, F; Gheata, A; Gheata, M; Goulas, I; Nikitina, T; Bhattacharyya, A; Mohanty, A; Canal, P; Elvira, D; Jun, S Y; Lima, G; Duhem, L

    2016-01-01

    The recent emergence of hardware architectures characterized by many-core or accelerated processors has opened new opportunities for concurrent programming models taking advantage of both SIMD and SIMT architectures. GeantV, a next generation detector simulation, has been designed to exploit both the vector capability of mainstream CPUs and multi-threading capabilities of coprocessors including NVidia GPUs and Intel Xeon Phi. The characteristics of these architectures are very different in terms of the vectorization depth and type of parallelization needed to achieve optimal performance. In this paper we describe implementation of electromagnetic physics models developed for parallel computing architectures as a part of the GeantV project. Results of preliminary performance evaluation and physics validation are presented as well. (paper)

  20. Electromagnetic Physics Models for Parallel Computing Architectures

    Science.gov (United States)

    Amadio, G.; Ananya, A.; Apostolakis, J.; Aurora, A.; Bandieramonte, M.; Bhattacharyya, A.; Bianchini, C.; Brun, R.; Canal, P.; Carminati, F.; Duhem, L.; Elvira, D.; Gheata, A.; Gheata, M.; Goulas, I.; Iope, R.; Jun, S. Y.; Lima, G.; Mohanty, A.; Nikitina, T.; Novak, M.; Pokorski, W.; Ribon, A.; Seghal, R.; Shadura, O.; Vallecorsa, S.; Wenzel, S.; Zhang, Y.

    2016-10-01

    The recent emergence of hardware architectures characterized by many-core or accelerated processors has opened new opportunities for concurrent programming models taking advantage of both SIMD and SIMT architectures. GeantV, a next generation detector simulation, has been designed to exploit both the vector capability of mainstream CPUs and multi-threading capabilities of coprocessors including NVidia GPUs and Intel Xeon Phi. The characteristics of these architectures are very different in terms of the vectorization depth and type of parallelization needed to achieve optimal performance. In this paper we describe implementation of electromagnetic physics models developed for parallel computing architectures as a part of the GeantV project. Results of preliminary performance evaluation and physics validation are presented as well.

  1. Flexibility and Performance of Parallel File Systems

    Science.gov (United States)

    Kotz, David; Nieuwejaar, Nils

    1996-01-01

    As we gain experience with parallel file systems, it becomes increasingly clear that a single solution does not suit all applications. For example, it appears to be impossible to find a single appropriate interface, caching policy, file structure, or disk-management strategy. Furthermore, the proliferation of file-system interfaces and abstractions make applications difficult to port. We propose that the traditional functionality of parallel file systems be separated into two components: a fixed core that is standard on all platforms, encapsulating only primitive abstractions and interfaces, and a set of high-level libraries to provide a variety of abstractions and application-programmer interfaces (API's). We present our current and next-generation file systems as examples of this structure. Their features, such as a three-dimensional file structure, strided read and write interfaces, and I/O-node programs, are specifically designed with the flexibility and performance necessary to support a wide range of applications.

  2. Oxytocin: parallel processing in the social brain?

    Science.gov (United States)

    Dölen, Gül

    2015-06-01

    Early studies attempting to disentangle the network complexity of the brain exploited the accessibility of sensory receptive fields to reveal circuits made up of synapses connected both in series and in parallel. More recently, extension of this organisational principle beyond the sensory systems has been made possible by the advent of modern molecular, viral and optogenetic approaches. Here, evidence supporting parallel processing of social behaviours mediated by oxytocin is reviewed. Understanding oxytocinergic signalling from this perspective has significant implications for the design of oxytocin-based therapeutic interventions aimed at disorders such as autism, where disrupted social function is a core clinical feature. Moreover, identification of opportunities for novel technology development will require a better appreciation of the complexity of the circuit-level organisation of the social brain. © 2015 The Authors. Journal of Neuroendocrinology published by John Wiley & Sons Ltd on behalf of British Society for Neuroendocrinology.

  3. Integrated Task And Data Parallel Programming: Language Design

    Science.gov (United States)

    Grimshaw, Andrew S.; West, Emily A.

    1998-01-01

    his research investigates the combination of task and data parallel language constructs within a single programming language. There are an number of applications that exhibit properties which would be well served by such an integrated language. Examples include global climate models, aircraft design problems, and multidisciplinary design optimization problems. Our approach incorporates data parallel language constructs into an existing, object oriented, task parallel language. The language will support creation and manipulation of parallel classes and objects of both types (task parallel and data parallel). Ultimately, the language will allow data parallel and task parallel classes to be used either as building blocks or managers of parallel objects of either type, thus allowing the development of single and multi-paradigm parallel applications. 1995 Research Accomplishments In February I presented a paper at Frontiers '95 describing the design of the data parallel language subset. During the spring I wrote and defended my dissertation proposal. Since that time I have developed a runtime model for the language subset. I have begun implementing the model and hand-coding simple examples which demonstrate the language subset. I have identified an astrophysical fluid flow application which will validate the data parallel language subset. 1996 Research Agenda Milestones for the coming year include implementing a significant portion of the data parallel language subset over the Legion system. Using simple hand-coded methods, I plan to demonstrate (1) concurrent task and data parallel objects and (2) task parallel objects managing both task and data parallel objects. My next steps will focus on constructing a compiler and implementing the fluid flow application with the language. Concurrently, I will conduct a search for a real-world application exhibiting both task and data parallelism within the same program m. Additional 1995 Activities During the fall I collaborated

  4. Rubus: A compiler for seamless and extensible parallelism

    Science.gov (United States)

    Adnan, Muhammad; Aslam, Faisal; Sarwar, Syed Mansoor

    2017-01-01

    Nowadays, a typical processor may have multiple processing cores on a single chip. Furthermore, a special purpose processing unit called Graphic Processing Unit (GPU), originally designed for 2D/3D games, is now available for general purpose use in computers and mobile devices. However, the traditional programming languages which were designed to work with machines having single core CPUs, cannot utilize the parallelism available on multi-core processors efficiently. Therefore, to exploit the extraordinary processing power of multi-core processors, researchers are working on new tools and techniques to facilitate parallel programming. To this end, languages like CUDA and OpenCL have been introduced, which can be used to write code with parallelism. The main shortcoming of these languages is that programmer needs to specify all the complex details manually in order to parallelize the code across multiple cores. Therefore, the code written in these languages is difficult to understand, debug and maintain. Furthermore, to parallelize legacy code can require rewriting a significant portion of code in CUDA or OpenCL, which can consume significant time and resources. Thus, the amount of parallelism achieved is proportional to the skills of the programmer and the time spent in code optimizations. This paper proposes a new open source compiler, Rubus, to achieve seamless parallelism. The Rubus compiler relieves the programmer from manually specifying the low-level details. It analyses and transforms a sequential program into a parallel program automatically, without any user intervention. This achieves massive speedup and better utilization of the underlying hardware without a programmer’s expertise in parallel programming. For five different benchmarks, on average a speedup of 34.54 times has been achieved by Rubus as compared to Java on a basic GPU having only 96 cores. Whereas, for a matrix multiplication benchmark the average execution speedup of 84 times has been

  5. Parallelization characteristics of the DeCART code

    International Nuclear Information System (INIS)

    Cho, J. Y.; Joo, H. G.; Kim, H. Y.; Lee, C. C.; Chang, M. H.; Zee, S. Q.

    2003-12-01

    This report is to describe the parallelization characteristics of the DeCART code and also examine its parallel performance. Parallel computing algorithms are implemented to DeCART to reduce the tremendous computational burden and memory requirement involved in the three-dimensional whole core transport calculation. In the parallelization of the DeCART code, the axial domain decomposition is first realized by using MPI (Message Passing Interface), and then the azimuthal angle domain decomposition by using either MPI or OpenMP. When using the MPI for both the axial and the angle domain decomposition, the concept of MPI grouping is employed for convenient communication in each communication world. For the parallel computation, most of all the computing modules except for the thermal hydraulic module are parallelized. These parallelized computing modules include the MOC ray tracing, CMFD, NEM, region-wise cross section preparation and cell homogenization modules. For the distributed allocation, most of all the MOC and CMFD/NEM variables are allocated only for the assigned planes, which reduces the required memory by a ratio of the number of the assigned planes to the number of all planes. The parallel performance of the DeCART code is evaluated by solving two problems, a rodded variation of the C5G7 MOX three-dimensional benchmark problem and a simplified three-dimensional SMART PWR core problem. In the aspect of parallel performance, the DeCART code shows a good speedup of about 40.1 and 22.4 in the ray tracing module and about 37.3 and 20.2 in the total computing time when using 48 CPUs on the IBM Regatta and 24 CPUs on the LINUX cluster, respectively. In the comparison between the MPI and OpenMP, OpenMP shows a somewhat better performance than MPI. Therefore, it is concluded that the first priority in the parallel computation of the DeCART code is in the axial domain decomposition by using MPI, and then in the angular domain using OpenMP, and finally the angular

  6. Core BPEL

    DEFF Research Database (Denmark)

    Hallwyl, Tim; Højsgaard, Espen

    The Web Services Business Process Execution Language (WS-BPEL) is a language for expressing business process behaviour based on web services. The language is intentionally not minimal but provides a rich set of constructs, allows omission of constructs by relying on defaults, and supports language......, does not allow omissions, and does not contain ignorable elements. We do so by identifying syntactic sugar, including default values, and ignorable elements in WS-BPEL. The analysis results in a translation from the full language to the core subset. Thus, we reduce the effort needed for working...

  7. Resistor Combinations for Parallel Circuits.

    Science.gov (United States)

    McTernan, James P.

    1978-01-01

    To help simplify both teaching and learning of parallel circuits, a high school electricity/electronics teacher presents and illustrates the use of tables of values for parallel resistive circuits in which total resistances are whole numbers. (MF)

  8. SOFTWARE FOR DESIGNING PARALLEL APPLICATIONS

    Directory of Open Access Journals (Sweden)

    M. K. Bouza

    2017-01-01

    Full Text Available The object of research is the tools to support the development of parallel programs in C/C ++. The methods and software which automates the process of designing parallel applications are proposed.

  9. From sequential to parallel programming with patterns

    CERN Document Server

    CERN. Geneva

    2018-01-01

    To increase in both performance and efficiency, our programming models need to adapt to better exploit modern processors. The classic idioms and patterns for programming such as loops, branches or recursion are the pillars of almost every code and are well known among all programmers. These patterns all have in common that they are sequential in nature. Embracing parallel programming patterns, which allow us to program for multi- and many-core hardware in a natural way, greatly simplifies the task of designing a program that scales and performs on modern hardware, independently of the used programming language, and in a generic way.

  10. Accelerating Climate and Weather Simulations through Hybrid Computing

    Science.gov (United States)

    Zhou, Shujia; Cruz, Carlos; Duffy, Daniel; Tucker, Robert; Purcell, Mark

    2011-01-01

    Unconventional multi- and many-core processors (e.g. IBM (R) Cell B.E.(TM) and NVIDIA (R) GPU) have emerged as effective accelerators in trial climate and weather simulations. Yet these climate and weather models typically run on parallel computers with conventional processors (e.g. Intel, AMD, and IBM) using Message Passing Interface. To address challenges involved in efficiently and easily connecting accelerators to parallel computers, we investigated using IBM's Dynamic Application Virtualization (TM) (IBM DAV) software in a prototype hybrid computing system with representative climate and weather model components. The hybrid system comprises two Intel blades and two IBM QS22 Cell B.E. blades, connected with both InfiniBand(R) (IB) and 1-Gigabit Ethernet. The system significantly accelerates a solar radiation model component by offloading compute-intensive calculations to the Cell blades. Systematic tests show that IBM DAV can seamlessly offload compute-intensive calculations from Intel blades to Cell B.E. blades in a scalable, load-balanced manner. However, noticeable communication overhead was observed, mainly due to IP over the IB protocol. Full utilization of IB Sockets Direct Protocol and the lower latency production version of IBM DAV will reduce this overhead.

  11. Core cooling system for reactor

    International Nuclear Information System (INIS)

    Kondo, Ryoichi; Amada, Tatsuo.

    1976-01-01

    Purpose: To improve the function of residual heat dissipation from the reactor core in case of emergency by providing a secondary cooling system flow channel, through which fluid having been subjected to heat exchange with the fluid flowing in a primary cooling system flow channel flows, with a core residual heat removal system in parallel with a main cooling system provided with a steam generator. Constitution: Heat generated in the core during normal reactor operation is transferred from a primary cooling system flow channel to a secondary cooling system flow channel through a main heat exchanger and then transferred through a steam generator to a water-steam system flow channel. In the event if removal of heat from the core by the main cooling system becomes impossible due to such cause as breakage of the duct line of the primary cooling system flow channel or a trouble in a primary cooling system pump, a flow control valve is opened, and steam generator inlet and outlet valves are closed, thus increasing the flow rate in the core residual heat removal system. Thereafter, a blower is started to cause dissipation of the core residual heat from the flow channel of a system for heat dissipation to atmosphere. (Seki, T.)

  12. Parallel External Memory Graph Algorithms

    DEFF Research Database (Denmark)

    Arge, Lars Allan; Goodrich, Michael T.; Sitchinava, Nodari

    2010-01-01

    In this paper, we study parallel I/O efficient graph algorithms in the Parallel External Memory (PEM) model, one o f the private-cache chip multiprocessor (CMP) models. We study the fundamental problem of list ranking which leads to efficient solutions to problems on trees, such as computing lowest...... an optimal speedup of ¿(P) in parallel I/O complexity and parallel computation time, compared to the single-processor external memory counterparts....

  13. Parallel inter channel interaction mechanisms

    International Nuclear Information System (INIS)

    Jovic, V.; Afgan, N.; Jovic, L.

    1995-01-01

    Parallel channels interactions are examined. For experimental researches of nonstationary regimes flow in three parallel vertical channels results of phenomenon analysis and mechanisms of parallel channel interaction for adiabatic condition of one-phase fluid and two-phase mixture flow are shown. (author)

  14. Core microbiomes for sustainable agroecosystems.

    Science.gov (United States)

    Toju, Hirokazu; Peay, Kabir G; Yamamichi, Masato; Narisawa, Kazuhiko; Hiruma, Kei; Naito, Ken; Fukuda, Shinji; Ushio, Masayuki; Nakaoka, Shinji; Onoda, Yusuke; Yoshida, Kentaro; Schlaeppi, Klaus; Bai, Yang; Sugiura, Ryo; Ichihashi, Yasunori; Minamisawa, Kiwamu; Kiers, E Toby

    2018-05-01

    In an era of ecosystem degradation and climate change, maximizing microbial functions in agroecosystems has become a prerequisite for the future of global agriculture. However, managing species-rich communities of plant-associated microbiomes remains a major challenge. Here, we propose interdisciplinary research strategies to optimize microbiome functions in agroecosystems. Informatics now allows us to identify members and characteristics of 'core microbiomes', which may be deployed to organize otherwise uncontrollable dynamics of resident microbiomes. Integration of microfluidics, robotics and machine learning provides novel ways to capitalize on core microbiomes for increasing resource-efficiency and stress-resistance of agroecosystems.

  15. [Constructing climate. From classical climatology to modern climate research].

    Science.gov (United States)

    Heymann, Matthias

    2009-01-01

    Both climate researchers and historians of climate science have conceived climate as a stable and well defined category. This article argues that such a conception is flawed. In the course of the 19th and 20th century the very concept of climate changed considerably. Scientists came up with different definitions and concepts of climate, which implied different understandings, interests, and research approaches. Understanding climate shifted from a timeless, spatial concept at the end of the 19th century to a spaceless, temporal concept at the end of the 20th. Climatologists in the 19th and early 20th centuries considered climate as a set of atmospheric characteristics associated with specific places or regions. In this context, while the weather was subject to change, climate remained largely stable. Of particular interest was the impact of climate on human beings and the environment. In modern climate research at the close of the 20th century, the concept of climate lost its temporal stability. Instead, climate change has become a core feature of the understanding of climate and a focus of research interests. Climate has also lost its immediate association with specific geographical places and become global. The interest is now focused on the impact of human beings on climate. The paper attempts to investigate these conceptual shifts and their origins and impacts in order to provide a more comprehensive perspective on the history of climate research.

  16. Massively Parallel QCD

    International Nuclear Information System (INIS)

    Soltz, R; Vranas, P; Blumrich, M; Chen, D; Gara, A; Giampap, M; Heidelberger, P; Salapura, V; Sexton, J; Bhanot, G

    2007-01-01

    The theory of the strong nuclear force, Quantum Chromodynamics (QCD), can be numerically simulated from first principles on massively-parallel supercomputers using the method of Lattice Gauge Theory. We describe the special programming requirements of lattice QCD (LQCD) as well as the optimal supercomputer hardware architectures that it suggests. We demonstrate these methods on the BlueGene massively-parallel supercomputer and argue that LQCD and the BlueGene architecture are a natural match. This can be traced to the simple fact that LQCD is a regular lattice discretization of space into lattice sites while the BlueGene supercomputer is a discretization of space into compute nodes, and that both are constrained by requirements of locality. This simple relation is both technologically important and theoretically intriguing. The main result of this paper is the speedup of LQCD using up to 131,072 CPUs on the largest BlueGene/L supercomputer. The speedup is perfect with sustained performance of about 20% of peak. This corresponds to a maximum of 70.5 sustained TFlop/s. At these speeds LQCD and BlueGene are poised to produce the next generation of strong interaction physics theoretical results

  17. Exact parallel maximum clique algorithm for general and protein graphs.

    Science.gov (United States)

    Depolli, Matjaž; Konc, Janez; Rozman, Kati; Trobec, Roman; Janežič, Dušanka

    2013-09-23

    A new exact parallel maximum clique algorithm MaxCliquePara, which finds the maximum clique (the fully connected subgraph) in undirected general and protein graphs, is presented. First, a new branch and bound algorithm for finding a maximum clique on a single computer core, which builds on ideas presented in two published state of the art sequential algorithms is implemented. The new sequential MaxCliqueSeq algorithm is faster than the reference algorithms on both DIMACS benchmark graphs as well as on protein-derived product graphs used for protein structural comparisons. Next, the MaxCliqueSeq algorithm is parallelized by splitting the branch-and-bound search tree to multiple cores, resulting in MaxCliquePara algorithm. The ability to exploit all cores efficiently makes the new parallel MaxCliquePara algorithm markedly superior to other tested algorithms. On a 12-core computer, the parallelization provides up to 2 orders of magnitude faster execution on the large DIMACS benchmark graphs and up to an order of magnitude faster execution on protein product graphs. The algorithms are freely accessible on http://commsys.ijs.si/~matjaz/maxclique.

  18. Integrated Current Balancing Transformer for Primary Parallel Isolated Boost Converter

    DEFF Research Database (Denmark)

    Sen, Gökhan; Ouyang, Ziwei; Thomsen, Ole Cornelius

    2011-01-01

    A simple, PCB compatible integrated solution is proposed for the current balancing requirement of the primary parallel isolated boost converter (PPIBC). Input inductor and the current balancing transformer are merged into the same core, which reduces the number of components allowing a cheaper...

  19. Scaling up machine learning: parallel and distributed approaches

    National Research Council Canada - National Science Library

    Bekkerman, Ron; Bilenko, Mikhail; Langford, John

    2012-01-01

    ... presented in the book cover a range of parallelization platforms from FPGAs and GPUs to multi-core systems and commodity clusters; concurrent programming frameworks that include CUDA, MPI, MapReduce, and DryadLINQ; and various learning settings: supervised, unsupervised, semi-supervised, and online learning. Extensive coverage of parallelizat...

  20. A Parallel Algebraic Multigrid Solver on Graphics Processing Units

    KAUST Repository

    Haase, Gundolf; Liebmann, Manfred; Douglas, Craig C.; Plank, Gernot

    2010-01-01

    -vector multiplication scheme underlying the PCG-AMG algorithm is presented for the many-core GPU architecture. A performance comparison of the parallel solver shows that a singe Nvidia Tesla C1060 GPU board delivers the performance of a sixteen node Infiniband cluster

  1. Method for orienting a borehole core

    International Nuclear Information System (INIS)

    Henry, W.

    1980-01-01

    A method is described for longitudinally orienting a borehold core with respect to the longitudinal axis of the drill string which drilled said borehold core in such a manner that the original longitudinal attitude of said borehold core within the earth may be determined. At least a portion of said borehold core is partialy demagnetized in steps to thereby at least partially remove in steps the artificial remanent magnetism imparted to said borehole core by said drill string. The artifical remanent magnetism is oriented substantially parallel to the longitudinal axis of said drill string. The direction and intensity of the total magnetism of said borehold core is measured at desired intervals during the partial demagnetizing procedure. An artificial remanent magnetism vector is established which extends from the final measurement of the direction and intensity of the total magnetism of said borehole core taken during said partial demagnetizing procedure towards the initial measurement of the direction and intensity of the total magnetism of said borehold core taken during said partial demagnetizing procedure. The borehold core is oriented in such a manner that said artificial remanent magnetism vector points at least substantially downwardly towards the bottom of said borehold core for a borehold in the northern hemisphere and points at least substantailly upwardly towards the top of said borehole core for a borehole in the southern hemisphere

  2. Parallel computing by Monte Carlo codes MVP/GMVP

    International Nuclear Information System (INIS)

    Nagaya, Yasunobu; Nakagawa, Masayuki; Mori, Takamasa

    2001-01-01

    General-purpose Monte Carlo codes MVP/GMVP are well-vectorized and thus enable us to perform high-speed Monte Carlo calculations. In order to achieve more speedups, we parallelized the codes on the different types of parallel computing platforms or by using a standard parallelization library MPI. The platforms used for benchmark calculations are a distributed-memory vector-parallel computer Fujitsu VPP500, a distributed-memory massively parallel computer Intel paragon and a distributed-memory scalar-parallel computer Hitachi SR2201, IBM SP2. As mentioned generally, linear speedup could be obtained for large-scale problems but parallelization efficiency decreased as the batch size per a processing element(PE) was smaller. It was also found that the statistical uncertainty for assembly powers was less than 0.1% by the PWR full-core calculation with more than 10 million histories and it took about 1.5 hours by massively parallel computing. (author)

  3. Parallel Computing in SCALE

    International Nuclear Information System (INIS)

    DeHart, Mark D.; Williams, Mark L.; Bowman, Stephen M.

    2010-01-01

    The SCALE computational architecture has remained basically the same since its inception 30 years ago, although constituent modules and capabilities have changed significantly. This SCALE concept was intended to provide a framework whereby independent codes can be linked to provide a more comprehensive capability than possible with the individual programs - allowing flexibility to address a wide variety of applications. However, the current system was designed originally for mainframe computers with a single CPU and with significantly less memory than today's personal computers. It has been recognized that the present SCALE computation system could be restructured to take advantage of modern hardware and software capabilities, while retaining many of the modular features of the present system. Preliminary work is being done to define specifications and capabilities for a more advanced computational architecture. This paper describes the state of current SCALE development activities and plans for future development. With the release of SCALE 6.1 in 2010, a new phase of evolutionary development will be available to SCALE users within the TRITON and NEWT modules. The SCALE (Standardized Computer Analyses for Licensing Evaluation) code system developed by Oak Ridge National Laboratory (ORNL) provides a comprehensive and integrated package of codes and nuclear data for a wide range of applications in criticality safety, reactor physics, shielding, isotopic depletion and decay, and sensitivity/uncertainty (S/U) analysis. Over the last three years, since the release of version 5.1 in 2006, several important new codes have been introduced within SCALE, and significant advances applied to existing codes. Many of these new features became available with the release of SCALE 6.0 in early 2009. However, beginning with SCALE 6.1, a first generation of parallel computing is being introduced. In addition to near-term improvements, a plan for longer term SCALE enhancement

  4. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2006-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. (author). 27 refs., 18 figs., 2 tabs

  5. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2005-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. (author). 27 refs., 18 figs., 3 tabs

  6. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2007-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. (author). 27 refs., 18 figs., 2 tabs

  7. Parallel Polarization State Generation.

    Science.gov (United States)

    She, Alan; Capasso, Federico

    2016-05-17

    The control of polarization, an essential property of light, is of wide scientific and technological interest. The general problem of generating arbitrary time-varying states of polarization (SOP) has always been mathematically formulated by a series of linear transformations, i.e. a product of matrices, imposing a serial architecture. Here we show a parallel architecture described by a sum of matrices. The theory is experimentally demonstrated by modulating spatially-separated polarization components of a laser using a digital micromirror device that are subsequently beam combined. This method greatly expands the parameter space for engineering devices that control polarization. Consequently, performance characteristics, such as speed, stability, and spectral range, are entirely dictated by the technologies of optical intensity modulation, including absorption, reflection, emission, and scattering. This opens up important prospects for polarization state generation (PSG) with unique performance characteristics with applications in spectroscopic ellipsometry, spectropolarimetry, communications, imaging, and security.

  8. Parallelization of a Monte Carlo particle transport simulation code

    Science.gov (United States)

    Hadjidoukas, P.; Bousis, C.; Emfietzoglou, D.

    2010-05-01

    We have developed a high performance version of the Monte Carlo particle transport simulation code MC4. The original application code, developed in Visual Basic for Applications (VBA) for Microsoft Excel, was first rewritten in the C programming language for improving code portability. Several pseudo-random number generators have been also integrated and studied. The new MC4 version was then parallelized for shared and distributed-memory multiprocessor systems using the Message Passing Interface. Two parallel pseudo-random number generator libraries (SPRNG and DCMT) have been seamlessly integrated. The performance speedup of parallel MC4 has been studied on a variety of parallel computing architectures including an Intel Xeon server with 4 dual-core processors, a Sun cluster consisting of 16 nodes of 2 dual-core AMD Opteron processors and a 200 dual-processor HP cluster. For large problem size, which is limited only by the physical memory of the multiprocessor server, the speedup results are almost linear on all systems. We have validated the parallel implementation against the serial VBA and C implementations using the same random number generator. Our experimental results on the transport and energy loss of electrons in a water medium show that the serial and parallel codes are equivalent in accuracy. The present improvements allow for studying of higher particle energies with the use of more accurate physical models, and improve statistics as more particles tracks can be simulated in low response time.

  9. Parallel imaging microfluidic cytometer.

    Science.gov (United States)

    Ehrlich, Daniel J; McKenna, Brian K; Evans, James G; Belkina, Anna C; Denis, Gerald V; Sherr, David H; Cheung, Man Ching

    2011-01-01

    By adding an additional degree of freedom from multichannel flow, the parallel microfluidic cytometer (PMC) combines some of the best features of fluorescence-activated flow cytometry (FCM) and microscope-based high-content screening (HCS). The PMC (i) lends itself to fast processing of large numbers of samples, (ii) adds a 1D imaging capability for intracellular localization assays (HCS), (iii) has a high rare-cell sensitivity, and (iv) has an unusual capability for time-synchronized sampling. An inability to practically handle large sample numbers has restricted applications of conventional flow cytometers and microscopes in combinatorial cell assays, network biology, and drug discovery. The PMC promises to relieve a bottleneck in these previously constrained applications. The PMC may also be a powerful tool for finding rare primary cells in the clinic. The multichannel architecture of current PMC prototypes allows 384 unique samples for a cell-based screen to be read out in ∼6-10 min, about 30 times the speed of most current FCM systems. In 1D intracellular imaging, the PMC can obtain protein localization using HCS marker strategies at many times for the sample throughput of charge-coupled device (CCD)-based microscopes or CCD-based single-channel flow cytometers. The PMC also permits the signal integration time to be varied over a larger range than is practical in conventional flow cytometers. The signal-to-noise advantages are useful, for example, in counting rare positive cells in the most difficult early stages of genome-wide screening. We review the status of parallel microfluidic cytometry and discuss some of the directions the new technology may take. Copyright © 2011 Elsevier Inc. All rights reserved.

  10. Separated core turbofan engine; Core bunrigata turbofan engine

    Energy Technology Data Exchange (ETDEWEB)

    Saito, Y; Endo, M; Matsuda, Y; Sugiyama, N; Sugahara, N; Yamamoto, K [National Aerospace Laboratory, Tokyo (Japan)

    1996-04-01

    This report outlines the separated core turbofan engine. This engine is featured by parallel separated arrangement of a fan and core engine which are integrated into one unit in the conventional turbofan engine. In general, cruising efficiency improvement and noise reduction are achieved by low fan pressure ratio and low exhaust speed due to high bypass ratio, however, it causes various problems such as large fan and nacelle weight due to large air flow rate of a fan, and shift of an operating point affected by flight speed. The parallel separated arrangement is thus adopted. The stable operation of a fan and core engine is easily retained by independently operating air inlet unaffected by fan. The large degree of freedom of combustion control is also obtained by independent combustor. Fast response, simple structure and optimum aerodynamic design are easily achieved. This arrangement is also featured by flexibility of development and easy maintenance, and by various merits superior to conventional turbofan engines. It has no technological problems difficult to be overcome, and is also suitable for high-speed VTOL transport aircraft. 4 refs., 5 figs.

  11. ONE SEGMENT OF THE BULGARIAN-ENGLISH PAREMIOLOGICAL CORE

    Directory of Open Access Journals (Sweden)

    KOTOVA M.Y.

    2015-12-01

    Full Text Available The English proverbial parallels of the Russian-Bulgarian paremiological core are analysed in the article. The comparison of current Bulgarian proverbs and their English proverbial parallels is based upon the material of the author’s multi-lingual dictionary and her collection of Bulgarian-Russian proverbial parallels published as a result of her sociolinguistic paremiological experiment from 2003 (on the basis of 100 questionnaires filled by 100 Bulgarian respondents and supported in 2013 with the current Bulgarian contexts from the Bulgarian Internet. The number of 'alive' Bulgarian-English proverbial parallels, constructed from the paremiological questionnaires (pointed out by 70 % - 100 % respondents is 62, the biggest part of which belongs to the proverbial parallels with a similar inner form (35, i.e. the biggest part of the segment of the current Bulgarian-English paremiological core (reflecting the Russian paremiological minimum contains proverbial parallels with a similar inner form.

  12. Out-of-order parallel discrete event simulation for electronic system-level design

    CERN Document Server

    Chen, Weiwei

    2014-01-01

    This book offers readers a set of new approaches and tools a set of tools and techniques for facing challenges in parallelization with design of embedded systems.? It provides an advanced parallel simulation infrastructure for efficient and effective system-level model validation and development so as to build better products in less time.? Since parallel discrete event simulation (PDES) has the potential to exploit the underlying parallel computational capability in today's multi-core simulation hosts, the author begins by reviewing the parallelization of discrete event simulation, identifyin

  13. Domain Specific Language for Geant4 Parallelization for Space-based Applications, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — A major limiting factor in HPC growth is the requirement to parallelize codes to leverage emerging architectures, especially as single core performance has plateaued...

  14. About Parallel Programming: Paradigms, Parallel Execution and Collaborative Systems

    Directory of Open Access Journals (Sweden)

    Loredana MOCEAN

    2009-01-01

    Full Text Available In the last years, there were made efforts for delineation of a stabile and unitary frame, where the problems of logical parallel processing must find solutions at least at the level of imperative languages. The results obtained by now are not at the level of the made efforts. This paper wants to be a little contribution at these efforts. We propose an overview in parallel programming, parallel execution and collaborative systems.

  15. SCORPIO - WWER core surveillance system

    International Nuclear Information System (INIS)

    Hornaes, Arne; Bodal, Terje; Sunde, Svein; Zalesky, K.; Lehman, M.; Pecka, M.; Svarny, J.; Krysl, V.; Juzova, Z.; Sedlak, A.; Semmler, M.

    1998-01-01

    The Institut for energiteknikk has developed the core surveillance system SCORPIO, which has two parallel modes of operation: the Core Follow Mode and the Predictive Mode. The main motivation behind the development of SCORPIO is to make a practical tool for reactor operators which can increase the quality and quantity of information presented on core status and dynamic behavior. This can first of all improve plant safety, as undesired core conditions are detected and prevented. Secondly, more flexible and efficient plant operation is made possible. The system has been implemented on western PWRs, but the basic concept is applicable to a wide range of reactors including WWERs. The main differences between WWERs and typical western PWRs with respect to core surveillance requirements are outlined. The development of a WWER version of SCORPIO has been done in co-operation with the Nuclear Research Institute Rez, and industry partners in the Czech Republic. The first system is installed at Dukovany NPP, where the Site Acceptance Test was completed 6. March 1998.(Authors)

  16. SCORPIO - VVER core surveillance system

    International Nuclear Information System (INIS)

    Hornaes, A.; Bodal, T.; Sunde, S.

    1998-01-01

    The Institutt for energiteknikk has developed the core surveillance system SCORPIO, which has two parallel modes of operation: the Core Follow Mode and the Predictive Mode. The main motivation behind the development of SCORPIO is to make a practical tool for reactor operators, which can increase the quality and quantity of information presented on core status and dynamic behavior. This can first of all improve plant safety, as undesired core conditions are detected and prevented. Secondly, more flexible and efficient plant operation is made possible. The system has been implemented on western PWRs, but the basic concept is applicable to a wide range of reactors including VVERs. The main differences between VVERs and typical western PWRs with respect to core surveillance requirements are outlined. The development of a VVER version of SCORPIO has been done in co-operation with the Nuclear Research Institute Rez, and industry partners in the Czech Republic. The first system is installed at Dukovany NPP, where the Site Acceptance Test was completed 6. March 1998.(author)

  17. A 21 000-year record of fluorescent organic matter markers in the WAIS Divide ice core

    Science.gov (United States)

    D'Andrilli, Juliana; Foreman, Christine M.; Sigl, Michael; Priscu, John C.; McConnell, Joseph R.

    2017-05-01

    Englacial ice contains a significant reservoir of organic material (OM), preserving a chronological record of materials from Earth's past. Here, we investigate if OM composition surveys in ice core research can provide paleoecological information on the dynamic nature of our Earth through time. Temporal trends in OM composition from the early Holocene extending back to the Last Glacial Maximum (LGM) of the West Antarctic Ice Sheet Divide (WD) ice core were measured by fluorescence spectroscopy. Multivariate parallel factor (PARAFAC) analysis is widely used to isolate the chemical components that best describe the observed variation across three-dimensional fluorescence spectroscopy (excitation-emission matrices; EEMs) assays. Fluorescent OM markers identified by PARAFAC modeling of the EEMs from the LGM (27.0-18.0 kyr BP; before present 1950) through the last deglaciation (LD; 18.0-11.5 kyr BP), to the mid-Holocene (11.5-6.0 kyr BP) provided evidence of different types of fluorescent OM composition and origin in the WD ice core over 21.0 kyr. Low excitation-emission wavelength fluorescent PARAFAC component one (C1), associated with chemical species similar to simple lignin phenols was the greatest contributor throughout the ice core, suggesting a strong signature of terrestrial OM in all climate periods. The component two (C2) OM marker, encompassed distinct variability in the ice core describing chemical species similar to tannin- and phenylalanine-like material. Component three (C3), associated with humic-like terrestrial material further resistant to biodegradation, was only characteristic of the Holocene, suggesting that more complex organic polymers such as lignins or tannins may be an ecological marker of warmer climates. We suggest that fluorescent OM markers observed during the LGM were the result of greater continental dust loading of lignin precursor (monolignol) material in a drier climate, with lower marine influences when sea ice extent was higher and

  18. Intelligent spatial ecosystem modeling using parallel processors

    International Nuclear Information System (INIS)

    Maxwell, T.; Costanza, R.

    1993-01-01

    Spatial modeling of ecosystems is essential if one's modeling goals include developing a relatively realistic description of past behavior and predictions of the impacts of alternative management policies on future ecosystem behavior. Development of these models has been limited in the past by the large amount of input data required and the difficulty of even large mainframe serial computers in dealing with large spatial arrays. These two limitations have begun to erode with the increasing availability of remote sensing data and GIS systems to manipulate it, and the development of parallel computer systems which allow computation of large, complex, spatial arrays. Although many forms of dynamic spatial modeling are highly amenable to parallel processing, the primary focus in this project is on process-based landscape models. These models simulate spatial structure by first compartmentalizing the landscape into some geometric design and then describing flows within compartments and spatial processes between compartments according to location-specific algorithms. The authors are currently building and running parallel spatial models at the regional scale for the Patuxent River region in Maryland, the Everglades in Florida, and Barataria Basin in Louisiana. The authors are also planning a project to construct a series of spatially explicit linked ecological and economic simulation models aimed at assessing the long-term potential impacts of global climate change

  19. Beam dynamics simulations using a parallel version of PARMILA

    International Nuclear Information System (INIS)

    Ryne, R.D.

    1996-01-01

    The computer code PARMILA has been the primary tool for the design of proton and ion linacs in the United States for nearly three decades. Previously it was sufficient to perform simulations with of order 10000 particles, but recently the need to perform high resolution halo studies for next-generation, high intensity linacs has made it necessary to perform simulations with of order 100 million particles. With the advent of massively parallel computers such simulations are now within reach. Parallel computers already make it possible, for example, to perform beam dynamics calculations with tens of millions of particles, requiring over 10 GByte of core memory, in just a few hours. Also, parallel computers are becoming easier to use thanks to the availability of mature, Fortran-like languages such as Connection Machine Fortran and High Performance Fortran. We will describe our experience developing a parallel version of PARMILA and the performance of the new code

  20. Beam dynamics simulations using a parallel version of PARMILA

    International Nuclear Information System (INIS)

    Ryne, Robert

    1996-01-01

    The computer code PARMILA has been the primary tool for the design of proton and ion linacs in the United States for nearly three decades. Previously it was sufficient to perform simulations with of order 10000 particles, but recently the need to perform high resolution halo studies for next-generation, high intensity linacs has made it necessary to perform simulations with of order 100 million particles. With the advent of massively parallel computers such simulations are now within reach. Parallel computers already make it possible, for example, to perform beam dynamics calculations with tens of millions of particles, requiring over 10 GByte of core memory, in just a few hours. Also, parallel computers are becoming easier to use thanks to the availability of mature, Fortran-like languages such as Connection Machine Fortran and High Performance Fortran. We will describe our experience developing a parallel version of PARMILA and the performance of the new code. (author)

  1. Parallel Framework for Cooperative Processes

    Directory of Open Access Journals (Sweden)

    Mitică Craus

    2005-01-01

    Full Text Available This paper describes the work of an object oriented framework designed to be used in the parallelization of a set of related algorithms. The idea behind the system we are describing is to have a re-usable framework for running several sequential algorithms in a parallel environment. The algorithms that the framework can be used with have several things in common: they have to run in cycles and the work should be possible to be split between several "processing units". The parallel framework uses the message-passing communication paradigm and is organized as a master-slave system. Two applications are presented: an Ant Colony Optimization (ACO parallel algorithm for the Travelling Salesman Problem (TSP and an Image Processing (IP parallel algorithm for the Symmetrical Neighborhood Filter (SNF. The implementations of these applications by means of the parallel framework prove to have good performances: approximatively linear speedup and low communication cost.

  2. Parallel Monte Carlo reactor neutronics

    International Nuclear Information System (INIS)

    Blomquist, R.N.; Brown, F.B.

    1994-01-01

    The issues affecting implementation of parallel algorithms for large-scale engineering Monte Carlo neutron transport simulations are discussed. For nuclear reactor calculations, these include load balancing, recoding effort, reproducibility, domain decomposition techniques, I/O minimization, and strategies for different parallel architectures. Two codes were parallelized and tested for performance. The architectures employed include SIMD, MIMD-distributed memory, and workstation network with uneven interactive load. Speedups linear with the number of nodes were achieved

  3. Anti-parallel triplexes

    DEFF Research Database (Denmark)

    Kosbar, Tamer R.; Sofan, Mamdouh A.; Waly, Mohamed A.

    2015-01-01

    about 6.1 °C when the TFO strand was modified with Z and the Watson-Crick strand with adenine-LNA (AL). The molecular modeling results showed that, in case of nucleobases Y and Z a hydrogen bond (1.69 and 1.72 Å, respectively) was formed between the protonated 3-aminopropyn-1-yl chain and one...... of the phosphate groups in Watson-Crick strand. Also, it was shown that the nucleobase Y made a good stacking and binding with the other nucleobases in the TFO and Watson-Crick duplex, respectively. In contrast, the nucleobase Z with LNA moiety was forced to twist out of plane of Watson-Crick base pair which......The phosphoramidites of DNA monomers of 7-(3-aminopropyn-1-yl)-8-aza-7-deazaadenine (Y) and 7-(3-aminopropyn-1-yl)-8-aza-7-deazaadenine LNA (Z) are synthesized, and the thermal stability at pH 7.2 and 8.2 of anti-parallel triplexes modified with these two monomers is determined. When, the anti...

  4. Parallel consensual neural networks.

    Science.gov (United States)

    Benediktsson, J A; Sveinsson, J R; Ersoy, O K; Swain, P H

    1997-01-01

    A new type of a neural-network architecture, the parallel consensual neural network (PCNN), is introduced and applied in classification/data fusion of multisource remote sensing and geographic data. The PCNN architecture is based on statistical consensus theory and involves using stage neural networks with transformed input data. The input data are transformed several times and the different transformed data are used as if they were independent inputs. The independent inputs are first classified using the stage neural networks. The output responses from the stage networks are then weighted and combined to make a consensual decision. In this paper, optimization methods are used in order to weight the outputs from the stage networks. Two approaches are proposed to compute the data transforms for the PCNN, one for binary data and another for analog data. The analog approach uses wavelet packets. The experimental results obtained with the proposed approach show that the PCNN outperforms both a conjugate-gradient backpropagation neural network and conventional statistical methods in terms of overall classification accuracy of test data.

  5. Climate scenarios for California

    Science.gov (United States)

    Cayan, Daniel R.; Maurer, Ed; Dettinger, Mike; Tyree, Mary; Hayhoe, Katharine; Bonfils, Celine; Duffy, Phil; Santer, Ben

    2006-01-01

    Possible future climate changes in California are investigated from a varied set of climate change model simulations. These simulations, conducted by three state-of-the-art global climate models, provide trajectories from three greenhouse gas (GHG) emission scenarios. These scenarios and the resulting climate simulations are not “predictions,” but rather are a limited sample from among the many plausible pathways that may affect California’s climate. Future GHG concentrations are uncertain because they depend on future social, political, and technological pathways, and thus the IPCC has produced four “families” of emission scenarios. To explore some of these uncertainties, emissions scenarios A2 (a medium-high emissions) and B1 (low emissions) were selected from the current IPCC Fourth climate assessment, which provides several recent model simulations driven by A2 and B1 emissions. The global climate model simulations addressed here were from PCM1, the Parallel Climate Model from the National Center for Atmospheric Research (NCAR) and U.S. Department of Energy (DOE) group, and CM2.1 from the National Oceanic and Atmospheric Administration (NOAA) Geophysical Fluids Dynamics Laboratory (GFDL).

  6. A Parallel Particle Swarm Optimizer

    National Research Council Canada - National Science Library

    Schutte, J. F; Fregly, B .J; Haftka, R. T; George, A. D

    2003-01-01

    .... Motivated by a computationally demanding biomechanical system identification problem, we introduce a parallel implementation of a stochastic population based global optimizer, the Particle Swarm...

  7. Patterns for Parallel Software Design

    CERN Document Server

    Ortega-Arjona, Jorge Luis

    2010-01-01

    Essential reading to understand patterns for parallel programming Software patterns have revolutionized the way we think about how software is designed, built, and documented, and the design of parallel software requires you to consider other particular design aspects and special skills. From clusters to supercomputers, success heavily depends on the design skills of software developers. Patterns for Parallel Software Design presents a pattern-oriented software architecture approach to parallel software design. This approach is not a design method in the classic sense, but a new way of managin

  8. Seeing or moving in parallel

    DEFF Research Database (Denmark)

    Christensen, Mark Schram; Ehrsson, H Henrik; Nielsen, Jens Bo

    2013-01-01

    a different network, involving bilateral dorsal premotor cortex (PMd), primary motor cortex, and SMA, was more active when subjects viewed parallel movements while performing either symmetrical or parallel movements. Correlations between behavioral instability and brain activity were present in right lateral...... adduction-abduction movements symmetrically or in parallel with real-time congruent or incongruent visual feedback of the movements. One network, consisting of bilateral superior and middle frontal gyrus and supplementary motor area (SMA), was more active when subjects performed parallel movements, whereas...

  9. Parallel Task Processing on a Multicore Platform in a PC-based Control System for Parallel Kinematics

    Directory of Open Access Journals (Sweden)

    Harald Michalik

    2009-02-01

    Full Text Available Multicore platforms are such that have one physical processor chip with multiple cores interconnected via a chip level bus. Because they deliver a greater computing power through concurrency, offer greater system density multicore platforms provide best qualifications to address the performance bottleneck encountered in PC-based control systems for parallel kinematic robots with heavy CPU-load. Heavy load control tasks are generated by new control approaches that include features like singularity prediction, structure control algorithms, vision data integration and similar tasks. In this paper we introduce the parallel task scheduling extension of a communication architecture specially tailored for the development of PC-based control of parallel kinematics. The Sche-duling is specially designed for the processing on a multicore platform. It breaks down the serial task processing of the robot control cycle and extends it with parallel task processing paths in order to enhance the overall control performance.

  10. CITYZEN climate impact studies

    Energy Technology Data Exchange (ETDEWEB)

    Schutz, Martin (ed.)

    2011-07-01

    We have estimated the impact of climate change on the chemical composition of the troposphere due to changes in climate from current climate (2000-2010) looking 40 years ahead (2040-2050). The climate projection has been made by the ECHAM5 model and was followed by chemistry-transport modelling using a global model, Oslo CTM2 (Isaksen et al., 2005; Srvde et al., 2008), and a regional model, EMEP. In this report we focus on carbon monoxide (CO) and surface ozone (O3) which are measures of primary and secondary air pollution. In parallel we have estimated the change in the same air pollutants resulting from changes in emissions over the same time period. (orig.)

  11. An overview of European efforts in generating climate data records

    NARCIS (Netherlands)

    Su, Z.; Timmermans, W.J.; Zeng, Y.; Schulz, J.; John, V.O.; Roebeling, R.A.; Poli, P.; Tan, D.; Kaspar, F.; Kaiser-Weiss, A.; Swinnen, E.; Tote, C.; Gregow, H.; Manninen, T.; Riihela, A.; Calvet, J.C.; Ma, Yaoming; Wen, Jun

    2018-01-01

    The Coordinating Earth Observation Data Validation for Reanalysis for Climate Services project (CORE-CLIMAX) aimed to substantiate how Copernicus observations and products can contribute to climate change analyses. CORE-CLIMAX assessed the European capability to provide climate data records (CDRs)

  12. Core clamping device for a nuclear reactor

    International Nuclear Information System (INIS)

    Guenther, R.W.

    1974-01-01

    The core clamping device for a fast neutron reactor includes clamps to support the fuel zone against the pressure vessel. The clamps are arranged around the circumference of the core. They consist of torsion bars arranged parallel at some distance around the core with lever arms attached to the ends whose force is directed in the opposite direction, pressing against the wall of the pressure vessel. The lever arms and pressure plates also actuated by the ends of the torsion bars transfer the stress, the pressure plates acting upon the fuel elements or fuel assemblies. Coupling between the ends of the torsion bars and the pressure plates is achieved by end carrier plates directly attached to the torsion bars and radially movable. This clamping device follows the thermal expansions of the core, allows specific elements to be disengaged in sections and saves space between the core and the neutron reflectors. (DG) [de

  13. Side core lifter

    Energy Technology Data Exchange (ETDEWEB)

    Edelman, Ya A

    1982-01-01

    A side core lifter is proposed which contains a housing with guide slits and a removable core lifter with side projections on the support section connected to the core receiver. In order to preserve the structure of the rock in the core sample by means of guaranteeing rectilinear movement of the core lifter in the rock, the support and core receiver sections are hinged. The device is equipped with a spring for angular shift in the core-reception part.

  14. PERFORMANCE EVALUATION OF OR1200 PROCESSOR WITH EVOLUTIONARY PARALLEL HPRC USING GEP

    Directory of Open Access Journals (Sweden)

    R. Maheswari

    2012-04-01

    Full Text Available In this fast computing era, most of the embedded system requires more computing power to complete the complex function/ task at the lesser amount of time. One way to achieve this is by boosting up the processor performance which allows processor core to run faster. This paper presents a novel technique of increasing the performance by parallel HPRC (High Performance Reconfigurable Computing in the CPU/DSP (Digital Signal Processor unit of OR1200 (Open Reduced Instruction Set Computer (RISC 1200 using Gene Expression Programming (GEP an evolutionary programming model. OR1200 is a soft-core RISC processor of the Intellectual Property cores that can efficiently run any modern operating system. In the manufacturing process of OR1200 a parallel HPRC is placed internally in the Integer Execution Pipeline unit of the CPU/DSP core to increase the performance. The GEP Parallel HPRC is activated /deactivated by triggering the signals i HPRC_Gene_Start ii HPRC_Gene_End. A Verilog HDL(Hardware Description language functional code for Gene Expression Programming parallel HPRC is developed and synthesised using XILINX ISE in the former part of the work and a CoreMark processor core benchmark is used to test the performance of the OR1200 soft core in the later part of the work. The result of the implementation ensures the overall speed-up increased to 20.59% by GEP based parallel HPRC in the execution unit of OR1200.

  15. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2009-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. Ice core records provide an annual-scale, 'instrumental-quality' baseline of atmospheric temperature and circulation changes back many thousands of years. (author). 45 refs., 16 figs., 2 tabs.

  16. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2009-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. Ice core records provide an annual-scale, 'instrumental-quality' baseline of atmospheric temperature and circulation changes back many thousands of years. (author). 27 refs., 18 figs., 2 tabs

  17. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.A.N.

    2012-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. Ice core records provide an annual-scale, 'instrumental-quality' baseline of atmospheric temperature and circulation changes back many thousands of years. (author). 28 refs., 20 figs., 1 tab.

  18. Stable isotope analysis in ice core paleoclimatology

    International Nuclear Information System (INIS)

    Bertler, N.

    2008-01-01

    Ice cores from New Zealand and the Antarctic margin provide an excellent means of addressing the lack of longer-term climate observations in the Southern Hemisphere with near instrumental quality. Their study helps us to improve our understanding of regional patterns of climate behaviour in Antarctica and its influence on New Zealand, leading to more realistic regional climate models. Such models are needed to sensibly interpret current Antarctic and New Zealand climate variability and for the development of appropriate mitigation strategies for New Zealand. Ice core records provide an annual-scale, 'instrumental-quality' baseline of atmospheric temperature and circulation changes back many thousands of years. (author). 27 refs., 18 figs., 2 tabs

  19. Optimizations of Unstructured Aerodynamics Computations for Many-core Architectures

    KAUST Repository

    Al Farhan, Mohammed Ahmed; Keyes, David E.

    2018-01-01

    involving thread and data-level parallelism. Our approach is based upon a multi-level hierarchical distribution of work and data across both the threads and the SIMD units within every hardware core. On a 64-core KNL chip, we achieve nearly 2.9x speedup

  20. Fast ℓ1-SPIRiT Compressed Sensing Parallel Imaging MRI: Scalable Parallel Implementation and Clinically Feasible Runtime

    Science.gov (United States)

    Murphy, Mark; Alley, Marcus; Demmel, James; Keutzer, Kurt; Vasanawala, Shreyas; Lustig, Michael

    2012-01-01

    We present ℓ1-SPIRiT, a simple algorithm for auto calibrating parallel imaging (acPI) and compressed sensing (CS) that permits an efficient implementation with clinically-feasible runtimes. We propose a CS objective function that minimizes cross-channel joint sparsity in the Wavelet domain. Our reconstruction minimizes this objective via iterative soft-thresholding, and integrates naturally with iterative Self-Consistent Parallel Imaging (SPIRiT). Like many iterative MRI reconstructions, ℓ1-SPIRiT’s image quality comes at a high computational cost. Excessively long runtimes are a barrier to the clinical use of any reconstruction approach, and thus we discuss our approach to efficiently parallelizing ℓ1-SPIRiT and to achieving clinically-feasible runtimes. We present parallelizations of ℓ1-SPIRiT for both multi-GPU systems and multi-core CPUs, and discuss the software optimization and parallelization decisions made in our implementation. The performance of these alternatives depends on the processor architecture, the size of the image matrix, and the number of parallel imaging channels. Fundamentally, achieving fast runtime requires the correct trade-off between cache usage and parallelization overheads. We demonstrate image quality via a case from our clinical experimentation, using a custom 3DFT Spoiled Gradient Echo (SPGR) sequence with up to 8× acceleration via poisson-disc undersampling in the two phase-encoded directions. PMID:22345529

  1. HPC parallel programming model for gyrokinetic MHD simulation

    International Nuclear Information System (INIS)

    Naitou, Hiroshi; Yamada, Yusuke; Tokuda, Shinji; Ishii, Yasutomo; Yagi, Masatoshi

    2011-01-01

    The 3-dimensional gyrokinetic PIC (particle-in-cell) code for MHD simulation, Gpic-MHD, was installed on SR16000 (“Plasma Simulator”), which is a scalar cluster system consisting of 8,192 logical cores. The Gpic-MHD code advances particle and field quantities in time. In order to distribute calculations over large number of logical cores, the total simulation domain in cylindrical geometry was broken up into N DD-r × N DD-z (number of radial decomposition times number of axial decomposition) small domains including approximately the same number of particles. The axial direction was uniformly decomposed, while the radial direction was non-uniformly decomposed. N RP replicas (copies) of each decomposed domain were used (“particle decomposition”). The hybrid parallelization model of multi-threads and multi-processes was employed: threads were parallelized by the auto-parallelization and N DD-r × N DD-z × N RP processes were parallelized by MPI (message-passing interface). The parallelization performance of Gpic-MHD was investigated for the medium size system of N r × N θ × N z = 1025 × 128 × 128 mesh with 4.196 or 8.192 billion particles. The highest speed for the fixed number of logical cores was obtained for two threads, the maximum number of N DD-z , and optimum combination of N DD-r and N RP . The observed optimum speeds demonstrated good scaling up to 8,192 logical cores. (author)

  2. PARALLEL IMPORT: REALITY FOR RUSSIA

    Directory of Open Access Journals (Sweden)

    Т. А. Сухопарова

    2014-01-01

    Full Text Available Problem of parallel import is urgent question at now. Parallel import legalization in Russia is expedient. Such statement based on opposite experts opinion analysis. At the same time it’s necessary to negative consequences consider of this decision and to apply remedies to its minimization.Purchase on Elibrary.ru > Buy now

  3. Toward an ultra-high resolution community climate system model for the BlueGene platform

    Energy Technology Data Exchange (ETDEWEB)

    Dennis, John M [Computer Science Section, National Center for Atmospheric Research, Boulder, CO (United States); Jacob, Robert [Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, IL (United States); Vertenstein, Mariana [Climate and Global Dynamics Division, National Center for Atmospheric Research, Boulder, CO (United States); Craig, Tony [Climate and Global Dynamics Division, National Center for Atmospheric Research, Boulder, CO (United States); Loy, Raymond [Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, IL (United States)

    2007-07-15

    Global climate models need to simulate several small, regional-scale processes which affect the global circulation in order to accurately simulate the climate. This is particularly important in the ocean where small scale features such as oceanic eddies are currently represented with adhoc parameterizations. There is also a need for higher resolution to provide climate predictions at small, regional scales. New high-performance computing platforms such as the IBM BlueGene can provide the necessary computational power to perform ultra-high resolution climate model integrations. We have begun to investigate the scaling of the individual components of the Community Climate System Model to prepare it for integrations on BlueGene and similar platforms. Our investigations show that it is possible to successfully utilize O(32K) processors. We describe the scalability of five models: the Parallel Ocean Program (POP), the Community Ice CodE (CICE), the Community Land Model (CLM), and the new CCSM sequential coupler (CPL7) which are components of the next generation Community Climate System Model (CCSM); as well as the High-Order Method Modeling Environment (HOMME) which is a dynamical core currently being evaluated within the Community Atmospheric Model. For our studies we concentrate on 1/10{sup 0} resolution for CICE, POP, and CLM models and 1/4{sup 0} resolution for HOMME. The ability to simulate high resolutions on the massively parallel petascale systems that will dominate high-performance computing for the foreseeable future is essential to the advancement of climate science.

  4. Shared memory parallelism for 3D cartesian discrete ordinates solver

    International Nuclear Information System (INIS)

    Moustafa, S.; Dutka-Malen, I.; Plagne, L.; Poncot, A.; Ramet, P.

    2013-01-01

    This paper describes the design and the performance of DOMINO, a 3D Cartesian SN solver that implements two nested levels of parallelism (multi-core + SIMD - Single Instruction on Multiple Data) on shared memory computation nodes. DOMINO is written in C++, a multi-paradigm programming language that enables the use of powerful and generic parallel programming tools such as Intel TBB and Eigen. These two libraries allow us to combine multi-thread parallelism with vector operations in an efficient and yet portable way. As a result, DOMINO can exploit the full power of modern multi-core processors and is able to tackle very large simulations, that usually require large HPC clusters, using a single computing node. For example, DOMINO solves a 3D full core PWR eigenvalue problem involving 26 energy groups, 288 angular directions (S16), 46*10 6 spatial cells and 1*10 12 DoFs within 11 hours on a single 32-core SMP node. This represents a sustained performance of 235 GFlops and 40.74% of the SMP node peak performance for the DOMINO sweep implementation. The very high Flops/Watt ratio of DOMINO makes it a very interesting building block for a future many-nodes nuclear simulation tool. (authors)

  5. Provably optimal parallel transport sweeps on regular grids

    Energy Technology Data Exchange (ETDEWEB)

    Adams, M. P.; Adams, M. L.; Hawkins, W. D. [Dept. of Nuclear Engineering, Texas A and M University, 3133 TAMU, College Station, TX 77843-3133 (United States); Smith, T.; Rauchwerger, L.; Amato, N. M. [Dept. of Computer Science and Engineering, Texas A and M University, 3133 TAMU, College Station, TX 77843-3133 (United States); Bailey, T. S.; Falgout, R. D. [Lawrence Livermore National Laboratory (United States)

    2013-07-01

    We have found provably optimal algorithms for full-domain discrete-ordinate transport sweeps on regular grids in 3D Cartesian geometry. We describe these algorithms and sketch a 'proof that they always execute the full eight-octant sweep in the minimum possible number of stages for a given P{sub x} x P{sub y} x P{sub z} partitioning. Computational results demonstrate that our optimal scheduling algorithms execute sweeps in the minimum possible stage count. Observed parallel efficiencies agree well with our performance model. An older version of our PDT transport code achieves almost 80% parallel efficiency on 131,072 cores, on a weak-scaling problem with only one energy group, 80 directions, and 4096 cells/core. A newer version is less efficient at present-we are still improving its implementation - but achieves almost 60% parallel efficiency on 393,216 cores. These results conclusively demonstrate that sweeps can perform with high efficiency on core counts approaching 10{sup 6}. (authors)

  6. Provably optimal parallel transport sweeps on regular grids

    International Nuclear Information System (INIS)

    Adams, M. P.; Adams, M. L.; Hawkins, W. D.; Smith, T.; Rauchwerger, L.; Amato, N. M.; Bailey, T. S.; Falgout, R. D.

    2013-01-01

    We have found provably optimal algorithms for full-domain discrete-ordinate transport sweeps on regular grids in 3D Cartesian geometry. We describe these algorithms and sketch a 'proof that they always execute the full eight-octant sweep in the minimum possible number of stages for a given P x x P y x P z partitioning. Computational results demonstrate that our optimal scheduling algorithms execute sweeps in the minimum possible stage count. Observed parallel efficiencies agree well with our performance model. An older version of our PDT transport code achieves almost 80% parallel efficiency on 131,072 cores, on a weak-scaling problem with only one energy group, 80 directions, and 4096 cells/core. A newer version is less efficient at present-we are still improving its implementation - but achieves almost 60% parallel efficiency on 393,216 cores. These results conclusively demonstrate that sweeps can perform with high efficiency on core counts approaching 10 6 . (authors)

  7. The Galley Parallel File System

    Science.gov (United States)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    Most current multiprocessor file systems are designed to use multiple disks in parallel, using the high aggregate bandwidth to meet the growing I/0 requirements of parallel scientific applications. Many multiprocessor file systems provide applications with a conventional Unix-like interface, allowing the application to access multiple disks transparently. This interface conceals the parallelism within the file system, increasing the ease of programmability, but making it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. In addition to providing an insufficient interface, most current multiprocessor file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic scientific multiprocessor workloads. We discuss Galley's file structure and application interface, as well as the performance advantages offered by that interface.

  8. Parallelization of the FLAPW method

    International Nuclear Information System (INIS)

    Canning, A.; Mannstadt, W.; Freeman, A.J.

    1999-01-01

    The FLAPW (full-potential linearized-augmented plane-wave) method is one of the most accurate first-principles methods for determining electronic and magnetic properties of crystals and surfaces. Until the present work, the FLAPW method has been limited to systems of less than about one hundred atoms due to a lack of an efficient parallel implementation to exploit the power and memory of parallel computers. In this work we present an efficient parallelization of the method by division among the processors of the plane-wave components for each state. The code is also optimized for RISC (reduced instruction set computer) architectures, such as those found on most parallel computers, making full use of BLAS (basic linear algebra subprograms) wherever possible. Scaling results are presented for systems of up to 686 silicon atoms and 343 palladium atoms per unit cell, running on up to 512 processors on a CRAY T3E parallel computer

  9. Parallelization of the FLAPW method

    Science.gov (United States)

    Canning, A.; Mannstadt, W.; Freeman, A. J.

    2000-08-01

    The FLAPW (full-potential linearized-augmented plane-wave) method is one of the most accurate first-principles methods for determining structural, electronic and magnetic properties of crystals and surfaces. Until the present work, the FLAPW method has been limited to systems of less than about a hundred atoms due to the lack of an efficient parallel implementation to exploit the power and memory of parallel computers. In this work, we present an efficient parallelization of the method by division among the processors of the plane-wave components for each state. The code is also optimized for RISC (reduced instruction set computer) architectures, such as those found on most parallel computers, making full use of BLAS (basic linear algebra subprograms) wherever possible. Scaling results are presented for systems of up to 686 silicon atoms and 343 palladium atoms per unit cell, running on up to 512 processors on a CRAY T3E parallel supercomputer.

  10. Exploiting Thread Parallelism for Ocean Modeling on Cray XC Supercomputers

    Energy Technology Data Exchange (ETDEWEB)

    Sarje, Abhinav [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Jacobsen, Douglas W. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Williams, Samuel W. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Ringler, Todd [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Oliker, Leonid [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)

    2016-05-01

    The incorporation of increasing core counts in modern processors used to build state-of-the-art supercomputers is driving application development towards exploitation of thread parallelism, in addition to distributed memory parallelism, with the goal of delivering efficient high-performance codes. In this work we describe the exploitation of threading and our experiences with it with respect to a real-world ocean modeling application code, MPAS-Ocean. We present detailed performance analysis and comparisons of various approaches and configurations for threading on the Cray XC series supercomputers.

  11. Development of massively parallel quantum chemistry program SMASH

    Energy Technology Data Exchange (ETDEWEB)

    Ishimura, Kazuya [Department of Theoretical and Computational Molecular Science, Institute for Molecular Science 38 Nishigo-Naka, Myodaiji, Okazaki, Aichi 444-8585 (Japan)

    2015-12-31

    A massively parallel program for quantum chemistry calculations SMASH was released under the Apache License 2.0 in September 2014. The SMASH program is written in the Fortran90/95 language with MPI and OpenMP standards for parallelization. Frequently used routines, such as one- and two-electron integral calculations, are modularized to make program developments simple. The speed-up of the B3LYP energy calculation for (C{sub 150}H{sub 30}){sub 2} with the cc-pVDZ basis set (4500 basis functions) was 50,499 on 98,304 cores of the K computer.

  12. A privacy-preserving parallel and homomorphic encryption scheme

    Directory of Open Access Journals (Sweden)

    Min Zhaoe

    2017-04-01

    Full Text Available In order to protect data privacy whilst allowing efficient access to data in multi-nodes cloud environments, a parallel homomorphic encryption (PHE scheme is proposed based on the additive homomorphism of the Paillier encryption algorithm. In this paper we propose a PHE algorithm, in which plaintext is divided into several blocks and blocks are encrypted with a parallel mode. Experiment results demonstrate that the encryption algorithm can reach a speed-up ratio at about 7.1 in the MapReduce environment with 16 cores and 4 nodes.

  13. Development of massively parallel quantum chemistry program SMASH

    International Nuclear Information System (INIS)

    Ishimura, Kazuya

    2015-01-01

    A massively parallel program for quantum chemistry calculations SMASH was released under the Apache License 2.0 in September 2014. The SMASH program is written in the Fortran90/95 language with MPI and OpenMP standards for parallelization. Frequently used routines, such as one- and two-electron integral calculations, are modularized to make program developments simple. The speed-up of the B3LYP energy calculation for (C 150 H 30 ) 2 with the cc-pVDZ basis set (4500 basis functions) was 50,499 on 98,304 cores of the K computer

  14. The boat hull model : adapting the roofline model to enable performance prediction for parallel computing

    NARCIS (Netherlands)

    Nugteren, C.; Corporaal, H.

    2012-01-01

    Multi-core and many-core were already major trends for the past six years, and are expected to continue for the next decades. With these trends of parallel computing, it becomes increasingly difficult to decide on which architecture to run a given application. In this work, we use an algorithm

  15. The boat hull model : enabling performance prediction for parallel computing prior to code development

    NARCIS (Netherlands)

    Nugteren, C.; Corporaal, H.

    2012-01-01

    Multi-core and many-core were already major trends for the past six years and are expected to continue for the next decade. With these trends of parallel computing, it becomes increasingly difficult to decide on which processor to run a given application, mainly because the programming of these

  16. Parallel computing solution of Boltzmann neutron transport equation

    International Nuclear Information System (INIS)

    Ansah-Narh, T.

    2010-01-01

    The focus of the research was on developing parallel computing algorithm for solving Eigen-values of the Boltzmam Neutron Transport Equation (BNTE) in a slab geometry using multi-grid approach. In response to the problem of slow execution of serial computing when solving large problems, such as BNTE, the study was focused on the design of parallel computing systems which was an evolution of serial computing that used multiple processing elements simultaneously to solve complex physical and mathematical problems. Finite element method (FEM) was used for the spatial discretization scheme, while angular discretization was accomplished by expanding the angular dependence in terms of Legendre polynomials. The eigenvalues representing the multiplication factors in the BNTE were determined by the power method. MATLAB Compiler Version 4.1 (R2009a) was used to compile the MATLAB codes of BNTE. The implemented parallel algorithms were enabled with matlabpool, a Parallel Computing Toolbox function. The option UseParallel was set to 'always' and the default value of the option was 'never'. When those conditions held, the solvers computed estimated gradients in parallel. The parallel computing system was used to handle all the bottlenecks in the matrix generated from the finite element scheme and each domain of the power method generated. The parallel algorithm was implemented on a Symmetric Multi Processor (SMP) cluster machine, which had Intel 32 bit quad-core x 86 processors. Convergence rates and timings for the algorithm on the SMP cluster machine were obtained. Numerical experiments indicated the designed parallel algorithm could reach perfect speedup and had good stability and scalability. (au)

  17. Performing an allreduce operation on a plurality of compute nodes of a parallel computer

    Science.gov (United States)

    Faraj, Ahmad [Rochester, MN

    2012-04-17

    Methods, apparatus, and products are disclosed for performing an allreduce operation on a plurality of compute nodes of a parallel computer. Each compute node includes at least two processing cores. Each processing core has contribution data for the allreduce operation. Performing an allreduce operation on a plurality of compute nodes of a parallel computer includes: establishing one or more logical rings among the compute nodes, each logical ring including at least one processing core from each compute node; performing, for each logical ring, a global allreduce operation using the contribution data for the processing cores included in that logical ring, yielding a global allreduce result for each processing core included in that logical ring; and performing, for each compute node, a local allreduce operation using the global allreduce results for each processing core on that compute node.

  18. Is Monte Carlo embarrassingly parallel?

    Energy Technology Data Exchange (ETDEWEB)

    Hoogenboom, J. E. [Delft Univ. of Technology, Mekelweg 15, 2629 JB Delft (Netherlands); Delft Nuclear Consultancy, IJsselzoom 2, 2902 LB Capelle aan den IJssel (Netherlands)

    2012-07-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  19. Is Monte Carlo embarrassingly parallel?

    International Nuclear Information System (INIS)

    Hoogenboom, J. E.

    2012-01-01

    Monte Carlo is often stated as being embarrassingly parallel. However, running a Monte Carlo calculation, especially a reactor criticality calculation, in parallel using tens of processors shows a serious limitation in speedup and the execution time may even increase beyond a certain number of processors. In this paper the main causes of the loss of efficiency when using many processors are analyzed using a simple Monte Carlo program for criticality. The basic mechanism for parallel execution is MPI. One of the bottlenecks turn out to be the rendez-vous points in the parallel calculation used for synchronization and exchange of data between processors. This happens at least at the end of each cycle for fission source generation in order to collect the full fission source distribution for the next cycle and to estimate the effective multiplication factor, which is not only part of the requested results, but also input to the next cycle for population control. Basic improvements to overcome this limitation are suggested and tested. Also other time losses in the parallel calculation are identified. Moreover, the threading mechanism, which allows the parallel execution of tasks based on shared memory using OpenMP, is analyzed in detail. Recommendations are given to get the maximum efficiency out of a parallel Monte Carlo calculation. (authors)

  20. Parallel integer sorting with medium and fine-scale parallelism

    Science.gov (United States)

    Dagum, Leonardo

    1993-01-01

    Two new parallel integer sorting algorithms, queue-sort and barrel-sort, are presented and analyzed in detail. These algorithms do not have optimal parallel complexity, yet they show very good performance in practice. Queue-sort designed for fine-scale parallel architectures which allow the queueing of multiple messages to the same destination. Barrel-sort is designed for medium-scale parallel architectures with a high message passing overhead. The performance results from the implementation of queue-sort on a Connection Machine CM-2 and barrel-sort on a 128 processor iPSC/860 are given. The two implementations are found to be comparable in performance but not as good as a fully vectorized bucket sort on the Cray YMP.

  1. Template based parallel checkpointing in a massively parallel computer system

    Science.gov (United States)

    Archer, Charles Jens [Rochester, MN; Inglett, Todd Alan [Rochester, MN

    2009-01-13

    A method and apparatus for a template based parallel checkpoint save for a massively parallel super computer system using a parallel variation of the rsync protocol, and network broadcast. In preferred embodiments, the checkpoint data for each node is compared to a template checkpoint file that resides in the storage and that was previously produced. Embodiments herein greatly decrease the amount of data that must be transmitted and stored for faster checkpointing and increased efficiency of the computer system. Embodiments are directed to a parallel computer system with nodes arranged in a cluster with a high speed interconnect that can perform broadcast communication. The checkpoint contains a set of actual small data blocks with their corresponding checksums from all nodes in the system. The data blocks may be compressed using conventional non-lossy data compression algorithms to further reduce the overall checkpoint size.

  2. Parallel education: what is it?

    OpenAIRE

    Amos, Michelle Peta

    2017-01-01

    In the history of education it has long been discussed that single-sex and coeducation are the two models of education present in schools. With the introduction of parallel schools over the last 15 years, there has been very little research into this 'new model'. Many people do not understand what it means for a school to be parallel or they confuse a parallel model with co-education, due to the presence of both boys and girls within the one institution. Therefore, the main obj...

  3. Balanced, parallel operation of flashlamps

    International Nuclear Information System (INIS)

    Carder, B.M.; Merritt, B.T.

    1979-01-01

    A new energy store, the Compensated Pulsed Alternator (CPA), promises to be a cost effective substitute for capacitors to drive flashlamps that pump large Nd:glass lasers. Because the CPA is large and discrete, it will be necessary that it drive many parallel flashlamp circuits, presenting a problem in equal current distribution. Current division to +- 20% between parallel flashlamps has been achieved, but this is marginal for laser pumping. A method is presented here that provides equal current sharing to about 1%, and it includes fused protection against short circuit faults. The method was tested with eight parallel circuits, including both open-circuit and short-circuit fault tests

  4. Animal MRI Core

    Data.gov (United States)

    Federal Laboratory Consortium — The Animal Magnetic Resonance Imaging (MRI) Core develops and optimizes MRI methods for cardiovascular imaging of mice and rats. The Core provides imaging expertise,...

  5. On efficiency of fire simulation realization: parallelization with greater number of computational meshes

    Science.gov (United States)

    Valasek, Lukas; Glasa, Jan

    2017-12-01

    Current fire simulation systems are capable to utilize advantages of high-performance computer (HPC) platforms available and to model fires efficiently in parallel. In this paper, efficiency of a corridor fire simulation on a HPC computer cluster is discussed. The parallel MPI version of Fire Dynamics Simulator is used for testing efficiency of selected strategies of allocation of computational resources of the cluster using a greater number of computational cores. Simulation results indicate that if the number of cores used is not equal to a multiple of the total number of cluster node cores there are allocation strategies which provide more efficient calculations.

  6. Implementation of parallel processing in the basf2 framework for Belle II

    International Nuclear Information System (INIS)

    Itoh, Ryosuke; Lee, Soohyung; Katayama, N; Mineo, S; Moll, A; Kuhr, T; Heck, M

    2012-01-01

    Recent PC servers are equipped with multi-core CPUs and it is desired to utilize the full processing power of them for the data analysis in large scale HEP experiments. A software framework basf2 is being developed for the use in the Belle II experiment, a new generation B-factory experiment at KEK, and the parallel event processing to utilize the multi-core CPUs is in its design for the use in the massive data production. The details of the implementation of event parallel processing in the basf2 framework are discussed with the report of preliminary performance study in the realistic use on a 32 core PC server.

  7. Greenland climate change

    DEFF Research Database (Denmark)

    Masson-Delmotte, Valérie; Swingedouw, Didier; Landais, Amaëlle

    2012-01-01

    Climate archives available from deep-sea and marine shelf sediments, glaciers, lakes and ice cores in and around Greenland allow us to place the current trends in regional climate, ice sheet dynamics, and land surface changes in a broader perspective. We show that during the last decade (2000s......), atmospheric and sea-surface temperatures are reaching levels last encountered millennia ago when northern high latitude summer insolation was higher due to a different orbital configuration. Concurrently, records from lake sediments in southern Greenland document major environmental and climatic conditions...... regional climate and ice sheet dynamics. The magnitude and rate of future changes in Greenland temperature, in response to increasing greenhouse gas emissions, may be faster than any past abrupt events occurring under interglacial conditions. Projections indicate that within one century Greenland may...

  8. Hierarchical approach to optimization of parallel matrix multiplication on large-scale platforms

    KAUST Repository

    Hasanov, Khalid

    2014-03-04

    © 2014, Springer Science+Business Media New York. Many state-of-the-art parallel algorithms, which are widely used in scientific applications executed on high-end computing systems, were designed in the twentieth century with relatively small-scale parallelism in mind. Indeed, while in 1990s a system with few hundred cores was considered a powerful supercomputer, modern top supercomputers have millions of cores. In this paper, we present a hierarchical approach to optimization of message-passing parallel algorithms for execution on large-scale distributed-memory systems. The idea is to reduce the communication cost by introducing hierarchy and hence more parallelism in the communication scheme. We apply this approach to SUMMA, the state-of-the-art parallel algorithm for matrix–matrix multiplication, and demonstrate both theoretically and experimentally that the modified Hierarchical SUMMA significantly improves the communication cost and the overall performance on large-scale platforms.

  9. Climate Change

    Science.gov (United States)

    Climate is the average weather in a place over a period of time. Climate change is major change in temperature, rainfall, snow, ... by natural factors or by human activities. Today climate changes are occurring at an increasingly rapid rate. ...

  10. Agriculture: Climate

    Science.gov (United States)

    Climate change affects agricultural producers because agriculture and fisheries depend on specific climate conditions. Temperature changes can cause crop planting dates to shift. Droughts and floods due to climate change may hinder farming practices.

  11. Discontinuous Galerkin Dynamical Core in HOMME

    Energy Technology Data Exchange (ETDEWEB)

    Nair, R. D. [Univ. of Colorado, Boulder, CO (United States); Tufo, Henry [Univ. of Colorado, Boulder, CO (United States)

    2012-08-14

    Atmospheric numerical modeling has been going through radical changes over the past decade. One major reason for this trend is due to the recent paradigm change in scientific computing , triggered by the arrival of petascale computing resources with core counts in the tens of thousands to hundreds of thousands range. Modern atmospheric modelers must adapt grid systems and numerical algorithms to facilitate an unprecedented levels of scalability on these modern highly parallel computer architectures. The numerical algorithms which can address these challenges should have the local properties such as high on-processor floating-point operation count to bytes moved and minimum parallel communication overhead.

  12. Workspace Analysis for Parallel Robot

    Directory of Open Access Journals (Sweden)

    Ying Sun

    2013-05-01

    Full Text Available As a completely new-type of robot, the parallel robot possesses a lot of advantages that the serial robot does not, such as high rigidity, great load-carrying capacity, small error, high precision, small self-weight/load ratio, good dynamic behavior and easy control, hence its range is extended in using domain. In order to find workspace of parallel mechanism, the numerical boundary-searching algorithm based on the reverse solution of kinematics and limitation of link length has been introduced. This paper analyses position workspace, orientation workspace of parallel robot of the six degrees of freedom. The result shows: It is a main means to increase and decrease its workspace to change the length of branch of parallel mechanism; The radius of the movement platform has no effect on the size of workspace, but will change position of workspace.

  13. "Feeling" Series and Parallel Resistances.

    Science.gov (United States)

    Morse, Robert A.

    1993-01-01

    Equipped with drinking straws and stirring straws, a teacher can help students understand how resistances in electric circuits combine in series and in parallel. Follow-up suggestions are provided. (ZWH)

  14. Parallel encoders for pixel detectors

    International Nuclear Information System (INIS)

    Nikityuk, N.M.

    1991-01-01

    A new method of fast encoding and determining the multiplicity and coordinates of fired pixels is described. A specific example construction of parallel encodes and MCC for n=49 and t=2 is given. 16 refs.; 6 figs.; 2 tabs

  15. Event monitoring of parallel computations

    Directory of Open Access Journals (Sweden)

    Gruzlikov Alexander M.

    2015-06-01

    Full Text Available The paper considers the monitoring of parallel computations for detection of abnormal events. It is assumed that computations are organized according to an event model, and monitoring is based on specific test sequences

  16. Parallelizing AT with MatlabMPI

    International Nuclear Information System (INIS)

    2011-01-01

    The Accelerator Toolbox (AT) is a high-level collection of tools and scripts specifically oriented toward solving problems dealing with computational accelerator physics. It is integrated into the MATLAB environment, which provides an accessible, intuitive interface for accelerator physicists, allowing researchers to focus the majority of their efforts on simulations and calculations, rather than programming and debugging difficulties. Efforts toward parallelization of AT have been put in place to upgrade its performance to modern standards of computing. We utilized the packages MatlabMPI and pMatlab, which were developed by MIT Lincoln Laboratory, to set up a message-passing environment that could be called within MATLAB, which set up the necessary pre-requisites for multithread processing capabilities. On local quad-core CPUs, we were able to demonstrate processor efficiencies of roughly 95% and speed increases of nearly 380%. By exploiting the efficacy of modern-day parallel computing, we were able to demonstrate incredibly efficient speed increments per processor in AT's beam-tracking functions. Extrapolating from prediction, we can expect to reduce week-long computation runtimes to less than 15 minutes. This is a huge performance improvement and has enormous implications for the future computing power of the accelerator physics group at SSRL. However, one of the downfalls of parringpass is its current lack of transparency; the pMatlab and MatlabMPI packages must first be well-understood by the user before the system can be configured to run the scripts. In addition, the instantiation of argument parameters requires internal modification of the source code. Thus, parringpass, cannot be directly run from the MATLAB command line, which detracts from its flexibility and user-friendliness. Future work in AT's parallelization will focus on development of external functions and scripts that can be called from within MATLAB and configured on multiple nodes, while

  17. WASCAL - West African Science Service Center on Climate Change and Adapted Land Use Regional Climate Simulations and Land-Atmosphere Simulations for West Africa at DKRZ and elsewhere

    Science.gov (United States)

    Hamann, Ilse; Arnault, Joel; Bliefernicht, Jan; Klein, Cornelia; Heinzeller, Dominikus; Kunstmann, Harald

    2014-05-01

    accompanied by the WASCAL Graduate Research Program on the West African Climate System. The GRP-WACS provides ten scholarships per year for West African PhD students with a duration of three years. Present and future WASCAL PhD students will constitute one important user group of the Linux cluster that will be installed at the Competence Center in Ouagadougou, Burkina Faso. Regional Land-Atmosphere Simulations A key research activity of the WASCAL Core Research Program is the analysis of interactions between the land surface and the atmosphere to investigate how land surface changes affect hydro-meteorological surface fluxes such as evapotranspiration. Since current land surface models of global and regional climate models neglect dominant lateral hydrological processes such as surface runoff, a novel land surface model is used, the NCAR Distributed Hydrological Modeling System (NDHMS). This model can be coupled to WRF (WRF-Hydro) to perform two-way coupled atmospheric-hydrological simulations for the watershed of interest. Hardware and network prerequisites include a HPC cluster, network switches, internal storage media, Internet connectivity of sufficient bandwidth. Competences needed are HPC, storage, and visualization systems optimized for climate research, parallelization and optimization of climate models and workflows, efficient management of highest data volumes.

  18. Writing parallel programs that work

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    Serial algorithms typically run inefficiently on parallel machines. This may sound like an obvious statement, but it is the root cause of why parallel programming is considered to be difficult. The current state of the computer industry is still that almost all programs in existence are serial. This talk will describe the techniques used in the Intel Parallel Studio to provide a developer with the tools necessary to understand the behaviors and limitations of the existing serial programs. Once the limitations are known the developer can refactor the algorithms and reanalyze the resulting programs with the tools in the Intel Parallel Studio to create parallel programs that work. About the speaker Paul Petersen is a Sr. Principal Engineer in the Software and Solutions Group (SSG) at Intel. He received a Ph.D. degree in Computer Science from the University of Illinois in 1993. After UIUC, he was employed at Kuck and Associates, Inc. (KAI) working on auto-parallelizing compiler (KAP), and was involved in th...

  19. Exploiting Symmetry on Parallel Architectures.

    Science.gov (United States)

    Stiller, Lewis Benjamin

    1995-01-01

    This thesis describes techniques for the design of parallel programs that solve well-structured problems with inherent symmetry. Part I demonstrates the reduction of such problems to generalized matrix multiplication by a group-equivariant matrix. Fast techniques for this multiplication are described, including factorization, orbit decomposition, and Fourier transforms over finite groups. Our algorithms entail interaction between two symmetry groups: one arising at the software level from the problem's symmetry and the other arising at the hardware level from the processors' communication network. Part II illustrates the applicability of our symmetry -exploitation techniques by presenting a series of case studies of the design and implementation of parallel programs. First, a parallel program that solves chess endgames by factorization of an associated dihedral group-equivariant matrix is described. This code runs faster than previous serial programs, and discovered it a number of results. Second, parallel algorithms for Fourier transforms for finite groups are developed, and preliminary parallel implementations for group transforms of dihedral and of symmetric groups are described. Applications in learning, vision, pattern recognition, and statistics are proposed. Third, parallel implementations solving several computational science problems are described, including the direct n-body problem, convolutions arising from molecular biology, and some communication primitives such as broadcast and reduce. Some of our implementations ran orders of magnitude faster than previous techniques, and were used in the investigation of various physical phenomena.

  20. Parallel algorithms for continuum dynamics

    International Nuclear Information System (INIS)

    Hicks, D.L.; Liebrock, L.M.

    1987-01-01

    Simply porting existing parallel programs to a new parallel processor may not achieve the full speedup possible; to achieve the maximum efficiency may require redesigning the parallel algorithms for the specific architecture. The authors discuss here parallel algorithms that were developed first for the HEP processor and then ported to the CRAY X-MP/4, the ELXSI/10, and the Intel iPSC/32. Focus is mainly on the most recent parallel processing results produced, i.e., those on the Intel Hypercube. The applications are simulations of continuum dynamics in which the momentum and stress gradients are important. Examples of these are inertial confinement fusion experiments, severe breaks in the coolant system of a reactor, weapons physics, shock-wave physics. Speedup efficiencies on the Intel iPSC Hypercube are very sensitive to the ratio of communication to computation. Great care must be taken in designing algorithms for this machine to avoid global communication. This is much more critical on the iPSC than it was on the three previous parallel processors

  1. Parallel Distributed Processing theory in the age of deep networks

    OpenAIRE

    Bowers, Jeffrey

    2017-01-01

    Parallel Distributed Processing (PDP) models in psychology are the precursors of deep networks used in computer science. However, only PDP models are associated with two core psychological claims, namely, that all knowledge is coded in a distributed format, and cognition is mediated by non-symbolic computations. These claims have long been debated within cognitive science, and recent work with deep networks speaks to this debate. Specifically, single-unit recordings show that deep networks le...

  2. Large-Scale, Parallel, Multi-Sensor Atmospheric Data Fusion Using Cloud Computing

    Science.gov (United States)

    Wilson, B. D.; Manipon, G.; Hua, H.; Fetzer, E. J.

    2013-12-01

    NASA's Earth Observing System (EOS) is an ambitious facility for studying global climate change. The mandate now is to combine measurements from the instruments on the 'A-Train' platforms (AIRS, AMSR-E, MODIS, MISR, MLS, and CloudSat) and other Earth probes to enable large-scale studies of climate change over decades. Moving to multi-sensor, long-duration analyses of important climate variables presents serious challenges for large-scale data mining and fusion. For example, one might want to compare temperature and water vapor retrievals from one instrument (AIRS) to another (MODIS), and to a model (MERRA), stratify the comparisons using a classification of the 'cloud scenes' from CloudSat, and repeat the entire analysis over 10 years of data. To efficiently assemble such datasets, we are utilizing Elastic Computing in the Cloud and parallel map/reduce-based algorithms. However, these problems are Data Intensive computing so the data transfer times and storage costs (for caching) are key issues. SciReduce is a Hadoop-like parallel analysis system, programmed in parallel python, that is designed from the ground up for Earth science. SciReduce executes inside VMWare images and scales to any number of nodes in the Cloud. Unlike Hadoop, SciReduce operates on bundles of named numeric arrays, which can be passed in memory or serialized to disk in netCDF4 or HDF5. Figure 1 shows the architecture of the full computational system, with SciReduce at the core. Multi-year datasets are automatically 'sharded' by time and space across a cluster of nodes so that years of data (millions of files) can be processed in a massively parallel way. Input variables (arrays) are pulled on-demand into the Cloud using OPeNDAP URLs or other subsetting services, thereby minimizing the size of the cached input and intermediate datasets. We are using SciReduce to automate the production of multiple versions of a ten-year A-Train water vapor climatology under a NASA MEASURES grant. We will

  3. Fundamental Parallel Algorithms for Private-Cache Chip Multiprocessors

    DEFF Research Database (Denmark)

    Arge, Lars Allan; Goodrich, Michael T.; Nelson, Michael

    2008-01-01

    about the way cores are interconnected, for we assume that all inter-processor communication occurs through the memory hierarchy. We study several fundamental problems, including prefix sums, selection, and sorting, which often form the building blocks of other parallel algorithms. Indeed, we present...... two sorting algorithms, a distribution sort and a mergesort. Our algorithms are asymptotically optimal in terms of parallel cache accesses and space complexity under reasonable assumptions about the relationships between the number of processors, the size of memory, and the size of cache blocks....... In addition, we study sorting lower bounds in a computational model, which we call the parallel external-memory (PEM) model, that formalizes the essential properties of our algorithms for private-cache CMPs....

  4. Engineering-Based Thermal CFD Simulations on Massive Parallel Systems

    KAUST Repository

    Frisch, Jérôme

    2015-05-22

    The development of parallel Computational Fluid Dynamics (CFD) codes is a challenging task that entails efficient parallelization concepts and strategies in order to achieve good scalability values when running those codes on modern supercomputers with several thousands to millions of cores. In this paper, we present a hierarchical data structure for massive parallel computations that supports the coupling of a Navier–Stokes-based fluid flow code with the Boussinesq approximation in order to address complex thermal scenarios for energy-related assessments. The newly designed data structure is specifically designed with the idea of interactive data exploration and visualization during runtime of the simulation code; a major shortcoming of traditional high-performance computing (HPC) simulation codes. We further show and discuss speed-up values obtained on one of Germany’s top-ranked supercomputers with up to 140,000 processes and present simulation results for different engineering-based thermal problems.

  5. Parallel halftoning technique using dot diffusion optimization

    Science.gov (United States)

    Molina-Garcia, Javier; Ponomaryov, Volodymyr I.; Reyes-Reyes, Rogelio; Cruz-Ramos, Clara

    2017-05-01

    In this paper, a novel approach for halftone images is proposed and implemented for images that are obtained by the Dot Diffusion (DD) method. Designed technique is based on an optimization of the so-called class matrix used in DD algorithm and it consists of generation new versions of class matrix, which has no baron and near-baron in order to minimize inconsistencies during the distribution of the error. Proposed class matrix has different properties and each is designed for two different applications: applications where the inverse-halftoning is necessary, and applications where this method is not required. The proposed method has been implemented in GPU (NVIDIA GeForce GTX 750 Ti), multicore processors (AMD FX(tm)-6300 Six-Core Processor and in Intel core i5-4200U), using CUDA and OpenCV over a PC with linux. Experimental results have shown that novel framework generates a good quality of the halftone images and the inverse halftone images obtained. The simulation results using parallel architectures have demonstrated the efficiency of the novel technique when it is implemented in real-time processing.

  6. Acceleration and parallelization calculation of EFEN-SP_3 method

    International Nuclear Information System (INIS)

    Yang Wen; Zheng Youqi; Wu Hongchun; Cao Liangzhi; Li Yunzhao

    2013-01-01

    Due to the fact that the exponential function expansion nodal-SP_3 (EFEN-SP_3) method needs further improvement in computational efficiency to routinely carry out PWR whole core pin-by-pin calculation, the coarse mesh acceleration and spatial parallelization were investigated in this paper. The coarse mesh acceleration was built by considering discontinuity factor on each coarse mesh interface and preserving neutron balance within each coarse mesh in space, angle and energy. The spatial parallelization based on MPI was implemented by guaranteeing load balancing and minimizing communications cost to fully take advantage of the modern computing and storage abilities. Numerical results based on a commercial nuclear power reactor demonstrate an speedup ratio of about 40 for the coarse mesh acceleration and a parallel efficiency of higher than 60% with 40 CPUs for the spatial parallelization. With these two improvements, the EFEN code can complete a PWR whole core pin-by-pin calculation with 289 × 289 × 218 meshes and 4 energy groups within 100 s by using 48 CPUs (2.40 GHz frequency). (authors)

  7. User-friendly parallelization of GAUDI applications with Python

    International Nuclear Information System (INIS)

    Mato, Pere; Smith, Eoin

    2010-01-01

    GAUDI is a software framework in C++ used to build event data processing applications using a set of standard components with well-defined interfaces. Simulation, high-level trigger, reconstruction, and analysis programs used by several experiments are developed using GAUDI. These applications can be configured and driven by simple Python scripts. Given the fact that a considerable amount of existing software has been developed using serial methodology, and has existed in some cases for many years, implementation of parallelisation techniques at the framework level may offer a way of exploiting current multi-core technologies to maximize performance and reduce latencies without re-writing thousands/millions of lines of code. In the solution we have developed, the parallelization techniques are introduced to the high level Python scripts which configure and drive the applications, such that the core C++ application code requires no modification, and that end users need make only minimal changes to their scripts. The developed solution leverages from existing generic Python modules that support parallel processing. Naturally, the parallel version of a given program should produce results consistent with its serial execution. The evaluation of several prototypes incorporating various parallelization techniques are presented and discussed.

  8. User-friendly parallelization of GAUDI applications with Python

    Energy Technology Data Exchange (ETDEWEB)

    Mato, Pere; Smith, Eoin, E-mail: pere.mato@cern.c [PH Department, CERN, 1211 Geneva 23 (Switzerland)

    2010-04-01

    GAUDI is a software framework in C++ used to build event data processing applications using a set of standard components with well-defined interfaces. Simulation, high-level trigger, reconstruction, and analysis programs used by several experiments are developed using GAUDI. These applications can be configured and driven by simple Python scripts. Given the fact that a considerable amount of existing software has been developed using serial methodology, and has existed in some cases for many years, implementation of parallelisation techniques at the framework level may offer a way of exploiting current multi-core technologies to maximize performance and reduce latencies without re-writing thousands/millions of lines of code. In the solution we have developed, the parallelization techniques are introduced to the high level Python scripts which configure and drive the applications, such that the core C++ application code requires no modification, and that end users need make only minimal changes to their scripts. The developed solution leverages from existing generic Python modules that support parallel processing. Naturally, the parallel version of a given program should produce results consistent with its serial execution. The evaluation of several prototypes incorporating various parallelization techniques are presented and discussed.

  9. Endpoint-based parallel data processing in a parallel active messaging interface of a parallel computer

    Science.gov (United States)

    Archer, Charles J.; Blocksome, Michael A.; Ratterman, Joseph D.; Smith, Brian E.

    2014-08-12

    Endpoint-based parallel data processing in a parallel active messaging interface (`PAMI`) of a parallel computer, the PAMI composed of data communications endpoints, each endpoint including a specification of data communications parameters for a thread of execution on a compute node, including specifications of a client, a context, and a task, the compute nodes coupled for data communications through the PAMI, including establishing a data communications geometry, the geometry specifying, for tasks representing processes of execution of the parallel application, a set of endpoints that are used in collective operations of the PAMI including a plurality of endpoints for one of the tasks; receiving in endpoints of the geometry an instruction for a collective operation; and executing the instruction for a collective operation through the endpoints in dependence upon the geometry, including dividing data communications operations among the plurality of endpoints for one of the tasks.

  10. Parallel Implicit Algorithms for CFD

    Science.gov (United States)

    Keyes, David E.

    1998-01-01

    The main goal of this project was efficient distributed parallel and workstation cluster implementations of Newton-Krylov-Schwarz (NKS) solvers for implicit Computational Fluid Dynamics (CFD.) "Newton" refers to a quadratically convergent nonlinear iteration using gradient information based on the true residual, "Krylov" to an inner linear iteration that accesses the Jacobian matrix only through highly parallelizable sparse matrix-vector products, and "Schwarz" to a domain decomposition form of preconditioning the inner Krylov iterations with primarily neighbor-only exchange of data between the processors. Prior experience has established that Newton-Krylov methods are competitive solvers in the CFD context and that Krylov-Schwarz methods port well to distributed memory computers. The combination of the techniques into Newton-Krylov-Schwarz was implemented on 2D and 3D unstructured Euler codes on the parallel testbeds that used to be at LaRC and on several other parallel computers operated by other agencies or made available by the vendors. Early implementations were made directly in Massively Parallel Integration (MPI) with parallel solvers we adapted from legacy NASA codes and enhanced for full NKS functionality. Later implementations were made in the framework of the PETSC library from Argonne National Laboratory, which now includes pseudo-transient continuation Newton-Krylov-Schwarz solver capability (as a result of demands we made upon PETSC during our early porting experiences). A secondary project pursued with funding from this contract was parallel implicit solvers in acoustics, specifically in the Helmholtz formulation. A 2D acoustic inverse problem has been solved in parallel within the PETSC framework.

  11. Second derivative parallel block backward differentiation type ...

    African Journals Online (AJOL)

    Second derivative parallel block backward differentiation type formulas for Stiff ODEs. ... Log in or Register to get access to full text downloads. ... and the methods are inherently parallel and can be distributed over parallel processors. They are ...

  12. A Parallel Approach to Fractal Image Compression

    OpenAIRE

    Lubomir Dedera

    2004-01-01

    The paper deals with a parallel approach to coding and decoding algorithms in fractal image compressionand presents experimental results comparing sequential and parallel algorithms from the point of view of achieved bothcoding and decoding time and effectiveness of parallelization.

  13. EEA core set of indicators. Guide

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2005-07-01

    This guide provides information on the quality of the 37 indicators in the EEA core set. Its primary role is to support improved implementation of the core set in the EEA, European topic centres and the European environment information and observation network (Eionet). In parallel, it is aimed at helping users outside the EEA/Eionet system make best use of the indicators in their own work. It is hoped that the guide will promote cooperation on improving indicator methodologies and data quality as part of the wider process to streamline and improve environmental reporting in the European Union and beyond. (au)

  14. Parallel fabrication of macroporous scaffolds.

    Science.gov (United States)

    Dobos, Andrew; Grandhi, Taraka Sai Pavan; Godeshala, Sudhakar; Meldrum, Deirdre R; Rege, Kaushal

    2018-07-01

    Scaffolds generated from naturally occurring and synthetic polymers have been investigated in several applications because of their biocompatibility and tunable chemo-mechanical properties. Existing methods for generation of 3D polymeric scaffolds typically cannot be parallelized, suffer from low throughputs, and do not allow for quick and easy removal of the fragile structures that are formed. Current molds used in hydrogel and scaffold fabrication using solvent casting and porogen leaching are often single-use and do not facilitate 3D scaffold formation in parallel. Here, we describe a simple device and related approaches for the parallel fabrication of macroporous scaffolds. This approach was employed for the generation of macroporous and non-macroporous materials in parallel, in higher throughput and allowed for easy retrieval of these 3D scaffolds once formed. In addition, macroporous scaffolds with interconnected as well as non-interconnected pores were generated, and the versatility of this approach was employed for the generation of 3D scaffolds from diverse materials including an aminoglycoside-derived cationic hydrogel ("Amikagel"), poly(lactic-co-glycolic acid) or PLGA, and collagen. Macroporous scaffolds generated using the device were investigated for plasmid DNA binding and cell loading, indicating the use of this approach for developing materials for different applications in biotechnology. Our results demonstrate that the device-based approach is a simple technology for generating scaffolds in parallel, which can enhance the toolbox of current fabrication techniques. © 2018 Wiley Periodicals, Inc.

  15. Evaluating parallel optimization on transputers

    Directory of Open Access Journals (Sweden)

    A.G. Chalmers

    2003-12-01

    Full Text Available The faster processing power of modern computers and the development of efficient algorithms have made it possible for operations researchers to tackle a much wider range of problems than ever before. Further improvements in processing speed can be achieved utilising relatively inexpensive transputers to process components of an algorithm in parallel. The Davidon-Fletcher-Powell method is one of the most successful and widely used optimisation algorithms for unconstrained problems. This paper examines the algorithm and identifies the components that can be processed in parallel. The results of some experiments with these components are presented which indicates under what conditions parallel processing with an inexpensive configuration is likely to be faster than the traditional sequential implementations. The performance of the whole algorithm with its parallel components is then compared with the original sequential algorithm. The implementation serves to illustrate the practicalities of speeding up typical OR algorithms in terms of difficulty, effort and cost. The results give an indication of the savings in time a given parallel implementation can be expected to yield.

  16. Pattern-Driven Automatic Parallelization

    Directory of Open Access Journals (Sweden)

    Christoph W. Kessler

    1996-01-01

    Full Text Available This article describes a knowledge-based system for automatic parallelization of a wide class of sequential numerical codes operating on vectors and dense matrices, and for execution on distributed memory message-passing multiprocessors. Its main feature is a fast and powerful pattern recognition tool that locally identifies frequently occurring computations and programming concepts in the source code. This tool also works for dusty deck codes that have been "encrypted" by former machine-specific code transformations. Successful pattern recognition guides sophisticated code transformations including local algorithm replacement such that the parallelized code need not emerge from the sequential program structure by just parallelizing the loops. It allows access to an expert's knowledge on useful parallel algorithms, available machine-specific library routines, and powerful program transformations. The partially restored program semantics also supports local array alignment, distribution, and redistribution, and allows for faster and more exact prediction of the performance of the parallelized target code than is usually possible.

  17. The Destabilization of Protected Soil Organic Carbon Following Experimental Drought at the Pore and Core scale

    Science.gov (United States)

    Smith, A. P.; Bond-Lamberty, B. P.; Tfaily, M. M.; Todd-Brown, K. E.; Bailey, V. L.

    2015-12-01

    The movement of water and solutes through the pore matrix controls the distribution and transformation of carbon (C) in soils. Thus, a change in the hydrologic connectivity, such as increased saturation, disturbance or drought, may alter C mineralization and greenhouse gas (GHG) fluxes to the atmosphere. While these processes occur at the pore scale, they are often investigated at coarser scale. This project investigates pore- and core-scale soil C dynamics with varying hydrologic factors (simulated precipitation, groundwater-led saturation, and drought) to assess how climate-change induced shifts in hydrologic connectivity influences the destabilization of protected C in soils. Surface soil cores (0-15 cm depth) were collected from the Disney Wilderness Preserve, Florida, USA where water dynamics, particularly water table rise and fall, appear to exert a strong control on the emissions of GHGs and the persistence of soil organic matter in these soils. We measured CO2 and CH4 from soils allowed to freely imbibe water from below to a steady state starting from either field moist conditions or following experimental drought. Parallel treatments included the addition of similar quantities of water from above to simulate precipitation. Overall respiration increased in soil cores subjected to drought compared to field moist cores independent of wetting type. Cumulative CH4 production was higher in drought-induced soils, especially in the soils subjected to experimental groundwater-led saturation. Overall, the more C (from CO2 and CH4) was lost in drought-induced soils compared to field moist cores. Our results indicate that future drought events could have profound effects on the destabilization of protected C, especially in groundwater-fed soils. Our next steps focus on how to accurately capture drought-induced C destabilization mechanisms in earth system models.

  18. Running climate model on a commercial cloud computing environment: A case study using Community Earth System Model (CESM) on Amazon AWS

    Science.gov (United States)

    Chen, Xiuhong; Huang, Xianglei; Jiao, Chaoyi; Flanner, Mark G.; Raeker, Todd; Palen, Brock

    2017-01-01

    The suites of numerical models used for simulating climate of our planet are usually run on dedicated high-performance computing (HPC) resources. This study investigates an alternative to the usual approach, i.e. carrying out climate model simulations on commercially available cloud computing environment. We test the performance and reliability of running the CESM (Community Earth System Model), a flagship climate model in the United States developed by the National Center for Atmospheric Research (NCAR), on Amazon Web Service (AWS) EC2, the cloud computing environment by Amazon.com, Inc. StarCluster is used to create virtual computing cluster on the AWS EC2 for the CESM simulations. The wall-clock time for one year of CESM simulation on the AWS EC2 virtual cluster is comparable to the time spent for the same simulation on a local dedicated high-performance computing cluster with InfiniBand connections. The CESM simulation can be efficiently scaled with the number of CPU cores on the AWS EC2 virtual cluster environment up to 64 cores. For the standard configuration of the CESM at a spatial resolution of 1.9° latitude by 2.5° longitude, increasing the number of cores from 16 to 64 reduces the wall-clock running time by more than 50% and the scaling is nearly linear. Beyond 64 cores, the communication latency starts to outweigh the benefit of distributed computing and the parallel speedup becomes nearly unchanged.

  19. k-core covers and the core

    NARCIS (Netherlands)

    Sanchez-Rodriguez, E.; Borm, Peter; Estevez-Fernandez, A.; Fiestras-Janeiro, G.; Mosquera, M.A.

    This paper extends the notion of individual minimal rights for a transferable utility game (TU-game) to coalitional minimal rights using minimal balanced families of a specific type, thus defining a corresponding minimal rights game. It is shown that the core of a TU-game coincides with the core of

  20. GPU based numerical simulation of core shooting process

    Directory of Open Access Journals (Sweden)

    Yi-zhong Zhang

    2017-11-01

    Full Text Available Core shooting process is the most widely used technique to make sand cores and it plays an important role in the quality of sand cores. Although numerical simulation can hopefully optimize the core shooting process, research on numerical simulation of the core shooting process is very limited. Based on a two-fluid model (TFM and a kinetic-friction constitutive correlation, a program for 3D numerical simulation of the core shooting process has been developed and achieved good agreements with in-situ experiments. To match the needs of engineering applications, a graphics processing unit (GPU has also been used to improve the calculation efficiency. The parallel algorithm based on the Compute Unified Device Architecture (CUDA platform can significantly decrease computing time by multi-threaded GPU. In this work, the program accelerated by CUDA parallelization method was developed and the accuracy of the calculations was ensured by comparing with in-situ experimental results photographed by a high-speed camera. The design and optimization of the parallel algorithm were discussed. The simulation result of a sand core test-piece indicated the improvement of the calculation efficiency by GPU. The developed program has also been validated by in-situ experiments with a transparent core-box, a high-speed camera, and a pressure measuring system. The computing time of the parallel program was reduced by nearly 95% while the simulation result was still quite consistent with experimental data. The GPU parallelization method can successfully solve the problem of low computational efficiency of the 3D sand shooting simulation program, and thus the developed GPU program is appropriate for engineering applications.

  1. Examination of Speed Contribution of Parallelization for Several Fingerprint Pre-Processing Algorithms

    Directory of Open Access Journals (Sweden)

    GORGUNOGLU, S.

    2014-05-01

    Full Text Available In analysis of minutiae based fingerprint systems, fingerprints needs to be pre-processed. The pre-processing is carried out to enhance the quality of the fingerprint and to obtain more accurate minutiae points. Reducing the pre-processing time is important for identification and verification in real time systems and especially for databases holding large fingerprints information. Parallel processing and parallel CPU computing can be considered as distribution of processes over multi core processor. This is done by using parallel programming techniques. Reducing the execution time is the main objective in parallel processing. In this study, pre-processing of minutiae based fingerprint system is implemented by parallel processing on multi core computers using OpenMP and on graphics processor using CUDA to improve execution time. The execution times and speedup ratios are compared with the one that of single core processor. The results show that by using parallel processing, execution time is substantially improved. The improvement ratios obtained for different pre-processing algorithms allowed us to make suggestions on the more suitable approaches for parallelization.

  2. PARALLEL IMPLEMENTATION OF MORPHOLOGICAL PROFILE BASED SPECTRAL-SPATIAL CLASSIFICATION SCHEME FOR HYPERSPECTRAL IMAGERY

    Directory of Open Access Journals (Sweden)

    B. Kumar

    2016-06-01

    Full Text Available Extended morphological profile (EMP is a good technique for extracting spectral-spatial information from the images but large size of hyperspectral images is an important concern for creating EMPs. However, with the availability of modern multi-core processors and commodity parallel processing systems like graphics processing units (GPUs at desktop level, parallel computing provides a viable option to significantly accelerate execution of such computations. In this paper, parallel implementation of an EMP based spectralspatial classification method for hyperspectral imagery is presented. The parallel implementation is done both on multi-core CPU and GPU. The impact of parallelization on speed up and classification accuracy is analyzed. For GPU, the implementation is done in compute unified device architecture (CUDA C. The experiments are carried out on two well-known hyperspectral images. It is observed from the experimental results that GPU implementation provides a speed up of about 7 times, while parallel implementation on multi-core CPU resulted in speed up of about 3 times. It is also observed that parallel implementation has no adverse impact on the classification accuracy.

  3. Parallel artificial liquid membrane extraction

    DEFF Research Database (Denmark)

    Gjelstad, Astrid; Rasmussen, Knut Einar; Parmer, Marthe Petrine

    2013-01-01

    This paper reports development of a new approach towards analytical liquid-liquid-liquid membrane extraction termed parallel artificial liquid membrane extraction. A donor plate and acceptor plate create a sandwich, in which each sample (human plasma) and acceptor solution is separated by an arti......This paper reports development of a new approach towards analytical liquid-liquid-liquid membrane extraction termed parallel artificial liquid membrane extraction. A donor plate and acceptor plate create a sandwich, in which each sample (human plasma) and acceptor solution is separated...... by an artificial liquid membrane. Parallel artificial liquid membrane extraction is a modification of hollow-fiber liquid-phase microextraction, where the hollow fibers are replaced by flat membranes in a 96-well plate format....

  4. Parallel algorithms for mapping pipelined and parallel computations

    Science.gov (United States)

    Nicol, David M.

    1988-01-01

    Many computational problems in image processing, signal processing, and scientific computing are naturally structured for either pipelined or parallel computation. When mapping such problems onto a parallel architecture it is often necessary to aggregate an obvious problem decomposition. Even in this context the general mapping problem is known to be computationally intractable, but recent advances have been made in identifying classes of problems and architectures for which optimal solutions can be found in polynomial time. Among these, the mapping of pipelined or parallel computations onto linear array, shared memory, and host-satellite systems figures prominently. This paper extends that work first by showing how to improve existing serial mapping algorithms. These improvements have significantly lower time and space complexities: in one case a published O(nm sup 3) time algorithm for mapping m modules onto n processors is reduced to an O(nm log m) time complexity, and its space requirements reduced from O(nm sup 2) to O(m). Run time complexity is further reduced with parallel mapping algorithms based on these improvements, which run on the architecture for which they create the mappings.

  5. Parallel family trees for transfer matrices in the Potts model

    Science.gov (United States)

    Navarro, Cristobal A.; Canfora, Fabrizio; Hitschfeld, Nancy; Navarro, Gonzalo

    2015-02-01

    The computational cost of transfer matrix methods for the Potts model is related to the question in how many ways can two layers of a lattice be connected? Answering the question leads to the generation of a combinatorial set of lattice configurations. This set defines the configuration space of the problem, and the smaller it is, the faster the transfer matrix can be computed. The configuration space of generic (q , v) transfer matrix methods for strips is in the order of the Catalan numbers, which grows asymptotically as O(4m) where m is the width of the strip. Other transfer matrix methods with a smaller configuration space indeed exist but they make assumptions on the temperature, number of spin states, or restrict the structure of the lattice. In this paper we propose a parallel algorithm that uses a sub-Catalan configuration space of O(3m) to build the generic (q , v) transfer matrix in a compressed form. The improvement is achieved by grouping the original set of Catalan configurations into a forest of family trees, in such a way that the solution to the problem is now computed by solving the root node of each family. As a result, the algorithm becomes exponentially faster than the Catalan approach while still highly parallel. The resulting matrix is stored in a compressed form using O(3m ×4m) of space, making numerical evaluation and decompression to be faster than evaluating the matrix in its O(4m ×4m) uncompressed form. Experimental results for different sizes of strip lattices show that the parallel family trees (PFT) strategy indeed runs exponentially faster than the Catalan Parallel Method (CPM), especially when dealing with dense transfer matrices. In terms of parallel performance, we report strong-scaling speedups of up to 5.7 × when running on an 8-core shared memory machine and 28 × for a 32-core cluster. The best balance of speedup and efficiency for the multi-core machine was achieved when using p = 4 processors, while for the cluster

  6. Parallel computing for homogeneous diffusion and transport equations in neutronics

    International Nuclear Information System (INIS)

    Pinchedez, K.

    1999-06-01

    Parallel computing meets the ever-increasing requirements for neutronic computer code speed and accuracy. In this work, two different approaches have been considered. We first parallelized the sequential algorithm used by the neutronics code CRONOS developed at the French Atomic Energy Commission. The algorithm computes the dominant eigenvalue associated with PN simplified transport equations by a mixed finite element method. Several parallel algorithms have been developed on distributed memory machines. The performances of the parallel algorithms have been studied experimentally by implementation on a T3D Cray and theoretically by complexity models. A comparison of various parallel algorithms has confirmed the chosen implementations. We next applied a domain sub-division technique to the two-group diffusion Eigen problem. In the modal synthesis-based method, the global spectrum is determined from the partial spectra associated with sub-domains. Then the Eigen problem is expanded on a family composed, on the one hand, from eigenfunctions associated with the sub-domains and, on the other hand, from functions corresponding to the contribution from the interface between the sub-domains. For a 2-D homogeneous core, this modal method has been validated and its accuracy has been measured. (author)

  7. Models of parallel computation :a survey and classification

    Institute of Scientific and Technical Information of China (English)

    ZHANG Yunquan; CHEN Guoliang; SUN Guangzhong; MIAO Qiankun

    2007-01-01

    In this paper,the state-of-the-art parallel computational model research is reviewed.We will introduce various models that were developed during the past decades.According to their targeting architecture features,especially memory organization,we classify these parallel computational models into three generations.These models and their characteristics are discussed based on three generations classification.We believe that with the ever increasing speed gap between the CPU and memory systems,incorporating non-uniform memory hierarchy into computational models will become unavoidable.With the emergence of multi-core CPUs,the parallelism hierarchy of current computing platforms becomes more and more complicated.Describing this complicated parallelism hierarchy in future computational models becomes more and more important.A semi-automatic toolkit that can extract model parameters and their values on real computers can reduce the model analysis complexity,thus allowing more complicated models with more parameters to be adopted.Hierarchical memory and hierarchical parallelism will be two very important features that should be considered in future model design and research.

  8. PLAST: parallel local alignment search tool for database comparison

    Directory of Open Access Journals (Sweden)

    Lavenier Dominique

    2009-10-01

    Full Text Available Abstract Background Sequence similarity searching is an important and challenging task in molecular biology and next-generation sequencing should further strengthen the need for faster algorithms to process such vast amounts of data. At the same time, the internal architecture of current microprocessors is tending towards more parallelism, leading to the use of chips with two, four and more cores integrated on the same die. The main purpose of this work was to design an effective algorithm to fit with the parallel capabilities of modern microprocessors. Results A parallel algorithm for comparing large genomic banks and targeting middle-range computers has been developed and implemented in PLAST software. The algorithm exploits two key parallel features of existing and future microprocessors: the SIMD programming model (SSE instruction set and the multithreading concept (multicore. Compared to multithreaded BLAST software, tests performed on an 8-processor server have shown speedup ranging from 3 to 6 with a similar level of accuracy. Conclusion A parallel algorithmic approach driven by the knowledge of the internal microprocessor architecture allows significant speedup to be obtained while preserving standard sensitivity for similarity search problems.

  9. Parallel Breadth-First Search on Distributed Memory Systems

    Energy Technology Data Exchange (ETDEWEB)

    Computational Research Division; Buluc, Aydin; Madduri, Kamesh

    2011-04-15

    Data-intensive, graph-based computations are pervasive in several scientific applications, and are known to to be quite challenging to implement on distributed memory systems. In this work, we explore the design space of parallel algorithms for Breadth-First Search (BFS), a key subroutine in several graph algorithms. We present two highly-tuned par- allel approaches for BFS on large parallel systems: a level-synchronous strategy that relies on a simple vertex-based partitioning of the graph, and a two-dimensional sparse matrix- partitioning-based approach that mitigates parallel commu- nication overhead. For both approaches, we also present hybrid versions with intra-node multithreading. Our novel hybrid two-dimensional algorithm reduces communication times by up to a factor of 3.5, relative to a common vertex based approach. Our experimental study identifies execu- tion regimes in which these approaches will be competitive, and we demonstrate extremely high performance on lead- ing distributed-memory parallel systems. For instance, for a 40,000-core parallel execution on Hopper, an AMD Magny- Cours based system, we achieve a BFS performance rate of 17.8 billion edge visits per second on an undirected graph of 4.3 billion vertices and 68.7 billion edges with skewed degree distribution.

  10. Cellular automata a parallel model

    CERN Document Server

    Mazoyer, J

    1999-01-01

    Cellular automata can be viewed both as computational models and modelling systems of real processes. This volume emphasises the first aspect. In articles written by leading researchers, sophisticated massive parallel algorithms (firing squad, life, Fischer's primes recognition) are treated. Their computational power and the specific complexity classes they determine are surveyed, while some recent results in relation to chaos from a new dynamic systems point of view are also presented. Audience: This book will be of interest to specialists of theoretical computer science and the parallelism challenge.

  11. Parallel PDE-Based Simulations Using the Common Component Architecture

    International Nuclear Information System (INIS)

    McInnes, Lois C.; Allan, Benjamin A.; Armstrong, Robert; Benson, Steven J.; Bernholdt, David E.; Dahlgren, Tamara L.; Diachin, Lori; Krishnan, Manoj Kumar; Kohl, James A.; Larson, J. Walter; Lefantzi, Sophia; Nieplocha, Jarek; Norris, Boyana; Parker, Steven G.; Ray, Jaideep; Zhou, Shujia

    2006-01-01

    The complexity of parallel PDE-based simulations continues to increase as multimodel, multiphysics, and multi-institutional projects become widespread. A goal of component based software engineering in such large-scale simulations is to help manage this complexity by enabling better interoperability among various codes that have been independently developed by different groups. The Common Component Architecture (CCA) Forum is defining a component architecture specification to address the challenges of high-performance scientific computing. In addition, several execution frameworks, supporting infrastructure, and general purpose components are being developed. Furthermore, this group is collaborating with others in the high-performance computing community to design suites of domain-specific component interface specifications and underlying implementations. This chapter discusses recent work on leveraging these CCA efforts in parallel PDE-based simulations involving accelerator design, climate modeling, combustion, and accidental fires and explosions. We explain how component technology helps to address the different challenges posed by each of these applications, and we highlight how component interfaces built on existing parallel toolkits facilitate the reuse of software for parallel mesh manipulation, discretization, linear algebra, integration, optimization, and parallel data redistribution. We also present performance data to demonstrate the suitability of this approach, and we discuss strategies for applying component technologies to both new and existing applications

  12. Calibration of transfer functions between phytolith, vegetation and climate for integration of grassland dynamics in vegetation models. Application to a 50,000 yr crater lake core in Tanzania.

    Science.gov (United States)

    Bremond, L.; Alexandre, A.; Hely, C.; Vincens, A.; Williamson, D.; Guiot, J.

    2004-12-01

    Global vegetation models provide a way to translate the outputs from climate models into maps of potential vegetation distribution for present, past and future. Validation of these models goes through the comparison between model outputs and vegetation proxies for well constrained past climatic periods. Grass-dominated biomes are widespread and numerous. This diversity is hardly mirrored by common proxies such as pollen, charcoal or carbon isotopes. Phytoliths are amorphous silica that precipitate in and/or between living plant cells. They are commonly used to trace grasslands dynamics. However, calibration between phytolith assemblages, vegetation, and climate parameters are scarce. This work introduces transfer functions between phytolith indices, inter-tropical grassland physiognomy, and bio-climatic data that will be available for model/data comparisons. The Iph phytolith index discriminates tall from short grass savannas in West Africa. A transfer function allows to estimate evapo-transpiration AET/PET. The Ic phytolith index accurately estimates the proportion of Pooideae and Panicoideae grass sub-families, and potentially the C4/C3 grass dominance on East African mountains. The D/P index appears as a good proxy of Leaf Area Index (LAI) in tropical areas. These environmental parameters are commonly used as vegetation model outputs, but have been, up to now, hardly estimated by vegetation proxies. These transfer functions are applied to a 50,000 yr phytolith sequence from a crater lake (9°S; 33°E Tanzania). The record is compared to the pollen vegetation reconstruction and confronted to simulations of the LPJ-GUESS vegetation model (Stitch et. al, 2003).

  13. A parallel solver for huge dense linear systems

    Science.gov (United States)

    Badia, J. M.; Movilla, J. L.; Climente, J. I.; Castillo, M.; Marqués, M.; Mayo, R.; Quintana-Ortí, E. S.; Planelles, J.

    2011-11-01

    HDSS (Huge Dense Linear System Solver) is a Fortran Application Programming Interface (API) to facilitate the parallel solution of very large dense systems to scientists and engineers. The API makes use of parallelism to yield an efficient solution of the systems on a wide range of parallel platforms, from clusters of processors to massively parallel multiprocessors. It exploits out-of-core strategies to leverage the secondary memory in order to solve huge linear systems O(100.000). The API is based on the parallel linear algebra library PLAPACK, and on its Out-Of-Core (OOC) extension POOCLAPACK. Both PLAPACK and POOCLAPACK use the Message Passing Interface (MPI) as the communication layer and BLAS to perform the local matrix operations. The API provides a friendly interface to the users, hiding almost all the technical aspects related to the parallel execution of the code and the use of the secondary memory to solve the systems. In particular, the API can automatically select the best way to store and solve the systems, depending of the dimension of the system, the number of processes and the main memory of the platform. Experimental results on several parallel platforms report high performance, reaching more than 1 TFLOP with 64 cores to solve a system with more than 200 000 equations and more than 10 000 right-hand side vectors. New version program summaryProgram title: Huge Dense System Solver (HDSS) Catalogue identifier: AEHU_v1_1 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEHU_v1_1.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 87 062 No. of bytes in distributed program, including test data, etc.: 1 069 110 Distribution format: tar.gz Programming language: Fortran90, C Computer: Parallel architectures: multiprocessors, computer clusters Operating system

  14. Parallel Sparse Matrix - Vector Product

    DEFF Research Database (Denmark)

    Alexandersen, Joe; Lazarov, Boyan Stefanov; Dammann, Bernd

    This technical report contains a case study of a sparse matrix-vector product routine, implemented for parallel execution on a compute cluster with both pure MPI and hybrid MPI-OpenMP solutions. C++ classes for sparse data types were developed and the report shows how these class can be used...

  15. [Falsified medicines in parallel trade].

    Science.gov (United States)

    Muckenfuß, Heide

    2017-11-01

    The number of falsified medicines on the German market has distinctly increased over the past few years. In particular, stolen pharmaceutical products, a form of falsified medicines, have increasingly been introduced into the legal supply chain via parallel trading. The reasons why parallel trading serves as a gateway for falsified medicines are most likely the complex supply chains and routes of transport. It is hardly possible for national authorities to trace the history of a medicinal product that was bought and sold by several intermediaries in different EU member states. In addition, the heterogeneous outward appearance of imported and relabelled pharmaceutical products facilitates the introduction of illegal products onto the market. Official batch release at the Paul-Ehrlich-Institut offers the possibility of checking some aspects that might provide an indication of a falsified medicine. In some circumstances, this may allow the identification of falsified medicines before they come onto the German market. However, this control is only possible for biomedicinal products that have not received a waiver regarding official batch release. For improved control of parallel trade, better networking among the EU member states would be beneficial. European-wide regulations, e. g., for disclosure of the complete supply chain, would help to minimise the risks of parallel trading and hinder the marketing of falsified medicines.

  16. The parallel adult education system

    DEFF Research Database (Denmark)

    Wahlgren, Bjarne

    2015-01-01

    for competence development. The Danish university educational system includes two parallel programs: a traditional academic track (candidatus) and an alternative practice-based track (master). The practice-based program was established in 2001 and organized as part time. The total program takes half the time...

  17. Where are the parallel algorithms?

    Science.gov (United States)

    Voigt, R. G.

    1985-01-01

    Four paradigms that can be useful in developing parallel algorithms are discussed. These include computational complexity analysis, changing the order of computation, asynchronous computation, and divide and conquer. Each is illustrated with an example from scientific computation, and it is shown that computational complexity must be used with great care or an inefficient algorithm may be selected.

  18. Parallel imaging with phase scrambling.

    Science.gov (United States)

    Zaitsev, Maxim; Schultz, Gerrit; Hennig, Juergen; Gruetter, Rolf; Gallichan, Daniel

    2015-04-01

    Most existing methods for accelerated parallel imaging in MRI require additional data, which are used to derive information about the sensitivity profile of each radiofrequency (RF) channel. In this work, a method is presented to avoid the acquisition of separate coil calibration data for accelerated Cartesian trajectories. Quadratic phase is imparted to the image to spread the signals in k-space (aka phase scrambling). By rewriting the Fourier transform as a convolution operation, a window can be introduced to the convolved chirp function, allowing a low-resolution image to be reconstructed from phase-scrambled data without prominent aliasing. This image (for each RF channel) can be used to derive coil sensitivities to drive existing parallel imaging techniques. As a proof of concept, the quadratic phase was applied by introducing an offset to the x(2) - y(2) shim and the data were reconstructed using adapted versions of the image space-based sensitivity encoding and GeneRalized Autocalibrating Partially Parallel Acquisitions algorithms. The method is demonstrated in a phantom (1 × 2, 1 × 3, and 2 × 2 acceleration) and in vivo (2 × 2 acceleration) using a 3D gradient echo acquisition. Phase scrambling can be used to perform parallel imaging acceleration without acquisition of separate coil calibration data, demonstrated here for a 3D-Cartesian trajectory. Further research is required to prove the applicability to other 2D and 3D sampling schemes. © 2014 Wiley Periodicals, Inc.

  19. Default Parallels Plesk Panel Page

    Science.gov (United States)

    services that small businesses want and need. Our software includes key building blocks of cloud service virtualized servers Service Provider Products Parallels® Automation Hosting, SaaS, and cloud computing , the leading hosting automation software. You see this page because there is no Web site at this

  20. Parallel plate transmission line transformer

    NARCIS (Netherlands)

    Voeten, S.J.; Brussaard, G.J.H.; Pemen, A.J.M.

    2011-01-01

    A Transmission Line Transformer (TLT) can be used to transform high-voltage nanosecond pulses. These transformers rely on the fact that the length of the pulse is shorter than the transmission lines used. This allows connecting the transmission lines in parallel at the input and in series at the

  1. Matpar: Parallel Extensions for MATLAB

    Science.gov (United States)

    Springer, P. L.

    1998-01-01

    Matpar is a set of client/server software that allows a MATLAB user to take advantage of a parallel computer for very large problems. The user can replace calls to certain built-in MATLAB functions with calls to Matpar functions.

  2. Massively parallel quantum computer simulator

    NARCIS (Netherlands)

    De Raedt, K.; Michielsen, K.; De Raedt, H.; Trieu, B.; Arnold, G.; Richter, M.; Lippert, Th.; Watanabe, H.; Ito, N.

    2007-01-01

    We describe portable software to simulate universal quantum computers on massive parallel Computers. We illustrate the use of the simulation software by running various quantum algorithms on different computer architectures, such as a IBM BlueGene/L, a IBM Regatta p690+, a Hitachi SR11000/J1, a Cray

  3. Reactor core fuel management

    International Nuclear Information System (INIS)

    Silvennoinen, P.

    1976-01-01

    The subject is covered in chapters, entitled: concepts of reactor physics; neutron diffusion; core heat transfer; reactivity; reactor operation; variables of core management; computer code modules; alternative reactor concepts; methods of optimization; general system aspects. (U.K.)

  4. Nuclear reactor core catcher

    International Nuclear Information System (INIS)

    1977-01-01

    A nuclear reactor core catcher is described for containing debris resulting from an accident causing core meltdown and which incorporates a method of cooling the debris by the circulation of a liquid coolant. (U.K.)

  5. Compiling the functional data-parallel language SaC for Microgrids of Self-Adaptive Virtual Processors

    NARCIS (Netherlands)

    Grelck, C.; Herhut, S.; Jesshope, C.; Joslin, C.; Lankamp, M.; Scholz, S.-B.; Shafarenko, A.

    2009-01-01

    We present preliminary results from compiling the high-level, functional and data-parallel programming language SaC into a novel multi-core design: Microgrids of Self-Adaptive Virtual Processors (SVPs). The side-effect free nature of SaC in conjunction with its data-parallel foundation make it an

  6. Parallel computing: numerics, applications, and trends

    National Research Council Canada - National Science Library

    Trobec, Roman; Vajteršic, Marián; Zinterhof, Peter

    2009-01-01

    ... and/or distributed systems. The contributions to this book are focused on topics most concerned in the trends of today's parallel computing. These range from parallel algorithmics, programming, tools, network computing to future parallel computing. Particular attention is paid to parallel numerics: linear algebra, differential equations, numerica...

  7. Experiments with parallel algorithms for combinatorial problems

    NARCIS (Netherlands)

    G.A.P. Kindervater (Gerard); H.W.J.M. Trienekens

    1985-01-01

    textabstractIn the last decade many models for parallel computation have been proposed and many parallel algorithms have been developed. However, few of these models have been realized and most of these algorithms are supposed to run on idealized, unrealistic parallel machines. The parallel machines

  8. Parallel R-matrix computation

    International Nuclear Information System (INIS)

    Heggarty, J.W.

    1999-06-01

    For almost thirty years, sequential R-matrix computation has been used by atomic physics research groups, from around the world, to model collision phenomena involving the scattering of electrons or positrons with atomic or molecular targets. As considerable progress has been made in the understanding of fundamental scattering processes, new data, obtained from more complex calculations, is of current interest to experimentalists. Performing such calculations, however, places considerable demands on the computational resources to be provided by the target machine, in terms of both processor speed and memory requirement. Indeed, in some instances the computational requirements are so great that the proposed R-matrix calculations are intractable, even when utilising contemporary classic supercomputers. Historically, increases in the computational requirements of R-matrix computation were accommodated by porting the problem codes to a more powerful classic supercomputer. Although this approach has been successful in the past, it is no longer considered to be a satisfactory solution due to the limitations of current (and future) Von Neumann machines. As a consequence, there has been considerable interest in the high performance multicomputers, that have emerged over the last decade which appear to offer the computational resources required by contemporary R-matrix research. Unfortunately, developing codes for these machines is not as simple a task as it was to develop codes for successive classic supercomputers. The difficulty arises from the considerable differences in the computing models that exist between the two types of machine and results in the programming of multicomputers to be widely acknowledged as a difficult, time consuming and error-prone task. Nevertheless, unless parallel R-matrix computation is realised, important theoretical and experimental atomic physics research will continue to be hindered. This thesis describes work that was undertaken in

  9. Seismic core shroud

    International Nuclear Information System (INIS)

    Puri, A.; Mullooly, J.F.

    1981-01-01

    A core shroud is provided, comprising: a coolant boundary, following the shape of the core boundary, for channeling the coolant through the fuel assemblies; a cylindrical band positioned inside the core barrel and surrounding the coolant boundary; and support members extending from the coolant boundary to the band, for transferring load from the coolant boundary to the band. The shroud may be assembled in parts using automated welding techniques, and it may be adjusted to fit the reactor core easily

  10. Denali Ice Core Record of North Pacific Sea Surface Temperatures and Marine Primary Productivity

    Science.gov (United States)

    Polashenski, D.; Osterberg, E. C.; Kreutz, K. J.; Winski, D.; Wake, C. P.; Ferris, D. G.; Introne, D.; Campbell, S. W.

    2016-12-01

    Chemical analyses of precipitation preserved in glacial ice cores provide a unique opportunity to study changes in atmospheric circulation patterns and ocean surface conditions through time. In this study, we aim to investigate changes in both the physical and biological parameters of the north-central Pacific Ocean and Bering Sea over the twentieth century using the deuterium excess (d-excess) and methanesulfonic acid (MSA) records from the Mt. Hunter ice cores drilled in Denali National Park, Alaska. These parallel, 208 m-long ice cores were drilled to bedrock during the 2013 field season on the Mt. Hunter plateau (63° N, 151° W, 3,900 m above sea level) by a collaborative research team consisting of members from Dartmouth College and the Universities of Maine and New Hampshire. The cores were sampled on a continuous melter system at Dartmouth College and analyzed for the concentrations major ions (Dionex IC) and trace metals (Element2 ICPMS), and for stable water isotope ratios (Picarro). The depth-age scale has been accurately dated to 400 AD using annual layer counting of several chemical species and further validated using known historical volcanic eruptions and the Cesium-137 spike associated with nuclear weapons testing in 1963. We use HYSPLIT back trajectory modeling to identify likely source areas of moisture and aerosol MSA being transported to the core site. Satellite imagery allows for a direct comparison between chlorophyll a concentrations in these source areas and MSA concentrations in the core record. Preliminary analysis of chlorophyll a and MSA concentrations, both derived almost exclusively from marine biota, suggest that the Mt. Hunter ice cores reflect changes in North Pacific and Bering Sea marine primary productivity. Analysis of the water isotope and MSA data in conjunction with climate reanalysis products shows significant correlations (psea surface temperatures in the Bering Sea and North Central Pacific. These findings, coupled with

  11. The numerical parallel computing of photon transport

    International Nuclear Information System (INIS)

    Huang Qingnan; Liang Xiaoguang; Zhang Lifa

    1998-12-01

    The parallel computing of photon transport is investigated, the parallel algorithm and the parallelization of programs on parallel computers both with shared memory and with distributed memory are discussed. By analyzing the inherent law of the mathematics and physics model of photon transport according to the structure feature of parallel computers, using the strategy of 'to divide and conquer', adjusting the algorithm structure of the program, dissolving the data relationship, finding parallel liable ingredients and creating large grain parallel subtasks, the sequential computing of photon transport into is efficiently transformed into parallel and vector computing. The program was run on various HP parallel computers such as the HY-1 (PVP), the Challenge (SMP) and the YH-3 (MPP) and very good parallel speedup has been gotten

  12. Core Values | NREL

    Science.gov (United States)

    Core Values Core Values NREL's core values are rooted in a safe and supportive work environment guide our everyday actions and efforts: Safe and supportive work environment Respect for the rights physical and social environment Integrity Maintain the highest standard of ethics, honesty, and integrity

  13. Sidewall coring shell

    Energy Technology Data Exchange (ETDEWEB)

    Edelman, Ya A; Konstantinov, L P; Martyshin, A N

    1966-12-12

    A sidewall coring shell consists of a housing and a detachable core catcher. The core lifter is provided with projections, the ends of which are situated in another plane, along the longitudinal axis of the lifter. The chamber has corresponding projections.

  14. Carbon dioxide and climate

    International Nuclear Information System (INIS)

    1991-10-01

    Global climate change is a serious environmental concern, and the US has developed ''An Action Agenda'' to deal with it. At the heart of the US effort is the US Global Change Research Program (USGCRP), which has been developed by the Committee on Earth and Environmental Sciences (CEES) of the Federal Coordinating Council for Sciences, Engineering, and Technology (FCCSET). The USGCRP will provide the scientific basis for sound policy making on the climate-change issue. The DOE contribution to the USGCRP is the Carbon Dioxide Research Program, which now places particular emphasis on the rapid improvement of the capability to predict global and regional climate change. DOE's Carbon Dioxide Research Program has been addressing the carbon dioxide-climate change connection for more than twelve years and has provided a solid scientific foundation for the USGCRP. The expansion of the DOE effort reflects the increased attention that the Department has placed on the issue and is reflected in the National Energy Strategy (NES) that was released in 1991. This Program Summary describes projects funded by the Carbon Dioxide Research Program during FY 1991 and gives a brief overview of objectives, organization, and accomplishments. The Environmental Sciences Division of the Office of Health and Environmental Research, Office of Energy Research supports a Carbon Dioxide Research Program to determine the scientific linkage between the rise of greenhouse gases in the atmosphere, especially carbon dioxide, and climate and vegetation change. One facet is the Core CO 2 Program, a pioneering program that DOE established more than 10 years ago to understand and predict the ways that fossil-fuel burning could affect atmospheric CO 2 concentration, global climate, and the Earth's biosphere. Major research areas are: global carbon cycle; climate detection and models of climate change; vegetation research; resource analysis; and, information and integration

  15. 1500-year Record of trans-Pacific Dust Flux collected from the Denali Ice Core, Mt. Hunter, Alaska

    Science.gov (United States)

    Saylor, P. L.; Osterberg, E. C.; Koffman, B. G.; Winski, D.; Ferris, D. G.; Kreutz, K. J.; Wake, C. P.; Handley, M.; Campbell, S. W.

    2016-12-01

    Mineral dust aerosols are a critical component of the climate system through their influence on atmospheric radiative forcing, ocean productivity, and surface albedo. Dust aerosols derived from Asian deserts are known to reach as far as Europe through efficient transport in the upper tropospheric westerlies. While centennially-to-millennially resolved Asian dust records exist over the late Holocene from North Pacific marine sediment cores and Asian loess deposits, a high-resolution (sub-annual to decadal) record of trans-Pacific dust flux will significantly improve our understanding of North Pacific dust-climate interactions and provide paleoclimatological context for 20th century dust activity. Here we present an annually resolved 1500-year record of trans-Pacific dust transport based on chemical and physical dust measurements in parallel Alaskan ice cores (208 m to bedrock) collected from the summit plateau of Mt. Hunter in Denali National Park. The cores were sampled at high resolution using a continuous melter system with discrete analyses for major ions (Dionex ion chromatograph), trace elements (Element2 inductively coupled plasma mass spectrometer), and stable water isotope ratios (Picarro laser ringdown spectroscopy), and continuous flow analysis for dust concentration and size distribution (Klotz Abakus). We compare the ice core dust record to instrumental aerosol stations, satellite observations, and dust model data from the instrumental period, and evaluate climatic controls on dust emission and trans-Pacific transport using climate reanalysis data, to inform dust-climate relationships over the past 1500 years. Physical particulate and chemical data demonstrate remarkable fidelity at sub-annual resolution, with both displaying a strong springtime peak consistent with periods of high dust activity over Asian desert source regions. Preliminary results suggest volumetric mode typically ranges from 4.5 - 6.5 um, with a mean value of 5.5 um. Preliminary

  16. Rotary core drills

    Energy Technology Data Exchange (ETDEWEB)

    1967-11-30

    The design of a rotary core drill is described. Primary consideration is given to the following component parts of the drill: the inner and outer tube, the core bit, an adapter, and the core lifter. The adapter has the form of a downward-converging sleeve and is mounted to the lower end of the inner tube. The lifter, extending from the adapter, is split along each side so that it can be held open to permit movement of a core. It is possible to grip a core by allowing the lifter to assume a closed position.

  17. Climate Kids

    Science.gov (United States)

    ... What Is Permafrost? How Do We Predict Future Climate? Green Career: Earth Scientist 10 Things About Ecosystems ... study Earth? What can trees tell us about climate change? Why does NASA care about food? Games ...

  18. Three dimensional Burn-up program parallelization using socket programming

    International Nuclear Information System (INIS)

    Haliyati R, Evi; Su'ud, Zaki

    2002-01-01

    A computer parallelization process was built with a purpose to decrease execution time of a physics program. In this case, a multi computer system was built to be used to analyze burn-up process of a nuclear reactor. This multi computer system was design need using a protocol communication among sockets, i.e. TCP/IP. This system consists of computer as a server and the rest as clients. The server has a main control to all its clients. The server also divides the reactor core geometrically to in parts in accordance with the number of clients, each computer including the server has a task to conduct burn-up analysis of 1/n part of the total reactor core measure. This burn-up analysis was conducted simultaneously and in a parallel way by all computers, so a faster program execution time was achieved close to 1/n times that of one computer. Then an analysis was carried out and states that in order to calculate the density of atoms in a reactor of 91 cm x 91 cm x 116 cm, the usage of a parallel system of 2 computers has the highest efficiency

  19. Parallel file system performances in fusion data storage

    International Nuclear Information System (INIS)

    Iannone, F.; Podda, S.; Bracco, G.; Manduchi, G.; Maslennikov, A.; Migliori, S.; Wolkersdorfer, K.

    2012-01-01

    High I/O flow rates, up to 10 GB/s, are required in large fusion Tokamak experiments like ITER where hundreds of nodes store simultaneously large amounts of data acquired during the plasma discharges. Typical network topologies such as linear arrays (systolic), rings, meshes (2-D arrays), tori (3-D arrays), trees, butterfly, hypercube in combination with high speed data transports like Infiniband or 10G-Ethernet, are the main areas in which the effort to overcome the so-called parallel I/O bottlenecks is most focused. The high I/O flow rates were modelled in an emulated testbed based on the parallel file systems such as Lustre and GPFS, commonly used in High Performance Computing. The test runs on High Performance Computing–For Fusion (8640 cores) and ENEA CRESCO (3392 cores) supercomputers. Message Passing Interface based applications were developed to emulate parallel I/O on Lustre and GPFS using data archival and access solutions like MDSPLUS and Universal Access Layer. These methods of data storage organization are widely diffused in nuclear fusion experiments and are being developed within the EFDA Integrated Tokamak Modelling – Task Force; the authors tried to evaluate their behaviour in a realistic emulation setup.

  20. Parallel efficient rate control methods for JPEG 2000

    Science.gov (United States)

    Martínez-del-Amor, Miguel Á.; Bruns, Volker; Sparenberg, Heiko

    2017-09-01

    Since the introduction of JPEG 2000, several rate control methods have been proposed. Among them, post-compression rate-distortion optimization (PCRD-Opt) is the most widely used, and the one recommended by the standard. The approach followed by this method is to first compress the entire image split in code blocks, and subsequently, optimally truncate the set of generated bit streams according to the maximum target bit rate constraint. The literature proposes various strategies on how to estimate ahead of time where a block will get truncated in order to stop the execution prematurely and save time. However, none of them have been defined bearing in mind a parallel implementation. Today, multi-core and many-core architectures are becoming popular for JPEG 2000 codecs implementations. Therefore, in this paper, we analyze how some techniques for efficient rate control can be deployed in GPUs. In order to do that, the design of our GPU-based codec is extended, allowing stopping the process at a given point. This extension also harnesses a higher level of parallelism on the GPU, leading to up to 40% of speedup with 4K test material on a Titan X. In a second step, three selected rate control methods are adapted and implemented in our parallel encoder. A comparison is then carried out, and used to select the best candidate to be deployed in a GPU encoder, which gave an extra 40% of speedup in those situations where it was really employed.

  1. Block-Parallel Data Analysis with DIY2

    Energy Technology Data Exchange (ETDEWEB)

    Morozov, Dmitriy [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Peterka, Tom [Argonne National Lab. (ANL), Argonne, IL (United States)

    2017-08-30

    DIY2 is a programming model and runtime for block-parallel analytics on distributed-memory machines. Its main abstraction is block-structured data parallelism: data are decomposed into blocks; blocks are assigned to processing elements (processes or threads); computation is described as iterations over these blocks, and communication between blocks is defined by reusable patterns. By expressing computation in this general form, the DIY2 runtime is free to optimize the movement of blocks between slow and fast memories (disk and flash vs. DRAM) and to concurrently execute blocks residing in memory with multiple threads. This enables the same program to execute in-core, out-of-core, serial, parallel, single-threaded, multithreaded, or combinations thereof. This paper describes the implementation of the main features of the DIY2 programming model and optimizations to improve performance. DIY2 is evaluated on benchmark test cases to establish baseline performance for several common patterns and on larger complete analysis codes running on large-scale HPC machines.

  2. Parallel optimization of IDW interpolation algorithm on multicore platform

    Science.gov (United States)

    Guan, Xuefeng; Wu, Huayi

    2009-10-01

    Due to increasing power consumption, heat dissipation, and other physical issues, the architecture of central processing unit (CPU) has been turning to multicore rapidly in recent years. Multicore processor is packaged with multiple processor cores in the same chip, which not only offers increased performance, but also presents significant challenges to application developers. As a matter of fact, in GIS field most of current GIS algorithms were implemented serially and could not best exploit the parallelism potential on such multicore platforms. In this paper, we choose Inverse Distance Weighted spatial interpolation algorithm (IDW) as an example to study how to optimize current serial GIS algorithms on multicore platform in order to maximize performance speedup. With the help of OpenMP, threading methodology is introduced to split and share the whole interpolation work among processor cores. After parallel optimization, execution time of interpolation algorithm is greatly reduced and good performance speedup is achieved. For example, performance speedup on Intel Xeon 5310 is 1.943 with 2 execution threads and 3.695 with 4 execution threads respectively. An additional output comparison between pre-optimization and post-optimization is carried out and shows that parallel optimization does to affect final interpolation result.

  3. Parallel file system performances in fusion data storage

    Energy Technology Data Exchange (ETDEWEB)

    Iannone, F., E-mail: francesco.iannone@enea.it [Associazione EURATOM-ENEA sulla Fusione, C.R.ENEA Frascati, via E.Fermi, 45 - 00044 Frascati, Rome (Italy); Podda, S.; Bracco, G. [ENEA Information Communication Tecnologies, Lungotevere Thaon di Revel, 76 - 00196 Rome (Italy); Manduchi, G. [Associazione EURATOM-ENEA sulla Fusione, Consorzio RFX, Corso Stati Uniti, 4 - 35127 Padua (Italy); Maslennikov, A. [CASPUR Inter-University Consortium for the Application of Super-Computing for Research, via dei Tizii, 6b - 00185 Rome (Italy); Migliori, S. [ENEA Information Communication Tecnologies, Lungotevere Thaon di Revel, 76 - 00196 Rome (Italy); Wolkersdorfer, K. [Juelich Supercomputing Centre-FZJ, D-52425 Juelich (Germany)

    2012-12-15

    High I/O flow rates, up to 10 GB/s, are required in large fusion Tokamak experiments like ITER where hundreds of nodes store simultaneously large amounts of data acquired during the plasma discharges. Typical network topologies such as linear arrays (systolic), rings, meshes (2-D arrays), tori (3-D arrays), trees, butterfly, hypercube in combination with high speed data transports like Infiniband or 10G-Ethernet, are the main areas in which the effort to overcome the so-called parallel I/O bottlenecks is most focused. The high I/O flow rates were modelled in an emulated testbed based on the parallel file systems such as Lustre and GPFS, commonly used in High Performance Computing. The test runs on High Performance Computing-For Fusion (8640 cores) and ENEA CRESCO (3392 cores) supercomputers. Message Passing Interface based applications were developed to emulate parallel I/O on Lustre and GPFS using data archival and access solutions like MDSPLUS and Universal Access Layer. These methods of data storage organization are widely diffused in nuclear fusion experiments and are being developed within the EFDA Integrated Tokamak Modelling - Task Force; the authors tried to evaluate their behaviour in a realistic emulation setup.

  4. QR-decomposition based SENSE reconstruction using parallel architecture.

    Science.gov (United States)

    Ullah, Irfan; Nisar, Habab; Raza, Haseeb; Qasim, Malik; Inam, Omair; Omer, Hammad

    2018-04-01

    Magnetic Resonance Imaging (MRI) is a powerful medical imaging technique that provides essential clinical information about the human body. One major limitation of MRI is its long scan time. Implementation of advance MRI algorithms on a parallel architecture (to exploit inherent parallelism) has a great potential to reduce the scan time. Sensitivity Encoding (SENSE) is a Parallel Magnetic Resonance Imaging (pMRI) algorithm that utilizes receiver coil sensitivities to reconstruct MR images from the acquired under-sampled k-space data. At the heart of SENSE lies inversion of a rectangular encoding matrix. This work presents a novel implementation of GPU based SENSE algorithm, which employs QR decomposition for the inversion of the rectangular encoding matrix. For a fair comparison, the performance of the proposed GPU based SENSE reconstruction is evaluated against single and multicore CPU using openMP. Several experiments against various acceleration factors (AFs) are performed using multichannel (8, 12 and 30) phantom and in-vivo human head and cardiac datasets. Experimental results show that GPU significantly reduces the computation time of SENSE reconstruction as compared to multi-core CPU (approximately 12x speedup) and single-core CPU (approximately 53x speedup) without any degradation in the quality of the reconstructed images. Copyright © 2018 Elsevier Ltd. All rights reserved.

  5. Feasibility Study of Core Design with a Monte Carlo Code for APR1400 Initial core

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jinsun; Chang, Do Ik; Seong, Kibong [KEPCO NF, Daejeon (Korea, Republic of)

    2014-10-15

    The Monte Carlo calculation becomes more popular and useful nowadays due to the rapid progress in computing power and parallel calculation techniques. There have been many attempts to analyze a commercial core by Monte Carlo transport code using the enhanced computer capability, recently. In this paper, Monte Carlo calculation of APR1400 initial core has been performed and the results are compared with the calculation results of conventional deterministic code to find out the feasibility of core design using Monte Carlo code. SERPENT, a 3D continuous-energy Monte Carlo reactor physics burnup calculation code is used for this purpose and the KARMA-ASTRA code system, which is used for a deterministic code of comparison. The preliminary investigation for the feasibility of commercial core design with Monte Carlo code was performed in this study. Simplified core geometry modeling was performed for the reactor core surroundings and reactor coolant model is based on two region model. The reactivity difference at HZP ARO condition between Monte Carlo code and the deterministic code is consistent with each other and the reactivity difference during the depletion could be reduced by adopting the realistic moderator temperature. The reactivity difference calculated at HFP, BOC, ARO equilibrium condition was 180 ±9 pcm, with axial moderator temperature of a deterministic code. The computing time will be a significant burden at this time for the application of Monte Carlo code to the commercial core design even with the application of parallel computing because numerous core simulations are required for actual loading pattern search. One of the remedy will be a combination of Monte Carlo code and the deterministic code to generate the physics data. The comparison of physics parameters with sophisticated moderator temperature modeling and depletion will be performed for a further study.

  6. Understanding climate

    International Nuclear Information System (INIS)

    1995-01-01

    In this article the following question is answered. What is the climate? What factors do determine our climate? What is solar radiation? How does solar radiation relate to the earth's energy? What is greenhouse effect? What role does the greenhouse effect play in the global ecosystem? How does the water cycle affect climate? What is drought? What role do oceans play in influencing climate. (author)

  7. Parallel Computing for Terrestrial Ecosystem Carbon Modeling

    International Nuclear Information System (INIS)

    Wang, Dali; Post, Wilfred M.; Ricciuto, Daniel M.; Berry, Michael

    2011-01-01

    Terrestrial ecosystems are a primary component of research on global environmental change. Observational and modeling research on terrestrial ecosystems at the global scale, however, has lagged behind their counterparts for oceanic and atmospheric systems, largely because the unique challenges associated with the tremendous diversity and complexity of terrestrial ecosystems. There are 8 major types of terrestrial ecosystem: tropical rain forest, savannas, deserts, temperate grassland, deciduous forest, coniferous forest, tundra, and chaparral. The carbon cycle is an important mechanism in the coupling of terrestrial ecosystems with climate through biological fluxes of CO 2 . The influence of terrestrial ecosystems on atmospheric CO 2 can be modeled via several means at different timescales. Important processes include plant dynamics, change in land use, as well as ecosystem biogeography. Over the past several decades, many terrestrial ecosystem models (see the 'Model developments' section) have been developed to understand the interactions between terrestrial carbon storage and CO 2 concentration in the atmosphere, as well as the consequences of these interactions. Early TECMs generally adapted simple box-flow exchange models, in which photosynthetic CO 2 uptake and respiratory CO 2 release are simulated in an empirical manner with a small number of vegetation and soil carbon pools. Demands on kinds and amount of information required from global TECMs have grown. Recently, along with the rapid development of parallel computing, spatially explicit TECMs with detailed process based representations of carbon dynamics become attractive, because those models can readily incorporate a variety of additional ecosystem processes (such as dispersal, establishment, growth, mortality etc.) and environmental factors (such as landscape position, pest populations, disturbances, resource manipulations, etc.), and provide information to frame policy options for climate change

  8. Climate Change: The Evidence and Our Options

    Science.gov (United States)

    Thompson, Lonnie G.

    2010-01-01

    Glaciers serve as early indicators of climate change. Over the last 35 years, our research team has recovered ice-core records of climatic and environmental variations from the polar regions and from low-latitude high-elevation ice fields from 16 countries. The ongoing widespread melting of high-elevation glaciers and ice caps, particularly in low…

  9. Structural synthesis of parallel robots

    CERN Document Server

    Gogu, Grigore

    This book represents the fifth part of a larger work dedicated to the structural synthesis of parallel robots. The originality of this work resides in the fact that it combines new formulae for mobility, connectivity, redundancy and overconstraints with evolutionary morphology in a unified structural synthesis approach that yields interesting and innovative solutions for parallel robotic manipulators.  This is the first book on robotics that presents solutions for coupled, decoupled, uncoupled, fully-isotropic and maximally regular robotic manipulators with Schönflies motions systematically generated by using the structural synthesis approach proposed in Part 1.  Overconstrained non-redundant/overactuated/redundantly actuated solutions with simple/complex limbs are proposed. Many solutions are presented here for the first time in the literature. The author had to make a difficult and challenging choice between protecting these solutions through patents and releasing them directly into the public domain. T...

  10. GPU Parallel Bundle Block Adjustment

    Directory of Open Access Journals (Sweden)

    ZHENG Maoteng

    2017-09-01

    Full Text Available To deal with massive data in photogrammetry, we introduce the GPU parallel computing technology. The preconditioned conjugate gradient and inexact Newton method are also applied to decrease the iteration times while solving the normal equation. A brand new workflow of bundle adjustment is developed to utilize GPU parallel computing technology. Our method can avoid the storage and inversion of the big normal matrix, and compute the normal matrix in real time. The proposed method can not only largely decrease the memory requirement of normal matrix, but also largely improve the efficiency of bundle adjustment. It also achieves the same accuracy as the conventional method. Preliminary experiment results show that the bundle adjustment of a dataset with about 4500 images and 9 million image points can be done in only 1.5 minutes while achieving sub-pixel accuracy.

  11. Test model of WWER core

    International Nuclear Information System (INIS)

    Tikhomirov, A. V.; Gorokhov, A. K.

    2007-01-01

    The objective of this paper is creation of precision test model for WWER RP neutron-physics calculations. The model is considered as a tool for verification of deterministic computer codes that enables to reduce conservatism of design calculations and enhance WWER RP competitiveness. Precision calculations were performed using code MCNP5/1/ (Monte Carlo method). Engineering computer package Sapfir 9 5andRC V VER/2/ is used in comparative analysis of the results, it was certified for design calculations of WWER RU neutron-physics characteristic. The object of simulation is the first fuel loading of Volgodon NPP RP. Peculiarities of transition in calculation using MCNP5 from 2D geometry to 3D geometry are shown on the full-scale model. All core components as well as radial and face reflectors, automatic regulation in control and protection system control rod are represented in detail description according to the design. The first stage of application of the model is assessment of accuracy of calculation of the core power. At the second stage control and protection system control rod worth was assessed. Full scale RP representation in calculation using code MCNP5 is time consuming that calls for parallelization of computational problem on multiprocessing computer (Authors)

  12. HYDRATE CORE DRILLING TESTS

    Energy Technology Data Exchange (ETDEWEB)

    John H. Cohen; Thomas E. Williams; Ali G. Kadaster; Bill V. Liddell

    2002-11-01

    The ''Methane Hydrate Production from Alaskan Permafrost'' project is a three-year endeavor being conducted by Maurer Technology Inc. (MTI), Noble, and Anadarko Petroleum, in partnership with the U.S. DOE National Energy Technology Laboratory (NETL). The project's goal is to build on previous and ongoing R&D in the area of onshore hydrate deposition. The project team plans to design and implement a program to safely and economically drill, core and produce gas from arctic hydrates. The current work scope includes drilling and coring one well on Anadarko leases in FY 2003 during the winter drilling season. A specially built on-site core analysis laboratory will be used to determine some of the physical characteristics of the hydrates and surrounding rock. Prior to going to the field, the project team designed and conducted a controlled series of coring tests for simulating coring of hydrate formations. A variety of equipment and procedures were tested and modified to develop a practical solution for this special application. This Topical Report summarizes these coring tests. A special facility was designed and installed at MTI's Drilling Research Center (DRC) in Houston and used to conduct coring tests. Equipment and procedures were tested by cutting cores from frozen mixtures of sand and water supported by casing and designed to simulate hydrate formations. Tests were conducted with chilled drilling fluids. Tests showed that frozen core can be washed out and reduced in size by the action of the drilling fluid. Washing of the core by the drilling fluid caused a reduction in core diameter, making core recovery very difficult (if not impossible). One successful solution was to drill the last 6 inches of core dry (without fluid circulation). These tests demonstrated that it will be difficult to capture core when drilling in permafrost or hydrates without implementing certain safeguards. Among the coring tests was a simulated hydrate

  13. A tandem parallel plate analyzer

    International Nuclear Information System (INIS)

    Hamada, Y.; Fujisawa, A.; Iguchi, H.; Nishizawa, A.; Kawasumi, Y.

    1996-11-01

    By a new modification of a parallel plate analyzer the second-order focus is obtained in an arbitrary injection angle. This kind of an analyzer with a small injection angle will have an advantage of small operational voltage, compared to the Proca and Green analyzer where the injection angle is 30 degrees. Thus, the newly proposed analyzer will be very useful for the precise energy measurement of high energy particles in MeV range. (author)

  14. High-speed parallel counter

    International Nuclear Information System (INIS)

    Gus'kov, B.N.; Kalinnikov, V.A.; Krastev, V.R.; Maksimov, A.N.; Nikityuk, N.M.

    1985-01-01

    This paper describes a high-speed parallel counter that contains 31 inputs and 15 outputs and is implemented by integrated circuits of series 500. The counter is designed for fast sampling of events according to the number of particles that pass simultaneously through the hodoscopic plane of the detector. The minimum delay of the output signals relative to the input is 43 nsec. The duration of the output signals can be varied from 75 to 120 nsec

  15. An anthropologist in parallel structure

    Directory of Open Access Journals (Sweden)

    Noelle Molé Liston

    2016-08-01

    Full Text Available The essay examines the parallels between Molé Liston’s studies on labor and precarity in Italy and the United States’ anthropology job market. Probing the way economic shift reshaped the field of anthropology of Europe in the late 2000s, the piece explores how the neoliberalization of the American academy increased the value in studying the hardships and daily lives of non-western populations in Europe.

  16. Combinatorics of spreads and parallelisms

    CERN Document Server

    Johnson, Norman

    2010-01-01

    Partitions of Vector Spaces Quasi-Subgeometry Partitions Finite Focal-SpreadsGeneralizing André SpreadsThe Going Up Construction for Focal-SpreadsSubgeometry Partitions Subgeometry and Quasi-Subgeometry Partitions Subgeometries from Focal-SpreadsExtended André SubgeometriesKantor's Flag-Transitive DesignsMaximal Additive Partial SpreadsSubplane Covered Nets and Baer Groups Partial Desarguesian t-Parallelisms Direct Products of Affine PlanesJha-Johnson SL(2,

  17. New algorithms for parallel MRI

    International Nuclear Information System (INIS)

    Anzengruber, S; Ramlau, R; Bauer, F; Leitao, A

    2008-01-01

    Magnetic Resonance Imaging with parallel data acquisition requires algorithms for reconstructing the patient's image from a small number of measured lines of the Fourier domain (k-space). In contrast to well-known algorithms like SENSE and GRAPPA and its flavors we consider the problem as a non-linear inverse problem. However, in order to avoid cost intensive derivatives we will use Landweber-Kaczmarz iteration and in order to improve the overall results some additional sparsity constraints.

  18. Wakefield calculations on parallel computers

    International Nuclear Information System (INIS)

    Schoessow, P.

    1990-01-01

    The use of parallelism in the solution of wakefield problems is illustrated for two different computer architectures (SIMD and MIMD). Results are given for finite difference codes which have been implemented on a Connection Machine and an Alliant FX/8 and which are used to compute wakefields in dielectric loaded structures. Benchmarks on code performance are presented for both cases. 4 refs., 3 figs., 2 tabs

  19. Aspects of computation on asynchronous parallel processors

    International Nuclear Information System (INIS)

    Wright, M.

    1989-01-01

    The increasing availability of asynchronous parallel processors has provided opportunities for original and useful work in scientific computing. However, the field of parallel computing is still in a highly volatile state, and researchers display a wide range of opinion about many fundamental questions such as models of parallelism, approaches for detecting and analyzing parallelism of algorithms, and tools that allow software developers and users to make effective use of diverse forms of complex hardware. This volume collects the work of researchers specializing in different aspects of parallel computing, who met to discuss the framework and the mechanics of numerical computing. The far-reaching impact of high-performance asynchronous systems is reflected in the wide variety of topics, which include scientific applications (e.g. linear algebra, lattice gauge simulation, ordinary and partial differential equations), models of parallelism, parallel language features, task scheduling, automatic parallelization techniques, tools for algorithm development in parallel environments, and system design issues

  20. Parallel processing of genomics data

    Science.gov (United States)

    Agapito, Giuseppe; Guzzi, Pietro Hiram; Cannataro, Mario

    2016-10-01

    The availability of high-throughput experimental platforms for the analysis of biological samples, such as mass spectrometry, microarrays and Next Generation Sequencing, have made possible to analyze a whole genome in a single experiment. Such platforms produce an enormous volume of data per single experiment, thus the analysis of this enormous flow of data poses several challenges in term of data storage, preprocessing, and analysis. To face those issues, efficient, possibly parallel, bioinformatics software needs to be used to preprocess and analyze data, for instance to highlight genetic variation associated with complex diseases. In this paper we present a parallel algorithm for the parallel preprocessing and statistical analysis of genomics data, able to face high dimension of data and resulting in good response time. The proposed system is able to find statistically significant biological markers able to discriminate classes of patients that respond to drugs in different ways. Experiments performed on real and synthetic genomic datasets show good speed-up and scalability.

  1. Climate change

    International Nuclear Information System (INIS)

    Anon.

    1990-01-01

    In this paper, the authors discuss in brief the magnitude and rate of past changes in climate and examine the various factors influencing climate in order to place the potential warming due to increasing greenhouse gas concentrations in context. Feedback mechanisms that can amplify or lessen imposed climate changes are discussed next. The overall sensitivity of climate to changes in forcing is then considered, followed by a discussion of the time-dependent response of the Earth system. The focus is on global temperature as an indicator for the magnitude of climatic change

  2. Parallel processing implementation for the coupled transport of photons and electrons using OpenMP

    Science.gov (United States)

    Doerner, Edgardo

    2016-05-01

    In this work the use of OpenMP to implement the parallel processing of the Monte Carlo (MC) simulation of the coupled transport for photons and electrons is presented. This implementation was carried out using a modified EGSnrc platform which enables the use of the Microsoft Visual Studio 2013 (VS2013) environment, together with the developing tools available in the Intel Parallel Studio XE 2015 (XE2015). The performance study of this new implementation was carried out in a desktop PC with a multi-core CPU, taking as a reference the performance of the original platform. The results were satisfactory, both in terms of scalability as parallelization efficiency.

  3. Parallel implementation of DNA sequences matching algorithms using PWM on GPU architecture.

    Science.gov (United States)

    Sharma, Rahul; Gupta, Nitin; Narang, Vipin; Mittal, Ankush

    2011-01-01

    Positional Weight Matrices (PWMs) are widely used in representation and detection of Transcription Factor Of Binding Sites (TFBSs) on DNA. We implement online PWM search algorithm over parallel architecture. A large PWM data can be processed on Graphic Processing Unit (GPU) systems in parallel which can help in matching sequences at a faster rate. Our method employs extensive usage of highly multithreaded architecture and shared memory of multi-cored GPU. An efficient use of shared memory is required to optimise parallel reduction in CUDA. Our optimised method has a speedup of 230-280x over linear implementation on GPU named GeForce GTX 280.

  4. Parallel algorithms for online trackfinding at PANDA

    Energy Technology Data Exchange (ETDEWEB)

    Bianchi, Ludovico; Ritman, James; Stockmanns, Tobias [IKP, Forschungszentrum Juelich GmbH (Germany); Herten, Andreas [JSC, Forschungszentrum Juelich GmbH (Germany); Collaboration: PANDA-Collaboration

    2016-07-01

    The PANDA experiment, one of the four scientific pillars of the FAIR facility currently in construction in Darmstadt, is a next-generation particle detector that will study collisions of antiprotons with beam momenta of 1.5-15 GeV/c on a fixed proton target. Because of the broad physics scope and the similar signature of signal and background events, PANDA's strategy for data acquisition is to continuously record data from the whole detector and use this global information to perform online event reconstruction and filtering. A real-time rejection factor of up to 1000 must be achieved to match the incoming data rate for offline storage, making all components of the data processing system computationally very challenging. Online particle track identification and reconstruction is an essential step, since track information is used as input in all following phases. Online tracking algorithms must ensure a delicate balance between high tracking efficiency and quality, and minimal computational footprint. For this reason, a massively parallel solution exploiting multiple Graphic Processing Units (GPUs) is under investigation. The talk presents the core concepts of the algorithms being developed for primary trackfinding, along with details of their implementation on GPUs.

  5. Parallel algorithms for testing finite state machines:Generating UIO sequences

    OpenAIRE

    Hierons, RM; Turker, UC

    2016-01-01

    This paper describes an efficient parallel algorithm that uses many-core GPUs for automatically deriving Unique Input Output sequences (UIOs) from Finite State Machines. The proposed algorithm uses the global scope of the GPU's global memory through coalesced memory access and minimises the transfer between CPU and GPU memory. The results of experiments indicate that the proposed method yields considerably better results compared to a single core UIO construction algorithm. Our algorithm is s...

  6. A Parallel Sweeping Preconditioner for Heterogeneous 3D Helmholtz Equations

    KAUST Repository

    Poulson, Jack

    2013-05-02

    A parallelization of a sweeping preconditioner for three-dimensional Helmholtz equations without large cavities is introduced and benchmarked for several challenging velocity models. The setup and application costs of the sequential preconditioner are shown to be O(γ2N4/3) and O(γN logN), where γ(ω) denotes the modestly frequency-dependent number of grid points per perfectly matched layer. Several computational and memory improvements are introduced relative to using black-box sparse-direct solvers for the auxiliary problems, and competitive runtimes and iteration counts are reported for high-frequency problems distributed over thousands of cores. Two open-source packages are released along with this paper: Parallel Sweeping Preconditioner (PSP) and the underlying distributed multifrontal solver, Clique. © 2013 Society for Industrial and Applied Mathematics.

  7. Data-Parallel Mesh Connected Components Labeling and Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Harrison, Cyrus; Childs, Hank; Gaither, Kelly

    2011-04-10

    We present a data-parallel algorithm for identifying and labeling the connected sub-meshes within a domain-decomposed 3D mesh. The identification task is challenging in a distributed-memory parallel setting because connectivity is transitive and the cells composing each sub-mesh may span many or all processors. Our algorithm employs a multi-stage application of the Union-find algorithm and a spatial partitioning scheme to efficiently merge information across processors and produce a global labeling of connected sub-meshes. Marking each vertex with its corresponding sub-mesh label allows us to isolate mesh features based on topology, enabling new analysis capabilities. We briefly discuss two specific applications of the algorithm and present results from a weak scaling study. We demonstrate the algorithm at concurrency levels up to 2197 cores and analyze meshes containing up to 68 billion cells.

  8. Concurrent, parallel, multiphysics coupling in the FACETS project

    Energy Technology Data Exchange (ETDEWEB)

    Cary, J R; Carlsson, J A; Hakim, A H; Kruger, S E; Miah, M; Pletzer, A; Shasharina, S [Tech-X Corporation, 5621 Arapahoe Avenue, Suite A, Boulder, CO 80303 (United States); Candy, J; Groebner, R J [General Atomics (United States); Cobb, J; Fahey, M R [Oak Ridge National Laboratory (United States); Cohen, R H; Epperly, T [Lawrence Livermore National Laboratory (United States); Estep, D J [Colorado State University (United States); Krasheninnikov, S [University of California at San Diego (United States); Malony, A D [ParaTools, Inc (United States); McCune, D C [Princeton Plasma Physics Laboratory (United States); McInnes, L; Balay, S [Argonne National Laboratory (United States); Pankin, A, E-mail: cary@txcorp.co [Lehigh University (United States)

    2009-07-01

    FACETS (Framework Application for Core-Edge Transport Simulations), is now in its third year. The FACETS team has developed a framework for concurrent coupling of parallel computational physics for use on Leadership Class Facilities (LCFs). In the course of the last year, FACETS has tackled many of the difficult problems of moving to parallel, integrated modeling by developing algorithms for coupled systems, extracting legacy applications as components, modifying them to run on LCFs, and improving the performance of all components. The development of FACETS abides by rigorous engineering standards, including cross platform build and test systems, with the latter covering regression, performance, and visualization. In addition, FACETS has demonstrated the ability to incorporate full turbulence computations for the highest fidelity transport computations. Early indications are that the framework, using such computations, scales to multiple tens of thousands of processors. These accomplishments were a result of an interdisciplinary collaboration among computational physics, computer scientists and applied mathematicians on the team.

  9. A scalable implementation of RI-SCF on parallel computers

    International Nuclear Information System (INIS)

    Fruechtl, H.A.; Kendall, R.A.; Harrison, R.J.

    1996-01-01

    In order to avoid the integral bottleneck of conventional SCF calculations, the Resolution of the Identity (RI) method is used to obtain an approximate solution to the Hartree-Fock equations. In this approximation only three-center integrals are needed to build the Fock matrix. It has been implemented as part of the NWChem package of portable and scalable ab initio programs for parallel computers. Utilizing the V-approximation, both the Coulomb and exchange contribution to the Fock matrix can be calculated from a transformed set of three-center integrals which have to be precalculated and stored. A distributed in-core method as well as a disk based implementation have been programmed. Details of the implementation as well as the parallel programming tools used are described. We also give results and timings from benchmark calculations

  10. Performance evaluations of advanced massively parallel platforms based on gyrokinetic toroidal five-dimensional Eulerian code GT5D

    International Nuclear Information System (INIS)

    Idomura, Yasuhiro; Jolliet, Sebastien

    2010-01-01

    A gyrokinetic toroidal five dimensional Eulerian code GT5D is ported on six advanced massively parallel platforms and comprehensive benchmark tests are performed. A parallelisation technique based on physical properties of the gyrokinetic equation is presented. By extending the parallelisation technique with a hybrid parallel model, the scalability of the code is improved on platforms with multi-core processors. In the benchmark tests, a good salability is confirmed up to several thousands cores on every platforms, and the maximum sustained performance of ∼18.6 Tflops is achieved using 16384 cores of BX900. (author)

  11. The core paradox.

    Science.gov (United States)

    Kennedy, G. C.; Higgins, G. H.

    1973-01-01

    Rebuttal of suggestions from various critics attempting to provide an escape from the seeming paradox originated by Higgins and Kennedy's (1971) proposed possibility that the liquid in the outer core was thermally stably stratified and that this stratification might prove a powerful inhibitor to circulation of the outer core fluid of the kind postulated for the generation of the earth's magnetic field. These suggestions are examined and shown to provide no reasonable escape from the core paradox.

  12. Efficient sequential and parallel algorithms for record linkage.

    Science.gov (United States)

    Mamun, Abdullah-Al; Mi, Tian; Aseltine, Robert; Rajasekaran, Sanguthevar

    2014-01-01

    Integrating data from multiple sources is a crucial and challenging problem. Even though there exist numerous algorithms for record linkage or deduplication, they suffer from either large time needs or restrictions on the number of datasets that they can integrate. In this paper we report efficient sequential and parallel algorithms for record linkage which handle any number of datasets and outperform previous algorithms. Our algorithms employ hierarchical clustering algorithms as the basis. A key idea that we use is radix sorting on certain attributes to eliminate identical records before any further processing. Another novel idea is to form a graph that links similar records and find the connected components. Our sequential and parallel algorithms have been tested on a real dataset of 1,083,878 records and synthetic datasets ranging in size from 50,000 to 9,000,000 records. Our sequential algorithm runs at least two times faster, for any dataset, than the previous best-known algorithm, the two-phase algorithm using faster computation of the edit distance (TPA (FCED)). The speedups obtained by our parallel algorithm are almost linear. For example, we get a speedup of 7.5 with 8 cores (residing in a single node), 14.1 with 16 cores (residing in two nodes), and 26.4 with 32 cores (residing in four nodes). We have compared the performance of our sequential algorithm with TPA (FCED) and found that our algorithm outperforms the previous one. The accuracy is the same as that of this previous best-known algorithm.

  13. GPU: the biggest key processor for AI and parallel processing

    Science.gov (United States)

    Baji, Toru

    2017-07-01

    Two types of processors exist in the market. One is the conventional CPU and the other is Graphic Processor Unit (GPU). Typical CPU is composed of 1 to 8 cores while GPU has thousands of cores. CPU is good for sequential processing, while GPU is good to accelerate software with heavy parallel executions. GPU was initially dedicated for 3D graphics. However from 2006, when GPU started to apply general-purpose cores, it was noticed that this architecture can be used as a general purpose massive-parallel processor. NVIDIA developed a software framework Compute Unified Device Architecture (CUDA) that make it possible to easily program the GPU for these application. With CUDA, GPU started to be used in workstations and supercomputers widely. Recently two key technologies are highlighted in the industry. The Artificial Intelligence (AI) and Autonomous Driving Cars. AI requires a massive parallel operation to train many-layers of neural networks. With CPU alone, it was impossible to finish the training in a practical time. The latest multi-GPU system with P100 makes it possible to finish the training in a few hours. For the autonomous driving cars, TOPS class of performance is required to implement perception, localization, path planning processing and again SoC with integrated GPU will play a key role there. In this paper, the evolution of the GPU which is one of the biggest commercial devices requiring state-of-the-art fabrication technology will be introduced. Also overview of the GPU demanding key application like the ones described above will be introduced.

  14. Nuclear reactor core flow baffling

    International Nuclear Information System (INIS)

    Berringer, R.T.

    1979-01-01

    A flow baffling arrangement is disclosed for the core of a nuclear reactor. A plurality of core formers are aligned with the grids of the core fuel assemblies such that the high pressure drop areas in the core are at the same elevations as the high pressure drop areas about the core periphery. The arrangement minimizes core bypass flow, maintains cooling of the structure surrounding the core, and allows the utilization of alternative beneficial components such as neutron reflectors positioned near the core

  15. Sediment Core Laboratory

    Data.gov (United States)

    Federal Laboratory Consortium — FUNCTION: Provides instrumentation and expertise for physical and geoacoustic characterization of marine sediments.DESCRIPTION: The multisensor core logger measures...

  16. Big climate data analysis

    Science.gov (United States)

    Mudelsee, Manfred

    2015-04-01

    The Big Data era has begun also in the climate sciences, not only in economics or molecular biology. We measure climate at increasing spatial resolution by means of satellites and look farther back in time at increasing temporal resolution by means of natural archives and proxy data. We use powerful supercomputers to run climate models. The model output of the calculations made for the IPCC's Fifth Assessment Report amounts to ~650 TB. The 'scientific evolution' of grid computing has started, and the 'scientific revolution' of quantum computing is being prepared. This will increase computing power, and data amount, by several orders of magnitude in the future. However, more data does not automatically mean more knowledge. We need statisticians, who are at the core of transforming data into knowledge. Statisticians notably also explore the limits of our knowledge (uncertainties, that is, confidence intervals and P-values). Mudelsee (2014 Climate Time Series Analysis: Classical Statistical and Bootstrap Methods. Second edition. Springer, Cham, xxxii + 454 pp.) coined the term 'optimal estimation'. Consider the hyperspace of climate estimation. It has many, but not infinite, dimensions. It consists of the three subspaces Monte Carlo design, method and measure. The Monte Carlo design describes the data generating process. The method subspace describes the estimation and confidence interval construction. The measure subspace describes how to detect the optimal estimation method for the Monte Carlo experiment. The envisaged large increase in computing power may bring the following idea of optimal climate estimation into existence. Given a data sample, some prior information (e.g. measurement standard errors) and a set of questions (parameters to be estimated), the first task is simple: perform an initial estimation on basis of existing knowledge and experience with such types of estimation problems. The second task requires the computing power: explore the hyperspace to

  17. Overview of the Force Scientific Parallel Language

    Directory of Open Access Journals (Sweden)

    Gita Alaghband

    1994-01-01

    Full Text Available The Force parallel programming language designed for large-scale shared-memory multiprocessors is presented. The language provides a number of parallel constructs as extensions to the ordinary Fortran language and is implemented as a two-level macro preprocessor to support portability across shared memory multiprocessors. The global parallelism model on which the Force is based provides a powerful parallel language. The parallel constructs, generic synchronization, and freedom from process management supported by the Force has resulted in structured parallel programs that are ported to the many multiprocessors on which the Force is implemented. Two new parallel constructs for looping and functional decomposition are discussed. Several programming examples to illustrate some parallel programming approaches using the Force are also presented.

  18. Parallel kinematics type, kinematics, and optimal design

    CERN Document Server

    Liu, Xin-Jun

    2014-01-01

    Parallel Kinematics- Type, Kinematics, and Optimal Design presents the results of 15 year's research on parallel mechanisms and parallel kinematics machines. This book covers the systematic classification of parallel mechanisms (PMs) as well as providing a large number of mechanical architectures of PMs available for use in practical applications. It focuses on the kinematic design of parallel robots. One successful application of parallel mechanisms in the field of machine tools, which is also called parallel kinematics machines, has been the emerging trend in advanced machine tools. The book describes not only the main aspects and important topics in parallel kinematics, but also references novel concepts and approaches, i.e. type synthesis based on evolution, performance evaluation and optimization based on screw theory, singularity model taking into account motion and force transmissibility, and others.   This book is intended for researchers, scientists, engineers and postgraduates or above with interes...

  19. Applied Parallel Computing Industrial Computation and Optimization

    DEFF Research Database (Denmark)

    Madsen, Kaj; NA NA NA Olesen, Dorte

    Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)......Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)...

  20. Radiation-hard/high-speed parallel optical links

    Energy Technology Data Exchange (ETDEWEB)

    Gan, K.K., E-mail: gan@mps.ohio-state.edu [Department of Physics, The Ohio State University, Columbus, OH 43210 (United States); Buchholz, P.; Heidbrink, S. [Fachbereich Physik, Universität Siegen, Siegen (Germany); Kagan, H.P.; Kass, R.D.; Moore, J.; Smith, D.S. [Department of Physics, The Ohio State University, Columbus, OH 43210 (United States); Vogt, M.; Ziolkowski, M. [Fachbereich Physik, Universität Siegen, Siegen (Germany)

    2016-09-21

    We have designed and fabricated a compact parallel optical engine for transmitting data at 5 Gb/s. The device consists of a 4-channel ASIC driving a VCSEL (Vertical Cavity Surface Emitting Laser) array in an optical package. The ASIC is designed using only core transistors in a 65 nm CMOS process to enhance the radiation-hardness. The ASIC contains an 8-bit DAC to control the bias and modulation currents of the individual channels in the VCSEL array. The performance of the optical engine up at 5 Gb/s is satisfactory.