WorldWideScience

Sample records for macroscale unit processes

  1. Simplified Process for Manufacturing Macroscale Patterns to Enhance Voltage Generation by a Triboelectric Generator

    Directory of Open Access Journals (Sweden)

    Jun-Ho Yang

    2015-11-01

    Full Text Available This paper proposes a simple, yet effective and affordable, manufacturing process to enhance the overall efficiency of voltage generation by a triboelectric generator (TEG using 3D printers for energy-harvesting applications. The proposed method can be classified as macroscale surface patterning, in contrast to micro- and nanoscale patterning of TEG proposed in previous studies. Experiments were conducted using a designed test-bed system that allowed the control of external factors, such as the magnitude and frequency of the frictional force and the relative humidity, and an output voltage increase of up to 67% was obtained from a TEG with macroscale patterns that increased the surface area by 14%. The peak voltage generated by the TEG was as high as 18 V, and the addition of a designed analog circuit that uses no external power enabled storage of a DC voltage of 0.4 V. In comparison with previous methods that employ micro- or nanoscale patterns, the proposed patterning method is faster and more suitable for mass production.

  2. Investigation of Micro- and Macro-Scale Transport Processes for Improved Fuel Cell Performance

    Energy Technology Data Exchange (ETDEWEB)

    Gu, Wenbin

    2015-02-05

    This report documents the work performed by General Motors (GM) under the Cooperative agreement No. DE-EE0000470, “Investigation of Micro- and Macro-Scale Transport Processes for Improved Fuel Cell Performance,” in collaboration with the Penn State University (PSU), University of Tennessee Knoxville (UTK), Rochester Institute of Technology (RIT), and University of Rochester (UR) via subcontracts. The overall objectives of the project are to investigate and synthesize fundamental understanding of transport phenomena at both the macro- and micro-scales for the development of a down-the-channel model that accounts for all transport domains in a broad operating space. GM as a prime contractor focused on cell level experiments and modeling, and the Universities as subcontractors worked toward fundamental understanding of each component and associated interface.

  3. Investigation of Micro- and Macro-Scale Transport Processes for Improved Fuel Cell Performance

    Energy Technology Data Exchange (ETDEWEB)

    Gu, Wenbin [General Motors LLC, Pontiac, MI (United States)

    2014-08-29

    This report documents the work performed by General Motors (GM) under the Cooperative agreement No. DE-EE0000470, “Investigation of Micro- and Macro-Scale Transport Processes for Improved Fuel Cell Performance,” in collaboration with the Penn State University (PSU), University of Tennessee Knoxville (UTK), Rochester Institute of Technology (RIT), and University of Rochester (UR) via subcontracts. The overall objectives of the project are to investigate and synthesize fundamental understanding of transport phenomena at both the macro- and micro-scales for the development of a down-the-channel model that accounts for all transport domains in a broad operating space. GM as a prime contractor focused on cell level experiments and modeling, and the Universities as subcontractors worked toward fundamental understanding of each component and associated interface.

  4. Semi-automated image processing system for micro- to macro-scale analysis of immunohistopathology: application to ischemic brain tissue.

    Science.gov (United States)

    Wu, Chunyan; Zhao, Weizhao; Lin, Baowan; Ginsberg, Myron D

    2005-04-01

    Immunochemical staining techniques are commonly used to assess neuronal, astrocytic and microglial alterations in experimental neuroscience research, and in particular, are applied to tissues from animals subjected to ischemic stroke. Immunoreactivity of brain sections can be measured from digitized immunohistology slides so that quantitative assessment can be carried out by computer-assisted analysis. Conventional methods of analyzing immunohistology are based on image classification techniques applied to a specific anatomic location at high magnification. Such micro-scale localized image analysis limits one for further correlative studies with other imaging modalities on whole brain sections, which are of particular interest in experimental stroke research. This report presents a semi-automated image analysis method that performs convolution-based image classification on micro-scale images, extracts numerical data representing positive immunoreactivity from the processed micro-scale images and creates a corresponding quantitative macro-scale image. The present method utilizes several image-processing techniques to cope with variances in intensity distribution, as well as artifacts caused by light scattering or heterogeneity of antigen expression, which are commonly encountered in immunohistology. Micro-scale images are composed by a tiling function in a mosaic manner. Image classification is accomplished by the K-means clustering method at the relatively low-magnification micro-scale level in order to increase computation efficiency. The quantitative macro-scale image is suitable for correlative analysis with other imaging modalities. This method was applied to different immunostaining antibodies, such as endothelial barrier antigen (EBA), lectin, and glial fibrillary acidic protein (GFAP), on histology slides from animals subjected to middle cerebral artery occlusion by the intraluminal suture method. Reliability tests show that the results obtained from

  5. The study of the process of alkaline precipitation purification of solutions from silicon macroscales in the ammonium-fluoride processing of beryllium-containing materials

    Directory of Open Access Journals (Sweden)

    Dyachenko Alexander

    2016-01-01

    Full Text Available The ammonium-fluoride method of beryllium materials processing is examined. An analysis of the existing sulfuric-acid and fluoride processing scheme of beryllium-containing concentrates processing is described; advantages of the proposed ammonium-fluoride scheme and possible problems that may occur when testing a new technique are discussed. Studies on determining the effect of silicon macroscales on the behavior of beryllium in fluoride solutions at an increased pH, as well as on establishing the distribution of beryllium and silicon by phases in laboratory testing of the technology on model mixtures, are described. As a result of the studies, it is found that precipitation purification of solutions from silicon with a solution of ammonia is possible at a temperature of 20-25 °C. The paper presents data on the effect of an excessive concentration of ion-fluoride in a solution on the pH level of silicon precipitation; and on the hydrolysis degree of beryllium and silicon during the precipitation.

  6. Strategy to Achieve Highly Porous/Biocompatible Macroscale Cell Blocks, Using a Collagen/Genipin-bioink and an Optimal 3D Printing Process.

    Science.gov (United States)

    Kim, Yong Bok; Lee, Hyeongjin; Kim, Geun Hyung

    2016-11-30

    Recently, a three-dimensional (3D) bioprinting process for obtaining a cell-laden structure has been widely applied because of its ability to fabricate biomimetic complex structures embedded with and without cells. To successfully obtain a cell-laden porous block, the cell-delivering vehicle, bioink, is one of the significant factors. Until now, various biocompatible hydrogels (synthetic and natural biopolymers) have been utilized in the cell-printing process, but a bioink satisfying both biocompatibility and print-ability requirements to achieve a porous structure with reasonable mechanical strength has not been issued. Here, we propose a printing strategy with optimal conditions including a safe cross-linking procedure for obtaining a 3D porous cell block composed of a biocompatible collagen-bioink and genipin, a cross-linking agent. To obtain the optimal processing conditions, we modified the 3D printing machine and selected an optimal cross-linking condition (∼1 mM and 1 h) of genipin solution. To show the feasibility of the process, 3D pore-interconnected cell-laden constructs were manufactured using osteoblast-like cells (MG63) and human adipose stem cells (hASCs). Under these processing conditions, a macroscale 3D collagen-based cell block of 21 × 21 × 12 mm(3) and over 95% cell viability was obtained. In vitro biological testing of the cell-laden 3D porous structure showed that the embedded cells were sufficiently viable, and their proliferation was significantly higher; the cells also exhibited increased osteogenic activities compared to the conventional alginate-based bioink (control). The results indicated the fabrication process using the collagen-bioink would be an innovative platform to design highly biocompatible and mechanically stable cell blocks.

  7. Role of 3D force networks in linking grain scale to macroscale processes in sheared granular debris

    Science.gov (United States)

    Mair, K.; Jettestuen, E.; Abe, S.

    2013-12-01

    Active faults, landslides and subglacial tills contain accumulations of granular debris that evolve during sliding. The macroscopic motion in these environments is at least to some extent determined by processes operating in this sheared granular material. A valid question is how the local behavior at the individual granular contacts actually sums up to influence macroscopic sliding. Laboratory experiments and numerical modeling can potentially help elucidate this. Observations of jamming (stick) and unjamming (flow) as well as concentrated shear bands on the scale of 5-10 grains suggest that a simple continuum description may be insufficient to capture important elements of the behavior. We therefore seek a measure of the organization of the granular fabric and the 3D structure of the load bearing skeleton that effectively demonstrates how the individual grain interactions are manifested in the macroscopic sliding behavior we observe. Contact force networks are an expression of this. Here we investigate the structure and variability of the most connected system spanning force networks produced in 3D discrete element models of granular layers under shear. We use percolation measures to identify, characterize, compare and track the evolution of these strongly connected contact force networks. We show that specific topological measures used in describing the networks, such as number of contacts and coordination number, are sensitive to grain size distribution (and likely the grain shape) of the material as well as loading conditions. Hence, faults of different maturity would be expected to accommodate shear in different ways. Distinct changes in the topological characteristics i.e. the geometry of strong force networks with accumulated strain are directly correlated to fluctuations in macroscopic shearing resistance. This suggests that 3D force networks play an important bridging role between individual grain scale processes and macroscopic sliding behavior.

  8. Meso-Scale and Macro-Scale Analysis of the Geochemical and Physical Processes Responsible for Acid Rock Drainage

    Science.gov (United States)

    Otwinowski, Matthew

    1997-08-01

    We have developed a computer model which describes the geochemical and physical processes responsible for acid drainage from waste rock piles. The model is in the form of coupled nonlinear PDEs which describe: the kinetics of the chemical reactions, the release of contaminants, the generation of energy due to the exothermic oxidation of sulphides, the diffusive and convective transport of oxygen and water, and the transport of energy by conduction and convection. The meso-scale and large-scale characterization of waste rock and waste rock piles is discussed. We show that long-term leaching rates are inversely proportional to the square of particle diameter and that the previously used models underestimate the particle size effect on long-term sulphide oxidation. Experimental data on rock fragmentation are used for a fractal statistical characterization of waste rock piles. The acid generation rates, oxygen consumption rates and temperature profiles have been determined for piles containing from fifty thousand to five hundred thousand tonnes of waste rock. The thermodynamic instabilities, which occur at certain critical values of pile height, are responsible for thermodynamic catastrophes which result in a rapid increase of acid generation rates. The critical height is determined by the values of sulphide concentration, particle size, pile porosity and other factors. The numerical code is based on the finite elements method with an adaptive grid generator. abstract.

  9. Signal processing unit

    Energy Technology Data Exchange (ETDEWEB)

    Boswell, J.

    1983-01-01

    The architecture of the signal processing unit (SPU) comprises an ROM connected to a program bus, and an input-output bus connected to a data bus and register through a pipeline multiplier accumulator (pmac) and a pipeline arithmetic logic unit (palu), each associated with a random access memory (ram1,2). The system pulse frequency is from 20 mhz. The pmac is further detailed, and has a capability of 20 mega operations per second. There is also a block diagram for the palu, showing interconnections between the register block (rbl), separator for bus (bs), register (reg), shifter (sh) and combination unit. The first and second rams have formats 64*16 and 32*32 bits, respectively. Further data are a 5-v power supply and 2.5 micron n-channel silicon gate mos technology with about 50000 transistors.

  10. THOR Particle Processing Unit PPU

    Science.gov (United States)

    Federica Marcucci, Maria; Bruno, Roberto; Consolini, Giuseppe; D'Amicis, Raffaella; De Lauretis, Marcello; De Marco, Rossana; De Michelis, Paola; Francia, Patrizia; Laurenza, Monica; Materassi, Massimo; Vellante, Massimo; Valentini, Francesco

    2016-04-01

    Turbulence Heating ObserveR (THOR) is the first mission ever flown in space dedicated to plasma turbulence. On board THOR, data collected by the Turbulent Electron Analyser, the Ion Mass Spectrum analyser and the Cold Solar Wind ion analyser instruments will be processed by a common digital processor unit, the Particle Processing Unit (PPU). PPU architecture will be based on the state of the art space flight processors and will be fully redundant, in order to efficiently and safely handle the data from the numerous sensors of the instruments suite. The approach of a common processing unit for particle instruments is very important for the enabling of an efficient management for correlative plasma measurements, also facilitating interoperation with other instruments on the spacecraft. Moreover, it permits technical and programmatic synergies giving the possibility to optimize and save spacecraft resources.

  11. Micro- to macroscale perspectives on space plasmas

    Science.gov (United States)

    Eastman, Timothy E.

    1993-01-01

    The Earth's magnetosphere is the most accessible of natural collisionless plasma environments; an astrophysical plasma 'laboratory'. Magnetospheric physics has been in an exploration phase since its origin 35 years ago but new coordinated, multipoint observations, theory, modeling, and simulations are moving this highly interdisciplinary field of plasma science into a new phase of synthesis and understanding. Plasma systems are ones in which binary collisions are relatively negligible and collective behavior beyond the microscale emerges. Most readily accessible natural plasma systems are collisional and nearest-neighbor classical interactions compete with longer-range plasma effects. Except for stars, most space plasmas are collisionless, however, and the effects of electrodynamic coupling dominate. Basic physical processes in such collisionless plasmas occur at micro-, meso-, and macroscales that are not merely reducible to each other in certain crucial ways as illustrated for the global coupling of the Earth's magnetosphere and for the nonlinear dynamics of charged particle motion in the magnetotail. Such global coupling and coherence makes the geospace environment, the domain of solar-terrestrial science, the most highly coupled of all physical geospheres.

  12. Linking macroscale graph analytical organization to microscale neuroarchitectonics in the macaque connectome.

    Science.gov (United States)

    Scholtens, Lianne H; Schmidt, Ruben; de Reus, Marcel A; van den Heuvel, Martijn P

    2014-09-03

    Macroscale connectivity of the mammalian brain has been shown to display several characteristics of an efficient communication network architecture. In parallel, at the microscopic scale, histological studies have extensively revealed large interregional variation in cortical neural architectonics. However, how these two "scales" of cerebrum organization are linked remains an open question. Collating and combining data across multiple studies on the cortical cytoarchitecture of the macaque cortex with information on macroscale anatomical wiring derived from tract tracing studies, this study focuses on examining the interplay between macroscale organization of the macaque connectome and microscale cortical neuronal architecture. Our findings show that both macroscale degree as well as the topological role in the overall network are related to the level of neuronal complexity of cortical regions at the microscale, showing (among several effects) a positive overall association between macroscale degree and metrics of microscale pyramidal complexity. Macroscale hub regions, together forming a densely interconnected "rich club," are noted to display a high level of neuronal complexity, findings supportive of a high level of integrative neuronal processes to occur in these regions. Together, we report on cross-scale observations that jointly suggest that a region's microscale neuronal architecture is tuned to its role in the global brain network. Copyright © 2014 the authors 0270-6474/14/3412192-14$15.00/0.

  13. Multi-unit Integration in Microfluidic Processes: Current Status and Future Horizons

    Directory of Open Access Journals (Sweden)

    Pratap R. Patnaik

    2011-07-01

    Full Text Available Microfluidic processes, mainly for biological and chemical applications, have expanded rapidly in recent years. While the initial focus was on single units, principally microreactors, technological and economic considerations have caused a shift to integrated microchips in which a number of microdevices function coherently. These integrated devices have many advantages over conventional macro-scale processes. However, the small scale of operation, complexities in the underlying physics and chemistry, and differences in the time constants of the participating units, in the interactions among them and in the outputs of interest make it difficult to design and optimize integrated microprocesses. These aspects are discussed here, current research and applications are reviewed, and possible future directions are considered.

  14. Local biogeomorphic feedbacks and macroscale drivers shape coastal wetland distributions

    Science.gov (United States)

    Braswell, A. E.; Heffernan, J. B.

    2016-12-01

    Recent models have demonstrated that lateral biogeomorphic processes are important for the persistence of coastal wetlands in the face of sea level rise and other anthropogenic pressures. Yet empirical studies of marsh ecomorphodynamics have largely focused on vertical accretion. Moreover, local vertical and lateral processes of marsh-building depend on external sediment supply and the wave energy environment, and thus are connected to macroscale characteristics such as estuarine morphology and watershed size. These broad scale drivers, combined with local biogeomorphic feedbacks within wetlands, determine wetland extent. Our goal is to understand the scales at which local biogeomorphic feedbacks and macroscale estuarine and watershed characteristics influence the distribution of coastal marshes. To that end, we examined the distribution of wetland extent and its potential watershed and estuarine drivers at multiple scales along the Atlantic and Gulf coasts, USA. Using existing GIS resources, we delineated extents of coastal wetlands, and generated proxies of sediment input, estuarine energy, and human alteration. We found that distributions of wetland extent were bi-modal at the finest scale of our analysis (approx. 1-100 km2), a finding that is consistent with theoretical models of local marsh feedbacks. At larger spatial scales, distributions of marsh extent were associated with both estuarine size and drainage ratio. These relationships indicate that sediment supply and erosion ultimately constrain the extent of marsh development and persistence, while local feedbacks operate at smaller scales. Our findings support and extend theory and observation at the scale of marsh platforms and lagoons, but also demonstrate the importance of macroscale watershed and estuarine characteristics for wetland establishment and persistence.

  15. Monitoring of frozen soil hydrology in macro-scale in the Qinghai-Xizang Plateau

    Institute of Scientific and Technical Information of China (English)

    2000-01-01

    Monitoring of frozen soil hydrology in macro-scale was performed by Chinese and Japanese scientists from 1997 to 1998. Quality measured data were obtained. Measured data on soil moisture and temperature are preliminarily analyzed. Based on profiles of soil temperature and moisture in individual measured sites, intra-annual freezing and melting process of soil is discussed. Maximum frozen and thawed depths and frozen days in various depths are estimated. The work emphasized the spatial distribution on soil temperature and moisture in macro-scale and the effect of topography on conditions of soil water and heat.

  16. Cortical chemoarchitecture shapes macroscale effective functional connectivity patterns in macaque cerebral cortex

    NARCIS (Netherlands)

    Turk, Elise; Scholtens, Lianne H.; van den Heuvel, Martijn P.|info:eu-repo/dai/nl/304820466

    2016-01-01

    The mammalian cortex is a complex system of-at the microscale level-interconnected neurons and-at the macroscale level-interconnected areas, forming the infrastructure for local and global neural processing and information integration. While the effects of regional chemoarchitecture on local cortica

  17. ON DEVELOPING CLEANER ORGANIC UNIT PROCESSES

    Science.gov (United States)

    Organic waste products, potentially harmful to the human health and the environment, are primarily produced in the synthesis stage of manufacturing processes. Many such synthetic unit processes, such as halogenation, oxidation, alkylation, nitration, and sulfonation are common to...

  18. Data Sorting Using Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    M. J. Mišić

    2012-06-01

    Full Text Available Graphics processing units (GPUs have been increasingly used for general-purpose computation in recent years. The GPU accelerated applications are found in both scientific and commercial domains. Sorting is considered as one of the very important operations in many applications, so its efficient implementation is essential for the overall application performance. This paper represents an effort to analyze and evaluate the implementations of the representative sorting algorithms on the graphics processing units. Three sorting algorithms (Quicksort, Merge sort, and Radix sort were evaluated on the Compute Unified Device Architecture (CUDA platform that is used to execute applications on NVIDIA graphics processing units. Algorithms were tested and evaluated using an automated test environment with input datasets of different characteristics. Finally, the results of this analysis are briefly discussed.

  19. Analysis and Optimization of Central Processing Unit Process Parameters

    Science.gov (United States)

    Kaja Bantha Navas, R.; Venkata Chaitana Vignan, Budi; Durganadh, Margani; Rama Krishna, Chunduri

    2017-05-01

    The rapid growth of computer has made processing more data capable, which increase the heat dissipation. Hence the system unit CPU must be cooled against operating temperature. This paper presents a novel approach for the optimization of operating parameters on Central Processing Unit with single response based on response graph method. These methods have a series of steps from of proposed approach which are capable of decreasing uncertainty caused by engineering judgment in the Taguchi method. Orthogonal Array value was taken from ANSYS report. The method shows a good convergence with the experimental and the optimum process parameters.

  20. Quantum Central Processing Unit and Quantum Algorithm

    Institute of Scientific and Technical Information of China (English)

    王安民

    2002-01-01

    Based on a scalable and universal quantum network, quantum central processing unit, proposed in our previous paper [Chin. Phys. Left. 18 (2001)166], the whole quantum network for the known quantum algorithms,including quantum Fourier transformation, Shor's algorithm and Grover's algorithm, is obtained in a unitied way.

  1. Syllables as Processing Units in Handwriting Production

    Science.gov (United States)

    Kandel, Sonia; Alvarez, Carlos J.; Vallee, Nathalie

    2006-01-01

    This research focused on the syllable as a processing unit in handwriting. Participants wrote, in uppercase letters, words that had been visually presented. The interletter intervals provide information on the timing of motor production. In Experiment 1, French participants wrote words that shared the initial letters but had different syllable…

  2. Graphics processing unit-assisted lossless decompression

    Science.gov (United States)

    Loughry, Thomas A.

    2016-04-12

    Systems and methods for decompressing compressed data that has been compressed by way of a lossless compression algorithm are described herein. In a general embodiment, a graphics processing unit (GPU) is programmed to receive compressed data packets and decompress such packets in parallel. The compressed data packets are compressed representations of an image, and the lossless compression algorithm is a Rice compression algorithm.

  3. Graphics processing unit-assisted lossless decompression

    Energy Technology Data Exchange (ETDEWEB)

    Loughry, Thomas A.

    2016-04-12

    Systems and methods for decompressing compressed data that has been compressed by way of a lossless compression algorithm are described herein. In a general embodiment, a graphics processing unit (GPU) is programmed to receive compressed data packets and decompress such packets in parallel. The compressed data packets are compressed representations of an image, and the lossless compression algorithm is a Rice compression algorithm.

  4. Characteristics of soil water retention curve at macro-scale

    Institute of Scientific and Technical Information of China (English)

    2009-01-01

    Scale adaptable hydrological models have attracted more and more attentions in the hydrological modeling research community, and the constitutive relationship at the macro-scale is one of the most important issues, upon which there are not enough research activities yet. Taking the constitutive relationships of soil water movement--soil water retention curve (SWRC) as an example, this study extends the definition of SWRC at the micro-scale to that at the macro-scale, and aided by Monte Carlo method we demonstrate that soil property and the spatial distribution of soil moisture will affect the features of SWRC greatly. Furthermore, we assume that the spatial distribution of soil moisture is the result of self-organization of climate, soil, ground water and soil water movement under the specific boundary conditions, and we also carry out numerical experiments of soil water movement at the vertical direction in order to explore the relationship between SWRC at the macro-scale and the combinations of climate, soil, and groundwater. The results show that SWRCs at the macro-scale and micro-scale presents totally different features, e.g., the essential hysteresis phenomenon which is exaggerated with increasing aridity index and rising groundwater table. Soil property plays an important role in the shape of SWRC which will even lead to a rectangular shape under drier conditions, and power function form of SWRC widely adopted in hydrological model might be revised for most situations at the macro-scale.

  5. Numerical Integration with Graphical Processing Unit for QKD Simulation

    Science.gov (United States)

    2014-03-27

    existing and proposed Quantum Key Distribution (QKD) systems. This research investigates using graphical processing unit ( GPU ) technology to more...Time Pad GPU graphical processing unit API application programming interface CUDA Compute Unified Device Architecture SIMD single-instruction-stream...and can be passed by value or reference [2]. 2.3 Graphical Processing Units Programming with graphical processing unit ( GPU ) requires a different

  6. Macroscale hydrologic modeling of ecologically relevant flow metrics

    Science.gov (United States)

    Seth J. Wenger; Charles H. Luce; Alan F. Hamlet; Daniel J. Isaak; Helen M. Neville

    2010-01-01

    Stream hydrology strongly affects the structure of aquatic communities. Changes to air temperature and precipitation driven by increased greenhouse gas concentrations are shifting timing and volume of streamflows potentially affecting these communities. The variable infiltration capacity (VIC) macroscale hydrologic model has been employed at regional scales to describe...

  7. A Spotlight on Bridging Microscale and Macroscale Human Brain Architecture

    NARCIS (Netherlands)

    van den Heuvel, Martijn P; Yeo, B T Thomas

    2017-01-01

    We place a spotlight on the emerging trend of jointly studying the micro- and macroscale organization of nervous systems. We discuss the pioneering studies of Ding et al. (2016) and Glasser et al. (2016) in the context of growing efforts to combine and integrate multiple features of brain

  8. Temperature of the Central Processing Unit

    Directory of Open Access Journals (Sweden)

    Ivan Lavrov

    2016-10-01

    Full Text Available Heat is inevitably generated in the semiconductors during operation. Cooling in a computer, and in its main part – the Central Processing Unit (CPU, is crucial, allowing the proper functioning without overheating, malfunctioning, and damage. In order to estimate the temperature as a function of time, it is important to solve the differential equations describing the heat flow and to understand how it depends on the physical properties of the system. This project aims to answer these questions by considering a simplified model of the CPU + heat sink. A similarity with the electrical circuit and certain methods from electrical circuit analysis are discussed.

  9. Graphics Processing Unit Assisted Thermographic Compositing

    Science.gov (United States)

    Ragasa, Scott; McDougal, Matthew; Russell, Sam

    2013-01-01

    Objective: To develop a software application utilizing general purpose graphics processing units (GPUs) for the analysis of large sets of thermographic data. Background: Over the past few years, an increasing effort among scientists and engineers to utilize the GPU in a more general purpose fashion is allowing for supercomputer level results at individual workstations. As data sets grow, the methods to work them grow at an equal, and often greater, pace. Certain common computations can take advantage of the massively parallel and optimized hardware constructs of the GPU to allow for throughput that was previously reserved for compute clusters. These common computations have high degrees of data parallelism, that is, they are the same computation applied to a large set of data where the result does not depend on other data elements. Signal (image) processing is one area were GPUs are being used to greatly increase the performance of certain algorithms and analysis techniques.

  10. Relativistic hydrodynamics on graphics processing units

    CERN Document Server

    Sikorski, Jan; Porter-Sobieraj, Joanna; Słodkowski, Marcin; Krzyżanowski, Piotr; Książek, Natalia; Duda, Przemysław

    2016-01-01

    Hydrodynamics calculations have been successfully used in studies of the bulk properties of the Quark-Gluon Plasma, particularly of elliptic flow and shear viscosity. However, there are areas (for instance event-by-event simulations for flow fluctuations and higher-order flow harmonics studies) where further advancement is hampered by lack of efficient and precise 3+1D~program. This problem can be solved by using Graphics Processing Unit (GPU) computing, which offers unprecedented increase of the computing power compared to standard CPU simulations. In this work, we present an implementation of 3+1D ideal hydrodynamics simulations on the Graphics Processing Unit using Nvidia CUDA framework. MUSTA-FORCE (MUlti STAge, First ORder CEntral, with a~slope limiter and MUSCL reconstruction) and WENO (Weighted Essentially Non-Oscillating) schemes are employed in the simulations, delivering second (MUSTA-FORCE), fifth and seventh (WENO) order of accuracy. Third order Runge-Kutta scheme was used for integration in the t...

  11. Accelerating the Fourier split operator method via graphics processing units

    CERN Document Server

    Bauke, Heiko

    2010-01-01

    Current generations of graphics processing units have turned into highly parallel devices with general computing capabilities. Thus, graphics processing units may be utilized, for example, to solve time dependent partial differential equations by the Fourier split operator method. In this contribution, we demonstrate that graphics processing units are capable to calculate fast Fourier transforms much more efficiently than traditional central processing units. Thus, graphics processing units render efficient implementations of the Fourier split operator method possible. Performance gains of more than an order of magnitude as compared to implementations for traditional central processing units are reached in the solution of the time dependent Schr\\"odinger equation and the time dependent Dirac equation.

  12. Direct Numerical Simulations in Solid Mechanics for Quantifying the Macroscale Effects of Microstructure and Material Model-Form Error

    Science.gov (United States)

    Bishop, Joseph E.; Emery, John M.; Battaile, Corbett C.; Littlewood, David J.; Baines, Andrew J.

    2016-05-01

    Two fundamental approximations in macroscale solid-mechanics modeling are (1) the assumption of scale separation in homogenization theory and (2) the use of a macroscopic plasticity material model that represents, in a mean sense, the multitude of inelastic processes occurring at the microscale. With the goal of quantifying the errors induced by these approximations on engineering quantities of interest, we perform a set of direct numerical simulations (DNS) in which polycrystalline microstructures are embedded throughout a macroscale structure. The largest simulations model over 50,000 grains. The microstructure is idealized using a randomly close-packed Voronoi tessellation in which each polyhedral Voronoi cell represents a grain. An face centered cubic crystal-plasticity model is used to model the mechanical response of each grain. The overall grain structure is equiaxed, and each grain is randomly oriented with no overall texture. The detailed results from the DNS simulations are compared to results obtained from conventional macroscale simulations that use homogeneous isotropic plasticity models. The macroscale plasticity models are calibrated using a representative volume element of the idealized microstructure. Ultimately, we envision that DNS modeling will be used to gain new insights into the mechanics of material deformation and failure.

  13. Magnetohydrodynamics simulations on graphics processing units

    CERN Document Server

    Wong, Hon-Cheng; Feng, Xueshang; Tang, Zesheng

    2009-01-01

    Magnetohydrodynamics (MHD) simulations based on the ideal MHD equations have become a powerful tool for modeling phenomena in a wide range of applications including laboratory, astrophysical, and space plasmas. In general, high-resolution methods for solving the ideal MHD equations are computationally expensive and Beowulf clusters or even supercomputers are often used to run the codes that implemented these methods. With the advent of the Compute Unified Device Architecture (CUDA), modern graphics processing units (GPUs) provide an alternative approach to parallel computing for scientific simulations. In this paper we present, to the authors' knowledge, the first implementation to accelerate computation of MHD simulations on GPUs. Numerical tests have been performed to validate the correctness of our GPU MHD code. Performance measurements show that our GPU-based implementation achieves speedups of 2 (1D problem with 2048 grids), 106 (2D problem with 1024^2 grids), and 43 (3D problem with 128^3 grids), respec...

  14. Graphics Processing Units for HEP trigger systems

    Science.gov (United States)

    Ammendola, R.; Bauce, M.; Biagioni, A.; Chiozzi, S.; Cotta Ramusino, A.; Fantechi, R.; Fiorini, M.; Giagu, S.; Gianoli, A.; Lamanna, G.; Lonardo, A.; Messina, A.; Neri, I.; Paolucci, P. S.; Piandani, R.; Pontisso, L.; Rescigno, M.; Simula, F.; Sozzi, M.; Vicini, P.

    2016-07-01

    General-purpose computing on GPUs (Graphics Processing Units) is emerging as a new paradigm in several fields of science, although so far applications have been tailored to the specific strengths of such devices as accelerator in offline computation. With the steady reduction of GPU latencies, and the increase in link and memory throughput, the use of such devices for real-time applications in high-energy physics data acquisition and trigger systems is becoming ripe. We will discuss the use of online parallel computing on GPU for synchronous low level trigger, focusing on CERN NA62 experiment trigger system. The use of GPU in higher level trigger system is also briefly considered.

  15. Kernel density estimation using graphical processing unit

    Science.gov (United States)

    Sunarko, Su'ud, Zaki

    2015-09-01

    Kernel density estimation for particles distributed over a 2-dimensional space is calculated using a single graphical processing unit (GTX 660Ti GPU) and CUDA-C language. Parallel calculations are done for particles having bivariate normal distribution and by assigning calculations for equally-spaced node points to each scalar processor in the GPU. The number of particles, blocks and threads are varied to identify favorable configuration. Comparisons are obtained by performing the same calculation using 1, 2 and 4 processors on a 3.0 GHz CPU using MPICH 2.0 routines. Speedups attained with the GPU are in the range of 88 to 349 times compared the multiprocessor CPU. Blocks of 128 threads are found to be the optimum configuration for this case.

  16. Graphics Processing Units for HEP trigger systems

    Energy Technology Data Exchange (ETDEWEB)

    Ammendola, R. [INFN Sezione di Roma “Tor Vergata”, Via della Ricerca Scientifica 1, 00133 Roma (Italy); Bauce, M. [INFN Sezione di Roma “La Sapienza”, P.le A. Moro 2, 00185 Roma (Italy); University of Rome “La Sapienza”, P.lee A.Moro 2, 00185 Roma (Italy); Biagioni, A. [INFN Sezione di Roma “La Sapienza”, P.le A. Moro 2, 00185 Roma (Italy); Chiozzi, S.; Cotta Ramusino, A. [INFN Sezione di Ferrara, Via Saragat 1, 44122 Ferrara (Italy); University of Ferrara, Via Saragat 1, 44122 Ferrara (Italy); Fantechi, R. [INFN Sezione di Pisa, Largo B. Pontecorvo 3, 56127 Pisa (Italy); CERN, Geneve (Switzerland); Fiorini, M. [INFN Sezione di Ferrara, Via Saragat 1, 44122 Ferrara (Italy); University of Ferrara, Via Saragat 1, 44122 Ferrara (Italy); Giagu, S. [INFN Sezione di Roma “La Sapienza”, P.le A. Moro 2, 00185 Roma (Italy); University of Rome “La Sapienza”, P.lee A.Moro 2, 00185 Roma (Italy); Gianoli, A. [INFN Sezione di Ferrara, Via Saragat 1, 44122 Ferrara (Italy); University of Ferrara, Via Saragat 1, 44122 Ferrara (Italy); Lamanna, G., E-mail: gianluca.lamanna@cern.ch [INFN Sezione di Pisa, Largo B. Pontecorvo 3, 56127 Pisa (Italy); INFN Laboratori Nazionali di Frascati, Via Enrico Fermi 40, 00044 Frascati (Roma) (Italy); Lonardo, A. [INFN Sezione di Roma “La Sapienza”, P.le A. Moro 2, 00185 Roma (Italy); Messina, A. [INFN Sezione di Roma “La Sapienza”, P.le A. Moro 2, 00185 Roma (Italy); University of Rome “La Sapienza”, P.lee A.Moro 2, 00185 Roma (Italy); and others

    2016-07-11

    General-purpose computing on GPUs (Graphics Processing Units) is emerging as a new paradigm in several fields of science, although so far applications have been tailored to the specific strengths of such devices as accelerator in offline computation. With the steady reduction of GPU latencies, and the increase in link and memory throughput, the use of such devices for real-time applications in high-energy physics data acquisition and trigger systems is becoming ripe. We will discuss the use of online parallel computing on GPU for synchronous low level trigger, focusing on CERN NA62 experiment trigger system. The use of GPU in higher level trigger system is also briefly considered.

  17. Energy Efficient Iris Recognition With Graphics Processing Units

    National Research Council Canada - National Science Library

    Rakvic, Ryan; Broussard, Randy; Ngo, Hau

    2016-01-01

    .... In the past few years, however, this growth has slowed for central processing units (CPUs). Instead, there has been a shift to multicore computing, specifically with the general purpose graphic processing units (GPUs...

  18. Abiotic Controls on Macroscale Variations of Humid Tropical Forest Height

    Directory of Open Access Journals (Sweden)

    Yan Yang

    2016-06-01

    Full Text Available Spatial variation of tropical forest tree height is a key indicator of ecological processes associated with forest growth and carbon dynamics. Here we examine the macroscale variations of tree height of humid tropical forests across three continents and quantify the climate and edaphic controls on these variations. Forest tree heights are systematically sampled across global humid tropical forests with more than 2.5 million measurements from Geoscience Laser Altimeter System (GLAS satellite observations (2004–2008. We used top canopy height (TCH of GLAS footprints to grid the statistical mean and variance and the 90 percentile height of samples at 0.5 degrees to capture the regional variability of average and large trees globally. We used the spatial regression method (spatial eigenvector mapping-SEVM to evaluate the contributions of climate, soil and topography in explaining and predicting the regional variations of forest height. Statistical models suggest that climate, soil, topography, and spatial contextual information together can explain more than 60% of the observed forest height variation, while climate and soil jointly explain 30% of the height variations. Soil basics, including physical compositions such as clay and sand contents, chemical properties such as PH values and cation-exchange capacity, as well as biological variables such as the depth of organic matter, all present independent but statistically significant relationships to forest height across three continents. We found significant relations between the precipitation and tree height with shorter trees on the average in areas of higher annual water stress, and large trees occurring in areas with low stress and higher annual precipitation but with significant differences across the continents. Our results confirm other landscape and regional studies by showing that soil fertility, topography and climate may jointly control a significant variation of forest height and

  19. Micro and Macroscale Drivers of Nutrient Concentrations in Urban Streams in South, Central and North America

    Science.gov (United States)

    Loiselle, Steven A.; Gasparini Fernandes Cunha, Davi; Shupe, Scott; Valiente, Elsa; Rocha, Luciana; Heasley, Eleanore; Belmont, Patricia Pérez; Baruch, Avinoam

    2016-01-01

    Global metrics of land cover and land use provide a fundamental basis to examine the spatial variability of human-induced impacts on freshwater ecosystems. However, microscale processes and site specific conditions related to bank vegetation, pollution sources, adjacent land use and water uses can have important influences on ecosystem conditions, in particular in smaller tributary rivers. Compared to larger order rivers, these low-order streams and rivers are more numerous, yet often under-monitored. The present study explored the relationship of nutrient concentrations in 150 streams in 57 hydrological basins in South, Central and North America (Buenos Aires, Curitiba, São Paulo, Rio de Janeiro, Mexico City and Vancouver) with macroscale information available from global datasets and microscale data acquired by trained citizen scientists. Average sub-basin phosphate (P-PO4) concentrations were found to be well correlated with sub-basin attributes on both macro and microscales, while the relationships between sub-basin attributes and nitrate (N-NO3) concentrations were limited. A phosphate threshold for eutrophic conditions (>0.1 mg L-1 P-PO4) was exceeded in basins where microscale point source discharge points (eg. residential, industrial, urban/road) were identified in more than 86% of stream reaches monitored by citizen scientists. The presence of bankside vegetation covaried (rho = –0.53) with lower phosphate concentrations in the ecosystems studied. Macroscale information on nutrient loading allowed for a strong separation between basins with and without eutrophic conditions. Most importantly, the combination of macroscale and microscale information acquired increased our ability to explain sub-basin variability of P-PO4 concentrations. The identification of microscale point sources and bank vegetation conditions by citizen scientists provided important information that local authorities could use to improve their management of lower order river

  20. A new global river network database for macroscale hydrologic modeling

    Science.gov (United States)

    Wu, Huan; Kimball, John S.; Li, Hongyi; Huang, Maoyi; Leung, L. Ruby; Adler, Robert F.

    2012-09-01

    Coarse-resolution (upscaled) river networks are critical inputs for runoff routing in macroscale hydrologic models. Recently, Wu et al. (2011) developed a hierarchical dominant river tracing (DRT) algorithm for automated extraction and spatial upscaling of river networks using fine-scale hydrography inputs. We applied the DRT algorithms using combined HydroSHEDS and HYDRO1k global fine-scale hydrography inputs and produced a new series of upscaled global river network data at multiple (1/16° to 2°) spatial resolutions. The new upscaled results are internally consistent and congruent with the baseline fine-scale inputs and should facilitate improved regional to global scale hydrologic simulations.

  1. Quantum manifestation of systems on the macro-scale – the concept of transition state and transition amplitude wave

    Indian Academy of Sciences (India)

    Ram K Varma

    2007-06-01

    Quantum effects which have usually been associated with micro-scale phenomena can also arise on the macro-scale in situations other than the well-known macro-quantum phenomena of superconductivity and superfluidity. Such situations have been shown here to arise in processes involving inelastic scattering with bound or partially bound systems (not bound in all degrees of freedom), and the macro-quantum behaviour is associated with the state of the total system in transition in the process of scattering. Such a state is designated as a `transition-state'. It is pointed out that we have already observed such manifestations for a particular system, the charged particles in a magnetic field where interference effects involving macro-scale matter waves along the magnetic field have been reported [R K Varma et al, Phys. Rev. E65, 026503 (2002)].

  2. Integration Process for the Habitat Demonstration Unit

    Science.gov (United States)

    Gill, Tracy; Merbitz, Jerad; Kennedy, Kriss; Tn, Terry; Toups, Larry; Howe, A. Scott; Smitherman, David

    2011-01-01

    The Habitat Demonstration Unit (HDU) is an experimental exploration habitat technology and architecture test platform designed for analog demonstration activities. The HDU previously served as a test bed for testing technologies and sub-systems in a terrestrial surface environment. in 2010 in the Pressurized Excursion Module (PEM) configuration. Due to the amount of work involved to make the HDU project successful, the HDU project has required a team to integrate a variety of contributions from NASA centers and outside collaborators The size of the team and number of systems involved With the HDU makes Integration a complicated process. However, because the HDU shell manufacturing is complete, the team has a head start on FY--11 integration activities and can focus on integrating upgrades to existing systems as well as integrating new additions. To complete the development of the FY-11 HDU from conception to rollout for operations in July 2011, a cohesive integration strategy has been developed to integrate the various systems of HDU and the payloads. The highlighted HDU work for FY-11 will focus on performing upgrades to the PEM configuration, adding the X-Hab as a second level, adding a new porch providing the astronauts a larger work area outside the HDU for EVA preparations, and adding a Hygiene module. Together these upgrades result in a prototype configuration of the Deep Space Habitat (DSH), an element under evaluation by NASA's Human Exploration Framework Team (HEFT) Scheduled activates include early fit-checks and the utilization of a Habitat avionics test bed prior to installation into HDU. A coordinated effort to utilize modeling and simulation systems has aided in design and integration concept development. Modeling tools have been effective in hardware systems layout, cable routing, sub-system interface length estimation and human factors analysis. Decision processes on integration and use of all new subsystems will be defined early in the project to

  3. Mesh-particle interpolations on graphics processing units and multicore central processing units.

    Science.gov (United States)

    Rossinelli, Diego; Conti, Christian; Koumoutsakos, Petros

    2011-06-13

    Particle-mesh interpolations are fundamental operations for particle-in-cell codes, as implemented in vortex methods, plasma dynamics and electrostatics simulations. In these simulations, the mesh is used to solve the field equations and the gradients of the fields are used in order to advance the particles. The time integration of particle trajectories is performed through an extensive resampling of the flow field at the particle locations. The computational performance of this resampling turns out to be limited by the memory bandwidth of the underlying computer architecture. We investigate how mesh-particle interpolation can be efficiently performed on graphics processing units (GPUs) and multicore central processing units (CPUs), and we present two implementation techniques. The single-precision results for the multicore CPU implementation show an acceleration of 45-70×, depending on system size, and an acceleration of 85-155× for the GPU implementation over an efficient single-threaded C++ implementation. In double precision, we observe a performance improvement of 30-40× for the multicore CPU implementation and 20-45× for the GPU implementation. With respect to the 16-threaded standard C++ implementation, the present CPU technique leads to a performance increase of roughly 2.8-3.7× in single precision and 1.7-2.4× in double precision, whereas the GPU technique leads to an improvement of 9× in single precision and 2.2-2.8× in double precision.

  4. Modeling of biopharmaceutical processes. Part 2: Process chromatography unit operation

    DEFF Research Database (Denmark)

    Kaltenbrunner, Oliver; McCue, Justin; Engel, Philip;

    2008-01-01

    Process modeling can be a useful tool to aid in process development, process optimization, and process scale-up. When modeling a chromatography process, one must first select the appropriate models that describe the mass transfer and adsorption that occurs within the porous adsorbent...

  5. Proton Testing of Advanced Stellar Compass Digital Processing Unit

    DEFF Research Database (Denmark)

    Thuesen, Gøsta; Denver, Troelz; Jørgensen, Finn E

    1999-01-01

    The Advanced Stellar Compass Digital Processing Unit was radiation tested with 300 MeV protons at Proton Irradiation Facility (PIF), Paul Scherrer Institute, Switzerland.......The Advanced Stellar Compass Digital Processing Unit was radiation tested with 300 MeV protons at Proton Irradiation Facility (PIF), Paul Scherrer Institute, Switzerland....

  6. Optimized Technology for Residuum Processing in the ARGG Unit

    Institute of Scientific and Technical Information of China (English)

    Pan Luoqi; Yuan hongxing; Nie Baiqiu

    2006-01-01

    The influence of feedstock property on operation in the FCC unit was studied to identify the cause leading to deteriorated products distribution related with increasingly heavier feedstock for the ARGG unit. In order to maximize the economic benefits of the ARGG unit a string of measures, including the modification of catalyst formulation, retention of high catalyst activity, application of mixed termination agents to control the reaction temperature and once-through operation, and optimization of catalyst regeneration technique, were adopted to adapt the ARGG unit to processing of the heavy feedstock with its carbon residue equating to 7% on an average. The heavy oil processing technology has brought about apparent economic benefits.

  7. Parallelization of heterogeneous reactor calculations on a graphics processing unit

    Energy Technology Data Exchange (ETDEWEB)

    Malofeev, V. M., E-mail: vm-malofeev@mail.ru; Pal’shin, V. A. [National Research Center Kurchatov Institute (Russian Federation)

    2016-12-15

    Parallelization is applied to the neutron calculations performed by the heterogeneous method on a graphics processing unit. The parallel algorithm of the modified TREC code is described. The efficiency of the parallel algorithm is evaluated.

  8. Diffusion tensor fiber tracking on graphics processing units.

    Science.gov (United States)

    Mittmann, Adiel; Comunello, Eros; von Wangenheim, Aldo

    2008-10-01

    Diffusion tensor magnetic resonance imaging has been successfully applied to the process of fiber tracking, which determines the location of fiber bundles within the human brain. This process, however, can be quite lengthy when run on a regular workstation. We present a means of executing this process by making use of the graphics processing units of computers' video cards, which provide a low-cost parallel execution environment that algorithms like fiber tracking can benefit from. With this method we have achieved performance gains varying from 14 to 40 times on common computers. Because of accuracy issues inherent to current graphics processing units, we define a variation index in order to assess how close the results obtained with our method are to those generated by programs running on the central processing units of computers. This index shows that results produced by our method are acceptable when compared to those of traditional programs.

  9. Business Process Compliance through Reusable Units of Compliant Processes

    NARCIS (Netherlands)

    Shumm, D.; Turetken, O.; Kokash, N.; Elgammal, A.; Leymann, F.; Heuvel, J. van den

    2010-01-01

    Compliance management is essential for ensuring that organizational business processes and supporting information systems are in accordance with a set of prescribed requirements originating from laws, regulations, and various legislative or technical documents such as Sarbanes-Oxley Act or ISO 17799

  10. High Input Voltage, Silicon Carbide Power Processing Unit Performance Demonstration

    Science.gov (United States)

    Bozak, Karin E.; Pinero, Luis R.; Scheidegger, Robert J.; Aulisio, Michael V.; Gonzalez, Marcelo C.; Birchenough, Arthur G.

    2015-01-01

    A silicon carbide brassboard power processing unit has been developed by the NASA Glenn Research Center in Cleveland, Ohio. The power processing unit operates from two sources: a nominal 300 Volt high voltage input bus and a nominal 28 Volt low voltage input bus. The design of the power processing unit includes four low voltage, low power auxiliary supplies, and two parallel 7.5 kilowatt (kW) discharge power supplies that are capable of providing up to 15 kilowatts of total power at 300 to 500 Volts (V) to the thruster. Additionally, the unit contains a housekeeping supply, high voltage input filter, low voltage input filter, and master control board, such that the complete brassboard unit is capable of operating a 12.5 kilowatt Hall effect thruster. The performance of the unit was characterized under both ambient and thermal vacuum test conditions, and the results demonstrate exceptional performance with full power efficiencies exceeding 97%. The unit was also tested with a 12.5kW Hall effect thruster to verify compatibility and output filter specifications. With space-qualified silicon carbide or similar high voltage, high efficiency power devices, this would provide a design solution to address the need for high power electric propulsion systems.

  11. Adaptive-optics Optical Coherence Tomography Processing Using a Graphics Processing Unit*

    Science.gov (United States)

    Shafer, Brandon A.; Kriske, Jeffery E.; Kocaoglu, Omer P.; Turner, Timothy L.; Liu, Zhuolin; Lee, John Jaehwan; Miller, Donald T.

    2015-01-01

    Graphics processing units are increasingly being used for scientific computing for their powerful parallel processing abilities, and moderate price compared to super computers and computing grids. In this paper we have used a general purpose graphics processing unit to process adaptive-optics optical coherence tomography (AOOCT) images in real time. Increasing the processing speed of AOOCT is an essential step in moving the super high resolution technology closer to clinical viability. PMID:25570838

  12. Adaptive-optics optical coherence tomography processing using a graphics processing unit.

    Science.gov (United States)

    Shafer, Brandon A; Kriske, Jeffery E; Kocaoglu, Omer P; Turner, Timothy L; Liu, Zhuolin; Lee, John Jaehwan; Miller, Donald T

    2014-01-01

    Graphics processing units are increasingly being used for scientific computing for their powerful parallel processing abilities, and moderate price compared to super computers and computing grids. In this paper we have used a general purpose graphics processing unit to process adaptive-optics optical coherence tomography (AOOCT) images in real time. Increasing the processing speed of AOOCT is an essential step in moving the super high resolution technology closer to clinical viability.

  13. Precise Nanoscale Surface Modification and Coating of Macroscale Objects: Open-Environment in Loco Atomic Layer Deposition on an Automobile.

    Science.gov (United States)

    Mousa, Moataz Bellah M; Oldham, Christopher J; Parsons, Gregory N

    2015-09-09

    The fundamental chemical reaction conditions that define atomic layer deposition (ALD) can be achieved in an open environment on a macroscale surface too large and complex for typical laboratory reactor-based ALD. We describe the concept of in loco ALD using conventional modulated reactant flow through a surface-mounted "ALD delivery head" to form a precise nanoscale Al2O3 film on the window of a parked automobile. Analysis confirms that the processes eliminated ambient water contamination and met other conditions that define ALD growth. Using this tool, we demonstrate open-ambient patterned deposition, metal corrosion protection, and polymer surface modification.

  14. Present and future contribution of glacier storage change to runoff from macroscale drainage basins in Europe

    National Research Council Canada - National Science Library

    Matthias Huss

    2011-01-01

      Glaciers make a significant runoff contribution in macroscale drainage basins The impact of glacial melt water is recognizable with very small glacierization The retreat of alpine glaciers plays...

  15. Advancing Tissue Engineering: A Tale of Nano-, Micro-, and Macroscale Integration

    NARCIS (Netherlands)

    Leijten, Jeroen Christianus Hermanus; Rouwkema, Jeroen; Zhang, Y.S.; Nasajpour, A.; Dokmeci, M.R.; Khademhosseini, A.

    2016-01-01

    Tissue engineering has the potential to revolutionize the health care industry. Delivering on this promise requires the generation of efficient, controllable and predictable implants. The integration of nano- and microtechnologies into macroscale regenerative biomaterials plays an essential role in

  16. Unit Operations for the Food Industry: Equilibrium Processes & Mechanical Operations

    OpenAIRE

    Guiné, Raquel

    2013-01-01

    Unit operations are an area of engineering that is at the same time very fascinating and most essential for the industry in general and the food industry in particular. This book was prepared in a way to achieve simultaneously the academic and practical perspectives. It is organized into two parts: the unit operations based on equilibrium processes and the mechanical operations. Each topic starts with a presentation of the fundamental concepts and principles, followed by a discussion of ...

  17. Formalizing the Process of Constructing Chains of Lexical Units

    Directory of Open Access Journals (Sweden)

    Grigorij Chetverikov

    2015-06-01

    Full Text Available Formalizing the Process of Constructing Chains of Lexical Units The paper investigates mathematical aspects of describing the construction of chains of lexical units on the basis of finite-predicate algebra. Analyzing the construction peculiarities is carried out and application of the method of finding the power of linear logical transformation for removing characteristic words of a dictionary entry is given. Analysis and perspectives of the results of the study are provided.

  18. Modeling Units of Assessment for Sharing Assessment Process Information: towards an Assessment Process Specification

    NARCIS (Netherlands)

    Miao, Yongwu; Sloep, Peter; Koper, Rob

    2009-01-01

    Miao, Y., Sloep, P. B., & Koper, R. (2008). Modeling Units of Assessment for Sharing Assessment Process Information: towards an Assessment Process Specification. Presentation at the ICWL 2008 conference. August, 20, 2008, Jinhua, China.

  19. COST ESTIMATION MODELS FOR DRINKING WATER TREATMENT UNIT PROCESSES

    Science.gov (United States)

    Cost models for unit processes typically utilized in a conventional water treatment plant and in package treatment plant technology are compiled in this paper. The cost curves are represented as a function of specified design parameters and are categorized into four major catego...

  20. Determinants of profitability of smallholder palm oil processing units ...

    African Journals Online (AJOL)

    ... of profitability of smallholder palm oil processing units in Ogun state, Nigeria. ... as well as their geographical spread covering the entire land space of the state. ... The F-ratio value is statistically significant (P<0.01) implying that the model is ...

  1. Reflector antenna analysis using physical optics on Graphics Processing Units

    DEFF Research Database (Denmark)

    Borries, Oscar Peter; Sørensen, Hans Henrik Brandenborg; Dammann, Bernd

    2014-01-01

    The Physical Optics approximation is a widely used asymptotic method for calculating the scattering from electrically large bodies. It requires significant computational work and little memory, and is thus well suited for application on a Graphics Processing Unit. Here, we investigate...... the performance of an implementation and demonstrate that while there are some implementational pitfalls, a careful implementation can result in impressive improvements....

  2. Utilizing Graphics Processing Units for Network Anomaly Detection

    Science.gov (United States)

    2012-09-13

    matching system using deterministic finite automata and extended finite automata resulting in a speedup of 9x over the CPU implementation [SGO09]. Kovach ...pages 14–18, 2009. [Kov10] Nicholas S. Kovach . Accelerating malware detection via a graphics processing unit, 2010. http://www.dtic.mil/dtic/tr

  3. Acceleration of option pricing technique on graphics processing units

    NARCIS (Netherlands)

    Zhang, B.; Oosterlee, C.W.

    2010-01-01

    The acceleration of an option pricing technique based on Fourier cosine expansions on the Graphics Processing Unit (GPU) is reported. European options, in particular with multiple strikes, and Bermudan options will be discussed. The influence of the number of terms in the Fourier cosine series expan

  4. Acceleration of option pricing technique on graphics processing units

    NARCIS (Netherlands)

    Zhang, B.; Oosterlee, C.W.

    2014-01-01

    The acceleration of an option pricing technique based on Fourier cosine expansions on the graphics processing unit (GPU) is reported. European options, in particular with multiple strikes, and Bermudan options will be discussed. The influence of the number of terms in the Fourier cosine series expan

  5. Point process models for household distributions within small areal units

    Directory of Open Access Journals (Sweden)

    Zack W. Almquist

    2012-06-01

    Full Text Available Spatio-demographic data sets are increasingly available worldwide, permitting ever more realistic modeling and analysis of social processes ranging from mobility to disease trans- mission. The information provided by these data sets is typically aggregated by areal unit, for reasons of both privacy and administrative cost. Unfortunately, such aggregation does not permit fine-grained assessment of geography at the level of individual households. In this paper, we propose to partially address this problem via the development of point pro- cess models that can be used to effectively simulate the location of individual households within small areal units.

  6. Turbulent velocity and concentration measurements in a macro-scale multi-inlet vortex nanoprecipitation reactor

    Science.gov (United States)

    Liu, Zhenping; Fox, Rodney; Hill, James; Olsen, Michael

    2013-11-01

    Flash Nanoprecipitation (FNP) is a technique to produce monodisperse functional nanoparticles. Microscale multi-inlet vortex reactors (MIVR) have been effectively applied to FNP due to their ability to provide rapid mixing and flexibility of inlet flow conditions. A scaled-up MIVR could potentially generate large quantities of functional nanoparticles, giving FNP wider applicability in industry. In the presented research, the turbulent velocity field inside a scaled-up, macroscale MIVR is measured by particle image velocimetry (PIV). Within the reactor, velocity is measured using both two-dimensional and stereoscopic PIV at two Reynolds numbers (3500 and 8750) based on the flow at each inlet. Data have been collected at numerous locations in the inlet channels, the reaction chamber, and the reactor outlet. Mean velocity and Reynolds stresses have been obtained based on 5000 instantaneous velocity realizations at each measurement location. The turbulent mixing process has also been investigated with passive scalar planar laser-induced fluorescence and simultaneous PIV/PLIF. Velocity and concentration results are compared to results from previous experiments in a microscale MIVR. Scaled profiles of turbulent quantities are similar to those previously found in the microscale MIVR.

  7. Accelerated space object tracking via graphic processing unit

    Science.gov (United States)

    Jia, Bin; Liu, Kui; Pham, Khanh; Blasch, Erik; Chen, Genshe

    2016-05-01

    In this paper, a hybrid Monte Carlo Gauss mixture Kalman filter is proposed for the continuous orbit estimation problem. Specifically, the graphic processing unit (GPU) aided Monte Carlo method is used to propagate the uncertainty of the estimation when the observation is not available and the Gauss mixture Kalman filter is used to update the estimation when the observation sequences are available. A typical space object tracking problem using the ground radar is used to test the performance of the proposed algorithm. The performance of the proposed algorithm is compared with the popular cubature Kalman filter (CKF). The simulation results show that the ordinary CKF diverges in 5 observation periods. In contrast, the proposed hybrid Monte Carlo Gauss mixture Kalman filter achieves satisfactory performance in all observation periods. In addition, by using the GPU, the computational time is over 100 times less than that using the conventional central processing unit (CPU).

  8. Ising Processing Units: Potential and Challenges for Discrete Optimization

    Energy Technology Data Exchange (ETDEWEB)

    Coffrin, Carleton James [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Nagarajan, Harsha [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Bent, Russell Whitford [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2017-07-05

    The recent emergence of novel computational devices, such as adiabatic quantum computers, CMOS annealers, and optical parametric oscillators, presents new opportunities for hybrid-optimization algorithms that leverage these kinds of specialized hardware. In this work, we propose the idea of an Ising processing unit as a computational abstraction for these emerging tools. Challenges involved in using and bench- marking these devices are presented, and open-source software tools are proposed to address some of these challenges. The proposed benchmarking tools and methodology are demonstrated by conducting a baseline study of established solution methods to a D-Wave 2X adiabatic quantum computer, one example of a commercially available Ising processing unit.

  9. A Universal Quantum Network Quantum Central Processing Unit

    Institute of Scientific and Technical Information of China (English)

    WANG An-Min

    2001-01-01

    A new construction scheme of a universal quantum network which is compatible with the known quantum gate- assembly schemes is proposed. Our quantum network is standard, easy-assemble, reusable, scalable and even potentially programmable. Moreover, we can construct a whole quantum network to implement the generalquantum algorithm and quantum simulation procedure. In the above senses, it is a realization of the quantum central processing unit.

  10. Accelerating Malware Detection via a Graphics Processing Unit

    Science.gov (United States)

    2010-09-01

    Processing Unit . . . . . . . . . . . . . . . . . . 4 PE Portable Executable . . . . . . . . . . . . . . . . . . . . . 4 COFF Common Object File Format...operating systems for the future [Szo05]. The PE format is an updated version of the common object file format ( COFF ) [Mic06]. Microsoft released a new...pro.mspx, Accessed July 2010, 2001. 79 Mic06. Microsoft. Common object file format ( coff ). MSDN, November 2006. Re- vision 4.1. Mic07a. Microsoft

  11. An Architecture of Deterministic Quantum Central Processing Unit

    OpenAIRE

    Xue, Fei; Chen, Zeng-Bing; Shi, Mingjun; Zhou, Xianyi; Du, Jiangfeng; Han, Rongdian

    2002-01-01

    We present an architecture of QCPU(Quantum Central Processing Unit), based on the discrete quantum gate set, that can be programmed to approximate any n-qubit computation in a deterministic fashion. It can be built efficiently to implement computations with any required accuracy. QCPU makes it possible to implement universal quantum computation with a fixed, general purpose hardware. Thus the complexity of the quantum computation can be put into the software rather than the hardware.

  12. BitTorrent Processing Unit BPU发展观望

    Institute of Scientific and Technical Information of China (English)

    Zone; 杨原青

    2007-01-01

    在电脑发展的早期,无论是运算处理、还是图形处理、还是输入、输出处理,都由CPU(Central Processing Unit,中央处理器)一力承担,然而随着处理专用化发展,1999年NVIDIA率先将图形处理独立出来,提出了GPU(Graphics Processing unit,绘图处理单元)概念。八年过去,现在GPU已经成为图形处理的中坚力量,并让所玩家耳熟能详。而近期,台湾2家公刊则提出了BPU(BitTorrent Processing Unit,BT处理单元)概念。下面,就让我们一起看看,这款极为新鲜的概念产品。

  13. Optimized Laplacian image sharpening algorithm based on graphic processing unit

    Science.gov (United States)

    Ma, Tinghuai; Li, Lu; Ji, Sai; Wang, Xin; Tian, Yuan; Al-Dhelaan, Abdullah; Al-Rodhaan, Mznah

    2014-12-01

    In classical Laplacian image sharpening, all pixels are processed one by one, which leads to large amount of computation. Traditional Laplacian sharpening processed on CPU is considerably time-consuming especially for those large pictures. In this paper, we propose a parallel implementation of Laplacian sharpening based on Compute Unified Device Architecture (CUDA), which is a computing platform of Graphic Processing Units (GPU), and analyze the impact of picture size on performance and the relationship between the processing time of between data transfer time and parallel computing time. Further, according to different features of different memory, an improved scheme of our method is developed, which exploits shared memory in GPU instead of global memory and further increases the efficiency. Experimental results prove that two novel algorithms outperform traditional consequentially method based on OpenCV in the aspect of computing speed.

  14. Modeling Units of Assessment for Sharing Assessment Process Information: towards an Assessment Process Specification

    NARCIS (Netherlands)

    Miao, Yongwu; Sloep, Peter; Koper, Rob

    2008-01-01

    Miao, Y., Sloep, P. B., & Koper, R. (2008). Modeling Units of Assessment for Sharing Assessment Process Information: towards an Assessment Process Specification. In F. W. B. Li, J. Zhao, T. K. Shih, R. W. H. Lau, Q. Li & D. McLeod (Eds.), Advances in Web Based Learning - Proceedings of the 7th

  15. Fast calculation of HELAS amplitudes using graphics processing unit (GPU)

    CERN Document Server

    Hagiwara, K; Okamura, N; Rainwater, D L; Stelzer, T

    2009-01-01

    We use the graphics processing unit (GPU) for fast calculations of helicity amplitudes of physics processes. As our first attempt, we compute $u\\overline{u}\\to n\\gamma$ ($n=2$ to 8) processes in $pp$ collisions at $\\sqrt{s} = 14$TeV by transferring the MadGraph generated HELAS amplitudes (FORTRAN) into newly developed HEGET ({\\bf H}ELAS {\\bf E}valuation with {\\bf G}PU {\\bf E}nhanced {\\bf T}echnology) codes written in CUDA, a C-platform developed by NVIDIA for general purpose computing on the GPU. Compared with the usual CPU programs, we obtain 40-150 times better performance on the GPU.

  16. Product- and Process Units in the CRITT Translation Process Research Database

    DEFF Research Database (Denmark)

    Carl, Michael

    The first version of the "Translation Process Research Database" (TPR DB v1.0) was released In August 2012, containing logging data of more than 400 translation and text production sessions. The current version of the TPR DB, (v1.4), contains data from more than 940 sessions, which represents more...... than 300 hours of text production. The database provides the raw logging data, as well as Tables of pre-processed product- and processing units. The TPR-DB includes various types of simple and composed product and process units that are intended to support the analysis and modelling of human text...... reception, production, and translation processes. In this talk I describe some of the functions and features of the TPR-DB v1.4, and how they can be deployed in empirical human translation process research....

  17. Use of general purpose graphics processing units with MODFLOW.

    Science.gov (United States)

    Hughes, Joseph D; White, Jeremy T

    2013-01-01

    To evaluate the use of general-purpose graphics processing units (GPGPUs) to improve the performance of MODFLOW, an unstructured preconditioned conjugate gradient (UPCG) solver has been developed. The UPCG solver uses a compressed sparse row storage scheme and includes Jacobi, zero fill-in incomplete, and modified-incomplete lower-upper (LU) factorization, and generalized least-squares polynomial preconditioners. The UPCG solver also includes options for sequential and parallel solution on the central processing unit (CPU) using OpenMP. For simulations utilizing the GPGPU, all basic linear algebra operations are performed on the GPGPU; memory copies between the central processing unit CPU and GPCPU occur prior to the first iteration of the UPCG solver and after satisfying head and flow criteria or exceeding a maximum number of iterations. The efficiency of the UPCG solver for GPGPU and CPU solutions is benchmarked using simulations of a synthetic, heterogeneous unconfined aquifer with tens of thousands to millions of active grid cells. Testing indicates GPGPU speedups on the order of 2 to 8, relative to the standard MODFLOW preconditioned conjugate gradient (PCG) solver, can be achieved when (1) memory copies between the CPU and GPGPU are optimized, (2) the percentage of time performing memory copies between the CPU and GPGPU is small relative to the calculation time, (3) high-performance GPGPU cards are utilized, and (4) CPU-GPGPU combinations are used to execute sequential operations that are difficult to parallelize. Furthermore, UPCG solver testing indicates GPGPU speedups exceed parallel CPU speedups achieved using OpenMP on multicore CPUs for preconditioners that can be easily parallelized. Published 2013. This article is a U.S. Government work and is in the public domain in the USA.

  18. Fast analytical scatter estimation using graphics processing units.

    Science.gov (United States)

    Ingleby, Harry; Lippuner, Jonas; Rickey, Daniel W; Li, Yue; Elbakri, Idris

    2015-01-01

    To develop a fast patient-specific analytical estimator of first-order Compton and Rayleigh scatter in cone-beam computed tomography, implemented using graphics processing units. The authors developed an analytical estimator for first-order Compton and Rayleigh scatter in a cone-beam computed tomography geometry. The estimator was coded using NVIDIA's CUDA environment for execution on an NVIDIA graphics processing unit. Performance of the analytical estimator was validated by comparison with high-count Monte Carlo simulations for two different numerical phantoms. Monoenergetic analytical simulations were compared with monoenergetic and polyenergetic Monte Carlo simulations. Analytical and Monte Carlo scatter estimates were compared both qualitatively, from visual inspection of images and profiles, and quantitatively, using a scaled root-mean-square difference metric. Reconstruction of simulated cone-beam projection data of an anthropomorphic breast phantom illustrated the potential of this method as a component of a scatter correction algorithm. The monoenergetic analytical and Monte Carlo scatter estimates showed very good agreement. The monoenergetic analytical estimates showed good agreement for Compton single scatter and reasonable agreement for Rayleigh single scatter when compared with polyenergetic Monte Carlo estimates. For a voxelized phantom with dimensions 128 × 128 × 128 voxels and a detector with 256 × 256 pixels, the analytical estimator required 669 seconds for a single projection, using a single NVIDIA 9800 GX2 video card. Accounting for first order scatter in cone-beam image reconstruction improves the contrast to noise ratio of the reconstructed images. The analytical scatter estimator, implemented using graphics processing units, provides rapid and accurate estimates of single scatter and with further acceleration and a method to account for multiple scatter may be useful for practical scatter correction schemes.

  19. Congestion estimation technique in the optical network unit registration process.

    Science.gov (United States)

    Kim, Geunyong; Yoo, Hark; Lee, Dongsoo; Kim, Youngsun; Lim, Hyuk

    2016-07-01

    We present a congestion estimation technique (CET) to estimate the optical network unit (ONU) registration success ratio for the ONU registration process in passive optical networks. An optical line terminal (OLT) estimates the number of collided ONUs via the proposed scheme during the serial number state. The OLT can obtain congestion level among ONUs to be registered such that this information may be exploited to change the size of a quiet window to decrease the collision probability. We verified the efficiency of the proposed method through simulation and experimental results.

  20. Heterogeneous Multicore Parallel Programming for Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Francois Bodin

    2009-01-01

    Full Text Available Hybrid parallel multicore architectures based on graphics processing units (GPUs can provide tremendous computing power. Current NVIDIA and AMD Graphics Product Group hardware display a peak performance of hundreds of gigaflops. However, exploiting GPUs from existing applications is a difficult task that requires non-portable rewriting of the code. In this paper, we present HMPP, a Heterogeneous Multicore Parallel Programming workbench with compilers, developed by CAPS entreprise, that allows the integration of heterogeneous hardware accelerators in a unintrusive manner while preserving the legacy code.

  1. Porting a Hall MHD Code to a Graphic Processing Unit

    Science.gov (United States)

    Dorelli, John C.

    2011-01-01

    We present our experience porting a Hall MHD code to a Graphics Processing Unit (GPU). The code is a 2nd order accurate MUSCL-Hancock scheme which makes use of an HLL Riemann solver to compute numerical fluxes and second-order finite differences to compute the Hall contribution to the electric field. The divergence of the magnetic field is controlled with Dedner?s hyperbolic divergence cleaning method. Preliminary benchmark tests indicate a speedup (relative to a single Nehalem core) of 58x for a double precision calculation. We discuss scaling issues which arise when distributing work across multiple GPUs in a CPU-GPU cluster.

  2. Line-by-line spectroscopic simulations on graphics processing units

    Science.gov (United States)

    Collange, Sylvain; Daumas, Marc; Defour, David

    2008-01-01

    We report here on software that performs line-by-line spectroscopic simulations on gases. Elaborate models (such as narrow band and correlated-K) are accurate and efficient for bands where various components are not simultaneously and significantly active. Line-by-line is probably the most accurate model in the infrared for blends of gases that contain high proportions of H 2O and CO 2 as this was the case for our prototype simulation. Our implementation on graphics processing units sustains a speedup close to 330 on computation-intensive tasks and 12 on memory intensive tasks compared to implementations on one core of high-end processors. This speedup is due to data parallelism, efficient memory access for specific patterns and some dedicated hardware operators only available in graphics processing units. It is obtained leaving most of processor resources available and it would scale linearly with the number of graphics processing units in parallel machines. Line-by-line simulation coupled with simulation of fluid dynamics was long believed to be economically intractable but our work shows that it could be done with some affordable additional resources compared to what is necessary to perform simulations on fluid dynamics alone. Program summaryProgram title: GPU4RE Catalogue identifier: ADZY_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADZY_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 62 776 No. of bytes in distributed program, including test data, etc.: 1 513 247 Distribution format: tar.gz Programming language: C++ Computer: x86 PC Operating system: Linux, Microsoft Windows. Compilation requires either gcc/g++ under Linux or Visual C++ 2003/2005 and Cygwin under Windows. It has been tested using gcc 4.1.2 under Ubuntu Linux 7.04 and using Visual C

  3. Active microchannel fluid processing unit and method of making

    Science.gov (United States)

    Bennett, Wendy D [Kennewick, WA; Martin, Peter M [Kennewick, WA; Matson, Dean W [Kennewick, WA; Roberts, Gary L [West Richland, WA; Stewart, Donald C [Richland, WA; Tonkovich, Annalee Y [Pasco, WA; Zilka, Jennifer L [Pasco, WA; Schmitt, Stephen C [Dublin, OH; Werner, Timothy M [Columbus, OH

    2001-01-01

    The present invention is an active microchannel fluid processing unit and method of making, both relying on having (a) at least one inner thin sheet; (b) at least one outer thin sheet; (c) defining at least one first sub-assembly for performing at least one first unit operation by stacking a first of the at least one inner thin sheet in alternating contact with a first of the at least one outer thin sheet into a first stack and placing an end block on the at least one inner thin sheet, the at least one first sub-assembly having at least a first inlet and a first outlet; and (d) defining at least one second sub-assembly for performing at least one second unit operation either as a second flow path within the first stack or by stacking a second of the at least one inner thin sheet in alternating contact with second of the at least one outer thin sheet as a second stack, the at least one second sub-assembly having at least a second inlet and a second outlet.

  4. Steady electrodiffusion in hydrogel-colloid composites: macroscale properties from microscale electrokinetics

    OpenAIRE

    Hill, Reghan J.

    2010-01-01

    A rigorous microscale electrokinetic model for hydrogel-colloid composites is adopted to compute macroscale profiles of electrolyte concentration, electrostatic potential, and hydrostatic pressure across membranes that separate electrolytes with different concentrations. The membranes are uncharged polymeric hydrogels in which charged spherical colloidal particles are immobilized and randomly dispersed with a low solid volume fraction. Bulk membrane characteristics and performance are calcula...

  5. Accelerating Radio Astronomy Cross-Correlation with Graphics Processing Units

    CERN Document Server

    Clark, M A; Greenhill, L J

    2011-01-01

    We present a highly parallel implementation of the cross-correlation of time-series data using graphics processing units (GPUs), which is scalable to hundreds of independent inputs and suitable for the processing of signals from "Large-N" arrays of many radio antennas. The computational part of the algorithm, the X-engine, is implementated efficiently on Nvidia's Fermi architecture, sustaining up to 79% of the peak single precision floating-point throughput. We compare performance obtained for hardware- and software-managed caches, observing significantly better performance for the latter. The high performance reported involves use of a multi-level data tiling strategy in memory and use of a pipelined algorithm with simultaneous computation and transfer of data from host to device memory. The speed of code development, flexibility, and low cost of the GPU implementations compared to ASIC and FPGA implementations have the potential to greatly shorten the cycle of correlator development and deployment, for case...

  6. Significantly reducing registration time in IGRT using graphics processing units

    DEFF Research Database (Denmark)

    Noe, Karsten Østergaard; Denis de Senneville, Baudouin; Tanderup, Kari

    2008-01-01

    Purpose/Objective For online IGRT, rapid image processing is needed. Fast parallel computations using graphics processing units (GPUs) have recently been made more accessible through general purpose programming interfaces. We present a GPU implementation of the Horn and Schunck method...... respiration phases in a free breathing volunteer and 41 anatomical landmark points in each image series. The registration method used is a multi-resolution GPU implementation of the 3D Horn and Schunck algorithm. It is based on the CUDA framework from Nvidia. Results On an Intel Core 2 CPU at 2.4GHz each...... registration took 30 minutes. On an Nvidia Geforce 8800GTX GPU in the same machine this registration took 37 seconds, making the GPU version 48.7 times faster. The nine image series of different respiration phases were registered to the same reference image (full inhale). Accuracy was evaluated on landmark...

  7. Fast free-form deformation using graphics processing units.

    Science.gov (United States)

    Modat, Marc; Ridgway, Gerard R; Taylor, Zeike A; Lehmann, Manja; Barnes, Josephine; Hawkes, David J; Fox, Nick C; Ourselin, Sébastien

    2010-06-01

    A large number of algorithms have been developed to perform non-rigid registration and it is a tool commonly used in medical image analysis. The free-form deformation algorithm is a well-established technique, but is extremely time consuming. In this paper we present a parallel-friendly formulation of the algorithm suitable for graphics processing unit execution. Using our approach we perform registration of T1-weighted MR images in less than 1 min and show the same level of accuracy as a classical serial implementation when performing segmentation propagation. This technology could be of significant utility in time-critical applications such as image-guided interventions, or in the processing of large data sets. Copyright 2009 Elsevier Ireland Ltd. All rights reserved.

  8. Conceptual Design for the Pilot-Scale Plutonium Oxide Processing Unit in the Radiochemical Processing Laboratory

    Energy Technology Data Exchange (ETDEWEB)

    Lumetta, Gregg J.; Meier, David E.; Tingey, Joel M.; Casella, Amanda J.; Delegard, Calvin H.; Edwards, Matthew K.; Jones, Susan A.; Rapko, Brian M.

    2014-08-05

    This report describes a conceptual design for a pilot-scale capability to produce plutonium oxide for use as exercise and reference materials, and for use in identifying and validating nuclear forensics signatures associated with plutonium production. This capability is referred to as the Pilot-scale Plutonium oxide Processing Unit (P3U), and it will be located in the Radiochemical Processing Laboratory at the Pacific Northwest National Laboratory. The key unit operations are described, including plutonium dioxide (PuO2) dissolution, purification of the Pu by ion exchange, precipitation, and conversion to oxide by calcination.

  9. Exploiting graphics processing units for computational biology and bioinformatics.

    Science.gov (United States)

    Payne, Joshua L; Sinnott-Armstrong, Nicholas A; Moore, Jason H

    2010-09-01

    Advances in the video gaming industry have led to the production of low-cost, high-performance graphics processing units (GPUs) that possess more memory bandwidth and computational capability than central processing units (CPUs), the standard workhorses of scientific computing. With the recent release of generalpurpose GPUs and NVIDIA's GPU programming language, CUDA, graphics engines are being adopted widely in scientific computing applications, particularly in the fields of computational biology and bioinformatics. The goal of this article is to concisely present an introduction to GPU hardware and programming, aimed at the computational biologist or bioinformaticist. To this end, we discuss the primary differences between GPU and CPU architecture, introduce the basics of the CUDA programming language, and discuss important CUDA programming practices, such as the proper use of coalesced reads, data types, and memory hierarchies. We highlight each of these topics in the context of computing the all-pairs distance between instances in a dataset, a common procedure in numerous disciplines of scientific computing. We conclude with a runtime analysis of the GPU and CPU implementations of the all-pairs distance calculation. We show our final GPU implementation to outperform the CPU implementation by a factor of 1700.

  10. The First Prototype for the FastTracker Processing Unit

    CERN Document Server

    Andreani, A; The ATLAS collaboration; Beretta, M; Bogdan, M; Citterio, M; Alberti, F; Giannetti, P; Lanza, A; Magalotti, D; Piendibene, M; Shochet, M; Stabile, A; Tang, J; Tompkins, L

    2012-01-01

    Modern experiments search for extremely rare processes hidden in much larger background levels. As the experiment complexity and the accelerator backgrounds and luminosity increase we need increasingly complex and exclusive selections. We present the first prototype of a new Processing Unit, the core of the FastTracker processor for Atlas, whose computing power is such that a couple of hundreds of them will be able to reconstruct all the tracks with transverse momentum above 1 GeV in the ATLAS events up to Phase II instantaneous luminosities (5×1034 cm-2 s-1) with an event input rate of 100 kHz and a latency below hundreds of microseconds. We plan extremely powerful, very compact and low consumption units for the far future, essential to increase efficiency and purity of the Level 2 selected samples through the intensive use of tracking. This strategy requires massive computing power to minimize the online execution time of complex tracking algorithms. The time consuming pattern recognition problem, generall...

  11. Graphics processing units in bioinformatics, computational biology and systems biology.

    Science.gov (United States)

    Nobile, Marco S; Cazzaniga, Paolo; Tangherloni, Andrea; Besozzi, Daniela

    2016-07-08

    Several studies in Bioinformatics, Computational Biology and Systems Biology rely on the definition of physico-chemical or mathematical models of biological systems at different scales and levels of complexity, ranging from the interaction of atoms in single molecules up to genome-wide interaction networks. Traditional computational methods and software tools developed in these research fields share a common trait: they can be computationally demanding on Central Processing Units (CPUs), therefore limiting their applicability in many circumstances. To overcome this issue, general-purpose Graphics Processing Units (GPUs) are gaining an increasing attention by the scientific community, as they can considerably reduce the running time required by standard CPU-based software, and allow more intensive investigations of biological systems. In this review, we present a collection of GPU tools recently developed to perform computational analyses in life science disciplines, emphasizing the advantages and the drawbacks in the use of these parallel architectures. The complete list of GPU-powered tools here reviewed is available at http://bit.ly/gputools. © The Author 2016. Published by Oxford University Press.

  12. Macro-scale complexity of nano- to micro-scale architecture of olivine crystals through an iodine vapour transport mechanism

    Indian Academy of Sciences (India)

    Raymond L D Whitby; Takahiro Fukuda; Toru Maekawa

    2014-04-01

    The production of nano- to micro-scale olivine (magnesium and iron silicate) crystals has been achieved at relatively low temperatures through an iodine vapour transport of the metal onto amorphous silicon dioxide. The process occurs down a temperature gradient from 800 to 600°C yielding high quality crystals with long range crystallinity, highly complex interconnectivity and intricate macroscale architecture. Scanning electron microscopy (SEM) imaging of the substrate before and after the reaction reveals that the amorphous silicon oxide species is mobile, due to the lack of correlation between the silicon oxide layer and the final olivine particles, leading to a vapour–liquid–solid or vapour–solid growth mechanism. This technique demonstrates a facile, low temperature synthetic route towards olivine crystals with nano- to micro-scale dimensions.

  13. Macroscale patterns in body size of intertidal crustaceans provide insights on climate change effects.

    Science.gov (United States)

    Jaramillo, Eduardo; Dugan, Jenifer E; Hubbard, David M; Contreras, Heraldo; Duarte, Cristian; Acuña, Emilio; Schoeman, David S

    2017-01-01

    Predicting responses of coastal ecosystems to altered sea surface temperatures (SST) associated with global climate change, requires knowledge of demographic responses of individual species. Body size is an excellent metric because it scales strongly with growth and fecundity for many ectotherms. These attributes can underpin demographic as well as community and ecosystem level processes, providing valuable insights for responses of vulnerable coastal ecosystems to changing climate. We investigated contemporary macroscale patterns in body size among widely distributed crustaceans that comprise the majority of intertidal abundance and biomass of sandy beach ecosystems of the eastern Pacific coasts of Chile and California, USA. We focused on ecologically important species representing different tidal zones, trophic guilds and developmental modes, including a high-shore macroalga-consuming talitrid amphipod (Orchestoidea tuberculata), two mid-shore scavenging cirolanid isopods (Excirolana braziliensis and E. hirsuticauda), and a low-shore suspension-feeding hippid crab (Emerita analoga) with an amphitropical distribution. Significant latitudinal patterns in body sizes were observed for all species in Chile (21° - 42°S), with similar but steeper patterns in Emerita analoga, in California (32°- 41°N). Sea surface temperature was a strong predictor of body size (-4% to -35% °C-1) in all species. Beach characteristics were subsidiary predictors of body size. Alterations in ocean temperatures of even a few degrees associated with global climate change are likely to affect body sizes of important intertidal ectotherms, with consequences for population demography, life history, community structure, trophic interactions, food-webs, and indirect effects such as ecosystem function. The consistency of results for body size and temperature across species with different life histories, feeding modes, ecological roles, and microhabitats inhabiting a single widespread coastal

  14. Macro-Scale Correction of Precipitation Undercatch in the Midwest/Great Lakes Region

    Science.gov (United States)

    Chiu, C. M.; Hamlet, A. F.

    2015-12-01

    Precipitation gauge undercatch is a serious problem in the context of using observed meteorological data sets for hydrologic modeling studies in regions with cold winters, such as the Midwest. Attention to this matter is urgently needed to support hydroclimatological research efforts in the region. To support hydrologic modeling studies, a new hybrid gridded meteorological dataset at 1/16 degree resolution based on data from CO-OP station records, the U. S. Historical Climatology Network, the Historical Canadian Climate Database, and Precipitation Regression on Independent Slopes Method has been assembled over the Great Lakes and Midwest regions from 1915-2013 at daily time step. Preliminary hydrologic simulations results using the Variable Infiltration Capacity hydrology model with this hybrid gridded meteorological dataset showed that precipitation gauge undercatch was a very significant issue throughout the region, especially for winter snowfall and simulated streamflow, which were both grossly underpredicted. Correction of primary CO-OP station data is generally infeasible due to missing station meta data and lack of local-scale wind speed measurements. Instead, macro-scale post processing techniques were developed to adjust the regridded precipitation product from CO-OP station records from 1950-2013 forwards, accounting for undercatch as a function of regridded wind speed simulations obtained from NCAR Reanalysis. Comparisons of simulated and observed streamflow over seven river basins in the Midwest were used to evaluate the datasets constructed using different combinations of meteorological station inputs, with and without undercatch corrections. The comparisons show promise in producing corrected precipitation data sets from 1950-2013 for hydrologic modeling studies, with substantial improvements in streamflow simulation from the uncalibrated VIC model when gauge undercatch corrections are included.

  15. Real-time radar signal processing using GPGPU (general-purpose graphic processing unit)

    Science.gov (United States)

    Kong, Fanxing; Zhang, Yan Rockee; Cai, Jingxiao; Palmer, Robert D.

    2016-05-01

    This study introduces a practical approach to develop real-time signal processing chain for general phased array radar on NVIDIA GPUs(Graphical Processing Units) using CUDA (Compute Unified Device Architecture) libraries such as cuBlas and cuFFT, which are adopted from open source libraries and optimized for the NVIDIA GPUs. The processed results are rigorously verified against those from the CPUs. Performance benchmarked in computation time with various input data cube sizes are compared across GPUs and CPUs. Through the analysis, it will be demonstrated that GPGPUs (General Purpose GPU) real-time processing of the array radar data is possible with relatively low-cost commercial GPUs.

  16. MASSIVELY PARALLEL LATENT SEMANTIC ANALYSES USING A GRAPHICS PROCESSING UNIT

    Energy Technology Data Exchange (ETDEWEB)

    Cavanagh, J.; Cui, S.

    2009-01-01

    Latent Semantic Analysis (LSA) aims to reduce the dimensions of large term-document datasets using Singular Value Decomposition. However, with the ever-expanding size of datasets, current implementations are not fast enough to quickly and easily compute the results on a standard PC. A graphics processing unit (GPU) can solve some highly parallel problems much faster than a traditional sequential processor or central processing unit (CPU). Thus, a deployable system using a GPU to speed up large-scale LSA processes would be a much more effective choice (in terms of cost/performance ratio) than using a PC cluster. Due to the GPU’s application-specifi c architecture, harnessing the GPU’s computational prowess for LSA is a great challenge. We presented a parallel LSA implementation on the GPU, using NVIDIA® Compute Unifi ed Device Architecture and Compute Unifi ed Basic Linear Algebra Subprograms software. The performance of this implementation is compared to traditional LSA implementation on a CPU using an optimized Basic Linear Algebra Subprograms library. After implementation, we discovered that the GPU version of the algorithm was twice as fast for large matrices (1 000x1 000 and above) that had dimensions not divisible by 16. For large matrices that did have dimensions divisible by 16, the GPU algorithm ran fi ve to six times faster than the CPU version. The large variation is due to architectural benefi ts of the GPU for matrices divisible by 16. It should be noted that the overall speeds for the CPU version did not vary from relative normal when the matrix dimensions were divisible by 16. Further research is needed in order to produce a fully implementable version of LSA. With that in mind, the research we presented shows that the GPU is a viable option for increasing the speed of LSA, in terms of cost/performance ratio.

  17. Efficient magnetohydrodynamic simulations on graphics processing units with CUDA

    Science.gov (United States)

    Wong, Hon-Cheng; Wong, Un-Hong; Feng, Xueshang; Tang, Zesheng

    2011-10-01

    Magnetohydrodynamic (MHD) simulations based on the ideal MHD equations have become a powerful tool for modeling phenomena in a wide range of applications including laboratory, astrophysical, and space plasmas. In general, high-resolution methods for solving the ideal MHD equations are computationally expensive and Beowulf clusters or even supercomputers are often used to run the codes that implemented these methods. With the advent of the Compute Unified Device Architecture (CUDA), modern graphics processing units (GPUs) provide an alternative approach to parallel computing for scientific simulations. In this paper we present, to the best of the author's knowledge, the first implementation of MHD simulations entirely on GPUs with CUDA, named GPU-MHD, to accelerate the simulation process. GPU-MHD supports both single and double precision computations. A series of numerical tests have been performed to validate the correctness of our code. Accuracy evaluation by comparing single and double precision computation results is also given. Performance measurements of both single and double precision are conducted on both the NVIDIA GeForce GTX 295 (GT200 architecture) and GTX 480 (Fermi architecture) graphics cards. These measurements show that our GPU-based implementation achieves between one and two orders of magnitude of improvement depending on the graphics card used, the problem size, and the precision when comparing to the original serial CPU MHD implementation. In addition, we extend GPU-MHD to support the visualization of the simulation results and thus the whole MHD simulation and visualization process can be performed entirely on GPUs.

  18. Accelerating sparse linear algebra using graphics processing units

    Science.gov (United States)

    Spagnoli, Kyle E.; Humphrey, John R.; Price, Daniel K.; Kelmelis, Eric J.

    2011-06-01

    The modern graphics processing unit (GPU) found in many standard personal computers is a highly parallel math processor capable of over 1 TFLOPS of peak computational throughput at a cost similar to a high-end CPU with excellent FLOPS-to-watt ratio. High-level sparse linear algebra operations are computationally intense, often requiring large amounts of parallel operations and would seem a natural fit for the processing power of the GPU. Our work is on a GPU accelerated implementation of sparse linear algebra routines. We present results from both direct and iterative sparse system solvers. The GPU execution model featured by NVIDIA GPUs based on CUDA demands very strong parallelism, requiring between hundreds and thousands of simultaneous operations to achieve high performance. Some constructs from linear algebra map extremely well to the GPU and others map poorly. CPUs, on the other hand, do well at smaller order parallelism and perform acceptably during low-parallelism code segments. Our work addresses this via hybrid a processing model, in which the CPU and GPU work simultaneously to produce results. In many cases, this is accomplished by allowing each platform to do the work it performs most naturally. For example, the CPU is responsible for graph theory portion of the direct solvers while the GPU simultaneously performs the low level linear algebra routines.

  19. GENETIC ALGORITHM ON GENERAL PURPOSE GRAPHICS PROCESSING UNIT: PARALLELISM REVIEW

    Directory of Open Access Journals (Sweden)

    A.J. Umbarkar

    2013-01-01

    Full Text Available Genetic Algorithm (GA is effective and robust method for solving many optimization problems. However, it may take more runs (iterations and time to get optimal solution. The execution time to find the optimal solution also depends upon the niching-technique applied to evolving population. This paper provides the information about how various authors, researchers, scientists have implemented GA on GPGPU (General purpose Graphics Processing Units with and without parallelism. Many problems have been solved on GPGPU using GA. GA is easy to parallelize because of its SIMD nature and therefore can be implemented well on GPGPU. Thus, speedup can definitely be achieved if bottleneck in GAs are identified and implemented effectively on GPGPU. Paper gives review of various applications solved using GAs on GPGPU with the future scope in the area of optimization.

  20. Centralization of Intensive Care Units: Process Reengineering in a Hospital

    Directory of Open Access Journals (Sweden)

    Arun Kumar

    2010-03-01

    Full Text Available Centralization of intensive care units (ICUs is a concept that has been around for several decades and the OECD countries have led the way in adopting this in their operations. Singapore Hospital was built in 1981, before the concept of centralization of ICUs took off. The hospital's ICUs were never centralized and were spread out across eight different blocks with the specialization they were associated with. Coupled with the acquisitions of the new concept of centralization and its benefits, the hospital recognizes the importance of having a centralized ICU to better handle major disasters. Using simulation models, this paper attempts to study the feasibility of centralization of ICUs in Singapore Hospital, subject to space constraints. The results will prove helpful to those who consider reengineering the intensive care process in hospitals.

  1. Simulating Lattice Spin Models on Graphics Processing Units

    CERN Document Server

    Levy, Tal; Rabani, Eran; 10.1021/ct100385b

    2012-01-01

    Lattice spin models are useful for studying critical phenomena and allow the extraction of equilibrium and dynamical properties. Simulations of such systems are usually based on Monte Carlo (MC) techniques, and the main difficulty is often the large computational effort needed when approaching critical points. In this work, it is shown how such simulations can be accelerated with the use of NVIDIA graphics processing units (GPUs) using the CUDA programming architecture. We have developed two different algorithms for lattice spin models, the first useful for equilibrium properties near a second-order phase transition point and the second for dynamical slowing down near a glass transition. The algorithms are based on parallel MC techniques, and speedups from 70- to 150-fold over conventional single-threaded computer codes are obtained using consumer-grade hardware.

  2. Molecular Dynamics Simulation of Macromolecules Using Graphics Processing Unit

    CERN Document Server

    Xu, Ji; Ge, Wei; Yu, Xiang; Yang, Xiaozhen; Li, Jinghai

    2010-01-01

    Molecular dynamics (MD) simulation is a powerful computational tool to study the behavior of macromolecular systems. But many simulations of this field are limited in spatial or temporal scale by the available computational resource. In recent years, graphics processing unit (GPU) provides unprecedented computational power for scientific applications. Many MD algorithms suit with the multithread nature of GPU. In this paper, MD algorithms for macromolecular systems that run entirely on GPU are presented. Compared to the MD simulation with free software GROMACS on a single CPU core, our codes achieve about 10 times speed-up on a single GPU. For validation, we have performed MD simulations of polymer crystallization on GPU, and the results observed perfectly agree with computations on CPU. Therefore, our single GPU codes have already provided an inexpensive alternative for macromolecular simulations on traditional CPU clusters and they can also be used as a basis to develop parallel GPU programs to further spee...

  3. Integrating post-Newtonian equations on graphics processing units

    Energy Technology Data Exchange (ETDEWEB)

    Herrmann, Frank; Tiglio, Manuel [Department of Physics, Center for Fundamental Physics, and Center for Scientific Computation and Mathematical Modeling, University of Maryland, College Park, MD 20742 (United States); Silberholz, John [Center for Scientific Computation and Mathematical Modeling, University of Maryland, College Park, MD 20742 (United States); Bellone, Matias [Facultad de Matematica, Astronomia y Fisica, Universidad Nacional de Cordoba, Cordoba 5000 (Argentina); Guerberoff, Gustavo, E-mail: tiglio@umd.ed [Facultad de Ingenieria, Instituto de Matematica y Estadistica ' Prof. Ing. Rafael Laguardia' , Universidad de la Republica, Montevideo (Uruguay)

    2010-02-07

    We report on early results of a numerical and statistical study of binary black hole inspirals. The two black holes are evolved using post-Newtonian approximations starting with initially randomly distributed spin vectors. We characterize certain aspects of the distribution shortly before merger. In particular we note the uniform distribution of black hole spin vector dot products shortly before merger and a high correlation between the initial and final black hole spin vector dot products in the equal-mass, maximally spinning case. More than 300 million simulations were performed on graphics processing units, and we demonstrate a speed-up of a factor 50 over a more conventional CPU implementation. (fast track communication)

  4. Air pollution modelling using a graphics processing unit with CUDA

    CERN Document Server

    Molnar, Ferenc; Meszaros, Robert; Lagzi, Istvan; 10.1016/j.cpc.2009.09.008

    2010-01-01

    The Graphics Processing Unit (GPU) is a powerful tool for parallel computing. In the past years the performance and capabilities of GPUs have increased, and the Compute Unified Device Architecture (CUDA) - a parallel computing architecture - has been developed by NVIDIA to utilize this performance in general purpose computations. Here we show for the first time a possible application of GPU for environmental studies serving as a basement for decision making strategies. A stochastic Lagrangian particle model has been developed on CUDA to estimate the transport and the transformation of the radionuclides from a single point source during an accidental release. Our results show that parallel implementation achieves typical acceleration values in the order of 80-120 times compared to CPU using a single-threaded implementation on a 2.33 GHz desktop computer. Only very small differences have been found between the results obtained from GPU and CPU simulations, which are comparable with the effect of stochastic tran...

  5. PO*WW*ER mobile treatment unit process hazards analysis

    Energy Technology Data Exchange (ETDEWEB)

    Richardson, R.B.

    1996-06-01

    The objective of this report is to demonstrate that a thorough assessment of the risks associated with the operation of the Rust Geotech patented PO*WW*ER mobile treatment unit (MTU) has been performed and documented. The MTU was developed to treat aqueous mixed wastes at the US Department of Energy (DOE) Albuquerque Operations Office sites. The MTU uses evaporation to separate organics and water from radionuclides and solids, and catalytic oxidation to convert the hazardous into byproducts. This process hazards analysis evaluated a number of accident scenarios not directly related to the operation of the MTU, such as natural phenomena damage and mishandling of chemical containers. Worst case accident scenarios were further evaluated to determine the risk potential to the MTU and to workers, the public, and the environment. The overall risk to any group from operation of the MTU was determined to be very low; the MTU is classified as a Radiological Facility with low hazards.

  6. Iterative Methods for MPC on Graphical Processing Units

    DEFF Research Database (Denmark)

    2012-01-01

    The high oating point performance and memory bandwidth of Graphical Processing Units (GPUs) makes them ideal for a large number of computations which often arises in scientic computing, such as matrix operations. GPUs achieve this performance by utilizing massive par- allelism, which requires...... on their applicability for GPUs. We examine published techniques for iterative methods in interior points methods (IPMs) by applying them to simple test cases, such as a system of masses connected by springs. Iterative methods allows us deal with the ill-conditioning occurring in the later iterations of the IPM as well...... as to avoid the use of dense matrices, which may be too large for the limited memory capacity of current graphics cards....

  7. Graphics processing units accelerated semiclassical initial value representation molecular dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Tamascelli, Dario; Dambrosio, Francesco Saverio [Dipartimento di Fisica, Università degli Studi di Milano, via Celoria 16, 20133 Milano (Italy); Conte, Riccardo [Department of Chemistry and Cherry L. Emerson Center for Scientific Computation, Emory University, Atlanta, Georgia 30322 (United States); Ceotto, Michele, E-mail: michele.ceotto@unimi.it [Dipartimento di Chimica, Università degli Studi di Milano, via Golgi 19, 20133 Milano (Italy)

    2014-05-07

    This paper presents a Graphics Processing Units (GPUs) implementation of the Semiclassical Initial Value Representation (SC-IVR) propagator for vibrational molecular spectroscopy calculations. The time-averaging formulation of the SC-IVR for power spectrum calculations is employed. Details about the GPU implementation of the semiclassical code are provided. Four molecules with an increasing number of atoms are considered and the GPU-calculated vibrational frequencies perfectly match the benchmark values. The computational time scaling of two GPUs (NVIDIA Tesla C2075 and Kepler K20), respectively, versus two CPUs (Intel Core i5 and Intel Xeon E5-2687W) and the critical issues related to the GPU implementation are discussed. The resulting reduction in computational time and power consumption is significant and semiclassical GPU calculations are shown to be environment friendly.

  8. Polymer Field-Theory Simulations on Graphics Processing Units

    CERN Document Server

    Delaney, Kris T

    2012-01-01

    We report the first CUDA graphics-processing-unit (GPU) implementation of the polymer field-theoretic simulation framework for determining fully fluctuating expectation values of equilibrium properties for periodic and select aperiodic polymer systems. Our implementation is suitable both for self-consistent field theory (mean-field) solutions of the field equations, and for fully fluctuating simulations using the complex Langevin approach. Running on NVIDIA Tesla T20 series GPUs, we find double-precision speedups of up to 30x compared to single-core serial calculations on a recent reference CPU, while single-precision calculations proceed up to 60x faster than those on the single CPU core. Due to intensive communications overhead, an MPI implementation running on 64 CPU cores remains two times slower than a single GPU.

  9. Graphics Processing Units and High-Dimensional Optimization.

    Science.gov (United States)

    Zhou, Hua; Lange, Kenneth; Suchard, Marc A

    2010-08-01

    This paper discusses the potential of graphics processing units (GPUs) in high-dimensional optimization problems. A single GPU card with hundreds of arithmetic cores can be inserted in a personal computer and dramatically accelerates many statistical algorithms. To exploit these devices fully, optimization algorithms should reduce to multiple parallel tasks, each accessing a limited amount of data. These criteria favor EM and MM algorithms that separate parameters and data. To a lesser extent block relaxation and coordinate descent and ascent also qualify. We demonstrate the utility of GPUs in nonnegative matrix factorization, PET image reconstruction, and multidimensional scaling. Speedups of 100 fold can easily be attained. Over the next decade, GPUs will fundamentally alter the landscape of computational statistics. It is time for more statisticians to get on-board.

  10. Graphics Processing Unit Enhanced Parallel Document Flocking Clustering

    Energy Technology Data Exchange (ETDEWEB)

    Cui, Xiaohui [ORNL; Potok, Thomas E [ORNL; ST Charles, Jesse Lee [ORNL

    2010-01-01

    Analyzing and clustering documents is a complex problem. One explored method of solving this problem borrows from nature, imitating the flocking behavior of birds. One limitation of this method of document clustering is its complexity O(n2). As the number of documents grows, it becomes increasingly difficult to generate results in a reasonable amount of time. In the last few years, the graphics processing unit (GPU) has received attention for its ability to solve highly-parallel and semi-parallel problems much faster than the traditional sequential processor. In this paper, we have conducted research to exploit this archi- tecture and apply its strengths to the flocking based document clustering problem. Using the CUDA platform from NVIDIA, we developed a doc- ument flocking implementation to be run on the NVIDIA GEFORCE GPU. Performance gains ranged from thirty-six to nearly sixty times improvement of the GPU over the CPU implementation.

  11. Implementing wide baseline matching algorithms on a graphics processing unit.

    Energy Technology Data Exchange (ETDEWEB)

    Rothganger, Fredrick H.; Larson, Kurt W.; Gonzales, Antonio Ignacio; Myers, Daniel S.

    2007-10-01

    Wide baseline matching is the state of the art for object recognition and image registration problems in computer vision. Though effective, the computational expense of these algorithms limits their application to many real-world problems. The performance of wide baseline matching algorithms may be improved by using a graphical processing unit as a fast multithreaded co-processor. In this paper, we present an implementation of the difference of Gaussian feature extractor, based on the CUDA system of GPU programming developed by NVIDIA, and implemented on their hardware. For a 2000x2000 pixel image, the GPU-based method executes nearly thirteen times faster than a comparable CPU-based method, with no significant loss of accuracy.

  12. Macroscale cobalt-MOFs derived metallic Co nanoparticles embedded in N-doped porous carbon layers as efficient oxygen electrocatalysts

    Science.gov (United States)

    Lu, Hai-Sheng; Zhang, Haimin; Liu, Rongrong; Zhang, Xian; Zhao, Huijun; Wang, Guozhong

    2017-01-01

    Metal-organic frameworks (MOFs) materials have aroused great research interest in different areas owing to their unique properties, such as high surface area, various composition, well-organized framework and controllable porous structure. Controllable fabrication of MOFs materials at macro-scale may be more promising for their large-scale practical applications. Here we report the synthesis of macro-scale Co-MOFs crystals using 1,3,5-benzenetricarboxylic acid (H3BTC) linker in the presence of Co2+, triethylamine (TEA) and nonanoic acid by a facile solvothermal reaction. Further, the as-fabricated Co-MOFs as precursor was pyrolytically treated at different temperatures in N2 atmosphere to obtain metallic Co nanoparticles embedded in N-doped porous carbon layers (denoted as Co@NPC). The results demonstrate that the Co-MOFs derived sample obtained at 900 °C (Co@NPC-900) shows a porous structure (including micropore and mesopore) with a surface area of 110.8 m2 g-1 and an N doping level of 1.62 at.% resulted from TEA in the pyrolysis process. As electrocatalyst, the Co@NPC-900 exhibits bifunctional electrocatalytic activities toward the oxygen reduction reaction (ORR) and oxygen evolution reaction (OER) in alkaline media which are key reactions in some renewable energy technologies such as fuel cells and rechargeable metal-air batteries. The results indicate that the Co@NPC-900 can afford an onset potential of 1.50 V (vs. RHE) and a potential value of 1.61 V (vs. RHE) at a current density of 10 mA cm-2 for ORR and OER with high applicable stability, respectively. The efficient catalytic activity of Co@NPC-900 as bifunctional oxygen electrocatalyst can be ascribed to N doping and embedded metallic Co nanoparticles in carbon structure providing catalytic active sites and porous structure favourable for electrocatalysis-related mass transport.

  13. Soft matter approaches as enablers for food macroscale simulation

    NARCIS (Netherlands)

    Datta, A.K.; Sman, van der R.G.M.; Gulati, T.; Warning, A.

    2012-01-01

    Macroscopic deformable multiphase porous media models have been successful in describing many complex food processes. However, the properties needed for such detailed physics-based models are scarce and consist of primarily empirical models obtained from experiment. Likewise, driving forces such as

  14. Minimization of entropy production in separate and connected process units

    Energy Technology Data Exchange (ETDEWEB)

    Roesjorde, Audun

    2004-08-01

    The objective of this thesis was to further develop a methodology for minimizing the entropy production of single and connected chemical process units. When chemical process equipment is designed and operated at the lowest entropy production possible, the energy efficiency of the equipment is enhanced. We have found for single process units that the entropy production could be reduced with up to 20-40%, given the degrees of freedom in the optimization. In processes, our results indicated that even bigger reductions were possible. The states of minimum entropy production were studied and important painter's for obtaining significant reductions in the entropy production were identified. Both from sustain ability and economical viewpoints knowledge of energy efficient design and operation are important. In some of the systems we studied, nonequilibrium thermodynamics was used to model the entropy production. In Chapter 2, we gave a brief introduction to different industrial applications of nonequilibrium thermodynamics. The link between local transport phenomena and overall system description makes nonequilibrium thermodynamics a useful tool for understanding design of chemical process units. We developed the methodology of minimization of entropy production in several steps. First, we analyzed and optimized the entropy production of single units: Two alternative concepts of adiabatic distillation; diabatic and heat-integrated distillation, were analyzed and optimized in Chapter 3 to 5. In diabatic distillation, heat exchange is allowed along the column, and it is this feature that increases the energy efficiency of the distillation column. In Chapter 3, we found how a given area of heat transfer should be optimally distributed among the trays in a column separating a mixture of propylene and propane. The results showed that heat exchange was most important on the trays close to the re boiler and condenser. In Chapter 4 and 5, we studied how the entropy

  15. Microscale physiological and ecological studies of aquatic cyanobacteria: macroscale implications.

    Science.gov (United States)

    Paerl, H W

    1996-01-01

    Cyanobacteria have had a profound and unparalleled biogeochemical impact on the earth's biosphere. As the first oxygenic phototrophs, cyanobacteria were responsible for the transition from anaerobic to aerobic life. Ironically, molecular oxygen (O2) is inhibitory to critical components of cyanobacterial metabolism, including photosynthesis and nitrogen fixation. Cyanobacteria have developed a great variety of biochemical, structural, and biotic adaptations ensuring optimal growth and proliferation in diverse oxic environments to counter this difficult situation. Structurally, cyanobacteria reveal remarkable diversity, including the formation of highly differentiated, O2-deplete cells (heterocysts), multicellularity as trichomes, and aggregates, that, among N2-fixing genera, facilitate division of labor between aerobic and anaerobic processes. Cyanobacteria enjoy unique consortial and symbiotic associations with other microorganisms, higher plants, and animals, in which O2 consumption is closely coupled in time and space to its production. Because as prokaryotes they are devoid of O2-consuming organelles (e.g., mitochondria), cyanobacteria have developed alternative strategies for locally protecting O2-sensitive processes, including consortial relationships with other microorganisms. Specific organic compounds released by cyanobacteria are capable of chemotactically attracting bacterial consorts, which in turn attach to the host cyanobacteria, consume O2, and recycle inorganic nutrients within the cyanobacterial "phycosphere." Multicellularity and aggregation lead to localized O2 gradients and hypoxic/anoxic microzones in which O2-sensitive processes can coexist. Microscale partitioning of O2-producing and O2-inhibited processes promotes contiguous and effective metabolite and nutrient exchange between these processes in oxygenated waters, representing a bulk of the world's oceanic and freshwater ecosystems.

  16. The ATLAS Fast Tracker Processing Units - track finding and fitting

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00384270; The ATLAS collaboration; Alison, John; Ancu, Lucian Stefan; Andreani, Alessandro; Annovi, Alberto; Beccherle, Roberto; Beretta, Matteo; Biesuz, Nicolo Vladi; Bogdan, Mircea Arghir; Bryant, Patrick; Calabro, Domenico; Citraro, Saverio; Crescioli, Francesco; Dell'Orso, Mauro; Donati, Simone; Gentsos, Christos; Giannetti, Paola; Gkaitatzis, Stamatios; Gramling, Johanna; Greco, Virginia; Horyn, Lesya Anna; Iovene, Alessandro; Kalaitzidis, Panagiotis; Kim, Young-Kee; Kimura, Naoki; Kordas, Kostantinos; Kubota, Takashi; Lanza, Agostino; Liberali, Valentino; Luciano, Pierluigi; Magnin, Betty; Sakellariou, Andreas; Sampsonidis, Dimitrios; Saxon, James; Shojaii, Seyed Ruhollah; Sotiropoulou, Calliope Louisa; Stabile, Alberto; Swiatlowski, Maximilian; Volpi, Guido; Zou, Rui; Shochet, Mel

    2016-01-01

    The Fast Tracker is a hardware upgrade to the ATLAS trigger and data-acquisition system, with the goal of providing global track reconstruction by the start of the High Level Trigger starts. The Fast Tracker can process incoming data from the whole inner detector at full first level trigger rate, up to 100 kHz, using custom electronic boards. At the core of the system is a Processing Unit installed in a VMEbus crate, formed by two sets of boards: the Associative Memory Board and a powerful rear transition module called the Auxiliary card, while the second set is the Second Stage board. The associative memories perform the pattern matching looking for correlations within the incoming data, compatible with track candidates at coarse resolution. The pattern matching task is performed using custom application specific integrated circuits, called associative memory chips. The auxiliary card prepares the input and reject bad track candidates obtained from from the Associative Memory Board using the full precision a...

  17. The ATLAS Fast TracKer Processing Units

    CERN Document Server

    Krizka, Karol; The ATLAS collaboration

    2016-01-01

    The Fast Tracker is a hardware upgrade to the ATLAS trigger and data-acquisition system, with the goal of providing global track reconstruction by the start of the High Level Trigger starts. The Fast Tracker can process incoming data from the whole inner detector at full first level trigger rate, up to 100 kHz, using custom electronic boards. At the core of the system is a Processing Unit installed in a VMEbus crate, formed by two sets of boards: the Associative Memory Board and a powerful rear transition module called the Auxiliary card, while the second set is the Second Stage board. The associative memories perform the pattern matching looking for correlations within the incoming data, compatible with track candidates at coarse resolution. The pattern matching task is performed using custom application specific integrated circuits, called associative memory chips. The auxiliary card prepares the input and reject bad track candidates obtained from from the Associative Memory Board using the full precision a...

  18. Beowulf Distributed Processing and the United States Geological Survey

    Science.gov (United States)

    Maddox, Brian G.

    2002-01-01

    Introduction In recent years, the United States Geological Survey's (USGS) National Mapping Discipline (NMD) has expanded its scientific and research activities. Work is being conducted in areas such as emergency response research, scientific visualization, urban prediction, and other simulation activities. Custom-produced digital data have become essential for these types of activities. High-resolution, remotely sensed datasets are also seeing increased use. Unfortunately, the NMD is also finding that it lacks the resources required to perform some of these activities. Many of these projects require large amounts of computer processing resources. Complex urban-prediction simulations, for example, involve large amounts of processor-intensive calculations on large amounts of input data. This project was undertaken to learn and understand the concepts of distributed processing. Experience was needed in developing these types of applications. The idea was that this type of technology could significantly aid the needs of the NMD scientific and research programs. Porting a numerically intensive application currently being used by an NMD science program to run in a distributed fashion would demonstrate the usefulness of this technology. There are several benefits that this type of technology can bring to the USGS's research programs. Projects can be performed that were previously impossible due to a lack of computing resources. Other projects can be performed on a larger scale than previously possible. For example, distributed processing can enable urban dynamics research to perform simulations on larger areas without making huge sacrifices in resolution. The processing can also be done in a more reasonable amount of time than with traditional single-threaded methods (a scaled version of Chester County, Pennsylvania, took about fifty days to finish its first calibration phase with a single-threaded program). This paper has several goals regarding distributed processing

  19. Density functional theory calculation on many-cores hybrid central processing unit-graphic processing unit architectures.

    Science.gov (United States)

    Genovese, Luigi; Ospici, Matthieu; Deutsch, Thierry; Méhaut, Jean-François; Neelov, Alexey; Goedecker, Stefan

    2009-07-21

    We present the implementation of a full electronic structure calculation code on a hybrid parallel architecture with graphic processing units (GPUs). This implementation is performed on a free software code based on Daubechies wavelets. Such code shows very good performances, systematic convergence properties, and an excellent efficiency on parallel computers. Our GPU-based acceleration fully preserves all these properties. In particular, the code is able to run on many cores which may or may not have a GPU associated, and thus on parallel and massive parallel hybrid machines. With double precision calculations, we may achieve considerable speedup, between a factor of 20 for some operations and a factor of 6 for the whole density functional theory code.

  20. Numerical modeling of macroscale brittle rock crushing during impacts

    Energy Technology Data Exchange (ETDEWEB)

    Badr, Salah A.; Abdelhaffez, Gamal S. [King Abdulaziz Univ., Jeddah (Saudi Arabia)

    2014-02-01

    Several machines, such as crushers use the physical effect of compression to cause fragmentation 'crushing' of brittle rocks. As a consequence of the complex fragmentation process, crushers are still sized by empirical approaches. This paper present the results of a numerical study to understand some aspects of rock crushing phenomenon in terms of energy consumption. The study uses the discrete element approach of PFC2D code to simulate a stamp mill. The stamp mill has a simple crushing mechanism of a fixed kinetic energy delivered by a rigid ram impact. A single rock fragment crushing process dependent on the number of stamp mill ram blows is numerically examined. Both amount and type of energy generated by a ram blow are monitored besides the type of fractures generated. The model results indicate that the ram impact energy is mainly consumed in form of friction energy (up to 61 %) while strain energy stays at about 5 % of delivered energy. The energy consumed by crushing the rock represents only 32 % to 45 % of stamp mill energy and tends to decrease as the number of impacts increases. The rock fragmented matrix tends to convert into more friction energy with reduced number of new fractures as number of blows increase. The fragmentation caused by tensile is more often compared to those caused by shear, this behaviour increased with increasing number of ram blows. (orig.)

  1. Comparative analysis of the macroscale structural connectivity in the macaque and human brain.

    Directory of Open Access Journals (Sweden)

    Alexandros Goulas

    2014-03-01

    Full Text Available The macaque brain serves as a model for the human brain, but its suitability is challenged by unique human features, including connectivity reconfigurations, which emerged during primate evolution. We perform a quantitative comparative analysis of the whole brain macroscale structural connectivity of the two species. Our findings suggest that the human and macaque brain as a whole are similarly wired. A region-wise analysis reveals many interspecies similarities of connectivity patterns, but also lack thereof, primarily involving cingulate regions. We unravel a common structural backbone in both species involving a highly overlapping set of regions. This structural backbone, important for mediating information across the brain, seems to constitute a feature of the primate brain persevering evolution. Our findings illustrate novel evolutionary aspects at the macroscale connectivity level and offer a quantitative translational bridge between macaque and human research.

  2. Microbial control of mineral–groundwater equilibria:Macroscale to microscale

    Science.gov (United States)

    Bennett, Philip C.; Hiebert, Franz K.; Roger, Jennifer Roberts

    2000-01-01

    macroscaleprocesses that perturb general groundwater chemistry and therefore mineral–water equilibria; and microscale interactions, where attached organisms locally perturb mineral–water equilibria, potentially releasing limiting trace nutrients from the dissolving mineral.In the contaminated unconfined glacio-fluvial aquifer near Bemidji, Minnesota, USA, carbonate chemistry is influenced primarily at the macroscale. Under oxic conditions, respiration by native aerobic heterotrophs produces excess carbon dioxide that promotes calcite and dolomite dissolution. Aerobic microorganisms do not colonize dolomite surfaces and few occur on calcite. Within the anoxic groundwater, calcite overgrowths form on uncolonized calcite cleavage surfaces, possibly due to the consumption of acidity by dissimilatory iron-reducing bacteria. As molecular oxygen concentration increases downgradient of the oil pool, aerobes again dominate and residual hydrocarbons and ferrous iron are oxidized, resulting in macroscale carbonate-mineral dissolution and iron precipitation.

  3. Thermochemical Process Development Unit: Researching Fuels from Biomass, Bioenergy Technologies (Fact Sheet)

    Energy Technology Data Exchange (ETDEWEB)

    2009-01-01

    The Thermochemical Process Development Unit (TCPDU) at the National Renewable Energy Laboratory (NREL) is a unique facility dedicated to researching thermochemical processes to produce fuels from biomass.

  4. Analysis of Unit Process Cost for an Engineering-Scale Pyroprocess Facility Using a Process Costing Method in Korea

    National Research Council Canada - National Science Library

    Sungki Kim; Wonil Ko; Sungsig Bang

    2015-01-01

    ...) metal ingots in a high-temperature molten salt phase. This paper provides the unit process cost of a pyroprocess facility that can process up to 10 tons of pyroprocessing product per year by utilizing the process costing method...

  5. Monte Carlo MP2 on Many Graphical Processing Units.

    Science.gov (United States)

    Doran, Alexander E; Hirata, So

    2016-10-11

    In the Monte Carlo second-order many-body perturbation (MC-MP2) method, the long sum-of-product matrix expression of the MP2 energy, whose literal evaluation may be poorly scalable, is recast into a single high-dimensional integral of functions of electron pair coordinates, which is evaluated by the scalable method of Monte Carlo integration. The sampling efficiency is further accelerated by the redundant-walker algorithm, which allows a maximal reuse of electron pairs. Here, a multitude of graphical processing units (GPUs) offers a uniquely ideal platform to expose multilevel parallelism: fine-grain data-parallelism for the redundant-walker algorithm in which millions of threads compute and share orbital amplitudes on each GPU; coarse-grain instruction-parallelism for near-independent Monte Carlo integrations on many GPUs with few and infrequent interprocessor communications. While the efficiency boost by the redundant-walker algorithm on central processing units (CPUs) grows linearly with the number of electron pairs and tends to saturate when the latter exceeds the number of orbitals, on a GPU it grows quadratically before it increases linearly and then eventually saturates at a much larger number of pairs. This is because the orbital constructions are nearly perfectly parallelized on a GPU and thus completed in a near-constant time regardless of the number of pairs. In consequence, an MC-MP2/cc-pVDZ calculation of a benzene dimer is 2700 times faster on 256 GPUs (using 2048 electron pairs) than on two CPUs, each with 8 cores (which can use only up to 256 pairs effectively). We also numerically determine that the cost to achieve a given relative statistical uncertainty in an MC-MP2 energy increases as O(n(3)) or better with system size n, which may be compared with the O(n(5)) scaling of the conventional implementation of deterministic MP2. We thus establish the scalability of MC-MP2 with both system and computer sizes.

  6. Carbon MEMS from the nanoscale to the macroscale: Novel fabrication techniques and applications in electrochemistry

    Science.gov (United States)

    Zaouk, Rabih Bachir

    Micro electromechanical systems (MEMS) have strongly impacted our way of life in the last two decades. From accelerometers and gyroscopes that ensure your driving safety, to inkjet printer cartridges that transpose your ideas onto paper, to micromirrors that enable your small projectors. MEMS have become more and more ubiquitous. Silicon, the material on which the semiconductor industry based its revolution, has so far been the material of choice for MEMS. While silicon is a great platform for constructing electronics, it is less than ideal for applications that involve electrodes exposed to aggressive liquid and gaseous environments. Carbon is one of the most commonly used materials when it comes to electrochemical applications, it is therefore the best candidate to carry over the trend of miniaturization in arenas such as smart chemical sensing, biological microdevices, miniature power, etc. Recent advances in engineering nanoscale structures show great promise towards delivering higher performance sensors, detectors, transistors, displays, etc. In order to leverage the power of nanostructures in general, new manufacturing processes that can bridge between the nanoscale and the macroscale are needed. Such integrated fabrication methods are essential in enabling the transfer of the advantages boasted by nanostructures from the research labs towards mass manufacturing. The present work starts by introducing the basic photolithography technique that has been used so far to fabricate Carbon MEMS (C-MEMS). Several novel techniques stemming for the original process are then described in details and lithium-ion microbattery anodes are presented as an example application of these novel fabrication methods. These Carbon MEMS anodes are characterized through a combination of cyclic voltammetry and electrochemical impedance spectroscopy (OS). A new finite element analysis (FEA) technique is then proposed to more accurately model the current density distributions of 3

  7. Perspective on Continuum Modeling of Mesoscale/ Macroscale Phenomena

    Science.gov (United States)

    Bammann, D. J.

    The attempt to model or predict the inelastic response or permanent deformation and failure observed in metals dates back over 180 years. Various descriptions of the post elastic response of metals have been proposed from the fields of physics, materials science (metallurgy), engineering, mechanics, and applied mathematics. The communication between these fields has improved and many of the modeling efforts today involve concepts from most or all of these fields. Early engineering description of post yield response treated the material as perfectly plastic — the material continues to deform with zero additional increase in load. These models became the basis of the mathematical theory of plasticity and were extended to account for hardening, unloading, and directional hardening. In contradistinction, rheological models treated the finite deformation of a solid similar to the deformation of a viscous fluid. In many cases of large deformation, rheological models have provided both adequate and accurate information about the deformed shape of a metal during many manufacturing processes. The treatment of geometric defects in solid bodies initiated within the mathematical theory of elasticity, the dislocation, introduced as an incompatible "cut" in a continuum body. This resulted in a very large body of literature devoted to the linear elastic study of dislocations, dislocation structures, and their interactions, and has provided essential information in the understanding of the "state" of a deformed material.

  8. Accelerating chemical database searching using graphics processing units.

    Science.gov (United States)

    Liu, Pu; Agrafiotis, Dimitris K; Rassokhin, Dmitrii N; Yang, Eric

    2011-08-22

    The utility of chemoinformatics systems depends on the accurate computer representation and efficient manipulation of chemical compounds. In such systems, a small molecule is often digitized as a large fingerprint vector, where each element indicates the presence/absence or the number of occurrences of a particular structural feature. Since in theory the number of unique features can be exceedingly large, these fingerprint vectors are usually folded into much shorter ones using hashing and modulo operations, allowing fast "in-memory" manipulation and comparison of molecules. There is increasing evidence that lossless fingerprints can substantially improve retrieval performance in chemical database searching (substructure or similarity), which have led to the development of several lossless fingerprint compression algorithms. However, any gains in storage and retrieval afforded by compression need to be weighed against the extra computational burden required for decompression before these fingerprints can be compared. Here we demonstrate that graphics processing units (GPU) can greatly alleviate this problem, enabling the practical application of lossless fingerprints on large databases. More specifically, we show that, with the help of a ~$500 ordinary video card, the entire PubChem database of ~32 million compounds can be searched in ~0.2-2 s on average, which is 2 orders of magnitude faster than a conventional CPU. If multiple query patterns are processed in batch, the speedup is even more dramatic (less than 0.02-0.2 s/query for 1000 queries). In the present study, we use the Elias gamma compression algorithm, which results in a compression ratio as high as 0.097.

  9. Massively Parallel Latent Semantic Analyzes using a Graphics Processing Unit

    Energy Technology Data Exchange (ETDEWEB)

    Cavanagh, Joseph M [ORNL; Cui, Xiaohui [ORNL

    2009-01-01

    Latent Semantic Indexing (LSA) aims to reduce the dimensions of large Term-Document datasets using Singular Value Decomposition. However, with the ever expanding size of data sets, current implementations are not fast enough to quickly and easily compute the results on a standard PC. The Graphics Processing Unit (GPU) can solve some highly parallel problems much faster than the traditional sequential processor (CPU). Thus, a deployable system using a GPU to speedup large-scale LSA processes would be a much more effective choice (in terms of cost/performance ratio) than using a computer cluster. Due to the GPU s application-specific architecture, harnessing the GPU s computational prowess for LSA is a great challenge. We present a parallel LSA implementation on the GPU, using NVIDIA Compute Unified Device Architecture and Compute Unified Basic Linear Algebra Subprograms. The performance of this implementation is compared to traditional LSA implementation on CPU using an optimized Basic Linear Algebra Subprograms library. After implementation, we discovered that the GPU version of the algorithm was twice as fast for large matrices (1000x1000 and above) that had dimensions not divisible by 16. For large matrices that did have dimensions divisible by 16, the GPU algorithm ran five to six times faster than the CPU version. The large variation is due to architectural benefits the GPU has for matrices divisible by 16. It should be noted that the overall speeds for the CPU version did not vary from relative normal when the matrix dimensions were divisible by 16. Further research is needed in order to produce a fully implementable version of LSA. With that in mind, the research we presented shows that the GPU is a viable option for increasing the speed of LSA, in terms of cost/performance ratio.

  10. Towards a Unified Sentiment Lexicon Based on Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Liliana Ibeth Barbosa-Santillán

    2014-01-01

    Full Text Available This paper presents an approach to create what we have called a Unified Sentiment Lexicon (USL. This approach aims at aligning, unifying, and expanding the set of sentiment lexicons which are available on the web in order to increase their robustness of coverage. One problem related to the task of the automatic unification of different scores of sentiment lexicons is that there are multiple lexical entries for which the classification of positive, negative, or neutral {P,N,Z} depends on the unit of measurement used in the annotation methodology of the source sentiment lexicon. Our USL approach computes the unified strength of polarity of each lexical entry based on the Pearson correlation coefficient which measures how correlated lexical entries are with a value between 1 and −1, where 1 indicates that the lexical entries are perfectly correlated, 0 indicates no correlation, and −1 means they are perfectly inversely correlated and so is the UnifiedMetrics procedure for CPU and GPU, respectively. Another problem is the high processing time required for computing all the lexical entries in the unification task. Thus, the USL approach computes a subset of lexical entries in each of the 1344 GPU cores and uses parallel processing in order to unify 155802 lexical entries. The results of the analysis conducted using the USL approach show that the USL has 95.430 lexical entries, out of which there are 35.201 considered to be positive, 22.029 negative, and 38.200 neutral. Finally, the runtime was 10 minutes for 95.430 lexical entries; this allows a reduction of the time computing for the UnifiedMetrics by 3 times.

  11. High-throughput sequence alignment using Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Trapnell Cole

    2007-12-01

    Full Text Available Abstract Background The recent availability of new, less expensive high-throughput DNA sequencing technologies has yielded a dramatic increase in the volume of sequence data that must be analyzed. These data are being generated for several purposes, including genotyping, genome resequencing, metagenomics, and de novo genome assembly projects. Sequence alignment programs such as MUMmer have proven essential for analysis of these data, but researchers will need ever faster, high-throughput alignment tools running on inexpensive hardware to keep up with new sequence technologies. Results This paper describes MUMmerGPU, an open-source high-throughput parallel pairwise local sequence alignment program that runs on commodity Graphics Processing Units (GPUs in common workstations. MUMmerGPU uses the new Compute Unified Device Architecture (CUDA from nVidia to align multiple query sequences against a single reference sequence stored as a suffix tree. By processing the queries in parallel on the highly parallel graphics card, MUMmerGPU achieves more than a 10-fold speedup over a serial CPU version of the sequence alignment kernel, and outperforms the exact alignment component of MUMmer on a high end CPU by 3.5-fold in total application time when aligning reads from recent sequencing projects using Solexa/Illumina, 454, and Sanger sequencing technologies. Conclusion MUMmerGPU is a low cost, ultra-fast sequence alignment program designed to handle the increasing volume of data produced by new, high-throughput sequencing technologies. MUMmerGPU demonstrates that even memory-intensive applications can run significantly faster on the relatively low-cost GPU than on the CPU.

  12. Kinematic modelling of disc galaxies using graphics processing units

    Science.gov (United States)

    Bekiaris, G.; Glazebrook, K.; Fluke, C. J.; Abraham, R.

    2016-01-01

    With large-scale integral field spectroscopy (IFS) surveys of thousands of galaxies currently under-way or planned, the astronomical community is in need of methods, techniques and tools that will allow the analysis of huge amounts of data. We focus on the kinematic modelling of disc galaxies and investigate the potential use of massively parallel architectures, such as the graphics processing unit (GPU), as an accelerator for the computationally expensive model-fitting procedure. We review the algorithms involved in model-fitting and evaluate their suitability for GPU implementation. We employ different optimization techniques, including the Levenberg-Marquardt and nested sampling algorithms, but also a naive brute-force approach based on nested grids. We find that the GPU can accelerate the model-fitting procedure up to a factor of ˜100 when compared to a single-threaded CPU, and up to a factor of ˜10 when compared to a multithreaded dual CPU configuration. Our method's accuracy, precision and robustness are assessed by successfully recovering the kinematic properties of simulated data, and also by verifying the kinematic modelling results of galaxies from the GHASP and DYNAMO surveys as found in the literature. The resulting GBKFIT code is available for download from: http://supercomputing.swin.edu.au/gbkfit.

  13. Graphics processing unit-accelerated quantitative trait Loci detection.

    Science.gov (United States)

    Chapuis, Guillaume; Filangi, Olivier; Elsen, Jean-Michel; Lavenier, Dominique; Le Roy, Pascale

    2013-09-01

    Mapping quantitative trait loci (QTL) using genetic marker information is a time-consuming analysis that has interested the mapping community in recent decades. The increasing amount of genetic marker data allows one to consider ever more precise QTL analyses while increasing the demand for computation. Part of the difficulty of detecting QTLs resides in finding appropriate critical values or threshold values, above which a QTL effect is considered significant. Different approaches exist to determine these thresholds, using either empirical methods or algebraic approximations. In this article, we present a new implementation of existing software, QTLMap, which takes advantage of the data parallel nature of the problem by offsetting heavy computations to a graphics processing unit (GPU). Developments on the GPU were implemented using Cuda technology. This new implementation performs up to 75 times faster than the previous multicore implementation, while maintaining the same results and level of precision (Double Precision) and computing both QTL values and thresholds. This speedup allows one to perform more complex analyses, such as linkage disequilibrium linkage analyses (LDLA) and multiQTL analyses, in a reasonable time frame.

  14. Accelerating VASP electronic structure calculations using graphic processing units

    KAUST Repository

    Hacene, Mohamed

    2012-08-20

    We present a way to improve the performance of the electronic structure Vienna Ab initio Simulation Package (VASP) program. We show that high-performance computers equipped with graphics processing units (GPUs) as accelerators may reduce drastically the computation time when offloading these sections to the graphic chips. The procedure consists of (i) profiling the performance of the code to isolate the time-consuming parts, (ii) rewriting these so that the algorithms become better-suited for the chosen graphic accelerator, and (iii) optimizing memory traffic between the host computer and the GPU accelerator. We chose to accelerate VASP with NVIDIA GPU using CUDA. We compare the GPU and original versions of VASP by evaluating the Davidson and RMM-DIIS algorithms on chemical systems of up to 1100 atoms. In these tests, the total time is reduced by a factor between 3 and 8 when running on n (CPU core + GPU) compared to n CPU cores only, without any accuracy loss. © 2012 Wiley Periodicals, Inc.

  15. Parallelizing the Cellular Potts Model on graphics processing units

    Science.gov (United States)

    Tapia, José Juan; D'Souza, Roshan M.

    2011-04-01

    The Cellular Potts Model (CPM) is a lattice based modeling technique used for simulating cellular structures in computational biology. The computational complexity of the model means that current serial implementations restrict the size of simulation to a level well below biological relevance. Parallelization on computing clusters enables scaling the size of the simulation but marginally addresses computational speed due to the limited memory bandwidth between nodes. In this paper we present new data-parallel algorithms and data structures for simulating the Cellular Potts Model on graphics processing units. Our implementations handle most terms in the Hamiltonian, including cell-cell adhesion constraint, cell volume constraint, cell surface area constraint, and cell haptotaxis. We use fine level checkerboards with lock mechanisms using atomic operations to enable consistent updates while maintaining a high level of parallelism. A new data-parallel memory allocation algorithm has been developed to handle cell division. Tests show that our implementation enables simulations of >10 cells with lattice sizes of up to 256 3 on a single graphics card. Benchmarks show that our implementation runs ˜80× faster than serial implementations, and ˜5× faster than previous parallel implementations on computing clusters consisting of 25 nodes. The wide availability and economy of graphics cards mean that our techniques will enable simulation of realistically sized models at a fraction of the time and cost of previous implementations and are expected to greatly broaden the scope of CPM applications.

  16. Kinematic Modelling of Disc Galaxies using Graphics Processing Units

    CERN Document Server

    Bekiaris, Georgios; Fluke, Christopher J; Abraham, Roberto

    2015-01-01

    With large-scale Integral Field Spectroscopy (IFS) surveys of thousands of galaxies currently under-way or planned, the astronomical community is in need of methods, techniques and tools that will allow the analysis of huge amounts of data. We focus on the kinematic modelling of disc galaxies and investigate the potential use of massively parallel architectures, such as the Graphics Processing Unit (GPU), as an accelerator for the computationally expensive model-fitting procedure. We review the algorithms involved in model-fitting and evaluate their suitability for GPU implementation. We employ different optimization techniques, including the Levenberg-Marquardt and Nested Sampling algorithms, but also a naive brute-force approach based on Nested Grids. We find that the GPU can accelerate the model-fitting procedure up to a factor of ~100 when compared to a single-threaded CPU, and up to a factor of ~10 when compared to a multi-threaded dual CPU configuration. Our method's accuracy, precision and robustness a...

  17. Efficient graphics processing unit-based voxel carving for surveillance

    Science.gov (United States)

    Ober-Gecks, Antje; Zwicker, Marius; Henrich, Dominik

    2016-07-01

    A graphics processing unit (GPU)-based implementation of a space carving method for the reconstruction of the photo hull is presented. In particular, the generalized voxel coloring with item buffer approach is transferred to the GPU. The fast computation on the GPU is realized by an incrementally calculated standard deviation within the likelihood ratio test, which is applied as color consistency criterion. A fast and efficient computation of complete voxel-pixel projections is provided using volume rendering methods. This generates a speedup of the iterative carving procedure while considering all given pixel color information. Different volume rendering methods, such as texture mapping and raycasting, are examined. The termination of the voxel carving procedure is controlled through an anytime concept. The photo hull algorithm is examined for its applicability to real-world surveillance scenarios as an online reconstruction method. For this reason, a GPU-based redesign of a visual hull algorithm is provided that utilizes geometric knowledge about known static occluders of the scene in order to create a conservative and complete visual hull that includes all given objects. This visual hull approximation serves as input for the photo hull algorithm.

  18. Undergraduate Game Degree Programs in the United Kingdom and United States: A Comparison of the Curriculum Planning Process

    Science.gov (United States)

    McGill, Monica M.

    2010-01-01

    Digital games are marketed, mass-produced, and consumed by an increasing number of people and the game industry is only expected to grow. In response, post-secondary institutions in the United Kingdom (UK) and the United States (US) have started to create game degree programs. Though curriculum theorists provide insight into the process of…

  19. Undergraduate Game Degree Programs in the United Kingdom and United States: A Comparison of the Curriculum Planning Process

    Science.gov (United States)

    McGill, Monica M.

    2010-01-01

    Digital games are marketed, mass-produced, and consumed by an increasing number of people and the game industry is only expected to grow. In response, post-secondary institutions in the United Kingdom (UK) and the United States (US) have started to create game degree programs. Though curriculum theorists provide insight into the process of…

  20. Simulating Land Surface Hydrology at a 30-meter Spatial Resolution over the Contiguous United States

    Science.gov (United States)

    Wood, E. F.; Pan, M.; Cai, X.; Chaney, N.

    2016-12-01

    Big data, high performance computing, and recent advances in hydrologic similarity present a unique opportunity for macroscale hydrology: the land surface hydrology can be modeled at field scales over continental extents while ensuring computational efficiency to enable robust ensemble frameworks. In this presentation we will illustrate this potential breakthrough in macroscale hydrology by discussing results from a 30-meter simulation over the contiguous United States using the HydroBlocks land surface model. HydroBlocks is a novel land surface model that represents field-scale spatial heterogeneity of land surface processes through interacting hydrologic response units (HRUs) [Chaney et al., 2016]. The model is a coupling between the Noah-MP land surface model and the Dynamic TOPMODEL hydrologic model. The HRUs are defined by clustering proxies of the drivers of spatial heterogeneity using hyperresolution land data. For the simulations over CONUS, HydroBlocks is run at every HUC10 catchment using 100 HRUs per catchment between 2004 and 2014. The simulations are forced with the 4 km Stage IV radar rainfall product and a spatially downscaled version of NLDAS-2. We will show how this approach to macroscale hydrology ensures computational efficiency while providing field-scale hydrologic information over continental extents. We will illustrate how this approach provides a novel approach in both the application and validation of macroscale land surface and hydrologic models. Finally, using these results, we will discuss the important role that big data and high performance computing can play in providing solutions to longstanding challenges to not only flood and drought monitoring systems but also to numerical weather prediction, seasonal forecasting, and climate prediction. References Chaney, N., P. Metcalfe, and E. F. Wood (2016), HydroBlocks: A Field-scale Resolving Land Surface Model for Application Over Continental Extents, Hydrological Processes, (in press.)

  1. Remote Maintenance Design Guide for Compact Processing Units

    Energy Technology Data Exchange (ETDEWEB)

    Draper, J.V.

    2000-07-13

    Oak Ridge National Laboratory (ORNL) Robotics and Process Systems (RPSD) personnel have extensive experience working with remotely operated and maintained systems. These systems require expert knowledge in teleoperation, human factors, telerobotics, and other robotic devices so that remote equipment may be manipulated, operated, serviced, surveyed, and moved about in a hazardous environment. The RPSD staff has a wealth of experience in this area, including knowledge in the broad topics of human factors, modular electronics, modular mechanical systems, hardware design, and specialized tooling. Examples of projects that illustrate and highlight RPSD's unique experience in remote systems design and application include the following: (1) design of a remote shear and remote dissolver systems in support of U.S. Department of Energy (DOE) fuel recycling research and nuclear power missions; (2) building remotely operated mobile systems for metrology and characterizing hazardous facilities in support of remote operations within those facilities; (3) construction of modular robotic arms, including the Laboratory Telerobotic Manipulator, which was designed for the National Aeronautics and Space Administration (NASA) and the Advanced ServoManipulator, which was designed for the DOE; (4) design of remotely operated laboratories, including chemical analysis and biochemical processing laboratories; (5) construction of remote systems for environmental clean up and characterization, including underwater, buried waste, underground storage tank (UST) and decontamination and dismantlement (D&D) applications. Remote maintenance has played a significant role in fuel reprocessing because of combined chemical and radiological contamination. Furthermore, remote maintenance is expected to play a strong role in future waste remediation. The compact processing units (CPUs) being designed for use in underground waste storage tank remediation are examples of improvements in systems

  2. A Block-Asynchronous Relaxation Method for Graphics Processing Units

    Energy Technology Data Exchange (ETDEWEB)

    Antz, Hartwig [Karlsruhe Inst. of Technology (KIT) (Germany); Tomov, Stanimire [Univ. of Tennessee, Knoxville, TN (United States); Dongarra, Jack [Univ. of Tennessee, Knoxville, TN (United States); Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Univ. of Manchester (United Kingdom); Heuveline, Vincent [Karlsruhe Inst. of Technology (KIT) (Germany)

    2011-11-30

    In this paper, we analyze the potential of asynchronous relaxation methods on Graphics Processing Units (GPUs). For this purpose, we developed a set of asynchronous iteration algorithms in CUDA and compared them with a parallel implementation of synchronous relaxation methods on CPU-based systems. For a set of test matrices taken from the University of Florida Matrix Collection we monitor the convergence behavior, the average iteration time and the total time-to-solution time. Analyzing the results, we observe that even for our most basic asynchronous relaxation scheme, despite its lower convergence rate compared to the Gauss-Seidel relaxation (that we expected), the asynchronous iteration running on GPUs is still able to provide solution approximations of certain accuracy in considerably shorter time then Gauss- Seidel running on CPUs. Hence, it overcompensates for the slower convergence by exploiting the scalability and the good fit of the asynchronous schemes for the highly parallel GPU architectures. Further, enhancing the most basic asynchronous approach with hybrid schemes – using multiple iterations within the ”subdomain” handled by a GPU thread block and Jacobi-like asynchronous updates across the ”boundaries”, subject to tuning various parameters – we manage to not only recover the loss of global convergence but often accelerate convergence of up to two times (compared to the effective but difficult to parallelize Gauss-Seidel type of schemes), while keeping the execution time of a global iteration practically the same. This shows the high potential of the asynchronous methods not only as a stand alone numerical solver for linear systems of equations fulfilling certain convergence conditions but more importantly as a smoother in multigrid methods. Due to the explosion of parallelism in todays architecture designs, the significance and the need for asynchronous methods, as the ones described in this work, is expected to grow.

  3. Flocking-based Document Clustering on the Graphics Processing Unit

    Energy Technology Data Exchange (ETDEWEB)

    Cui, Xiaohui [ORNL; Potok, Thomas E [ORNL; Patton, Robert M [ORNL; ST Charles, Jesse Lee [ORNL

    2008-01-01

    Abstract?Analyzing and grouping documents by content is a complex problem. One explored method of solving this problem borrows from nature, imitating the flocking behavior of birds. Each bird represents a single document and flies toward other documents that are similar to it. One limitation of this method of document clustering is its complexity O(n2). As the number of documents grows, it becomes increasingly difficult to receive results in a reasonable amount of time. However, flocking behavior, along with most naturally inspired algorithms such as ant colony optimization and particle swarm optimization, are highly parallel and have found increased performance on expensive cluster computers. In the last few years, the graphics processing unit (GPU) has received attention for its ability to solve highly-parallel and semi-parallel problems much faster than the traditional sequential processor. Some applications see a huge increase in performance on this new platform. The cost of these high-performance devices is also marginal when compared with the price of cluster machines. In this paper, we have conducted research to exploit this architecture and apply its strengths to the document flocking problem. Our results highlight the potential benefit the GPU brings to all naturally inspired algorithms. Using the CUDA platform from NIVIDA? we developed a document flocking implementation to be run on the NIVIDA?GEFORCE 8800. Additionally, we developed a similar but sequential implementation of the same algorithm to be run on a desktop CPU. We tested the performance of each on groups of news articles ranging in size from 200 to 3000 documents. The results of these tests were very significant. Performance gains ranged from three to nearly five times improvement of the GPU over the CPU implementation. This dramatic improvement in runtime makes the GPU a potentially revolutionary platform for document clustering algorithms.

  4. Handling geophysical flows: Numerical modelling using Graphical Processing Units

    Science.gov (United States)

    Garcia-Navarro, Pilar; Lacasta, Asier; Juez, Carmelo; Morales-Hernandez, Mario

    2016-04-01

    Computational tools may help engineers in the assessment of sediment transport during the decision-making processes. The main requirements are that the numerical results have to be accurate and simulation models must be fast. The present work is based on the 2D shallow water equations in combination with the 2D Exner equation [1]. The resulting numerical model accuracy was already discussed in previous work. Regarding the speed of the computation, the Exner equation slows down the already costly 2D shallow water model as the number of variables to solve is increased and the numerical stability is more restrictive. On the other hand, the movement of poorly sorted material over steep areas constitutes a hazardous environmental problem. Computational tools help in the predictions of such landslides [2]. In order to overcome this problem, this work proposes the use of Graphical Processing Units (GPUs) for decreasing significantly the simulation time [3, 4]. The numerical scheme implemented in GPU is based on a finite volume scheme. The mathematical model and the numerical implementation are compared against experimental and field data. In addition, the computational times obtained with the Graphical Hardware technology are compared against Single-Core (sequential) and Multi-Core (parallel) CPU implementations. References [Juez et al.(2014)] Juez, C., Murillo, J., & Garca-Navarro, P. (2014) A 2D weakly-coupled and efficient numerical model for transient shallow flow and movable bed. Advances in Water Resources. 71 93-109. [Juez et al.(2013)] Juez, C., Murillo, J., & Garca-Navarro, P. (2013) . 2D simulation of granular flow over irregular steep slopes using global and local coordinates. Journal of Computational Physics. 225 166-204. [Lacasta et al.(2014)] Lacasta, A., Morales-Hernndez, M., Murillo, J., & Garca-Navarro, P. (2014) An optimized GPU implementation of a 2D free surface simulation model on unstructured meshes Advances in Engineering Software. 78 1-15. [Lacasta

  5. Viscoelastic Finite Difference Modeling Using Graphics Processing Units

    Science.gov (United States)

    Fabien-Ouellet, G.; Gloaguen, E.; Giroux, B.

    2014-12-01

    Full waveform seismic modeling requires a huge amount of computing power that still challenges today's technology. This limits the applicability of powerful processing approaches in seismic exploration like full-waveform inversion. This paper explores the use of Graphics Processing Units (GPU) to compute a time based finite-difference solution to the viscoelastic wave equation. The aim is to investigate whether the adoption of the GPU technology is susceptible to reduce significantly the computing time of simulations. The code presented herein is based on the freely accessible software of Bohlen (2002) in 2D provided under a General Public License (GNU) licence. This implementation is based on a second order centred differences scheme to approximate time differences and staggered grid schemes with centred difference of order 2, 4, 6, 8, and 12 for spatial derivatives. The code is fully parallel and is written using the Message Passing Interface (MPI), and it thus supports simulations of vast seismic models on a cluster of CPUs. To port the code from Bohlen (2002) on GPUs, the OpenCl framework was chosen for its ability to work on both CPUs and GPUs and its adoption by most of GPU manufacturers. In our implementation, OpenCL works in conjunction with MPI, which allows computations on a cluster of GPU for large-scale model simulations. We tested our code for model sizes between 1002 and 60002 elements. Comparison shows a decrease in computation time of more than two orders of magnitude between the GPU implementation run on a AMD Radeon HD 7950 and the CPU implementation run on a 2.26 GHz Intel Xeon Quad-Core. The speed-up varies depending on the order of the finite difference approximation and generally increases for higher orders. Increasing speed-ups are also obtained for increasing model size, which can be explained by kernel overheads and delays introduced by memory transfers to and from the GPU through the PCI-E bus. Those tests indicate that the GPU memory size

  6. Implementation and adaptation of a macro-scale methodology to calculate direct economic losses

    Science.gov (United States)

    Natho, Stephanie; Thieken, Annegret

    2017-04-01

    As one of the 195 member countries of the United Nations, Germany signed the Sendai Framework for Disaster Risk Reduction 2015-2030 (SFDRR). With this, though voluntary and non-binding, Germany agreed to report on achievements to reduce disaster impacts. Among other targets, the SFDRR aims at reducing direct economic losses in relation to the global gross domestic product by 2030 - but how to measure this without a standardized approach? The United Nations Office for Disaster Risk Reduction (UNISDR) has hence proposed a methodology to estimate direct economic losses per event and country on the basis of the number of damaged or destroyed items in different sectors. The method bases on experiences from developing countries. However, its applicability in industrial countries has not been investigated so far. Therefore, this study presents the first implementation of this approach in Germany to test its applicability for the costliest natural hazards and suggests adaptations. The approach proposed by UNISDR considers assets in the sectors agriculture, industry, commerce, housing, and infrastructure by considering roads, medical and educational facilities. The asset values are estimated on the basis of sector and event specific number of affected items, sector specific mean sizes per item, their standardized construction costs per square meter and a loss ratio of 25%. The methodology was tested for the three costliest natural hazard types in Germany, i.e. floods, storms and hail storms, considering 13 case studies on the federal or state scale between 1984 and 2016. Not any complete calculation of all sectors necessary to describe the total direct economic loss was possible due to incomplete documentation. Therefore, the method was tested sector-wise. Three new modules were developed to better adapt this methodology to German conditions covering private transport (cars), forestry and paved roads. Unpaved roads in contrast were integrated into the agricultural and

  7. Spatial variation in nutrient and water color effects on lake chlorophyll at macroscales

    Science.gov (United States)

    Fergus, C. Emi; Finley, Andrew O.; Soranno, Patricia A.; Wagner, Tyler

    2016-01-01

    positive effect such that a unit increase in water color resulted in a 2 μg/L increase in CHL and other locations where it had a negative effect such that a unit increase in water color resulted in a 2 μg/L decrease in CHL. In addition, the spatial scales that captured variation in TP and water color effects were different for our study lakes. Variation in TP–CHL relationships was observed at intermediate distances (~20 km) compared to variation in water color–CHL relationships that was observed at regional distances (~200 km). These results demonstrate that there are lake-to-lake differences in the effects of TP and water color on lake CHL and that this variation is spatially structured. Quantifying spatial structure in these relationships furthers our understanding of the variability in these relationships at macroscales and would improve model prediction of chlorophyll a to better meet lake management goals.

  8. Fast Pyrolysis Process Development Unit for Validating Bench Scale Data

    Energy Technology Data Exchange (ETDEWEB)

    Brown, Robert C. [Iowa State Univ., Ames, IA (United States). Biorenewables Research Lab.. Center for Sustainable Environmental Technologies. Bioeconomy Inst.; Jones, Samuel T. [Iowa State Univ., Ames, IA (United States). Biorenewables Research Lab.. Center for Sustainable Environmental Technologies. Bioeconomy Inst.

    2010-03-31

    The purpose of this project was to prepare and operate a fast pyrolysis process development unit (PDU) that can validate experimental data generated at the bench scale. In order to do this, a biomass preparation system, a modular fast pyrolysis fluidized bed reactor, modular gas clean-up systems, and modular bio-oil recovery systems were designed and constructed. Instrumentation for centralized data collection and process control were integrated. The bio-oil analysis laboratory was upgraded with the addition of analytical equipment needed to measure C, H, O, N, S, P, K, and Cl. To provide a consistent material for processing through the fluidized bed fast pyrolysis reactor, the existing biomass preparation capabilities of the ISU facility needed to be upgraded. A stationary grinder was installed to reduce biomass from bale form to 5-10 cm lengths. A 25 kg/hr rotary kiln drier was installed. It has the ability to lower moisture content to the desired level of less than 20% wt. An existing forage chopper was upgraded with new screens. It is used to reduce biomass to the desired particle size of 2-25 mm fiber length. To complete the material handling between these pieces of equipment, a bucket elevator and two belt conveyors must be installed. The bucket elevator has been installed. The conveyors are being procured using other funding sources. Fast pyrolysis bio-oil, char and non-condensable gases were produced from an 8 kg/hr fluidized bed reactor. The bio-oil was collected in a fractionating bio-oil collection system that produced multiple fractions of bio-oil. This bio-oil was fractionated through two separate, but equally important, mechanisms within the collection system. The aerosols and vapors were selectively collected by utilizing laminar flow conditions to prevent aerosol collection and electrostatic precipitators to collect the aerosols. The vapors were successfully collected through a selective condensation process. The combination of these two mechanisms

  9. Calculation of HELAS amplitudes for QCD processes using graphics processing unit (GPU)

    CERN Document Server

    Hagiwara, K; Okamura, N; Rainwater, D L; Stelzer, T

    2009-01-01

    We use a graphics processing unit (GPU) for fast calculations of helicity amplitudes of quark and gluon scattering processes in massless QCD. New HEGET ({\\bf H}ELAS {\\bf E}valuation with {\\bf G}PU {\\bf E}nhanced {\\bf T}echnology) codes for gluon self-interactions are introduced, and a C++ program to convert the MadGraph generated FORTRAN codes into HEGET codes in CUDA (a C-platform for general purpose computing on GPU) is created. Because of the proliferation of the number of Feynman diagrams and the number of independent color amplitudes, the maximum number of final state jets we can evaluate on a GPU is limited to 4 for pure gluon processes ($gg\\to 4g$), or 5 for processes with one or more quark lines such as $q\\bar{q}\\to 5g$ and $qq\\to qq+3g$. Compared with the usual CPU-based programs, we obtain 60-100 times better performance on the GPU, except for 5-jet production processes and the $gg\\to 4g$ processes for which the GPU gain over the CPU is about 20.

  10. Design of the Laboratory-Scale Plutonium Oxide Processing Unit in the Radiochemical Processing Laboratory

    Energy Technology Data Exchange (ETDEWEB)

    Lumetta, Gregg J. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Meier, David E. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Tingey, Joel M. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Casella, Amanda J. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Delegard, Calvin H. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Edwards, Matthew K. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Orton, Robert D. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Rapko, Brian M. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Smart, John E. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2015-05-01

    This report describes a design for a laboratory-scale capability to produce plutonium oxide (PuO2) for use in identifying and validating nuclear forensics signatures associated with plutonium production, as well as for use as exercise and reference materials. This capability will be located in the Radiochemical Processing Laboratory at the Pacific Northwest National Laboratory. The key unit operations are described, including PuO2 dissolution, purification of the Pu by ion exchange, precipitation, and re-conversion to PuO2 by calcination.

  11. Reproducibility of Mammography Units, Film Processing and Quality Imaging

    Science.gov (United States)

    Gaona, Enrique

    2003-09-01

    The purpose of this study was to carry out an exploratory survey of the problems of quality control in mammography and processors units as a diagnosis of the current situation of mammography facilities. Measurements of reproducibility, optical density, optical difference and gamma index are included. Breast cancer is the most frequently diagnosed cancer and is the second leading cause of cancer death among women in the Mexican Republic. Mammography is a radiographic examination specially designed for detecting breast pathology. We found that the problems of reproducibility of AEC are smaller than the problems of processors units because almost all processors fall outside of the acceptable variation limits and they can affect the mammography quality image and the dose to breast. Only four mammography units agree with the minimum score established by ACR and FDA for the phantom image.

  12. Improving the Quotation Process of an After-Sales Unit

    OpenAIRE

    Matilainen, Janne

    2013-01-01

    The purpose of this study was to model and analyze the quotation process of area managers at a global company. Process improvement requires understanding the fundamentals of the process. The study was conducted a case study. Data comprised of internal documentation of the case company, literature, and semi-structured, themed interviews of process performers and stakeholders. The objective was to produce model of the current state of the process. The focus was to establish a holistic view o...

  13. Co-occurrence of Photochemical and Microbiological Transformation Processes in Open-Water Unit Process Wetlands.

    Science.gov (United States)

    Prasse, Carsten; Wenk, Jannis; Jasper, Justin T; Ternes, Thomas A; Sedlak, David L

    2015-12-15

    The fate of anthropogenic trace organic contaminants in surface waters can be complex due to the occurrence of multiple parallel and consecutive transformation processes. In this study, the removal of five antiviral drugs (abacavir, acyclovir, emtricitabine, lamivudine and zidovudine) via both bio- and phototransformation processes, was investigated in laboratory microcosm experiments simulating an open-water unit process wetland receiving municipal wastewater effluent. Phototransformation was the main removal mechanism for abacavir, zidovudine, and emtricitabine, with half-lives (t1/2,photo) in wetland water of 1.6, 7.6, and 25 h, respectively. In contrast, removal of acyclovir and lamivudine was mainly attributable to slower microbial processes (t1/2,bio = 74 and 120 h, respectively). Identification of transformation products revealed that bio- and phototransformation reactions took place at different moieties. For abacavir and zidovudine, rapid transformation was attributable to high reactivity of the cyclopropylamine and azido moieties, respectively. Despite substantial differences in kinetics of different antiviral drugs, biotransformation reactions mainly involved oxidation of hydroxyl groups to the corresponding carboxylic acids. Phototransformation rates of parent antiviral drugs and their biotransformation products were similar, indicating that prior exposure to microorganisms (e.g., in a wastewater treatment plant or a vegetated wetland) would not affect the rate of transformation of the part of the molecule susceptible to phototransformation. However, phototransformation strongly affected the rates of biotransformation of the hydroxyl groups, which in some cases resulted in greater persistence of phototransformation products.

  14. Unit Operation Experiment Linking Classroom with Industrial Processing

    Science.gov (United States)

    Benson, Tracy J.; Richmond, Peyton C.; LeBlanc, Weldon

    2013-01-01

    An industrial-type distillation column, including appropriate pumps, heat exchangers, and automation, was used as a unit operations experiment to provide a link between classroom teaching and real-world applications. Students were presented with an open-ended experiment where they defined the testing parameters to solve a generalized problem. The…

  15. Effect of energetic dissipation processes on the friction unit tribological

    Directory of Open Access Journals (Sweden)

    Moving V. V.

    2007-01-01

    Full Text Available In article presented temperature influence on reological and fric-tion unit coefficients cast iron elements. It has been found that surface layer formed in the temperature friction has good rub off resistance. The surface layer structural hardening and capacity stress relaxation make up.

  16. On the hazard rate process for imperfectly monitored multi-unit systems

    Energy Technology Data Exchange (ETDEWEB)

    Barros, A. [Institut des Sciences et Techonologies de l' Information de Troyes (ISTIT-CNRS), Equipe de Modelisation et Surete des Systemes, Universite de Technologie de Troyes (UTT), 12, rue Marie Curie, BP2060, 10010 Troyes cedex (France)]. E-mail: anne.barros@utt.fr; Berenguer, C. [Institut des Sciences et Techonologies de l' Information de Troyes (ISTIT-CNRS), Equipe de Modelisation et Surete des Systemes, Universite de Technologie de Troyes (UTT), 12, rue Marie Curie, BP2060, 10010 Troyes cedex (France); Grall, A. [Institut des Sciences et Techonologies de l' Information de Troyes (ISTIT-CNRS), Equipe de Modelisation et Surete des Systemes, Universite de Technologie de Troyes (UTT), 12, rue Marie Curie, BP2060, 10010 Troyes cedex (France)

    2005-12-01

    The aim of this paper is to present a stochastic model to characterize the failure distribution of multi-unit systems when the current units state is imperfectly monitored. The definition of the hazard rate process existing with perfect monitoring is extended to the realistic case where the units failure time are not always detected (non-detection events). The so defined observed hazard rate process gives a better representation of the system behavior than the classical failure rate calculated without any information on the units state and than the hazard rate process based on perfect monitoring information. The quality of this representation is, however, conditioned by the monotony property of the process. This problem is mainly discussed and illustrated on a practical example (two parallel units). The results obtained motivate the use of the observed hazard rate process to characterize the stochastic behavior of the multi-unit systems and to optimize for example preventive maintenance policies.

  17. The United States Military Entrance Processing Command (USMEPCOM) Uses Six Sigma Process to Develop and Improve Data Quality

    Science.gov (United States)

    2007-06-01

    mecpom.army.mil Original title on 712 A/B: The United States Military Entrance Processing Command (USMEPCOM) uses Six Sigma process to develop and...Entrance Processing Command (USMEPCOM) uses Six Sigma Process to Develop and Improve Data Quality 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM...Processing Command (USMEPCOM) uses Six Sigma Process to Develop and Improve Data Quality 3 • USMEPCOM Overview/History • Purpose • Define: What is Important

  18. Nanoarchitecturing of Natural Melanin Nanospheres by Layer-by-Layer Assembly: Macroscale Anti-inflammatory Conductive Coatings with Optoelectronic Tunability.

    Science.gov (United States)

    Eom, Taesik; Woo, Kyungbae; Cho, Whirang; Heo, Jae Eun; Jang, Daseul; Shin, Jae In; Martin, David C; Wie, Jeong Jae; Shim, Bong Sup

    2017-06-12

    Natural melanins are biocompatible conductors with versatile functionalities. Here, we report fabrication of multifunctional poly(vinyl alcohol)/melanin nanocomposites by layer-by-layer (LBL) assembly using melanin nanoparticles (MNPs) directly extracted from sepia officinalis inks. The LBL assembly offers facile manipulation of nanotextures as well as nm-thickness control of the macroscale film by varying solvent qualities. The time-resolved absorption was monitored during the process and quantitatively studied by fractal dimension and lacunarity analysis. The capability of nanoarchitecturing provides confirmation of complete monolayer formation and leads to tunable iridescent reflective colors of the MNP films. In addition, the MNP films have durable electrochemical conductivities as evidenced by enhanced charge storage capacities for 1000 cycles. Moreover, the MNP covered ITO (indium tin oxide) substrates significantly reduced secretion of inflammatory cytokines, TNF-α, by raw 264.7 macrophage cells compared to bare ITO, by a factor of 5 and 1.8 with and without lipopolysaccharide endotoxins, respectively. These results highlight the optoelectronic device-level tunability along with the anti-inflammatory biocompatibility of the MNP LBL film. This combination of performance should make these films particularly interesting for bioelectronic device applications such as electroceuticals, artificial bionic organs, biosensors, and implantable devices.

  19. Amorphous Solid Water (ASW): Macroscale Environmentally-Neutral Application for Remediation of Hazardous Pollutants using Condensed-Phase Cryogenic Fluids

    Science.gov (United States)

    de Strulle, Ronald; Rheinhart, Maximilian

    2012-03-01

    We report macroscale environmentally-neutral use of cryogenic fluids to induce phase transitions from crystalline water-ices to amorphous solid water (ASW). New IP and uses in remediation of oil-spills and hazardous immiscibles from aquatic environments. We display high-resolution images of the transitions from hexagonal to cubic crystalline water-ice, then to hydrophobic ASW. Accretion and encapsulation of viscous pollutants within crystalline water-ice, and sequestration of condensed volatiles (PAH, methane) and low viscosity fluids within the interstitial cavities of ASW are shown and differentiated for: crude oils, diesel (heating) and blended oils, petroleum byproducts, vegetable and mineral oils, lipids, and light immiscible fluids. The effects of PdV work and thermal energy transfers during phase changes are shown, along with the sequestration efficiencies for hexagonal and cubic ice lattices vs. non-crystalline ASW, for a range of pollutant substances. The viability of ASW as a medium for study of quantum criticality phases is also proposed. The process is environmentally-neutral in that only substantially condensed-phase air liquefaction products, e.g. nitrogen in >90% liquid phase are employed as an active agent. The applications are also presented in terms of the scale-up of experiments performed at the nanoscale.

  20. Effects of macro-scale uncertainties on the imaging and automatic manipulation of nanoparticles

    Energy Technology Data Exchange (ETDEWEB)

    Korayem, M. H., E-mail: hkorayem@iust.ac.ir; Sadeghzadeh, S.; Homayooni, A. [Iran University of Science and Technology, Robotic Research Laboratory, School of Mechanical Engineering (Iran, Islamic Republic of)

    2013-01-15

    The steering, positioning, and fabrication operations in nano scale have been hampered by the uncertainties which come from the macro parts of nano-positioners. Among those uncertainties, the nonlinearities of piezo scanners have the highest contribution, which should be identified and compensated. On the other hand, the recognition of the effects of macro-scale nonlinearities on small-scale dynamics requires the simultaneous consideration of both the macro- and small-scale dynamics. This necessitates the implementation of multi-scale methods. In this article, a fixed interfacial multi-scale method (FIMM) that includes the effects of hysteresis has been used for the computationally and mathematically efficient modeling of nano-positioners. This method presents an improved coupling approach that can be used to study the imaging and manipulation of nanoparticles (from one to several hundred nanometers in diameter) subjected to nonlinear as well as linear positioning schemes. After comparing the applied hysteresis model with some previous experimental works, the dynamics of imaging and automatic manipulation of nanoparticles have been studied and some useful results have been presented. This paper opens a new window to the recognition and compensation of the errors of macro-scale nonlinearities imposed on small-scale dynamics.

  1. From the Nano- to the Macroscale - Bridging Scales for the Moving Contact Line Problem

    Science.gov (United States)

    Nold, Andreas; Sibley, David; Goddard, Benjamin; Kalliadasis, Serafim; Complex Multiscale Systems Team

    2016-11-01

    The moving contact line problem remains an unsolved fundamental problem in fluid mechanics. At the heart of the problem is its multiscale nature: a nanoscale region close to the solid boundary where the continuum hypothesis breaks down, must be resolved before effective macroscale parameters such as contact line friction and slip can be obtained. To capture nanoscale properties very close to the contact line and to establish a link to the macroscale behaviour, we employ classical density-functional theory (DFT), in combination with extended Navier-Stokes-like equations. Using simple models for viscosity and slip at the wall, we compare our computations with the Molecular Kinetic Theory, by extracting the contact line friction, depending on the imposed temperature of the fluid. A key fluid property captured by DFT is the fluid layering at the wall-fluid interface, which has a large effect on the shearing properties of a fluid. To capture this crucial property, we propose an anisotropic model for the viscosity, which also allows us to scrutinize the effect of fluid layering on contact line friction.

  2. Meshing complex macro-scale objects into self-assembling bricks

    Science.gov (United States)

    Hacohen, Adar; Hanniel, Iddo; Nikulshin, Yasha; Wolfus, Shuki; Abu-Horowitz, Almogit; Bachelet, Ido

    2015-07-01

    Self-assembly provides an information-economical route to the fabrication of objects at virtually all scales. However, there is no known algorithm to program self-assembly in macro-scale, solid, complex 3D objects. Here such an algorithm is described, which is inspired by the molecular assembly of DNA, and based on bricks designed by tetrahedral meshing of arbitrary objects. Assembly rules are encoded by topographic cues imprinted on brick faces while attraction between bricks is provided by embedded magnets. The bricks can then be mixed in a container and agitated, leading to properly assembled objects at high yields and zero errors. The system and its assembly dynamics were characterized by video and audio analysis, enabling the precise time- and space-resolved characterization of its performance and accuracy. Improved designs inspired by our system could lead to successful implementation of self-assembly at the macro-scale, allowing rapid, on-demand fabrication of objects without the need for assembly lines.

  3. A Ten-Step Process for Developing Teaching Units

    Science.gov (United States)

    Butler, Geoffrey; Heslup, Simon; Kurth, Lara

    2015-01-01

    Curriculum design and implementation can be a daunting process. Questions quickly arise, such as who is qualified to design the curriculum and how do these people begin the design process. According to Graves (2008), in many contexts the design of the curriculum and the implementation of the curricular product are considered to be two mutually…

  4. 76 FR 13973 - United States Warehouse Act; Processed Agricultural Products Licensing Agreement

    Science.gov (United States)

    2011-03-15

    ... Farm Service Agency United States Warehouse Act; Processed Agricultural Products Licensing Agreement... warehouse licenses may be issued under the United States Warehouse Act (USWA). Through this notice, FSA is... processed agricultural products that are stored in climate controlled, cooler, and freezer warehouses....

  5. A macro-scale perspective on within-farm management: how climate and topography alter the effect of farming practices.

    Science.gov (United States)

    Amano, Tatsuya; Kusumoto, Yoshinobu; Okamura, Hiroshi; Baba, Yuki G; Hamasaki, Kenji; Tanaka, Koichi; Yamamoto, Shori

    2011-12-01

    Organic farming has the potential to reverse biodiversity loss in farmland and benefit agriculture by enhancing ecosystem services. Although the mixed success of organic farming in enhancing biodiversity has been attributed to differences in taxa and landscape context, no studies have focused on the effect of macro-scale factors such as climate and topography. This study provides the first assessment of the impact of macro-scale factors on the effectiveness of within-farm management on biodiversity, using spiders in Japan as an example. A multilevel modelling approach revealed that reducing pesticide applications increases spider abundance, particularly in areas with high precipitation, which were also associated with high potential spider abundance. Using the model we identified areas throughout Japan that can potentially benefit from organic farming. The alteration of local habitat-abundance relations by macro-scale factors could explain the reported low spatial generality in the effects of organic farming and patterns of habitat association.

  6. Reactive-Separator Process Unit for Lunar Regolith Project

    Data.gov (United States)

    National Aeronautics and Space Administration — NASA's plans for a lunar habitation outpost call out for process technologies to separate hydrogen sulfide and sulfur dioxide gases from regolith product gas...

  7. Evidence of a sensory processing unit in the mammalian macula

    Science.gov (United States)

    Chimento, T. C.; Ross, M. D.

    1996-01-01

    We cut serial sections through the medial part of the rat vestibular macula for transmission electron microscopic (TEM) examination, computer-assisted 3-D reconstruction, and compartmental modeling. The ultrastructural research showed that many primary vestibular neurons have an unmyelinated segment, often branched, that extends between the heminode (putative site of the spike initiation zone) and the expanded terminal(s) (calyx, calyces). These segments, termed the neuron branches, and the calyces frequently have spine-like processes of various dimensions with bouton endings that morphologically are afferent, efferent, or reciprocal to other macular neural elements. The major questions posed by this study were whether small details of morphology, such as the size and location of neuronal processes or synapses, could influence the output of a vestibular afferent, and whether a knowledge of morphological details could guide the selection of values for simulation parameters. The conclusions from our simulations are (1) values of 5.0 k omega cm2 for membrane resistivity and 1.0 nS for synaptic conductance yield simulations that best match published physiological results; (2) process morphology has little effect on orthodromic spread of depolarization from the head (bouton) to the spike initiation zone (SIZ); (3) process morphology has no effect on antidromic spread of depolarization to the process head; (4) synapses do not sum linearly; (5) synapses are electrically close to the SIZ; and (6) all whole-cell simulations should be run with an active SIZ.

  8. Option pricing with COS method on Graphics Processing Units

    NARCIS (Netherlands)

    B. Zhang (Bo); C.W. Oosterlee (Cornelis)

    2009-01-01

    htmlabstractIn this paper, acceleration on the GPU for option pricing by the COS method is demonstrated. In particular, both European and Bermudan options will be discussed in detail. For Bermudan options, we consider both the Black-Scholes model and Levy processes of infinite activity. Moreover, th

  9. Option pricing with COS method on Graphics Processing Units

    NARCIS (Netherlands)

    Zhang, B.; Oosterlee, C.W.

    2009-01-01

    In this paper, acceleration on the GPU for option pricing by the COS method is demonstrated. In particular, both European and Bermudan options will be discussed in detail. For Bermudan options, we consider both the Black-Scholes model and Levy processes of infinite activity. Moreover, the influence

  10. The Cilium: Cellular Antenna and Central Processing Unit

    OpenAIRE

    Malicki, Jarema J.; Johnson, Colin A.

    2017-01-01

    Cilia mediate an astonishing diversity of processes. Recent advances provide unexpected insights into the regulatory mechanisms of cilium formation, and reveal diverse regulatory inputs that are related to the cell cycle, cytoskeleton, proteostasis, and cilia-mediated signaling itself. Ciliogenesis and cilia maintenance are regulated by reciprocal antagonistic or synergistic influences, often acting in parallel to each other. By receiving parallel inputs, cilia appear to integrate multiple si...

  11. Uniting Gradual and Abrupt set Processes in Resistive Switching Oxides

    Science.gov (United States)

    Fleck, Karsten; La Torre, Camilla; Aslam, Nabeel; Hoffmann-Eifert, Susanne; Böttger, Ulrich; Menzel, Stephan

    2016-12-01

    Identifying limiting factors is crucial for a better understanding of the dynamics of the resistive switching phenomenon in transition-metal oxides. This improved understanding is important for the design of fast-switching, energy-efficient, and long-term stable redox-based resistive random-access memory devices. Therefore, this work presents a detailed study of the set kinetics of valence change resistive switches on a time scale from 10 ns to 104 s , taking Pt /SrTiO3/TiN nanocrossbars as a model material. The analysis of the transient currents reveals that the switching process can be subdivided into a linear-degradation process that is followed by a thermal runaway. The comparison with a dynamical electrothermal model of the memory cell allows the deduction of the physical origin of the degradation. The origin is an electric-field-induced increase of the oxygen-vacancy concentration near the Schottky barrier of the Pt /SrTiO3 interface that is accompanied by a steadily rising local temperature due to Joule heating. The positive feedback of the temperature increase on the oxygen-vacancy mobility, and thereby on the conductivity of the filament, leads to a self-acceleration of the set process.

  12. Predator-prey interactions as macro-scale drivers of species diversity in mammals

    DEFF Research Database (Denmark)

    Sandom, Christopher James; Sandel, Brody Steven; Dalby, Lars

    mechanistic drivers of mammal species richness at macro-scales for two trophic levels: predators and prey. To disentangle biotic (i.e. functional predator-prey interactions) from abiotic (i.e. environmental) and bottom-up from top-down determinants we considered three hypotheses: 1) environmental factors......-down). We gathered distributional range, mass and diet data for 4,091 terrestrial mammal species, excluding bats. Species richness maps were created for predators and prey and structural equation modelling was used to test the three hypotheses at continental and global scales. We also explored...... the importance of functional trait composition by analyzing richness of large and small mass categories for prey (division at 10 kg) and predators (division at 21.5 kg). Results/Conclusions Mammal species richness increased from the poles to the equator, supporting the classic latitudinal richness gradient...

  13. Analysis of Unit Process Cost for an Engineering-Scale Pyroprocess Facility Using a Process Costing Method in Korea

    Directory of Open Access Journals (Sweden)

    Sungki Kim

    2015-08-01

    Full Text Available Pyroprocessing, which is a dry recycling method, converts spent nuclear fuel into U (Uranium/TRU (TRansUranium metal ingots in a high-temperature molten salt phase. This paper provides the unit process cost of a pyroprocess facility that can process up to 10 tons of pyroprocessing product per year by utilizing the process costing method. Toward this end, the pyroprocess was classified into four kinds of unit processes: pretreatment, electrochemical reduction, electrorefining and electrowinning. The unit process cost was calculated by classifying the cost consumed at each process into raw material and conversion costs. The unit process costs of the pretreatment, electrochemical reduction, electrorefining and electrowinning were calculated as 195 US$/kgU-TRU, 310 US$/kgU-TRU, 215 US$/kgU-TRU and 231 US$/kgU-TRU, respectively. Finally the total pyroprocess cost was calculated as 951 US$/kgU-TRU. In addition, the cost driver for the raw material cost was identified as the cost for Li3PO4, needed for the LiCl-KCl purification process, and platinum as an anode electrode in the electrochemical reduction process.

  14. COSTS AND PROFITABILITY IN FOOD PROCESSING: PASTRY TYPE UNITS

    Directory of Open Access Journals (Sweden)

    DUMITRANA MIHAELA

    2013-08-01

    Full Text Available For each company, profitability, products quality and customer satisfaction are the most importanttargets. To attaint these targets, managers need to know all about costs that are used in decision making. Whatkind of costs? How these costs are calculated for a specific sector such as food processing? These are only a fewquestions with answers in our paper. We consider that a case study for this sector may be relevant for all peoplethat are interested to increase the profitability of this specific activity sector.

  15. ENTREPRENEURIAL OPPORTUNITIES IN FOOD PROCESSING UNITS (WITH SPECIAL REFERENCES TO BYADGI RED CHILLI COLD STORAGE UNITS IN THE KARNATAKA STATE

    Directory of Open Access Journals (Sweden)

    P. ISHWARA

    2010-01-01

    Full Text Available After the green revolution, we are now ushering in the evergreen revolution in the country; food processing is an evergreen activity. It is the key to the agricultural sector. In this paper an attempt has been made to study the workings of food processing units with special references to Red Chilli Cold Storage units in the Byadgi district of Karnataka State. Byadgi has been famous for Red Chilli since the days it’s of antiquity. The vast and extensive market yard in Byadagi taluk is famous as the second largest Red Chilli dealing market in the country. However, the most common and recurring problem faced by the farmer is inability to store enough red chilli from one harvest to another. Red chilli that was locally abundant for only a short period of time had to be stored against times of scarcity. In recent years, due to Oleoresin, demand for Red Chilli has grow from other countries like Sri Lanka, Bangladesh, America, Europe, Nepal, Indonesia, Mexico etc. The study reveals that all the cold storage units of the study area have been using vapour compression refrigeration system or method. All entrepreneurs have satisfied with their turnover and profit and they are in a good economic position. Even though the average turnover and profits are increased, few units have shown negligible amount of decrease in turnover and profit. This is due to the competition from increasing number of cold storages and early established units. The cold storages of the study area have been storing Red chilli, Chilli seeds, Chilli powder, Tamarind, Jeera, Dania, Turmeric, Sunflower, Zinger, Channa, Flower seeds etc,. But the 80 per cent of the each cold storage is filled by the red chilli this is due to the existence of vast and extensivered chilli market yard in the Byadgi. There is no business without problems. In the same way the entrepreneurs who are chosen for the study are facing a few problems in their business like skilled labour, technical and management

  16. The Cilium: Cellular Antenna and Central Processing Unit.

    Science.gov (United States)

    Malicki, Jarema J; Johnson, Colin A

    2017-02-01

    Cilia mediate an astonishing diversity of processes. Recent advances provide unexpected insights into the regulatory mechanisms of cilium formation, and reveal diverse regulatory inputs that are related to the cell cycle, cytoskeleton, proteostasis, and cilia-mediated signaling itself. Ciliogenesis and cilia maintenance are regulated by reciprocal antagonistic or synergistic influences, often acting in parallel to each other. By receiving parallel inputs, cilia appear to integrate multiple signals into specific outputs and may have functions similar to logic gates of digital systems. Some combinations of input signals appear to impose higher hierarchical control related to the cell cycle. An integrated view of these regulatory inputs will be necessary to understand ciliogenesis and its wider relevance to human biology. Copyright © 2016 The Authors. Published by Elsevier Ltd.. All rights reserved.

  17. Professional Competence of the Head of External Relations Unit and its Development in the Study Process

    OpenAIRE

    Turuševa, Larisa

    2010-01-01

    Dissertation Annotation Larisa Turuševa’s promotion paper „Professional Competence of the Head of External Relations Unit and its Development in the Study Process” is a fulfilled research on the development of Professional competence of the heads of external relations units, conditions for the study programme development. A model of professional competence of the head of external relations unit is worked out, its indicators and levels are described. A study process model for th...

  18. Low cost solar array project production process and equipment task. A Module Experimental Process System Development Unit (MEPSDU)

    Science.gov (United States)

    1981-01-01

    Technical readiness for the production of photovoltaic modules using single crystal silicon dendritic web sheet material is demonstrated by: (1) selection, design and implementation of solar cell and photovoltaic module process sequence in a Module Experimental Process System Development Unit; (2) demonstration runs; (3) passing of acceptance and qualification tests; and (4) achievement of a cost effective module.

  19. The Aluminum Deep Processing Project of North United Aluminum Landed in Qijiang

    Institute of Scientific and Technical Information of China (English)

    2014-01-01

    <正>On April 10,North United Aluminum Company respectively signed investment cooperation agreements with Qijiang Industrial Park and Qineng Electricity&Aluminum Co.,Ltd,signifying the landing of North United Aluminum’s aluminum deep processing project in Qijiang.

  20. Modellierung of meso- and macroscale river basins - a workshop held at Lauenburg; Modellierung in meso- bis makroskaligen Flusseinzugsgebieten - Tagungsband zum gleichnamigen Workshop

    Energy Technology Data Exchange (ETDEWEB)

    Sutmoeller, J.; Raschke, E. (eds.) [GKSS-Forschungszentrum Geesthacht GmbH (Germany). Inst. fuer Atmosphaerenphysik

    2001-07-01

    During the past decade measuring and modelling of global and regional processes that exchange energy and water in the climate system of the Earth became a focus in hydrological and meteorological research. Besides climate research many more applications will gain from this effort, e.g. as weather forecasting, water management and agriculture. As large scale weather and climate applications diversify to water related issues such as water resources, reservoir management, and flood and drought forecasting hydrologists and meteorologists are challenged to work interdisciplinary. The workshop 'Modelling of meso- and macroscale river basins' brought together various current aspects of this issue, ranging from coupled atmosphere-hydrology models to integrated river basin management to land use change. Recent results are introduced and summarised in this report. (orig.)

  1. Thermal/Heat Transfer Analysis Using a Graphic Processing Unit (GPU) Enabled Computing Environment Project

    Data.gov (United States)

    National Aeronautics and Space Administration — The objective of this project was to use GPU enabled computing to accelerate the analyses of heat transfer and thermal effects. Graphical processing unit (GPU)...

  2. Advanced In-Space Propulsion (AISP): High Temperature Boost Power Processing Unit (PPU) Project

    Data.gov (United States)

    National Aeronautics and Space Administration — The task is to investigate the technology path to develop a 10kW modular Silicon Carbide (SiC) based power processing unit (PPU). The PPU utilizes the high...

  3. Silicon Carbide (SiC) Power Processing Unit (PPU) for Hall Effect Thrusters Project

    Data.gov (United States)

    National Aeronautics and Space Administration — In this SBIR project, APEI, Inc. is proposing to develop a high efficiency, rad-hard 3.8 kW silicon carbide (SiC) Power Processing Unit (PPU) for Hall Effect...

  4. Performance Analysis of the United States Marine Corps War Reserve Materiel Program Process Flow

    Science.gov (United States)

    2016-12-01

    NAVAL POSTGRADUATE SCHOOL MONTEREY, CALIFORNIA MBA PROFESSIONAL REPORT PERFORMANCE ANALYSIS OF THE UNITED STATES MARINE CORPS...PERFORMANCE ANALYSIS OF THE UNITED STATES MARINE CORPS WAR RESERVE MATERIEL PROGRAM PROCESS FLOW 5. FUNDING NUMBERS 6. AUTHOR(S) Nathan A. Campbell...an item is requested but not maintained in the WRM inventory. By conducting a process analysis and using computer modeling, our recommendations are

  5. From micro-scale 3D simulations to macro-scale model of periodic porous media

    Science.gov (United States)

    Crevacore, Eleonora; Tosco, Tiziana; Marchisio, Daniele; Sethi, Rajandrea; Messina, Francesca

    2015-04-01

    In environmental engineering, the transport of colloidal suspensions in porous media is studied to understand the fate of potentially harmful nano-particles and to design new remediation technologies. In this perspective, averaging techniques applied to micro-scale numerical simulations are a powerful tool to extrapolate accurate macro-scale models. Choosing two simplified packing configurations of soil grains and starting from a single elementary cell (module), it is possible to take advantage of the periodicity of the structures to reduce the computation costs of full 3D simulations. Steady-state flow simulations for incompressible fluid in laminar regime are implemented. Transport simulations are based on the pore-scale advection-diffusion equation, that can be enriched introducing also the Stokes velocity (to consider the gravity effect) and the interception mechanism. Simulations are carried on a domain composed of several elementary modules, that serve as control volumes in a finite volume method for the macro-scale method. The periodicity of the medium involves the periodicity of the flow field and this will be of great importance during the up-scaling procedure, allowing relevant simplifications. Micro-scale numerical data are treated in order to compute the mean concentration (volume and area averages) and fluxes on each module. The simulation results are used to compare the micro-scale averaged equation to the integral form of the macroscopic one, making a distinction between those terms that could be computed exactly and those for which a closure in needed. Of particular interest it is the investigation of the origin of macro-scale terms such as the dispersion and tortuosity, trying to describe them with micro-scale known quantities. Traditionally, to study the colloidal transport many simplifications are introduced, such those concerning ultra-simplified geometry that usually account for a single collector. Gradual removal of such hypothesis leads to a

  6. Automated processing of whole blood units: operational value and in vitro quality of final blood components

    Science.gov (United States)

    Jurado, Marisa; Algora, Manuel; Garcia-Sanchez, Félix; Vico, Santiago; Rodriguez, Eva; Perez, Sonia; Barbolla, Luz

    2012-01-01

    Background The Community Transfusion Centre in Madrid currently processes whole blood using a conventional procedure (Compomat, Fresenius) followed by automated processing of buffy coats with the OrbiSac system (CaridianBCT). The Atreus 3C system (CaridianBCT) automates the production of red blood cells, plasma and an interim platelet unit from a whole blood unit. Interim platelet unit are pooled to produce a transfusable platelet unit. In this study the Atreus 3C system was evaluated and compared to the routine method with regards to product quality and operational value. Materials and methods Over a 5-week period 810 whole blood units were processed using the Atreus 3C system. The attributes of the automated process were compared to those of the routine method by assessing productivity, space, equipment and staffing requirements. The data obtained were evaluated in order to estimate the impact of implementing the Atreus 3C system in the routine setting of the blood centre. Yield and in vitro quality of the final blood components processed with the two systems were evaluated and compared. Results The Atreus 3C system enabled higher throughput while requiring less space and employee time by decreasing the amount of equipment and processing time per unit of whole blood processed. Whole blood units processed on the Atreus 3C system gave a higher platelet yield, a similar amount of red blood cells and a smaller volume of plasma. Discussion These results support the conclusion that the Atreus 3C system produces blood components meeting quality requirements while providing a high operational efficiency. Implementation of the Atreus 3C system could result in a large organisational improvement. PMID:22044958

  7. THE ASYMPTOTIC PROPERTIES OF SUPERCRITICAL BISEXUAL GALTON-WATSON BRANCHING PROCESSES WITH IMMIGRATION OF MATING UNITS

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    In this article the supercritical bisexual Galton-Watson branching processes with the immigration of mating units is considered. A necessary condition for the almost sure convergence, and a sufficient condition for the L1 convergence are given for the process with the suitably normed condition.

  8. A Framework for Smart Distribution of Bio-signal Processing Units in M-Health

    NARCIS (Netherlands)

    Mei, Hailiang; Widya, Ing; Broens, Tom; Pawar, Pravin; Halteren, van Aart; Shishkov, Boris; Sinderen, van Marten

    2007-01-01

    This paper introduces the Bio-Signal Processing Unit (BSPU) as a functional component that hosts (part of ) the bio-signal information processing algorithms that are needed for an m-health application. With our approach, the BSPUs can be dynamically assigned to available nodes between the bio-signal

  9. Characterization of suspended bacteria from processing units in an advanced drinking water treatment plant of China.

    Science.gov (United States)

    Wang, Feng; Li, Weiying; Zhang, Junpeng; Qi, Wanqi; Zhou, Yanyan; Xiang, Yuan; Shi, Nuo

    2017-05-01

    For the drinking water treatment plant (DWTP), the organic pollutant removal was the primary focus, while the suspended bacterial was always neglected. In this study, the suspended bacteria from each processing unit in a DWTP employing an ozone-biological activated carbon process was mainly characterized by using heterotrophic plate counts (HPCs), a flow cytometer, and 454-pyrosequencing methods. The results showed that an adverse changing tendency of HPC and total cell counts was observed in the sand filtration tank (SFT), where the cultivability of suspended bacteria increased to 34%. However, the cultivability level of other units stayed below 3% except for ozone contact tank (OCT, 13.5%) and activated carbon filtration tank (ACFT, 34.39%). It meant that filtration processes promoted the increase in cultivability of suspended bacteria remarkably, which indicated biodegrading capability. In the unit of OCT, microbial diversity indexes declined drastically, and the dominant bacteria were affiliated to Proteobacteria phylum (99.9%) and Betaproteobacteria class (86.3%), which were also the dominant bacteria in the effluent of other units. Besides, the primary genus was Limnohabitans in the effluents of SFT (17.4%) as well as ACFT (25.6%), which was inferred to be the crucial contributors for the biodegradable function in the filtration units. Overall, this paper provided an overview of community composition of each processing units in a DWTP as well as reference for better developing microbial function for drinking water treatment in the future.

  10. Acceleration of integral imaging based incoherent Fourier hologram capture using graphic processing unit.

    Science.gov (United States)

    Jeong, Kyeong-Min; Kim, Hee-Seung; Hong, Sung-In; Lee, Sung-Keun; Jo, Na-Young; Kim, Yong-Soo; Lim, Hong-Gi; Park, Jae-Hyeung

    2012-10-01

    Speed enhancement of integral imaging based incoherent Fourier hologram capture using a graphic processing unit is reported. Integral imaging based method enables exact hologram capture of real-existing three-dimensional objects under regular incoherent illumination. In our implementation, we apply parallel computation scheme using the graphic processing unit, accelerating the processing speed. Using enhanced speed of hologram capture, we also implement a pseudo real-time hologram capture and optical reconstruction system. The overall operation speed is measured to be 1 frame per second.

  11. Optimization Solutions for Improving the Performance of the Parallel Reduction Algorithm Using Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Ion LUNGU

    2012-01-01

    Full Text Available In this paper, we research, analyze and develop optimization solutions for the parallel reduction function using graphics processing units (GPUs that implement the Compute Unified Device Architecture (CUDA, a modern and novel approach for improving the software performance of data processing applications and algorithms. Many of these applications and algorithms make use of the reduction function in their computational steps. After having designed the function and its algorithmic steps in CUDA, we have progressively developed and implemented optimization solutions for the reduction function. In order to confirm, test and evaluate the solutions' efficiency, we have developed a custom tailored benchmark suite. We have analyzed the obtained experimental results regarding: the comparison of the execution time and bandwidth when using graphic processing units covering the main CUDA architectures (Tesla GT200, Fermi GF100, Kepler GK104 and a central processing unit; the data type influence; the binary operator's influence.

  12. Detecting benzoyl peroxide in wheat flour by line-scan macro-scale Raman chemical imaging

    Science.gov (United States)

    Qin, Jianwei; Kim, Moon S.; Chao, Kuanglin; Gonzalez, Maria; Cho, Byoung-Kwan

    2017-05-01

    Excessive use of benzoyl peroxide (BPO, a bleaching agent) in wheat flour can destroy flour nutrients and cause diseases to consumers. A macro-scale Raman chemical imaging method was developed for direct detection of BPO mixed in the wheat flour. A 785 nm line laser was used in a line-scan Hyperspectral Raman imaging system. Raman images were collected from wheat flour mixed with BPO at eight concentrations (w/w) from 50 to 6,400 ppm. A sample holder (150×100×2 mm3) was used to present a thin layer (2 mm thick) of the powdered sample for image acquisition. A baseline correction method was used to correct the fluctuating fluorescence signals from the wheat flour. To isolate BPO particles from the flour background, a simple thresholding method was applied to the single-band fluorescence-free images at a unique Raman peak wavenumber (i.e., 1001 cm-1) preselected for the BPO detection. Chemical images were created to detect and map the BPO particles. Limit of detection for the BPO was estimated in the order of 50 ppm, which is on the same level with regulatory standards.

  13. Linking microstructural evolution and macro-scale friction behavior in metals

    Science.gov (United States)

    Argibay, N.; Chandross, M.; Cheng, S.; Michael, J. R.

    2017-03-01

    A correlation is established between the macro-scale friction regimes of metals and a transition between two dominant atomistic mechanisms of deformation. Metals tend to exhibit bi-stable friction behavior -- low and converging or high and diverging. These general trends in behavior are shown to be largely explained using a simplified model based on grain size evolution, as a function of contact stress and temperature, and are demonstrated for pure copper and gold. Specifically, the low friction regime is linked to the formation of ultra-nanocrystalline surface films (10 to 20 nm), driving toward shear accommodation by grain boundary sliding. Above a critical combination of stress and temperature -- demonstrated to be a material property -- shear accommodation transitions to dislocation dominated plasticity and high friction. We utilize a combination of experimental and computational methods to develop and validate the proposed structure-property relationship. This quantitative framework provides a shift from phenomenological to mechanistic and predictive fundamental understanding of friction for crystalline materials, including engineering alloys.

  14. Command decoder unit. [performance tests of data processing terminals and data converters for space shuttle orbiters

    Science.gov (United States)

    1976-01-01

    The design and testing of laboratory hardware (a command decoder unit) used in evaluating space shuttle instrumentation, data processing, and ground check-out operations is described. The hardware was a modification of another similar instrumentation system. A data bus coupler was designed and tested to interface the equipment to a central bus controller (computer). A serial digital data transfer mechanism was also designed. Redundant power supplies and overhead modules were provided to minimize the probability of a single component failure causing a catastrophic failure. The command decoder unit is packaged in a modular configuration to allow maximum user flexibility in configuring a system. Test procedures and special test equipment for use in testing the hardware are described. Results indicate that the unit will allow NASA to evaluate future software systems for use in space shuttles. The units were delivered to NASA and appear to be adequately performing their intended function. Engineering sketches and photographs of the command decoder unit are included.

  15. Activation process in excitable systems with multiple noise sources: Large number of units

    CERN Document Server

    Franović, Igor; Todorović, Kristina; Kostić, Srđan; Burić, Nikola

    2015-01-01

    We study the activation process in large assemblies of type II excitable units whose dynamics is influenced by two independent noise terms. The mean-field approach is applied to explicitly demonstrate that the assembly of excitable units can itself exhibit macroscopic excitable behavior. In order to facilitate the comparison between the excitable dynamics of a single unit and an assembly, we introduce three distinct formulations of the assembly activation event. Each formulation treats different aspects of the relevant phenomena, including the threshold-like behavior and the role of coherence of individual spikes. Statistical properties of the assembly activation process, such as the mean time-to-first pulse and the associated coefficient of variation, are found to be qualitatively analogous for all three formulations, as well as to resemble the results for a single unit. These analogies are shown to derive from the fact that global variables undergo a stochastic bifurcation from the stochastically stable fix...

  16. Unit Process Wetlands for Removal of Trace Organic Contaminants and Pathogens from Municipal Wastewater Effluents

    Science.gov (United States)

    Jasper, Justin T.; Nguyen, Mi T.; Jones, Zackary L.; Ismail, Niveen S.; Sedlak, David L.; Sharp, Jonathan O.; Luthy, Richard G.; Horne, Alex J.; Nelson, Kara L.

    2013-01-01

    Abstract Treatment wetlands have become an attractive option for the removal of nutrients from municipal wastewater effluents due to their low energy requirements and operational costs, as well as the ancillary benefits they provide, including creating aesthetically appealing spaces and wildlife habitats. Treatment wetlands also hold promise as a means of removing other wastewater-derived contaminants, such as trace organic contaminants and pathogens. However, concerns about variations in treatment efficacy of these pollutants, coupled with an incomplete mechanistic understanding of their removal in wetlands, hinder the widespread adoption of constructed wetlands for these two classes of contaminants. A better understanding is needed so that wetlands as a unit process can be designed for their removal, with individual wetland cells optimized for the removal of specific contaminants, and connected in series or integrated with other engineered or natural treatment processes. In this article, removal mechanisms of trace organic contaminants and pathogens are reviewed, including sorption and sedimentation, biotransformation and predation, photolysis and photoinactivation, and remaining knowledge gaps are identified. In addition, suggestions are provided for how these treatment mechanisms can be enhanced in commonly employed unit process wetland cells or how they might be harnessed in novel unit process cells. It is hoped that application of the unit process concept to a wider range of contaminants will lead to more widespread application of wetland treatment trains as components of urban water infrastructure in the United States and around the globe. PMID:23983451

  17. A Shipping Container-Based Sterile Processing Unit for Low Resources Settings.

    Science.gov (United States)

    Boubour, Jean; Jenson, Katherine; Richter, Hannah; Yarbrough, Josiah; Oden, Z Maria; Schuler, Douglas A

    2016-01-01

    Deficiencies in the sterile processing of medical instruments contribute to poor outcomes for patients, such as surgical site infections, longer hospital stays, and deaths. In low resources settings, such as some rural and semi-rural areas and secondary and tertiary cities of developing countries, deficiencies in sterile processing are accentuated due to the lack of access to sterilization equipment, improperly maintained and malfunctioning equipment, lack of power to operate equipment, poor protocols, and inadequate quality control over inventory. Inspired by our sterile processing fieldwork at a district hospital in Sierra Leone in 2013, we built an autonomous, shipping-container-based sterile processing unit to address these deficiencies. The sterile processing unit, dubbed "the sterile box," is a full suite capable of handling instruments from the moment they leave the operating room to the point they are sterile and ready to be reused for the next surgery. The sterile processing unit is self-sufficient in power and water and features an intake for contaminated instruments, decontamination, sterilization via non-electric steam sterilizers, and secure inventory storage. To validate efficacy, we ran tests of decontamination and sterilization performance. Results of 61 trials validate convincingly that our sterile processing unit achieves satisfactory outcomes for decontamination and sterilization and as such holds promise to support healthcare facilities in low resources settings.

  18. A Shipping Container-Based Sterile Processing Unit for Low Resources Settings.

    Directory of Open Access Journals (Sweden)

    Jean Boubour

    Full Text Available Deficiencies in the sterile processing of medical instruments contribute to poor outcomes for patients, such as surgical site infections, longer hospital stays, and deaths. In low resources settings, such as some rural and semi-rural areas and secondary and tertiary cities of developing countries, deficiencies in sterile processing are accentuated due to the lack of access to sterilization equipment, improperly maintained and malfunctioning equipment, lack of power to operate equipment, poor protocols, and inadequate quality control over inventory. Inspired by our sterile processing fieldwork at a district hospital in Sierra Leone in 2013, we built an autonomous, shipping-container-based sterile processing unit to address these deficiencies. The sterile processing unit, dubbed "the sterile box," is a full suite capable of handling instruments from the moment they leave the operating room to the point they are sterile and ready to be reused for the next surgery. The sterile processing unit is self-sufficient in power and water and features an intake for contaminated instruments, decontamination, sterilization via non-electric steam sterilizers, and secure inventory storage. To validate efficacy, we ran tests of decontamination and sterilization performance. Results of 61 trials validate convincingly that our sterile processing unit achieves satisfactory outcomes for decontamination and sterilization and as such holds promise to support healthcare facilities in low resources settings.

  19. High Power Silicon Carbide (SiC) Power Processing Unit Development

    Science.gov (United States)

    Scheidegger, Robert J.; Santiago, Walter; Bozak, Karin E.; Pinero, Luis R.; Birchenough, Arthur G.

    2015-01-01

    NASA GRC successfully designed, built and tested a technology-push power processing unit for electric propulsion applications that utilizes high voltage silicon carbide (SiC) technology. The development specifically addresses the need for high power electronics to enable electric propulsion systems in the 100s of kilowatts. This unit demonstrated how high voltage combined with superior semiconductor components resulted in exceptional converter performance.

  20. Experience in design and startup of distillation towers in primary crude oil processing unit

    Energy Technology Data Exchange (ETDEWEB)

    Lebedev, Y.N.; D' yakov, V.G.; Mamontov, G.V.; Sheinman, V.A.; Ukhin, V.V.

    1985-11-01

    This paper describes a refinery in the city of Mathura, India, with a capacity of 7 million metric tons of crude per year, designed and constructed to include the following units: AVT for primary crude oil processing; catalytic cracking; visbreaking; asphalt; and other units. A diagram of the atmospheric tower with stripping sections is shown, and the stabilizer tower is illustrated. The startup and operation of the AVT and visbreaking units are described, and they demonstrate the high reliability and efficiency of the equipment.

  1. Program note: applying the UN process indicators for emergency obstetric care to the United States.

    Science.gov (United States)

    Lobis, S; Fry, D; Paxton, A

    2005-02-01

    The United Nations Process Indicators for emergency obstetric care (EmOC) have been used extensively in countries with high maternal mortality ratios (MMR) to assess the availability, utilization and quality of EmOC services. To compare the situation in high MMR countries to that of a low MMR country, data from the United States were used to determine EmOC service availability, utilization and quality. As was expected, the United States was found to have an adequate amount of good-quality EmOC services that are used by the majority of women with life-threatening obstetric complications.

  2. A macroscale mixture theory analysis of deposition and sublimation rates during heat and mass transfer in dry snow

    Directory of Open Access Journals (Sweden)

    A. C. Hansen

    2015-09-01

    Full Text Available The microstructure of a dry alpine snowpack is a dynamic environment where microstructural evolution is driven by seasonal density profiles and weather conditions. Notably, temperature gradients on the order of 10–20 K m−1, or larger, are known to produce a faceted snow microstructure exhibiting little strength. However, while strong temperature gradients are widely accepted as the primary driver for kinetic growth, they do not fully account for the range of experimental observations. An additional factor influencing snow metamorphism is believed to be the rate of mass transfer at the macroscale. We develop a mixture theory capable of predicting macroscale deposition and/or sublimation in a snow cover under temperature gradient conditions. Temperature gradients and mass exchange are tracked over periods ranging from 1 to 10 days. Interesting heat and mass transfer behavior is observed near the ground, near the surface, as well as immediately above and below dense ice crusts. Information about deposition (condensation and sublimation rates may help explain snow metamorphism phenomena that cannot be accounted for by temperature gradients alone. The macroscale heat and mass transfer analysis requires accurate representations of the effective thermal conductivity and the effective mass diffusion coefficient for snow. We develop analytical models for these parameters based on first principles at the microscale. The expressions derived contain no empirical adjustments, and further, provide self consistent values for effective thermal conductivity and the effective diffusion coefficient for the limiting cases of air and solid ice. The predicted values for these macroscale material parameters are also in excellent agreement with numerical results based on microscale finite element analyses of representative volume elements generated from X-ray tomography.

  3. A macroscale mixture theory analysis of deposition and sublimation rates during heat and mass transfer in snow

    Directory of Open Access Journals (Sweden)

    A. C. Hansen

    2015-03-01

    Full Text Available The microstructure of a dry alpine snowpack is a dynamic environment where microstructural evolution is driven by seasonal density profiles and weather conditions. Notably, temperature gradients on the order of 10–20 K m−1, or larger, are known to produce a faceted snow microstructure exhibiting little strength. However, while strong temperature gradients are widely accepted as the primary driver for kinetic growth, they do not fully account for the range of experimental observations. An additional factor influencing snow metamorphism is believed to be the rate of mass transfer at the macroscale. We develop a mixture theory capable of predicting macroscale deposition and/or sublimation in a snow cover under temperature gradient conditions. Temperature gradients and mass exchange are tracked over periods ranging from 1 to 10 days. Interesting heat and mass transfer behavior is observed near the ground, near the surface, as well as immediately above and below dense ice crusts. Information about deposition (condensation and sublimation rates may help explain snow metamorphism phenomena that cannot be accounted for by temperature gradients alone. The macroscale heat and mass transfer analysis requires accurate representations of the thermal conductivity and the effective mass diffusion coefficient for snow. We develop analytical models for these parameters based on first principles at the microscale. The expressions derived contain no empirical adjustments, and further, provide self consistent values for thermal conductivity and the effective diffusion coefficient for the limiting cases of air and solid ice. The predicted values for these macroscale material parameters are also in excellent agreement with numerical results based on microscale finite element analyses of representative volume elements generated from X-ray tomography.

  4. NUMATH: a nuclear-material-holdup estimator for unit operations and chemical processes

    Energy Technology Data Exchange (ETDEWEB)

    Krichinsky, A.M.

    1983-02-01

    A computer program, NUMATH (Nuclear Material Holdup Estimator), has been developed to estimate compositions of materials in vessels involved in unit operations and chemical processes. This program has been implemented in a remotely operated nuclear fuel processing plant. NUMATH provides estimates of the steady-state composition of materials residing in process vessels until representative samples can be obtained and chemical analyses can be performed. Since these compositions are used for inventory estimations, the results are determined for the cataloged in container-oriented files. The estimated compositions represent materials collected in applicable vessels - including consideration for materials previously acknowledged in these vessels. The program utilizes process measurements and simple performance models to estimate material holdup and distribution within unit operations. In simulated run-testing, NUMATH typically produced estimates within 5% of the measured inventories for uranium and within 8% of the measured inventories for thorium during steady-state process operation.

  5. Grace: a Cross-platform Micromagnetic Simulator On Graphics Processing Units

    CERN Document Server

    Zhu, Ru

    2014-01-01

    A micromagnetic simulator running on graphics processing unit (GPU) is presented. It achieves significant performance boost as compared to previous central processing unit (CPU) simulators, up to two orders of magnitude for large input problems. Different from GPU implementations of other research groups, this simulator is developed with C++ Accelerated Massive Parallelism (C++ AMP) and is hardware platform compatible. It runs on GPU from venders include NVidia, AMD and Intel, which paved the way for fast micromagnetic simulation on both high-end workstations with dedicated graphics cards and low-end personal computers with integrated graphics card. A copy of the simulator software is publicly available.

  6. Fast extended focused imaging in digital holography using a graphics processing unit.

    Science.gov (United States)

    Wang, Le; Zhao, Jianlin; Di, Jianglei; Jiang, Hongzhen

    2011-05-01

    We present a simple and effective method for reconstructing extended focused images in digital holography using a graphics processing unit (GPU). The Fresnel transform method is simplified by an algorithm named fast Fourier transform pruning with frequency shift. Then the pixel size consistency problem is solved by coordinate transformation and combining the subpixel resampling and the fast Fourier transform pruning with frequency shift. With the assistance of the GPU, we implemented an improved parallel version of this method, which obtained about a 300-500-fold speedup compared with central processing unit codes.

  7. Fast high-resolution computer-generated hologram computation using multiple graphics processing unit cluster system.

    Science.gov (United States)

    Takada, Naoki; Shimobaba, Tomoyoshi; Nakayama, Hirotaka; Shiraki, Atsushi; Okada, Naohisa; Oikawa, Minoru; Masuda, Nobuyuki; Ito, Tomoyoshi

    2012-10-20

    To overcome the computational complexity of a computer-generated hologram (CGH), we implement an optimized CGH computation in our multi-graphics processing unit cluster system. Our system can calculate a CGH of 6,400×3,072 pixels from a three-dimensional (3D) object composed of 2,048 points in 55 ms. Furthermore, in the case of a 3D object composed of 4096 points, our system is 553 times faster than a conventional central processing unit (using eight threads).

  8. Massively Parallel Signal Processing using the Graphics Processing Unit for Real-Time Brain-Computer Interface Feature Extraction.

    Science.gov (United States)

    Wilson, J Adam; Williams, Justin C

    2009-01-01

    The clock speeds of modern computer processors have nearly plateaued in the past 5 years. Consequently, neural prosthetic systems that rely on processing large quantities of data in a short period of time face a bottleneck, in that it may not be possible to process all of the data recorded from an electrode array with high channel counts and bandwidth, such as electrocorticographic grids or other implantable systems. Therefore, in this study a method of using the processing capabilities of a graphics card [graphics processing unit (GPU)] was developed for real-time neural signal processing of a brain-computer interface (BCI). The NVIDIA CUDA system was used to offload processing to the GPU, which is capable of running many operations in parallel, potentially greatly increasing the speed of existing algorithms. The BCI system records many channels of data, which are processed and translated into a control signal, such as the movement of a computer cursor. This signal processing chain involves computing a matrix-matrix multiplication (i.e., a spatial filter), followed by calculating the power spectral density on every channel using an auto-regressive method, and finally classifying appropriate features for control. In this study, the first two computationally intensive steps were implemented on the GPU, and the speed was compared to both the current implementation and a central processing unit-based implementation that uses multi-threading. Significant performance gains were obtained with GPU processing: the current implementation processed 1000 channels of 250 ms in 933 ms, while the new GPU method took only 27 ms, an improvement of nearly 35 times.

  9. Process and structure: resource management and the development of sub-unit organisational structure.

    Science.gov (United States)

    Packwood, T; Keen, J; Buxton, M

    1992-03-01

    Resource Management (RM) requires hospital units to manage their work in new ways, and the new management processes affect, and are affected by, organisation structure. This paper is concerned with these effects, reporting on the basis of a three-year evaluation of the national RM experiment that was commissioned by the DH. After briefly indicating some of the major characteristics of the RM process, the two main types of unit structures existing in the pilot sites at the beginning of the experiment, unit disciplinary structure and clinical directorates, are analysed. At the end of the experiment, while clinical directorates had become more popular, another variant, clinical grouping, had replaced the unit disciplinary structure. Both types of structure represent a movement towards sub-unit organisation, bringing the work and interests of the service providers and unit managers closer together. Their properties are likewise analysed and their implications, particularly in terms of training and organisational development (OD), are then considered. The paper concludes by considering the causes for these structural changes, which, in the immediate time-scale, appear to owe as much to the NHS Review as to RM.

  10. [Applying graphics processing unit in real-time signal processing and visualization of ophthalmic Fourier-domain OCT system].

    Science.gov (United States)

    Liu, Qiaoyan; Li, Yuejie; Xu, Qiujing; Zhao, Jincheng; Wang, Liwei; Gao, Yonghe

    2013-01-01

    This investigation introduces GPU (Graphics Processing Unit)- based CUDA (Compute Unified Device Architecture) technology into signal processing of ophthalmic FD-OCT (Fourier-Domain Optical Coherence Tomography) imaging system, can realize parallel data processing, using CUDA to optimize relevant operations and algorithms, in order to solve the technical bottlenecks that currently affect ophthalmic real-time imaging in OCT system. Laboratory results showed that with GPU as a general parallel computing processor, the speed of imaging data processing using GPU+CPU mode is more than dozens times faster than traditional CPU platform based serial computing and imaging mode when executing the same data processing, which reaches the clinical requirements for two dimensional real-time imaging.

  11. Fast blood flow visualization of high-resolution laser speckle imaging data using graphics processing unit.

    Science.gov (United States)

    Liu, Shusen; Li, Pengcheng; Luo, Qingming

    2008-09-15

    Laser speckle contrast analysis (LASCA) is a non-invasive, full-field optical technique that produces two-dimensional map of blood flow in biological tissue by analyzing speckle images captured by CCD camera. Due to the heavy computation required for speckle contrast analysis, video frame rate visualization of blood flow which is essentially important for medical usage is hardly achieved for the high-resolution image data by using the CPU (Central Processing Unit) of an ordinary PC (Personal Computer). In this paper, we introduced GPU (Graphics Processing Unit) into our data processing framework of laser speckle contrast imaging to achieve fast and high-resolution blood flow visualization on PCs by exploiting the high floating-point processing power of commodity graphics hardware. By using GPU, a 12-60 fold performance enhancement is obtained in comparison to the optimized CPU implementations.

  12. Liquid phase methanol LaPorte process development unit: Modification, operation, and support studies

    Energy Technology Data Exchange (ETDEWEB)

    1991-02-02

    The primary focus of this Process Development Unit operating program was to prepare for a confident move to the next scale of operation with a simplified and optimized process. The main purpose of these runs was the evaluation of the alternate commercial catalyst (F21/0E75-43) that had been identified in the laboratory under a different subtask of the program. If the catalyst proved superior to the previous catalyst, then the evaluation run would be continued into a 120-day life run. Also, minor changes were made to the Process Development Unit system to improve operations and reliability. The damaged reactor demister from a previous run was replaced, and a new demister was installed in the intermediate V/L separator. The internal heat exchanger was equipped with an expansion loop to relieve thermal stresses so operation at higher catalyst loadings and gas velocities would be possible. These aggressive conditions are important for improving process economics. (VC)

  13. Proposal of a Novel Approach to Developing Material Models for Micro-scale Composites Based on Testing and Modeling of Macro-scale Composites

    Energy Technology Data Exchange (ETDEWEB)

    Siranosian, Antranik Antonio [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Schembri, Philip Edward [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Luscher, Darby Jon [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-04-20

    The Los Alamos National Laboratory's Weapon Systems Engineering division's Advanced Engineering Analysis group employs material constitutive models of composites for use in simulations of components and assemblies of interest. Experimental characterization, modeling and prediction of the macro-scale (i.e. continuum) behaviors of these composite materials is generally difficult because they exhibit nonlinear behaviors on the meso- (e.g. micro-) and macro-scales. Furthermore, it can be difficult to measure and model the mechanical responses of the individual constituents and constituent interactions in the composites of interest. Current efforts to model such composite materials rely on semi-empirical models in which meso-scale properties are inferred from continuum level testing and modeling. The proposed approach involves removing the difficulties of interrogating and characterizing micro-scale behaviors by scaling-up the problem to work with macro-scale composites, with the intention of developing testing and modeling capabilities that will be applicable to the mesoscale. This approach assumes that the physical mechanisms governing the responses of the composites on the meso-scale are reproducible on the macro-scale. Working on the macro-scale simplifies the quantification of composite constituents and constituent interactions so that efforts can be focused on developing material models and the testing techniques needed for calibration and validation. Other benefits to working with macro-scale composites include the ability to engineer and manufacture—potentially using additive manufacturing techniques—composites that will support the application of advanced measurement techniques such as digital volume correlation and three-dimensional computed tomography imaging, which would aid in observing and quantifying complex behaviors that are exhibited in the macro-scale composites of interest. Ultimately, the goal of this new approach is to develop a meso

  14. [Work process of the nurse who works in child care in family health units].

    Science.gov (United States)

    de Assis, Wesley Dantas; Collet, Neusa; Reichert, Altamira Pereira da Silva; de Sá, Lenilde Duarte

    2011-01-01

    This is a qualitative research, which purpose was to analyse the working process of nurse in child care actions in family health units. Nurses are the subjects and empirical data was achieved by the means of participant observation, and interviews. Data analysis followed thematic analysis fundaments. Results reveal that working process organization of nurses still remains centered in proceedings with an offert of assistance based in client illness, showing obstacles to puericulture practice in health basic attention.

  15. 75 FR 74005 - Fisheries of the Northeastern United States; Monkfish Fishery; Scoping Process

    Science.gov (United States)

    2010-11-30

    ... National Oceanic and Atmospheric Administration RIN 0648-BA50 Fisheries of the Northeastern United States; Monkfish Fishery; Scoping Process AGENCY: National Marine Fisheries Service (NMFS), National Oceanic and... statement (EIS) and scoping meetings; request for comments. SUMMARY: The New England Fishery...

  16. ECO LOGIC INTERNATIONAL GAS-PHASE CHEMICAL REDUCTION PROCESS - THE THERMAL DESORPTION UNIT - APPLICATIONS ANALYSIS REPORT

    Science.gov (United States)

    ELI ECO Logic International, Inc.'s Thermal Desorption Unit (TDU) is specifically designed for use with Eco Logic's Gas Phase Chemical Reduction Process. The technology uses an externally heated bath of molten tin in a hydrogen atmosphere to desorb hazardous organic compounds fro...

  17. Process methods and levels of automation of wood pallet repair in the United States

    Science.gov (United States)

    Jonghun Park; Laszlo Horvath; Robert J. Bush

    2016-01-01

    This study documented the current status of wood pallet repair in the United States by identifying the types of processing and equipment usage in repair operations from an automation prespective. The wood pallet repair firms included in the sudy received an average of approximately 1.28 million cores (i.e., used pallets) for recovery in 2012. A majority of the cores...

  18. Catalyzed steam gasification of biomass. Phase 3: Biomass Process Development Unit (PDU) construction and initial operation

    Science.gov (United States)

    Healey, J. J.; Hooverman, R. H.

    1981-12-01

    The design and construction of the process development unit (PDU) are described in detail, examining each system and component in order. Siting, the chip handling system, the reactor feed system, the reactor, the screw conveyor, the ash dump system, the PDU support equipment, control and information management, and shakedown runs are described.

  19. Silicon Carbide (SiC) Power Processing Unit (PPU) for Hall Effect Thrusters Project

    Data.gov (United States)

    National Aeronautics and Space Administration — In this SBIR project, APEI, Inc. is proposing to develop a high efficiency, rad-hard 3.8 kW silicon carbide (SiC) power supply for the Power Processing Unit (PPU) of...

  20. Parallelized CCHE2D flow model with CUDA Fortran on Graphics Process Units

    Science.gov (United States)

    This paper presents the CCHE2D implicit flow model parallelized using CUDA Fortran programming technique on Graphics Processing Units (GPUs). A parallelized implicit Alternating Direction Implicit (ADI) solver using Parallel Cyclic Reduction (PCR) algorithm on GPU is developed and tested. This solve...

  1. Liquid phase methanol LaPorte process development unit: Modification, operation, and support studies

    Energy Technology Data Exchange (ETDEWEB)

    1991-02-02

    This report consists of Detailed Data Acquisition Sheets for Runs E-6 and E-7 for Task 2.2 of the Modification, Operation, and Support Studies of the Liquid Phase Methanol Laporte Process Development Unit. (Task 2.2: Alternate Catalyst Run E-6 and Catalyst Activity Maintenance Run E-7).

  2. Sodium content of popular commercially processed and restaurant foods in the United States

    Science.gov (United States)

    Nutrient Data Laboratory (NDL) of the U.S. Department of Agriculture (USDA) in close collaboration with U.S. Center for Disease Control and Prevention is monitoring the sodium content of commercially processed and restaurant foods in the United States. The main purpose of this manuscript is to prov...

  3. On the use of graphics processing units (GPUs) for molecular dynamics simulation of spherical particles

    NARCIS (Netherlands)

    Hidalgo, R.C.; Kanzaki, T.; Alonso-Marroquin, F.; Luding, S.; Yu, A.; Dong, K.; Yang, R.; Luding, S.

    2013-01-01

    General-purpose computation on Graphics Processing Units (GPU) on personal computers has recently become an attractive alternative to parallel computing on clusters and supercomputers. We present the GPU-implementation of an accurate molecular dynamics algorithm for a system of spheres. The new hybr

  4. Graphics Processing Unit-Based Bioheat Simulation to Facilitate Rapid Decision Making Associated with Cryosurgery Training.

    Science.gov (United States)

    Keelan, Robert; Zhang, Hong; Shimada, Kenji; Rabin, Yoed

    2016-04-01

    This study focuses on the implementation of an efficient numerical technique for cryosurgery simulations on a graphics processing unit as an alternative means to accelerate runtime. This study is part of an ongoing effort to develop computerized training tools for cryosurgery, with prostate cryosurgery as a developmental model. The ability to perform rapid simulations of various test cases is critical to facilitate sound decision making associated with medical training. Consistent with clinical practice, the training tool aims at correlating the frozen region contour and the corresponding temperature field with the target region shape. The current study focuses on the feasibility of graphics processing unit-based computation using C++ accelerated massive parallelism, as one possible implementation. Benchmark results on a variety of computation platforms display between 3-fold acceleration (laptop) and 13-fold acceleration (gaming computer) of cryosurgery simulation, in comparison with the more common implementation on a multicore central processing unit. While the general concept of graphics processing unit-based simulations is not new, its application to phase-change problems, combined with the unique requirements for cryosurgery optimization, represents the core contribution of the current study.

  5. Fast, multi-channel real-time processing of signals with microsecond latency using graphics processing units.

    Science.gov (United States)

    Rath, N; Kato, S; Levesque, J P; Mauel, M E; Navratil, G A; Peng, Q

    2014-04-01

    Fast, digital signal processing (DSP) has many applications. Typical hardware options for performing DSP are field-programmable gate arrays (FPGAs), application-specific integrated DSP chips, or general purpose personal computer systems. This paper presents a novel DSP platform that has been developed for feedback control on the HBT-EP tokamak device. The system runs all signal processing exclusively on a Graphics Processing Unit (GPU) to achieve real-time performance with latencies below 8 μs. Signals are transferred into and out of the GPU using PCI Express peer-to-peer direct-memory-access transfers without involvement of the central processing unit or host memory. Tests were performed on the feedback control system of the HBT-EP tokamak using forty 16-bit floating point inputs and outputs each and a sampling rate of up to 250 kHz. Signals were digitized by a D-TACQ ACQ196 module, processing done on an NVIDIA GTX 580 GPU programmed in CUDA, and analog output was generated by D-TACQ AO32CPCI modules.

  6. Fast, multi-channel real-time processing of signals with microsecond latency using graphics processing units

    Science.gov (United States)

    Rath, N.; Kato, S.; Levesque, J. P.; Mauel, M. E.; Navratil, G. A.; Peng, Q.

    2014-04-01

    Fast, digital signal processing (DSP) has many applications. Typical hardware options for performing DSP are field-programmable gate arrays (FPGAs), application-specific integrated DSP chips, or general purpose personal computer systems. This paper presents a novel DSP platform that has been developed for feedback control on the HBT-EP tokamak device. The system runs all signal processing exclusively on a Graphics Processing Unit (GPU) to achieve real-time performance with latencies below 8 μs. Signals are transferred into and out of the GPU using PCI Express peer-to-peer direct-memory-access transfers without involvement of the central processing unit or host memory. Tests were performed on the feedback control system of the HBT-EP tokamak using forty 16-bit floating point inputs and outputs each and a sampling rate of up to 250 kHz. Signals were digitized by a D-TACQ ACQ196 module, processing done on an NVIDIA GTX 580 GPU programmed in CUDA, and analog output was generated by D-TACQ AO32CPCI modules.

  7. 从图形处理器到基于GPU的通用计算%From Graphic Processing Unit to General Purpose Graphic Processing Unit

    Institute of Scientific and Technical Information of China (English)

    刘金硕; 刘天晓; 吴慧; 曾秋梅; 任梦菲; 顾宜淳

    2013-01-01

    对GPU(graphic process unit)、基于GPU的通用计算(general purpose GPU,GPGPU)、基于GPU的编程模型与环境进行了界定;将GPU的发展分为4个阶段,阐述了GPU的架构由非统一的渲染架构到统一的渲染架构,再到新一代的费米架构的变化;通过对基于GPU的通用计算的架构与多核CPU架构、分布式集群架构进行了软硬件的对比.分析表明:当进行中粒度的线程级数据密集型并行运算时,采用多核多线程并行;当进行粗粒度的网络密集型并行运算时,采用集群并行;当进行细粒度的计算密集型并行运算时,采用GPU通用计算并行.最后本文展示了未来的GPGPU的研究热点和发展方向--GPGPU自动并行化、CUDA对多种语言的支持、CUDA的性能优化,并介绍了GPGPU的一些典型应用.%This paper defines the outline of GPU(graphic processing unit) , the general purpose computation, the programming model and the environment for GPU. Besides, it introduces the evolution process from GPU to GPGPU (general purpose graphic processing unit) , and the change from non-uniform render architecture to the unified render architecture and the next Fermi architecture in details. Then it compares GPGPU architecture with multi-core GPU architecture and distributed cluster architecture from the perspective of software and hardware. When doing the middle grain level thread data intensive parallel computing, the multi-core and multi-thread should be utilized. When doing the coarse grain network computing, the cluster computing should be utilized. When doing the fine grained compute intensive parallel computing, the general purpose computation should be adopted. Meanwhile, some classical applications of GPGPU have been mentioned. At last, this paper demonstrates the further developments and research hotspots of GPGPU, which are automatic parallelization of GPGPU, multi-language support and performance optimization of CUDA, and introduces the classic

  8. Bandwidth Enhancement between Graphics Processing Units on the Peripheral Component Interconnect Bus

    Directory of Open Access Journals (Sweden)

    ANTON Alin

    2015-10-01

    Full Text Available General purpose computing on graphics processing units is a new trend in high performance computing. Present day applications require office and personal supercomputers which are mostly based on many core hardware accelerators communicating with the host system through the Peripheral Component Interconnect (PCI bus. Parallel data compression is a difficult topic but compression has been used successfully to improve the communication between parallel message passing interface (MPI processes on high performance computing clusters. In this paper we show that special pur pose compression algorithms designed for scientific floating point data can be used to enhance the bandwidth between 2 graphics processing unit (GPU devices on the PCI Express (PCIe 3.0 x16 bus in a homebuilt personal supercomputer (PSC.

  9. Parallel computing for simultaneous iterative tomographic imaging by graphics processing units

    Science.gov (United States)

    Bello-Maldonado, Pedro D.; López, Ricardo; Rogers, Colleen; Jin, Yuanwei; Lu, Enyue

    2016-05-01

    In this paper, we address the problem of accelerating inversion algorithms for nonlinear acoustic tomographic imaging by parallel computing on graphics processing units (GPUs). Nonlinear inversion algorithms for tomographic imaging often rely on iterative algorithms for solving an inverse problem, thus computationally intensive. We study the simultaneous iterative reconstruction technique (SIRT) for the multiple-input-multiple-output (MIMO) tomography algorithm which enables parallel computations of the grid points as well as the parallel execution of multiple source excitation. Using graphics processing units (GPUs) and the Compute Unified Device Architecture (CUDA) programming model an overall improvement of 26.33x was achieved when combining both approaches compared with sequential algorithms. Furthermore we propose an adaptive iterative relaxation factor and the use of non-uniform weights to improve the overall convergence of the algorithm. Using these techniques, fast computations can be performed in parallel without the loss of image quality during the reconstruction process.

  10. Rapid learning-based video stereolization using graphic processing unit acceleration

    Science.gov (United States)

    Sun, Tian; Jung, Cheolkon; Wang, Lei; Kim, Joongkyu

    2016-09-01

    Video stereolization has received much attention in recent years due to the lack of stereoscopic three-dimensional (3-D) contents. Although video stereolization can enrich stereoscopic 3-D contents, it is hard to achieve automatic two-dimensional-to-3-D conversion with less computational cost. We proposed rapid learning-based video stereolization using a graphic processing unit (GPU) acceleration. We first generated an initial depth map based on learning from examples. Then, we refined the depth map using saliency and cross-bilateral filtering to make object boundaries clear. Finally, we performed depth-image-based-rendering to generate stereoscopic 3-D views. To accelerate the computation of video stereolization, we provided a parallelizable hybrid GPU-central processing unit (CPU) solution to be suitable for running on GPU. Experimental results demonstrate that the proposed method is nearly 180 times faster than CPU-based processing and achieves a good performance comparable to the-state-of-the-art ones.

  11. Molecular dynamics for long-range interacting systems on Graphic Processing Units

    CERN Document Server

    Filho, Tarcísio M Rocha

    2012-01-01

    We present implementations of a fourth-order symplectic integrator on graphic processing units for three $N$-body models with long-range interactions of general interest: the Hamiltonian Mean Field, Ring and two-dimensional self-gravitating models. We discuss the algorithms, speedups and errors using one and two GPU units. Speedups can be as high as 140 compared to a serial code, and the overall relative error in the total energy is of the same order of magnitude as for the CPU code. The number of particles used in the tests range from 10,000 to 50,000,000 depending on the model.

  12. Software Graphics Processing Unit (sGPU) for Deep Space Applications

    Science.gov (United States)

    McCabe, Mary; Salazar, George; Steele, Glen

    2015-01-01

    A graphics processing capability will be required for deep space missions and must include a range of applications, from safety-critical vehicle health status to telemedicine for crew health. However, preliminary radiation testing of commercial graphics processing cards suggest they cannot operate in the deep space radiation environment. Investigation into an Software Graphics Processing Unit (sGPU)comprised of commercial-equivalent radiation hardened/tolerant single board computers, field programmable gate arrays, and safety-critical display software shows promising results. Preliminary performance of approximately 30 frames per second (FPS) has been achieved. Use of multi-core processors may provide a significant increase in performance.

  13. Intra- versus inter-site macroscale variation in biogeochemical properties along a paddy soil chronosequence

    Directory of Open Access Journals (Sweden)

    C. Mueller-Niggemann

    2012-03-01

    Full Text Available In order to assess the intrinsic heterogeneity of paddy soils, a set of biogeochemical soil parameters was investigated in five field replicates of seven paddy fields (50, 100, 300, 500, 700, 1000, and 2000 yr of wetland rice cultivation, one flooded paddy nursery, one tidal wetland (TW, and one freshwater site (FW from a coastal area at Hangzhou Bay, Zhejiang Province, China. All soils evolved from a marine tidal flat substrate due to land reclamation. The biogeochemical parameters based on their properties were differentiated into (i a group behaving conservatively (TC, TOC, TN, TS, magnetic susceptibility, soil lightness and colour parameters, δ13C, δ15N, lipids and n-alkanes and (ii one encompassing more labile properties or fast cycling components (Nmic, Cmic, nitrate, ammonium, DON and DOC. The macroscale heterogeneity in paddy soils was assessed by evaluating intra- versus inter-site spatial variability of biogeochemical properties using statistical data analysis (descriptive, explorative and non-parametric. Results show that the intrinsic heterogeneity of paddy soil organic and minerogenic components per field is smaller than between study sites. The coefficient of variation (CV values of conservative parameters varied in a low range (10% to 20%, decreasing from younger towards older paddy soils. This indicates a declining variability of soil biogeochemical properties in longer used cropping sites according to progress in soil evolution. A generally higher variation of CV values (>20–40% observed for labile parameters implies a need for substantially higher sampling frequency when investigating these as compared to more conservative parameters. Since the representativeness of the sampling strategy could be sufficiently demonstrated, an investigation of long-term carbon accumulation/sequestration trends in topsoils of the 2000 yr paddy chronosequence under wetland rice cultivation

  14. Case Studies of Internationalization in Adult and Higher Education: Inside the Processes of Four Universities in the United States and the United Kingdom

    Science.gov (United States)

    Coryell, Joellen Elizabeth; Durodoye, Beth A.; Wright, Robin Redmon; Pate, P. Elizabeth; Nguyen, Shelbee

    2012-01-01

    This report outlines a method for learning about the internationalization processes at institutions of adult and higher education and then provides the analysis of data gathered from the researchers' own institution and from site visits to three additional universities in the United States and the United Kingdom. It was found that campus…

  15. Monitoring and assessment of soil erosion at micro-scale and macro-scale in forests affected by fire damage in northern Iran.

    Science.gov (United States)

    Akbarzadeh, Ali; Ghorbani-Dashtaki, Shoja; Naderi-Khorasgani, Mehdi; Kerry, Ruth; Taghizadeh-Mehrjardi, Ruhollah

    2016-12-01

    Understanding the occurrence of erosion processes at large scales is very difficult without studying them at small scales. In this study, soil erosion parameters were investigated at micro-scale and macro-scale in forests in northern Iran. Surface erosion and some vegetation attributes were measured at the watershed scale in 30 parcels of land which were separated into 15 fire-affected (burned) forests and 15 original (unburned) forests adjacent to the burned sites. The soil erodibility factor and splash erosion were also determined at the micro-plot scale within each burned and unburned site. Furthermore, soil sampling and infiltration studies were carried out at 80 other sites, as well as the 30 burned and unburned sites, (a total of 110 points) to create a map of the soil erodibility factor at the regional scale. Maps of topography, rainfall, and cover-management were also determined for the study area. The maps of erosion risk and erosion risk potential were finally prepared for the study area using the Revised Universal Soil Loss Equation (RUSLE) procedure. Results indicated that destruction of the protective cover of forested areas by fire had significant effects on splash erosion and the soil erodibility factor at the micro-plot scale and also on surface erosion, erosion risk, and erosion risk potential at the watershed scale. Moreover, the results showed that correlation coefficients between different variables at the micro-plot and watershed scales were positive and significant. Finally, assessment and monitoring of the erosion maps at the regional scale showed that the central and western parts of the study area were more susceptible to erosion compared with the western regions due to more intense crop-management, greater soil erodibility, and more rainfall. The relationships between erosion parameters and the most important vegetation attributes were also used to provide models with equations that were specific to the study region. The results of this

  16. Designing and Implementing an OVERFLOW Reader for ParaView and Comparing Performance Between Central Processing Units and Graphical Processing Units

    Science.gov (United States)

    Chawner, David M.; Gomez, Ray J.

    2010-01-01

    In the Applied Aerosciences and CFD branch at Johnson Space Center, computational simulations are run that face many challenges. Two of which are the ability to customize software for specialized needs and the need to run simulations as fast as possible. There are many different tools that are used for running these simulations and each one has its own pros and cons. Once these simulations are run, there needs to be software capable of visualizing the results in an appealing manner. Some of this software is called open source, meaning that anyone can edit the source code to make modifications and distribute it to all other users in a future release. This is very useful, especially in this branch where many different tools are being used. File readers can be written to load any file format into a program, to ease the bridging from one tool to another. Programming such a reader requires knowledge of the file format that is being read as well as the equations necessary to obtain the derived values after loading. When running these CFD simulations, extremely large files are being loaded and having values being calculated. These simulations usually take a few hours to complete, even on the fastest machines. Graphics processing units (GPUs) are usually used to load the graphics for computers; however, in recent years, GPUs are being used for more generic applications because of the speed of these processors. Applications run on GPUs have been known to run up to forty times faster than they would on normal central processing units (CPUs). If these CFD programs are extended to run on GPUs, the amount of time they would require to complete would be much less. This would allow more simulations to be run in the same amount of time and possibly perform more complex computations.

  17. The safety and regulatory process for low calorie sweeteners in the United States.

    Science.gov (United States)

    Roberts, Ashley

    2016-10-01

    Low calorie sweeteners are some of the most thoroughly tested and evaluated of all food additives. Products including aspartame and saccharin, have undergone several rounds of risk assessment by the United States Food and Drug Administration (FDA) and the European Food Safety Authority (EFSA), in relation to a number of potential safety concerns, including carcinogenicity and more recently, effects on body weight gain, glycemic control and effects on the gut microbiome. The majority of the modern day sweeteners; acesulfame K, advantame, aspartame, neotame and sucralose have been approved in the United States through the food additive process, whereas the most recent sweetener approvals for steviol glycosides and lo han guo have occurred through the Generally Recognized as Safe (GRAS) system, based on scientific procedures. While the regulatory process and review time of these two types of sweetener evaluations by the FDA differ, the same level of scientific evidence is required to support safety, so as to ensure a reasonable certainty of no harm.

  18. Fast crustal deformation computing method for multiple computations accelerated by a graphics processing unit cluster

    Science.gov (United States)

    Yamaguchi, Takuma; Ichimura, Tsuyoshi; Yagi, Yuji; Agata, Ryoichiro; Hori, Takane; Hori, Muneo

    2017-08-01

    As high-resolution observational data become more common, the demand for numerical simulations of crustal deformation using 3-D high-fidelity modelling is increasing. To increase the efficiency of performing numerical simulations with high computation costs, we developed a fast solver using heterogeneous computing, with graphics processing units (GPUs) and central processing units, and then used the solver in crustal deformation computations. The solver was based on an iterative solver and was devised so that a large proportion of the computation was calculated more quickly using GPUs. To confirm the utility of the proposed solver, we demonstrated a numerical simulation of the coseismic slip distribution estimation, which requires 360 000 crustal deformation computations with 82 196 106 degrees of freedom.

  19. Using Graphics Processing Units to solve the classical N-body problem in physics and astrophysics

    CERN Document Server

    Spera, Mario

    2014-01-01

    Graphics Processing Units (GPUs) can speed up the numerical solution of various problems in astrophysics including the dynamical evolution of stellar systems; the performance gain can be more than a factor 100 compared to using a Central Processing Unit only. In this work I describe some strategies to speed up the classical N-body problem using GPUs. I show some features of the N-body code HiGPUs as template code. In this context, I also give some hints on the parallel implementation of a regularization method and I introduce the code HiGPUs-R. Although the main application of this work concerns astrophysics, some of the presented techniques are of general validity and can be applied to other branches of physics such as electrodynamics and QCD.

  20. Advanced Investigation and Comparative Study of Graphics Processing Unit-queries Countered

    Directory of Open Access Journals (Sweden)

    A. Baskar

    2014-10-01

    Full Text Available GPU, Graphics Processing Unit, is the buzz word ruling the market these days. What is that and how has it gained that much importance is what to be answered in this research work. The study has been constructed with full attention paid towards answering the following question. What is a GPU? How is it different from a CPU? How good/bad it is computationally when comparing to CPU? Can GPU replace CPU, or it is a day dream? How significant is arrival of APU (Accelerated Processing Unit in market? What tools are needed to make GPU work? What are the improvement/focus areas for GPU to stand in the market? All the above questions are discussed and answered well in this study with relevant explanations.

  1. Ocean feedback to tropical cyclones: Climatology and processes

    Digital Repository Service at National Institute of Oceanography (India)

    Jullien, S.; Marchesiello, P.; Menkes, C.E.; Lefevre, J.; Jourdain, N.C.; Samson, G.; Lengaigne, M.

    is not instantaneous but accumulated over time within the TC inner-core. These results thus contradict the classical evaporation-wind feedback process as being essential to intensification and rather emphasize the role of macro-scale dynamics...

  2. Accelerated molecular dynamics force evaluation on graphics processing units for thermal conductivity calculations

    OpenAIRE

    Fan, Zheyong; Siro, Topi; Harju, Ari

    2012-01-01

    In this paper, we develop a highly efficient molecular dynamics code fully implemented on graphics processing units for thermal conductivity calculations using the Green-Kubo formula. We compare two different schemes for force evaluation, a previously used thread-scheme where a single thread is used for one particle and each thread calculates the total force for the corresponding particle, and a new block-scheme where a whole block is used for one particle and each thread in the block calcula...

  3. State-Level Comparison of Processes and Timelines for Distributed Photovoltaic Interconnection in the United States

    Energy Technology Data Exchange (ETDEWEB)

    Ardani, K.; Davidson, C.; Margolis, R.; Nobler, E.

    2015-01-01

    This report presents results from an analysis of distributed photovoltaic (PV) interconnection and deployment processes in the United States. Using data from more than 30,000 residential (up to 10 kilowatts) and small commercial (10-50 kilowatts) PV systems, installed from 2012 to 2014, we assess the range in project completion timelines nationally (across 87 utilities in 16 states) and in five states with active solar markets (Arizona, California, New Jersey, New York, and Colorado).

  4. Sodium content of popular commercially processed and restaurant foods in the United States ☆

    OpenAIRE

    Ahuja, Jaspreet K.C.; Shirley Wasswa-Kintu; Haytowitz, David B; Marlon Daniel; Robin Thomas; Bethany Showell; Melissa Nickle; Roseland, Janet M.; Janelle Gunn; Mary Cogswell; Pehrsson, Pamela R

    2015-01-01

    Purpose: The purpose of this study was to provide baseline estimates of sodium levels in 125 popular, sodium-contributing, commercially processed and restaurant foods in the U.S., to assess future changes as manufacturers reformulate foods. Methods: In 2010–2013, we obtained ~5200 sample units from up to 12 locations and analyzed 1654 composites for sodium and related nutrients (potassium, total dietary fiber, total and saturated fat, and total sugar), as part of the U.S. Department of Agr...

  5. Architectural and performance considerations for a 10(7)-instruction/sec optoelectronic central processing unit.

    Science.gov (United States)

    Arrathoon, R; Kozaitis, S

    1987-11-01

    Architectural considerations for a multiple-instruction, single-data-based optoelectronic central processing unit operating at 10(7) instructions per second are detailed. Central to the operation of this device is a giant fiber-optic content-addressable memory in a programmable logic array configuration. The design includes four instructions and emphasizes the fan-in and fan-out capabilities of optical systems. Interconnection limitations and scaling issues are examined.

  6. Data Handling and Processing Unit for Alphabus/Alphasat TDP-8

    Science.gov (United States)

    Habinc, Sandi; Martins, Rodolfo; Costa Pinto, Joao; Furano, Gianluca

    2011-08-01

    ESA's and Inmarsat's ARTES 8 Alphabus/Alphasat is a specific programme dedicated to the development and deployment of Alphasat. It encompasses several technology demonstration payloads (TDPs), of which the TDP8 is an Environment effects facility to monitor the GEO radiation environment and its effects on electronic components and sensors. This paper will discuss the rapid development of the processor and board for TDP8's data handling and processing unit.

  7. Orthographic units in the absence of visual processing: Evidence from sublexical structure in braille.

    Science.gov (United States)

    Fischer-Baum, Simon; Englebretson, Robert

    2016-08-01

    Reading relies on the recognition of units larger than single letters and smaller than whole words. Previous research has linked sublexical structures in reading to properties of the visual system, specifically on the parallel processing of letters that the visual system enables. But whether the visual system is essential for this to happen, or whether the recognition of sublexical structures may emerge by other means, is an open question. To address this question, we investigate braille, a writing system that relies exclusively on the tactile rather than the visual modality. We provide experimental evidence demonstrating that adult readers of (English) braille are sensitive to sublexical units. Contrary to prior assumptions in the braille research literature, we find strong evidence that braille readers do indeed access sublexical structure, namely the processing of multi-cell contractions as single orthographic units and the recognition of morphemes within morphologically-complex words. Therefore, we conclude that the recognition of sublexical structure is not exclusively tied to the visual system. However, our findings also suggest that there are aspects of morphological processing on which braille and print readers differ, and that these differences may, crucially, be related to reading using the tactile rather than the visual sensory modality.

  8. IPULOC - Exploring Dynamic Program Locality with the Instruction Processing Unit for Filling Memory Gap

    Institute of Scientific and Technical Information of China (English)

    黄震春; 李三立

    2002-01-01

    Memory gap has become an essential factor influencing the peak performance of high-speed CPU-based systems. To fill this gap, enlarging cache capacity has been a traditional method based on static program locality principle. However, the order of instructions stored in I-Cache before being sent to Data Processing Unit (DPU) is a kind of useful information that has not ever been utilized before. So an architecture containing an Instruction Processing Unit (IPU) in parallel with the ordinary DPU is proposed. The IPU can prefetch,analyze and preprocess a large amount of instructions otherwise lying in the I-Cache untouched.It is more efficient than the conventional prefetch buffer that can only store several instructions for previewing. By IPU, Load Instructions can be preprocessed while the DPU is executing on data simultaneously. It is termed as "Instruction Processing Unit with LOokahead Cache"(IPULOC for short) in which the idea of dynamic program locality is presented. This paper describes the principle of IPULOC and illustrates the quantitative parameters for evaluation.Tools for simulating the IPULOC have been developed. The simulation result shows that it can improve program locality during program execution, and hence can improve the cache hit ratio correspondingly without further enlarging the on-chip cache that occupies a large portion of chip area.

  9. FEATURES OF THE SOCIO-POLITICAL PROCESS IN THE UNITED STATES

    Directory of Open Access Journals (Sweden)

    Tatyana Evgenevna Beydina

    2017-06-01

    Full Text Available The subject of this article is the study of political and social developments of the USA at the present stage. There are four stages of the American tradition of studying political processes. The first stage is connected with substantiation of the Executive, Legislative and Judicial branches of political system (works of F. Pollack and R. Sili. The second one includes behavioral studies of politics. Besides studying political processes Charles Merriam has studied their similarities and differences. The third stage is characterized by political system studies – the works of T. Parsons, D. Easton, R. Aron, G. Almond and K. Deutsch. The fourth stage is characterized by superpower and the systems democratization problem (S. Huntington, Zb. Bzhezinsky. American social processes were qualified by R. Park, P. Sorokin, E. Giddens. The work is concentrated on the divided explanation of social and political processes of the us and the reflection of unity of American social-political reality. Academic novelty is composed of substantiation of the US social-political process concept and characterization of its features. The US social-political process is characterized by two channels: soft power and aggression. Soft power appears in the US economy dominancy. The main results of the research are features of the socio-political process in the United States. Purpose: the main goal of the research is to systematize the definition of social-political process of the USA and estimate the line of its study within American political tradition. Methodology: in this article have used methods: such as system, comparison and historical analysis, structural-functional analysis. Results: during the research the analysis of the dynamics of social and political processes of the United States had been made. Practical implications it is expedient to apply the received results in the international relation theory and practice.

  10. Evaluating Mobile Graphics Processing Units (GPUs) for Real-Time Resource Constrained Applications

    Energy Technology Data Exchange (ETDEWEB)

    Meredith, J; Conger, J; Liu, Y; Johnson, J

    2005-11-11

    Modern graphics processing units (GPUs) can provide tremendous performance boosts for some applications beyond what a single CPU can accomplish, and their performance is growing at a rate faster than CPUs as well. Mobile GPUs available for laptops have the small form factor and low power requirements suitable for use in embedded processing. We evaluated several desktop and mobile GPUs and CPUs on traditional and non-traditional graphics tasks, as well as on the most time consuming pieces of a full hyperspectral imaging application. Accuracy remained high despite small differences in arithmetic operations like rounding. Performance improvements are summarized here relative to a desktop Pentium 4 CPU.

  11. All-optical quantum computing with a hybrid solid-state processing unit

    CERN Document Server

    Pei, Pei; Li, Chong

    2011-01-01

    We develop an architecture of hybrid quantum solid-state processing unit for universal quantum computing. The architecture allows distant and nonidentical solid-state qubits in distinct physical systems to interact and work collaboratively. All the quantum computing procedures are controlled by optical methods using classical fields and cavity QED. Our methods have prominent advantage of the insensitivity to dissipation process due to the virtual excitation of subsystems. Moreover, the QND measurements and state transfer for the solid-state qubits are proposed. The architecture opens promising perspectives for implementing scalable quantum computation in a broader sense that different solid systems can merge and be integrated into one quantum processor afterwards.

  12. Graphics processing unit-based quantitative second-harmonic generation imaging.

    Science.gov (United States)

    Kabir, Mohammad Mahfuzul; Jonayat, A S M; Patel, Sanjay; Toussaint, Kimani C

    2014-09-01

    We adapt a graphics processing unit (GPU) to dynamic quantitative second-harmonic generation imaging. We demonstrate the temporal advantage of the GPU-based approach by computing the number of frames analyzed per second from SHG image videos showing varying fiber orientations. In comparison to our previously reported CPU-based approach, our GPU-based image analysis results in ∼10× improvement in computational time. This work can be adapted to other quantitative, nonlinear imaging techniques and provides a significant step toward obtaining quantitative information from fast in vivo biological processes.

  13. Real-time resampling in Fourier domain optical coherence tomography using a graphics processing unit.

    Science.gov (United States)

    Van der Jeught, Sam; Bradu, Adrian; Podoleanu, Adrian Gh

    2010-01-01

    Fourier domain optical coherence tomography (FD-OCT) requires either a linear-in-wavenumber spectrometer or a computationally heavy software algorithm to recalibrate the acquired optical signal from wavelength to wavenumber. The first method is sensitive to the position of the prism in the spectrometer, while the second method drastically slows down the system speed when it is implemented on a serially oriented central processing unit. We implement the full resampling process on a commercial graphics processing unit (GPU), distributing the necessary calculations to many stream processors that operate in parallel. A comparison between several recalibration methods is made in terms of performance and image quality. The GPU is also used to accelerate the fast Fourier transform (FFT) and to remove the background noise, thereby achieving full GPU-based signal processing without the need for extra resampling hardware. A display rate of 25 framessec is achieved for processed images (1,024 x 1,024 pixels) using a line-scan charge-coupled device (CCD) camera operating at 25.6 kHz.

  14. A Comparative Study on Retirement Process in Korea, Germany, and the United States: Identifying Determinants of Retirement Process.

    Science.gov (United States)

    Cho, Joonmo; Lee, Ayoung; Woo, Kwangho

    2016-10-01

    This study classifies the retirement process and empirically identifies the individual and institutional characteristics determining the retirement process of the aged in South Korea, Germany, and the United States. Using data from the Cross-National Equivalent File, we use a multinomial logistic regression with individual factors, public pension, and an interaction term between an occupation and an education level. We found that in Germany, the elderly with a higher education level were more likely to continue work after retirement with a relatively well-developed social support system, while in Korea, the elderly, with a lower education level in almost all occupation sectors, tended to work off and on after retirement. In the United States, the public pension and the interaction terms have no statistically significant impact on work after retirement. In both Germany and Korea, receiving a higher pension decreased the probability of working after retirement, but the influence of a pension in Korea was much greater than that of Germany. In South Korea, the elderly workers, with lower education levels, tended to work off and on repeatedly because there is no proper security in both the labor market and pension system.

  15. General Purpose Graphics Processing Unit Based High-Rate Rice Decompression and Reed-Solomon Decoding.

    Energy Technology Data Exchange (ETDEWEB)

    Loughry, Thomas A.

    2015-02-01

    As the volume of data acquired by space-based sensors increases, mission data compression/decompression and forward error correction code processing performance must likewise scale. This competency development effort was explored using the General Purpose Graphics Processing Unit (GPGPU) to accomplish high-rate Rice Decompression and high-rate Reed-Solomon (RS) decoding at the satellite mission ground station. Each algorithm was implemented and benchmarked on a single GPGPU. Distributed processing across one to four GPGPUs was also investigated. The results show that the GPGPU has considerable potential for performing satellite communication Data Signal Processing, with three times or better performance improvements and up to ten times reduction in cost over custom hardware, at least in the case of Rice Decompression and Reed-Solomon Decoding.

  16. General Purpose Graphics Processing Unit Based High-Rate Rice Decompression and Reed-Solomon Decoding

    Energy Technology Data Exchange (ETDEWEB)

    Loughry, Thomas A. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-02-01

    As the volume of data acquired by space-based sensors increases, mission data compression/decompression and forward error correction code processing performance must likewise scale. This competency development effort was explored using the General Purpose Graphics Processing Unit (GPGPU) to accomplish high-rate Rice Decompression and high-rate Reed-Solomon (RS) decoding at the satellite mission ground station. Each algorithm was implemented and benchmarked on a single GPGPU. Distributed processing across one to four GPGPUs was also investigated. The results show that the GPGPU has considerable potential for performing satellite communication Data Signal Processing, with three times or better performance improvements and up to ten times reduction in cost over custom hardware, at least in the case of Rice Decompression and Reed-Solomon Decoding.

  17. Using real time process measurements to reduce catheter related bloodstream infections in the intensive care unit

    Science.gov (United States)

    Wall, R; Ely, E; Elasy, T; Dittus, R; Foss, J; Wilkerson, K; Speroff, T

    2005-01-01

    

Problem: Measuring a process of care in real time is essential for continuous quality improvement (CQI). Our inability to measure the process of central venous catheter (CVC) care in real time prevented CQI efforts aimed at reducing catheter related bloodstream infections (CR-BSIs) from these devices. Design: A system was developed for measuring the process of CVC care in real time. We used these new process measurements to continuously monitor the system, guide CQI activities, and deliver performance feedback to providers. Setting: Adult medical intensive care unit (MICU). Key measures for improvement: Measured process of CVC care in real time; CR-BSI rate and time between CR-BSI events; and performance feedback to staff. Strategies for change: An interdisciplinary team developed a standardized, user friendly nursing checklist for CVC insertion. Infection control practitioners scanned the completed checklists into a computerized database, thereby generating real time measurements for the process of CVC insertion. Armed with these new process measurements, the team optimized the impact of a multifaceted intervention aimed at reducing CR-BSIs. Effects of change: The new checklist immediately provided real time measurements for the process of CVC insertion. These process measures allowed the team to directly monitor adherence to evidence-based guidelines. Through continuous process measurement, the team successfully overcame barriers to change, reduced the CR-BSI rate, and improved patient safety. Two years after the introduction of the checklist the CR-BSI rate remained at a historic low. Lessons learnt: Measuring the process of CVC care in real time is feasible in the ICU. When trying to improve care, real time process measurements are an excellent tool for overcoming barriers to change and enhancing the sustainability of efforts. To continually improve patient safety, healthcare organizations should continually measure their key clinical processes in real

  18. Steady electrodiffusion in hydrogel-colloid composites: macroscale properties from microscale electrokinetics

    Directory of Open Access Journals (Sweden)

    Reghan J. Hill

    2010-03-01

    Full Text Available A rigorous microscale electrokinetic model for hydrogel-colloid composites is adopted to compute macroscale profiles of electrolyte concentration, electrostatic potential, and hydrostatic pressure across membranes that separate electrolytes with different concentrations. The membranes are uncharged polymeric hydrogels in which charged spherical colloidal particles are immobilized and randomly dispersed with a low solid volume fraction. Bulk membrane characteristics and performance are calculated from a continuum microscale electrokinetic model (Hill 2006b, c. The computations undertaken in this paper quantify the streaming and membrane potentials. For the membrane potential, increasing the volume fraction of negatively charged inclusions decreases the differential electrostatic potential across the membrane under conditions where there is zero convective flow and zero electrical current. With low electrolyte concentration and highly charged nanoparticles, the membrane potential is very sensitive to the particle volume fraction. Accordingly, the membrane potential - and changes brought about by the inclusion size, charge and concentration - could be a useful experimental diagnostic to complement more recent applications of the microscale electrokinetic model for electrical microrheology and electroacoustics (Hill and Ostoja-Starzewski 2008, Wang and Hill 2008.Um modelo eletrocinético rigoroso para compósitos formados por um hidrogel e um colóide é adotado para computar os perfis macroscópicos de concentração eletrolítica, potencial eletrostático e pressão hidrostática através de uma membrana que separa soluções com diferentes concentrações eletrolíticas. A membrana é composta por um hidrogel polimérico sem carga elétrica onde partículas esféricas são imobilizadas e dispersas aleatoriamente com baixa fração de volume do sólido. As características da membrana e a sua performance são calculadas a partir de um modelo

  19. Toward a formal verification of a floating-point coprocessor and its composition with a central processing unit

    Science.gov (United States)

    Pan, Jing; Levitt, Karl N.; Cohen, Gerald C.

    1991-01-01

    Discussed here is work to formally specify and verify a floating point coprocessor based on the MC68881. The HOL verification system developed at Cambridge University was used. The coprocessor consists of two independent units: the bus interface unit used to communicate with the cpu and the arithmetic processing unit used to perform the actual calculation. Reasoning about the interaction and synchronization among processes using higher order logic is demonstrated.

  20. Large scale neural circuit mapping data analysis accelerated with the graphical processing unit (GPU)

    Science.gov (United States)

    Shi, Yulin; Veidenbaum, Alexander V.; Nicolau, Alex; Xu, Xiangmin

    2014-01-01

    Background Modern neuroscience research demands computing power. Neural circuit mapping studies such as those using laser scanning photostimulation (LSPS) produce large amounts of data and require intensive computation for post-hoc processing and analysis. New Method Here we report on the design and implementation of a cost-effective desktop computer system for accelerated experimental data processing with recent GPU computing technology. A new version of Matlab software with GPU enabled functions is used to develop programs that run on Nvidia GPUs to harness their parallel computing power. Results We evaluated both the central processing unit (CPU) and GPU-enabled computational performance of our system in benchmark testing and practical applications. The experimental results show that the GPU-CPU co-processing of simulated data and actual LSPS experimental data clearly outperformed the multi-core CPU with up to a 22x speedup, depending on computational tasks. Further, we present a comparison of numerical accuracy between GPU and CPU computation to verify the precision of GPU computation. In addition, we show how GPUs can be effectively adapted to improve the performance of commercial image processing software such as Adobe Photoshop. Comparison with Existing Method(s) To our best knowledge, this is the first demonstration of GPU application in neural circuit mapping and electrophysiology-based data processing. Conclusions Together, GPU enabled computation enhances our ability to process large-scale data sets derived from neural circuit mapping studies, allowing for increased processing speeds while retaining data precision. PMID:25277633

  1. 40 CFR Appendix Xiii to Part 266 - Mercury Bearing Wastes That May Be Processed in Exempt Mercury Recovery Units

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 26 2010-07-01 2010-07-01 false Mercury Bearing Wastes That May Be Processed in Exempt Mercury Recovery Units XIII Appendix XIII to Part 266 Protection of Environment... XIII to Part 266—Mercury Bearing Wastes That May Be Processed in Exempt Mercury Recovery Units These...

  2. 40 CFR Table 6 to Subpart Ppp of... - Process Vents From Continuous Unit Operations-Monitoring, Recordkeeping, and Reporting Requirements

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 11 2010-07-01 2010-07-01 true Process Vents From Continuous Unit Operations-Monitoring, Recordkeeping, and Reporting Requirements 6 Table 6 to Subpart PPP of Part 63... Subpart PPP of Part 63—Process Vents From Continuous Unit Operations—Monitoring, Recordkeeping, and...

  3. Using Loop Heat Pipes to Minimize Survival Heater Power for NASA's Evolutionary Xenon Thruster Power Processing Units

    Science.gov (United States)

    Choi, Michael K.

    2017-01-01

    A thermal design concept of using propylene loop heat pipes to minimize survival heater power for NASA's Evolutionary Xenon Thruster power processing units is presented. It reduces the survival heater power from 183 W to 35 W per power processing unit. The reduction is 81%.

  4. Performance Recognition for Sulphur Flotation Process Based on Froth Texture Unit Distribution

    Directory of Open Access Journals (Sweden)

    Mingfang He

    2013-01-01

    Full Text Available As an important indicator of flotation performance, froth texture is believed to be related to operational condition in sulphur flotation process. A novel fault detection method based on froth texture unit distribution (TUD is proposed to recognize the fault condition of sulphur flotation in real time. The froth texture unit number is calculated based on texture spectrum, and the probability density function (PDF of froth texture unit number is defined as texture unit distribution, which can describe the actual textual feature more accurately than the grey level dependence matrix approach. As the type of the froth TUD is unknown, a nonparametric kernel estimation method based on the fixed kernel basis is proposed, which can overcome the difficulty when comparing different TUDs under various conditions is impossible using the traditional varying kernel basis. Through transforming nonparametric description into dynamic kernel weight vectors, a principle component analysis (PCA model is established to reduce the dimensionality of the vectors. Then a threshold criterion determined by the TQ statistic based on the PCA model is proposed to realize the performance recognition. The industrial application results show that the accurate performance recognition of froth flotation can be achieved by using the proposed method.

  5. Using Discrete Event Simulation for Programming Model Exploration at Extreme-Scale: Macroscale Components for the Structural Simulation Toolkit (SST).

    Energy Technology Data Exchange (ETDEWEB)

    Wilke, Jeremiah J [Sandia National Laboratories (SNL-CA), Livermore, CA (United States); Kenny, Joseph P. [Sandia National Laboratories (SNL-CA), Livermore, CA (United States)

    2015-02-01

    Discrete event simulation provides a powerful mechanism for designing and testing new extreme- scale programming models for high-performance computing. Rather than debug, run, and wait for results on an actual system, design can first iterate through a simulator. This is particularly useful when test beds cannot be used, i.e. to explore hardware or scales that do not yet exist or are inaccessible. Here we detail the macroscale components of the structural simulation toolkit (SST). Instead of depending on trace replay or state machines, the simulator is architected to execute real code on real software stacks. Our particular user-space threading framework allows massive scales to be simulated even on small clusters. The link between the discrete event core and the threading framework allows interesting performance metrics like call graphs to be collected from a simulated run. Performance analysis via simulation can thus become an important phase in extreme-scale programming model and runtime system design via the SST macroscale components.

  6. Recent accelerating mass loss of southeast Tibetan glaciers and the relationship with changes in macroscale atmospheric circulations

    Science.gov (United States)

    Yang, Wei; Guo, Xiaofeng; Yao, Tandong; Zhu, Meilin; Wang, Yongjie

    2016-08-01

    The mass balance history (1980-2010) of a monsoon-dominated glacier in the southeast Tibetan Plateau is reconstructed using an energy balance model and later interpreted with regard to macroscale atmospheric variables. The results show that this glacier is characterized by significant interannual mass fluctuations over the past three decades, with a remarkably high mass loss during the recent period of 2003-2010. Analysis of the relationships between glacier mass balance and climatic variables shows that interannual temperature variability in the monsoonal season (June-September) is a primary driver of its mass balance fluctuations, but monsoonal precipitation tends to play an accentuated role for driving the observed glacier mass changes due to their covariation (concurrence of warm/dry and cold/wet climates) in the monsoon-influenced southeast Tibetan Plateau. Analysis of the atmospheric circulation pattern reveals that the predominance of anticyclonic/cyclonic circulations prevailing in the southeastern/northern Tibetan Plateau during 2003-2010 contributes to increased air temperature and decreased precipitation in the southeast Tibetan Plateau. Regionally contrasting atmospheric circulations explain the distinct mass changes between in the monsoon-influenced southeast Tibetan Plateau and in the north Tibetan Plateau/Tien Shan Mountains during 2003-2010. The macroscale climate change seems to be linked with the Europe-Asia teleconnection.

  7. Simulation of abrasive water jet cutting process: Part 1. Unit event approach

    Science.gov (United States)

    Lebar, Andrej; Junkar, Mihael

    2004-11-01

    Abrasive water jet (AWJ) machined surfaces exhibit the texture typical of machining with high energy density beam processing technologies. It has a superior surface quality in the upper region and rough surface in the lower zone with pronounced texture marks called striations. The nature of the mechanisms involved in the domain of AWJ machining is still not well understood but is essential for AWJ control improvement. In this paper, the development of an AWJ machining simulation is reported on. It is based on an AWJ process unit event, which in this case represents the impact of a particular abrasive grain. The geometrical characteristics of the unit event are measured on a physical model of the AWJ process. The measured dependences and the proposed model relations are then implemented in the AWJ machining process simulation. The obtained results are in good agreement in the engraving regime of AWJ machining. To expand the validity of the simulation further, a cellular automata approach is explored in the second part of the paper.

  8. Factors associated with the process of adaptation among Pakistani adolescent females living in United States.

    Science.gov (United States)

    Khuwaja, Salma A; Selwyn, Beatrice J; Mgbere, Osaro; Khuwaja, Alam; Kapadia, Asha; McCurdy, Sheryl; Hsu, Chiehwen E

    2013-04-01

    This study explored post-migration experiences of recently migrated Pakistani Muslim adolescent females residing in the United States. In-depth, semi-structured interviews were conducted with thirty Pakistani Muslim adolescent females between the ages of 15 and 18 years living with their families in Houston, Texas. Data obtained from the interviews were evaluated using discourse analysis to identify major reoccurring themes. Participants discussed factors associated with the process of adaptation to the American culture. The results revealed that the main factors associated with adaptation process included positive motivation for migration, family bonding, social support networks, inter-familial communication, aspiration of adolescents to learn other cultures, availability of English-as-second-language programs, participation in community rebuilding activities, and faith practices, English proficiency, peer pressure, and inter-generational conflicts. This study provided much needed information on factors associated with adaptation process of Pakistani Muslim adolescent females in the United States. The results have important implications for improving the adaptation process of this group and offer potential directions for intervention and counseling services.

  9. Implementation and practical application of the nutrition care process in the dialysis unit.

    Science.gov (United States)

    Memmer, Debra

    2013-01-01

    The Nutrition Care Process (NCP) was introduced in 2003 (Lacey and Pritchett, J Am Diet Assoc. 2003;103:1061-1071). Since then, dietitians have been encouraged to incorporate the NCP into their daily practice, yet it has not been totally adopted in all dialysis units (Dent and McDuffie, J Ren Nutr. 2011;1:205-207). The renal dietitian has the benefit of being able to follow-up with the dialysis patient on a monthly basis. During these monthly visits, as information unfolds, a unique relationship culminates with the dialysis patient. The NCP allows the dietitian to make precise nutrition diagnoses, which reflect the complexity of the renal dietitian's involvement with the dialysis patient. The purpose of this article is to provide a brief description of the NCP as it relates to dialysis, offer a framework on how to begin using the NCP in the dialysis unit, and provide an example of a monthly nutrition note.

  10. Usability of computerized nursing process from the ICNP® in intensive care units

    Directory of Open Access Journals (Sweden)

    Daniela Couto Carvalho Barra

    2015-04-01

    Full Text Available OBJECTIVE To analyze the usability of Computerized Nursing Process (CNP from the ICNP® 1.0 in Intensive Care Units in accordance with the criteria established by the standards of the International Organization for Standardization and the Brazilian Association of Technical Standards of systems. METHOD This is a before-and-after semi-experimental quantitative study, with a sample of 34 participants (nurses, professors and systems programmers, carried out in three Intensive Care Units. RESULTS The evaluated criteria (use, content and interface showed that CNP has usability criteria, as it integrates a logical data structure, clinical assessment, diagnostics and nursing interventions. CONCLUSION The CNP is a source of information and knowledge that provide nurses with new ways of learning in intensive care, for it is a place that provides complete, comprehensive, and detailed content, supported by current and relevant data and scientific research information for Nursing practices.

  11. Free and open source simulation tools for the design of power processing units for photovoltaic systems

    Directory of Open Access Journals (Sweden)

    Sergio Morales-Hernández

    2015-06-01

    Full Text Available Renewable energy sources, including solar photovoltaic, require electronic circuits that serve as interface between the transducer device and the device or system that uses energy. Moreover, the energy efficiency and the cost of the system can be compromised if such electronic circuit is not designed properly. Given that the electrical characteristics of the photovoltaic devices are nonlinear and that the most efficient electronic circuits for power processing are naturally discontinuous, a detailed dynamic analysis to optimize the design is required. This analysis should be supported by computer simulation tools. In this paper a comparison between two software tools for dynamic system simulation is performed to determinate its usefulness in the design process of photovoltaic systems, mainly in what corresponds to the power processing units. Using as a case of study a photovoltaic system for battery charging it was determined that Scicoslab tool was the most suitable.

  12. Future evolution of the Fast TracKer (FTK) processing unit

    CERN Document Server

    Gentsos, C; The ATLAS collaboration; Giannetti, P; Magalotti, D; Nikolaidis, S

    2014-01-01

    The Fast Tracker (FTK) processor [1] for the ATLAS experiment has a computing core made of 128 Processing Units that reconstruct tracks in the silicon detector in a ~100 μsec deep pipeline. The track parameter resolution provided by FTK enables the HLT trigger to identify efficiently and reconstruct significant samples of fermionic Higgs decays. Data processing speed is achieved with custom VLSI pattern recognition, linearized track fitting executed inside modern FPGAs, pipelining, and parallel processing. One large FPGA executes full resolution track fitting inside low resolution candidate tracks found by a set of 16 custom Asic devices, called Associative Memories (AM chips) [2]. The FTK dual structure, based on the cooperation of VLSI dedicated AM and programmable FPGAs, is maintained to achieve further technology performance, miniaturization and integration of the current state of the art prototypes. This allows to fully exploit new applications within and outside the High Energy Physics field. We plan t...

  13. [Variations in the diagnostic confirmation process between breast cancer mass screening units].

    Science.gov (United States)

    Natal, Carmen; Fernández-Somoano, Ana; Torá-Rocamora, Isabel; Tardón, Adonina; Castells, Xavier

    2016-01-01

    To analyse variations in the diagnostic confirmation process between screening units, variations in the outcome of each episode and the relationship between the use of the different diagnostic confirmation tests and the lesion detection rate. Observational study of variability of the standardised use of diagnostic and lesion detection tests in 34 breast cancer mass screening units participating in early-detection programmes in three Spanish regions from 2002-2011. The diagnostic test variation ratio in percentiles 25-75 ranged from 1.68 (further appointments) to 3.39 (fine-needle aspiration). The variation ratio in detection rates of benign lesions, ductal carcinoma in situ and invasive cancer were 2.79, 1.99 and 1.36, respectively. A positive relationship between rates of testing and detection rates was found with fine-needle aspiration-benign lesions (R(2): 0.53), fine-needle aspiration-invasive carcinoma (R(2): 0 28), core biopsy-benign lesions (R(2): 0.64), core biopsy-ductal carcinoma in situ (R(2): 0.61) and core biopsy-invasive carcinoma (R(2): 0.48). Variation in the use of invasive tests between the breast cancer screening units participating in early-detection programmes was found to be significantly higher than variations in lesion detection. Units which conducted more fine-needle aspiration tests had higher benign lesion detection rates, while units that conducted more core biopsies detected more benign lesions and cancer. Copyright © 2016 SESPAS. Published by Elsevier Espana. All rights reserved.

  14. HTS current lead units prepared by the TFA-MOD processed YBCO coated conductors

    Energy Technology Data Exchange (ETDEWEB)

    Shiohara, K.; Sakai, S.; Ishii, Y. [Tokai University, 1117 Kita-Kaname, Hiratsuka, Kanagawa 259-1292 (Japan); Yamada, Y., E-mail: yyamaday@keyaki.cc.u-tokai.ac.j [Tokai University, 1117 Kita-Kaname, Hiratsuka, Kanagawa 259-1292 (Japan); Tachikawa, K. [Tokai University, 1117 Kita-Kaname, Hiratsuka, Kanagawa 259-1292 (Japan); Koizumi, T.; Aoki, Y.; Hasegawa, T. [SWCC Showa Cable System Co., LTD, 4-1-1 Minami-Hashimoto, Sagamihara, Kanagawa 229-1133 (Japan); Tamura, H.; Mito, T. [NIFS National Institute for Fusion Science, 322-6 Oroshi, Toki, Gifu 509-5292 (Japan)

    2010-11-01

    Two superconducting current lead units have been prepared using ten coated conductors of the Tri-Fluoro-Acetate - Metal Organic Deposition (TFA-MOD) processed Y{sub 1}Ba{sub 2}Cu{sub 3}O{sub 7-{delta}}(YBCO) coated conductors with critical current (I{sub c}) of about 170 A at 77 K in self-field. The coated conductors are 5 mm in width, 190 mm in length and about 120 {mu}m in overall thickness. The 1.5 {mu}m thick superconducting YBCO layer was synthesized through the TFA-MOD process on Hastelloy{sup TM} C-276 substrate tape with two buffer oxide layers of Gd{sub 2}Zr{sub 2}O{sub 7} and CeO{sub 2}. The five YBCO coated conductors are attached on a 1 mm thick Glass Fiber Reinforced Plastics (GFRP) board and soldered to Cu caps at the both ends. We prepared two 500 A-class current lead units. The DC transport current of 800 A was stably applied at 77 K without any voltage generation in all coated conductors. The voltage between both Cu caps linearly increased with increasing the applied current, and was about 350 {mu}V at 500 A in both current lead units. According to the estimated values of the heat leakage from 77 K to 4.2 K, the heat leakage for the current lead unit was 46.5 mW. We successfully attained reduction of the heat leakage because of improvement of the transport current performance (I{sub c}), a thinner Ag layer of YBCO coated conductor and usage of the GFRP board for reinforcement instead of a stainless steel board used in the previous study. The DC transport current of 1400 A was stably applied when the two current lead units were joined in parallel. The sum of the heat leakages from 77 K to 4.2 K for the combined the current lead units was 93 mW. In comparison with the conventional Cu current leads by gas-cooling, it could be noted that the heat leakage of the current lead is about one order of magnitude smaller than that of the Cu current lead.

  15. Motor unit action potential conduction velocity estimated from surface electromyographic signals using image processing techniques.

    Science.gov (United States)

    Soares, Fabiano Araujo; Carvalho, João Luiz Azevedo; Miosso, Cristiano Jacques; de Andrade, Marcelino Monteiro; da Rocha, Adson Ferreira

    2015-09-17

    In surface electromyography (surface EMG, or S-EMG), conduction velocity (CV) refers to the velocity at which the motor unit action potentials (MUAPs) propagate along the muscle fibers, during contractions. The CV is related to the type and diameter of the muscle fibers, ion concentration, pH, and firing rate of the motor units (MUs). The CV can be used in the evaluation of contractile properties of MUs, and of muscle fatigue. The most popular methods for CV estimation are those based on maximum likelihood estimation (MLE). This work proposes an algorithm for estimating CV from S-EMG signals, using digital image processing techniques. The proposed approach is demonstrated and evaluated, using both simulated and experimentally-acquired multichannel S-EMG signals. We show that the proposed algorithm is as precise and accurate as the MLE method in typical conditions of noise and CV. The proposed method is not susceptible to errors associated with MUAP propagation direction or inadequate initialization parameters, which are common with the MLE algorithm. Image processing -based approaches may be useful in S-EMG analysis to extract different physiological parameters from multichannel S-EMG signals. Other new methods based on image processing could also be developed to help solving other tasks in EMG analysis, such as estimation of the CV for individual MUs, localization and tracking of innervation zones, and study of MU recruitment strategies.

  16. Parallel particle swarm optimization on a graphics processing unit with application to trajectory optimization

    Science.gov (United States)

    Wu, Q.; Xiong, F.; Wang, F.; Xiong, Y.

    2016-10-01

    In order to reduce the computational time, a fully parallel implementation of the particle swarm optimization (PSO) algorithm on a graphics processing unit (GPU) is presented. Instead of being executed on the central processing unit (CPU) sequentially, PSO is executed in parallel via the GPU on the compute unified device architecture (CUDA) platform. The processes of fitness evaluation, updating of velocity and position of all particles are all parallelized and introduced in detail. Comparative studies on the optimization of four benchmark functions and a trajectory optimization problem are conducted by running PSO on the GPU (GPU-PSO) and CPU (CPU-PSO). The impact of design dimension, number of particles and size of the thread-block in the GPU and their interactions on the computational time is investigated. The results show that the computational time of the developed GPU-PSO is much shorter than that of CPU-PSO, with comparable accuracy, which demonstrates the remarkable speed-up capability of GPU-PSO.

  17. THE APPLICATION OF THE “UNIVERSAL” MATHEMATICAL MODELS WITHIN THE ANALYSIS OF MUNICIPAL UNIT PROCESSES

    Directory of Open Access Journals (Sweden)

    Tatyana Nikolayevna Gordeeva

    2017-06-01

    Full Text Available The article describes the application possibility of the well-known Lotka-Volterra interaction model to the analysis of processes within a municipal unit as an administrative territory operating as a lower managerial level, i.e. local self-government. Object: to prove the application of the mathematical models in management sociology and local self-government sociology. The approach realization is possible due to universality of the mathematical models which can be applied in different areas regardless of their specificity. Methods or methodology of the research: deals with features of ODE solutions without their solving. Its basis was formed in the classical researches of H. Poincare and A.M. Lyapunov in the late XIX century. Nowadays the methods are widely used to analyze evolution systems describing dynamic processes in mechanics and physics, as well as in economics, ecology, medicine, and sociology. Results: The article proves applicability of the developed mathematical models within the analysis of municipal unit processes. Application field for the results: the interpretation of the model analysis results gives an opportunity to amplify and extend scientific basis in the field of management sociology and local self-government sociology.

  18. Real-time display on Fourier domain optical coherence tomography system using a graphics processing unit.

    Science.gov (United States)

    Watanabe, Yuuki; Itagaki, Toshiki

    2009-01-01

    Fourier domain optical coherence tomography (FD-OCT) requires resampling of spectrally resolved depth information from wavelength to wave number, and the subsequent application of the inverse Fourier transform. The display rates of OCT images are much slower than the image acquisition rates due to processing speed limitations on most computers. We demonstrate a real-time display of processed OCT images using a linear-in-wave-number (linear-k) spectrometer and a graphics processing unit (GPU). We use the linear-k spectrometer with the combination of a diffractive grating with 1200 lines/mm and a F2 equilateral prism in the 840-nm spectral region to avoid calculating the resampling process. The calculations of the fast Fourier transform (FFT) are accelerated by the GPU with many stream processors, which realizes highly parallel processing. A display rate of 27.9 frames/sec for processed images (2048 FFT size x 1000 lateral A-scans) is achieved in our OCT system using a line scan CCD camera operated at 27.9 kHz.

  19. An Optimization System with Parallel Processing for Reducing Common-Mode Current on Electronic Control Unit

    Science.gov (United States)

    Okazaki, Yuji; Uno, Takanori; Asai, Hideki

    In this paper, we propose an optimization system with parallel processing for reducing electromagnetic interference (EMI) on electronic control unit (ECU). We adopt simulated annealing (SA), genetic algorithm (GA) and taboo search (TS) to seek optimal solutions, and a Spice-like circuit simulator to analyze common-mode current. Therefore, the proposed system can determine the adequate combinations of the parasitic inductance and capacitance values on printed circuit board (PCB) efficiently and practically, to reduce EMI caused by the common-mode current. Finally, we apply the proposed system to an example circuit to verify the validity and efficiency of the system.

  20. Employing OpenCL to Accelerate Ab Initio Calculations on Graphics Processing Units.

    Science.gov (United States)

    Kussmann, Jörg; Ochsenfeld, Christian

    2017-06-13

    We present an extension of our graphics processing units (GPU)-accelerated quantum chemistry package to employ OpenCL compute kernels, which can be executed on a wide range of computing devices like CPUs, Intel Xeon Phi, and AMD GPUs. Here, we focus on the use of AMD GPUs and discuss differences as compared to CUDA-based calculations on NVIDIA GPUs. First illustrative timings are presented for hybrid density functional theory calculations using serial as well as parallel compute environments. The results show that AMD GPUs are as fast or faster than comparable NVIDIA GPUs and provide a viable alternative for quantum chemical applications.

  1. Monte Carlo Simulations of Random Frustrated Systems on Graphics Processing Units

    Science.gov (United States)

    Feng, Sheng; Fang, Ye; Hall, Sean; Papke, Ariane; Thomasson, Cade; Tam, Ka-Ming; Moreno, Juana; Jarrell, Mark

    2012-02-01

    We study the implementation of the classical Monte Carlo simulation for random frustrated models using the multithreaded computing environment provided by the the Compute Unified Device Architecture (CUDA) on modern Graphics Processing Units (GPU) with hundreds of cores and high memory bandwidth. The key for optimizing the performance of the GPU computing is in the proper handling of the data structure. Utilizing the multi-spin coding, we obtain an efficient GPU implementation of the parallel tempering Monte Carlo simulation for the Edwards-Anderson spin glass model. In the typical simulations, we find over two thousand times of speed-up over the single threaded CPU implementation.

  2. Uncontracted Rys Quadrature Implementation of up to G Functions on Graphical Processing Units.

    Science.gov (United States)

    Asadchev, Andrey; Allada, Veerendra; Felder, Jacob; Bode, Brett M; Gordon, Mark S; Windus, Theresa L

    2010-03-09

    An implementation is presented of an uncontracted Rys quadrature algorithm for electron repulsion integrals, including up to g functions on graphical processing units (GPUs). The general GPU programming model, the challenges associated with implementing the Rys quadrature on these highly parallel emerging architectures, and a new approach to implementing the quadrature are outlined. The performance of the implementation is evaluated for single and double precision on two different types of GPU devices. The performance obtained is on par with the matrix-vector routine from the CUDA basic linear algebra subroutines (CUBLAS) library.

  3. Efficient neighbor list calculation for molecular simulation of colloidal systems using graphics processing units

    Science.gov (United States)

    Howard, Michael P.; Anderson, Joshua A.; Nikoubashman, Arash; Glotzer, Sharon C.; Panagiotopoulos, Athanassios Z.

    2016-06-01

    We present an algorithm based on linear bounding volume hierarchies (LBVHs) for computing neighbor (Verlet) lists using graphics processing units (GPUs) for colloidal systems characterized by large size disparities. We compare this to a GPU implementation of the current state-of-the-art CPU algorithm based on stenciled cell lists. We report benchmarks for both neighbor list algorithms in a Lennard-Jones binary mixture with synthetic interaction range disparity and a realistic colloid solution. LBVHs outperformed the stenciled cell lists for systems with moderate or large size disparity and dilute or semidilute fractions of large particles, conditions typical of colloidal systems.

  4. Conversion of a deasphalting unit for use in the process of supercritical solvent recovery

    Directory of Open Access Journals (Sweden)

    Waintraub S.

    2000-01-01

    Full Text Available In order to reduce energy consumption and to increase deasphalted oil yield, an old PETROBRAS deasphalting unit was converted for use in the process of supercritical solvent recovery. In-plant and pilot tests were performed to determine the ideal solvent-to-oil ratio. The optimum conditions for separation of the supercritical solvent from the solvent-plus-oil liquid mixture were determined by experimental tests in PVT cells. These tests also allowed measurement of the dew and bubble points, determination of the retrograde region, observation of supercritical fluid compressibility and as a result construction of a phase equilibrium diagram.

  5. Accelerated 3D Monte Carlo light dosimetry using a graphics processing unit (GPU) cluster

    Science.gov (United States)

    Lo, William Chun Yip; Lilge, Lothar

    2010-11-01

    This paper presents a basic computational framework for real-time, 3-D light dosimetry on graphics processing unit (GPU) clusters. The GPU-based approach offers a direct solution to overcome the long computation time preventing Monte Carlo simulations from being used in complex optimization problems such as treatment planning, particularly if simulated annealing is employed as the optimization algorithm. The current multi- GPU implementation is validated using a commercial light modelling software (ASAP from Breault Research Organization). It also supports the latest Fermi GPU architecture and features an interactive 3-D visualization interface. The software is available for download at http://code.google.com/p/gpu3d.

  6. The Open Physiology workflow: modeling processes over physiology circuitboards of interoperable tissue units

    Science.gov (United States)

    de Bono, Bernard; Safaei, Soroush; Grenon, Pierre; Nickerson, David P.; Alexander, Samuel; Helvensteijn, Michiel; Kok, Joost N.; Kokash, Natallia; Wu, Alan; Yu, Tommy; Hunter, Peter; Baldock, Richard A.

    2015-01-01

    A key challenge for the physiology modeling community is to enable the searching, objective comparison and, ultimately, re-use of models and associated data that are interoperable in terms of their physiological meaning. In this work, we outline the development of a workflow to modularize the simulation of tissue-level processes in physiology. In particular, we show how, via this approach, we can systematically extract, parcellate and annotate tissue histology data to represent component units of tissue function. These functional units are semantically interoperable, in terms of their physiological meaning. In particular, they are interoperable with respect to [i] each other and with respect to [ii] a circuitboard representation of long-range advective routes of fluid flow over which to model long-range molecular exchange between these units. We exemplify this approach through the combination of models for physiology-based pharmacokinetics and pharmacodynamics to quantitatively depict biological mechanisms across multiple scales. Links to the data, models and software components that constitute this workflow are found at http://open-physiology.org/. PMID:25759670

  7. Ultra-processed food consumption in children from a Basic Health Unit.

    Science.gov (United States)

    Sparrenberger, Karen; Friedrich, Roberta Roggia; Schiffner, Mariana Dihl; Schuch, Ilaine; Wagner, Mário Bernardes

    2015-01-01

    To evaluate the contribution of ultra-processed food (UPF) on the dietary consumption of children treated at a Basic Health Unit and the associated factors. Cross-sectional study carried out with a convenience sample of 204 children, aged 2-10 years old, in Southern Brazil. Children's food intake was assessed using a 24-h recall questionnaire. Food items were classified as minimally processed, processed for culinary use, and ultra-processed. A semi-structured questionnaire was applied to collect socio-demographic and anthropometric variables. Overweight in children was classified using a Z score >2 for children younger than 5 and Z score >+1 for those aged between 5 and 10 years, using the body mass index for age. Overweight frequency was 34% (95% CI: 28-41%). Mean energy consumption was 1672.3 kcal/day, with 47% (95% CI: 45-49%) coming from ultra-processed food. In the multiple linear regression model, maternal education (r=0.23; p=0.001) and child age (r=0.40; pfactors associated with a greater percentage of UPF in the diet (r=0.42; pfactor for the consumption of such products. Copyright © 2015 Sociedade Brasileira de Pediatria. Published by Elsevier Editora Ltda. All rights reserved.

  8. Effect of a "Lean" intervention to improve safety processes and outcomes on a surgical emergency unit.

    Science.gov (United States)

    McCulloch, Peter; Kreckler, Simon; New, Steve; Sheena, Yezen; Handa, Ashok; Catchpole, Ken

    2010-11-02

    Emergency surgical patients are at high risk for harm because of errors in care. Quality improvement methods that involve process redesign, such as “Lean,” appear to improve service reliability and efficiency in healthcare. Interrupted time series. The emergency general surgery ward of a university hospital in the United Kingdom. Seven safety relevant care processes. A Lean intervention targeting five of the seven care processes relevant to patient safety. 969 patients were admitted during the four month study period before the introduction of the Lean intervention (May to August 2007), and 1114 were admitted during the four month period after completion of the intervention (May to August 2008). Compliance with the five process measures targeted for Lean intervention (but not the two that were not) improved significantly (relative improvement 28% to 149%; PLean can substantially and simultaneously improve compliance with a bundle of safety related processes. Given the interconnected nature of hospital care, this strategy might not translate into improvements in safety outcomes unless a system-wide approach is adopted to remove barriers to change.

  9. United States Department of Energy Integrated Manufacturing & Processing Predoctoral Fellowships. Final Report

    Energy Technology Data Exchange (ETDEWEB)

    Petrochenkov, M.

    2003-03-31

    The objective of the program was threefold: to create a pool of PhDs trained in the integrated approach to manufacturing and processing, to promote academic interest in the field, and to attract talented professionals to this challenging area of engineering. It was anticipated that the program would result in the creation of new manufacturing methods that would contribute to improved energy efficiency, to better utilization of scarce resources, and to less degradation of the environment. Emphasis in the competition was on integrated systems of manufacturing and the integration of product design with manufacturing processes. Research addressed such related areas as aspects of unit operations, tooling and equipment, intelligent sensors, and manufacturing systems as they related to product design.

  10. ANALYSIS OF THE PROCESS OF ORGANIC CERTIFICATION IN THE UNITED KINGDOM

    Directory of Open Access Journals (Sweden)

    Alexandra MUSCĂNESCU

    2013-01-01

    Full Text Available Every product marketed under ’organic’ must comply with a set of UK, European Union and international rules. These rules (also known as standards assure the consumers that they bought original organic products able to be traced back all the way to the farm. Organic standards cover all of the aspects of organic food certification, including production and packaging, animal welfare, wildlife preservation and interdict all useless and dangerous food additives in processing organic foods. The British organic control bodies authorize individual organic operators. These are approved by DEFRA (The Department for Environment, Food and Rural Affairs. In order for a producer to have his merchandise marketed as organic, he must be affiliated to one of the control bodies and comply to its standards and requirements. The present article presents the certification control bodies in organic agriculture as well as the steps to be taken in the organic certification process within the United Kingdom.

  11. Productivity Gap and Asymmetric Trade Relations: The Canada-United States of America Integration Process

    Directory of Open Access Journals (Sweden)

    Germán H. Gonzalez

    2014-08-01

    Full Text Available The usefulness of the European model of integration is currently subject to debate and the North American integration process has been largely ignored as a comparative framework. The asymmetrical relationship between Canada and the United States began a long time before NAFTA, and the study of this process could shed light on the usual problems faced by Latin American countries. This article attempts to encourage discussion about this topic. Particularly,there is evidence for a substantial and positive change in Canadian productivity at the time of the Canada-US Free Trade Agreement (CUFTA. However, the enactment of the North American Free Trade Agreement (NAFTA does not seem to have had the same effect as the earlier treaty.

  12. Extreme Environment Capable, Modular and Scalable Power Processing Unit for Solar Electric Propulsion

    Science.gov (United States)

    Carr, Gregory A.; Iannello, Christopher J.; Chen, Yuan; Hunter, Don J.; Del Castillo, Linda; Bradley, Arthur T.; Stell, Christopher; Mojarradi, Mohammad M.

    2013-01-01

    This paper is to present a concept of a modular and scalable High Temperature Boost (HTB) Power Processing Unit (PPU) capable of operating at temperatures beyond the standard military temperature range. The various extreme environments technologies are also described as the fundamental technology path to this concept. The proposed HTB PPU is intended for power processing in the area of space solar electric propulsion, where the reduction of in-space mass and volume are desired, and sometimes even critical, to achieve the goals of future space flight missions. The concept of the HTB PPU can also be applied to other extreme environment applications, such as geothermal and petroleum deep-well drilling, where higher temperature operation is required.

  13. Silicon-Carbide Power MOSFET Performance in High Efficiency Boost Power Processing Unit for Extreme Environments

    Science.gov (United States)

    Ikpe, Stanley A.; Lauenstein, Jean-Marie; Carr, Gregory A.; Hunter, Don; Ludwig, Lawrence L.; Wood, William; Del Castillo, Linda Y.; Fitzpatrick, Fred; Chen, Yuan

    2016-01-01

    Silicon-Carbide device technology has generated much interest in recent years. With superior thermal performance, power ratings and potential switching frequencies over its Silicon counterpart, Silicon-Carbide offers a greater possibility for high powered switching applications in extreme environment. In particular, Silicon-Carbide Metal-Oxide- Semiconductor Field-Effect Transistors' (MOSFETs) maturing process technology has produced a plethora of commercially available power dense, low on-state resistance devices capable of switching at high frequencies. A novel hard-switched power processing unit (PPU) is implemented utilizing Silicon-Carbide power devices. Accelerated life data is captured and assessed in conjunction with a damage accumulation model of gate oxide and drain-source junction lifetime to evaluate potential system performance at high temperature environments.

  14. Fast direct reconstruction strategy of dynamic fluorescence molecular tomography using graphics processing units

    Science.gov (United States)

    Chen, Maomao; Zhang, Jiulou; Cai, Chuangjian; Gao, Yang; Luo, Jianwen

    2016-06-01

    Dynamic fluorescence molecular tomography (DFMT) is a valuable method to evaluate the metabolic process of contrast agents in different organs in vivo, and direct reconstruction methods can improve the temporal resolution of DFMT. However, challenges still remain due to the large time consumption of the direct reconstruction methods. An acceleration strategy using graphics processing units (GPU) is presented. The procedure of conjugate gradient optimization in the direct reconstruction method is programmed using the compute unified device architecture and then accelerated on GPU. Numerical simulations and in vivo experiments are performed to validate the feasibility of the strategy. The results demonstrate that, compared with the traditional method, the proposed strategy can reduce the time consumption by ˜90% without a degradation of quality.

  15. Acceleration of Early-Photon Fluorescence Molecular Tomography with Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Xin Wang

    2013-01-01

    Full Text Available Fluorescence molecular tomography (FMT with early-photons can improve the spatial resolution and fidelity of the reconstructed results. However, its computing scale is always large which limits its applications. In this paper, we introduced an acceleration strategy for the early-photon FMT with graphics processing units (GPUs. According to the procedure, the whole solution of FMT was divided into several modules and the time consumption for each module is studied. In this strategy, two most time consuming modules (Gd and W modules were accelerated with GPU, respectively, while the other modules remained coded in the Matlab. Several simulation studies with a heterogeneous digital mouse atlas were performed to confirm the performance of the acceleration strategy. The results confirmed the feasibility of the strategy and showed that the processing speed was improved significantly.

  16. Processing techniques for data from the Kuosheng Unit 1 shakedown safety-relief-valve tests

    Energy Technology Data Exchange (ETDEWEB)

    McCauley, E.W.; Rompel, S.L.; Weaver, H.J.; Altenbach, T.J.

    1982-08-01

    This report describes techniques developed at the Lawrence Livermore National Laobratory, Livermore, CA for processing original data from the Taiwan Power Company's Kuosheng MKIII Unit 1 Safety Relief Valve Shakedown Tests conducted in April/May 1981. The computer codes used, TPSORT, TPPLOT, and TPPSD, form a special evaluation system for treating the data from its original packed binary form to ordered, calibrated ASCII transducer files and then to production of time-history plots, numerical output files, and spectral analyses. Using the data processing techniques described, a convenient means of independently examining and analyzing a unique data base for steam condensation phenomena in the MARKIII wetwell is described. The techniques developed for handling these data are applicable to the treatment of similar, but perhaps differently structured, experiment data sets.

  17. Speedup for quantum optimal control from automatic differentiation based on graphics processing units

    Science.gov (United States)

    Leung, Nelson; Abdelhafez, Mohamed; Koch, Jens; Schuster, David

    2017-04-01

    We implement a quantum optimal control algorithm based on automatic differentiation and harness the acceleration afforded by graphics processing units (GPUs). Automatic differentiation allows us to specify advanced optimization criteria and incorporate them in the optimization process with ease. We show that the use of GPUs can speedup calculations by more than an order of magnitude. Our strategy facilitates efficient numerical simulations on affordable desktop computers and exploration of a host of optimization constraints and system parameters relevant to real-life experiments. We demonstrate optimization of quantum evolution based on fine-grained evaluation of performance at each intermediate time step, thus enabling more intricate control on the evolution path, suppression of departures from the truncated model subspace, as well as minimization of the physical time needed to perform high-fidelity state preparation and unitary gates.

  18. ACTION OF UNIFORM SEARCH ALGORITHM WHEN SELECTING LANGUAGE UNITS IN THE PROCESS OF SPEECH

    Directory of Open Access Journals (Sweden)

    Nekipelova Irina Mikhaylovna

    2013-04-01

    Full Text Available The article is devoted to research of action of uniform search algorithm when selecting by human of language units for speech produce. The process is connected with a speech optimization phenomenon. This makes it possible to shorten the time of cogitation something that human want to say, and to achieve the maximum precision in thoughts expression. The algorithm of uniform search works at consciousness and subconsciousness levels. It favours the forming of automatism produce and perception of speech. Realization of human's cognitive potential in the process of communication starts up complicated mechanism of self-organization and self-regulation of language. In turn, it results in optimization of language system, servicing needs not only human's self-actualization but realization of communication in society. The method of problem-oriented search is used for researching of optimization mechanisms, which are distinctive to speech producing and stabilization of language.

  19. Prospects for expanded mohair and cashmere production and processing in the United States of America.

    Science.gov (United States)

    Lupton, C J

    1996-05-01

    Mohair from Angora goats has been produced in the United States since the introduction of these animals from Turkey in 1849. Cashmere on Texas meat goats was reported in 1973, but domestic interest in commercial production did not occur until the mid-1980s. Since 1982, the average prices of U.S.-produced mohair and cashmere (de-haired) have ranged from $1.81 to $9.48/kg and approximately $55 to $200/kg, respectively. However, return to producers from mohair has been relatively constant, averaging $10.21/kg, due to the federal incentive program. Because this program is scheduled to terminate with final payment in 1996, the future of mohair profitability is questionable. Prospects for expanded mohair and cashmere production and processing in the United States are influenced by numerous interacting factors and potential constraints. These include the prospect that the goat and textile industries may no longer be profitable in the absence of clear government policies. Although selection may have slightly increased fiber production by Angoras (long term) and domestic meat goats (short term), availability of genetic resources may prove to be a constraint to increased fiber production by cashmere goats and improved meat production by both types of goat. Land resources are plentiful unless new government policies prohibit goats from vast tracts of rangeland and forest because of environmental concerns. Future demand is an unknown, but with increasing world population and affluence, prospects for long-term improved demand for luxury fibers seem good. Competition from foreign cashmere growers is expected, whereas, in the short term, mohair production overseas is declining. However, increased processing of cashmere in its country of origin is expected to result in shortages of raw materials for European and U.S. processors. The amount of scouring, worsted, and woolen equipment in the United States is adequate to accommodate major increases in domestic processing of goat

  20. Lightweight concrete masonry units based on processed granulate of corn cob as aggregate

    Directory of Open Access Journals (Sweden)

    Faustino, J.

    2015-06-01

    Full Text Available A research work was performed in order to assess the potential application of processed granulate of corn cob (PCC as an alternative lightweight aggregate for the manufacturing process of lightweight concrete masonry units (CMU. Therefore, CMU-PCC were prepared in a factory using a typical lightweight concrete mixture for non-structural purposes. Additionally, lightweight concrete masonry units based on a currently applied lightweight aggregate such as expanded clay (CMU-EC were also manufactured. An experimental work allowed achieving a set of results that suggest that the proposed building product presents interesting material properties within the masonry wall context. Therefore, this unit is promising for both interior and exterior applications. This conclusion is even more relevant considering that corn cob is an agricultural waste product.En este trabajo de investigación se evaluó la posible aplicación de granulado procesado de la mazorca de maiz como un árido ligero alternativo en el proceso de fabricación de unidades de mampostería de hormigón ligero. Con esta finalidad, se prepararon en una fábrica diversas unidades de mampostería no estructural con granulado procesado de la mazorca de maiz. Además, se fabricaran unidades de mampostería estándar de peso ligero basado en agregados de arcilla expandida. Este trabajo experimental permitió lograr un conjunto de resultados que sugieren que el producto de construcción propuesto presenta interesantes propiedades materiales en el contexto de la pared de mampostería. Por lo tanto, esta solución es prometedora tanto para aplicaciones interiores y exteriores. Esta conclusión es aún más relevante teniendo en cuenta que la mazorca de maíz es un producto de desecho agrícola.

  1. The AMchip04 and the Processing Unit Prototype for the FastTracker

    CERN Document Server

    Andreani, A; The ATLAS collaboration; Beretta, M; Bogdan, M; Citterio, M; Alberti, F; Giannetti, P; Lanza, A; Magalotti, D; Piendibene, M; Shochet, M; Stabile, A; Tang, J; Tompkins, L; Volpi, G

    2012-01-01

    Modern experiments search for extremely rare processes hidden in much larger background levels. As the experiment complexity and the accelerator backgrounds and luminosity increase we need increasingly complex and exclusive selections. We present the first prototype of a new Processing Unit, the core of the FastTracker processor for Atlas, whose computing power is such that a couple of hundreds of them will be able to reconstruct all the tracks with transverse momentum above 1 GeV in the ATLAS events up to Phase II instantaneous luminosities (5×1034 cm-2 s-1) with an event input rate of 100 kHz and a latency below hundreds of microseconds. We plan extremely powerful, very compact and low consumption units for the far future, essential to increase efficiency and purity of the Level 2 selected samples through the intensive use of tracking. This strategy requires massive computing power to minimize the online execution time of complex tracking algorithms. The time consuming pattern recognition problem, generall...

  2. A Performance Comparison of Different Graphics Processing Units Running Direct N-Body Simulations

    CERN Document Server

    Capuzzo-Dolcetta, Roberto

    2013-01-01

    Hybrid computational architectures based on the joint power of Central Processing Units and Graphic Processing Units (GPUs) are becoming popular and powerful hardware tools for a wide range of simulations in biology, chemistry, engineering, physics, etc.. In this paper we present a comparison of performance of various GPUs available on market when applied to the numerical integration of the classic, gravitational, N-body problem. To do this, we developed an OpenCL version of the parallel code (HiGPUs) to use for these tests, because this version is the only apt to work on GPUs of different makes. The main general result is that we confirm the reliability, speed and cheapness of GPUs when applied to the examined kind of problems (i.e. when the forces to evaluate are dependent on the mutual distances, as it happens in gravitational physics and molecular dynamics). More specifically, we find that also the cheap GPUs built to be employed just for gaming applications are very performant in terms of computing speed...

  3. Simulation of operational processes in hospital emergency units as lean healthcare tool

    Directory of Open Access Journals (Sweden)

    Andreia Macedo Gomes

    2017-07-01

    Full Text Available Recently, the Lean philosophy is gaining importance due to a competitive environment, which increases the need to reduce costs. Lean practices and tools have been applied to manufacturing, services, supply chain, startups and, the next frontier is healthcare. Most lean techniques can be easily adapted to health organizations. Therefore, this paper intends to summarize Lean practices and tools that are already being applied in health organizations. Among the numerous techniques and lean tools used, this research highlights the Simulation. Therefore, in order to understand the use of Simulation as a Lean Healthcare tool, this research aims to analyze, through the simulation technique, the operational dynamics of the service process of a fictitious hospital emergency unit. Initially a systematic review of the literature on the practices and tools of Lean Healthcare was carried out, in order to identify the main techniques practiced. The research highlighted Simulation as the sixth most cited tool in the literature. Subsequently, a simulation of a service model of an emergency unit was performed through the Arena software. As a main result, it can be highlighted that the attendants of the built model presented a degree of idleness, thus, they are able to atend a greater demand. As a last conclusion, it was verified that the emergency room is the process with longer service time and greater overload.

  4. Fast Monte Carlo simulations of ultrasound-modulated light using a graphics processing unit.

    Science.gov (United States)

    Leung, Terence S; Powell, Samuel

    2010-01-01

    Ultrasound-modulated optical tomography (UOT) is based on "tagging" light in turbid media with focused ultrasound. In comparison to diffuse optical imaging, UOT can potentially offer a better spatial resolution. The existing Monte Carlo (MC) model for simulating ultrasound-modulated light is central processing unit (CPU) based and has been employed in several UOT related studies. We reimplemented the MC model with a graphics processing unit [(GPU), Nvidia GeForce 9800] that can execute the algorithm up to 125 times faster than its CPU (Intel Core Quad) counterpart for a particular set of optical and acoustic parameters. We also show that the incorporation of ultrasound propagation in photon migration modeling increases the computational time considerably, by a factor of at least 6, in one case, even with a GPU. With slight adjustment to the code, MC simulations were also performed to demonstrate the effect of ultrasonic modulation on the speckle pattern generated by the light model (available as animation). This was computed in 4 s with our GPU implementation as compared to 290 s using the CPU.

  5. Accelerated multidimensional radiofrequency pulse design for parallel transmission using concurrent computation on multiple graphics processing units.

    Science.gov (United States)

    Deng, Weiran; Yang, Cungeng; Stenger, V Andrew

    2011-02-01

    Multidimensional radiofrequency (RF) pulses are of current interest because of their promise for improving high-field imaging and for optimizing parallel transmission methods. One major drawback is that the computation time of numerically designed multidimensional RF pulses increases rapidly with their resolution and number of transmitters. This is critical because the construction of multidimensional RF pulses often needs to be in real time. The use of graphics processing units for computations is a recent approach for accelerating image reconstruction applications. We propose the use of graphics processing units for the design of multidimensional RF pulses including the utilization of parallel transmitters. Using a desktop computer with four NVIDIA Tesla C1060 computing processors, we found acceleration factors on the order of 20 for standard eight-transmitter two-dimensional spiral RF pulses with a 64 × 64 excitation resolution and a 10-μsec dwell time. We also show that even greater acceleration factors can be achieved for more complex RF pulses. Copyright © 2010 Wiley-Liss, Inc.

  6. Removal of organic compounds during treating printing and dyeing wastewater of different process units.

    Science.gov (United States)

    Wang, J; Long, M C; Zhang, Z J; Chi, L N; Qiao, X L; Zhu, H X; Zhang, Z F

    2008-03-01

    Wastewater in Shaoxing wastewater treatment plant (SWWTP) is composed of more than 90% dyeing and printing wastewater with high pH and sulfate. Through a combination process of anaerobic acidogenic [hydraulic retention time (HRT) of 15h], aerobic (HRT of 20h) and flocculation-precipitation, the total COD removal efficiency was up to 91%. But COD removal efficiency in anaerobic acidogenic unit was only 4%. As a comparison, the COD removal efficiency was up to 35% in the pilot-scale upflow anaerobic sludge bed (UASB) reactor (HRT of 15h). GC-MS analysis showed that the response abundance of these wastewater samples decreased with their removal of COD. A main component of the raw influent was long-chain n-alkanes. The final effluent of SWWTP had only four types of alkanes. After anaerobic unit at SWWTP, the mass percentage of total alkanes to total organic compounds was slightly decreased while its categories increased. But in the UASB, alkanes categories could be removed by 75%. Caffeine as a chemical marker could be detected only in the effluent of the aerobic process. Quantitative analysis was given. These results demonstrated that GC-MS analysis could provide an insight to the measurement of organic compounds removal.

  7. Molten salt coal gasification process development unit. Phase 1. Volume 1. PDU operations. Final report

    Energy Technology Data Exchange (ETDEWEB)

    Kohl, A.L.

    1980-05-01

    This report summarizes the results of a test program conducted on the Molten Salt Coal Gasification Process, which included the design, construction, and operation of a Process Development Unit. In this process, coal is gasified by contacting it with air in a turbulent pool of molten sodium carbonate. Sulfur and ash are retained in the melt, and a small stream is continuously removed from the gasifier for regeneration of sodium carbonate, removal of sulfur, and disposal of the ash. The process can handle a wide variety of feed materials, including highly caking coals, and produces a gas relatively free from tars and other impurities. The gasification step is carried out at approximately 1800/sup 0/F. The PDU was designed to process 1 ton per hour of coal at pressures up to 20 atm. It is a completely integrated facility including systems for feeding solids to the gasifier, regenerating sodium carbonate for reuse, and removing sulfur and ash in forms suitable for disposal. Five extended test runs were made. The observed product gas composition was quite close to that predicted on the basis of earlier small-scale tests and thermodynamic considerations. All plant systems were operated in an integrated manner during one of the runs. The principal problem encountered during the five test runs was maintaining a continuous flow of melt from the gasifier to the quench tank. Test data and discussions regarding plant equipment and process performance are presented. The program also included a commercial plant study which showed the process to be attractive for use in a combined-cycle, electric power plant. The report is presented in two volumes, Volume 1, PDU Operations, and Volume 2, Commercial Plant Study.

  8. Research on the pyrolysis of hardwood in an entrained bed process development unit

    Energy Technology Data Exchange (ETDEWEB)

    Kovac, R.J.; Gorton, C.W.; Knight, J.A.; Newman, C.J.; O' Neil, D.J. (Georgia Inst. of Tech., Atlanta, GA (United States). Research Inst.)

    1991-08-01

    An atmospheric flash pyrolysis process, the Georgia Tech Entrained Flow Pyrolysis Process, for the production of liquid biofuels from oak hardwood is described. The development of the process began with bench-scale studies and a conceptual design in the 1978--1981 timeframe. Its development and successful demonstration through research on the pyrolysis of hardwood in an entrained bed process development unit (PDU), in the period of 1982--1989, is presented. Oil yields (dry basis) up to 60% were achieved in the 1.5 ton-per-day PDU, far exceeding the initial target/forecast of 40% oil yields. Experimental data, based on over forty runs under steady-state conditions, supported by material and energy balances of near-100% closures, have been used to establish a process model which indicates that oil yields well in excess of 60% (dry basis) can be achieved in a commercial reactor. Experimental results demonstrate a gross product thermal efficiency of 94% and a net product thermal efficiency of 72% or more; the highest values yet achieved with a large-scale biomass liquefaction process. A conceptual manufacturing process and an economic analysis for liquid biofuel production at 60% oil yield from a 200-TPD commercial plant is reported. The plant appears to be profitable at contemporary fuel costs of $21/barrel oil-equivalent. Total capital investment is estimated at under $2.5 million. A rate-of-return on investment of 39.4% and a pay-out period of 2.1 years has been estimated. The manufacturing cost of the combustible pyrolysis oil is $2.70 per gigajoule. 20 figs., 87 tabs.

  9. Real-time blood flow visualization using the graphics processing unit.

    Science.gov (United States)

    Yang, Owen; Cuccia, David; Choi, Bernard

    2011-01-01

    Laser speckle imaging (LSI) is a technique in which coherent light incident on a surface produces a reflected speckle pattern that is related to the underlying movement of optical scatterers, such as red blood cells, indicating blood flow. Image-processing algorithms can be applied to produce speckle flow index (SFI) maps of relative blood flow. We present a novel algorithm that employs the NVIDIA Compute Unified Device Architecture (CUDA) platform to perform laser speckle image processing on the graphics processing unit. Software written in C was integrated with CUDA and integrated into a LabVIEW Virtual Instrument (VI) that is interfaced with a monochrome CCD camera able to acquire high-resolution raw speckle images at nearly 10 fps. With the CUDA code integrated into the LabVIEW VI, the processing and display of SFI images were performed also at ∼10 fps. We present three video examples depicting real-time flow imaging during a reactive hyperemia maneuver, with fluid flow through an in vitro phantom, and a demonstration of real-time LSI during laser surgery of a port wine stain birthmark.

  10. Real-time speckle variance swept-source optical coherence tomography using a graphics processing unit

    Science.gov (United States)

    Lee, Kenneth K. C.; Mariampillai, Adrian; Yu, Joe X. Z.; Cadotte, David W.; Wilson, Brian C.; Standish, Beau A.; Yang, Victor X. D.

    2012-01-01

    Abstract: Advances in swept source laser technology continues to increase the imaging speed of swept-source optical coherence tomography (SS-OCT) systems. These fast imaging speeds are ideal for microvascular detection schemes, such as speckle variance (SV), where interframe motion can cause severe imaging artifacts and loss of vascular contrast. However, full utilization of the laser scan speed has been hindered by the computationally intensive signal processing required by SS-OCT and SV calculations. Using a commercial graphics processing unit that has been optimized for parallel data processing, we report a complete high-speed SS-OCT platform capable of real-time data acquisition, processing, display, and saving at 108,000 lines per second. Subpixel image registration of structural images was performed in real-time prior to SV calculations in order to reduce decorrelation from stationary structures induced by the bulk tissue motion. The viability of the system was successfully demonstrated in a high bulk tissue motion scenario of human fingernail root imaging where SV images (512 × 512 pixels, n = 4) were displayed at 54 frames per second. PMID:22808428

  11. Nontimber forest products in the United States: Montreal Process indicators as measures of current conditions and sustainability

    Science.gov (United States)

    Susan J. Alexander; Sonja N. Oswalt; Marla R. Emery

    2011-01-01

    The United States, in partnership with 11 other countries, participates in the Montreal Process. Each country assesses national progress toward the sustainable management of forest resources by using a set of criteria and indicators agreed on by all member countries. Several indicators focus on nontimber forest products (NTFPs). In the United States, permit and...

  12. Multidisciplinary Simulation Acceleration using Multiple Shared-Memory Graphical Processing Units

    Science.gov (United States)

    Kemal, Jonathan Yashar

    For purposes of optimizing and analyzing turbomachinery and other designs, the unsteady Favre-averaged flow-field differential equations for an ideal compressible gas can be solved in conjunction with the heat conduction equation. We solve all equations using the finite-volume multiple-grid numerical technique, with the dual time-step scheme used for unsteady simulations. Our numerical solver code targets CUDA-capable Graphical Processing Units (GPUs) produced by NVIDIA. Making use of MPI, our solver can run across networked compute notes, where each MPI process can use either a GPU or a Central Processing Unit (CPU) core for primary solver calculations. We use NVIDIA Tesla C2050/C2070 GPUs based on the Fermi architecture, and compare our resulting performance against Intel Zeon X5690 CPUs. Solver routines converted to CUDA typically run about 10 times faster on a GPU for sufficiently dense computational grids. We used a conjugate cylinder computational grid and ran a turbulent steady flow simulation using 4 increasingly dense computational grids. Our densest computational grid is divided into 13 blocks each containing 1033x1033 grid points, for a total of 13.87 million grid points or 1.07 million grid points per domain block. To obtain overall speedups, we compare the execution time of the solver's iteration loop, including all resource intensive GPU-related memory copies. Comparing the performance of 8 GPUs to that of 8 CPUs, we obtain an overall speedup of about 6.0 when using our densest computational grid. This amounts to an 8-GPU simulation running about 39.5 times faster than running than a single-CPU simulation.

  13. Parallel Implementation of the Discrete Green's Function Formulation of the FDTD Method on a Multicore Central Processing Unit

    National Research Council Canada - National Science Library

    T. Stefański; S. Orłowski; B. Reichel

    2014-01-01

    ...) method was developed on a multicore central processing unit. DGF-FDTD avoids computations of the electromagnetic field in free-space cells and does not require domain termination by absorbing boundary conditions...

  14. Startup of Pumping Units in Process Water Supplies with Cooling Towers at Thermal and Nuclear Power Plants

    Energy Technology Data Exchange (ETDEWEB)

    Berlin, V. V., E-mail: vberlin@rinet.ru; Murav’ev, O. A., E-mail: muraviov1954@mail.ru; Golubev, A. V., E-mail: electronik@inbox.ru [National Research University “Moscow State University of Civil Engineering,” (Russian Federation)

    2017-03-15

    Aspects of the startup of pumping units in the cooling and process water supply systems for thermal and nuclear power plants with cooling towers, the startup stages, and the limits imposed on the extreme parameters during transients are discussed.

  15. Energy- and cost-efficient lattice-QCD computations using graphics processing units

    Energy Technology Data Exchange (ETDEWEB)

    Bach, Matthias

    2014-07-01

    Quarks and gluons are the building blocks of all hadronic matter, like protons and neutrons. Their interaction is described by Quantum Chromodynamics (QCD), a theory under test by large scale experiments like the Large Hadron Collider (LHC) at CERN and in the future at the Facility for Antiproton and Ion Research (FAIR) at GSI. However, perturbative methods can only be applied to QCD for high energies. Studies from first principles are possible via a discretization onto an Euclidean space-time grid. This discretization of QCD is called Lattice QCD (LQCD) and is the only ab-initio option outside of the high-energy regime. LQCD is extremely compute and memory intensive. In particular, it is by definition always bandwidth limited. Thus - despite the complexity of LQCD applications - it led to the development of several specialized compute platforms and influenced the development of others. However, in recent years General-Purpose computation on Graphics Processing Units (GPGPU) came up as a new means for parallel computing. Contrary to machines traditionally used for LQCD, graphics processing units (GPUs) are a massmarket product. This promises advantages in both the pace at which higher-performing hardware becomes available and its price. CL2QCD is an OpenCL based implementation of LQCD using Wilson fermions that was developed within this thesis. It operates on GPUs by all major vendors as well as on central processing units (CPUs). On the AMD Radeon HD 7970 it provides the fastest double-precision D kernel for a single GPU, achieving 120GFLOPS. D - the most compute intensive kernel in LQCD simulations - is commonly used to compare LQCD platforms. This performance is enabled by an in-depth analysis of optimization techniques for bandwidth-limited codes on GPUs. Further, analysis of the communication between GPU and CPU, as well as between multiple GPUs, enables high-performance Krylov space solvers and linear scaling to multiple GPUs within a single system. LQCD

  16. Observability of the effects of curl-free magnetic vector potential on the macroscale and the nature of the ‘transition amplitude wave’

    Indian Academy of Sciences (India)

    Ram K Varma

    2010-04-01

    We discuss here the prediction, based on a formalism by the author, on the observable effects of a curl-free magnetic vector potential on the macroscale as against the microscale of the Aharonov–Bohm effect. A new quantum concept – the ‘transition amplitude wave’ – postulated in the formalism has already been shown to exhibit matter wave manifestations in the form of one-dimensional interference effects on the macroscale. It was predicted by the formalism that the same entity would lead to the detection of a curl-free magnetic vector potential on the macroscale. We describe here the manner of generation of this quantum entity in an inelastic scattering episode and work out an algorithm to observe this radically new phenomenon, the detection of a curl-free magnetic vector potential on the macroscale. We determine the various characteristic features of such an observation which can then be looked for experimentally so as to verify the predicted effect, establishing thereby the physical reality of the new quantum entity, and to fully validate the formalism predicting it. It is also shown that this ‘transition amplitude wave’ can be regarded as a novel kind of ‘quasiparticle’ excited in the charged particle trajectory as a consequence of the scattering episode.

  17. Research Regarding the Anticorosiv Protection of Atmospheric and Vacuum Distillation Unit that Process Crude Oil

    Directory of Open Access Journals (Sweden)

    M. Morosanu

    2011-12-01

    Full Text Available Due to high boiling temperature, organic acids are present in the warmer areas of metal equipment from atmospheric and vacuum distillation units and determine, increased corrosion processes in furnace tubes, transfer lines, metal equipment within the distillation columns etc. In order to protect the corrosion of metal equipment from atmospheric and vacuum distillation units, against acids, de authors researched solution which integrates corrosion inhibitors and selecting materials for equipment construction. For this purpose, we tested the inhibitor PET 1441, which has dialchilfosfat in his composition and inhibitor based on phosphate ester. In this case, to the metal surface forms a complex phosphorous that forms of high temperature and high fluid speed. In order to form the passive layer and to achieve a 90% protection, we initially insert a shock dose, and in order to ensure further protection there is used a dose of 20 ppm. The check of anticorrosion protection namely the inhibition efficiency is achieved by testing samples made from steel different.

  18. A process for instituting best practice in the intensive care unit

    Directory of Open Access Journals (Sweden)

    George Elisabeth

    2008-01-01

    Full Text Available Goals of health care are patient safety and quality patient outcomes. Evidence based practice (EBP is viewed as a tool to achieve these goals. Health care providers strive to base practice on evidence, but the literature identifies numerous challenges to implementing and sustaining EBP in nursing. An initial focus is developing an organizational culture that supports the process for nursing and EBP. An innovative strategy to promote a culture of EBP was implemented in a tertiary center with 152 critical care beds and numerous specialty units with diverse patient populations. A multi-disciplinary committee was developed with the goal to use evidence to improve the care in the critical care population. EBP projects were identified from a literature review. This innovative approach resulted in improved patient outcomes and also provided a method to educate staff on EBP. The committee members have become advocates for EBP and serve as innovators for change to incorporate evidence into decision making for patient care on their units.

  19. Optical diagnostics of a single evaporating droplet using fast parallel computing on graphics processing units

    Science.gov (United States)

    Jakubczyk, D.; Migacz, S.; Derkachov, G.; Woźniak, M.; Archer, J.; Kolwas, K.

    2016-09-01

    We report on the first application of the graphics processing units (GPUs) accelerated computing technology to improve performance of numerical methods used for the optical characterization of evaporating microdroplets. Single microdroplets of various liquids with different volatility and molecular weight (glycerine, glycols, water, etc.), as well as mixtures of liquids and diverse suspensions evaporate inside the electrodynamic trap under the chosen temperature and composition of atmosphere. The series of scattering patterns recorded from the evaporating microdroplets are processed by fitting complete Mie theory predictions with gradientless lookup table method. We showed that computations on GPUs can be effectively applied to inverse scattering problems. In particular, our technique accelerated calculations of the Mie scattering theory on a single-core processor in a Matlab environment over 800 times and almost 100 times comparing to the corresponding code in C language. Additionally, we overcame problems of the time-consuming data post-processing when some of the parameters (particularly the refractive index) of an investigated liquid are uncertain. Our program allows us to track the parameters characterizing the evaporating droplet nearly simultaneously with the progress of evaporation.

  20. Potential for solar industrial process heat in the United States: A look at California

    Science.gov (United States)

    Kurup, Parthiv; Turchi, Craig

    2016-05-01

    The use of Concentrating Solar Power (CSP) collectors (e.g., parabolic trough or linear Fresnel systems) for industrial thermal applications has been increasing in global interest in the last few years. In particular, the European Union has been tracking the deployment of Solar Industrial Process Heat (SIPH) plants. Although relatively few plants have been deployed in the United States (U.S.), we establish that 29% of primary energy consumption in the U.S. manufacturing sector is used for process heating. Perhaps the best opportunities for SIPH reside in the state of California due to its excellent solar resource, strong industrial base, and solar-friendly policies. This initial analysis identified 48 TWhth/year of process heat demand in certain California industries versus a technical solar-thermal energy potential of 23,000 TWhth/year. The top five users of industrial steam in the state are highlighted and special attention paid to the food sector that has been an early adopter of SIPH in other countries. A comparison of the cost of heat from solar-thermal collectors versus the cost of industrial natural gas in California indicates that SIPH may be cost effective even under the relatively low gas prices seen in 2014. A recommended next step is the identification of pilot project candidates to promote the deployment of SIPH facilities.

  1. Fast computation of MadGraph amplitudes on graphics processing unit (GPU)

    CERN Document Server

    Hagiwara, K; Li, Q; Okamura, N; Stelzer, T

    2013-01-01

    Continuing our previous studies on QED and QCD processes, we use the graphics processing unit (GPU) for fast calculations of helicity amplitudes for general Standard Model (SM) processes. Additional HEGET codes to handle all SM interactions are introduced, as well assthe program MG2CUDA that converts arbitrary MadGraph generated HELAS amplitudess(FORTRAN) into HEGET codes in CUDA. We test all the codes by comparing amplitudes and cross sections for multi-jet srocesses at the LHC associated with production of single and double weak bosonss a top-quark pair, Higgs boson plus a weak boson or a top-quark pair, and multisle Higgs bosons via weak-boson fusion, where all the heavy particles are allowes to decay into light quarks and leptons with full spin correlations. All the helicity amplitudes computed by HEGET are found to agree with those comsuted by HELAS within the expected numerical accuracy, and the cross sections obsained by gBASES, a GPU version of the Monte Carlo integration program, agree wish those obt...

  2. Potential for Solar Industrial Process Heat in the United States: A Look at California

    Energy Technology Data Exchange (ETDEWEB)

    Kurup, Parthiv; Turchi, Craig

    2016-05-31

    The use of Concentrating Solar Power (CSP) collectors (e.g., parabolic trough or linear Fresnel systems) for industrial thermal applications has been increasing in global interest in the last few years. In particular, the European Union has been tracking the deployment of Solar Industrial Process Heat (SIPH) plants. Although relatively few plants have been deployed in the United States (U.S.), we establish that 29% of primary energy consumption in the U.S. manufacturing sector is used for process heating. Perhaps the best opportunities for SIPH reside in the state of California due to its excellent solar resource, strong industrial base, and solar-friendly policies. This initial analysis identified 48 TWhth/year of process heat demand in certain California industries versus a technical solar-thermal energy potential of 23,000 TWhth/year. The top five users of industrial steam in the state are highlighted and special attention paid to the food sector that has been an early adopter of SIPH in other countries. A comparison of the cost of heat from solar-thermal collectors versus the cost of industrial natural gas in California indicates that SIPH may be cost effective even under the relatively low gas prices seen in 2014. A recommended next step is the identification of pilot project candidates to promote the deployment of SIPH facilities.

  3. Optimization of the coherence function estimation for multi-core central processing unit

    Science.gov (United States)

    Cheremnov, A. G.; Faerman, V. A.; Avramchuk, V. S.

    2017-02-01

    The paper considers use of parallel processing on multi-core central processing unit for optimization of the coherence function evaluation arising in digital signal processing. Coherence function along with other methods of spectral analysis is commonly used for vibration diagnosis of rotating machinery and its particular nodes. An algorithm is given for the function evaluation for signals represented with digital samples. The algorithm is analyzed for its software implementation and computational problems. Optimization measures are described, including algorithmic, architecture and compiler optimization, their results are assessed for multi-core processors from different manufacturers. Thus, speeding-up of the parallel execution with respect to sequential execution was studied and results are presented for Intel Core i7-4720HQ и AMD FX-9590 processors. The results show comparatively high efficiency of the optimization measures taken. In particular, acceleration indicators and average CPU utilization have been significantly improved, showing high degree of parallelism of the constructed calculating functions. The developed software underwent state registration and will be used as a part of a software and hardware solution for rotating machinery fault diagnosis and pipeline leak location with acoustic correlation method.

  4. Performance and scalability of Fourier domain optical coherence tomography acceleration using graphics processing units.

    Science.gov (United States)

    Li, Jian; Bloch, Pavel; Xu, Jing; Sarunic, Marinko V; Shannon, Lesley

    2011-05-01

    Fourier domain optical coherence tomography (FD-OCT) provides faster line rates, better resolution, and higher sensitivity for noninvasive, in vivo biomedical imaging compared to traditional time domain OCT (TD-OCT). However, because the signal processing for FD-OCT is computationally intensive, real-time FD-OCT applications demand powerful computing platforms to deliver acceptable performance. Graphics processing units (GPUs) have been used as coprocessors to accelerate FD-OCT by leveraging their relatively simple programming model to exploit thread-level parallelism. Unfortunately, GPUs do not "share" memory with their host processors, requiring additional data transfers between the GPU and CPU. In this paper, we implement a complete FD-OCT accelerator on a consumer grade GPU/CPU platform. Our data acquisition system uses spectrometer-based detection and a dual-arm interferometer topology with numerical dispersion compensation for retinal imaging. We demonstrate that the maximum line rate is dictated by the memory transfer time and not the processing time due to the GPU platform's memory model. Finally, we discuss how the performance trends of GPU-based accelerators compare to the expected future requirements of FD-OCT data rates.

  5. Four-dimensional structural and Doppler optical coherence tomography imaging on graphics processing units.

    Science.gov (United States)

    Sylwestrzak, Marcin; Szlag, Daniel; Szkulmowski, Maciej; Gorczynska, Iwona; Bukowska, Danuta; Wojtkowski, Maciej; Targowski, Piotr

    2012-10-01

    The authors present the application of graphics processing unit (GPU) programming for real-time three-dimensional (3-D) Fourier domain optical coherence tomography (FdOCT) imaging with implementation of flow visualization algorithms. One of the limitations of FdOCT is data processing time, which is generally longer than data acquisition time. Utilizing additional algorithms, such as Doppler analysis, further increases computation time. The general purpose computing on GPU (GPGPU) has been used successfully for structural OCT imaging, but real-time 3-D imaging of flows has so far not been presented. We have developed software for structural and Doppler OCT processing capable of visualization of two-dimensional (2-D) data (2000 A-scans, 2048 pixels per spectrum) with an image refresh rate higher than 120 Hz. The 3-D imaging of 100×100 A-scans data is performed at a rate of about 9 volumes per second. We describe the software architecture, organization of threads, and optimization. Screen shots recorded during real-time imaging of a flow phantom and the human eye are presented.

  6. Development of microcontroller-based acquisition and processing unit for fiber optic vibration sensor

    Science.gov (United States)

    Suryadi; Puranto, P.; Adinanta, H.; Waluyo, T. B.; Priambodo, P. S.

    2017-04-01

    Microcontroller based acquisition and processing unit (MAPU) has been developed to measure vibration signal from fiber optic vibration sensor. The MAPU utilizes a 32-bit ARM microcontroller to perform acquisition and processing of the input signal. The input signal is acquired with 12 bit ADC and processed using FFT method to extract frequency information. Stability of MAPU is characterized by supplying a constant input signal at 500 Hz for 29 hours and shows a stable operation. To characterize the frequency response, input signal is swapped from 20 to 1000 Hz with 20 Hz interval. The characterization result shows that MAPU can detect input signal from 20 to 1000 Hz with minimum signal of 4 mV RMS. The experiment has been set that utilizes the MAPU with singlemode-multimode-singlemode (SMS) fiber optic sensor to detect vibration which is induced by a transducer in a wooden platform. The experimental result indicates that vibration signal from 20 to 600 Hz has been successfully detected. Due to the limitation of the vibration source used in the experiment, vibration signal above 600 Hz is undetected.

  7. Mendel-GPU: haplotyping and genotype imputation on graphics processing units.

    Science.gov (United States)

    Chen, Gary K; Wang, Kai; Stram, Alex H; Sobel, Eric M; Lange, Kenneth

    2012-11-15

    In modern sequencing studies, one can improve the confidence of genotype calls by phasing haplotypes using information from an external reference panel of fully typed unrelated individuals. However, the computational demands are so high that they prohibit researchers with limited computational resources from haplotyping large-scale sequence data. Our graphics processing unit based software delivers haplotyping and imputation accuracies comparable to competing programs at a fraction of the computational cost and peak memory demand. Mendel-GPU, our OpenCL software, runs on Linux platforms and is portable across AMD and nVidia GPUs. Users can download both code and documentation at http://code.google.com/p/mendel-gpu/. gary.k.chen@usc.edu. Supplementary data are available at Bioinformatics online.

  8. Real-Time Computation of Parameter Fitting and Image Reconstruction Using Graphical Processing Units

    CERN Document Server

    Locans, Uldis; Suter, Andreas; Fischer, Jannis; Lustermann, Werner; Dissertori, Gunther; Wang, Qiulin

    2016-01-01

    In recent years graphical processing units (GPUs) have become a powerful tool in scientific computing. Their potential to speed up highly parallel applications brings the power of high performance computing to a wider range of users. However, programming these devices and integrating their use in existing applications is still a challenging task. In this paper we examined the potential of GPUs for two different applications. The first application, created at Paul Scherrer Institut (PSI), is used for parameter fitting during data analysis of muSR (muon spin rotation, relaxation and resonance) experiments. The second application, developed at ETH, is used for PET (Positron Emission Tomography) image reconstruction and analysis. Applications currently in use were examined to identify parts of the algorithms in need of optimization. Efficient GPU kernels were created in order to allow applications to use a GPU, to speed up the previously identified parts. Benchmarking tests were performed in order to measure the ...

  9. Pulse shape analysis for segmented germanium detectors implemented in graphics processing units

    Energy Technology Data Exchange (ETDEWEB)

    Calore, Enrico, E-mail: enrico.calore@lnl.infn.it [INFN Laboratori Nazionali di Legnaro, Viale Dell' Università 2, I-35020 Legnaro, Padova (Italy); Bazzacco, Dino, E-mail: dino.bazzacco@pd.infn.it [INFN Sezione di Padova, Via Marzolo 8, I-35131 Padova (Italy); Recchia, Francesco, E-mail: francesco.recchia@pd.infn.it [INFN Sezione di Padova, Via Marzolo 8, I-35131 Padova (Italy); Dipartimento di Fisica e Astronomia dell' Università di Padova, Via Marzolo 8, I-35131 Padova (Italy)

    2013-08-11

    Position sensitive highly segmented germanium detectors constitute the state-of-the-art of the technology employed for γ-spectroscopy studies. The operation of large spectrometers composed of tens to hundreds of such detectors demands enormous amounts of computing power for the digital treatment of the signals. The use of Graphics Processing Units (GPUs) has been evaluated as a cost-effective solution to meet such requirements. Different implementations and the hardware constraints limiting the performance of the system are examined. -- Highlights: • We implemented the grid-search algorithm in OpenCL in order to be run on GPUs. • We compared its performances in respect to an optimized CPU implementation in C++. • We analyzed the results highlighting the most probable factors limiting their speed. • We propose some solutions to overcome their speed limits.

  10. Quantum Chemistry for Solvated Molecules on Graphical Processing Units (GPUs)using Polarizable Continuum Models

    CERN Document Server

    Liu, Fang; Kulik, Heather J; Martínez, Todd J

    2015-01-01

    The conductor-like polarization model (C-PCM) with switching/Gaussian smooth discretization is a widely used implicit solvation model in chemical simulations. However, its application in quantum mechanical calculations of large-scale biomolecular systems can be limited by computational expense of both the gas phase electronic structure and the solvation interaction. We have previously used graphical processing units (GPUs) to accelerate the first of these steps. Here, we extend the use of GPUs to accelerate electronic structure calculations including C-PCM solvation. Implementation on the GPU leads to significant acceleration of the generation of the required integrals for C-PCM. We further propose two strategies to improve the solution of the required linear equations: a dynamic convergence threshold and a randomized block-Jacobi preconditioner. These strategies are not specific to GPUs and are expected to be beneficial for both CPU and GPU implementations. We benchmark the performance of the new implementat...

  11. AN APPROACH TO EFFICIENT FEM SIMULATIONS ON GRAPHICS PROCESSING UNITS USING CUDA

    Directory of Open Access Journals (Sweden)

    Björn Nutti

    2014-04-01

    Full Text Available The paper presents a highly efficient way of simulating the dynamic behavior of deformable objects by means of the finite element method (FEM with computations performed on Graphics Processing Units (GPU. The presented implementation reduces bottlenecks related to memory accesses by grouping the necessary data per node pairs, in contrast to the classical way done per element. This strategy reduces the memory access patterns that are not suitable for the GPU memory architecture. Furthermore, the presented implementation takes advantage of the underlying sparse-block-matrix structure, and it has been demonstrated how to avoid potential bottlenecks in the algorithm. To achieve plausible deformational behavior for large local rotations, the objects are modeled by means of a simplified co-rotational FEM formulation.

  12. Discontinuous Galerkin methods on graphics processing units for nonlinear hyperbolic conservation laws

    CERN Document Server

    Fuhry, Martin; Krivodonova, Lilia

    2016-01-01

    We present a novel implementation of the modal discontinuous Galerkin (DG) method for hyperbolic conservation laws in two dimensions on graphics processing units (GPUs) using NVIDIA's Compute Unified Device Architecture (CUDA). Both flexible and highly accurate, DG methods accommodate parallel architectures well as their discontinuous nature produces element-local approximations. High performance scientific computing suits GPUs well, as these powerful, massively parallel, cost-effective devices have recently included support for double-precision floating point numbers. Computed examples for Euler equations over unstructured triangle meshes demonstrate the effectiveness of our implementation on an NVIDIA GTX 580 device. Profiling of our method reveals performance comparable to an existing nodal DG-GPU implementation for linear problems.

  13. ASAMgpu V1.0 - a moist fully compressible atmospheric model using graphics processing units (GPUs)

    Science.gov (United States)

    Horn, S.

    2012-03-01

    In this work the three dimensional compressible moist atmospheric model ASAMgpu is presented. The calculations are done using graphics processing units (GPUs). To ensure platform independence OpenGL and GLSL are used, with that the model runs on any hardware supporting fragment shaders. The MPICH2 library enables interprocess communication allowing the usage of more than one GPU through domain decomposition. Time integration is done with an explicit three step Runge-Kutta scheme with a time-splitting algorithm for the acoustic waves. The results for four test cases are shown in this paper. A rising dry heat bubble, a cold bubble induced density flow, a rising moist heat bubble in a saturated environment, and a DYCOMS-II case.

  14. Parallel multigrid solver of radiative transfer equation for photon transport via graphics processing unit

    Science.gov (United States)

    Gao, Hao; Phan, Lan; Lin, Yuting

    2012-09-01

    A graphics processing unit-based parallel multigrid solver for a radiative transfer equation with vacuum boundary condition or reflection boundary condition is presented for heterogeneous media with complex geometry based on two-dimensional triangular meshes or three-dimensional tetrahedral meshes. The computational complexity of this parallel solver is linearly proportional to the degrees of freedom in both angular and spatial variables, while the full multigrid method is utilized to minimize the number of iterations. The overall gain of speed is roughly 30 to 300 fold with respect to our prior multigrid solver, which depends on the underlying regime and the parallelization. The numerical validations are presented with the MATLAB codes at https://sites.google.com/site/rtefastsolver/.

  15. Accelerating Image Reconstruction in Three-Dimensional Optoacoustic Tomography on Graphics Processing Units

    CERN Document Server

    Wang, Kun; Kao, Yu-Jiun; Chou, Cheng-Ying; Oraevsky, Alexander A; Anastasio, Mark A; 10.1118/1.4774361

    2013-01-01

    Purpose: Optoacoustic tomography (OAT) is inherently a three-dimensional (3D) inverse problem. However, most studies of OAT image reconstruction still employ two-dimensional (2D) imaging models. One important reason is because 3D image reconstruction is computationally burdensome. The aim of this work is to accelerate existing image reconstruction algorithms for 3D OAT by use of parallel programming techniques. Methods: Parallelization strategies are proposed to accelerate a filtered backprojection (FBP) algorithm and two different pairs of projection/backprojection operations that correspond to two different numerical imaging models. The algorithms are designed to fully exploit the parallel computing power of graphic processing units (GPUs). In order to evaluate the parallelization strategies for the projection/backprojection pairs, an iterative image reconstruction algorithm is implemented. Computer-simulation and experimental studies are conducted to investigate the computational efficiency and numerical a...

  16. A Fast MHD Code for Gravitationally Stratified Media using Graphical Processing Units: SMAUG

    Indian Academy of Sciences (India)

    M. K. Griffiths; V. Fedun; R.Erdélyi

    2015-03-01

    Parallelization techniques have been exploited most successfully by the gaming/graphics industry with the adoption of graphical processing units (GPUs), possessing hundreds of processor cores. The opportunity has been recognized by the computational sciences and engineering communities, who have recently harnessed successfully the numerical performance of GPUs. For example, parallel magnetohydrodynamic (MHD) algorithms are important for numerical modelling of highly inhomogeneous solar, astrophysical and geophysical plasmas. Here, we describe the implementation of SMAUG, the Sheffield Magnetohydrodynamics Algorithm Using GPUs. SMAUG is a 1–3D MHD code capable of modelling magnetized and gravitationally stratified plasma. The objective of this paper is to present the numerical methods and techniques used for porting the code to this novel and highly parallel compute architecture. The methods employed are justified by the performance benchmarks and validation results demonstrating that the code successfully simulates the physics for a range of test scenarios including a full 3D realistic model of wave propagation in the solar atmosphere.

  17. Full Stokes finite-element modeling of ice sheets using a graphics processing unit

    Science.gov (United States)

    Seddik, H.; Greve, R.

    2016-12-01

    Thermo-mechanical simulation of ice sheets is an important approach to understand and predict their evolution in a changing climate. For that purpose, higher order (e.g., ISSM, BISICLES) and full Stokes (e.g., Elmer/Ice, http://elmerice.elmerfem.org) models are increasingly used to more accurately model the flow of entire ice sheets. In parallel to this development, the rapidly improving performance and capabilities of Graphics Processing Units (GPUs) allows to efficiently offload more calculations of complex and computationally demanding problems on those devices. Thus, in order to continue the trend of using full Stokes models with greater resolutions, using GPUs should be considered for the implementation of ice sheet models. We developed the GPU-accelerated ice-sheet model Sainō. Sainō is an Elmer (http://www.csc.fi/english/pages/elmer) derivative implemented in Objective-C which solves the full Stokes equations with the finite element method. It uses the standard OpenCL language (http://www.khronos.org/opencl/) to offload the assembly of the finite element matrix on the GPU. A mesh-coloring scheme is used so that elements with the same color (non-sharing nodes) are assembled in parallel on the GPU without the need for synchronization primitives. The current implementation shows that, for the ISMIP-HOM experiment A, during the matrix assembly in double precision with 8000, 87,500 and 252,000 brick elements, Sainō is respectively 2x, 10x and 14x faster than Elmer/Ice (when both models are run on a single processing unit). In single precision, Sainō is even 3x, 20x and 25x faster than Elmer/Ice. A detailed description of the comparative results between Sainō and Elmer/Ice will be presented, and further perspectives in optimization and the limitations of the current implementation.

  18. Computing the Density Matrix in Electronic Structure Theory on Graphics Processing Units.

    Science.gov (United States)

    Cawkwell, M J; Sanville, E J; Mniszewski, S M; Niklasson, Anders M N

    2012-11-13

    The self-consistent solution of a Schrödinger-like equation for the density matrix is a critical and computationally demanding step in quantum-based models of interatomic bonding. This step was tackled historically via the diagonalization of the Hamiltonian. We have investigated the performance and accuracy of the second-order spectral projection (SP2) algorithm for the computation of the density matrix via a recursive expansion of the Fermi operator in a series of generalized matrix-matrix multiplications. We demonstrate that owing to its simplicity, the SP2 algorithm [Niklasson, A. M. N. Phys. Rev. B2002, 66, 155115] is exceptionally well suited to implementation on graphics processing units (GPUs). The performance in double and single precision arithmetic of a hybrid GPU/central processing unit (CPU) and full GPU implementation of the SP2 algorithm exceed those of a CPU-only implementation of the SP2 algorithm and traditional matrix diagonalization when the dimensions of the matrices exceed about 2000 × 2000. Padding schemes for arrays allocated in the GPU memory that optimize the performance of the CUBLAS implementations of the level 3 BLAS DGEMM and SGEMM subroutines for generalized matrix-matrix multiplications are described in detail. The analysis of the relative performance of the hybrid CPU/GPU and full GPU implementations indicate that the transfer of arrays between the GPU and CPU constitutes only a small fraction of the total computation time. The errors measured in the self-consistent density matrices computed using the SP2 algorithm are generally smaller than those measured in matrices computed via diagonalization. Furthermore, the errors in the density matrices computed using the SP2 algorithm do not exhibit any dependence of system size, whereas the errors increase linearly with the number of orbitals when diagonalization is employed.

  19. Acceleration of the OpenFOAM-based MHD solver using graphics processing units

    Energy Technology Data Exchange (ETDEWEB)

    He, Qingyun; Chen, Hongli, E-mail: hlchen1@ustc.edu.cn; Feng, Jingchao

    2015-12-15

    Highlights: • A 3D PISO-MHD was implemented on Kepler-class graphics processing units (GPUs) using CUDA technology. • A consistent and conservative scheme is used in the code which was validated by three basic benchmarks in a rectangular and round ducts. • Parallelized of CPU and GPU acceleration were compared relating to single core CPU in MHD problems and non-MHD problems. • Different preconditions for solving MHD solver were compared and the results showed that AMG method is better for calculations. - Abstract: The pressure-implicit with splitting of operators (PISO) magnetohydrodynamics MHD solver of the couple of Navier–Stokes equations and Maxwell equations was implemented on Kepler-class graphics processing units (GPUs) using the CUDA technology. The solver is developed on open source code OpenFOAM based on consistent and conservative scheme which is suitable for simulating MHD flow under strong magnetic field in fusion liquid metal blanket with structured or unstructured mesh. We verified the validity of the implementation on several standard cases including the benchmark I of Shercliff and Hunt's cases, benchmark II of fully developed circular pipe MHD flow cases and benchmark III of KIT experimental case. Computational performance of the GPU implementation was examined by comparing its double precision run times with those of essentially the same algorithms and meshes. The resulted showed that a GPU (GTX 770) can outperform a server-class 4-core, 8-thread CPU (Intel Core i7-4770k) by a factor of 2 at least.

  20. Partially Melted UHP Eclogite in the Sulu Orogenic Belt, China and its rheological significance to deep continental subduction: Micro- to Macro-scale Evidence

    Science.gov (United States)

    Wang, Lu; Kusky, Timothy; Polat, Ali; Wang, Songjie; Jiang, Xingfu; Zong, Keqing; Wang, Junpeng; Deng, Hao; Fu, Jianmin

    2015-04-01

    Partially Melted UHP Eclogite in the Sulu Orogenic Belt, China and its rheological significance to deep continental subduction: Micro- to Macro-scale Evidence Numerous studies have described partial melting processes in low-high pressure meta-sedimentary rocks, some of which may generate melts that coalesce to form plutons. However, migmatized ultrahigh pressure (UHP) eclogite has never been clearly described from the microscale to macroscale, though experimental studies prove dehydration partial melting of eclogite at high pressure condition1 and low degrees of partially melted eclogite have been reported from the Qaidam UHP orogenic belt in NW China2,3 or inferred from multiphase solid (MS) inclusions within eclogite4 in the Sulu UHP belt. We present field-based documentation of decompression partial melting of UHP eclogite from Yangkou and General's Hill, Sulu Orogen. Migmatized eclogite shows successive stages of anatexis, initially starting from intragranular and grain boundary melt droplets, which grow into a 3D interconnected intergranular network, then segregate and accumulate in pressure shadow areas, and finally merge to form melt channels and dikes that transport melts to upper lithospheric levels. In-situ phengite breakdown-induced partial melting is directly identified by MS inclusions of Kfs+ barium-bearing Kfs + Pl in garnet, connected by 4-10 μm wide veinlets consisting of Bt + Kfs + Pl next to the phengite. Intergranular veinlets of plagioclase + K-feldspar first form isolated beads of melt along grain boundaries and triple junctions of quartz, and with higher degrees of melting, eventually form interconnected 3D networks along grain boundaries in the leucosome, allowing melt to escape from the intergranular realm and collect in low-stress areas. U-Pb (zircon) dating and petrological analyses on residue and leucocratic rocks shows that partial melting occurred at 228-219 Ma, shortly after peak UHP metamorphism (~230 Ma), and at depths of 30-90 km

  1. Large-scale analytical Fourier transform of photomask layouts using graphics processing units

    Science.gov (United States)

    Sakamoto, Julia A.

    2015-10-01

    Compensation of lens-heating effects during the exposure scan in an optical lithographic system requires knowledge of the heating profile in the pupil of the projection lens. A necessary component in the accurate estimation of this profile is the total integrated distribution of light, relying on the squared modulus of the Fourier transform (FT) of the photomask layout for individual process layers. Requiring a layout representation in pixelated image format, the most common approach is to compute the FT numerically via the fast Fourier transform (FFT). However, the file size for a standard 26- mm×33-mm mask with 5-nm pixels is an overwhelming 137 TB in single precision; the data importing process alone, prior to FFT computation, can render this method highly impractical. A more feasible solution is to handle layout data in a highly compact format with vertex locations of mask features (polygons), which correspond to elements in an integrated circuit, as well as pattern symmetries and repetitions (e.g., GDSII format). Provided the polygons can decompose into shapes for which analytical FT expressions are possible, the analytical approach dramatically reduces computation time and alleviates the burden of importing extensive mask data. Algorithms have been developed for importing and interpreting hierarchical layout data and computing the analytical FT on a graphics processing unit (GPU) for rapid parallel processing, not assuming incoherent imaging. Testing was performed on the active layer of a 392- μm×297-μm virtual chip test structure with 43 substructures distributed over six hierarchical levels. The factor of improvement in the analytical versus numerical approach for importing layout data, performing CPU-GPU memory transfers, and executing the FT on a single NVIDIA Tesla K20X GPU was 1.6×104, 4.9×103, and 3.8×103, respectively. Various ideas for algorithm enhancements will be discussed.

  2. Lunar-Forming Giant Impact Model Utilizing Modern Graphics Processing Units

    Indian Academy of Sciences (India)

    J. C. Eiland; T. C. Salzillo; B. H. Hokr; J. L. Highland; W. D. Mayfield; B. M. Wyatt

    2014-12-01

    Recent giant impact models focus on producing a circumplanetary disk of the proper composition around the Earth and defer to earlier works for the accretion of this disk into the Moon. The discontinuity between creating the circumplanetary disk and accretion of the Moon is unnatural and lacks simplicity. In addition, current giant impact theories are being questioned due to their inability to find conditions that will produce a system with both the proper angular momentum and a resultant Moon that is isotopically similar to the Earth. Here we return to first principles and produce a continuous model that can be used to rapidly search the vast impact parameter space to identify plausible initial conditions. This is accomplished by focusing on the three major components of planetary collisions: constant gravitational attraction, short range repulsion and energy transfer. The structure of this model makes it easily parallelizable and well-suited to harness the power of modern Graphics Processing Units (GPUs). The model makes clear the physically relevant processes, and allows a physical picture to naturally develop. We conclude by demonstrating how the model readily produces stable Earth–Moon systems from a single, continuous simulation. The resultant systems possess many desired characteristics such as an iron-deficient, heterogeneously-mixed Moon and accurate axial tilt of the Earth.

  3. BarraCUDA - a fast short read sequence aligner using graphics processing units

    LENUS (Irish Health Repository)

    Klus, Petr

    2012-01-13

    Abstract Background With the maturation of next-generation DNA sequencing (NGS) technologies, the throughput of DNA sequencing reads has soared to over 600 gigabases from a single instrument run. General purpose computing on graphics processing units (GPGPU), extracts the computing power from hundreds of parallel stream processors within graphics processing cores and provides a cost-effective and energy efficient alternative to traditional high-performance computing (HPC) clusters. In this article, we describe the implementation of BarraCUDA, a GPGPU sequence alignment software that is based on BWA, to accelerate the alignment of sequencing reads generated by these instruments to a reference DNA sequence. Findings Using the NVIDIA Compute Unified Device Architecture (CUDA) software development environment, we ported the most computational-intensive alignment component of BWA to GPU to take advantage of the massive parallelism. As a result, BarraCUDA offers a magnitude of performance boost in alignment throughput when compared to a CPU core while delivering the same level of alignment fidelity. The software is also capable of supporting multiple CUDA devices in parallel to further accelerate the alignment throughput. Conclusions BarraCUDA is designed to take advantage of the parallelism of GPU to accelerate the alignment of millions of sequencing reads generated by NGS instruments. By doing this, we could, at least in part streamline the current bioinformatics pipeline such that the wider scientific community could benefit from the sequencing technology. BarraCUDA is currently available from http:\\/\\/seqbarracuda.sf.net

  4. Deriving Hydrological Response Units (HRUs using a Web Processing Service implementation based on GRASS GIS

    Directory of Open Access Journals (Sweden)

    Christian Schwartze

    2009-04-01

    Full Text Available QGIS releases equal to or newer than 0.7 can easily connected to GRASS GIS by means of a toolbox that provides a wide range of standard GRASS modules you can launch – albeit only on data coming from GRASS. This QGIS plugin is expandable through XML configurations describing the assignment of options and inputs for a certain module. But how about embedding a precise workflow where the several processes don’t consist of a single GRASS module by force? Especially for a sequence of dependent tasks it makes sense to merge relevant GRASS functionality into an own and encapsulated QGIS extension. Its architecture and development is tested and combined with the Web Processing Service (WPS for remote execution using the concept of hydrological response units (HRUs as an example. The results of this assay may be suitable for discussing and planning other wizard-like geoprocessing plugins in QGIS that also should make use of an additional GRASS server.

  5. Spatial resolution recovery utilizing multi-ray tracing and graphic processing unit in PET image reconstruction.

    Science.gov (United States)

    Liang, Yicheng; Peng, Hao

    2015-02-07

    Depth-of-interaction (DOI) poses a major challenge for a PET system to achieve uniform spatial resolution across the field-of-view, particularly for small animal and organ-dedicated PET systems. In this work, we implemented an analytical method to model system matrix for resolution recovery, which was then incorporated in PET image reconstruction on a graphical processing unit platform, due to its parallel processing capacity. The method utilizes the concepts of virtual DOI layers and multi-ray tracing to calculate the coincidence detection response function for a given line-of-response. The accuracy of the proposed method was validated for a small-bore PET insert to be used for simultaneous PET/MR breast imaging. In addition, the performance comparisons were studied among the following three cases: 1) no physical DOI and no resolution modeling; 2) two physical DOI layers and no resolution modeling; and 3) no physical DOI design but with a different number of virtual DOI layers. The image quality was quantitatively evaluated in terms of spatial resolution (full-width-half-maximum and position offset), contrast recovery coefficient and noise. The results indicate that the proposed method has the potential to be used as an alternative to other physical DOI designs and achieve comparable imaging performances, while reducing detector/system design cost and complexity.

  6. OCTGRAV: Sparse Octree Gravitational N-body Code on Graphics Processing Units

    Science.gov (United States)

    Gaburov, Evghenii; Bédorf, Jeroen; Portegies Zwart, Simon

    2010-10-01

    Octgrav is a new very fast tree-code which runs on massively parallel Graphical Processing Units (GPU) with NVIDIA CUDA architecture. The algorithms are based on parallel-scan and sort methods. The tree-construction and calculation of multipole moments is carried out on the host CPU, while the force calculation which consists of tree walks and evaluation of interaction list is carried out on the GPU. In this way, a sustained performance of about 100GFLOP/s and data transfer rates of about 50GB/s is achieved. It takes about a second to compute forces on a million particles with an opening angle of heta approx 0.5. To test the performance and feasibility, we implemented the algorithms in CUDA in the form of a gravitational tree-code which completely runs on the GPU. The tree construction and traverse algorithms are portable to many-core devices which have support for CUDA or OpenCL programming languages. The gravitational tree-code outperforms tuned CPU code during the tree-construction and shows a performance improvement of more than a factor 20 overall, resulting in a processing rate of more than 2.8 million particles per second. The code has a convenient user interface and is freely available for use.

  7. Graphical processing unit implementation of an integrated shape-based active contour: Application to digital pathology

    Directory of Open Access Journals (Sweden)

    Sahirzeeshan Ali

    2011-01-01

    Full Text Available Commodity graphics hardware has become a cost-effective parallel platform to solve many general computational problems. In medical imaging and more so in digital pathology, segmentation of multiple structures on high-resolution images, is often a complex and computationally expensive task. Shape-based level set segmentation has recently emerged as a natural solution to segmenting overlapping and occluded objects. However the flexibility of the level set method has traditionally resulted in long computation times and therefore might have limited clinical utility. The processing times even for moderately sized images could run into several hours of computation time. Hence there is a clear need to accelerate these segmentations schemes. In this paper, we present a parallel implementation of a computationally heavy segmentation scheme on a graphical processing unit (GPU. The segmentation scheme incorporates level sets with shape priors to segment multiple overlapping nuclei from very large digital pathology images. We report a speedup of 19× compared to multithreaded C and MATLAB-based implementations of the same scheme, albeit with slight reduction in accuracy. Our GPU-based segmentation scheme was rigorously and quantitatively evaluated for the problem of nuclei segmentation and overlap resolution on digitized histopathology images corresponding to breast and prostate biopsy tissue specimens.

  8. Practical Implementation of Prestack Kirchhoff Time Migration on a General Purpose Graphics Processing Unit

    Directory of Open Access Journals (Sweden)

    Liu Guofeng

    2016-08-01

    Full Text Available In this study, we present a practical implementation of prestack Kirchhoff time migration (PSTM on a general purpose graphic processing unit. First, we consider the three main optimizations of the PSTM GPU code, i.e., designing a configuration based on a reasonable execution, using the texture memory for velocity interpolation, and the application of an intrinsic function in device code. This approach can achieve a speedup of nearly 45 times on a NVIDIA GTX 680 GPU compared with CPU code when a larger imaging space is used, where the PSTM output is a common reflection point that is gathered as I[nx][ny][nh][nt] in matrix format. However, this method requires more memory space so the limited imaging space cannot fully exploit the GPU sources. To overcome this problem, we designed a PSTM scheme with multi-GPUs for imaging different seismic data on different GPUs using an offset value. This process can achieve the peak speedup of GPU PSTM code and it greatly increases the efficiency of the calculations, but without changing the imaging result.

  9. Practical Implementation of Prestack Kirchhoff Time Migration on a General Purpose Graphics Processing Unit

    Science.gov (United States)

    Liu, Guofeng; Li, Chun

    2016-08-01

    In this study, we present a practical implementation of prestack Kirchhoff time migration (PSTM) on a general purpose graphic processing unit. First, we consider the three main optimizations of the PSTM GPU code, i.e., designing a configuration based on a reasonable execution, using the texture memory for velocity interpolation, and the application of an intrinsic function in device code. This approach can achieve a speedup of nearly 45 times on a NVIDIA GTX 680 GPU compared with CPU code when a larger imaging space is used, where the PSTM output is a common reflection point that is gathered as I[ nx][ ny][ nh][ nt] in matrix format. However, this method requires more memory space so the limited imaging space cannot fully exploit the GPU sources. To overcome this problem, we designed a PSTM scheme with multi-GPUs for imaging different seismic data on different GPUs using an offset value. This process can achieve the peak speedup of GPU PSTM code and it greatly increases the efficiency of the calculations, but without changing the imaging result.

  10. Graphics Processing Unit (GPU) Acceleration of the Goddard Earth Observing System Atmospheric Model

    Science.gov (United States)

    Putnam, Williama

    2011-01-01

    The Goddard Earth Observing System 5 (GEOS-5) is the atmospheric model used by the Global Modeling and Assimilation Office (GMAO) for a variety of applications, from long-term climate prediction at relatively coarse resolution, to data assimilation and numerical weather prediction, to very high-resolution cloud-resolving simulations. GEOS-5 is being ported to a graphics processing unit (GPU) cluster at the NASA Center for Climate Simulation (NCCS). By utilizing GPU co-processor technology, we expect to increase the throughput of GEOS-5 by at least an order of magnitude, and accelerate the process of scientific exploration across all scales of global modeling, including: The large-scale, high-end application of non-hydrostatic, global, cloud-resolving modeling at 10- to I-kilometer (km) global resolutions Intermediate-resolution seasonal climate and weather prediction at 50- to 25-km on small clusters of GPUs Long-range, coarse-resolution climate modeling, enabled on a small box of GPUs for the individual researcher After being ported to the GPU cluster, the primary physics components and the dynamical core of GEOS-5 have demonstrated a potential speedup of 15-40 times over conventional processor cores. Performance improvements of this magnitude reduce the required scalability of 1-km, global, cloud-resolving models from an unfathomable 6 million cores to an attainable 200,000 GPU-enabled cores.

  11. BarraCUDA - a fast short read sequence aligner using graphics processing units

    Directory of Open Access Journals (Sweden)

    Klus Petr

    2012-01-01

    Full Text Available Abstract Background With the maturation of next-generation DNA sequencing (NGS technologies, the throughput of DNA sequencing reads has soared to over 600 gigabases from a single instrument run. General purpose computing on graphics processing units (GPGPU, extracts the computing power from hundreds of parallel stream processors within graphics processing cores and provides a cost-effective and energy efficient alternative to traditional high-performance computing (HPC clusters. In this article, we describe the implementation of BarraCUDA, a GPGPU sequence alignment software that is based on BWA, to accelerate the alignment of sequencing reads generated by these instruments to a reference DNA sequence. Findings Using the NVIDIA Compute Unified Device Architecture (CUDA software development environment, we ported the most computational-intensive alignment component of BWA to GPU to take advantage of the massive parallelism. As a result, BarraCUDA offers a magnitude of performance boost in alignment throughput when compared to a CPU core while delivering the same level of alignment fidelity. The software is also capable of supporting multiple CUDA devices in parallel to further accelerate the alignment throughput. Conclusions BarraCUDA is designed to take advantage of the parallelism of GPU to accelerate the alignment of millions of sequencing reads generated by NGS instruments. By doing this, we could, at least in part streamline the current bioinformatics pipeline such that the wider scientific community could benefit from the sequencing technology. BarraCUDA is currently available from http://seqbarracuda.sf.net

  12. The Design Process of a Board Game for Exploring the Territories of the United States

    Directory of Open Access Journals (Sweden)

    Mehmet Kosa

    2017-06-01

    Full Text Available The paper reports the design experience of a board game with an educational aspect, which takes place on the location of states and territories of the United States. Based on a territorial acquisition dynamic, the goal was to articulate the design process of a board game that provides information for individuals who are willing to learn the locations of the U.S. states by playing a game. The game was developed using an iterative design process based on focus groups studies and brainstorming sessions. A mechanic-driven design approach was adopted instead of a theme or setting-driven alternative and a relatively abstract game was developed. The initial design idea was formed and refined according to the player feedback. The paper details play-testing sessions conducted and documents the design experience from a qualitative perspective. Our preliminary results suggest that the initial design is moderately balanced and despite the lack of quantitative evidence, our subjective observations indicate that participants’ knowledge about the location of states was improved in an entertaining and interactive way.

  13. Facilitatory Effects of Multi-Word Units in Lexical Processing and Word Learning: A Computational Investigation.

    Science.gov (United States)

    Grimm, Robert; Cassani, Giovanni; Gillis, Steven; Daelemans, Walter

    2017-01-01

    Previous studies have suggested that children and adults form cognitive representations of co-occurring word sequences. We propose (1) that the formation of such multi-word unit (MWU) representations precedes and facilitates the formation of single-word representations in children and thus benefits word learning, and (2) that MWU representations facilitate adult word recognition and thus benefit lexical processing. Using a modified version of an existing computational model (McCauley and Christiansen, 2014), we extract MWUs from a corpus of child-directed speech (CDS) and a corpus of conversations among adults. We then correlate the number of MWUs within which each word appears with (1) age of first production and (2) adult reaction times on a word recognition task. In doing so, we take care to control for the effect of word frequency, as frequent words will naturally tend to occur in many MWUs. We also compare results to a baseline model which randomly groups words into sequences-and find that MWUs have a unique facilitatory effect on both response variables, suggesting that they benefit word learning in children and word recognition in adults. The effect is strongest on age of first production, implying that MWUs are comparatively more important for word learning than for adult lexical processing. We discuss possible underlying mechanisms and formulate testable predictions.

  14. Accelerated rescaling of single Monte Carlo simulation runs with the Graphics Processing Unit (GPU).

    Science.gov (United States)

    Yang, Owen; Choi, Bernard

    2013-01-01

    To interpret fiber-based and camera-based measurements of remitted light from biological tissues, researchers typically use analytical models, such as the diffusion approximation to light transport theory, or stochastic models, such as Monte Carlo modeling. To achieve rapid (ideally real-time) measurement of tissue optical properties, especially in clinical situations, there is a critical need to accelerate Monte Carlo simulation runs. In this manuscript, we report on our approach using the Graphics Processing Unit (GPU) to accelerate rescaling of single Monte Carlo runs to calculate rapidly diffuse reflectance values for different sets of tissue optical properties. We selected MATLAB to enable non-specialists in C and CUDA-based programming to use the generated open-source code. We developed a software package with four abstraction layers. To calculate a set of diffuse reflectance values from a simulated tissue with homogeneous optical properties, our rescaling GPU-based approach achieves a reduction in computation time of several orders of magnitude as compared to other GPU-based approaches. Specifically, our GPU-based approach generated a diffuse reflectance value in 0.08ms. The transfer time from CPU to GPU memory currently is a limiting factor with GPU-based calculations. However, for calculation of multiple diffuse reflectance values, our GPU-based approach still can lead to processing that is ~3400 times faster than other GPU-based approaches.

  15. Fast ray-tracing of human eye optics on Graphics Processing Units.

    Science.gov (United States)

    Wei, Qi; Patkar, Saket; Pai, Dinesh K

    2014-05-01

    We present a new technique for simulating retinal image formation by tracing a large number of rays from objects in three dimensions as they pass through the optic apparatus of the eye to objects. Simulating human optics is useful for understanding basic questions of vision science and for studying vision defects and their corrections. Because of the complexity of computing such simulations accurately, most previous efforts used simplified analytical models of the normal eye. This makes them less effective in modeling vision disorders associated with abnormal shapes of the ocular structures which are hard to be precisely represented by analytical surfaces. We have developed a computer simulator that can simulate ocular structures of arbitrary shapes, for instance represented by polygon meshes. Topographic and geometric measurements of the cornea, lens, and retina from keratometer or medical imaging data can be integrated for individualized examination. We utilize parallel processing using modern Graphics Processing Units (GPUs) to efficiently compute retinal images by tracing millions of rays. A stable retinal image can be generated within minutes. We simulated depth-of-field, accommodation, chromatic aberrations, as well as astigmatism and correction. We also show application of the technique in patient specific vision correction by incorporating geometric models of the orbit reconstructed from clinical medical images. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.

  16. The ATLAS Fast Tracker Processing Units - input and output data preparation

    CERN Document Server

    Bolz, Arthur Eugen; The ATLAS collaboration

    2016-01-01

    The ATLAS Fast Tracker is a hardware processor built to reconstruct tracks at a rate of up to 100 kHz and provide them to the high level trigger system. The Fast Tracker will allow the trigger to utilize tracking information from the entire detector at an earlier event selection stage than ever before, allowing for more efficient event rejection. The connection of the system from to the detector read-outs and to the high level trigger computing farms are made through custom boards implementing Advanced Telecommunications Computing Technologies standard. The input is processed by the Input Mezzanines and Data Formatter boards, designed to receive and sort the data coming from the Pixel and Semi-conductor Tracker. The Fast Tracker to Level-2 Interface Card connects the system to the computing farm. The Input Mezzanines are 128 boards, performing clustering, placed on the 32 Data Formatter mother boards that sort the information into 64 logical regions required by the downstream processing units. This necessitat...

  17. ACTION OF UNIFORM SEARCH ALGORITHM WHEN SELECTING LANGUAGE UNITS IN THE PROCESS OF SPEECH

    Directory of Open Access Journals (Sweden)

    Ирина Михайловна Некипелова

    2013-05-01

    Full Text Available The article is devoted to research of action of uniform search algorithm when selecting by human of language units for speech produce. The process is connected with a speech optimization phenomenon. This makes it possible to shorten the time of cogitation something that human want to say, and to achieve the maximum precision in thoughts expression. The algorithm of uniform search works at consciousness  and subconsciousness levels. It favours the forming of automatism produce and perception of speech. Realization of human's cognitive potential in the process of communication starts up complicated mechanism of self-organization and self-regulation of language. In turn, it results in optimization of language system, servicing needs not only human's self-actualization but realization of communication in society. The method of problem-oriented search is used for researching of optimization mechanisms, which are distinctive to speech producing and stabilization of language.DOI: http://dx.doi.org/10.12731/2218-7405-2013-4-50

  18. 21st Century Parent-Child Sex Communication in the United States: A Process Review.

    Science.gov (United States)

    Flores, Dalmacio; Barroso, Julie

    2017-01-06

    Parent-child sex communication results in the transmission of family expectations, societal values, and role modeling of sexual health risk-reduction strategies. Parent-child sex communication's potential to curb negative sexual health outcomes has sustained a multidisciplinary effort to better understand the process and its impact on the development of healthy sexual attitudes and behaviors among adolescents. This review advances what is known about the process of sex communication in the United States by reviewing studies published from 2003 to 2015. We used the Cumulative Index to Nursing and Allied Health Literature (CINAHL), PsycINFO, SocINDEX, and PubMed, and the key terms "parent child" AND "sex education" for the initial query; we included 116 original articles for analysis. Our review underscores long-established factors that prevent parents from effectively broaching and sustaining talks about sex with their children and has also identified emerging concerns unique to today's parenting landscape. Parental factors salient to sex communication are established long before individuals become parents and are acted upon by influences beyond the home. Child-focused communication factors likewise describe a maturing audience that is far from captive. The identification of both enduring and emerging factors that affect how sex communication occurs will inform subsequent work that will result in more positive sexual health outcomes for adolescents.

  19. Tailoring Macroscale Response of Mechanical and Heat Transfer Systems by Topology Optimization of Microstructural Details

    DEFF Research Database (Denmark)

    Alexandersen, Joe; Lazarov, Boyan Stefanov

    2015-01-01

    -contrast material parameters is proposed to alleviate the high computational cost associated with solving the discrete systems arising during the topology optimization process. Problems within important engineering areas, heat transfer and linear elasticity, are considered for exemplifying the approach...

  20. 78 FR 1260 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2013-01-08

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: Prevailing Wage Rates for Certain Occupations Processed Under H-2A Special...

  1. An Examination of Individual Level Factors in Stress and Coping Processes: Perspectives of Chinese International Students in the United States

    Science.gov (United States)

    Yan, Kun; Berliner, David C.

    2011-01-01

    No empirical research has focused solely upon understanding the stress and coping processes of Chinese international students in the United States. This qualitative inquiry examines the individual-level variables that affect the stress-coping process of Chinese international students and how they conceptualize and adapt to their stress at an…

  2. An Examination of Individual Level Factors in Stress and Coping Processes: Perspectives of Chinese International Students in the United States

    Science.gov (United States)

    Yan, Kun; Berliner, David C.

    2011-01-01

    No empirical research has focused solely upon understanding the stress and coping processes of Chinese international students in the United States. This qualitative inquiry examines the individual-level variables that affect the stress-coping process of Chinese international students and how they conceptualize and adapt to their stress at an…

  3. Accelerating Wright-Fisher Forward Simulations on the Graphics Processing Unit.

    Science.gov (United States)

    Lawrie, David S

    2017-09-07

    Forward Wright-Fisher simulations are powerful in their ability to model complex demography and selection scenarios, but suffer from slow execution on the Central Processor Unit (CPU), thus limiting their usefulness. However, the single-locus Wright-Fisher forward algorithm is exceedingly parallelizable, with many steps that are so-called "embarrassingly parallel," consisting of a vast number of individual computations that are all independent of each other and thus capable of being performed concurrently. The rise of modern Graphics Processing Units (GPUs) and programming languages designed to leverage the inherent parallel nature of these processors have allowed researchers to dramatically speed up many programs that have such high arithmetic intensity and intrinsic concurrency. The presented GPU Optimized Wright-Fisher simulation, or "GO Fish" for short, can be used to simulate arbitrary selection and demographic scenarios while running over 250-fold faster than its serial counterpart on the CPU. Even modest GPU hardware can achieve an impressive speedup of over two orders of magnitude. With simulations so accelerated, one can not only do quick parametric bootstrapping of previously estimated parameters, but also use simulated results to calculate the likelihoods and summary statistics of demographic and selection models against real polymorphism data, all without restricting the demographic and selection scenarios that can be modeled or requiring approximations to the single-locus forward algorithm for efficiency. Further, as many of the parallel programming techniques used in this simulation can be applied to other computationally intensive algorithms important in population genetics, GO Fish serves as an exciting template for future research into accelerating computation in evolution. GO Fish is part of the Parallel PopGen Package available at: http://dl42.github.io/ParallelPopGen/. Copyright © 2017 Lawrie.

  4. Computerized nursing process in the Intensive Care Unit: ergonomics and usability.

    Science.gov (United States)

    Almeida, Sônia Regina Wagner de; Sasso, Grace Teresinha Marcon Dal; Barra, Daniela Couto Carvalho

    2016-01-01

    Analyzing the ergonomics and usability criteria of the Computerized Nursing Process based on the International Classification for Nursing Practice in the Intensive Care Unit according to International Organization for Standardization(ISO). A quantitative, quasi-experimental, before-and-after study with a sample of 16 participants performed in an Intensive Care Unit. Data collection was performed through the application of five simulated clinical cases and an evaluation instrument. Data analysis was performed by descriptive and inferential statistics. The organization, content and technical criteria were considered "excellent", and the interface criteria were considered "very good", obtaining means of 4.54, 4.60, 4.64 and 4.39, respectively. The analyzed standards obtained means above 4.0, being considered "very good" by the participants. The Computerized Nursing Processmet ergonomic and usability standards according to the standards set by ISO. This technology supports nurses' clinical decision-making by providing complete and up-to-date content for Nursing practice in the Intensive Care Unit. Analisar os critérios de ergonomia e usabilidade do Processo de Enfermagem Informatizado a partir da Classificação Internacional para as Práticas de Enfermagem, em Unidade de Terapia Intensiva, de acordo com os padrões da InternationalOrganization for Standardization (ISO). Pesquisa quantitativa, quase-experimental do tipo antes e depois, com uma amostra de 16 participantes, realizada em uma Unidade de Terapia Intensiva. Coleta de dados realizada por meio da aplicação de cinco casos clínicos simulados e instrumento de avaliação. A análise dos dados foi realizada pela estatística descritiva e inferencial. Os critérios organização, conteúdo e técnico foram considerados "excelentes", e o critério interface "muito bom", obtendo médias 4,54, 4,60, 4,64 e 4,39, respectivamente. Os padrões analisados obtiveram médias acima de 4,0, sendo considerados "muito bons

  5. Relationships between registered nurse staffing, processes of nursing care, and nurse-reported patient outcomes in chronic hemodialysis units.

    Science.gov (United States)

    Thomas-Hawkins, Charlotte; Flynn, Linda; Clarke, Sean P

    2008-01-01

    Little attention has been given to the effects of registered nurse (RN) staffing and processes of nursing care on patient outcomes in hemodialysis units. This research examined the effects of patient-to-RN ratios and necessary tasks left undone by RNs on the likelihood of nurse-reported frequent occurrences of adverse patient events in chronic hemodialysis units. Study findings revealed that high patient-to-RN ratios and increased numbers of tasks left undone by RNs were associated with an increased likelihood of frequent occurrences of dialysis hypotension, skipped dialysis treatments, shortened dialysis treatments, and patient complaints in hemodialysis units. These findings indicate that federal, state, and dialysis organization policies must foster staffing structures and processes of care in dialysis units that effectively utilize the invaluable skills and services of professional, registered nurses.

  6. Quantification of Cell-Free DNA in Red Blood Cell Units in Different Whole Blood Processing Methods

    Directory of Open Access Journals (Sweden)

    Andrew W. Shih

    2016-01-01

    Full Text Available Background. Whole blood donations in Canada are processed by either the red cell filtration (RCF or whole blood filtration (WBF methods, where leukoreduction is potentially delayed in WBF. Fresh WBF red blood cells (RBCs have been associated with increased in-hospital mortality after transfusion. Cell-free DNA (cfDNA is released by neutrophils prior to leukoreduction, degraded during RBC storage, and is associated with adverse patient outcomes. We explored cfDNA levels in RBCs prepared by RCF and WBF and different storage durations. Methods. Equal numbers of fresh (stored ≤14 days and older RBCs were sampled. cfDNA was quantified by spectrophotometry and PicoGreen. Separate regression models determined the association with processing method and storage duration and their interaction on cfDNA. Results. cfDNA in 120 RBC units (73 RCF, 47 WBF were measured. Using PicoGreen, WBF units overall had higher cfDNA than RCF units (p=0.0010; fresh WBF units had higher cfDNA than fresh RCF units (p=0.0093. Using spectrophotometry, fresh RBC units overall had higher cfDNA than older units (p=0.0031; fresh WBF RBCs had higher cfDNA than older RCF RBCs (p=0.024. Conclusion. Higher cfDNA in fresh WBF was observed compared to older RCF blood. Further study is required for association with patient outcomes.

  7. In-Situ Statistical Analysis of Autotune Simulation Data using Graphical Processing Units

    Energy Technology Data Exchange (ETDEWEB)

    Ranjan, Niloo [ORNL; Sanyal, Jibonananda [ORNL; New, Joshua Ryan [ORNL

    2013-08-01

    Developing accurate building energy simulation models to assist energy efficiency at speed and scale is one of the research goals of the Whole-Building and Community Integration group, which is a part of Building Technologies Research and Integration Center (BTRIC) at Oak Ridge National Laboratory (ORNL). The aim of the Autotune project is to speed up the automated calibration of building energy models to match measured utility or sensor data. The workflow of this project takes input parameters and runs EnergyPlus simulations on Oak Ridge Leadership Computing Facility s (OLCF) computing resources such as Titan, the world s second fastest supercomputer. Multiple simulations run in parallel on nodes having 16 processors each and a Graphics Processing Unit (GPU). Each node produces a 5.7 GB output file comprising 256 files from 64 simulations. Four types of output data covering monthly, daily, hourly, and 15-minute time steps for each annual simulation is produced. A total of 270TB+ of data has been produced. In this project, the simulation data is statistically analyzed in-situ using GPUs while annual simulations are being computed on the traditional processors. Titan, with its recent addition of 18,688 Compute Unified Device Architecture (CUDA) capable NVIDIA GPUs, has greatly extended its capability for massively parallel data processing. CUDA is used along with C/MPI to calculate statistical metrics such as sum, mean, variance, and standard deviation leveraging GPU acceleration. The workflow developed in this project produces statistical summaries of the data which reduces by multiple orders of magnitude the time and amount of data that needs to be stored. These statistical capabilities are anticipated to be useful for sensitivity analysis of EnergyPlus simulations.

  8. Monte Carlo simulation of photon migration in 3D turbid media accelerated by graphics processing units.

    Science.gov (United States)

    Fang, Qianqian; Boas, David A

    2009-10-26

    We report a parallel Monte Carlo algorithm accelerated by graphics processing units (GPU) for modeling time-resolved photon migration in arbitrary 3D turbid media. By taking advantage of the massively parallel threads and low-memory latency, this algorithm allows many photons to be simulated simultaneously in a GPU. To further improve the computational efficiency, we explored two parallel random number generators (RNG), including a floating-point-only RNG based on a chaotic lattice. An efficient scheme for boundary reflection was implemented, along with the functions for time-resolved imaging. For a homogeneous semi-infinite medium, good agreement was observed between the simulation output and the analytical solution from the diffusion theory. The code was implemented with CUDA programming language, and benchmarked under various parameters, such as thread number, selection of RNG and memory access pattern. With a low-cost graphics card, this algorithm has demonstrated an acceleration ratio above 300 when using 1792 parallel threads over conventional CPU computation. The acceleration ratio drops to 75 when using atomic operations. These results render the GPU-based Monte Carlo simulation a practical solution for data analysis in a wide range of diffuse optical imaging applications, such as human brain or small-animal imaging.

  9. Quantitative Estimation of Risks for Production Unit Based on OSHMS and Process Resilience

    Science.gov (United States)

    Nyambayar, D.; Koshijima, I.; Eguchi, H.

    2017-06-01

    Three principal elements in the production field of chemical/petrochemical industry are (i) Production Units, (ii) Production Plant Personnel and (iii) Production Support System (computer system introduced for improving productivity). Each principal element has production process resilience, i.e. a capability to restrain disruptive signals occurred in and out of the production field. In each principal element, risk assessment is indispensable for the production field. In a production facility, the occupational safety and health management system (Hereafter, referred to as OSHMS) has been introduced to reduce a risk of accidents and troubles that may occur during production. In OSHMS, a risk assessment is specified to reduce a potential risk in the production facility such as a factory, and PDCA activities are required for a continual improvement of safety production environments. However, there is no clear statement to adopt the OSHMS standard into the production field. This study introduces a metric to estimate the resilience of the production field by using the resilience generated by the production plant personnel and the result of the risk assessment in the production field. A method for evaluating how OSHMS functions are systematically installed in the production field is also discussed based on the resilience of the three principal elements.

  10. A software architecture for multi-cellular system simulations on graphics processing units.

    Science.gov (United States)

    Jeannin-Girardon, Anne; Ballet, Pascal; Rodin, Vincent

    2013-09-01

    The first aim of simulation in virtual environment is to help biologists to have a better understanding of the simulated system. The cost of such simulation is significantly reduced compared to that of in vivo simulation. However, the inherent complexity of biological system makes it hard to simulate these systems on non-parallel architectures: models might be made of sub-models and take several scales into account; the number of simulated entities may be quite large. Today, graphics cards are used for general purpose computing which has been made easier thanks to frameworks like CUDA or OpenCL. Parallelization of models may however not be easy: parallel computer programing skills are often required; several hardware architectures may be used to execute models. In this paper, we present the software architecture we built in order to implement various models able to simulate multi-cellular system. This architecture is modular and it implements data structures adapted for graphics processing units architectures. It allows efficient simulation of biological mechanisms.

  11. Transparent Runtime Migration of Loop-Based Traces of Processor Instructions to Reconfigurable Processing Units

    Directory of Open Access Journals (Sweden)

    João Bispo

    2013-01-01

    Full Text Available The ability to map instructions running in a microprocessor to a reconfigurable processing unit (RPU, acting as a coprocessor, enables the runtime acceleration of applications and ensures code and possibly performance portability. In this work, we focus on the mapping of loop-based instruction traces (called Megablocks to RPUs. The proposed approach considers offline partitioning and mapping stages without ignoring their future runtime applicability. We present a toolchain that automatically extracts specific trace-based loops, called Megablocks, from MicroBlaze instruction traces and generates an RPU for executing those loops. Our hardware infrastructure is able to move loop execution from the microprocessor to the RPU transparently, at runtime, and without changing the executable binaries. The toolchain and the system are fully operational. Three FPGA implementations of the system, differing in the hardware interfaces used, were tested and evaluated with a set of 15 application kernels. Speedups ranging from 1.26 to 3.69 were achieved for the best alternative using a MicroBlaze processor with local memory.

  12. Accelerated Molecular Dynamics Simulations with the AMOEBA Polarizable Force Field on Graphics Processing Units.

    Science.gov (United States)

    Lindert, Steffen; Bucher, Denis; Eastman, Peter; Pande, Vijay; McCammon, J Andrew

    2013-11-12

    The accelerated molecular dynamics (aMD) method has recently been shown to enhance the sampling of biomolecules in molecular dynamics (MD) simulations, often by several orders of magnitude. Here, we describe an implementation of the aMD method for the OpenMM application layer that takes full advantage of graphics processing units (GPUs) computing. The aMD method is shown to work in combination with the AMOEBA polarizable force field (AMOEBA-aMD), allowing the simulation of long time-scale events with a polarizable force field. Benchmarks are provided to show that the AMOEBA-aMD method is efficiently implemented and produces accurate results in its standard parametrization. For the BPTI protein, we demonstrate that the protein structure described with AMOEBA remains stable even on the extended time scales accessed at high levels of accelerations. For the DNA repair metalloenzyme endonuclease IV, we show that the use of the AMOEBA force field is a significant improvement over fixed charged models for describing the enzyme active-site. The new AMOEBA-aMD method is publicly available (http://wiki.simtk.org/openmm/VirtualRepository) and promises to be interesting for studying complex systems that can benefit from both the use of a polarizable force field and enhanced sampling.

  13. Developing a multiscale, multi-resolution agent-based brain tumor model by graphics processing units

    Directory of Open Access Journals (Sweden)

    Zhang Le

    2011-12-01

    Full Text Available Abstract Multiscale agent-based modeling (MABM has been widely used to simulate Glioblastoma Multiforme (GBM and its progression. At the intracellular level, the MABM approach employs a system of ordinary differential equations to describe quantitatively specific intracellular molecular pathways that determine phenotypic switches among cells (e.g. from migration to proliferation and vice versa. At the intercellular level, MABM describes cell-cell interactions by a discrete module. At the tissue level, partial differential equations are employed to model the diffusion of chemoattractants, which are the input factors of the intracellular molecular pathway. Moreover, multiscale analysis makes it possible to explore the molecules that play important roles in determining the cellular phenotypic switches that in turn drive the whole GBM expansion. However, owing to limited computational resources, MABM is currently a theoretical biological model that uses relatively coarse grids to simulate a few cancer cells in a small slice of brain cancer tissue. In order to improve this theoretical model to simulate and predict actual GBM cancer progression in real time, a graphics processing unit (GPU-based parallel computing algorithm was developed and combined with the multi-resolution design to speed up the MABM. The simulated results demonstrated that the GPU-based, multi-resolution and multiscale approach can accelerate the previous MABM around 30-fold with relatively fine grids in a large extracellular matrix. Therefore, the new model has great potential for simulating and predicting real-time GBM progression, if real experimental data are incorporated.

  14. An Optimized Multicolor Point-Implicit Solver for Unstructured Grid Applications on Graphics Processing Units

    Science.gov (United States)

    Zubair, Mohammad; Nielsen, Eric; Luitjens, Justin; Hammond, Dana

    2016-01-01

    In the field of computational fluid dynamics, the Navier-Stokes equations are often solved using an unstructuredgrid approach to accommodate geometric complexity. Implicit solution methodologies for such spatial discretizations generally require frequent solution of large tightly-coupled systems of block-sparse linear equations. The multicolor point-implicit solver used in the current work typically requires a significant fraction of the overall application run time. In this work, an efficient implementation of the solver for graphics processing units is proposed. Several factors present unique challenges to achieving an efficient implementation in this environment. These include the variable amount of parallelism available in different kernel calls, indirect memory access patterns, low arithmetic intensity, and the requirement to support variable block sizes. In this work, the solver is reformulated to use standard sparse and dense Basic Linear Algebra Subprograms (BLAS) functions. However, numerical experiments show that the performance of the BLAS functions available in existing CUDA libraries is suboptimal for matrices representative of those encountered in actual simulations. Instead, optimized versions of these functions are developed. Depending on block size, the new implementations show performance gains of up to 7x over the existing CUDA library functions.

  15. Parallel design of JPEG-LS encoder on graphics processing units

    Science.gov (United States)

    Duan, Hao; Fang, Yong; Huang, Bormin

    2012-01-01

    With recent technical advances in graphic processing units (GPUs), GPUs have outperformed CPUs in terms of compute capability and memory bandwidth. Many successful GPU applications to high performance computing have been reported. JPEG-LS is an ISO/IEC standard for lossless image compression which utilizes adaptive context modeling and run-length coding to improve compression ratio. However, adaptive context modeling causes data dependency among adjacent pixels and the run-length coding has to be performed in a sequential way. Hence, using JPEG-LS to compress large-volume hyperspectral image data is quite time-consuming. We implement an efficient parallel JPEG-LS encoder for lossless hyperspectral compression on a NVIDIA GPU using the computer unified device architecture (CUDA) programming technology. We use the block parallel strategy, as well as such CUDA techniques as coalesced global memory access, parallel prefix sum, and asynchronous data transfer. We also show the relation between GPU speedup and AVIRIS block size, as well as the relation between compression ratio and AVIRIS block size. When AVIRIS images are divided into blocks, each with 64×64 pixels, we gain the best GPU performance with 26.3x speedup over its original CPU code.

  16. Graphic processing unit accelerated real-time partially coherent beam generator

    Science.gov (United States)

    Ni, Xiaolong; Liu, Zhi; Chen, Chunyi; Jiang, Huilin; Fang, Hanhan; Song, Lujun; Zhang, Su

    2016-07-01

    A method of using liquid-crystals (LCs) to generate a partially coherent beam in real-time is described. An expression for generating a partially coherent beam is given and calculated using a graphic processing unit (GPU), i.e., the GeForce GTX 680. A liquid-crystal on silicon (LCOS) with 256 × 256 pixels is used as the partially coherent beam generator (PCBG). An optimizing method with partition convolution is used to improve the generating speed of our LC PCBG. The total time needed to generate a random phase map with a coherence width range from 0.015 mm to 1.5 mm is less than 2.4 ms for calculation and readout with the GPU; adding the time needed for the CPU to read and send to LCOS with the response time of the LC PCBG, the real-time partially coherent beam (PCB) generation frequency of our LC PCBG is up to 312 Hz. To our knowledge, it is the first real-time partially coherent beam generator. A series of experiments based on double pinhole interference are performed. The result shows that to generate a laser beam with a coherence width of 0.9 mm and 1.5 mm, with a mean error of approximately 1%, the RMS values needed 0.021306 and 0.020883 and the PV values required 0.073576 and 0.072998, respectively.

  17. Graphics processing unit (GPU)-accelerated particle filter framework for positron emission tomography image reconstruction.

    Science.gov (United States)

    Yu, Fengchao; Liu, Huafeng; Hu, Zhenghui; Shi, Pengcheng

    2012-04-01

    As a consequence of the random nature of photon emissions and detections, the data collected by a positron emission tomography (PET) imaging system can be shown to be Poisson distributed. Meanwhile, there have been considerable efforts within the tracer kinetic modeling communities aimed at establishing the relationship between the PET data and physiological parameters that affect the uptake and metabolism of the tracer. Both statistical and physiological models are important to PET reconstruction. The majority of previous efforts are based on simplified, nonphysical mathematical expression, such as Poisson modeling of the measured data, which is, on the whole, completed without consideration of the underlying physiology. In this paper, we proposed a graphics processing unit (GPU)-accelerated reconstruction strategy that can take both statistical model and physiological model into consideration with the aid of state-space evolution equations. The proposed strategy formulates the organ activity distribution through tracer kinetics models and the photon-counting measurements through observation equations, thus making it possible to unify these two constraints into a general framework. In order to accelerate reconstruction, GPU-based parallel computing is introduced. Experiments of Zubal-thorax-phantom data, Monte Carlo simulated phantom data, and real phantom data show the power of the method. Furthermore, thanks to the computing power of the GPU, the reconstruction time is practical for clinical application.

  18. Exploring Graphics Processing Unit (GPU Resource Sharing Efficiency for High Performance Computing

    Directory of Open Access Journals (Sweden)

    Teng Li

    2013-11-01

    Full Text Available The increasing incorporation of Graphics Processing Units (GPUs as accelerators has been one of the forefront High Performance Computing (HPC trends and provides unprecedented performance; however, the prevalent adoption of the Single-Program Multiple-Data (SPMD programming model brings with it challenges of resource underutilization. In other words, under SPMD, every CPU needs GPU capability available to it. However, since CPUs generally outnumber GPUs, the asymmetric resource distribution gives rise to overall computing resource underutilization. In this paper, we propose to efficiently share the GPU under SPMD and formally define a series of GPU sharing scenarios. We provide performance-modeling analysis for each sharing scenario with accurate experimentation validation. With the modeling basis, we further conduct experimental studies to explore potential GPU sharing efficiency improvements from multiple perspectives. Both further theoretical and experimental GPU sharing performance analysis and results are presented. Our results not only demonstrate the significant performance gain for SPMD programs with the proposed efficient GPU sharing, but also the further improved sharing efficiency with the optimization techniques based on our accurate modeling.

  19. High-Throughput Characterization of Porous Materials Using Graphics Processing Units

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Jihan; Martin, Richard L.; Rübel, Oliver; Haranczyk, Maciej; Smit, Berend

    2012-05-08

    We have developed a high-throughput graphics processing units (GPU) code that can characterize a large database of crystalline porous materials. In our algorithm, the GPU is utilized to accelerate energy grid calculations where the grid values represent interactions (i.e., Lennard-Jones + Coulomb potentials) between gas molecules (i.e., CH$_{4}$ and CO$_{2}$) and material's framework atoms. Using a parallel flood fill CPU algorithm, inaccessible regions inside the framework structures are identified and blocked based on their energy profiles. Finally, we compute the Henry coefficients and heats of adsorption through statistical Widom insertion Monte Carlo moves in the domain restricted to the accessible space. The code offers significant speedup over a single core CPU code and allows us to characterize a set of porous materials at least an order of magnitude larger than ones considered in earlier studies. For structures selected from such a prescreening algorithm, full adsorption isotherms can be calculated by conducting multiple grand canonical Monte Carlo simulations concurrently within the GPU.

  20. Fast data preprocessing with Graphics Processing Units for inverse problem solving in light-scattering measurements

    Science.gov (United States)

    Derkachov, G.; Jakubczyk, T.; Jakubczyk, D.; Archer, J.; Woźniak, M.

    2017-07-01

    Utilising Compute Unified Device Architecture (CUDA) platform for Graphics Processing Units (GPUs) enables significant reduction of computation time at a moderate cost, by means of parallel computing. In the paper [Jakubczyk et al., Opto-Electron. Rev., 2016] we reported using GPU for Mie scattering inverse problem solving (up to 800-fold speed-up). Here we report the development of two subroutines utilising GPU at data preprocessing stages for the inversion procedure: (i) A subroutine, based on ray tracing, for finding spherical aberration correction function. (ii) A subroutine performing the conversion of an image to a 1D distribution of light intensity versus azimuth angle (i.e. scattering diagram), fed from a movie-reading CPU subroutine running in parallel. All subroutines are incorporated in PikeReader application, which we make available on GitHub repository. PikeReader returns a sequence of intensity distributions versus a common azimuth angle vector, corresponding to the recorded movie. We obtained an overall ∼ 400 -fold speed-up of calculations at data preprocessing stages using CUDA codes running on GPU in comparison to single thread MATLAB-only code running on CPU.

  1. GPUmotif: an ultra-fast and energy-efficient motif analysis program using graphics processing units.

    Directory of Open Access Journals (Sweden)

    Pooya Zandevakili

    Full Text Available Computational detection of TF binding patterns has become an indispensable tool in functional genomics research. With the rapid advance of new sequencing technologies, large amounts of protein-DNA interaction data have been produced. Analyzing this data can provide substantial insight into the mechanisms of transcriptional regulation. However, the massive amount of sequence data presents daunting challenges. In our previous work, we have developed a novel algorithm called Hybrid Motif Sampler (HMS that enables more scalable and accurate motif analysis. Despite much improvement, HMS is still time-consuming due to the requirement to calculate matching probabilities position-by-position. Using the NVIDIA CUDA toolkit, we developed a graphics processing unit (GPU-accelerated motif analysis program named GPUmotif. We proposed a "fragmentation" technique to hide data transfer time between memories. Performance comparison studies showed that commonly-used model-based motif scan and de novo motif finding procedures such as HMS can be dramatically accelerated when running GPUmotif on NVIDIA graphics cards. As a result, energy consumption can also be greatly reduced when running motif analysis using GPUmotif. The GPUmotif program is freely available at http://sourceforge.net/projects/gpumotif/

  2. Efficient molecular dynamics simulations with many-body potentials on graphics processing units

    CERN Document Server

    Fan, Zheyong; Vierimaa, Ville; Harju, Ari

    2016-01-01

    Graphics processing units have been extensively used to accelerate classical molecular dynamics simulations. However, there is much less progress on the acceleration of force evaluations for many-body potentials compared to pairwise ones. In the conventional force evaluation algorithm for many-body potentials, the force, virial stress, and heat current for a given atom are accumulated within different loops, which could result in write conflict between different threads in a CUDA kernel. In this work, we provide a new force evaluation algorithm, which is based on an explicit pairwise force expression for many-body potentials derived recently [Phys. Rev. B 92 (2015) 094301]. In our algorithm, the force, virial stress, and heat current for a given atom can be accumulated within a single thread and is free of write conflicts. We discuss the formulations and algorithms and evaluate their performance. A new open-source code, GPUMD, is developed based on the proposed formulations. For the Tersoff many-body potentia...

  3. Accelerating large-scale protein structure alignments with graphics processing units

    Directory of Open Access Journals (Sweden)

    Pang Bin

    2012-02-01

    Full Text Available Abstract Background Large-scale protein structure alignment, an indispensable tool to structural bioinformatics, poses a tremendous challenge on computational resources. To ensure structure alignment accuracy and efficiency, efforts have been made to parallelize traditional alignment algorithms in grid environments. However, these solutions are costly and of limited accessibility. Others trade alignment quality for speedup by using high-level characteristics of structure fragments for structure comparisons. Findings We present ppsAlign, a parallel protein structure Alignment framework designed and optimized to exploit the parallelism of Graphics Processing Units (GPUs. As a general-purpose GPU platform, ppsAlign could take many concurrent methods, such as TM-align and Fr-TM-align, into the parallelized algorithm design. We evaluated ppsAlign on an NVIDIA Tesla C2050 GPU card, and compared it with existing software solutions running on an AMD dual-core CPU. We observed a 36-fold speedup over TM-align, a 65-fold speedup over Fr-TM-align, and a 40-fold speedup over MAMMOTH. Conclusions ppsAlign is a high-performance protein structure alignment tool designed to tackle the computational complexity issues from protein structural data. The solution presented in this paper allows large-scale structure comparisons to be performed using massive parallel computing power of GPU.

  4. Development Status of Power Processing Unit for 250mN-Class Hall Thruster

    Science.gov (United States)

    Osuga, H.; Suzuki, K.; Ozaki, T.; Nakagawa, T.; Suga, I.; Tamida, T.; Akuzawa, Y.; Suzuki, H.; Soga, Y.; Furuichi, T.; Maki, S.; Matui, K.

    2008-09-01

    Institute for Unmanned Space Experiment Free Flyer (USEF) and Mitsubishi Electric Corporation (MELCO) are developing the next generation ion engine system under the sponsorship of Ministry of Economy, Trade and Industry (METI) within six years. The system requirement specifications are a thrust level of over 250mN and specific impulse of over 1500 sec with a less than 5kW electric power supply, and a lifetime of over 3,000 hours. These target specifications required the development of both a Hall Thruster and a Power Processing Unit (PPU). In the 2007 fiscal year, the PPU called Second Engineering Model (EM2) consist of all power supplies was a model for the Hall Thruster system. The EM2 PPU showed the discharge efficiency was over 96.2% for 250V and 350V at output power between 1.8kW to 4.5kW. And also the Hall Thruster could start up quickly and smoothly to control the discharge voltage, the inner magnet current, the outer magnet current and the xenon flow rate. This paper reports on the design and test results of the EM2 PPU.

  5. Accelerating frequency-domain diffuse optical tomographic image reconstruction using graphics processing units.

    Science.gov (United States)

    Prakash, Jaya; Chandrasekharan, Venkittarayan; Upendra, Vishwajith; Yalavarthy, Phaneendra K

    2010-01-01

    Diffuse optical tomographic image reconstruction uses advanced numerical models that are computationally costly to be implemented in the real time. The graphics processing units (GPUs) offer desktop massive parallelization that can accelerate these computations. An open-source GPU-accelerated linear algebra library package is used to compute the most intensive matrix-matrix calculations and matrix decompositions that are used in solving the system of linear equations. These open-source functions were integrated into the existing frequency-domain diffuse optical image reconstruction algorithms to evaluate the acceleration capability of the GPUs (NVIDIA Tesla C 1060) with increasing reconstruction problem sizes. These studies indicate that single precision computations are sufficient for diffuse optical tomographic image reconstruction. The acceleration per iteration can be up to 40, using GPUs compared to traditional CPUs in case of three-dimensional reconstruction, where the reconstruction problem is more underdetermined, making the GPUs more attractive in the clinical settings. The current limitation of these GPUs in the available onboard memory (4 GB) that restricts the reconstruction of a large set of optical parameters, more than 13,377.

  6. Accelerating image reconstruction in three-dimensional optoacoustic tomography on graphics processing units.

    Science.gov (United States)

    Wang, Kun; Huang, Chao; Kao, Yu-Jiun; Chou, Cheng-Ying; Oraevsky, Alexander A; Anastasio, Mark A

    2013-02-01

    Optoacoustic tomography (OAT) is inherently a three-dimensional (3D) inverse problem. However, most studies of OAT image reconstruction still employ two-dimensional imaging models. One important reason is because 3D image reconstruction is computationally burdensome. The aim of this work is to accelerate existing image reconstruction algorithms for 3D OAT by use of parallel programming techniques. Parallelization strategies are proposed to accelerate a filtered backprojection (FBP) algorithm and two different pairs of projection/backprojection operations that correspond to two different numerical imaging models. The algorithms are designed to fully exploit the parallel computing power of graphics processing units (GPUs). In order to evaluate the parallelization strategies for the projection/backprojection pairs, an iterative image reconstruction algorithm is implemented. Computer simulation and experimental studies are conducted to investigate the computational efficiency and numerical accuracy of the developed algorithms. The GPU implementations improve the computational efficiency by factors of 1000, 125, and 250 for the FBP algorithm and the two pairs of projection/backprojection operators, respectively. Accurate images are reconstructed by use of the FBP and iterative image reconstruction algorithms from both computer-simulated and experimental data. Parallelization strategies for 3D OAT image reconstruction are proposed for the first time. These GPU-based implementations significantly reduce the computational time for 3D image reconstruction, complementing our earlier work on 3D OAT iterative image reconstruction.

  7. High-speed nonlinear finite element analysis for surgical simulation using graphics processing units.

    Science.gov (United States)

    Taylor, Z A; Cheng, M; Ourselin, S

    2008-05-01

    The use of biomechanical modelling, especially in conjunction with finite element analysis, has become common in many areas of medical image analysis and surgical simulation. Clinical employment of such techniques is hindered by conflicting requirements for high fidelity in the modelling approach, and fast solution speeds. We report the development of techniques for high-speed nonlinear finite element analysis for surgical simulation. We use a fully nonlinear total Lagrangian explicit finite element formulation which offers significant computational advantages for soft tissue simulation. However, the key contribution of the work is the presentation of a fast graphics processing unit (GPU) solution scheme for the finite element equations. To the best of our knowledge, this represents the first GPU implementation of a nonlinear finite element solver. We show that the present explicit finite element scheme is well suited to solution via highly parallel graphics hardware, and that even a midrange GPU allows significant solution speed gains (up to 16.8 x) compared with equivalent CPU implementations. For the models tested the scheme allows real-time solution of models with up to 16,000 tetrahedral elements. The use of GPUs for such purposes offers a cost-effective high-performance alternative to expensive multi-CPU machines, and may have important applications in medical image analysis and surgical simulation.

  8. permGPU: Using graphics processing units in RNA microarray association studies

    Directory of Open Access Journals (Sweden)

    George Stephen L

    2010-06-01

    Full Text Available Abstract Background Many analyses of microarray association studies involve permutation, bootstrap resampling and cross-validation, that are ideally formulated as embarrassingly parallel computing problems. Given that these analyses are computationally intensive, scalable approaches that can take advantage of multi-core processor systems need to be developed. Results We have developed a CUDA based implementation, permGPU, that employs graphics processing units in microarray association studies. We illustrate the performance and applicability of permGPU within the context of permutation resampling for a number of test statistics. An extensive simulation study demonstrates a dramatic increase in performance when using permGPU on an NVIDIA GTX 280 card compared to an optimized C/C++ solution running on a conventional Linux server. Conclusions permGPU is available as an open-source stand-alone application and as an extension package for the R statistical environment. It provides a dramatic increase in performance for permutation resampling analysis in the context of microarray association studies. The current version offers six test statistics for carrying out permutation resampling analyses for binary, quantitative and censored time-to-event traits.

  9. Multidimensional upwind hydrodynamics on unstructured meshes using graphics processing units - I. Two-dimensional uniform meshes

    Science.gov (United States)

    Paardekooper, S.-J.

    2017-08-01

    We present a new method for numerical hydrodynamics which uses a multidimensional generalization of the Roe solver and operates on an unstructured triangular mesh. The main advantage over traditional methods based on Riemann solvers, which commonly use one-dimensional flux estimates as building blocks for a multidimensional integration, is its inherently multidimensional nature, and as a consequence its ability to recognize multidimensional stationary states that are not hydrostatic. A second novelty is the focus on graphics processing units (GPUs). By tailoring the algorithms specifically to GPUs, we are able to get speedups of 100-250 compared to a desktop machine. We compare the multidimensional upwind scheme to a traditional, dimensionally split implementation of the Roe solver on several test problems, and we find that the new method significantly outperforms the Roe solver in almost all cases. This comes with increased computational costs per time-step, which makes the new method approximately a factor of 2 slower than a dimensionally split scheme acting on a structured grid.

  10. Efficient molecular dynamics simulations with many-body potentials on graphics processing units

    Science.gov (United States)

    Fan, Zheyong; Chen, Wei; Vierimaa, Ville; Harju, Ari

    2017-09-01

    Graphics processing units have been extensively used to accelerate classical molecular dynamics simulations. However, there is much less progress on the acceleration of force evaluations for many-body potentials compared to pairwise ones. In the conventional force evaluation algorithm for many-body potentials, the force, virial stress, and heat current for a given atom are accumulated within different loops, which could result in write conflict between different threads in a CUDA kernel. In this work, we provide a new force evaluation algorithm, which is based on an explicit pairwise force expression for many-body potentials derived recently (Fan et al., 2015). In our algorithm, the force, virial stress, and heat current for a given atom can be accumulated within a single thread and is free of write conflicts. We discuss the formulations and algorithms and evaluate their performance. A new open-source code, GPUMD, is developed based on the proposed formulations. For the Tersoff many-body potential, the double precision performance of GPUMD using a Tesla K40 card is equivalent to that of the LAMMPS (Large-scale Atomic/Molecular Massively Parallel Simulator) molecular dynamics code running with about 100 CPU cores (Intel Xeon CPU X5670 @ 2.93 GHz).

  11. The application of projected conjugate gradient solvers on graphical processing units

    Energy Technology Data Exchange (ETDEWEB)

    Lin, Youzuo [Los Alamos National Laboratory; Renaut, Rosemary [ARIZONA STATE UNIV.

    2011-01-26

    Graphical processing units introduce the capability for large scale computation at the desktop. Presented numerical results verify that efficiencies and accuracies of basic linear algebra subroutines of all levels when implemented in CUDA and Jacket are comparable. But experimental results demonstrate that the basic linear algebra subroutines of level three offer the greatest potential for improving efficiency of basic numerical algorithms. We consider the solution of the multiple right hand side set of linear equations using Krylov subspace-based solvers. Thus, for the multiple right hand side case, it is more efficient to make use of a block implementation of the conjugate gradient algorithm, rather than to solve each system independently. Jacket is used for the implementation. Furthermore, including projection from one system to another improves efficiency. A relevant example, for which simulated results are provided, is the reconstruction of a three dimensional medical image volume acquired from a positron emission tomography scanner. Efficiency of the reconstruction is improved by using projection across nearby slices.

  12. Space Object Collision Probability via Monte Carlo on the Graphics Processing Unit

    Science.gov (United States)

    Vittaldev, Vivek; Russell, Ryan P.

    2017-09-01

    Fast and accurate collision probability computations are essential for protecting space assets. Monte Carlo (MC) simulation is the most accurate but computationally intensive method. A Graphics Processing Unit (GPU) is used to parallelize the computation and reduce the overall runtime. Using MC techniques to compute the collision probability is common in literature as the benchmark. An optimized implementation on the GPU, however, is a challenging problem and is the main focus of the current work. The MC simulation takes samples from the uncertainty distributions of the Resident Space Objects (RSOs) at any time during a time window of interest and outputs the separations at closest approach. Therefore, any uncertainty propagation method may be used and the collision probability is automatically computed as a function of RSO collision radii. Integration using a fixed time step and a quartic interpolation after every Runge Kutta step ensures that no close approaches are missed. Two orders of magnitude speedups over a serial CPU implementation are shown, and speedups improve moderately with higher fidelity dynamics. The tool makes the MC approach tractable on a single workstation, and can be used as a final product, or for verifying surrogate and analytical collision probability methods.

  13. Seismic interpretation using Support Vector Machines implemented on Graphics Processing Units

    Energy Technology Data Exchange (ETDEWEB)

    Kuzma, H A; Rector, J W; Bremer, D

    2006-06-22

    Support Vector Machines (SVMs) estimate lithologic properties of rock formations from seismic data by interpolating between known models using synthetically generated model/data pairs. SVMs are related to kriging and radial basis function neural networks. In our study, we train an SVM to approximate an inverse to the Zoeppritz equations. Training models are sampled from distributions constructed from well-log statistics. Training data is computed via a physically realistic forward modeling algorithm. In our experiments, each training data vector is a set of seismic traces similar to a 2-d image. The SVM returns a model given by a weighted comparison of the new data to each training data vector. The method of comparison is given by a kernel function which implicitly transforms data into a high-dimensional feature space and performs a dot-product. The feature space of a Gaussian kernel is made up of sines and cosines and so is appropriate for band-limited seismic problems. Training an SVM involves estimating a set of weights from the training model/data pairs. It is designed to be an easy problem; at worst it is a quadratic programming problem on the order of the size of the training set. By implementing the slowest part of our SVM algorithm on a graphics processing unit (GPU), we improve the speed of the algorithm by two orders of magnitude. Our SVM/GPU combination achieves results that are similar to those of conventional iterative inversion in fractions of the time.

  14. Impact of climate change on flood characteristics in Brahmaputra basin using a macro-scale distributed hydrological model

    Indian Academy of Sciences (India)

    Shyamal Ghosh; Subashisa Dutta

    2012-06-01

    Being the highest specific discharge river system in the world, the Brahmaputra river experiences a number of long-duration flood waves during the monsoon season annually. In order to assess the flood characteristics at the basin and tributary scales, a physically based macro-scale distributed hydrological model (DHM) has been calibrated and validated for 9 wet years. The model performance has been evaluated in terms of prediction of the flood characteristics such as peak discharge, flood duration, arrival time of flood wave, timing of the peak flow and number of flood waves per season. Future changes in the flood wave characteristics of the basin have been evaluated using the validated model with bias-corrected future-projected meteorological scenario from a regional climate model (RCM). Likelihood analysis of the simulated flow time series reveals that significant increase in both peak discharge and flood duration is expected for both the pre-monsoonal and monsoonal seasons in the basin, but the number of flood waves per season would be reduced. Under the projected climate change scenario, it is expected that there will be more catastrophic floods in the basin.

  15. Nano- and macroscale structural and mechanical properties of in situ synthesized bacterial cellulose/PEO-b-PPO-b-PEO biocomposites.

    Science.gov (United States)

    Tercjak, Agnieszka; Gutierrez, Junkal; Barud, Hernane S; Domeneguetti, Rafael R; Ribeiro, Sidney J L

    2015-02-25

    Highly transparent biocomposite based on bacterial cellulose (BC) mat modified with poly(ethylene oxide-b-propylene oxide-b-ethylene oxide) block copolymer (EPE) were fabricated in situ during biosynthesis of bacterial cellulose in a static culture from Gluconacetobacter xylinum. The effect of the addition to the culture medium of water-soluble EPE block copolymer on structure, morphology, crystallinity, and final properties of the novel biocomposites was investigated at nano- and macroscale. High compatibility between components was confirmed by ATR-FTIR indicating hydrogen bond formation between the OH group of BC and the PEO block of EPE block copolymer. Structural properties of EPE/BC biocomposites showed a strong effect of EPE block copolymer on the morphology of the BC mats. Thus, the increase of the EPE block copolymer content lead to the generation of spherulites of PEO block, clearly visualized using AFM and MO technique, changing crystallinity of the final EPE/BC biocomposites investigated by XRD. Generally, EPE/BC biocomposites maintain thermal stability and mechanical properties of the BC mat being 1 wt % EPE/BC biocomposite material with the best properties. Biosynthesis of EPE/BC composites open new strategy to the utilization of water-soluble block copolymers in the preparation of BC mat based biocomposites with tunable properties.

  16. Three-dimensional macro-scale assessment of regional and temporal wall shear stress characteristics on aortic valve leaflets.

    Science.gov (United States)

    Cao, K; Bukač, M; Sucosky, P

    2016-01-01

    The aortic valve (AV) achieves unidirectional blood flow between the left ventricle and the aorta. Although hemodynamic stresses have been shown to regulate valvular biology, the native wall shear stress (WSS) experienced by AV leaflets remains largely unknown. The objective of this study was to quantify computationally the macro-scale leaflet WSS environment using fluid-structure interaction modeling. An arbitrary Lagrangian-Eulerian approach was implemented to predict valvular flow and leaflet dynamics in a three-dimensional AV geometry subjected to physiologic transvalvular pressure. Local WSS characteristics were quantified in terms of temporal shear magnitude (TSM), oscillatory shear index (OSI) and temporal shear gradient (TSG). The dominant radial WSS predicted on the leaflets exhibited high amplitude and unidirectionality on the ventricularis (TSM>7.50 dyn/cm(2), OSI 325.54 dyn/cm(2) s) but low amplitude and bidirectionality on the fibrosa (TSM 0.38, TSG 0.25). This study provides new insights into the role played by leaflet-blood flow interactions in valvular function and critical hemodynamic stress data for the assessment of the hemodynamic theory of AV disease.

  17. A global fingerprint of macro-scale changes in urban structure from 1999 to 2009

    Science.gov (United States)

    Frolking, Steve; Milliman, Tom; Seto, Karen C.; Friedl, Mark A.

    2013-06-01

    Urban population now exceeds rural population globally, and 60-80% of global energy consumption by households, businesses, transportation, and industry occurs in urban areas. There is growing evidence that built-up infrastructure contributes to carbon emissions inertia, and that investments in infrastructure today have delayed climate cost in the future. Although the United Nations statistics include data on urban population by country and select urban agglomerations, there are no empirical data on built-up infrastructure for a large sample of cities. Here we present the first study to examine changes in the structure of the world’s largest cities from 1999 to 2009. Combining data from two space-borne sensors—backscatter power (PR) from NASA’s SeaWinds microwave scatterometer, and nighttime lights (NL) from NOAA’s defense meteorological satellite program/operational linescan system (DMSP/OLS)—we report large increases in built-up infrastructure stock worldwide and show that cities are expanding both outward and upward. Our results reveal previously undocumented recent and rapid changes in urban areas worldwide that reflect pronounced shifts in the form and structure of cities. Increases in built-up infrastructure are highest in East Asian cities, with Chinese cities rapidly expanding their material infrastructure stock in both height and extent. In contrast, Indian cities are primarily building out and not increasing in verticality. This new dataset will help characterize the structure and form of cities, and ultimately improve our understanding of how cities affect regional-to-global energy use and greenhouse gas emissions.

  18. Reaction hotspots at micro- and macroscales: Challenges in early diagenetic modeling (Invited)

    Science.gov (United States)

    Meile, C. D.

    2010-12-01

    Recent experimental developments, including novel chemical characterizations of organic matter composition, innovative methods for tracing mineral transformations, and a wealth of data generated by the omic revolution, are providing new insights into early diagenetic processes. However, integrating and interpreting this new data in the context of a comprehensive model poses a formidable challenge. Key questions emerging from these studies are: (1) what are characteristics of the key functional groups of microorganisms responsible for the breakdown of organic matter; (2) what are the linkages between composition and reactivity of organic matter; (3) how do biogeochemical conditions, including mineral surfaces, mediate the lability of organic matter. In addition, each of the factors controlling organic matter diagenesis must be interpreted within a spatio-temporal context that likely includes significant compartmentalization at a variety of scales. The joint analysis of the many aspects influencing early diagenesis necessitates that we develop simplified approximations while retaining key environmental characteristics. In this presentation, I will show examples of our work which combines experimental and modeling efforts to assess three key features of early diagenesis: (i) the functioning and integration of the microbes as fundamental metabolic entities, (ii) an assessment of organic matter sources and carbon cycling from a range of experimental data, and (iii) the spatio-temporal context of degradation processes taking place. At the cellular scale, we build on the knowledge gained from genome sequencing of environmentally relevant organisms, e.g. Geobacter sulfurreducens, and the reconstruction of their metabolic networks. When combined with an experimental assessment of substrate uptake kinetics one can integrate these representations of in silico microbial cell models into reactive transport models in order to predict their response and distribution in the field

  19. Self-forming oriented layer slip and macroscale super-low friction of graphene

    Science.gov (United States)

    Song, Hui; Ji, Li; Li, Hongxuan; Wang, Jinqing; Liu, Xiaohong; Zhou, Huidi; Chen, Jianmin

    2017-02-01

    Graphite lubrication is not effective in vacuum, and the failure mechanism is still under debate. Here, we show that graphene as two-dimensional (2D) "graphite paper" can overcome this shortcoming of graphite. Graphene exhibits stable super-low friction in a vacuum environment at the engineering scale because it can self-form a highly ordered lamellar structure on the sliding interface during the friction process owing to its unique 2D nano-effects. Experimental observation of the layer-slip phenomenon on the low-energy-state outside layers provides direct evidence to understand the lubrication mechanism of graphitic materials.

  20. Evaluation of the Synthoil process. Volume III. Unit block flow diagrams for a 100,000 barrel/stream day facility

    Energy Technology Data Exchange (ETDEWEB)

    Salmon, R.; Edwards, M.S.; Ulrich, W.C.

    1977-06-01

    This volume consists of individual block flowsheets for the various units of the Synthoil facility, showing the overall flows into and out of each unit. Material balances for the following units are incomplete because these are proprietary processes and the information was not provided by the respective vendors: Unit 24-Claus Sulfur Plant; Unit 25-Oxygen Plant; Unit 27-Sulfur Plant (Redox Type); and Unit 28-Sour Water Stripper and Ammonia Recovery Plant. The process information in this form was specifically requested by ERDA/FE for inclusion in the final report.

  1. Real-time Graphics Processing Unit Based Fourier Domain Optical Coherence Tomography and Surgical Applications

    Science.gov (United States)

    Zhang, Kang

    2011-12-01

    In this dissertation, real-time Fourier domain optical coherence tomography (FD-OCT) capable of multi-dimensional micrometer-resolution imaging targeted specifically for microsurgical intervention applications was developed and studied. As a part of this work several ultra-high speed real-time FD-OCT imaging and sensing systems were proposed and developed. A real-time 4D (3D+time) OCT system platform using the graphics processing unit (GPU) to accelerate OCT signal processing, the imaging reconstruction, visualization, and volume rendering was developed. Several GPU based algorithms such as non-uniform fast Fourier transform (NUFFT), numerical dispersion compensation, and multi-GPU implementation were developed to improve the impulse response, SNR roll-off and stability of the system. Full-range complex-conjugate-free FD-OCT was also implemented on the GPU architecture to achieve doubled image range and improved SNR. These technologies overcome the imaging reconstruction and visualization bottlenecks widely exist in current ultra-high speed FD-OCT systems and open the way to interventional OCT imaging for applications in guided microsurgery. A hand-held common-path optical coherence tomography (CP-OCT) distance-sensor based microsurgical tool was developed and validated. Through real-time signal processing, edge detection and feed-back control, the tool was shown to be capable of track target surface and compensate motion. The micro-incision test using a phantom was performed using a CP-OCT-sensor integrated hand-held tool, which showed an incision error less than +/-5 microns, comparing to >100 microns error by free-hand incision. The CP-OCT distance sensor has also been utilized to enhance the accuracy and safety of optical nerve stimulation. Finally, several experiments were conducted to validate the system for surgical applications. One of them involved 4D OCT guided micro-manipulation using a phantom. Multiple volume renderings of one 3D data set were

  2. Towards improved parameterization of a macroscale hydrologic model in a discontinuous permafrost boreal forest ecosystem

    Directory of Open Access Journals (Sweden)

    A. Endalamaw

    2017-09-01

    Full Text Available Modeling hydrological processes in the Alaskan sub-arctic is challenging because of the extreme spatial heterogeneity in soil properties and vegetation communities. Nevertheless, modeling and predicting hydrological processes is critical in this region due to its vulnerability to the effects of climate change. Coarse-spatial-resolution datasets used in land surface modeling pose a new challenge in simulating the spatially distributed and basin-integrated processes since these datasets do not adequately represent the small-scale hydrological, thermal, and ecological heterogeneity. The goal of this study is to improve the prediction capacity of mesoscale to large-scale hydrological models by introducing a small-scale parameterization scheme, which better represents the spatial heterogeneity of soil properties and vegetation cover in the Alaskan sub-arctic. The small-scale parameterization schemes are derived from observations and a sub-grid parameterization method in the two contrasting sub-basins of the Caribou Poker Creek Research Watershed (CPCRW in Interior Alaska: one nearly permafrost-free (LowP sub-basin and one permafrost-dominated (HighP sub-basin. The sub-grid parameterization method used in the small-scale parameterization scheme is derived from the watershed topography. We found that observed soil thermal and hydraulic properties – including the distribution of permafrost and vegetation cover heterogeneity – are better represented in the sub-grid parameterization method than the coarse-resolution datasets. Parameters derived from the coarse-resolution datasets and from the sub-grid parameterization method are implemented into the variable infiltration capacity (VIC mesoscale hydrological model to simulate runoff, evapotranspiration (ET, and soil moisture in the two sub-basins of the CPCRW. Simulated hydrographs based on the small-scale parameterization capture most of the peak and low flows, with similar accuracy in both sub

  3. Towards improved parameterization of a macroscale hydrologic model in a discontinuous permafrost boreal forest ecosystem

    Science.gov (United States)

    Endalamaw, Abraham; Bolton, W. Robert; Young-Robertson, Jessica M.; Morton, Don; Hinzman, Larry; Nijssen, Bart

    2017-09-01

    Modeling hydrological processes in the Alaskan sub-arctic is challenging because of the extreme spatial heterogeneity in soil properties and vegetation communities. Nevertheless, modeling and predicting hydrological processes is critical in this region due to its vulnerability to the effects of climate change. Coarse-spatial-resolution datasets used in land surface modeling pose a new challenge in simulating the spatially distributed and basin-integrated processes since these datasets do not adequately represent the small-scale hydrological, thermal, and ecological heterogeneity. The goal of this study is to improve the prediction capacity of mesoscale to large-scale hydrological models by introducing a small-scale parameterization scheme, which better represents the spatial heterogeneity of soil properties and vegetation cover in the Alaskan sub-arctic. The small-scale parameterization schemes are derived from observations and a sub-grid parameterization method in the two contrasting sub-basins of the Caribou Poker Creek Research Watershed (CPCRW) in Interior Alaska: one nearly permafrost-free (LowP) sub-basin and one permafrost-dominated (HighP) sub-basin. The sub-grid parameterization method used in the small-scale parameterization scheme is derived from the watershed topography. We found that observed soil thermal and hydraulic properties - including the distribution of permafrost and vegetation cover heterogeneity - are better represented in the sub-grid parameterization method than the coarse-resolution datasets. Parameters derived from the coarse-resolution datasets and from the sub-grid parameterization method are implemented into the variable infiltration capacity (VIC) mesoscale hydrological model to simulate runoff, evapotranspiration (ET), and soil moisture in the two sub-basins of the CPCRW. Simulated hydrographs based on the small-scale parameterization capture most of the peak and low flows, with similar accuracy in both sub-basins, compared to

  4. Population vulnerability to biannual cholera outbreaks and associated macro-scale drivers in the Bengal Delta.

    Science.gov (United States)

    Akanda, Ali Shafqat; Jutla, Antarpreet S; Gute, David M; Sack, R Bradley; Alam, Munirul; Huq, Anwar; Colwell, Rita R; Islam, Shafiqul

    2013-11-01

    The highly populated floodplains of the Bengal Delta have a long history of endemic and epidemic cholera outbreaks, both coastal and inland. Previous studies have not addressed the spatio-temporal dynamics of population vulnerability related to the influence of underlying large-scale processes. We analyzed spatial and temporal variability of cholera incidence across six surveillance sites in the Bengal Delta and their association with regional hydroclimatic and environmental drivers. More specifically, we use salinity and flood inundation modeling across the vulnerable districts of Bangladesh to test earlier proposed hypotheses on the role of these environmental variables. Our results show strong influence of seasonal and interannual variability in estuarine salinity on spring outbreaks and inland flooding on fall outbreaks. A large segment of the population in the Bengal Delta floodplains remain vulnerable to these biannual cholera transmission mechanisms that provide ecologic and environmental conditions for outbreaks over large geographic regions.

  5. A Behavioral Analysis of the Laboratory Learning Process: Redesigning a Teaching Unit on Recrystallization.

    Science.gov (United States)

    Mulder, T.; Verdonk, A. H.

    1984-01-01

    Reports on a project in which observations of student and teaching assistant behavior were used to redesign a teaching unit on recrystallization. Comments on the instruction manual, starting points for teaching the unit, and list of objectives with related tasks are included. (JN)

  6. Plankton Communities of The Macroscale Anticyclonic Gyre of The South At Lantic Ocean

    Science.gov (United States)

    Piontkovski, S. A.; Finenko, Z. Z.; Kovalev, A. V.; Williams, R.; Gallienne, C. P.; Mishonov, A.; Skryabin, V. N.; Tokarev, Yu. N.; Nikolsky, V. N.

    centre of the gyre shared up to 60% of species with the South Equat orial Current and the Brazil Current, whereas the percentage was less for the Benguela Current (41-50%) and least for the Westerly Wing Current (31 -40%). The mesozooplankton biomass size spectra (calculated in carbon units), exhibited a fairly st able shape, with the slope of the curve getting gently steeper from the eastern periphery of the gyre to its centre. The calculated phyto-to-zooplankton ratio indicat ed that on the scale of the gyre, the mesozooplankton carbon biomass could be represented as the exp onential regression function of the phytoplankton carbon. In general, changes of species composition, size structure, primary production, biomass ratios for different ecological groups, along the current ring to the centre of the gyre, confirmed the development of the plankton community towards a mature stage.

  7. Embedded-Based Graphics Processing Unit Cluster Platform for Multiple Sequence Alignments

    Directory of Open Access Journals (Sweden)

    Jyh-Da Wei

    2017-08-01

    Full Text Available High-end graphics processing units (GPUs, such as NVIDIA Tesla/Fermi/Kepler series cards with thousands of cores per chip, are widely applied to high-performance computing fields in a decade. These desktop GPU cards should be installed in personal computers/servers with desktop CPUs, and the cost and power consumption of constructing a GPU cluster platform are very high. In recent years, NVIDIA releases an embedded board, called Jetson Tegra K1 (TK1, which contains 4 ARM Cortex-A15 CPUs and 192 Compute Unified Device Architecture cores (belong to Kepler GPUs. Jetson Tegra K1 has several advantages, such as the low cost, low power consumption, and high applicability, and it has been applied into several specific applications. In our previous work, a bioinformatics platform with a single TK1 (STK platform was constructed, and this previous work is also used to prove that the Web and mobile services can be implemented in the STK platform with a good cost-performance ratio by comparing a STK platform with the desktop CPU and GPU. In this work, an embedded-based GPU cluster platform will be constructed with multiple TK1s (MTK platform. Complex system installation and setup are necessary procedures at first. Then, 2 job assignment modes are designed for the MTK platform to provide services for users. Finally, ClustalW v2.0.11 and ClustalWtk will be ported to the MTK platform. The experimental results showed that the speedup ratios achieved 5.5 and 4.8 times for ClustalW v2.0.11 and ClustalWtk, respectively, by comparing 6 TK1s with a single TK1. The MTK platform is proven to be useful for multiple sequence alignments.

  8. Parallel flow accumulation algorithms for graphical processing units with application to RUSLE model

    Science.gov (United States)

    Sten, Johan; Lilja, Harri; Hyväluoma, Jari; Westerholm, Jan; Aspnäs, Mats

    2016-04-01

    Digital elevation models (DEMs) are widely used in the modeling of surface hydrology, which typically includes the determination of flow directions and flow accumulation. The use of high-resolution DEMs increases the accuracy of flow accumulation computation, but as a drawback, the computational time may become excessively long if large areas are analyzed. In this paper we investigate the use of graphical processing units (GPUs) for efficient flow accumulation calculations. We present two new parallel flow accumulation algorithms based on dependency transfer and topological sorting and compare them to previously published flow transfer and indegree-based algorithms. We benchmark the GPU implementations against industry standards, ArcGIS and SAGA. With the flow-transfer D8 flow routing model and binary input data, a speed up of 19 is achieved compared to ArcGIS and 15 compared to SAGA. We show that on GPUs the topological sort-based flow accumulation algorithm leads on average to a speedup by a factor of 7 over the flow-transfer algorithm. Thus a total speed up of the order of 100 is achieved. We test the algorithms by applying them to the Revised Universal Soil Loss Equation (RUSLE) erosion model. For this purpose we present parallel versions of the slope, LS factor and RUSLE algorithms and show that the RUSLE erosion results for an area of 12 km x 24 km containing 72 million cells can be calculated in less than a second. Since flow accumulation is needed in many hydrological models, the developed algorithms may find use in many other applications than RUSLE modeling. The algorithm based on topological sorting is particularly promising for dynamic hydrological models where flow accumulations are repeatedly computed over an unchanged DEM.

  9. A New Method Based on Graphics Processing Units for Fast Near-Infrared Optical Tomography.

    Science.gov (United States)

    Jiang, Jingjing; Ahnen, Linda; Kalyanov, Alexander; Lindner, Scott; Wolf, Martin; Majos, Salvador Sanchez

    2017-01-01

    The accuracy of images obtained by Diffuse Optical Tomography (DOT) could be substantially increased by the newly developed time resolved (TR) cameras. These devices result in unprecedented data volumes, which present a challenge to conventional image reconstruction techniques. In addition, many clinical applications require taking photons in air regions like the trachea into account, where the diffusion model fails. Image reconstruction techniques based on photon tracking are mandatory in those cases but have not been implemented so far due to computing demands. We aimed at designing an inversion algorithm which could be implemented on commercial graphics processing units (GPUs) by making use of information obtained with other imaging modalities. The method requires a segmented volume and an approximately uniform value for the reduced scattering coefficient in the volume under study. The complex photon path is reduced to a small number of partial path lengths within each segment resulting in drastically reduced memory usage and computation time. Our approach takes advantage of wavelength normalized data which renders it robust against instrumental biases and skin irregularities which is critical for realistic clinical applications. The accuracy of this method has been assessed with both simulated and experimental inhomogeneous phantoms showing good agreement with target values. The simulation study analyzed a phantom containing a tumor next to an air region. For the experimental test, a segmented cuboid phantom was illuminated by a supercontinuum laser and data were gathered by a state of the art TR camera. Reconstructions were obtained on a GPU-installed computer in less than 2 h. To our knowledge, it is the first time Monte Carlo methods have been successfully used for DOT based on TR cameras. This opens the door to applications such as accurate measurements of oxygenation in neck tumors where the presence of air regions is a problem for conventional approaches.

  10. FLOCKING-BASED DOCUMENT CLUSTERING ON THE GRAPHICS PROCESSING UNIT [Book Chapter

    Energy Technology Data Exchange (ETDEWEB)

    Charles, J S; Patton, R M; Potok, T E; Cui, X

    2008-01-01

    Analyzing and grouping documents by content is a complex problem. One explored method of solving this problem borrows from nature, imitating the fl ocking behavior of birds. Each bird represents a single document and fl ies toward other documents that are similar to it. One limitation of this method of document clustering is its complexity O(n2). As the number of documents grows, it becomes increasingly diffi cult to receive results in a reasonable amount of time. However, fl ocking behavior, along with most naturally inspired algorithms such as ant colony optimization and particle swarm optimization, are highly parallel and have experienced improved performance on expensive cluster computers. In the last few years, the graphics processing unit (GPU) has received attention for its ability to solve highly-parallel and semi-parallel problems much faster than the traditional sequential processor. Some applications see a huge increase in performance on this new platform. The cost of these high-performance devices is also marginal when compared with the price of cluster machines. In this paper, we have conducted research to exploit this architecture and apply its strengths to the document flocking problem. Our results highlight the potential benefi t the GPU brings to all naturally inspired algorithms. Using the CUDA platform from NVIDIA®, we developed a document fl ocking implementation to be run on the NVIDIA® GEFORCE 8800. Additionally, we developed a similar but sequential implementation of the same algorithm to be run on a desktop CPU. We tested the performance of each on groups of news articles ranging in size from 200 to 3,000 documents. The results of these tests were very signifi cant. Performance gains ranged from three to nearly fi ve times improvement of the GPU over the CPU implementation. This dramatic improvement in runtime makes the GPU a potentially revolutionary platform for document clustering algorithms.

  11. Macro-Scale Patterns in Upwelling/Downwelling Activity at North American West Coast

    Science.gov (United States)

    Saldívar-Lucio, Romeo; Di Lorenzo, Emanuele; Nakamura, Miguel; Villalobos, Héctor; Lluch-Cota, Daniel; Del Monte-Luna, Pablo

    2016-01-01

    The seasonal and interannual variability of vertical transport (upwelling/downwelling) has been relatively well studied, mainly for the California Current System, including low-frequency changes and latitudinal heterogeneity. The aim of this work was to identify potentially predictable patterns in upwelling/downwelling activity along the North American west coast and discuss their plausible mechanisms. To this purpose we applied the min/max Autocorrelation Factor technique and time series analysis. We found that spatial co-variation of seawater vertical movements present three dominant low-frequency signals in the range of 33, 19 and 11 years, resembling periodicities of: atmospheric circulation, nodal moon tides and solar activity. Those periodicities might be related to the variability of vertical transport through their influence on dominant wind patterns, the position/intensity of pressure centers and the strength of atmospheric circulation cells (wind stress). The low-frequency signals identified in upwelling/downwelling are coherent with temporal patterns previously reported at the study region: sea surface temperature along the Pacific coast of North America, catch fluctuations of anchovy Engraulis mordax and sardine Sardinops sagax, the Pacific Decadal Oscillation, changes in abundance and distribution of salmon populations, and variations in the position and intensity of the Aleutian low. Since the vertical transport is an oceanographic process with strong biological relevance, the recognition of their spatio-temporal patterns might allow for some reasonable forecasting capacity, potentially useful for marine resources management of the region. PMID:27893826

  12. Situating the default-mode network along a principal gradient of macroscale cortical organization.

    Science.gov (United States)

    Margulies, Daniel S; Ghosh, Satrajit S; Goulas, Alexandros; Falkiewicz, Marcel; Huntenburg, Julia M; Langs, Georg; Bezgin, Gleb; Eickhoff, Simon B; Castellanos, F Xavier; Petrides, Michael; Jefferies, Elizabeth; Smallwood, Jonathan

    2016-11-01

    Understanding how the structure of cognition arises from the topographical organization of the cortex is a primary goal in neuroscience. Previous work has described local functional gradients extending from perceptual and motor regions to cortical areas representing more abstract functions, but an overarching framework for the association between structure and function is still lacking. Here, we show that the principal gradient revealed by the decomposition of connectivity data in humans and the macaque monkey is anchored by, at one end, regions serving primary sensory/motor functions and at the other end, transmodal regions that, in humans, are known as the default-mode network (DMN). These DMN regions exhibit the greatest geodesic distance along the cortical surface-and are precisely equidistant-from primary sensory/motor morphological landmarks. The principal gradient also provides an organizing spatial framework for multiple large-scale networks and characterizes a spectrum from unimodal to heteromodal activity in a functional metaanalysis. Together, these observations provide a characterization of the topographical organization of cortex and indicate that the role of the DMN in cognition might arise from its position at one extreme of a hierarchy, allowing it to process transmodal information that is unrelated to immediate sensory input.

  13. Macroscale and microscale analysis of Anammox in anaerobic rotating biological contactor

    Institute of Scientific and Technical Information of China (English)

    Yongtao Lv; Lei Wang; Xudong Wang; Yongzhe Yang; Zhiying Wang; Jie Li

    2011-01-01

    Inoculated with conventional anaerobic activated sludge,the Anammox process was successfully developed in an anaerobic rotating biological contactor (AnRBC) fed with a low ratio of C/N synthetic wastewater.Operated in a single point feed mode,the AnRBC removed 92.1% (n =126) of the influent N at the highest surface load of 12 g/(m2.day).The biomass increased by 25% and 17.1 g/(m2.day) of maximum N removal surface load was achieved by elevating flow rate with another feed point.Fluorescence in situ hybridization and polymerase chain reaction analysis indicated that the Anammox genus Candidatus Kuenenia stuttgartiensis dominated the community.Both Anammox and denitrifying activity were detected in biofilm by the application of microelectrodes.In the outer layer of the biofilm (0-2500 μm),nitrite and ammonium consumed simultaneously in a ratio of 1.12/1,revealing the occurrence of Anammox.In the inner layer (> 2500 μm),a decrease of nitrate was caused by denitrification in the absence of nitrite and ammonium.

  14. Modification and Validation of an Automotive Data Processing Unit, Compessed Video System, and Communications Equipment

    Energy Technology Data Exchange (ETDEWEB)

    Carter, R.J.

    1997-04-01

    The primary purpose of the "modification and validation of an automotive data processing unit (DPU), compressed video system, and communications equipment" cooperative research and development agreement (CRADA) was to modify and validate both hardware and software, developed by Scientific Atlanta, Incorporated (S-A) for defense applications (e.g., rotary-wing airplanes), for the commercial sector surface transportation domain (i.e., automobiles and trucks). S-A also furnished a state-of-the-art compressed video digital storage and retrieval system (CVDSRS), and off-the-shelf data storage and transmission equipment to support the data acquisition system for crash avoidance research (DASCAR) project conducted by Oak Ridge National Laboratory (ORNL). In turn, S-A received access to hardware and technology related to DASCAR. DASCAR was subsequently removed completely and installation was repeated a number of times to gain an accurate idea of complete installation, operation, and removal of DASCAR. Upon satisfactory completion of the DASCAR construction and preliminary shakedown, ORNL provided NHTSA with an operational demonstration of DASCAR at their East Liberty, OH test facility. The demonstration included an on-the-road demonstration of the entire data acquisition system using NHTSA'S test track. In addition, the demonstration also consisted of a briefing, containing the following: ORNL generated a plan for validating the prototype data acquisition system with regard to: removal of DASCAR from an existing vehicle, and installation and calibration in other vehicles; reliability of the sensors and systems; data collection and transmission process (data integrity); impact on the drivability of the vehicle and obtrusiveness of the system to the driver; data analysis procedures; conspicuousness of the vehicle to other drivers; and DASCAR installation and removal training and documentation. In order to identify any operational problems not captured by the systems

  15. The divining root: moisture-driven responses of roots at the micro- and macro-scale.

    Science.gov (United States)

    Robbins, Neil E; Dinneny, José R

    2015-04-01

    Water is fundamental to plant life, but the mechanisms by which plant roots sense and respond to variations in water availability in the soil are poorly understood. Many studies of responses to water deficit have focused on large-scale effects of this stress, but have overlooked responses at the sub-organ or cellular level that give rise to emergent whole-plant phenotypes. We have recently discovered hydropatterning, an adaptive environmental response in which roots position new lateral branches according to the spatial distribution of available water across the circumferential axis. This discovery illustrates that roots are capable of sensing and responding to water availability at spatial scales far lower than those normally studied for such processes. This review will explore how roots respond to water availability with an emphasis on what is currently known at different spatial scales. Beginning at the micro-scale, there is a discussion of water physiology at the cellular level and proposed sensory mechanisms cells use to detect osmotic status. The implications of these principles are then explored in the context of cell and organ growth under non-stress and water-deficit conditions. Following this, several adaptive responses employed by roots to tailor their functionality to the local moisture environment are discussed, including patterning of lateral root development and generation of hydraulic barriers to limit water loss. We speculate that these micro-scale responses are necessary for optimal functionality of the root system in a heterogeneous moisture environment, allowing for efficient water uptake with minimal water loss during periods of drought. © The Author 2015. Published by Oxford University Press on behalf of the Society for Experimental Biology. All rights reserved. For permissions, please email: journals.permissions@oup.com.

  16. Fabrication of nanoscale to macroscale nickel-multiwall carbon nanotube hybrid materials with tunable material properties

    Science.gov (United States)

    Abdalla, Ahmed M.; Majdi, Tahereh; Ghosh, Suvojit; Puri, Ishwar K.

    2016-12-01

    To utilize their superior properties, multiwall carbon nanotubes (MWNTs) must be manipulated and aligned end-to-end. We describe a nondestructive method to magnetize MWNTs and provide a means to remotely manipulate them through the electroless deposition of magnetic nickel nanoparticles on their surfaces. The noncovalent bonds between Ni nanoparticles and MWNTs produce a Ni-MWNT hybrid material (NiCH) that is electrically conductive and has an enhanced magnetic susceptibility and elastic modulus. Our experiments show that MWNTs can be plated with Ni for Ni:MWNT weight ratios of γ = 1, 7, 14 and 30, to control the material properties. The phase, atom-level, and morphological information from x-ray diffraction, energy dispersive x-ray spectroscopy, scanning electron microscopy, transmission electron microscopy, dark field STEM, and atomic force microscopy clarify the plating process and reveal the mechanical properties of the synthesized material. Ni metalizes at the surface of the Pd catalyst, forming a continuous wavy layer that encapsulates the MWNT surfaces. Subsequently, Ni acts as an autocatalyst, allowing the plating to continue even after the original Pd catalyst has been completely covered. Raising γ increases the coating layer thickness from 10 to 150 nm, which influences the NiCH magnetic properties and tunes its elastic modulus from 12.5 to 58.7 GPa. The NiCH was used to fabricate Ni-MWNT macrostructures and tune their morphologies by changing the direction of an applied magnetic field. Leveraging the hydrophilic Ni-MWNT outer surface, a water-based conductive ink was created and used to print a conductive path that had an electrical resistivity of 5.9 Ω m, illustrating the potential of this material for printing electronic circuits.

  17. Characterizing Micro- and Macro-Scale Seismicity from Bayou Corne, Louisiana

    Science.gov (United States)

    Baig, A. M.; Urbancic, T.; Karimi, S.

    2013-12-01

    parameters for the larger magnitude events. Our presentation is focused on investigating this deformation, characterizing the scaling behaviour and the other source processes by taking advantage of the wide-band afforded to us through the deployment.

  18. 76 FR 11286 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2011-03-01

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: 2011 Adverse Effect Wage Rates, Allowable Charges for Agricultural Workers...

  19. 77 FR 12882 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2012-03-02

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: 2012 Allowable Charges for Agricultural Workers' Meals and Travel...

  20. 78 FR 15741 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2013-03-12

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: 2013 Allowable Charges for Agricultural Workers' Meals and Travel...

  1. 76 FR 79711 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2011-12-22

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: 2012 Adverse Effect Wage Rates AGENCY: Employment and Training...

  2. 78 FR 1259 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2013-01-08

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: 2013 Adverse Effect Wage Rates AGENCY: Employment and Training...

  3. 77 FR 13635 - Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United...

    Science.gov (United States)

    2012-03-07

    ... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF LABOR Employment and Training Administration Labor Certification Process for the Temporary Employment of Aliens in Agriculture in the United States: 2012 Allowable Charges for Agricultural Workers' Meals and Travel...

  4. Unit Testing Using Design by Contract and Equivalence Partitions, Extreme Programming and Agile Processes in Software Engineering

    DEFF Research Database (Denmark)

    Madsen, Per

    2003-01-01

    Extreme Programming [1] and in particular the idea of Unit Testing can improve the quality of the testing process. But still programmers need to do a lot of tiresome manual work writing test cases. If the programmers could get some automatic tool support enforcing the quality of test cases then t...... then the overall quality of the software would improve significantly....

  5. Engineering Encounters: The Cat in the Hat Builds Satellites. A Unit Promoting Scientific Literacy and the Engineering Design Process

    Science.gov (United States)

    Rehmat, Abeera P.; Owens, Marissa C.

    2016-01-01

    This column presents ideas and techniques to enhance your science teaching. This month's issue shares information about a unit promoting scientific literacy and the engineering design process. The integration of engineering with scientific practices in K-12 education can promote creativity, hands-on learning, and an improvement in students'…

  6. High performance direct gravitational N-body simulations on graphics processing units II: An implementation in CUDA

    NARCIS (Netherlands)

    Belleman, R.G.; Bédorf, J.; Portegies Zwart, S.F.

    2008-01-01

    We present the results of gravitational direct N-body simulations using the graphics processing unit (GPU) on a commercial NVIDIA GeForce 8800GTX designed for gaming computers. The force evaluation of the N-body problem is implemented in "Compute Unified Device Architecture" (CUDA) using the GPU to

  7. Unit Testing Using Design by Contract and Equivalence Partitions, Extreme Programming and Agile Processes in Software Engineering

    DEFF Research Database (Denmark)

    Madsen, Per

    2003-01-01

    Extreme Programming [1] and in particular the idea of Unit Testing can improve the quality of the testing process. But still programmers need to do a lot of tiresome manual work writing test cases. If the programmers could get some automatic tool support enforcing the quality of test cases...

  8. High performance direct gravitational N-body simulations on graphics processing units II: An implementation in CUDA

    NARCIS (Netherlands)

    Belleman, R.G.; Bédorf, J.; Portegies Zwart, S.F.

    2008-01-01

    We present the results of gravitational direct N-body simulations using the graphics processing unit (GPU) on a commercial NVIDIA GeForce 8800GTX designed for gaming computers. The force evaluation of the N-body problem is implemented in "Compute Unified Device Architecture" (CUDA) using the GPU to

  9. Massively Parallel Signal Processing using the Graphics Processing Unit for Real-Time Brain?Computer Interface Feature Extraction

    OpenAIRE

    J. Adam Wilson; Williams, Justin C.

    2009-01-01

    The clock speeds of modern computer processors have nearly plateaued in the past 5 years. Consequently, neural prosthetic systems that rely on processing large quantities of data in a short period of time face a bottleneck, in that it may not be possible to process all of the data recorded from an electrode array with high channel counts and bandwidth, such as electrocorticographic grids or other implantable systems. Therefore, in this study a method of using the processing capabilities of a ...

  10. Active Mines and Mineral Processing Plants in the United States in 2003

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — This data set includes mineral and metal operations in the United States. The data represent commodities monitored by the National Minerals Information Center of...

  11. Active Mines and Mineral Processing Plants in the United States in 2003

    Data.gov (United States)

    U.S. Geological Survey, Department of the Interior — This data set includes mineral and metal operations in the United States. The data represent commodities monitored by the National Minerals Information Center of the...

  12. STRATEGIC BUSINESS UNIT – THE CENTRAL ELEMENT OF THE BUSINESS PORTFOLIO STRATEGIC PLANNING PROCESS

    OpenAIRE

    FLORIN TUDOR IONESCU

    2011-01-01

    Over time, due to changes in the marketing environment, generated by the tightening competition, technological, social and political pressures the companies have adopted a new approach, by which the potential businesses began to be treated as strategic business units. A strategic business unit can be considered a part of a company, a product line within a division, and sometimes a single product or brand. From a strategic perspective, the diversified companies represent a collection of busine...

  13. Ab initio nonadiabatic dynamics of multichromophore complexes: a scalable graphical-processing-unit-accelerated exciton framework.

    Science.gov (United States)

    Sisto, Aaron; Glowacki, David R; Martinez, Todd J

    2014-09-16

    ("fragmenting") a molecular system and then stitching it back together. In this Account, we address both of these problems, the first by using graphical processing units (GPUs) and electronic structure algorithms tuned for these architectures and the second by using an exciton model as a framework in which to stitch together the solutions of the smaller problems. The multitiered parallel framework outlined here is aimed at nonadiabatic dynamics simulations on large supramolecular multichromophoric complexes in full atomistic detail. In this framework, the lowest tier of parallelism involves GPU-accelerated electronic structure theory calculations, for which we summarize recent progress in parallelizing the computation and use of electron repulsion integrals (ERIs), which are the major computational bottleneck in both density functional theory (DFT) and time-dependent density functional theory (TDDFT). The topmost tier of parallelism relies on a distributed memory framework, in which we build an exciton model that couples chromophoric units. Combining these multiple levels of parallelism allows access to ground and excited state dynamics for large multichromophoric assemblies. The parallel excitonic framework is in good agreement with much more computationally demanding TDDFT calculations of the full assembly.

  14. The Politics of Process Implementation: Explaining Variations Across Units in a High-Tech Firm

    DEFF Research Database (Denmark)

    Müller, Sune Dueholm

    political strategies for adopting consensus or conflict based leadership styles and management practices to process implementation depending on the actors' response patterns. The developed concepts and propositions contribute to the streams of literature on process innovation and Software Process...

  15. Massively parallel signal processing using the graphics processing unit for real-time brain-computer interface feature extraction

    Directory of Open Access Journals (Sweden)

    J. Adam Wilson

    2009-07-01

    Full Text Available The clock speeds of modern computer processors have nearly plateaued in the past five years. Consequently, neural prosthetic systems that rely on processing large quantities of data in a short period of time face a bottleneck, in that it may not be possible to process all of the data recorded from an electrode array with high channel counts and bandwidth, such as electrocorticographic grids or other implantable systems. Therefore, in this study a method of using the processing capabilities of a graphics card (GPU was developed for real-time neural signal processing of a brain-computer interface (BCI. The NVIDIA CUDA system was used to offload processing to the GPU, which is capable of running many operations in parallel, potentially greatly increasing the speed of existing algorithms. The BCI system records many channels of data, which are processed and translated into a control signal, such as the movement of a computer cursor. This signal processing chain involves computing a matrix-matrix multiplication (i.e., a spatial filter, followed by calculating the power spectral density on every channel using an auto-regressive method, and finally classifying appropriate features for control. In this study, the first two computationally-intensive steps were implemented on the GPU, and the speed was compared to both the current implementation and a CPU-based implementation that uses multi-threading. Significant performance gains were obtained with GPU processing: the current implementation processed 1000 channels in 933 ms, while the new GPU method took only 27 ms, an improvement of nearly 35 times.

  16. The Digital Signal Processing Platform for the Low Frequency Aperture Array: Preliminary Results on the Data Acquisition Unit

    Science.gov (United States)

    Naldi, Giovanni; Mattana, Andrea; Pastore, Sandro; Alderighi, Monica; Zarb Adami, Kristian; Schillirò, Francesco; Aminaei, Amin; Baker, Jeremy; Belli, Carolina; Comoretto, Gianni; Chiarucci, Simone; Chiello, Riccardo; D'Angelo, Sergio; Dalle Mura, Gabriele; De Marco, Andrea; Halsall, Rob; Magro, Alessio; Monari, Jader; Roberts, Matt; Perini, Federico; Poloni, Marco; Pupillo, Giuseppe; Rusticelli, Simone; Schiaffino, Marco; Zaccaro, Emanuele

    A signal processing hardware platform has been developed for the Low Frequency Aperture Array component of the Square Kilometre Array (SKA). The processing board, called an Analog Digital Unit (ADU), is able to acquire and digitize broadband (up to 500MHz bandwidth) radio-frequency streams from 16 dual polarized antennas, channel the data streams and then combine them flexibly as part of a larger beamforming system. It is envisaged that there will be more than 8000 of these signal processing platforms in the first phase of the SKA, so particular attention has been devoted to ensure the design is low-cost and low-power. This paper describes the main features of the data acquisition unit of such a platform and presents preliminary results characterizing its performance.

  17. Methodologies to maximize olefins in process unit of COMPERJ; Metodologias para maximizacao de olefinas nas unidades de processamento do COMPERJ

    Energy Technology Data Exchange (ETDEWEB)

    Santos, Maria Clara de C. dos; Seidl, Peter R.; Guimaraes, Maria Jose O.C. [Universidade Federal do Rio de Janeiro (UFRJ), RJ (Brazil). Escola de Quimica

    2008-07-01

    With the growth of the national and worldwide economy, there has been a considerable increase in demand for polyolefins, thus requiring an increase in the production of basic petrochemicals (primarily ethane and propane). Due the quality of the national oil, heavy and poor in light derivatives, it is necessary investments in processes of conversion of heavy fractions with intent to maximize production these olefins and alternative raw materials for obtaining these petrochemicals The possible alternatives studied were the expansion of the core petrochemical, changes in the refinery processing units and the construction of COMPERJ, the latter being a example of alternative that can change the current scenario. The work aims to the simulation of process units of COMPERJ with the intention of evaluate which solutions like COMPERJ can best meet the growing market of polyolefins. (author)

  18. [Analysis of the safety culture in a Cardiology Unit managed by processes].

    Science.gov (United States)

    Raso-Raso, Rafael; Uris-Selles, Joaquín; Nolasco-Bonmatí, Andreu; Grau-Jornet, Guillermo; Revert-Gandia, Rosa; Jiménez-Carreño, Rebeca; Sánchez-Soriano, Ruth M; Chamorro-Fernández, Carlos I; Marco-Francés, Elvira; Albero-Martínez, José V

    2017-04-03

    The safety culture is one of the requirements to prevent the occurrence of adverse effects, however has not been studied in the field of cardiology. The objective is to evaluate the safety culture in a cardiology unit has implemented and certified an integrated quality and risk management for patient safety system. A transversal observational study was made in 2 consecutive years using the survey "Hospital Survey on Patient Safety Culture" of the "Agency for Healthcare Research and Quality" in its Spanish version (42 items grouped into 12 dimensions) in all staff. The percentage of positive responses in each dimension in 2014 and 2015 were compared, as well as national data and United States data, following the established rules. The overall assessment of a possible 5, was 4.5 in 2014 and 4.7 in 2015. We identified seven dimensions as a fortress. The worst rated were: staffing, management support and teamwork between units. The comparison showed superiority in all dimensions respect to national data, and 8 respect to American data. The safety culture in a Cardiology Unit with an integrated quality and risk management and patient safety system is high, higher than the national in all its dimensions and in most of them respect to the United States. Copyright © 2017 Instituto Nacional de Cardiología Ignacio Chávez. Publicado por Masson Doyma México S.A. All rights reserved.

  19. 40 CFR 63.1104 - Process vents from continuous unit operations: applicability assessment procedures and methods.

    Science.gov (United States)

    2010-07-01

    ... vent. (iv) Design analysis based on accepted chemical engineering principles, measurable process... be monitored to ensure the process vent is operated in conformance with its design or process and... 40 Protection of Environment 10 2010-07-01 2010-07-01 false Process vents from continuous...

  20. Improved Inventory Models for the United States Coast Guard Requirements Determination Process

    Science.gov (United States)

    1993-10-01

    Trepp present two versions of 5-24 a multi-item, supply availability safety level model.9 They used the Method of Lagrange Multipliers to solve for ki...the safety-level factor for item i. The Presutti and Trepp models address units backordered. To convert their unit models to requisition models, that...requisition size. R sE-Tnw MODELING In their paper, Presutti and Trepp also gave two ve.sions of a multi-item, response-time, safety-level model

  1. Anaerobic bio-digestion of concentrate obtained in the process of ultra filtration of effluents from tilapia processing unit

    Directory of Open Access Journals (Sweden)

    Milena Alves de Souza

    2012-02-01

    Full Text Available The objective of the present study was to evaluate the efficiency of the process of biodigestion of the protein concentrate resulting from the ultrafiltration of the effluent from a slaughterhouse freezer of Nile tilapia. Bench digesters were used with excrements and water (control in comparison with a mixture of cattle manure and effluent from the stages of filleting and bleeding of tilapias. The effluent obtained in the continuous process (bleeding + filleting was the one with highest accumulated population from the 37th day, as well as greatest daily production. Gases composition did not differ between the protein concentrates, but the gas obtained with the use of the effluent from the filleting stage presented highest methane gas average (78.05% in comparison with those obtained in the bleeding stage (69.95% and in the continuous process (70.02% or by the control method (68.59%.

  2. Investigation of the Dynamic Melting Process in a Thermal Energy Storage Unit Using a Helical Coil Heat Exchanger

    Directory of Open Access Journals (Sweden)

    Xun Yang

    2017-08-01

    Full Text Available In this study, the dynamic melting process of the phase change material (PCM in a vertical cylindrical tube-in-tank thermal energy storage (TES unit was investigated through numerical simulations and experimental measurements. To ensure good heat exchange performance, a concentric helical coil was inserted into the TES unit to pipe the heat transfer fluid (HTF. A numerical model using the computational fluid dynamics (CFD approach was developed based on the enthalpy-porosity method to simulate the unsteady melting process including temperature and liquid fraction variations. Temperature measurements using evenly spaced thermocouples were conducted, and the temperature variation at three locations inside the TES unit was recorded. The effects of the HTF inlet parameters were investigated by parametric studies with different temperatures and flow rate values. Reasonably good agreement was achieved between the numerical prediction and the temperature measurement, which confirmed the numerical simulation accuracy. The numerical results showed the significance of buoyancy effect for the dynamic melting process. The system TES performance was very sensitive to the HTF inlet temperature. By contrast, no apparent influences can be found when changing the HTF flow rates. This study provides a comprehensive solution to investigate the heat exchange process of the TES system using PCM.

  3. Single-unit studies of visual motion processing in cat extrastriate areas

    NARCIS (Netherlands)

    Vajda, Ildiko

    2003-01-01

    Motion vision has high survival value and is a fundamental property of all visual systems. The old Greeks already studied motion vision, but the physiological basis of it first came under scrutiny in the late nineteenth century. Later, with the introduction of single-cell (single-unit) recordings ar

  4. Future forest aboveground carbon dynamics in the central United States: the importance of forest demographic processes

    Science.gov (United States)

    Wenchi Jin; Hong S. He; Frank R. Thompson; Wen J. Wang; Jacob S. Fraser; Stephen R. Shifley; Brice B. Hanberry; William D. Dijak

    2017-01-01

    The Central Hardwood Forest (CHF) in the United States is currently a major carbon sink, there are uncertainties in how long the current carbon sink will persist and if the CHF will eventually become a carbon source. We used a multi-model ensemble to investigate aboveground carbon density of the CHF from 2010 to 2300 under current climate. Simulations were done using...

  5. An Investigation Into the Feasibility of Merging Three Technical Processing Operations Into One Central Unit.

    Science.gov (United States)

    Burns, Robert W., Jr.

    Three contiguous schools in the upper midwest--a teacher's training college and a private four-year college in one state, and a land-grant university in another--were studied to see if their libraries could merge one of their major divisions--technical services--into a single administrative unit. Potential benefits from such a merger were felt to…

  6. How semantics can improve engineering processes: A case of units of measure and quantities

    NARCIS (Netherlands)

    Rijgersberg, H.; Wigham, M.; Top, J.L.

    2011-01-01

    Science and engineering heavily depend on the ability to share data and models. Ontologies provide such standards. We have found that existing ontologies only define subsets of the necessary concepts and relations. We therefore propose a new ontology, called OM (Ontology of units of Measure and rela

  7. Single-unit studies of visual motion processing in cat extrastriate areas

    NARCIS (Netherlands)

    Vajda, Ildiko

    2003-01-01

    Motion vision has high survival value and is a fundamental property of all visual systems. The old Greeks already studied motion vision, but the physiological basis of it first came under scrutiny in the late nineteenth century. Later, with the introduction of single-cell (single-unit) recordings ar

  8. Ileus development in the trauma/surgical intensive care unit: a process improvement evaluation.

    Science.gov (United States)

    Phipps, Marcy; Bush, Jeffrey A; Buhrow, Dianne; Tittle, Mary B; Singh, Deepak; Harcombe, Julianne; Riddle, Evanthia

    2011-01-01

    Ileus development has been associated with a wide range of complications among hospitalized patients, ranging from increased patient pain and discomfort to malnutrition, aspiration, delayed rehabilitation, and sepsis. This article examines factors that appeared to correlate with an increase in ileus development among patients in a trauma/surgical intensive care unit, with the goal of preventing the condition through nursing practice changes.

  9. How semantics can improve engineering processes: A case of units of measure and quantities

    NARCIS (Netherlands)

    Rijgersberg, H.; Wigham, M.; Top, J.L.

    2011-01-01

    Science and engineering heavily depend on the ability to share data and models. Ontologies provide such standards. We have found that existing ontologies only define subsets of the necessary concepts and relations. We therefore propose a new ontology, called OM (Ontology of units of Measure and

  10. Polycarbonate modified with crystallisable bis-ester tetra-amide units in a reaction extrusion process

    NARCIS (Netherlands)

    Zuiderduin, W.C.J.; Gaymans, R.J.

    2008-01-01

    Dry blends of polycarbonate (PC) and a bis-ester tetra-amide were extruded at 305 °C with a mini twin screw extruder. The bis-ester tetra-amide diamide (T6T6T-dimethyl) was composed of two and a half repeat units of Nylon 6,T and had methyl ester endgroups. During the extrusion, a

  11. Grey water treatment by a continuous process of an electrocoagulation unit and a submerged membrane bioreactor system

    KAUST Repository

    Bani-Melhem, Khalid

    2012-08-01

    This paper presents the performance of an integrated process consisting of an electro-coagulation (EC) unit and a submerged membrane bioreactor (SMBR) technology for grey water treatment. For comparison purposes, another SMBR process without electrocoagulation (EC) was operated in parallel with both processes operated under constant transmembrane pressure for 24. days in continuous operation mode. It was found that integrating EC process with SMBR (EC-SMBR) was not only an effective method for grey water treatment but also for improving the overall performance of the membrane filtration process. EC-SMBR process achieved up to 13% reduction in membrane fouling compared to SMBR without electrocoagulation. High average percent removals were attained by both processes for most wastewater parameters studied. The results demonstrated that EC-SMBR performance slightly exceeded that of SMBR for COD, turbidity, and colour. Both processes produced effluent free of suspended solids, and faecal coliforms were nearly (100%) removed in both processes. A substantial improvement was achieved in removal of phosphate in the EC-SMBR process. However, ammonia nitrogen was removed more effectively by the SMBR only. Accordingly, the electrolysis condition in the EC-SMBR process should be optimized so as not to impede biological treatment. © 2012 Elsevier B.V.

  12. United States Climate Reference Network (USCRN) Processed Data from the Version 2 USCRN Database

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — USCRN Processed data are interpreted values and derived geophysical parameters processed from raw data by the USCRN Team. Data were interpreted and ingested into a...

  13. Large eddy simulations of turbulent flows on graphics processing units: Application to film-cooling flows

    Science.gov (United States)

    Shinn, Aaron F.

    Computational Fluid Dynamics (CFD) simulations can be very computationally expensive, especially for Large Eddy Simulations (LES) and Direct Numerical Simulations (DNS) of turbulent ows. In LES the large, energy containing eddies are resolved by the computational mesh, but the smaller (sub-grid) scales are modeled. In DNS, all scales of turbulence are resolved, including the smallest dissipative (Kolmogorov) scales. Clusters of CPUs have been the standard approach for such simulations, but an emerging approach is the use of Graphics Processing Units (GPUs), which deliver impressive computing performance compared to CPUs. Recently there has been great interest in the scientific computing community to use GPUs for general-purpose computation (such as the numerical solution of PDEs) rather than graphics rendering. To explore the use of GPUs for CFD simulations, an incompressible Navier-Stokes solver was developed for a GPU. This solver is capable of simulating unsteady laminar flows or performing a LES or DNS of turbulent ows. The Navier-Stokes equations are solved via a fractional-step method and are spatially discretized using the finite volume method on a Cartesian mesh. An immersed boundary method based on a ghost cell treatment was developed to handle flow past complex geometries. The implementation of these numerical methods had to suit the architecture of the GPU, which is designed for massive multithreading. The details of this implementation will be described, along with strategies for performance optimization. Validation of the GPU-based solver was performed for fundamental bench-mark problems, and a performance assessment indicated that the solver was over an order-of-magnitude faster compared to a CPU. The GPU-based Navier-Stokes solver was used to study film-cooling flows via Large Eddy Simulation. In modern gas turbine engines, the film-cooling method is used to protect turbine blades from hot combustion gases. Therefore, understanding the physics of

  14. Production of advanced biofuels: Co-processing of upgraded pyrolysis oil in standard refinery units

    NARCIS (Netherlands)

    Miguel Mercader, de F.; Groeneveld, M.J.; Kersten, S.R.A.; Way, N.W.J.; Schaverien, C.J.; Hogendoorn, J.A.

    2010-01-01

    One of the possible process options for the production of advanced biofuels is the co-processing of upgraded pyrolysis oil in standard refineries. The applicability of hydrodeoxygenation (HDO) was studied as a pyrolysis oil upgrading step to allow FCC co-processing. Different HDO reaction end temper

  15. Graphics Processing Unit-Accelerated Code for Computing Second-Order Wiener Kernels and Spike-Triggered Covariance.

    Science.gov (United States)

    Mano, Omer; Clark, Damon A

    2017-01-01

    Sensory neuroscience seeks to understand and predict how sensory neurons respond to stimuli. Nonlinear components of neural responses are frequently characterized by the second-order Wiener kernel and the closely-related spike-triggered covariance (STC). Recent advances in data acquisition have made it increasingly common and computationally intensive to compute second-order Wiener kernels/STC matrices. In order to speed up this sort of analysis, we developed a graphics processing unit (GPU)-accelerated module that computes the second-order Wiener kernel of a system's response to a stimulus. The generated kernel can be easily transformed for use in standard STC analyses. Our code speeds up such analyses by factors of over 100 relative to current methods that utilize central processing units (CPUs). It works on any modern GPU and may be integrated into many data analysis workflows. This module accelerates data analysis so that more time can be spent exploring parameter space and interpreting data.

  16. Trends in lumber processing in the western United States. Part I: board foot Scribner volume per cubic foot of timber

    Science.gov (United States)

    Charles E. Keegan; Todd A. Morgan; Keith A. Blatner; Jean M. Daniels

    2010-01-01

    This article describes trends in board foot Scribner volume per cubic foot of timber for logs processed by sawmills in the western United States. Board foot to cubic foot (BF/CF) ratios for the period from 2000 through 2006 ranged from 3.70 in Montana to 5.71 in the Four Corners Region (Arizona, Colorado, New Mexico, and Utah). Sawmills in the Four Corners Region,...

  17. The United States Government Interagency Process and the Failure of Institution Building in Iraq

    Science.gov (United States)

    2008-06-12

    LTG Martin Dempsey, to make the case for an increase in the investment by the United States Government (USG) in the development of a self-sufficient...the exception of 5 Feil , Scott R. “The Failure of Incrementalism: Interagency Coordination Challenges and Responses.” In The Interagency and...society invested in establishing a stable democratic, and economically viable Iraq.” (SIGIR-06-045, 2007) 35 Interview, Scott Feil , 27 November 2007. 22

  18. Polycarbonate modified with crystallisable bis-ester tetra-amide units in a reaction extrusion process

    NARCIS (Netherlands)

    Zuiderduin, W.C.J.; Gaymans, R.J.

    2008-01-01

    Dry blends of polycarbonate (PC) and a bis-ester tetra-amide were extruded at 305 °C with a mini twin screw extruder. The bis-ester tetra-amide diamide (T6T6T-dimethyl) was composed of two and a half repeat units of Nylon 6,T and had methyl ester endgroups. During the extrusion, a trans-esterificati

  19. Geostatistical analysis to identify hydrogeochemical processes in complex aquifers: a case study (Aguadulce unit, Almeria, SE Spain).

    Science.gov (United States)

    Daniele, Linda; Pulido Bosch, Antonio; Vallejos, Angela; Molina, Luis

    2008-06-01

    The Aguadulce aquifer unit in southeastern Spain is a complex hydrogeological system because of the varied lithology of the aquifer strata and the variability of the processes that can take place within the unit. Factorial analysis of the data allowed the number of variables to be reduced to 3 factors, which were found to be related to such physico-chemical processes as marine intrusion and leaching of saline deposits. Variographic analysis was applied to these factors, culminating in a study of spatial distribution using ordinary kriging. Mapping of the factors allowed rapid differentiation of some of the processes that affect the waters of the Gador carbonate aquifer within the Aguadulce unit, without the need to recur to purely hydrogeochemical techniques. The results indicate the existence of several factors related to salinity: marine intrusion, paleowaters, and/or leaching of marls and evaporitic deposits. The techniques employed are effective, and the results conform to those obtained using hydrogeochemical methods (vertical records of conductivity and temperature, ion ratios, and others). The findings of this study confirm that the application of such analytical methods can provide a useful assessment of factors affecting groundwater composition.

  20. Compressed sensing reconstruction for whole-heart imaging with 3D radial trajectories: a graphics processing unit implementation.

    Science.gov (United States)

    Nam, Seunghoon; Akçakaya, Mehmet; Basha, Tamer; Stehning, Christian; Manning, Warren J; Tarokh, Vahid; Nezafat, Reza

    2013-01-01

    A disadvantage of three-dimensional (3D) isotropic acquisition in whole-heart coronary MRI is the prolonged data acquisition time. Isotropic 3D radial trajectories allow undersampling of k-space data in all three spatial dimensions, enabling accelerated acquisition of the volumetric data. Compressed sensing (CS) reconstruction can provide further acceleration in the acquisition by removing the incoherent artifacts due to undersampling and improving the image quality. However, the heavy computational overhead of the CS reconstruction has been a limiting factor for its application. In this article, a parallelized implementation of an iterative CS reconstruction method for 3D radial acquisitions using a commercial graphics processing unit is presented. The execution time of the graphics processing unit-implemented CS reconstruction was compared with that of the C++ implementation, and the efficacy of the undersampled 3D radial acquisition with CS reconstruction was investigated in both phantom and whole-heart coronary data sets. Subsequently, the efficacy of CS in suppressing streaking artifacts in 3D whole-heart coronary MRI with 3D radial imaging and its convergence properties were studied. The CS reconstruction provides improved image quality (in terms of vessel sharpness and suppression of noise-like artifacts) compared with the conventional 3D gridding algorithm, and the graphics processing unit implementation greatly reduces the execution time of CS reconstruction yielding 34-54 times speed-up compared with C++ implementation. Copyright © 2012 Wiley Periodicals, Inc.

  1. Low cost solar array project production process and equipment task: A Module Experimental Process System Development Unit (MEPSDU)

    Science.gov (United States)

    1981-01-01

    Several major modifications were made to the design presented at the PDR. The frame was deleted in favor of a "frameless" design which will provide a substantially improved cell packing factor. Potential shaded cell damage resulting from operation into a short circuit can be eliminated by a change in the cell series/parallel electrical interconnect configuration. The baseline process sequence defined for the MEPSON was refined and equipment design and specification work was completed. SAMICS cost analysis work accelerated, format A's were prepared and computer simulations completed. Design work on the automated cell interconnect station was focused on bond technique selection experiments.

  2. Leadership development revisited : an assessment of midshipmen learning processes at the United States Naval Academy

    OpenAIRE

    Kennedy, Robert

    1998-01-01

    This thesis attempts to answer the following research questions: (1) How do midshipmen learn about leadership? (2) How do officer learning processes differ from midshipmen learning processes? The learning literature and the leadership literature identified the following leadership learning processes: (1) experience, (2) observation, (3) reflection, (4) experimentation, (5) interpersonal interactions, (6) organizational cultural, (7) formal instruction (i.e., conceptualization), and (8) self-d...

  3. Investigation of scale effects and directionality dependence on friction and adhesion of human hair using AFM and macroscale friction test apparatus.

    Science.gov (United States)

    LaTorre, Carmen; Bhushan, Bharat

    2006-01-01

    Macroscale testing of human hair tribological properties has been widely used to aid in the development of better shampoos and conditioners. Recently, literature has focused on using the atomic force microscope (AFM) to study surface roughness, coefficient of friction, adhesive force, and wear (tribological properties) on the nanoscale in order to increase understanding about how shampoos and conditioners interact with the hair cuticle. Since there are both similarities and differences when comparing the tribological trends at both scales, it is thus recognized that scale effects are an important aspect of studying the tribology of hair. However, no microscale tribological data for hair exists in literature. This is unfortunate because many interactions between hair-skin, hair-comb, and hair-hair contact takes place at microasperities ranging from a few mum to hundreds of mum. Thus, to bridge the gap between the macro- and nanoscale data, as well as to gain a full understanding of the mechanisms behind the trends, it is now worthwhile to look at hair tribology on the microscale. Presented in this paper are coefficient of friction and adhesive force data on various scales for virgin and chemically damaged hair, both with and without conditioner treatment. Macroscale coefficient of friction was determined using a traditional friction test apparatus. Microscale and nanoscale tribological characterization was performed with AFM tips of various radii. The nano-, micro-, and macroscale trends are compared and the mechanisms behind the scale effects are discussed. Since the coefficient of friction changes drastically (on any scale) depending on whether the direction of motion is along or against the cuticle scales, the directionality dependence and responsible mechanisms are discussed.

  4. Investigation of scale effects and directionality dependence on friction and adhesion of human hair using AFM and macroscale friction test apparatus

    Energy Technology Data Exchange (ETDEWEB)

    LaTorre, Carmen [Nanotribology Laboratory for Information Storage and MEMS/NEMS (NLIM), Ohio State University, Suite 255, 650 Ackerman Road, Columbus, OH 43202 (United States); Bhushan, Bharat [Nanotribology Laboratory for Information Storage and MEMS/NEMS (NLIM), Ohio State University, Suite 255, 650 Ackerman Road, Columbus, OH 43202 (United States)]. E-mail: bhushan.2@osu.edu

    2006-06-15

    Macroscale testing of human hair tribological properties has been widely used to aid in the development of better shampoos and conditioners. Recently, literature has focused on using the atomic force microscope (AFM) to study surface roughness, coefficient of friction, adhesive force, and wear (tribological properties) on the nanoscale in order to increase understanding about how shampoos and conditioners interact with the hair cuticle. Since there are both similarities and differences when comparing the tribological trends at both scales, it is thus recognized that scale effects are an important aspect of studying the tribology of hair. However, no microscale tribological data for hair exists in literature. This is unfortunate because many interactions between hair-skin, hair-comb, and hair-hair contact takes place at microasperities ranging from a few {mu}m to hundreds of {mu}m. Thus, to bridge the gap between the macro- and nanoscale data, as well as to gain a full understanding of the mechanisms behind the trends, it is now worthwhile to look at hair tribology on the microscale. Presented in this paper are coefficient of friction and adhesive force data on various scales for virgin and chemically damaged hair, both with and without conditioner treatment. Macroscale coefficient of friction was determined using a traditional friction test apparatus. Microscale and nanoscale tribological characterization was performed with AFM tips of various radii. The nano-, micro-, and macroscale trends are compared and the mechanisms behind the scale effects are discussed. Since the coefficient of friction changes drastically (on any scale) depending on whether the direction of motion is along or against the cuticle scales, the directionality dependence and responsible mechanisms are discussed.

  5. Using real time process measurements to reduce catheter related bloodstream infections in the intensive care unit

    OpenAIRE

    Wall, R; Ely, E; Elasy, T; Dittus, R; Foss, J.; Wilkerson, K; Speroff, T

    2005-01-01

    

Problem: Measuring a process of care in real time is essential for continuous quality improvement (CQI). Our inability to measure the process of central venous catheter (CVC) care in real time prevented CQI efforts aimed at reducing catheter related bloodstream infections (CR-BSIs) from these devices.

  6. Development of High-Power Hall Thruster Power Processing Units at NASA GRC

    Science.gov (United States)

    Pinero, Luis R.; Bozak, Karin E.; Santiago, Walter; Scheidegger, Robert J.; Birchenough, Arthur G.

    2015-01-01

    NASA GRC successfully designed, built and tested four different power processor concepts for high power Hall thrusters. Each design satisfies unique goals including the evaluation of a novel silicon carbide semiconductor technology, validation of innovative circuits to overcome the problems with high input voltage converter design, development of a direct-drive unit to demonstrate potential benefits, or simply identification of lessonslearned from the development of a PPU using a conventional design approach. Any of these designs could be developed further to satisfy NASA's needs for high power electric propulsion in the near future.

  7. Functional redundancy and the process of professionalization: the case of registered nurses in the United States.

    Science.gov (United States)

    Levi, M

    1980-01-01

    Registered nurses have been attempting to achieve professional status for nearly a century. Historical investigation of their efforts in the United States and a case study of the 1976 Seattle Nurses' strike indicate major obstacles to the professionalizing project. The most important of these are the inability of the nurses to control the labor supply, and their failure to define or monopolize a distinct set of tasks. One result is functional redundancy: there is no job nurses perform that is not also performed by some other occupation.

  8. Parallelized multi-graphics processing unit framework for high-speed Gabor-domain optical coherence microscopy.

    Science.gov (United States)

    Tankam, Patrice; Santhanam, Anand P; Lee, Kye-Sung; Won, Jungeun; Canavesi, Cristina; Rolland, Jannick P

    2014-07-01

    Gabor-domain optical coherence microscopy (GD-OCM) is a volumetric high-resolution technique capable of acquiring three-dimensional (3-D) skin images with histological resolution. Real-time image processing is needed to enable GD-OCM imaging in a clinical setting. We present a parallelized and scalable multi-graphics processing unit (GPU) computing framework for real-time GD-OCM image processing. A parallelized control mechanism was developed to individually assign computation tasks to each of the GPUs. For each GPU, the optimal number of amplitude-scans (A-scans) to be processed in parallel was selected to maximize GPU memory usage and core throughput. We investigated five computing architectures for computational speed-up in processing 1000×1000 A-scans. The proposed parallelized multi-GPU computing framework enables processing at a computational speed faster than the GD-OCM image acquisition, thereby facilitating high-speed GD-OCM imaging in a clinical setting. Using two parallelized GPUs, the image processing of a 1×1×0.6  mm3 skin sample was performed in about 13 s, and the performance was benchmarked at 6.5 s with four GPUs. This work thus demonstrates that 3-D GD-OCM data may be displayed in real-time to the examiner using parallelized GPU processing.

  9. Parallelized multi–graphics processing unit framework for high-speed Gabor-domain optical coherence microscopy

    Science.gov (United States)

    Tankam, Patrice; Santhanam, Anand P.; Lee, Kye-Sung; Won, Jungeun; Canavesi, Cristina; Rolland, Jannick P.

    2014-01-01

    Abstract. Gabor-domain optical coherence microscopy (GD-OCM) is a volumetric high-resolution technique capable of acquiring three-dimensional (3-D) skin images with histological resolution. Real-time image processing is needed to enable GD-OCM imaging in a clinical setting. We present a parallelized and scalable multi-graphics processing unit (GPU) computing framework for real-time GD-OCM image processing. A parallelized control mechanism was developed to individually assign computation tasks to each of the GPUs. For each GPU, the optimal number of amplitude-scans (A-scans) to be processed in parallel was selected to maximize GPU memory usage and core throughput. We investigated five computing architectures for computational speed-up in processing 1000×1000 A-scans. The proposed parallelized multi-GPU computing framework enables processing at a computational speed faster than the GD-OCM image acquisition, thereby facilitating high-speed GD-OCM imaging in a clinical setting. Using two parallelized GPUs, the image processing of a 1×1×0.6  mm3 skin sample was performed in about 13 s, and the performance was benchmarked at 6.5 s with four GPUs. This work thus demonstrates that 3-D GD-OCM data may be displayed in real-time to the examiner using parallelized GPU processing. PMID:24695868

  10. Efficient particle-in-cell simulation of auroral plasma phenomena using a CUDA enabled graphics processing unit

    Science.gov (United States)

    Sewell, Stephen

    This thesis introduces a software framework that effectively utilizes low-cost commercially available Graphic Processing Units (GPUs) to simulate complex scientific plasma phenomena that are modeled using the Particle-In-Cell (PIC) paradigm. The software framework that was developed conforms to the Compute Unified Device Architecture (CUDA), a standard for general purpose graphic processing that was introduced by NVIDIA Corporation. This framework has been verified for correctness and applied to advance the state of understanding of the electromagnetic aspects of the development of the Aurora Borealis and Aurora Australis. For each phase of the PIC methodology, this research has identified one or more methods to exploit the problem's natural parallelism and effectively map it for execution on the graphic processing unit and its host processor. The sources of overhead that can reduce the effectiveness of parallelization for each of these methods have also been identified. One of the novel aspects of this research was the utilization of particle sorting during the grid interpolation phase. The final representation resulted in simulations that executed about 38 times faster than simulations that were run on a single-core general-purpose processing system. The scalability of this framework to larger problem sizes and future generation systems has also been investigated.

  11. UNITED STATES DEPARTMENT OF ENERGY OFFICE OF ENVIRONMENTAL MANAGEMENT WASTE PROCESSING ANNUAL TECHNOLOGY DEVELOPMENT REPORT 2008

    Energy Technology Data Exchange (ETDEWEB)

    Bush, S.

    2009-11-05

    The Office of Waste Processing identifies and reduces engineering and technical risks and uncertainties of the waste processing programs and projects of the Department of Energy's Environmental Management (EM) mission through the timely development of solutions to technical issues. The risks, and actions taken to mitigate those risks, are determined through technology readiness assessments, program reviews, technology information exchanges, external technical reviews, technical assistance, and targeted technology development and deployment. The Office of Waste Processing works with other DOE Headquarters offices and project and field organizations to proactively evaluate technical needs, identify multi-site solutions, and improve the technology and engineering associated with project and contract management. Participants in this program are empowered with the authority, resources, and training to implement their defined priorities, roles, and responsibilities. The Office of Waste Processing Multi-Year Program Plan (MYPP) supports the goals and objectives of the U.S. Department of Energy (DOE) - Office of Environmental Management Engineering and Technology Roadmap by providing direction for technology enhancement, development, and demonstration that will lead to a reduction of technical risks and uncertainties in EM waste processing activities. The MYPP summarizes the program areas and the scope of activities within each program area proposed for the next five years to improve safety and reduce costs and environmental impacts associated with waste processing; authorized budget levels will impact how much of the scope of activities can be executed, on a year-to-year basis. Waste Processing Program activities within the Roadmap and the MYPP are described in these seven program areas: (1) Improved Waste Storage Technology; (2) Reliable and Efficient Waste Retrieval Technologies; (3) Enhanced Tank Closure Processes; (4) Next-Generation Pretreatment Solutions; (5

  12. A Shipping Container-Based Sterile Processing Unit for Low Resources Settings

    OpenAIRE

    Jean Boubour; Katherine Jenson; Hannah Richter; Josiah Yarbrough; Maria Oden, Z.; Schuler, Douglas A.

    2016-01-01

    Deficiencies in the sterile processing of medical instruments contribute to poor outcomes for patients, such as surgical site infections, longer hospital stays, and deaths. In low resources settings, such as some rural and semi-rural areas and secondary and tertiary cities of developing countries, deficiencies in sterile processing are accentuated due to the lack of access to sterilization equipment, improperly maintained and malfunctioning equipment, lack of power to operate equipment, poor ...

  13. Application of multivariate analysis toward biotech processes: case study of a cell-culture unit operation.

    Science.gov (United States)

    Kirdar, Alime Ozlem; Conner, Jeremy S; Baclaski, Jeffrey; Rathore, Anurag S

    2007-01-01

    This paper examines the feasibility of using multivariate data analysis (MVDA) for supporting some of the key activities that are required for successful manufacturing of biopharmaceutical products. These activities include scale-up, process comparability, process characterization, and fault diagnosis. Multivariate data analysis and modeling were performed using representative data from small-scale (2 L) and large-scale (2000 L) batches of a cell-culture process. Several input parameters (pCO2, pO2, glucose, pH, lactate, ammonium ions) and output parameters (purity, viable cell density, viability, osmolality) were evaluated in this analysis. Score plots, loadings plots, and VIP plots were utilized for assessing scale-up and comparability of the cell-culture process. Batch control charts were found to be useful for fault diagnosis during routine manufacturing. Finally, observations made from reviewing VIP plots were found to be in agreement with conclusions from process characterization studies demonstrating the effectiveness of MVDA as a tool for extracting process knowledge.

  14. Distributed Leadership in Organizational Change Processes: A Qualitative Study in Public Hospital Units

    DEFF Research Database (Denmark)

    Kjeldsen, Anne Mette; Jonasson, Charlotte; Ovesen, Maria

    2015-01-01

    This paper proposes that the emergence and boundaries of distributed leadership (DL) are developed in a dynamic interplay with planned as well as emergent organizational change. The empirical findings are based on a qualitative, longitudinal case study with interviews conducted at two different...... hospital units in the context of a larger hospital merger within the Danish health care system. The paper adds to prior studies arguing that DL contributes positively to planned organizational change by instead providing examples of how ongoing changes in contextual conditions such as routine....../non-routine, various goals, and organizational planning affect a simultaneous widening or restriction of the local DL. In return, such local DL also leads to ongoing changes in the form of novel work routines for improved collaboration. Moreover, the findings show that restrictions of DL are in some cases considered...

  15. Experimental Investigation of a Vertical Tubular Desalination Unit Using Humidification Dehumidification Process

    Institute of Scientific and Technical Information of China (English)

    熊日华; 王世昌; 王志; 解利昕; 李凭力; 朱爱梅

    2005-01-01

    A vertical tubular desalination unit with shell and tube structure was built to perform humidification and dehumidification simultaneously on the tube and shell side of the column, respectively. The effects of several operating conditions on the productivity and thermal efficiency of the column were investigated. The results show that both the productivity and thermal efficiency of the column enhance with the elevation of the inlet water temperature. The flow rates of water and carrier gas both have optimal operating ranges, which are 10-30 kg·h-1 and 4-7kg·h-1 for the present column, respectively. Meanwhile, the increase of external steam flow rate will promote the productivity of the column but reduce its thermal efficiency.

  16. A Survey Study of Institutional Review Board Thought Processes in the United States and South Korea

    Directory of Open Access Journals (Sweden)

    Si-Kyung Jung

    2012-09-01

    Full Text Available Introduction: In the last several decades, South Korea has rapidly adopted Western customs and practices. Yet, cultural differences between South Korea and the United States exist. The purpose ofthis study was to identify and characterize potential cultural differences in the Korean and US institutional review board (IRB approach to certain topics.Methods: A qualitative analysis of a 9-item survey, describing 4 research study case scenarios, sent to IRB members from the United States and South Korea. The case scenarios involved the followingissues: (1 the need for consent for retrospective chart review when research subjects receive their care after the study is conceived; (2 child assent; (3 individual versus population benefit; and (4 exception from informed consent in emergency resuscitation research. The free-text responses were analyzed and abstracted for recurrent themes.Results: Twenty-three of the 45 survey recipients completed the survey, for an overall response rate of 51%. The themes that emerged were as follows: (1 the importance of parental authority among Korean participants versus the importance of child autonomy and child assent among US participants; (2 the recognition of the rights of a proxy or surrogate who can represent an individual’s values by all participants; and (3 the importance of the community, expressed by the Korean respondents, versus individualism, expressed by US respondents.Conclusion: Whereas US participants appear to emphasize the importance of the individual and the autonomy of a child, the Korean respondents stressed the importance of parental authority andbenefiting the community, above and beyond that of the individual person. However, there was substantial overlap in the themes expressed by respondents from both countries.

  17. Predicting Summer Dryness Under a Warmer Climate: Modeling Land Surface Processes in the Midwestern United States

    Science.gov (United States)

    Winter, J. M.; Eltahir, E. A.

    2009-12-01

    One of the most significant impacts of climate change is the potential alteration of local hydrologic cycles over agriculturally productive areas. As the world’s food supply continues to be taxed by its burgeoning population, a greater percentage of arable land will need to be utilized and land currently producing food must become more efficient. This study seeks to quantify the effects of climate change on soil moisture in the American Midwest. A series of 24-year numerical experiments were conducted to assess the ability of Regional Climate Model Version 3 coupled to Integrated Biosphere Simulator (RegCM3-IBIS) and Biosphere-Atmosphere Transfer Scheme 1e (RegCM3-BATS1e) to simulate the observed hydroclimatology of the midwestern United States. Model results were evaluated using NASA Surface Radiation Budget, NASA Earth Radiation Budget Experiment, Illinois State Water Survey, Climate Research Unit Time Series 2.1, Global Soil Moisture Data Bank, and regional-scale estimations of evapotranspiration. The response of RegCM3-IBIS and RegCM3-BATS1e to a surrogate climate change scenario, a warming of 3oC at the boundaries and doubling of CO2, was explored. Precipitation increased significantly during the spring and summer in both RegCM3-IBIS and RegCM3-BATS1e, leading to additional runoff. In contrast, enhancement of evapotranspiration and shortwave radiation were modest. Soil moisture remained relatively unchanged in RegCM3-IBIS, while RegCM3-BATS1e exhibited some fall and winter wetting.

  18. Living Organ Donation and Informed Consent in the United States: Strategies to Improve the Process.

    Science.gov (United States)

    Henderson, Macey L; Gross, Jed Adam

    2017-03-01

    About 6,000 individuals participate in the U.S. transplant system as a living organ donor each year. Organ donation (most commonly a kidney or part of liver) by living individuals is a unique procedure, where healthy patients undergo a major surgical operation without any direct functional benefit to themselves. In this article, the authors explore how the ideal of informed consent guides education and evaluation for living organ donation. The authors posit that informed consent for living organ donation is a process. Though the steps in this process are partially standardized through national health policy, they can be improved through institutional structures at the local, transplant center-level. Effective structures and practices aimed at supporting and promoting comprehensive informed consent provide more opportunities for candidates to ask questions about the risks and benefits of living donation and to opt out voluntarily Additionally, these practices could enable new ways of measuring knowledge and improving the consent process.

  19. Guidelines, processes and tools for coastal ecosystem restoration, with examples from the United States

    Energy Technology Data Exchange (ETDEWEB)

    Thom, Ronald M.; Diefenderfer, Heida L.; Adkins, Jeffery E.; Judd, Chaeli; Anderson, Michael G.; Buenau, Kate E.; Borde, Amy B.; Johnson, Gary E.

    2011-02-01

    This paper presents a systematic approach to coastal restoration projects in five phases: planning, implementation, performance assessment, adaptive management, and dissemination of results. Twenty features of the iterative planning process are synthesized. The planning process starts with a vision, a description of the ecosystem and landscape, and goals. A conceptual model and planning objectives are developed, a site is selected using prioritization techniques, and numerical models contribute to preliminary designs as needed. Performance criteria and reference sites are selected and the monitoring program is designed. The monitoring program is emphasized as a tool to assess project performance and identify problems affecting progression toward project goals, in an adaptive management framework. Key approaches to aspects of the monitoring program are reviewed and detailed with project examples. Within the planning process, cost analysis involves budgeting, scheduling, and financing. Finally, documentation is peer reviewed prior to making construction plans and final costing.

  20. USE OF SPACE TECHNOLOGY IN FEDERALLY FUNDED LAND PROCESSES RESEARCH IN THE UNITED STATES.

    Science.gov (United States)

    Thorley, G.A.; McArdle, R.

    1986-01-01

    A review of the use of space technology in federally funded earth science research in the US was carried out in 1985 by the President's Office of Science and Technology Policy. Five departments and three independent agencies, representing the primary earth science research agencies in the Federal government, participated in the review. The review by the subcommittee indicated that, while there is considerable overlap in the legislated missions of the earth science agencies, most of the space-related land processes research is complementary. Summaries are provided of the current and projected uses of space technology in land processes activities within the eight Federal organizations.

  1. Liquid phase methanol LaPorte Process Development Unit: Modification, operation, and support studies

    Energy Technology Data Exchange (ETDEWEB)

    1991-02-28

    The LPMEOH process was conceived and patented by Chem Systems Inc. in 1975. Initial research and studies on the process focused on two distinct modes of operation. The first was a liquid fluidized mode with relatively large catalyst pellets suspended in a fluidizing liquid, and the second was an entrained (slurry) mode with fine catalyst particles slurried in an inert liquid. The development of both operating modes progressed in parallel from bench scale reactors, through an intermediate scale lab PDU, and then to the LaPorte PDU in 1984. The slurry mode of operation was ultimately chosen as the operating mode of choice due to its superior performance.

  2. Pyrolysis oil upgrading for Co-processing in standard refinery units

    NARCIS (Netherlands)

    De Miguel Mercader, Ferran

    2010-01-01

    This thesis considers the route that comprises the upgrading of pyrolysis oil (produced from lingo-cellulosic biomass) and its further co-processing in standard refineries to produce transportation fuels. In the present concept, pyrolysis oil is produced where biomass is available and then transport

  3. Effectiveness of Information Technology Infrastructure Library Process Implementations by Information Technology Departments within United States Organizations

    Science.gov (United States)

    Persinger, Jon F.

    2010-01-01

    This research study examined whether the overall effectiveness of the Information Technology Infrastructure Library (ITIL) could be predicted by measuring multiple independent variables. The primary variables studied included the number of ITIL process modules adopted and the overall ITIL "maturity level." An online survey was used to…

  4. Applying unit process life cycle inventor (UPLCI) methodology in product/packaging combinatons

    NARCIS (Netherlands)

    Oude Luttikhuis, Ellen; Toxopeus, Marten E.; Overcash, M.; Nee, Andrew Y.C.; Song, Bin; Ong, Soh-Khim

    2013-01-01

    This paper discusses how the UPLCI approach can be used for determining the inventory of the manufacturing phases of product/packaging combinations. The UPLCI approach can make the inventory of the manufacturing process of the product that is investigated more accurate. The life cycle of

  5. Pyrolysis oil upgrading for Co-processing in standard refinery units

    NARCIS (Netherlands)

    De Miguel Mercader, F.

    2010-01-01

    This thesis considers the route that comprises the upgrading of pyrolysis oil (produced from lingo-cellulosic biomass) and its further co-processing in standard refineries to produce transportation fuels. In the present concept, pyrolysis oil is produced where biomass is available and then

  6. How semantics can improve engineering processes: A case of units of measure and quantities

    NARCIS (Netherlands)

    Rijgersberg, H.; Wigham, M.L.I.; Top, J.L.

    2011-01-01

    Science and engineering heavily depend on the ability to share data and models. The World Wide Web provides even greater opportunity to reuse such information from disparate sources. Moreover, if the information is digitized it can to a large extent be processed automatically. However, information s

  7. Effectiveness of Information Technology Infrastructure Library Process Implementations by Information Technology Departments within United States Organizations

    Science.gov (United States)

    Persinger, Jon F.

    2010-01-01

    This research study examined whether the overall effectiveness of the Information Technology Infrastructure Library (ITIL) could be predicted by measuring multiple independent variables. The primary variables studied included the number of ITIL process modules adopted and the overall ITIL "maturity level." An online survey was used to…

  8. Effects of chemical protective equipment on team process performance in small unit rescue operations.

    Science.gov (United States)

    Grugle, Nancy L; Kleiner, Brian M

    2007-09-01

    In the event of a nuclear, biological, or chemical terrorist attack against civilians, both military and civilian emergency response teams must be able to respond and operate efficiently while wearing protective equipment. Chemical protective equipment protects the user by providing a barrier between the individual and hazardous environment. Unfortunately, the same equipment that is designed to support the user can potentially cause heat stress, reduced task efficiency, and reduced range-of-motion. Targeted Acceptable Responses to Generated Events of Tasks (TARGETS), an event-based team performance measurement methodology was used to investigate the effects of Mission Oriented Protective Posture (MOPP) on the behavioral processes underlying team performance during simulated rescue tasks. In addition, this study determined which team processes were related to team performance outcomes. Results of six primary analyses indicated that team process performance was not degraded by MOPP 4 on any rescue task and that the team processes critical for successful task performance are task-dependent. This article discusses the implications of these results with respect to the study design and the limitations of using an event-based team performance measurement methodology.

  9. How semantics can improve engineering processes: A case of units of measure and quantities

    NARCIS (Netherlands)

    Rijgersberg, H.; Wigham, M.L.I.; Top, J.L.

    2011-01-01

    Science and engineering heavily depend on the ability to share data and models. The World Wide Web provides even greater opportunity to reuse such information from disparate sources. Moreover, if the information is digitized it can to a large extent be processed automatically. However, information

  10. Applying unit process life cycle inventor (UPLCI) methodology in product/packaging combinatons

    NARCIS (Netherlands)

    Oude Luttikhuis, E.J.; Toxopeus, M.E.; Overcash, M.; Nee, Andrew Y.C.; Song, Bin; Ong, Soh-Khim

    2013-01-01

    This paper discusses how the UPLCI approach can be used for determining the inventory of the manufacturing phases of product/packaging combinations. The UPLCI approach can make the inventory of the manufacturing process of the product that is investigated more accurate. The life cycle of product/pac

  11. The role of personnel marketing in the process of building corporate social responsibility strategy of a scientific unit

    Directory of Open Access Journals (Sweden)

    Sylwia Jarosławska-Sobór

    2015-09-01

    Full Text Available The goal of this article is to discuss the significance of human capital in the process of building the strategy of social responsibility and the role of personnel marketing in the process. Dynamically changing social environment has enforced a new way of looking at non-material resources. Organizations have understood that it is human capital and social competences that have a significant impact on the creation of an organization’s value, generating profits, as well as gaining competitive advantage in the 21st century. Personnel marketing is now a key element in the process of implementation of the CSR concept and building the value of contemporary organizations, especially such unique organizations as scientific units. In this article you will find a discussion concerning the basic values regarded as crucial by the Central Mining Institute in the context of their significance for the paradigm of social responsibility. Such an analysis was carried out on the basis of the experiences of Central Mining Institute (GIG in the development of strategic CSR, which takes into consideration the specific character of the Institute as a scientific unit.

  12. Genetic Algorithm Supported by Graphical Processing Unit Improves the Exploration of Effective Connectivity in Functional Brain Imaging

    Directory of Open Access Journals (Sweden)

    Lawrence Wing Chi Chan

    2015-05-01

    Full Text Available Brain regions of human subjects exhibit certain levels of associated activation upon specific environmental stimuli. Functional Magnetic Resonance Imaging (fMRI detects regional signals, based on which we could infer the direct or indirect neuronal connectivity between the regions. Structural Equation Modeling (SEM is an appropriate mathematical approach for analyzing the effective connectivity using fMRI data. A maximum likelihood (ML discrepancy function is minimized against some constrained coefficients of a path model. The minimization is an iterative process. The computing time is very long as the number of iterations increases geometrically with the number of path coefficients. Using regular Quad-Core Central Processing Unit (CPU platform, duration up to three months is required for the iterations from 0 to 30 path coefficients. This study demonstrates the application of Graphical Processing Unit (GPU with the parallel Genetic Algorithm (GA that replaces the Powell minimization in the standard program code of the analysis software package. It was found in the same example that GA under GPU reduced the duration to 20 hours and provided more accurate solution when compared with standard program code under CPU.

  13. UNITED STATES DEPARTMENT OF ENERGY WASTE PROCESSING ANNUAL TECHNOLOGY DEVELOPMENT REPORT 2007

    Energy Technology Data Exchange (ETDEWEB)

    Bush, S

    2008-08-12

    The Office of Environmental Management's (EM) Roadmap, U.S. Department of Energy--Office of Environmental Management Engineering & Technology Roadmap (Roadmap), defines the Department's intent to reduce the technical risk and uncertainty in its cleanup programs. The unique nature of many of the remaining facilities will require a strong and responsive engineering and technology program to improve worker and public safety, and reduce costs and environmental impacts while completing the cleanup program. The technical risks and uncertainties associated with cleanup program were identified through: (1) project risk assessments, (2) programmatic external technical reviews and technology readiness assessments, and (3) direct site input. In order to address these needs, the technical risks and uncertainties were compiled and divided into the program areas of: Waste Processing, Groundwater and Soil Remediation, and Deactivation and Decommissioning (D&D). Strategic initiatives were then developed within each program area to address the technical risks and uncertainties in that program area. These strategic initiatives were subsequently incorporated into the Roadmap, where they form the strategic framework of the EM Engineering & Technology Program. The EM-21 Multi-Year Program Plan (MYPP) supports the goals and objectives of the Roadmap by providing direction for technology enhancement, development, and demonstrations that will lead to a reduction of technical uncertainties in EM waste processing activities. The current MYPP summarizes the strategic initiatives and the scope of the activities within each initiative that are proposed for the next five years (FY2008-2012) to improve safety and reduce costs and environmental impacts associated with waste processing; authorized budget levels will impact how much of the scope of activities can be executed, on a year-to-year basis. As a result of the importance of reducing technical risk and uncertainty in the EM Waste

  14. Future forest aboveground carbon dynamics in the central United States: the importance of forest demographic processes

    Science.gov (United States)

    Jin, Wenchi; He, Hong S.; Thompson, Frank R.; Wang, Wen J.; Fraser, Jacob S.; Shifley, Stephen R.; Hanberry, Brice B.; Dijak, William D.

    2017-01-01

    The Central Hardwood Forest (CHF) in the United States is currently a major carbon sink, there are uncertainties in how long the current carbon sink will persist and if the CHF will eventually become a carbon source. We used a multi-model ensemble to investigate aboveground carbon density of the CHF from 2010 to 2300 under current climate. Simulations were done using one representative model for each of the simple, intermediate, and complex demographic approaches (ED2, LANDIS PRO, and LINKAGES, respectively). All approaches agreed that the current carbon sink would persist at least to 2100. However, carbon dynamics after current carbon sink diminishes to zero differ for different demographic modelling approaches. Both the simple and the complex demographic approaches predicted prolonged periods of relatively stable carbon densities after 2100, with minor declines, until the end of simulations in 2300. In contrast, the intermediate demographic approach predicted the CHF would become a carbon source between 2110 and 2260, followed by another carbon sink period. The disagreement between these patterns can be partly explained by differences in the capacity of models to simulate gross growth (both birth and subsequent growth) and mortality of short-lived, relatively shade-intolerant tree species. PMID:28165483

  15. Triboelectric Nanogenerator as a Self-Powered Communication Unit for Processing and Transmitting Information.

    Science.gov (United States)

    Yu, Aifang; Chen, Xiangyu; Wang, Rui; Liu, Jingyu; Luo, Jianjun; Chen, Libo; Zhang, Yang; Wu, Wei; Liu, Caihong; Yuan, Hongtao; Peng, Mingzeng; Hu, Weiguo; Zhai, Junyi; Wang, Zhong Lin

    2016-04-26

    In this paper, we demonstrate an application of a triboelectric nanogenerator (TENG) as a self-powered communication unit. An elaborately designed TENG is used to translate a series of environmental triggering signals into binary digital signals and drives an electronic-optical device to transmit binary digital data in real-time without an external power supply. The elaborately designed TENG is built in a membrane structure that can effectively drive the electronic-optical device in a bandwidth from 1.30 to 1.65 kHz. Two typical communication modes (amplitude-shift keying and frequency-shift keying) are realized through the resonant response of TENG to different frequencies, and two digital signals, i.e., "1001" and "0110", are successfully transmitted and received through this system, respectively. Hence, in this study, a simple but efficient method for directly transmitting ambient vibration to the receiver as a digital signal is established using an elaborately designed TENG and an optical communication technique. This type of the communication system, as well as the implementation method presented, exhibits great potential for applications in the smart city, smart home, password authentication, and so on.

  16. Identified auditory neurons in the cricket Gryllus rubens: temporal processing in calling song sensitive units.

    Science.gov (United States)

    Farris, Hamilton E; Mason, Andrew C; Hoy, Ronald R

    2004-07-01

    This study characterizes aspects of the anatomy and physiology of auditory receptors and certain interneurons in the cricket Gryllus rubens. We identified an 'L'-shaped ascending interneuron tuned to frequencies > 15 kHz (57 dB SPL threshold at 20 kHz). Also identified were two intrasegmental 'omega'-shaped interneurons that were broadly tuned to 3-65 kHz, with best sensitivity to frequencies of the male calling song (5 kHz, 52 dB SPL). The temporal sensitivity of units excited by calling song frequencies were measured using sinusoidally amplitude modulated stimuli that varied in both modulation rate and depth, parameters that vary with song propagation distance and the number of singing males. Omega cells responded like low-pass filters with a time constant of 42 ms. In contrast, receptors significantly coded modulation rates up to the maximum rate presented (85 Hz). Whereas omegas required approximately 65% modulation depth at 45 Hz (calling song AM) to elicit significant synchrony coding, receptors tolerated a approximately 50% reduction in modulation depth up to 85 Hz. These results suggest that omega cells in G. rubens might not play a role in detecting song modulation per se at increased distances from a singing male.

  17. Developmental integration in a functional unit: deciphering processes from adult dental morphology.

    Science.gov (United States)

    Labonne, Gaëlle; Navarro, Nicolas; Laffont, Rémi; Chateau-Smith, Carmela; Montuire, Sophie

    2014-01-01

    The evolution of mammalian dentition is constrained by functional necessity and by the non-independence of morphological structures. Efficient chewing implies coherent tooth coordination from development to motion, involving covariation patterns (integration) within dental parts. Using geometric morphometrics, we investigate the modular organization of the highly derived vole dentition. Integration patterns between and within the upper and lower molar rows are analyzed to identify potential modules and their origins (functional and developmental). Results support an integrated adult dentition pattern for both developmental and functional aspects. The integration patterns between opposing molar pairs suggest a transient role for the second upper and lower molars during the chewing motion. Upper and lower molar rows form coherent units but the relative integration of molar pairs is in contradiction with existing developmental models. Emphasis on the first three cusps to grow leads to a very different integration pattern, which would be congruent with developmental models. The early developmental architecture of traits is masked by later stages of growth, but may still be deciphered from the adult phenotype, if careful attention is paid to relevant features.

  18. Methane emissions from process equipment at natural gas production sites in the United States: pneumatic controllers.

    Science.gov (United States)

    Allen, David T; Pacsi, Adam P; Sullivan, David W; Zavala-Araiza, Daniel; Harrison, Matthew; Keen, Kindal; Fraser, Matthew P; Daniel Hill, A; Sawyer, Robert F; Seinfeld, John H

    2015-01-06

    Emissions from 377 gas actuated (pneumatic) controllers were measured at natural gas production sites and a small number of oil production sites, throughout the United States. A small subset of the devices (19%), with whole gas emission rates in excess of 6 standard cubic feet per hour (scf/h), accounted for 95% of emissions. More than half of the controllers recorded emissions of 0.001 scf/h or less during 15 min of measurement. Pneumatic controllers in level control applications on separators and in compressor applications had higher emission rates than controllers in other types of applications. Regional differences in emissions were observed, with the lowest emissions measured in the Rocky Mountains and the highest emissions in the Gulf Coast. Average methane emissions per controller reported in this work are 17% higher than the average emissions per controller in the 2012 EPA greenhouse gas national emission inventory (2012 GHG NEI, released in 2014); the average of 2.7 controllers per well observed in this work is higher than the 1.0 controllers per well reported in the 2012 GHG NEI.

  19. Methane emissions from process equipment at natural gas production sites in the United States: liquid unloadings.

    Science.gov (United States)

    Allen, David T; Sullivan, David W; Zavala-Araiza, Daniel; Pacsi, Adam P; Harrison, Matthew; Keen, Kindal; Fraser, Matthew P; Daniel Hill, A; Lamb, Brian K; Sawyer, Robert F; Seinfeld, John H

    2015-01-01

    Methane emissions from liquid unloadings were measured at 107 wells in natural gas production regions throughout the United States. Liquid unloadings clear wells of accumulated liquids to increase production, employing a variety of liquid lifting mechanisms. In this work, wells with and without plunger lifts were sampled. Most wells without plunger lifts unload less than 10 times per year with emissions averaging 21,000-35,000 scf methane (0.4-0.7 Mg) per event (95% confidence limits of 10,000-50,000 scf/event). For wells with plunger lifts, emissions averaged 1000-10,000 scf methane (0.02-0.2 Mg) per event (95% confidence limits of 500-12,000 scf/event). Some wells with plunger lifts are automatically triggered and unload thousands of times per year and these wells account for the majority of the emissions from all wells with liquid unloadings. If the data collected in this work are assumed to be representative of national populations, the data suggest that the central estimate of emissions from unloadings (270 Gg/yr, 95% confidence range of 190-400 Gg) are within a few percent of the emissions estimated in the EPA 2012 Greenhouse Gas National Emission Inventory (released in 2014), with emissions dominated by wells with high frequencies of unloadings.

  20. Accelerating resolution-of-the-identity second-order Møller-Plesset quantum chemistry calculations with graphical processing units.

    Science.gov (United States)

    Vogt, Leslie; Olivares-Amaya, Roberto; Kermes, Sean; Shao, Yihan; Amador-Bedolla, Carlos; Aspuru-Guzik, Alan

    2008-03-13

    The modification of a general purpose code for quantum mechanical calculations of molecular properties (Q-Chem) to use a graphical processing unit (GPU) is reported. A 4.3x speedup of the resolution-of-the-identity second-order Møller-Plesset perturbation theory (RI-MP2) execution time is observed in single point energy calculations of linear alkanes. The code modification is accomplished using the compute unified basic linear algebra subprograms (CUBLAS) library for an NVIDIA Quadro FX 5600 graphics card. Furthermore, speedups of other matrix algebra based electronic structure calculations are anticipated as a result of using a similar approach.