WorldWideScience

Sample records for high computational efficiency

  1. PRCA:A highly efficient computing architecture

    Institute of Scientific and Technical Information of China (English)

    Luo Xingguo

    2014-01-01

    Applications can only reach 8 %~15 % of utilization on modern computer systems. There are many obstacles to improving system efficiency. The key root is the conflict between the fixed general computer architecture and the variable requirements of applications. Proactive reconfigurable computing architecture (PRCA) is proposed to improve computing efficiency. PRCA dynamically constructs an efficient computing ar chitecture for a specific application via reconfigurable technology by perceiving requirements,workload and utilization of computing resources. Proactive decision support system (PDSS),hybrid reconfigurable computing array (HRCA) and reconfigurable interconnect (RIC) are intensively researched as the key technologies. The principles of PRCA have been verified with four applications on a test bed. It is shown that PRCA is feasible and highly efficient.

  2. Efficient High Performance Computing on Heterogeneous Platforms

    NARCIS (Netherlands)

    Shen, J.

    2015-01-01

    Heterogeneous platforms are mixes of different processing units in a compute node (e.g., CPUs+GPUs, CPU+MICs) or a chip package (e.g., APUs). This type of platforms keeps gaining popularity in various computer systems ranging from supercomputers to mobile devices. In this context, improving their

  3. Efficient High Performance Computing on Heterogeneous Platforms

    NARCIS (Netherlands)

    Shen, J.

    2015-01-01

    Heterogeneous platforms are mixes of different processing units in a compute node (e.g., CPUs+GPUs, CPU+MICs) or a chip package (e.g., APUs). This type of platforms keeps gaining popularity in various computer systems ranging from supercomputers to mobile devices. In this context, improving their ef

  4. A Highly Efficient Parallel Algorithm for Computing the Fiedler Vector

    CERN Document Server

    Manguoglu, Murat

    2010-01-01

    The eigenvector corresponding to the second smallest eigenvalue of the Laplacian of a graph, known as the Fiedler vector, has a number of applications in areas that include matrix reordering, graph partitioning, protein analysis, data mining, machine learning, and web search. The computation of the Fiedler vector has been regarded as an expensive process as it involves solving a large eigenvalue problem. We present a novel and efficient parallel algorithm for computing the Fiedler vector of large graphs based on the Trace Minimization algorithm (Sameh, et.al). We compare the parallel performance of our method with a multilevel scheme, designed specifically for computing the Fiedler vector, which is implemented in routine MC73\\_Fiedler of the Harwell Subroutine Library (HSL). In addition, we compare the quality of the Fiedler vector for the application of weighted matrix reordering and provide a metric for measuring the quality of reordering.

  5. Enabling Efficient Climate Science Workflows in High Performance Computing Environments

    Science.gov (United States)

    Krishnan, H.; Byna, S.; Wehner, M. F.; Gu, J.; O'Brien, T. A.; Loring, B.; Stone, D. A.; Collins, W.; Prabhat, M.; Liu, Y.; Johnson, J. N.; Paciorek, C. J.

    2015-12-01

    A typical climate science workflow often involves a combination of acquisition of data, modeling, simulation, analysis, visualization, publishing, and storage of results. Each of these tasks provide a myriad of challenges when running on a high performance computing environment such as Hopper or Edison at NERSC. Hurdles such as data transfer and management, job scheduling, parallel analysis routines, and publication require a lot of forethought and planning to ensure that proper quality control mechanisms are in place. These steps require effectively utilizing a combination of well tested and newly developed functionality to move data, perform analysis, apply statistical routines, and finally, serve results and tools to the greater scientific community. As part of the CAlibrated and Systematic Characterization, Attribution and Detection of Extremes (CASCADE) project we highlight a stack of tools our team utilizes and has developed to ensure that large scale simulation and analysis work are commonplace and provide operations that assist in everything from generation/procurement of data (HTAR/Globus) to automating publication of results to portals like the Earth Systems Grid Federation (ESGF), all while executing everything in between in a scalable environment in a task parallel way (MPI). We highlight the use and benefit of these tools by showing several climate science analysis use cases they have been applied to.

  6. Energy-efficient high performance computing measurement and tuning

    CERN Document Server

    III, James H Laros; Kelly, Sue

    2012-01-01

    In this work, the unique power measurement capabilities of the Cray XT architecture were exploited to gain an understanding of power and energy use, and the effects of tuning both CPU and network bandwidth. Modifications were made to deterministically halt cores when idle. Additionally, capabilities were added to alter operating P-state. At the application level, an understanding of the power requirements of a range of important DOE/NNSA production scientific computing applications running at large scale is gained by simultaneously collecting current and voltage measurements on the hosting nod

  7. A computationally efficient parallel Levenberg-Marquardt algorithm for highly parameterized inverse model analyses

    Science.gov (United States)

    Lin, Youzuo; O'Malley, Daniel; Vesselinov, Velimir V.

    2016-09-01

    Inverse modeling seeks model parameters given a set of observations. However, for practical problems because the number of measurements is often large and the model parameters are also numerous, conventional methods for inverse modeling can be computationally expensive. We have developed a new, computationally efficient parallel Levenberg-Marquardt method for solving inverse modeling problems with a highly parameterized model space. Levenberg-Marquardt methods require the solution of a linear system of equations which can be prohibitively expensive to compute for moderate to large-scale problems. Our novel method projects the original linear problem down to a Krylov subspace such that the dimensionality of the problem can be significantly reduced. Furthermore, we store the Krylov subspace computed when using the first damping parameter and recycle the subspace for the subsequent damping parameters. The efficiency of our new inverse modeling algorithm is significantly improved using these computational techniques. We apply this new inverse modeling method to invert for random transmissivity fields in 2-D and a random hydraulic conductivity field in 3-D. Our algorithm is fast enough to solve for the distributed model parameters (transmissivity) in the model domain. The algorithm is coded in Julia and implemented in the MADS computational framework (http://mads.lanl.gov). By comparing with Levenberg-Marquardt methods using standard linear inversion techniques such as QR or SVD methods, our Levenberg-Marquardt method yields a speed-up ratio on the order of ˜101 to ˜102 in a multicore computational environment. Therefore, our new inverse modeling method is a powerful tool for characterizing subsurface heterogeneity for moderate to large-scale problems.

  8. Exploring Graphics Processing Unit (GPU Resource Sharing Efficiency for High Performance Computing

    Directory of Open Access Journals (Sweden)

    Teng Li

    2013-11-01

    Full Text Available The increasing incorporation of Graphics Processing Units (GPUs as accelerators has been one of the forefront High Performance Computing (HPC trends and provides unprecedented performance; however, the prevalent adoption of the Single-Program Multiple-Data (SPMD programming model brings with it challenges of resource underutilization. In other words, under SPMD, every CPU needs GPU capability available to it. However, since CPUs generally outnumber GPUs, the asymmetric resource distribution gives rise to overall computing resource underutilization. In this paper, we propose to efficiently share the GPU under SPMD and formally define a series of GPU sharing scenarios. We provide performance-modeling analysis for each sharing scenario with accurate experimentation validation. With the modeling basis, we further conduct experimental studies to explore potential GPU sharing efficiency improvements from multiple perspectives. Both further theoretical and experimental GPU sharing performance analysis and results are presented. Our results not only demonstrate the significant performance gain for SPMD programs with the proposed efficient GPU sharing, but also the further improved sharing efficiency with the optimization techniques based on our accurate modeling.

  9. Low-cost, high-performance and efficiency computational photometer design

    Science.gov (United States)

    Siewert, Sam B.; Shihadeh, Jeries; Myers, Randall; Khandhar, Jay; Ivanov, Vitaly

    2014-05-01

    Researchers at the University of Alaska Anchorage and University of Colorado Boulder have built a low cost high performance and efficiency drop-in-place Computational Photometer (CP) to test in field applications ranging from port security and safety monitoring to environmental compliance monitoring and surveying. The CP integrates off-the-shelf visible spectrum cameras with near to long wavelength infrared detectors and high resolution digital snapshots in a single device. The proof of concept combines three or more detectors into a single multichannel imaging system that can time correlate read-out, capture, and image process all of the channels concurrently with high performance and energy efficiency. The dual-channel continuous read-out is combined with a third high definition digital snapshot capability and has been designed using an FPGA (Field Programmable Gate Array) to capture, decimate, down-convert, re-encode, and transform images from two standard definition CCD (Charge Coupled Device) cameras at 30Hz. The continuous stereo vision can be time correlated to megapixel high definition snapshots. This proof of concept has been fabricated as a fourlayer PCB (Printed Circuit Board) suitable for use in education and research for low cost high efficiency field monitoring applications that need multispectral and three dimensional imaging capabilities. Initial testing is in progress and includes field testing in ports, potential test flights in un-manned aerial systems, and future planned missions to image harsh environments in the arctic including volcanic plumes, ice formation, and arctic marine life.

  10. Computationally efficient method for Fourier transform of highly chirped pulses for laser and parametric amplifier modeling.

    Science.gov (United States)

    Andrianov, Alexey; Szabo, Aron; Sergeev, Alexander; Kim, Arkady; Chvykov, Vladimir; Kalashnikov, Mikhail

    2016-11-14

    We developed an improved approach to calculate the Fourier transform of signals with arbitrary large quadratic phase which can be efficiently implemented in numerical simulations utilizing Fast Fourier transform. The proposed algorithm significantly reduces the computational cost of Fourier transform of a highly chirped and stretched pulse by splitting it into two separate transforms of almost transform limited pulses, thereby reducing the required grid size roughly by a factor of the pulse stretching. The application of our improved Fourier transform algorithm in the split-step method for numerical modeling of CPA and OPCPA shows excellent agreement with standard algorithms.

  11. Computational design of high efficiency release targets for use at ISOL facilities

    CERN Document Server

    Liu, Y

    1999-01-01

    This report describes efforts made at the Oak Ridge National Laboratory to design high-efficiency-release targets that simultaneously incorporate the short diffusion lengths, high permeabilities, controllable temperatures, and heat-removal properties required for the generation of useful radioactive ion beam (RIB) intensities for nuclear physics and astrophysics research using the isotope separation on-line (ISOL) technique. Short diffusion lengths are achieved either by using thin fibrous target materials or by coating thin layers of selected target material onto low-density carbon fibers such as reticulated-vitreous-carbon fiber (RVCF) or carbon-bonded-carbon fiber (CBCF) to form highly permeable composite target matrices. Computational studies that simulate the generation and removal of primary beam deposited heat from target materials have been conducted to optimize the design of target/heat-sink systems for generating RIBs. The results derived from diffusion release-rate simulation studies for selected t...

  12. Computational model for a high temperature electrolyzer coupled to a HTTR for efficient nuclear hydrogen production

    Energy Technology Data Exchange (ETDEWEB)

    Gonzalez, Daniel; Rojas, Leorlen; Rosales, Jesus; Castro, Landy; Gamez, Abel; Brayner, Carlos, E-mail: danielgonro@gmail.com [Universidade Federal de Pernambuco (UFPE), Recife, PE (Brazil); Garcia, Lazaro; Garcia, Carlos; Torre, Raciel de la, E-mail: lgarcia@instec.cu [Instituto Superior de Tecnologias y Ciencias Aplicadas (InSTEC), La Habana (Cuba); Sanchez, Danny [Universidade Estadual de Santa Cruz (UESC), Ilheus, BA (Brazil)

    2015-07-01

    High temperature electrolysis process coupled to a very high temperature reactor (VHTR) is one of the most promising methods for hydrogen production using a nuclear reactor as the primary heat source. However there are not references in the scientific publications of a test facility that allow to evaluate the efficiency of the process and other physical parameters that has to be taken into consideration for its accurate application in the hydrogen economy as a massive production method. For this lack of experimental facilities, mathematical models are one of the most used tools to study this process and theirs flowsheets, in which the electrolyzer is the most important component because of its complexity and importance in the process. A computational fluid dynamic (CFD) model for the evaluation and optimization of the electrolyzer of a high temperature electrolysis hydrogen production process flowsheet was developed using ANSYS FLUENT®. Electrolyzer's operational and design parameters will be optimized in order to obtain the maximum hydrogen production and the higher efficiency in the module. This optimized model of the electrolyzer will be incorporated to a chemical process simulation (CPS) code to study the overall high temperature flowsheet coupled to a high temperature accelerator driven system (ADS) that offers advantages in the transmutation of the spent fuel. (author)

  13. Measuring and tuning energy efficiency on large scale high performance computing platforms.

    Energy Technology Data Exchange (ETDEWEB)

    Laros, James H., III

    2011-08-01

    Recognition of the importance of power in the field of High Performance Computing, whether it be as an obstacle, expense or design consideration, has never been greater and more pervasive. While research has been conducted on many related aspects, there is a stark absence of work focused on large scale High Performance Computing. Part of the reason is the lack of measurement capability currently available on small or large platforms. Typically, research is conducted using coarse methods of measurement such as inserting a power meter between the power source and the platform, or fine grained measurements using custom instrumented boards (with obvious limitations in scale). To collect the measurements necessary to analyze real scientific computing applications at large scale, an in-situ measurement capability must exist on a large scale capability class platform. In response to this challenge, we exploit the unique power measurement capabilities of the Cray XT architecture to gain an understanding of power use and the effects of tuning. We apply these capabilities at the operating system level by deterministically halting cores when idle. At the application level, we gain an understanding of the power requirements of a range of important DOE/NNSA production scientific computing applications running at large scale (thousands of nodes), while simultaneously collecting current and voltage measurements on the hosting nodes. We examine the effects of both CPU and network bandwidth tuning and demonstrate energy savings opportunities of up to 39% with little or no impact on run-time performance. Capturing scale effects in our experimental results was key. Our results provide strong evidence that next generation large-scale platforms should not only approach CPU frequency scaling differently, but could also benefit from the capability to tune other platform components, such as the network, to achieve energy efficient performance.

  14. Exploring Infiniband Hardware Virtualization in OpenNebula towards Efficient High-Performance Computing

    Energy Technology Data Exchange (ETDEWEB)

    Pais Pitta de Lacerda Ruivo, Tiago [IIT, Chicago; Bernabeu Altayo, Gerard [Fermilab; Garzoglio, Gabriele [Fermilab; Timm, Steven [Fermilab; Kim, Hyun-Woo [Fermilab; Noh, Seo-Young [KISTI, Daejeon; Raicu, Ioan [IIT, Chicago

    2014-11-11

    has been widely accepted that software virtualization has a big negative impact on high-performance computing (HPC) application performance. This work explores the potential use of Infiniband hardware virtualization in an OpenNebula cloud towards the efficient support of MPI-based workloads. We have implemented, deployed, and tested an Infiniband network on the FermiCloud private Infrastructure-as-a-Service (IaaS) cloud. To avoid software virtualization towards minimizing the virtualization overhead, we employed a technique called Single Root Input/Output Virtualization (SRIOV). Our solution spanned modifications to the Linux’s Hypervisor as well as the OpenNebula manager. We evaluated the performance of the hardware virtualization on up to 56 virtual machines connected by up to 8 DDR Infiniband network links, with micro-benchmarks (latency and bandwidth) as well as w a MPI-intensive application (the HPL Linpack benchmark).

  15. Improving the Eco-Efficiency of High Performance Computing Clusters Using EECluster

    Directory of Open Access Journals (Sweden)

    Alberto Cocaña-Fernández

    2016-03-01

    Full Text Available As data and supercomputing centres increase their performance to improve service quality and target more ambitious challenges every day, their carbon footprint also continues to grow, and has already reached the magnitude of the aviation industry. Also, high power consumptions are building up to a remarkable bottleneck for the expansion of these infrastructures in economic terms due to the unavailability of sufficient energy sources. A substantial part of the problem is caused by current energy consumptions of High Performance Computing (HPC clusters. To alleviate this situation, we present in this work EECluster, a tool that integrates with multiple open-source Resource Management Systems to significantly reduce the carbon footprint of clusters by improving their energy efficiency. EECluster implements a dynamic power management mechanism based on Computational Intelligence techniques by learning a set of rules through multi-criteria evolutionary algorithms. This approach enables cluster operators to find the optimal balance between a reduction in the cluster energy consumptions, service quality, and number of reconfigurations. Experimental studies using both synthetic and actual workloads from a real world cluster support the adoption of this tool to reduce the carbon footprint of HPC clusters.

  16. Computational Design of High Efficiency Release Targets for Use at ISOL Facilities

    Energy Technology Data Exchange (ETDEWEB)

    Alton, G.D.; Liu, Y.; Middleton, J.W.

    1998-11-04

    This report describes efforts made at the Oak Ridge National Laboratory to design high-efficiency-release targets that simultaneously incorporate the short diffusion lengths, high permeabilities, controllable temperatures, and heat removal properties required for the generation of useful radioactive ion beam (RIB) intensities for nuclear physics and astrophysics research using the isotope separation on-line (ISOL) technique. Short diffusion lengths are achieved either by using thin fibrous target materials or by coating thin layers of selected target material onto low-density carbon fibers such as reticulated vitreous carbon fiber (RVCF) or carbon-bonded-carbon-fiber (CBCF) to form highly permeable composite target matrices. Computational studies which simulate the generation and removal of primary beam deposited heat from target materials have been conducted to optimize the design of target/heat-sink systems for generating RIBs. The results derived tlom diffusion release-rate simulation studies for selected targets and thermal analyses of temperature distributions within a prototype target/heat-sink system subjected to primary ion beam irradiation will be presented in this report.

  17. Efficient Seeds Computation Revisited

    CERN Document Server

    Christou, Michalis; Iliopoulos, Costas S; Kubica, Marcin; Pissis, Solon P; Radoszewski, Jakub; Rytter, Wojciech; Szreder, Bartosz; Walen, Tomasz

    2011-01-01

    The notion of the cover is a generalization of a period of a string, and there are linear time algorithms for finding the shortest cover. The seed is a more complicated generalization of periodicity, it is a cover of a superstring of a given string, and the shortest seed problem is of much higher algorithmic difficulty. The problem is not well understood, no linear time algorithm is known. In the paper we give linear time algorithms for some of its versions --- computing shortest left-seed array, longest left-seed array and checking for seeds of a given length. The algorithm for the last problem is used to compute the seed array of a string (i.e., the shortest seeds for all the prefixes of the string) in $O(n^2)$ time. We describe also a simpler alternative algorithm computing efficiently the shortest seeds. As a by-product we obtain an $O(n\\log{(n/m)})$ time algorithm checking if the shortest seed has length at least $m$ and finding the corresponding seed. We also correct some important details missing in th...

  18. Many-core technologies: The move to energy-efficient, high-throughput x86 computing (TFLOPS on a chip)

    CERN Document Server

    CERN. Geneva

    2012-01-01

    With Moore's Law alive and well, more and more parallelism is introduced into all computing platforms at all levels of integration and programming to achieve higher performance and energy efficiency. Especially in the area of High-Performance Computing (HPC) users can entertain a combination of different hardware and software parallel architectures and programming environments. Those technologies range from vectorization and SIMD computation over shared memory multi-threading (e.g. OpenMP) to distributed memory message passing (e.g. MPI) on cluster systems. We will discuss HPC industry trends and Intel's approach to it from processor/system architectures and research activities to hardware and software tools technologies. This includes the recently announced new Intel(r) Many Integrated Core (MIC) architecture for highly-parallel workloads and general purpose, energy efficient TFLOPS performance, some of its architectural features and its programming environment. At the end we will have a br...

  19. Design High Efficiency-Minimum Rule Base PID Like Fuzzy Computed Torque Controller

    Directory of Open Access Journals (Sweden)

    Alireza Khalilian

    2014-06-01

    Full Text Available The minimum rule base Proportional Integral Derivative (PID Fuzzy Computed Torque Controller is presented in this research. The popularity of PID Fuzzy Computed Torque Controller can be attributed to their robust performance in a wide range of operating conditions and partly to their functional simplicity. The process of setting of PID Fuzzy Computed Torque Controller can be determined as an optimization task. Over the years, use of intelligent strategies for tuning of these controllers has been growing. PID methodology has three inputs and if any input is described with seven linguistic values, and any rule has three conditions we will need 343 rules. It is too much work to write 343 rules. In this research the PID-like fuzzy controller can be constructed as a parallel structure of a PD-like fuzzy controller and a PI controller to have the minimum rule base. However computed torque controller is work based on cancelling decoupling and nonlinear terms of dynamic parameters of each link, this controller is work based on manipulator dynamic model and this technique is highly sensitive to the knowledge of all parameters of nonlinear robot manipulator’s dynamic equation. This research is used to reduce or eliminate the computed torque controller problem based on minimum rule base fuzzy logic theory to control of flexible robot manipulator system and testing of the quality of process control in the simulation environment of MATLAB/SIMULINK Simulator.

  20. Computationally efficient algorithm for high sampling-frequency operation of active noise control

    Science.gov (United States)

    Rout, Nirmal Kumar; Das, Debi Prasad; Panda, Ganapati

    2015-05-01

    In high sampling-frequency operation of active noise control (ANC) system the length of the secondary path estimate and the ANC filter are very long. This increases the computational complexity of the conventional filtered-x least mean square (FXLMS) algorithm. To reduce the computational complexity of long order ANC system using FXLMS algorithm, frequency domain block ANC algorithms have been proposed in past. These full block frequency domain ANC algorithms are associated with some disadvantages such as large block delay, quantization error due to computation of large size transforms and implementation difficulties in existing low-end DSP hardware. To overcome these shortcomings, the partitioned block ANC algorithm is newly proposed where the long length filters in ANC are divided into a number of equal partitions and suitably assembled to perform the FXLMS algorithm in the frequency domain. The complexity of this proposed frequency domain partitioned block FXLMS (FPBFXLMS) algorithm is quite reduced compared to the conventional FXLMS algorithm. It is further reduced by merging one fast Fourier transform (FFT)-inverse fast Fourier transform (IFFT) combination to derive the reduced structure FPBFXLMS (RFPBFXLMS) algorithm. Computational complexity analysis for different orders of filter and partition size are presented. Systematic computer simulations are carried out for both the proposed partitioned block ANC algorithms to show its accuracy compared to the time domain FXLMS algorithm.

  1. High efficiency

    Energy Technology Data Exchange (ETDEWEB)

    Baumann, G.

    1984-05-18

    The surgeon wants to have phlebograms with good contrast, which should show only the deep venous system and leaks to the superficial system, that means, the insufficient communicating veins and the inflow of the big and small saphenous vein into the deep vein must be visible. The most frequent causes for X-ray-prints of bad quality are: a too high position of the stowing, too high puncture at the back of the foot, bad focussing without showing the ankle joint or the popliteal region and too narrow sections of the X-ray-films as well as too late exposures with fullfilling of the total superficial venous system and extreme superposition on the film.

  2. A novel class of highly efficient and accurate time-integrators in nonlinear computational mechanics

    Science.gov (United States)

    Wang, Xuechuan; Atluri, Satya N.

    2017-05-01

    A new class of time-integrators is presented for strongly nonlinear dynamical systems. These algorithms are far superior to the currently common time integrators in computational efficiency and accuracy. These three algorithms are based on a local variational iteration method applied over a finite interval of time. By using Chebyshev polynomials as trial functions and Dirac-Delta functions as the test functions over the finite time interval, the three algorithms are developed into three different discrete time-integrators through the collocation method. These time integrators are labeled as Chebyshev local iterative collocation methods. Through examples of the forced Duffing oscillator, the Lorenz system, and the multiple coupled Duffing equations (which arise as semi-discrete equations for beams, plates and shells undergoing large deformations), it is shown that the new algorithms are far superior to the 4th order Runge-Kutta and ODE45 of MATLAB, in predicting the chaotic responses of strongly nonlinear dynamical systems.

  3. A novel class of highly efficient and accurate time-integrators in nonlinear computational mechanics

    Science.gov (United States)

    Wang, Xuechuan; Atluri, Satya N.

    2017-01-01

    A new class of time-integrators is presented for strongly nonlinear dynamical systems. These algorithms are far superior to the currently common time integrators in computational efficiency and accuracy. These three algorithms are based on a local variational iteration method applied over a finite interval of time. By using Chebyshev polynomials as trial functions and Dirac-Delta functions as the test functions over the finite time interval, the three algorithms are developed into three different discrete time-integrators through the collocation method. These time integrators are labeled as Chebyshev local iterative collocation methods. Through examples of the forced Duffing oscillator, the Lorenz system, and the multiple coupled Duffing equations (which arise as semi-discrete equations for beams, plates and shells undergoing large deformations), it is shown that the new algorithms are far superior to the 4th order Runge-Kutta and ODE45 of MATLAB, in predicting the chaotic responses of strongly nonlinear dynamical systems.

  4. Interaction Entropy: A New Paradigm for Highly Efficient and Reliable Computation of Protein-Ligand Binding Free Energy.

    Science.gov (United States)

    Duan, Lili; Liu, Xiao; Zhang, John Z H

    2016-05-04

    Efficient and reliable calculation of protein-ligand binding free energy is a grand challenge in computational biology and is of critical importance in drug design and many other molecular recognition problems. The main challenge lies in the calculation of entropic contribution to protein-ligand binding or interaction systems. In this report, we present a new interaction entropy method which is theoretically rigorous, computationally efficient, and numerically reliable for calculating entropic contribution to free energy in protein-ligand binding and other interaction processes. Drastically different from the widely employed but extremely expensive normal mode method for calculating entropy change in protein-ligand binding, the new method calculates the entropic component (interaction entropy or -TΔS) of the binding free energy directly from molecular dynamics simulation without any extra computational cost. Extensive study of over a dozen randomly selected protein-ligand binding systems demonstrated that this interaction entropy method is both computationally efficient and numerically reliable and is vastly superior to the standard normal mode approach. This interaction entropy paradigm introduces a novel and intuitive conceptual understanding of the entropic effect in protein-ligand binding and other general interaction systems as well as a practical method for highly efficient calculation of this effect.

  5. Simple a posteriori slope limiter (Post Limiter) for high resolution and efficient flow computations

    Science.gov (United States)

    Kitamura, Keiichi; Hashimoto, Atsushi

    2017-07-01

    A simple and efficient a posteriori slope limiter (;Post Limiter;) is proposed for compressible Navier-Stokes and Euler equations, and examined in 1D and 2D. The Post Limiter tries to employ un-limited solutions where and when possible (even at shocks), and blend the un-limited and (1st-order) limited solutions smoothly, leading to equivalently four times resolution in 1D. This idea was inspired by a posteriori limiting approaches originally developed by Clain et al. (2011) [18] for higher-order flow computations, but proposed here is an alternative suitable and simplified for 2nd-order spatial accuracy with improved both solution and convergence. In fact, any iteration processes are no longer required to determine optimal orders of accuracy, since the limited and un-limited values are available at one time at 2nd-order. In 2D, several numerical examples have been dealt with, and both the κ = 1 / 3 MUSCL (in a structured solver) and Green-Gauss (in an unstructured solver) reconstructions demonstrated resolution improvement (nearly 4 × 4 times), convergence acceleration, and removal of numerical noises. Even on triangular meshes (on which least-squares reconstruction is used), the unstructured solver showed the improved solutions if cell geometries (cell-orientation angles) are properly taken into account. Therefore, the Post Limiter is readily incorporated into existing codes.

  6. Power-efficient computer architectures recent advances

    CERN Document Server

    Själander, Magnus; Kaxiras, Stefanos

    2014-01-01

    As Moore's Law and Dennard scaling trends have slowed, the challenges of building high-performance computer architectures while maintaining acceptable power efficiency levels have heightened. Over the past ten years, architecture techniques for power efficiency have shifted from primarily focusing on module-level efficiencies, toward more holistic design styles based on parallelism and heterogeneity. This work highlights and synthesizes recent techniques and trends in power-efficient computer architecture.Table of Contents: Introduction / Voltage and Frequency Management / Heterogeneity and Sp

  7. High throughput computing to improve efficiency of predicting protein stability change upon mutation.

    Science.gov (United States)

    Wu, Chao-Chin; Lai, Lien-Fu; Gromiha, M Michael; Huang, Liang-Tsung

    2014-01-01

    Predicting protein stability change upon mutation is important for protein design. Although several methods have been proposed to improve prediction accuracy it will be difficult to employ those methods when the required input information is incomplete. In this work, we integrated a fuzzy query model based on the knowledge-based approach to overcome this problem, and then we proposed a high throughput computing method based on parallel technologies in emerging cluster or grid systems to discriminate stability change. To improve the load balance of heterogeneous computing power in cluster and grid nodes, a variety of self-scheduling schemes have been implemented. Further, we have tested the method by performing different analyses and the results showed that the present method can process hundreds of predication queries in more reasonable response time and perform a super linear speedup to a maximum of 86.2 times. We have also established a website tool to implement the proposed method and it is available at http://bioinformatics.myweb.hinet.net/para.htm.

  8. Guest Editorial High Performance Computing (HPC) Applications for a More Resilient and Efficient Power Grid

    Energy Technology Data Exchange (ETDEWEB)

    Huang, Zhenyu Henry; Tate, Zeb; Abhyankar, Shrirang; Dong, Zhaoyang; Khaitan, Siddhartha; Min, Liang; Taylor, Gary

    2017-05-01

    The power grid has been evolving over the last 120 years, but it is seeing more changes in this decade and next than it has seen over the past century. In particular, the widespread deployment of intermittent renewable generation, smart loads and devices, hierarchical and distributed control technologies, phasor measurement units, energy storage, and widespread usage of electric vehicles will require fundamental changes in methods and tools for the operation and planning of the power grid. The resulting new dynamic and stochastic behaviors will demand the inclusion of more complexity in modeling the power grid. Solving such complex models in the traditional computing environment will be a major challenge. Along with the increasing complexity of power system models, the increasing complexity of smart grid data further adds to the prevailing challenges. In this environment, the myriad of smart sensors and meters in the power grid increase by multiple orders of magnitude, so do the volume and speed of the data. The information infrastructure will need to drastically change to support the exchange of enormous amounts of data as smart grid applications will need the capability to collect, assimilate, analyze and process the data, to meet real-time grid functions. High performance computing (HPC) holds the promise to enhance these functions, but it is a great resource that has not been fully explored and adopted for the power grid domain.

  9. Efficient computation of argumentation semantics

    CERN Document Server

    Liao, Beishui

    2013-01-01

    Efficient Computation of Argumentation Semantics addresses argumentation semantics and systems, introducing readers to cutting-edge decomposition methods that drive increasingly efficient logic computation in AI and intelligent systems. Such complex and distributed systems are increasingly used in the automation and transportation systems field, and particularly autonomous systems, as well as more generic intelligent computation research. The Series in Intelligent Systems publishes titles that cover state-of-the-art knowledge and the latest advances in research and development in intelligen

  10. Growing Cloud Computing Efficiency

    Directory of Open Access Journals (Sweden)

    Dr. Mohamed F. AlAjmi, Dr. Arun Sharma, Shakir Khan

    2012-05-01

    Full Text Available Cloud computing is basically altering the expectation for how and when computing, storage and networking assets should be allocated, managed and devoted. End-users are progressively more sensitive in response time of services they ingest. Service Developers wish for the Service Providers to make sure or give the ability for dynamically assigning and managing resources in respond to alter the demand patterns in real-time. Ultimately, Service Providers are under anxiety to build their infrastructure to facilitate real-time end-to-end visibility and energetic resource management with well grained control to decrease total cost of tenure for improving quickness. What is required to rethink of the underlying operating system and management infrastructure to put up the on-going renovation of data centre from the traditional server-centric architecture model to a cloud or network centric model? This paper projects and describes a indication model for a network centric data centre infrastructure management heap that make use of it and validates key ideas that have enabled dynamism, the quality of being scalable, reliability and security in the telecommunication industry to the computing engineering. Finally, the paper will explain a proof of concept classification that was implemented to show how dynamic resource management can be enforced to enable real-time service guarantee for network centric data centre architecture.

  11. Scalable Energy Efficiency with Resilience for High Performance Computing Systems: A Quantitative Methodology

    Energy Technology Data Exchange (ETDEWEB)

    Tan, Li [Univ. of California, Riverside, CA (United States); Chen, Zizhong [Univ. of California, Riverside, CA (United States); Song, Shuaiwen Leon [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2015-11-16

    Energy efficiency and resilience are two crucial challenges for HPC systems to reach exascale. While energy efficiency and resilience issues have been extensively studied individually, little has been done to understand the interplay between energy efficiency and resilience for HPC systems. Decreasing the supply voltage associated with a given operating frequency for processors and other CMOS-based components can significantly reduce power consumption. However, this often raises system failure rates and consequently increases application execution time. In this work, we present an energy saving undervolting approach that leverages the mainstream resilience techniques to tolerate the increased failures caused by undervolting.

  12. Scalable Energy Efficiency with Resilience for High Performance Computing Systems: A Quantitative Methodology

    Energy Technology Data Exchange (ETDEWEB)

    Tan, Li; Chen, Zizhong; Song, Shuaiwen

    2016-01-18

    Energy efficiency and resilience are two crucial challenges for HPC systems to reach exascale. While energy efficiency and resilience issues have been extensively studied individually, little has been done to understand the interplay between energy efficiency and resilience for HPC systems. Decreasing the supply voltage associated with a given operating frequency for processors and other CMOS-based components can significantly reduce power consumption. However, this often raises system failure rates and consequently increases application execution time. In this work, we present an energy saving undervolting approach that leverages the mainstream resilience techniques to tolerate the increased failures caused by undervolting.

  13. Investigating the Interplay between Energy Efficiency and Resilience in High Performance Computing

    Energy Technology Data Exchange (ETDEWEB)

    Tan, Li; Song, Shuaiwen; Wu, Panruo; Chen, Zizhong; Ge, Rong; Kerbyson, Darren J.

    2015-05-29

    Energy efficiency and resilience are two crucial challenges for HPC systems to reach exascale. While energy efficiency and resilience issues have been extensively studied individually, little has been done to understand the interplay between energy efficiency and resilience for HPC systems. Decreasing the supply voltage associated with a given operating frequency for processors and other CMOS-based components can significantly reduce power consumption. However, this often raises system failure rates and consequently increases application execution time. In this work, we present an energy saving undervolting approach that leverages the mainstream resilience techniques to tolerate the increased failures caused by undervolting.

  14. Highly accurate and efficient self-force computations using time-domain methods: Error estimates, validation, and optimization

    CERN Document Server

    Thornburg, Jonathan

    2010-01-01

    If a small "particle" of mass $\\mu M$ (with $\\mu \\ll 1$) orbits a Schwarzschild or Kerr black hole of mass $M$, the particle is subject to an $\\O(\\mu)$ radiation-reaction "self-force". Here I argue that it's valuable to compute this self-force highly accurately (relative error of $\\ltsim 10^{-6}$) and efficiently, and I describe techniques for doing this and for obtaining and validating error estimates for the computation. I use an adaptive-mesh-refinement (AMR) time-domain numerical integration of the perturbation equations in the Barack-Ori mode-sum regularization formalism; this is efficient, yet allows easy generalization to arbitrary particle orbits. I focus on the model problem of a scalar particle in a circular geodesic orbit in Schwarzschild spacetime. The mode-sum formalism gives the self-force as an infinite sum of regularized spherical-harmonic modes $\\sum_{\\ell=0}^\\infty F_{\\ell,\\reg}$, with $F_{\\ell,\\reg}$ (and an "internal" error estimate) computed numerically for $\\ell \\ltsim 30$ and estimated ...

  15. Efficient modelling of aerodynamic flows in the boundary layer for high performance computing

    CSIR Research Space (South Africa)

    Smith, L

    2011-01-01

    Full Text Available full viscous solution. The boundary-layer solution is coupled to an existing inviscid solver. Coupling occurs by moving the wall to a streamline at the computed boundary layer thickness and treating it as a slip boundary, then solving the flow again...

  16. Development and Modelling of High-Efficiency Computing Structure for Digital Signal Processing

    CERN Document Server

    Sharma, Annapurna; Lee, Hoon Jae

    2011-01-01

    The paper is devoted to problem of spline approximation. A new method of nodes location for curves and surfaces computer construction by means of B-splines and results of simulink-modeling is presented. The advantages of this paper is that we comprise the basic spline with classical polynomials both on accuracy, as well as degree of paralleling calculations are also shown.

  17. Toward Computational Design of High-Efficiency Photovoltaics from First-Principles

    Science.gov (United States)

    2016-08-15

    pursued in materials design. 2. Interfacial Oxygen Vacancies as a Potential Cause of Hysteresis in Perovskite Solar Cells 3 Figure 2: (a) Electron...function of time during the recombination process for intact TiO2 and TiO2 with different VO sites. Organometal halide perovskite solar cells ...and yellow, respectively. 4 Despite rapid progress in the efficiency of organohalide perovskite based solar cells , physical mechanisms underlying

  18. The feasibility of an efficient drug design method with high-performance computers.

    Science.gov (United States)

    Yamashita, Takefumi; Ueda, Akihiko; Mitsui, Takashi; Tomonaga, Atsushi; Matsumoto, Shunji; Kodama, Tatsuhiko; Fujitani, Hideaki

    2015-01-01

    In this study, we propose a supercomputer-assisted drug design approach involving all-atom molecular dynamics (MD)-based binding free energy prediction after the traditional design/selection step. Because this prediction is more accurate than the empirical binding affinity scoring of the traditional approach, the compounds selected by the MD-based prediction should be better drug candidates. In this study, we discuss the applicability of the new approach using two examples. Although the MD-based binding free energy prediction has a huge computational cost, it is feasible with the latest 10 petaflop-scale computer. The supercomputer-assisted drug design approach also involves two important feedback procedures: The first feedback is generated from the MD-based binding free energy prediction step to the drug design step. While the experimental feedback usually provides binding affinities of tens of compounds at one time, the supercomputer allows us to simultaneously obtain the binding free energies of hundreds of compounds. Because the number of calculated binding free energies is sufficiently large, the compounds can be classified into different categories whose properties will aid in the design of the next generation of drug candidates. The second feedback, which occurs from the experiments to the MD simulations, is important to validate the simulation parameters. To demonstrate this, we compare the binding free energies calculated with various force fields to the experimental ones. The results indicate that the prediction will not be very successful, if we use an inaccurate force field. By improving/validating such simulation parameters, the next prediction can be made more accurate.

  19. XSTREAM: A Highly Efficient High Speed Real-time Satellite Data Acquisition and Processing System using Heterogeneous Computing

    Science.gov (United States)

    Pramod Kumar, K.; Mahendra, P.; Ramakrishna rReddy, V.; Tirupathi, T.; Akilan, A.; Usha Devi, R.; Anuradha, R.; Ravi, N.; Solanki, S. S.; Achary, K. K.; Satish, A. L.; Anshu, C.

    2014-11-01

    In the last decade, the remote sensing community has observed a significant growth in number of satellites, sensors and their resolutions, thereby increasing the volume of data to be processed each day. Satellite data processing is a complex and time consuming activity. It consists of various tasks, such as decode, decrypt, decompress, radiometric normalization, stagger corrections, ephemeris data processing for geometric corrections etc., and finally writing of the product in the form of an image file. Each task in the processing chain is sequential in nature and has different computing needs. Conventionally the processes are cascaded in a well organized workflow to produce the data products, which are executed on general purpose high-end servers / workstations in an offline mode. Hence, these systems are considered to be ineffective for real-time applications that require quick response and just-intime decision making such as disaster management, home land security and so on. This paper discusses anovel approach to processthe data online (as the data is being acquired) using a heterogeneous computing platform namely XSTREAM which has COTS hardware of CPUs, GPUs and FPGA. This paper focuses on the process architecture, re-engineering aspects and mapping of tasks to the right computing devicewithin the XSTREAM system, which makes it an ideal cost-effective platform for acquiring, processing satellite payload data in real-time and displaying the products in original resolution for quick response. The system has been tested for IRS CARTOSAT and RESOURCESAT series of satellites which have maximum data downlink speed of 210 Mbps.

  20. Comparison of the Computational Efficiency of the Original Versus Reformulated High-Fidelity Generalized Method of Cells

    Science.gov (United States)

    Arnold, Steven M; Bednarcyk, Brett; Aboydi, Jacob

    2004-01-01

    The High-Fidelity Generalized Method of Cells (HFGMC) micromechanics model has recently been reformulated by Bansal and Pindera (in the context of elastic phases with perfect bonding) to maximize its computational efficiency. This reformulated version of HFGMC has now been extended to include both inelastic phases and imperfect fiber-matrix bonding. The present paper presents an overview of the HFGMC theory in both its original and reformulated forms and a comparison of the results of the two implementations. The objective is to establish the correlation between the two HFGMC formulations and document the improved efficiency offered by the reformulation. The results compare the macro and micro scale predictions of the continuous reinforcement (doubly-periodic) and discontinuous reinforcement (triply-periodic) versions of both formulations into the inelastic regime, and, in the case of the discontinuous reinforcement version, with both perfect and weak interfacial bonding. The results demonstrate that identical predictions are obtained using either the original or reformulated implementations of HFGMC aside from small numerical differences in the inelastic regime due to the different implementation schemes used for the inelastic terms present in the two formulations. Finally, a direct comparison of execution times is presented for the original formulation and reformulation code implementations. It is shown that as the discretization employed in representing the composite repeating unit cell becomes increasingly refined (requiring a larger number of sub-volumes), the reformulated implementation becomes significantly (approximately an order of magnitude at best) more computationally efficient in both the continuous reinforcement (doubly-periodic) and discontinuous reinforcement (triply-periodic) cases.

  1. Efficient Architectural Framework for Cloud Computing

    Directory of Open Access Journals (Sweden)

    Souvik Pal

    2012-06-01

    Full Text Available Cloud computing is that enables adaptive, favorable and on-demand network access to a collective pool of adjustable and configurable computing physical resources which networks, servers, bandwidth, storage that can be swiftly provisioned and released with negligible supervision endeavor or service provider interaction. From business prospective, the viable achievements of Cloud Computing and recent developments in Grid computing have brought the platform that has introduced virtualization technology into the era of high performance computing. However, clouds are Internet-based concept and try to disguise complexity overhead for end users. Cloud service providers (CSPs use many structural designs combined with self-service capabilities and ready-to-use facilities for computing resources, which are enabled through network infrastructure especially the internet which is an important consideration. This paper provides an efficient architectural Framework for cloud computing that may lead to better performance and faster access.

  2. Efficient GPU-based skyline computation

    DEFF Research Database (Denmark)

    Bøgh, Kenneth Sejdenfaden; Assent, Ira; Magnani, Matteo

    2013-01-01

    The skyline operator for multi-criteria search returns the most interesting points of a data set with respect to any monotone preference function. Existing work has almost exclusively focused on efficiently computing skylines on one or more CPUs, ignoring the high parallelism possible in GPUs. In...

  3. Efficient computation of spaced seeds

    Directory of Open Access Journals (Sweden)

    Ilie Silvana

    2012-02-01

    Full Text Available Abstract Background The most frequently used tools in bioinformatics are those searching for similarities, or local alignments, between biological sequences. Since the exact dynamic programming algorithm is quadratic, linear-time heuristics such as BLAST are used. Spaced seeds are much more sensitive than the consecutive seed of BLAST and using several seeds represents the current state of the art in approximate search for biological sequences. The most important aspect is computing highly sensitive seeds. Since the problem seems hard, heuristic algorithms are used. The leading software in the common Bernoulli model is the SpEED program. Findings SpEED uses a hill climbing method based on the overlap complexity heuristic. We propose a new algorithm for this heuristic that improves its speed by over one order of magnitude. We use the new implementation to compute improved seeds for several software programs. We compute as well multiple seeds of the same weight as MegaBLAST, that greatly improve its sensitivity. Conclusion Multiple spaced seeds are being successfully used in bioinformatics software programs. Enabling researchers to compute very fast high quality seeds will help expanding the range of their applications.

  4. Efficient computation of optimal actions.

    Science.gov (United States)

    Todorov, Emanuel

    2009-07-14

    Optimal choice of actions is a fundamental problem relevant to fields as diverse as neuroscience, psychology, economics, computer science, and control engineering. Despite this broad relevance the abstract setting is similar: we have an agent choosing actions over time, an uncertain dynamical system whose state is affected by those actions, and a performance criterion that the agent seeks to optimize. Solving problems of this kind remains hard, in part, because of overly generic formulations. Here, we propose a more structured formulation that greatly simplifies the construction of optimal control laws in both discrete and continuous domains. An exhaustive search over actions is avoided and the problem becomes linear. This yields algorithms that outperform Dynamic Programming and Reinforcement Learning, and thereby solve traditional problems more efficiently. Our framework also enables computations that were not possible before: composing optimal control laws by mixing primitives, applying deterministic methods to stochastic systems, quantifying the benefits of error tolerance, and inferring goals from behavioral data via convex optimization. Development of a general class of easily solvable problems tends to accelerate progress--as linear systems theory has done, for example. Our framework may have similar impact in fields where optimal choice of actions is relevant.

  5. Changing computing paradigms towards power efficiency.

    Science.gov (United States)

    Klavík, Pavel; Malossi, A Cristiano I; Bekas, Costas; Curioni, Alessandro

    2014-06-28

    Power awareness is fast becoming immensely important in computing, ranging from the traditional high-performance computing applications to the new generation of data centric workloads. In this work, we describe our efforts towards a power-efficient computing paradigm that combines low- and high-precision arithmetic. We showcase our ideas for the widely used kernel of solving systems of linear equations that finds numerous applications in scientific and engineering disciplines as well as in large-scale data analytics, statistics and machine learning. Towards this goal, we developed tools for the seamless power profiling of applications at a fine-grain level. In addition, we verify here previous work on post-FLOPS/W metrics and show that these can shed much more light in the power/energy profile of important applications.

  6. A Practical Evaluation of a High-Security Energy-Efficient Gateway for IoT Fog Computing Applications

    Science.gov (United States)

    Castedo, Luis

    2017-01-01

    Fog computing extends cloud computing to the edge of a network enabling new Internet of Things (IoT) applications and services, which may involve critical data that require privacy and security. In an IoT fog computing system, three elements can be distinguished: IoT nodes that collect data, the cloud, and interconnected IoT gateways that exchange messages with the IoT nodes and with the cloud. This article focuses on securing IoT gateways, which are assumed to be constrained in terms of computational resources, but that are able to offload some processing from the cloud and to reduce the latency in the responses to the IoT nodes. However, it is usually taken for granted that IoT gateways have direct access to the electrical grid, which is not always the case: in mission-critical applications like natural disaster relief or environmental monitoring, it is common to deploy IoT nodes and gateways in large areas where electricity comes from solar or wind energy that charge the batteries that power every device. In this article, how to secure IoT gateway communications while minimizing power consumption is analyzed. The throughput and power consumption of Rivest–Shamir–Adleman (RSA) and Elliptic Curve Cryptography (ECC) are considered, since they are really popular, but have not been thoroughly analyzed when applied to IoT scenarios. Moreover, the most widespread Transport Layer Security (TLS) cipher suites use RSA as the main public key-exchange algorithm, but the key sizes needed are not practical for most IoT devices and cannot be scaled to high security levels. In contrast, ECC represents a much lighter and scalable alternative. Thus, RSA and ECC are compared for equivalent security levels, and power consumption and data throughput are measured using a testbed of IoT gateways. The measurements obtained indicate that, in the specific fog computing scenario proposed, ECC is clearly a much better alternative than RSA, obtaining energy consumption reductions of up

  7. A Practical Evaluation of a High-Security Energy-Efficient Gateway for IoT Fog Computing Applications.

    Science.gov (United States)

    Suárez-Albela, Manuel; Fernández-Caramés, Tiago M; Fraga-Lamas, Paula; Castedo, Luis

    2017-08-29

    Fog computing extends cloud computing to the edge of a network enabling new Internet of Things (IoT) applications and services, which may involve critical data that require privacy and security. In an IoT fog computing system, three elements can be distinguished: IoT nodes that collect data, the cloud, and interconnected IoT gateways that exchange messages with the IoT nodes and with the cloud. This article focuses on securing IoT gateways, which are assumed to be constrained in terms of computational resources, but that are able to offload some processing from the cloud and to reduce the latency in the responses to the IoT nodes. However, it is usually taken for granted that IoT gateways have direct access to the electrical grid, which is not always the case: in mission-critical applications like natural disaster relief or environmental monitoring, it is common to deploy IoT nodes and gateways in large areas where electricity comes from solar or wind energy that charge the batteries that power every device. In this article, how to secure IoT gateway communications while minimizing power consumption is analyzed. The throughput and power consumption of Rivest-Shamir-Adleman (RSA) and Elliptic Curve Cryptography (ECC) are considered, since they are really popular, but have not been thoroughly analyzed when applied to IoT scenarios. Moreover, the most widespread Transport Layer Security (TLS) cipher suites use RSA as the main public key-exchange algorithm, but the key sizes needed are not practical for most IoT devices and cannot be scaled to high security levels. In contrast, ECC represents a much lighter and scalable alternative. Thus, RSA and ECC are compared for equivalent security levels, and power consumption and data throughput are measured using a testbed of IoT gateways. The measurements obtained indicate that, in the specific fog computing scenario proposed, ECC is clearly a much better alternative than RSA, obtaining energy consumption reductions of up to

  8. Parallel high-performance grid computing: capabilities and opportunities of a novel demanding service and business class allowing highest resource efficiency.

    Science.gov (United States)

    Kepper, Nick; Ettig, Ramona; Dickmann, Frank; Stehr, Rene; Grosveld, Frank G; Wedemann, Gero; Knoch, Tobias A

    2010-01-01

    Especially in the life-science and the health-care sectors the huge IT requirements are imminent due to the large and complex systems to be analysed and simulated. Grid infrastructures play here a rapidly increasing role for research, diagnostics, and treatment, since they provide the necessary large-scale resources efficiently. Whereas grids were first used for huge number crunching of trivially parallelizable problems, increasingly parallel high-performance computing is required. Here, we show for the prime example of molecular dynamic simulations how the presence of large grid clusters including very fast network interconnects within grid infrastructures allows now parallel high-performance grid computing efficiently and thus combines the benefits of dedicated super-computing centres and grid infrastructures. The demands for this service class are the highest since the user group has very heterogeneous requirements: i) two to many thousands of CPUs, ii) different memory architectures, iii) huge storage capabilities, and iv) fast communication via network interconnects, are all needed in different combinations and must be considered in a highly dedicated manner to reach highest performance efficiency. Beyond, advanced and dedicated i) interaction with users, ii) the management of jobs, iii) accounting, and iv) billing, not only combines classic with parallel high-performance grid usage, but more importantly is also able to increase the efficiency of IT resource providers. Consequently, the mere "yes-we-can" becomes a huge opportunity like e.g. the life-science and health-care sectors as well as grid infrastructures by reaching higher level of resource efficiency.

  9. High efficiency incandescent lighting

    Science.gov (United States)

    Bermel, Peter; Ilic, Ognjen; Chan, Walker R.; Musabeyoglu, Ahmet; Cukierman, Aviv Ruben; Harradon, Michael Robert; Celanovic, Ivan; Soljacic, Marin

    2014-09-02

    Incandescent lighting structure. The structure includes a thermal emitter that can, but does not have to, include a first photonic crystal on its surface to tailor thermal emission coupled to, in a high-view-factor geometry, a second photonic filter selected to reflect infrared radiation back to the emitter while passing visible light. This structure is highly efficient as compared to standard incandescent light bulbs.

  10. Energy efficient distributed computing systems

    CERN Document Server

    Lee, Young-Choon

    2012-01-01

    The energy consumption issue in distributed computing systems raises various monetary, environmental and system performance concerns. Electricity consumption in the US doubled from 2000 to 2005.  From a financial and environmental standpoint, reducing the consumption of electricity is important, yet these reforms must not lead to performance degradation of the computing systems.  These contradicting constraints create a suite of complex problems that need to be resolved in order to lead to 'greener' distributed computing systems.  This book brings together a group of outsta

  11. Quantum computing: Efficient fault tolerance

    Science.gov (United States)

    Gottesman, Daniel

    2016-12-01

    Dealing with errors in a quantum computer typically requires complex programming and many additional quantum bits. A technique for controlling errors has been proposed that alleviates both of these problems.

  12. Efficient computation of k-Nearest Neighbour Graphs for large high-dimensional data sets on GPU clusters.

    Directory of Open Access Journals (Sweden)

    Ali Dashti

    Full Text Available This paper presents an implementation of the brute-force exact k-Nearest Neighbor Graph (k-NNG construction for ultra-large high-dimensional data cloud. The proposed method uses Graphics Processing Units (GPUs and is scalable with multi-levels of parallelism (between nodes of a cluster, between different GPUs on a single node, and within a GPU. The method is applicable to homogeneous computing clusters with a varying number of nodes and GPUs per node. We achieve a 6-fold speedup in data processing as compared with an optimized method running on a cluster of CPUs and bring a hitherto impossible [Formula: see text]-NNG generation for a dataset of twenty million images with 15 k dimensionality into the realm of practical possibility.

  13. PathoQC: Computationally Efficient Read Preprocessing and Quality Control for High-Throughput Sequencing Data Sets.

    Science.gov (United States)

    Hong, Changjin; Manimaran, Solaiappan; Johnson, William Evan

    2014-01-01

    Quality control and read preprocessing are critical steps in the analysis of data sets generated from high-throughput genomic screens. In the most extreme cases, improper preprocessing can negatively affect downstream analyses and may lead to incorrect biological conclusions. Here, we present PathoQC, a streamlined toolkit that seamlessly combines the benefits of several popular quality control software approaches for preprocessing next-generation sequencing data. PathoQC provides a variety of quality control options appropriate for most high-throughput sequencing applications. PathoQC is primarily developed as a module in the PathoScope software suite for metagenomic analysis. However, PathoQC is also available as an open-source Python module that can run as a stand-alone application or can be easily integrated into any bioinformatics workflow. PathoQC achieves high performance by supporting parallel computation and is an effective tool that removes technical sequencing artifacts and facilitates robust downstream analysis. The PathoQC software package is available at http://sourceforge.net/projects/PathoScope/.

  14. Highly efficient high temperature electrolysis

    DEFF Research Database (Denmark)

    Hauch, Anne; Ebbesen, Sune; Jensen, Søren Højgaard;

    2008-01-01

    High temperature electrolysis of water and steam may provide an efficient, cost effective and environmentally friendly production of H-2 Using electricity produced from sustainable, non-fossil energy sources. To achieve cost competitive electrolysis cells that are both high performing i.e. minimum...... internal resistance of the cell, and long-term stable, it is critical to develop electrode materials that are optimal for steam electrolysis. In this article electrolysis cells for electrolysis of water or steam at temperatures above 200 degrees C for production of H-2 are reviewed. High temperature...... electrolysis is favourable from a thermodynamic point of view, because a part of the required energy can be supplied as thermal heat, and the activation barrier is lowered increasing the H-2 production rate. Only two types of cells operating at high temperature (above 200 degrees C) have been described...

  15. Highly efficient sorghum transformation

    OpenAIRE

    Liu, Guoquan; Godwin, Ian D.

    2012-01-01

    A highly efficient microprojectile transformation system for sorghum (Sorghum bicolor L.) has been developed by using immature embryos (IEs) of inbred line Tx430. Co-bombardment was performed with the neomycin phosphotransferase II (nptII) gene and the green fluorescent protein (gfp) gene, both under the control of the maize ubiquitin1 (ubi1) promoter. After optimization of both tissue culture media and parameters of microprojectile transformation, 25 independent transgenic events were obtain...

  16. Efficient Computational Model of Hysteresis

    Science.gov (United States)

    Shields, Joel

    2005-01-01

    A recently developed mathematical model of the output (displacement) versus the input (applied voltage) of a piezoelectric transducer accounts for hysteresis. For the sake of computational speed, the model is kept simple by neglecting the dynamic behavior of the transducer. Hence, the model applies to static and quasistatic displacements only. A piezoelectric transducer of the type to which the model applies is used as an actuator in a computer-based control system to effect fine position adjustments. Because the response time of the rest of such a system is usually much greater than that of a piezoelectric transducer, the model remains an acceptably close approximation for the purpose of control computations, even though the dynamics are neglected. The model (see Figure 1) represents an electrically parallel, mechanically series combination of backlash elements, each having a unique deadband width and output gain. The zeroth element in the parallel combination has zero deadband width and, hence, represents a linear component of the input/output relationship. The other elements, which have nonzero deadband widths, are used to model the nonlinear components of the hysteresis loop. The deadband widths and output gains of the elements are computed from experimental displacement-versus-voltage data. The hysteresis curve calculated by use of this model is piecewise linear beyond deadband limits.

  17. Efficient computational noise in GLSL

    CERN Document Server

    McEwan, Ian; Gustavson, Stefan; Richardson, Mark

    2012-01-01

    We present GLSL implementations of Perlin noise and Perlin simplex noise that run fast enough for practical consideration on current generation GPU hardware. The key benefits are that the functions are purely computational, i.e. they use neither textures nor lookup tables, and that they are implemented in GLSL version 1.20, which means they are compatible with all current GLSL-capable platforms, including OpenGL ES 2.0 and WebGL 1.0. Their performance is on par with previously presented GPU implementations of noise, they are very convenient to use, and they scale well with increasing parallelism in present and upcoming GPU architectures.

  18. Energy-efficient quantum computing

    Science.gov (United States)

    Ikonen, Joni; Salmilehto, Juha; Möttönen, Mikko

    2017-04-01

    In the near future, one of the major challenges in the realization of large-scale quantum computers operating at low temperatures is the management of harmful heat loads owing to thermal conduction of cabling and dissipation at cryogenic components. This naturally raises the question that what are the fundamental limitations of energy consumption in scalable quantum computing. In this work, we derive the greatest lower bound for the gate error induced by a single application of a bosonic drive mode of given energy. Previously, such an error type has been considered to be inversely proportional to the total driving power, but we show that this limitation can be circumvented by introducing a qubit driving scheme which reuses and corrects drive pulses. Specifically, our method serves to reduce the average energy consumption per gate operation without increasing the average gate error. Thus our work shows that precise, scalable control of quantum systems can, in principle, be implemented without the introduction of excessive heat or decoherence.

  19. Computationally efficient algorithm for fast transients detection

    CERN Document Server

    Soudlenkov, Gene

    2011-01-01

    Computationally inexpensive algorithm for detecting of dispersed transients has been developed using Cumulative Sums (CUSUM) scheme for detecting abrupt changes in statistical characteristics of the signal. The efficiency of the algorithm is demonstrated on pulsar PSR J0835-4510.

  20. High-efficiency CARM

    Energy Technology Data Exchange (ETDEWEB)

    Bratman, V.L.; Kol`chugin, B.D.; Samsonov, S.V.; Volkov, A.B. [Institute of Applied Physics, Nizhny Novgorod (Russian Federation)

    1995-12-31

    The Cyclotron Autoresonance Maser (CARM) is a well-known variety of FEMs. Unlike the ubitron in which electrons move in a periodical undulator field, in the CARM the particles move along helical trajectories in a uniform magnetic field. Since it is much simpler to generate strong homogeneous magnetic fields than periodical ones for a relatively low electron energy ({Brit_pounds}{le}1-3 MeV) the period of particles` trajectories in the CARM can be sufficiently smaller than in the undulator in which, moreover, the field decreases rapidly in the transverse direction. In spite of this evident advantage, the number of papers on CARM is an order less than on ubitron, which is apparently caused by the low (not more than 10 %) CARM efficiency in experiments. At the same time, ubitrons operating in two rather complicated regimes-trapping and adiabatic deceleration of particles and combined undulator and reversed guiding fields - yielded efficiencies of 34 % and 27 %, respectively. The aim of this work is to demonstrate that high efficiency can be reached even for a simplest version of the CARM. In order to reduce sensitivity to an axial velocity spread of particles, a short interaction length where electrons underwent only 4-5 cyclotron oscillations was used in this work. Like experiments, a narrow anode outlet of a field-emission electron gun cut out the {open_quotes}most rectilinear{close_quotes} near-axis part of the electron beam. Additionally, magnetic field of a small correcting coil compensated spurious electron oscillations pumped by the anode aperture. A kicker in the form of a sloping to the axis frame with current provided a control value of rotary velocity at a small additional velocity spread. A simple cavity consisting of a cylindrical waveguide section restricted by a cut-off waveguide on the cathode side and by a Bragg reflector on the collector side was used as the CARM-oscillator microwave system.

  1. The EORTC computer-adaptive tests measuring physical functioning and fatigue exhibited high levels of measurement precision and efficiency

    DEFF Research Database (Denmark)

    Petersen, Morten Aa; Aaronson, Neil K; Arraras, Juan I

    2013-01-01

    The European Organisation for Research and Treatment of Cancer (EORTC) Quality of Life Group is developing a computer-adaptive test (CAT) version of the EORTC Quality of Life Questionnaire (QLQ-C30). We evaluated the measurement properties of the CAT versions of physical functioning (PF) and fati...

  2. The EORTC computer-adaptive tests measuring physical functioning and fatigue exhibited high levels of measurement precision and efficiency

    DEFF Research Database (Denmark)

    Petersen, Morten Aa; Aaronson, Neil K; Arraras, Juan I;

    2013-01-01

    The European Organisation for Research and Treatment of Cancer (EORTC) Quality of Life Group is developing a computer-adaptive test (CAT) version of the EORTC Quality of Life Questionnaire (QLQ-C30). We evaluated the measurement properties of the CAT versions of physical functioning (PF...

  3. Efficient Multi-Party Computation over Rings

    DEFF Research Database (Denmark)

    Cramer, Ronald; Fehr, Serge; Ishai, Yuval;

    2003-01-01

    by (boolean or arithmetic) circuits over finite fields. We are motivated by two limitations of these techniques: – Generality. Existing protocols do not apply to computation over more general algebraic structures (except via a brute-force simulation of computation in these structures). – Efficiency. The best...... known constant-round protocols do not efficiently scale even to the case of large finite fields. Our contribution goes in these two directions. First, we propose a basis for unconditionally secure MPC over an arbitrary ginite ring, an algebraic object with a much less nice structure than a field...... the usefulness of the above results by presenting a novel application of MPC over (non-field) rings to the round-efficient secure computation of the maximum function. Basic Research in Computer Science (www.brics.dk), funded by the Danish National Research Foundation....

  4. Efficient Multi-Party Computation over Rings

    DEFF Research Database (Denmark)

    Cramer, Ronald; Fehr, Serge; Ishai, Yuval

    2003-01-01

    by (boolean or arithmetic) circuits over finite fields. We are motivated by two limitations of these techniques: – Generality. Existing protocols do not apply to computation over more general algebraic structures (except via a brute-force simulation of computation in these structures). – Efficiency. The best...... known constant-round protocols do not efficiently scale even to the case of large finite fields. Our contribution goes in these two directions. First, we propose a basis for unconditionally secure MPC over an arbitrary ginite ring, an algebraic object with a much less nice structure than a field...... the usefulness of the above results by presenting a novel application of MPC over (non-field) rings to the round-efficient secure computation of the maximum function. Basic Research in Computer Science (www.brics.dk), funded by the Danish National Research Foundation....

  5. Resource-efficient linear optical quantum computation.

    Science.gov (United States)

    Browne, Daniel E; Rudolph, Terry

    2005-07-01

    We introduce a scheme for linear optics quantum computation, that makes no use of teleported gates, and requires stable interferometry over only the coherence length of the photons. We achieve a much greater degree of efficiency and a simpler implementation than previous proposals. We follow the "cluster state" measurement based quantum computational approach, and show how cluster states may be efficiently generated from pairs of maximally polarization entangled photons using linear optical elements. We demonstrate the universality and usefulness of generic parity measurements, as well as introducing the use of redundant encoding of qubits to enable utilization of destructive measurements--both features of use in a more general context.

  6. Efficient 3D nonlinear warping of computed tomography: two high-performance implementations using OpenGL

    Science.gov (United States)

    Levin, David; Dey, Damini; Slomka, Piotr

    2005-04-01

    We have implemented two hardware accelerated Thin Plate Spline (TPS) warping algorithms. The first algorithm is a hardware-software approach (HW-TPS) that uses OpenGL Vertex Shaders to perform a grid warp. The second is a Graphics Processor based approach (GPU-TPS) that uses the OpenGL Shading Language to perform all warping calculations on the GPU. Comparison with a software TPS algorithm was used to gauge the speed and quality of both hardware algorithms. Quality was analyzed visually and using the Sum of Absolute Difference (SAD) similarity metric. Warping was performed using 92 user-defined displacement vectors for 512x512x173 serial lung CT studies, matching normal-breathing and deep-inspiration scans. On a Xeon 2.2 Ghz machine with an ATI Radeon 9800XT GPU the GPU-TPS required 26.1 seconds to perform a per-voxel warp compared to 148.2 seconds for the software algorithm. The HW-TPS needed 1.63 seconds to warp the same study while the GPU-TPS required 1.94 seconds and the software grid transform required 22.8 seconds. The SAD values calculated between the outputs of each algorithm and the target CT volume were 15.2%, 15.4% and 15.5% for the HW-TPS, GPU-TPS and both software algorithms respectively. The computing power of ubiquitous 3D graphics cards can be exploited in medical image processing to provide order of magnitude acceleration of nonlinear warping algorithms without sacrificing output quality.

  7. Complexity-aware high efficiency video coding

    CERN Document Server

    Correa, Guilherme; Agostini, Luciano; Cruz, Luis A da Silva

    2016-01-01

    This book discusses computational complexity of High Efficiency Video Coding (HEVC) encoders with coverage extending from the analysis of HEVC compression efficiency and computational complexity to the reduction and scaling of its encoding complexity. After an introduction to the topic and a review of the state-of-the-art research in the field, the authors provide a detailed analysis of the HEVC encoding tools compression efficiency and computational complexity.  Readers will benefit from a set of algorithms for scaling the computational complexity of HEVC encoders, all of which take advantage from the flexibility of the frame partitioning structures allowed by the standard.  The authors also provide a set of early termination methods based on data mining and machine learning techniques, which are able to reduce the computational complexity required to find the best frame partitioning structures. The applicability of the proposed methods is finally exemplified with an encoding time control system that emplo...

  8. An Efficient Approach for Computing Silhouette Coefficients

    Directory of Open Access Journals (Sweden)

    Moh'd B. Al- Zoubi

    2008-01-01

    Full Text Available One popular approach for finding the best number of clusters (K in a data set is through computing the silhouette coefficients. The silhouette coefficients for different values of K, are first found and then the maximum value of these coefficients is chosen. However, computing the silhouette coefficient for different Ks is a very time consuming process. This is due to the amount of CPU time spent on distance calculations. A proposed approach to compute the silhouette coefficient quickly had been presented. The approach was based on decreasing the number of addition operations when computing distances. The results were efficient and more than 50% of the CPU time was achieved when applied to different data sets.

  9. Computationally Efficient Multiconfigurational Reactive Molecular Dynamics.

    Science.gov (United States)

    Yamashita, Takefumi; Peng, Yuxing; Knight, Chris; Voth, Gregory A

    2012-12-11

    It is a computationally demanding task to explicitly simulate the electronic degrees of freedom in a system to observe the chemical transformations of interest, while at the same time sampling the time and length scales required to converge statistical properties and thus reduce artifacts due to initial conditions, finite-size effects, and limited sampling. One solution that significantly reduces the computational expense consists of molecular models in which effective interactions between particles govern the dynamics of the system. If the interaction potentials in these models are developed to reproduce calculated properties from electronic structure calculations and/or ab initio molecular dynamics simulations, then one can calculate accurate properties at a fraction of the computational cost. Multiconfigurational algorithms model the system as a linear combination of several chemical bonding topologies to simulate chemical reactions, also sometimes referred to as "multistate". These algorithms typically utilize energy and force calculations already found in popular molecular dynamics software packages, thus facilitating their implementation without significant changes to the structure of the code. However, the evaluation of energies and forces for several bonding topologies per simulation step can lead to poor computational efficiency if redundancy is not efficiently removed, particularly with respect to the calculation of long-ranged Coulombic interactions. This paper presents accurate approximations (effective long-range interaction and resulting hybrid methods) and multiple-program parallelization strategies for the efficient calculation of electrostatic interactions in reactive molecular simulations.

  10. Fast fundamental frequency estimation: Making a statistically efficient estimator computationally efficient

    DEFF Research Database (Denmark)

    Nielsen, Jesper Kjær; Jensen, Tobias Lindstrøm; Jensen, Jesper Rindom

    2017-01-01

    estimator has a high computational complexity. In this paper, we propose an algorithm for lowering this complexity significantly by showing that the NLS estimator can be computed efficiently by solving two Toeplitz-plus-Hankel systems of equations and by exploiting the recursive-in-order matrix structures...

  11. Numerical aspects for efficient welding computational mechanics

    Directory of Open Access Journals (Sweden)

    Aburuga Tarek Kh.S.

    2014-01-01

    Full Text Available The effect of the residual stresses and strains is one of the most important parameter in the structure integrity assessment. A finite element model is constructed in order to simulate the multi passes mismatched submerged arc welding SAW which used in the welded tensile test specimen. Sequentially coupled thermal mechanical analysis is done by using ABAQUS software for calculating the residual stresses and distortion due to welding. In this work, three main issues were studied in order to reduce the time consuming during welding simulation which is the major problem in the computational welding mechanics (CWM. The first issue is dimensionality of the problem. Both two- and three-dimensional models are constructed for the same analysis type, shell element for two dimension simulation shows good performance comparing with brick element. The conventional method to calculate residual stress is by using implicit scheme that because of the welding and cooling time is relatively high. In this work, the author shows that it could use the explicit scheme with the mass scaling technique, and time consuming during the analysis will be reduced very efficiently. By using this new technique, it will be possible to simulate relatively large three dimensional structures.

  12. Computer Architecture Techniques for Power-Efficiency

    CERN Document Server

    Kaxiras, Stefanos

    2008-01-01

    In the last few years, power dissipation has become an important design constraint, on par with performance, in the design of new computer systems. Whereas in the past, the primary job of the computer architect was to translate improvements in operating frequency and transistor count into performance, now power efficiency must be taken into account at every step of the design process. While for some time, architects have been successful in delivering 40% to 50% annual improvement in processor performance, costs that were previously brushed aside eventually caught up. The most critical of these

  13. A Computational View of Market Efficiency

    CERN Document Server

    Hasanhodzic, Jasmina; Viola, Emanuele

    2009-01-01

    We propose to study market efficiency from a computational viewpoint. Borrowing from theoretical computer science, we define a market to be \\emph{efficient with respect to resources $S$} (e.g., time, memory) if no strategy using resources $S$ can make a profit. As a first step, we consider memory-$m$ strategies whose action at time $t$ depends only on the $m$ previous observations at times $t-m,...,t-1$. We introduce and study a simple model of market evolution, where strategies impact the market by their decision to buy or sell. We show that the effect of optimal strategies using memory $m$ can lead to "market conditions" that were not present initially, such as (1) market bubbles and (2) the possibility for a strategy using memory $m' > m$ to make a bigger profit than was initially possible. We suggest ours as a framework to rationalize the technological arms race of quantitative trading firms.

  14. Efficient quantum computing using coherent photon conversion.

    Science.gov (United States)

    Langford, N K; Ramelow, S; Prevedel, R; Munro, W J; Milburn, G J; Zeilinger, A

    2011-10-12

    Single photons are excellent quantum information carriers: they were used in the earliest demonstrations of entanglement and in the production of the highest-quality entanglement reported so far. However, current schemes for preparing, processing and measuring them are inefficient. For example, down-conversion provides heralded, but randomly timed, single photons, and linear optics gates are inherently probabilistic. Here we introduce a deterministic process--coherent photon conversion (CPC)--that provides a new way to generate and process complex, multiquanta states for photonic quantum information applications. The technique uses classically pumped nonlinearities to induce coherent oscillations between orthogonal states of multiple quantum excitations. One example of CPC, based on a pumped four-wave-mixing interaction, is shown to yield a single, versatile process that provides a full set of photonic quantum processing tools. This set satisfies the DiVincenzo criteria for a scalable quantum computing architecture, including deterministic multiqubit entanglement gates (based on a novel form of photon-photon interaction), high-quality heralded single- and multiphoton states free from higher-order imperfections, and robust, high-efficiency detection. It can also be used to produce heralded multiphoton entanglement, create optically switchable quantum circuits and implement an improved form of down-conversion with reduced higher-order effects. Such tools are valuable building blocks for many quantum-enabled technologies. Finally, using photonic crystal fibres we experimentally demonstrate quantum correlations arising from a four-colour nonlinear process suitable for CPC and use these measurements to study the feasibility of reaching the deterministic regime with current technology. Our scheme, which is based on interacting bosonic fields, is not restricted to optical systems but could also be implemented in optomechanical, electromechanical and superconducting

  15. COMPUTATIONALLY EFFICIENT PRIVATE INFORMATION RETRIEVAL PROTOCOL

    Directory of Open Access Journals (Sweden)

    A. V. Afanasyeva

    2016-03-01

    Full Text Available This paper describes a new computationally efficient private information retrieval protocol for one q-ary symbol retrieving. The main advantage of the proposed solution lies in a low computational complexity of information extraction procedure, as well as the constructive simplicity and flexibility in choosing the system parameters. Such results are based on cosets properties. The proposed protocol has communication complexity slightly worse than the best schemes at the moment, which is based on locally decodable codes, but it can be easily built for any parameters of the system, as opposed to codes. In comparison with similar solutions based on polynomials, the proposed method gains in computational complexity, which is important especially for servers which must service multiple requests from multiple users.

  16. Efficient Resource Management in Cloud Computing

    Directory of Open Access Journals (Sweden)

    Rushikesh Shingade

    2015-12-01

    Full Text Available Cloud computing, one of the widely used technology to provide cloud services for users who are charged for receiving services. In the aspect of a maximum number of resources, evaluating the performance of Cloud resource management policies are difficult to optimize efficiently. There are different simulation toolkits available for simulation and modelling the Cloud computing environment like GridSim CloudAnalyst, CloudSim, GreenCloud, CloudAuction etc. In proposed Efficient Resource Management in Cloud Computing (EFRE model, CloudSim is used as a simulation toolkit that allows simulation of DataCenter in Cloud computing system. The CloudSim toolkit also supports the creation of multiple virtual machines (VMs on a node of a DataCenter where cloudlets (user requests are assigned to virtual machines by scheduling policies. This paper represents, allocation policies, Time-Shared and Space-Shared are used for scheduling the cloudlets and compared with the constraints (metrics like total execution time, a number of resources and resource allocation algorithm. CloudSim has been used for simulations and the result of simulation demonstrate that Resource Management is effective.

  17. Computing with memory for energy-efficient robust systems

    CERN Document Server

    Paul, Somnath

    2013-01-01

    This book analyzes energy and reliability as major challenges faced by designers of computing frameworks in the nanometer technology regime.  The authors describe the existing solutions to address these challenges and then reveal a new reconfigurable computing platform, which leverages high-density nanoscale memory for both data storage and computation to maximize the energy-efficiency and reliability. The energy and reliability benefits of this new paradigm are illustrated and the design challenges are discussed. Various hardware and software aspects of this exciting computing paradigm are de

  18. A computationally efficient fuzzy control s

    Directory of Open Access Journals (Sweden)

    Abdel Badie Sharkawy

    2013-12-01

    Full Text Available This paper develops a decentralized fuzzy control scheme for MIMO nonlinear second order systems with application to robot manipulators via a combination of genetic algorithms (GAs and fuzzy systems. The controller for each degree of freedom (DOF consists of a feedforward fuzzy torque computing system and a feedback fuzzy PD system. The feedforward fuzzy system is trained and optimized off-line using GAs, whereas not only the parameters but also the structure of the fuzzy system is optimized. The feedback fuzzy PD system, on the other hand, is used to keep the closed-loop stable. The rule base consists of only four rules per each DOF. Furthermore, the fuzzy feedback system is decentralized and simplified leading to a computationally efficient control scheme. The proposed control scheme has the following advantages: (1 it needs no exact dynamics of the system and the computation is time-saving because of the simple structure of the fuzzy systems and (2 the controller is robust against various parameters and payload uncertainties. The computational complexity of the proposed control scheme has been analyzed and compared with previous works. Computer simulations show that this controller is effective in achieving the control goals.

  19. Structured Parallel Programming Patterns for Efficient Computation

    CERN Document Server

    McCool, Michael; Robison, Arch

    2012-01-01

    Programming is now parallel programming. Much as structured programming revolutionized traditional serial programming decades ago, a new kind of structured programming, based on patterns, is relevant to parallel programming today. Parallel computing experts and industry insiders Michael McCool, Arch Robison, and James Reinders describe how to design and implement maintainable and efficient parallel algorithms using a pattern-based approach. They present both theory and practice, and give detailed concrete examples using multiple programming models. Examples are primarily given using two of th

  20. Energy Efficiency in Computing (1/2)

    CERN Document Server

    CERN. Geneva

    2016-01-01

    As manufacturers improve the silicon process, truly low energy computing is becoming a reality - both in servers and in the consumer space. This series of lectures covers a broad spectrum of aspects related to energy efficient computing - from circuits to datacentres. We will discuss common trade-offs and basic components, such as processors, memory and accelerators. We will also touch on the fundamentals of modern datacenter design and operation. Lecturer's short bio: Andrzej Nowak has 10 years of experience in computing technologies, primarily from CERN openlab and Intel. At CERN, he managed a research lab collaborating with Intel and was part of the openlab Chief Technology Office. Andrzej also worked closely and initiated projects with the private sector (e.g. HP and Google), as well as international research institutes, such as EPFL. Currently, Andrzej acts as a consultant on technology and innovation with TIK Services (http://tik.services), and runs a peer-to-peer lending start-up. NB! All Academic L...

  1. Energy efficiency indicators for high electric-load buildings

    Energy Technology Data Exchange (ETDEWEB)

    Aebischer, Bernard; Balmer, Markus A.; Kinney, Satkartar; Le Strat, Pascale; Shibata, Yoshiaki; Varone, Frederic

    2003-06-01

    Energy per unit of floor area is not an adequate indicator for energy efficiency in high electric-load buildings. For two activities, restaurants and computer centres, alternative indicators for energy efficiency are discussed.

  2. A primer on the energy efficiency of computing

    Energy Technology Data Exchange (ETDEWEB)

    Koomey, Jonathan G. [Research Fellow, Steyer-Taylor Center for Energy Policy and Finance, Stanford University (United States)

    2015-03-30

    The efficiency of computing at peak output has increased rapidly since the dawn of the computer age. This paper summarizes some of the key factors affecting the efficiency of computing in all usage modes. While there is still great potential for improving the efficiency of computing devices, we will need to alter how we do computing in the next few decades because we are finally approaching the limits of current technologies.

  3. Characterizing and Implementing Efficient Primitives for Privacy-Preserving Computation

    Science.gov (United States)

    2015-07-01

    4 Efficient Mobile Oblivious Computation ( EMOC ) .................................................................... 4 Memory...Assumptions and Procedures  Efficient Mobile Oblivious Computation ( EMOC )  Mobile applications increasingly require users to surrender private...In this effort, we developed Efficient Mobile Oblivious Computation ( EMOC ), a set of SFE protocols customized for the mobile platform. Using

  4. Computational Biology and High Performance Computing 2000

    Energy Technology Data Exchange (ETDEWEB)

    Simon, Horst D.; Zorn, Manfred D.; Spengler, Sylvia J.; Shoichet, Brian K.; Stewart, Craig; Dubchak, Inna L.; Arkin, Adam P.

    2000-10-19

    The pace of extraordinary advances in molecular biology has accelerated in the past decade due in large part to discoveries coming from genome projects on human and model organisms. The advances in the genome project so far, happening well ahead of schedule and under budget, have exceeded any dreams by its protagonists, let alone formal expectations. Biologists expect the next phase of the genome project to be even more startling in terms of dramatic breakthroughs in our understanding of human biology, the biology of health and of disease. Only today can biologists begin to envision the necessary experimental, computational and theoretical steps necessary to exploit genome sequence information for its medical impact, its contribution to biotechnology and economic competitiveness, and its ultimate contribution to environmental quality. High performance computing has become one of the critical enabling technologies, which will help to translate this vision of future advances in biology into reality. Biologists are increasingly becoming aware of the potential of high performance computing. The goal of this tutorial is to introduce the exciting new developments in computational biology and genomics to the high performance computing community.

  5. Energy Efficiency in Computing (2/2)

    CERN Document Server

    CERN. Geneva

    2016-01-01

    We will start the second day of our energy efficient computing series with a brief discussion of software and the impact it has on energy consumption. A second major point of this lecture will be the current state of research and a few future technologies, ranging from mainstream (e.g. the Internet of Things) to exotic. Lecturer's short bio: Andrzej Nowak has 10 years of experience in computing technologies, primarily from CERN openlab and Intel. At CERN, he managed a research lab collaborating with Intel and was part of the openlab Chief Technology Office. Andrzej also worked closely and initiated projects with the private sector (e.g. HP and Google), as well as international research institutes, such as EPFL. Currently, Andrzej acts as a consultant on technology and innovation with TIK Services (http://tik.services), and runs a peer-to-peer lending start-up. NB! All Academic Lectures are recorded. No webcast! Because of a problem of the recording equipment, this lecture will be repeated for recording pu...

  6. High Energy Efficiency Air Conditioning

    Energy Technology Data Exchange (ETDEWEB)

    Edward McCullough; Patrick Dhooge; Jonathan Nimitz

    2003-12-31

    This project determined the performance of a new high efficiency refrigerant, Ikon B, in a residential air conditioner designed to use R-22. The refrigerant R-22, used in residential and small commercial air conditioners, is being phased out of production in developed countries beginning this year because of concerns regarding its ozone depletion potential. Although a replacement refrigerant, R-410A, is available, it operates at much higher pressure than R-22 and requires new equipment. R-22 air conditioners will continue to be in use for many years to come. Air conditioning is a large part of expensive summer peak power use in many parts of the U.S. Previous testing and computer simulations of Ikon B indicated that it would have 20 - 25% higher coefficient of performance (COP, the amount of cooling obtained per energy used) than R-22 in an air-cooled air conditioner. In this project, a typical new R-22 residential air conditioner was obtained, installed in a large environmental chamber, instrumented, and run both with its original charge of R-22 and then with Ikon B. In the environmental chamber, controlled temperature and humidity could be maintained to obtain repeatable and comparable energy use results. Tests with Ikon B included runs with and without a power controller, and an extended run for several months with subsequent analyses to check compatibility of Ikon B with the air conditioner materials and lubricant. Baseline energy use of the air conditioner with its original R-22 charge was measured at 90 deg F and 100 deg F. After changeover to Ikon B and a larger expansion orifice, energy use was measured at 90 deg F and 100 deg F. Ikon B proved to have about 19% higher COP at 90 deg F and about 26% higher COP at 100 deg F versus R-22. Ikon B had about 20% lower cooling capacity at 90 deg F and about 17% lower cooling capacity at 100 deg F versus R-22 in this system. All results over multiple runs were within 1% relative standard deviation (RSD). All of these

  7. High Efficiency Engine Technologies Program

    Energy Technology Data Exchange (ETDEWEB)

    Rich Kruiswyk

    2010-07-13

    Caterpillar's Product Development and Global Technology Division carried out a research program on waste heat recovery with support from DOE (Department of Energy) and the DOE National Energy Technology Laboratory. The objective of the program was to develop a new air management and exhaust energy recovery system that would demonstrate a minimum 10% improvement in thermal efficiency over a base heavy-duty on-highway diesel truck engine. The base engine for this program was a 2007 C15 15.2L series-turbocharged on-highway truck engine with a LPL (low-pressure loop) exhaust recirculation system. The focus of the program was on the development of high efficiency turbomachinery and a high efficiency turbocompound waste heat recovery system. The focus of each area of development was as follows: (1) For turbine stages, the focus was on investigation and development of technologies that would improve on-engine exhaust energy utilization compared to the conventional radial turbines in widespread use today. (2) For compressor stages, the focus was on investigating compressor wheel design parameters beyond the range typically utilized in production, to determine the potential efficiency benefits thereof. (3) For turbocompound, the focus was on the development of a robust bearing system that would provide higher bearing efficiencies compared to systems used in turbocompound power turbines in production. None of the turbocharger technologies investigated involved addition of moving parts, actuators, or exotic materials, thereby increasing the likelihood of a favorable cost-value tradeoff for each technology. And the turbocompound system requires less hardware addition than competing bottoming cycle technologies, making it a more attractive solution from a cost and packaging standpoint. Main outcomes of the program are as follows: (1) Two turbine technologies that demonstrated up to 6% improvement in turbine efficiency on gas stand and 1-3% improvement in thermal efficiency

  8. High Efficiency Engine Technologies Program

    Energy Technology Data Exchange (ETDEWEB)

    Rich Kruiswyk

    2010-07-13

    Caterpillar's Product Development and Global Technology Division carried out a research program on waste heat recovery with support from DOE (Department of Energy) and the DOE National Energy Technology Laboratory. The objective of the program was to develop a new air management and exhaust energy recovery system that would demonstrate a minimum 10% improvement in thermal efficiency over a base heavy-duty on-highway diesel truck engine. The base engine for this program was a 2007 C15 15.2L series-turbocharged on-highway truck engine with a LPL (low-pressure loop) exhaust recirculation system. The focus of the program was on the development of high efficiency turbomachinery and a high efficiency turbocompound waste heat recovery system. The focus of each area of development was as follows: (1) For turbine stages, the focus was on investigation and development of technologies that would improve on-engine exhaust energy utilization compared to the conventional radial turbines in widespread use today. (2) For compressor stages, the focus was on investigating compressor wheel design parameters beyond the range typically utilized in production, to determine the potential efficiency benefits thereof. (3) For turbocompound, the focus was on the development of a robust bearing system that would provide higher bearing efficiencies compared to systems used in turbocompound power turbines in production. None of the turbocharger technologies investigated involved addition of moving parts, actuators, or exotic materials, thereby increasing the likelihood of a favorable cost-value tradeoff for each technology. And the turbocompound system requires less hardware addition than competing bottoming cycle technologies, making it a more attractive solution from a cost and packaging standpoint. Main outcomes of the program are as follows: (1) Two turbine technologies that demonstrated up to 6% improvement in turbine efficiency on gas stand and 1-3% improvement in thermal efficiency

  9. High Performance Computing Today

    Energy Technology Data Exchange (ETDEWEB)

    Dongarra, Jack; Meuer,Hans; Simon,Horst D.; Strohmaier,Erich

    2000-04-01

    In last 50 years, the field of scientific computing has seen a rapid change of vendors, architectures, technologies and the usage of systems. Despite all these changes the evolution of performance on a large scale however seems to be a very steady and continuous process. Moore's Law is often cited in this context. If the authors plot the peak performance of various computers of the last 5 decades in Figure 1 that could have been called the supercomputers of their time they indeed see how well this law holds for almost the complete lifespan of modern computing. On average they see an increase in performance of two magnitudes of order every decade.

  10. Unconventional, High-Efficiency Propulsors

    DEFF Research Database (Denmark)

    Andersen, Poul

    1996-01-01

    The development of ship propellers has generally been characterized by search for propellers with as high efficiency as possible and at the same time low noise and vibration levels and little or no cavitation. This search has lead to unconventional propulsors, like vane-wheel propulsors, contra-r...

  11. High-Productivity Computing in Computational Physics Education

    Science.gov (United States)

    Tel-Zur, Guy

    2011-03-01

    We describe the development of a new course in Computational Physics at the Ben-Gurion University. This elective course for 3rd year undergraduates and MSc. students is being taught during one semester. Computational Physics is by now well accepted as the Third Pillar of Science. This paper's claim is that modern Computational Physics education should deal also with High-Productivity Computing. The traditional approach of teaching Computational Physics emphasizes ``Correctness'' and then ``Accuracy'' and we add also ``Performance.'' Along with topics in Mathematical Methods and case studies in Physics the course deals a significant amount of time with ``Mini-Courses'' in topics such as: High-Throughput Computing - Condor, Parallel Programming - MPI and OpenMP, How to build a Beowulf, Visualization and Grid and Cloud Computing. The course does not intend to teach neither new physics nor new mathematics but it is focused on an integrated approach for solving problems starting from the physics problem, the corresponding mathematical solution, the numerical scheme, writing an efficient computer code and finally analysis and visualization.

  12. A Computationally Efficient Method for Polyphonic Pitch Estimation

    Science.gov (United States)

    Zhou, Ruohua; Reiss, Joshua D.; Mattavelli, Marco; Zoia, Giorgio

    2009-12-01

    This paper presents a computationally efficient method for polyphonic pitch estimation. The method employs the Fast Resonator Time-Frequency Image (RTFI) as the basic time-frequency analysis tool. The approach is composed of two main stages. First, a preliminary pitch estimation is obtained by means of a simple peak-picking procedure in the pitch energy spectrum. Such spectrum is calculated from the original RTFI energy spectrum according to harmonic grouping principles. Then the incorrect estimations are removed according to spectral irregularity and knowledge of the harmonic structures of the music notes played on commonly used music instruments. The new approach is compared with a variety of other frame-based polyphonic pitch estimation methods, and results demonstrate the high performance and computational efficiency of the approach.

  13. High assurance services computing

    CERN Document Server

    2009-01-01

    Covers service-oriented technologies in different domains including high assurance systemsAssists software engineers from industry and government laboratories who develop mission-critical software, and simultaneously provides academia with a practitioner's outlook on the problems of high-assurance software development

  14. High Efficiency Room Air Conditioner

    Energy Technology Data Exchange (ETDEWEB)

    Bansal, Pradeep [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

    2015-01-01

    This project was undertaken as a CRADA project between UT-Battelle and Geberal Electric Company and was funded by Department of Energy to design and develop of a high efficiency room air conditioner. A number of novel elements were investigated to improve the energy efficiency of a state-of-the-art WAC with base capacity of 10,000 BTU/h. One of the major modifications was made by downgrading its capacity from 10,000 BTU/hr to 8,000 BTU/hr by replacing the original compressor with a lower capacity (8,000 BTU/hr) but high efficiency compressor having an EER of 9.7 as compared with 9.3 of the original compressor. However, all heat exchangers from the original unit were retained to provide higher EER. The other subsequent major modifications included- (i) the AC fan motor was replaced by a brushless high efficiency ECM motor along with its fan housing, (ii) the capillary tube was replaced with a needle valve to better control the refrigerant flow and refrigerant set points, and (iii) the unit was tested with a drop-in environmentally friendly binary mixture of R32 (90% molar concentration)/R125 (10% molar concentration). The WAC was tested in the environmental chambers at ORNL as per the design rating conditions of AHAM/ASHRAE (Outdoor- 95F and 40%RH, Indoor- 80F, 51.5%RH). All these modifications resulted in enhancing the EER of the WAC by up to 25%.

  15. Quantum-enhanced Sensing and Efficient Quantum Computation

    Science.gov (United States)

    2015-07-27

    Quantum -enhanced sensing and efficient quantum computation Ian Walmsley THE UNIVERSITY OF...COVERED (From - To) 1 February 2013 - 31 January 2015 4. TITLE AND SUBTITLE Quantum -enhanced sensing and efficient quantum computation 5a. CONTRACT...1895616013 Standard Form 298 (Rev. 8/98) Prescribed by ANSI Std. Z39.18 Final report for “ Quantum ‐Enhanced Sensing and Efficient  Quantum   Computation

  16. Energy based Efficient Resource Scheduling in Green Computing

    Directory of Open Access Journals (Sweden)

    B.Vasumathi,

    2015-11-01

    Full Text Available Cloud Computing is an evolving area of efficient utilization of computing resources. Data centers accommodating Cloud applications ingest massive quantities of energy, contributing to high functioning expenditures and carbon footprints to the atmosphere. Hence, Green Cloud computing resolutions are required not only to save energy for the environment but also to decrease operating charges. In this paper, we emphasis on the development of energy based resource scheduling framework and present an algorithm that consider the synergy between various data center infrastructures (i.e., software, hardware, etc., and performance. In specific, this paper proposes (a architectural principles for energy efficient management of Clouds; (b energy efficient resource allocation strategies and scheduling algorithm considering Quality of Service (QoS outlooks. The performance of the proposed algorithm has been evaluated with the existing energy based scheduling algorithms. The experimental results demonstrate that this approach is effective in minimizing the cost and energy consumption of Cloud applications thus moving towards the achievement of Green Clouds.

  17. Convolutional networks for fast, energy-efficient neuromorphic computing.

    Science.gov (United States)

    Esser, Steven K; Merolla, Paul A; Arthur, John V; Cassidy, Andrew S; Appuswamy, Rathinakumar; Andreopoulos, Alexander; Berg, David J; McKinstry, Jeffrey L; Melano, Timothy; Barch, Davis R; di Nolfo, Carmelo; Datta, Pallab; Amir, Arnon; Taba, Brian; Flickner, Myron D; Modha, Dharmendra S

    2016-10-11

    Deep networks are now able to achieve human-level performance on a broad spectrum of recognition tasks. Independently, neuromorphic computing has now demonstrated unprecedented energy-efficiency through a new chip architecture based on spiking neurons, low precision synapses, and a scalable communication network. Here, we demonstrate that neuromorphic computing, despite its novel architectural primitives, can implement deep convolution networks that (i) approach state-of-the-art classification accuracy across eight standard datasets encompassing vision and speech, (ii) perform inference while preserving the hardware's underlying energy-efficiency and high throughput, running on the aforementioned datasets at between 1,200 and 2,600 frames/s and using between 25 and 275 mW (effectively >6,000 frames/s per Watt), and (iii) can be specified and trained using backpropagation with the same ease-of-use as contemporary deep learning. This approach allows the algorithmic power of deep learning to be merged with the efficiency of neuromorphic processors, bringing the promise of embedded, intelligent, brain-inspired computing one step closer.

  18. Computationally efficient Bayesian inference for inverse problems.

    Energy Technology Data Exchange (ETDEWEB)

    Marzouk, Youssef M.; Najm, Habib N.; Rahn, Larry A.

    2007-10-01

    Bayesian statistics provides a foundation for inference from noisy and incomplete data, a natural mechanism for regularization in the form of prior information, and a quantitative assessment of uncertainty in the inferred results. Inverse problems - representing indirect estimation of model parameters, inputs, or structural components - can be fruitfully cast in this framework. Complex and computationally intensive forward models arising in physical applications, however, can render a Bayesian approach prohibitive. This difficulty is compounded by high-dimensional model spaces, as when the unknown is a spatiotemporal field. We present new algorithmic developments for Bayesian inference in this context, showing strong connections with the forward propagation of uncertainty. In particular, we introduce a stochastic spectral formulation that dramatically accelerates the Bayesian solution of inverse problems via rapid evaluation of a surrogate posterior. We also explore dimensionality reduction for the inference of spatiotemporal fields, using truncated spectral representations of Gaussian process priors. These new approaches are demonstrated on scalar transport problems arising in contaminant source inversion and in the inference of inhomogeneous material or transport properties. We also present a Bayesian framework for parameter estimation in stochastic models, where intrinsic stochasticity may be intermingled with observational noise. Evaluation of a likelihood function may not be analytically tractable in these cases, and thus several alternative Markov chain Monte Carlo (MCMC) schemes, operating on the product space of the observations and the parameters, are introduced.

  19. Role of computational efficiency in process simulation

    Directory of Open Access Journals (Sweden)

    Kurt Strand

    1989-07-01

    Full Text Available It is demonstrated how efficient numerical algorithms may be combined to yield a powerful environment for analysing and simulating dynamic systems. The importance of using efficient numerical algorithms is emphasized and demonstrated through examples from the petrochemical industry.

  20. Study of Efficient Utilization of Power using green Computing

    OpenAIRE

    Ms .Dheera Jadhwani, Mr.Mayur Agrawal, Mr.Hemant Mande

    2012-01-01

    Green computing or green IT, basically concerns toenvironmentally sustainable computing or IT. Thefield of green computing is defined as "theknowledge and practice of designing,manufacturing, using, and disposing of computers,servers, and associated subsystems—which includeprinters, monitors, and networking, storage devicesand communications systems—efficiently andeffectively with minimal or no impact on theenvironment. this computing is similar to greenchemistry that is minimum utilization o...

  1. Multi-petascale highly efficient parallel supercomputer

    Energy Technology Data Exchange (ETDEWEB)

    Asaad, Sameh; Bellofatto, Ralph E.; Blocksome, Michael A.; Blumrich, Matthias A.; Boyle, Peter; Brunheroto, Jose R.; Chen, Dong; Cher, Chen -Yong; Chiu, George L.; Christ, Norman; Coteus, Paul W.; Davis, Kristan D.; Dozsa, Gabor J.; Eichenberger, Alexandre E.; Eisley, Noel A.; Ellavsky, Matthew R.; Evans, Kahn C.; Fleischer, Bruce M.; Fox, Thomas W.; Gara, Alan; Giampapa, Mark E.; Gooding, Thomas M.; Gschwind, Michael K.; Gunnels, John A.; Hall, Shawn A.; Haring, Rudolf A.; Heidelberger, Philip; Inglett, Todd A.; Knudson, Brant L.; Kopcsay, Gerard V.; Kumar, Sameer; Mamidala, Amith R.; Marcella, James A.; Megerian, Mark G.; Miller, Douglas R.; Miller, Samuel J.; Muff, Adam J.; Mundy, Michael B.; O' Brien, John K.; O' Brien, Kathryn M.; Ohmacht, Martin; Parker, Jeffrey J.; Poole, Ruth J.; Ratterman, Joseph D.; Salapura, Valentina; Satterfield, David L.; Senger, Robert M.; Smith, Brian; Steinmacher-Burow, Burkhard; Stockdell, William M.; Stunkel, Craig B.; Sugavanam, Krishnan; Sugawara, Yutaka; Takken, Todd E.; Trager, Barry M.; Van Oosten, James L.; Wait, Charles D.; Walkup, Robert E.; Watson, Alfred T.; Wisniewski, Robert W.; Wu, Peng

    2015-07-14

    A Multi-Petascale Highly Efficient Parallel Supercomputer of 100 petaOPS-scale computing, at decreased cost, power and footprint, and that allows for a maximum packaging density of processing nodes from an interconnect point of view. The Supercomputer exploits technological advances in VLSI that enables a computing model where many processors can be integrated into a single Application Specific Integrated Circuit (ASIC). Each ASIC computing node comprises a system-on-chip ASIC utilizing four or more processors integrated into one die, with each having full access to all system resources and enabling adaptive partitioning of the processors to functions such as compute or messaging I/O on an application by application basis, and preferably, enable adaptive partitioning of functions in accordance with various algorithmic phases within an application, or if I/O or other processors are underutilized, then can participate in computation or communication nodes are interconnected by a five dimensional torus network with DMA that optimally maximize the throughput of packet communications between nodes and minimize latency.

  2. Efficient computation of smoothing splines via adaptive basis sampling

    KAUST Repository

    Ma, Ping

    2015-06-24

    © 2015 Biometrika Trust. Smoothing splines provide flexible nonparametric regression estimators. However, the high computational cost of smoothing splines for large datasets has hindered their wide application. In this article, we develop a new method, named adaptive basis sampling, for efficient computation of smoothing splines in super-large samples. Except for the univariate case where the Reinsch algorithm is applicable, a smoothing spline for a regression problem with sample size n can be expressed as a linear combination of n basis functions and its computational complexity is generally O(n3). We achieve a more scalable computation in the multivariate case by evaluating the smoothing spline using a smaller set of basis functions, obtained by an adaptive sampling scheme that uses values of the response variable. Our asymptotic analysis shows that smoothing splines computed via adaptive basis sampling converge to the true function at the same rate as full basis smoothing splines. Using simulation studies and a large-scale deep earth core-mantle boundary imaging study, we show that the proposed method outperforms a sampling method that does not use the values of response variables.

  3. High-performance scientific computing

    CERN Document Server

    Berry, Michael W; Gallopoulos, Efstratios

    2012-01-01

    This book presents the state of the art in parallel numerical algorithms, applications, architectures, and system software. The book examines various solutions for issues of concurrency, scale, energy efficiency, and programmability, which are discussed in the context of a diverse range of applications. Features: includes contributions from an international selection of world-class authorities; examines parallel algorithm-architecture interaction through issues of computational capacity-based codesign and automatic restructuring of programs using compilation techniques; reviews emerging applic

  4. Efficient Parallel Engineering Computing on Linux Workstations

    Science.gov (United States)

    Lou, John Z.

    2010-01-01

    A C software module has been developed that creates lightweight processes (LWPs) dynamically to achieve parallel computing performance in a variety of engineering simulation and analysis applications to support NASA and DoD project tasks. The required interface between the module and the application it supports is simple, minimal and almost completely transparent to the user applications, and it can achieve nearly ideal computing speed-up on multi-CPU engineering workstations of all operating system platforms. The module can be integrated into an existing application (C, C++, Fortran and others) either as part of a compiled module or as a dynamically linked library (DLL).

  5. Factors Affecting Computer Anxiety in High School Computer Science Students.

    Science.gov (United States)

    Hayek, Linda M.; Stephens, Larry

    1989-01-01

    Examines factors related to computer anxiety measured by the Computer Anxiety Index (CAIN). Achievement in two programing courses was inversely related to computer anxiety. Students who had a home computer and had computer experience before high school had lower computer anxiety than those who had not. Lists 14 references. (YP)

  6. Computer proficiency questionnaire: assessing low and high computer proficient seniors.

    Science.gov (United States)

    Boot, Walter R; Charness, Neil; Czaja, Sara J; Sharit, Joseph; Rogers, Wendy A; Fisk, Arthur D; Mitzner, Tracy; Lee, Chin Chin; Nair, Sankaran

    2015-06-01

    Computers and the Internet have the potential to enrich the lives of seniors and aid in the performance of important tasks required for independent living. A prerequisite for reaping these benefits is having the skills needed to use these systems, which is highly dependent on proper training. One prerequisite for efficient and effective training is being able to gauge current levels of proficiency. We developed a new measure (the Computer Proficiency Questionnaire, or CPQ) to measure computer proficiency in the domains of computer basics, printing, communication, Internet, calendaring software, and multimedia use. Our aim was to develop a measure appropriate for individuals with a wide range of proficiencies from noncomputer users to extremely skilled users. To assess the reliability and validity of the CPQ, a diverse sample of older adults, including 276 older adults with no or minimal computer experience, was recruited and asked to complete the CPQ. The CPQ demonstrated excellent reliability (Cronbach's α = .98), with subscale reliabilities ranging from .86 to .97. Age, computer use, and general technology use all predicted CPQ scores. Factor analysis revealed three main factors of proficiency related to Internet and e-mail use; communication and calendaring; and computer basics. Based on our findings, we also developed a short-form CPQ (CPQ-12) with similar properties but 21 fewer questions. The CPQ and CPQ-12 are useful tools to gauge computer proficiency for training and research purposes, even among low computer proficient older adults. © The Author 2013. Published by Oxford University Press on behalf of The Gerontological Society of America. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  7. A Systems Approach to High Performance Buildings: A Computational Systems Engineering R&D Program to Increase DoD Energy Efficiency

    Science.gov (United States)

    2012-02-01

    Screening 3.1.1 Objectives and Background 3.1.1a) Background: Building Energy Efficiency Retrofit Process The key steps (see Figure 3.1.1) in the...current building energy efficiency retrofit, include 1) Facility Audit to collect building information such as: Building type (climate, usage...building. To further benefit the performance of the building, tools were developed for tractable design optimization which trades off building energy efficiency and

  8. High efficiency shale oil recovery

    Energy Technology Data Exchange (ETDEWEB)

    Adams, D.C.

    1993-04-22

    The overall project objective is to demonstrate the high efficiency of the Adams Counter-Current shale oil recovery process. The efficiency will first be demonstrated on a small scale, in the current phase, after which the demonstration will be extended to the operation of a small pilot plant. Thus the immediate project objective is to obtain data on oil shale retorting operations in a small batch rotary kiln that will be representative of operations in the proposed continuous process pilot plant. Although an oil shale batch sample is sealed in the batch kiln from the start until the end of the run, the process conditions for the batch are the same as the conditions that an element of oil shale would encounter in a continuous process kiln. Similar chemical and physical conditions (heating, mixing, pyrolysis, oxidation) exist in both systems.The two most important data objectives in this phase of the project are to demonstrate (1) that the heat recovery projected for this project is reasonable and (2) that an oil shale kiln will run well and not plug up due to sticking and agglomeration. The following was completed this quarter. (1) Twelve pyrolysis runs were made on five different oil shales. All of the runs exhibited a complete absence of any plugging, tendency. Heat transfer for Green River oil shale in the rotary kiln was 84.6 Btu/hr/ft[sup 2]/[degrees]F, and this will provide for ample heat exchange in the Adams kiln. (2) One retorted residue sample was oxidized at 1000[degrees]F. Preliminary indications are that the ash of this run appears to have been completely oxidized. (3) Further minor equipment repairs and improvements were required during the course of the several runs.

  9. Computationally efficient prediction of area per lipid

    DEFF Research Database (Denmark)

    Chaban, Vitaly V.

    2014-01-01

    Area per lipid (APL) is an important property of biological and artificial membranes. Newly constructed bilayers are characterized by their APL and newly elaborated force fields must reproduce APL. Computer simulations of APL are very expensive due to slow conformational dynamics. The simulated d....... Thus, sampling times to predict accurate APL are reduced by a factor of 10. (C) 2014 Elsevier B.V. All rights reserved....

  10. COMPUTATIONALLY EFFICIENT ESPRIT METHOD FOR DIRECTION FINDING

    Institute of Scientific and Technical Information of China (English)

    2007-01-01

    In this paper, a low complexity ESPRIT algorithm based on power method and Orthogonal-triangular (QR) decomposition is presented for direction finding, which does not require a priori knowledge of source number and the predetermined threshold (separates the signal and noise eigen-values). Firstly, according to the estimation of noise subspace obtained by the power method, a novel source number detection method without eigen-decomposition is proposed based on QR decomposition. Furthermore, the eigenvectors of signal subspace can be determined according to Q matrix,and then the directions of signals could be computed by the ESPRIT algorithm. To determine the source number and subspace, the computation complexity of the proposed algorithm is approximated as (2log2 n+ 2.67)M3, where n is the power of covariance matrix and M is the number of array elements. Compared with the Single Vector Decomposition (SVD) based algorithm, it has a substantial computational saving with the approximation performance. The simulation results demonstrate its effectiveness and robustness.

  11. The Efficient Use of Vector Computers with Emphasis on Computational Fluid Dynamics : a GAMM-Workshop

    CERN Document Server

    Gentzsch, Wolfgang

    1986-01-01

    The GAMM Committee for Numerical Methods in Fluid Mechanics organizes workshops which should bring together experts of a narrow field of computational fluid dynamics (CFD) to exchange ideas and experiences in order to speed-up the development in this field. In this sense it was suggested that a workshop should treat the solution of CFD problems on vector computers. Thus we organized a workshop with the title "The efficient use of vector computers with emphasis on computational fluid dynamics". The workshop took place at the Computing Centre of the University of Karlsruhe, March 13-15,1985. The participation had been restricted to 22 people of 7 countries. 18 papers have been presented. In the announcement of the workshop we wrote: "Fluid mechanics has actively stimulated the development of superfast vector computers like the CRAY's or CYBER 205. Now these computers on their turn stimulate the development of new algorithms which result in a high degree of vectorization (sca1ar/vectorized execution-time). But w...

  12. Experiences With Efficient Methodologies for Teaching Computer Programming to Geoscientists

    Science.gov (United States)

    Jacobs, Christian T.; Gorman, Gerard J.; Rees, Huw E.; Craig, Lorraine E.

    2016-08-01

    Computer programming was once thought of as a skill required only by professional software developers. But today, given the ubiquitous nature of computation and data science it is quickly becoming necessary for all scientists and engineers to have at least a basic knowledge of how to program. Teaching how to program, particularly to those students with little or no computing background, is well-known to be a difficult task. However, there is also a wealth of evidence-based teaching practices for teaching programming skills which can be applied to greatly improve learning outcomes and the student experience. Adopting these practices naturally gives rise to greater learning efficiency - this is critical if programming is to be integrated into an already busy geoscience curriculum. This paper considers an undergraduate computer programming course, run during the last 5 years in the Department of Earth Science and Engineering at Imperial College London. The teaching methodologies that were used each year are discussed alongside the challenges that were encountered, and how the methodologies affected student performance. Anonymised student marks and feedback are used to highlight this, and also how the adjustments made to the course eventually resulted in a highly effective learning environment.

  13. Improving the efficiency of abdominal aortic aneurysm wall stress computations.

    Science.gov (United States)

    Zelaya, Jaime E; Goenezen, Sevan; Dargon, Phong T; Azarbal, Amir-Farzin; Rugonyi, Sandra

    2014-01-01

    An abdominal aortic aneurysm is a pathological dilation of the abdominal aorta, which carries a high mortality rate if ruptured. The most commonly used surrogate marker of rupture risk is the maximal transverse diameter of the aneurysm. More recent studies suggest that wall stress from models of patient-specific aneurysm geometries extracted, for instance, from computed tomography images may be a more accurate predictor of rupture risk and an important factor in AAA size progression. However, quantification of wall stress is typically computationally intensive and time-consuming, mainly due to the nonlinear mechanical behavior of the abdominal aortic aneurysm walls. These difficulties have limited the potential of computational models in clinical practice. To facilitate computation of wall stresses, we propose to use a linear approach that ensures equilibrium of wall stresses in the aneurysms. This proposed linear model approach is easy to implement and eliminates the burden of nonlinear computations. To assess the accuracy of our proposed approach to compute wall stresses, results from idealized and patient-specific model simulations were compared to those obtained using conventional approaches and to those of a hypothetical, reference abdominal aortic aneurysm model. For the reference model, wall mechanical properties and the initial unloaded and unstressed configuration were assumed to be known, and the resulting wall stresses were used as reference for comparison. Our proposed linear approach accurately approximates wall stresses for varying model geometries and wall material properties. Our findings suggest that the proposed linear approach could be used as an effective, efficient, easy-to-use clinical tool to estimate patient-specific wall stresses.

  14. Efficient Associative Computation with Discrete Synapses.

    Science.gov (United States)

    Knoblauch, Andreas

    2016-01-01

    Neural associative networks are a promising computational paradigm for both modeling neural circuits of the brain and implementing associative memory and Hebbian cell assemblies in parallel VLSI or nanoscale hardware. Previous work has extensively investigated synaptic learning in linear models of the Hopfield type and simple nonlinear models of the Steinbuch/Willshaw type. Optimized Hopfield networks of size n can store a large number of about n(2)/k memories of size k (or associations between them) but require real-valued synapses, which are expensive to implement and can store at most C = 0.72 bits per synapse. Willshaw networks can store a much smaller number of about n(2)/k(2) memories but get along with much cheaper binary synapses. Here I present a learning model employing synapses with discrete synaptic weights. For optimal discretization parameters, this model can store, up to a factor ζ close to one, the same number of memories as for optimized Hopfield-type learning--for example, ζ = 0.64 for binary synapses, ζ = 0.88 for 2 bit (four-state) synapses, ζ = 0.96 for 3 bit (8-state) synapses, and ζ > 0.99 for 4 bit (16-state) synapses. The model also provides the theoretical framework to determine optimal discretization parameters for computer implementations or brainlike parallel hardware including structural plasticity. In particular, as recently shown for the Willshaw network, it is possible to store C(I) = 1 bit per computer bit and up to C(S) = log n bits per nonsilent synapse, whereas the absolute number of stored memories can be much larger than for the Willshaw model.

  15. Efficient Tate pairing computation using double-base chains

    Institute of Scientific and Technical Information of China (English)

    ZHAO ChangAn; ZHANG FangGuo; HUANG JiWu

    2008-01-01

    Pairing-based cryptosystems have developed very fast in the last few years. The effi-ciencies of these cryptosystems depend on the computation of the bilinear pairings. In this paper, a new efficient algorithm based on double-base chains for computing the Tate pairing is proposed for odd characteristic p > 3. The inherent sparseness of double-base number system reduces the computational cost for computing the Tate pairing evidently. The new algorithm is 9% faster than the previous fastest method for the embedding degree k = 6.

  16. Skyline View: Efficient Distributed Subspace Skyline Computation

    Science.gov (United States)

    Kim, Jinhan; Lee, Jongwuk; Hwang, Seung-Won

    Skyline queries have gained much attention as alternative query semantics with pros (e.g.low query formulation overhead) and cons (e.g.large control over result size). To overcome the cons, subspace skyline queries have been recently studied, where users iteratively specify relevant feature subspaces on search space. However, existing works mainly focuss on centralized databases. This paper aims to extend subspace skyline computation to distributed environments such as the Web, where the most important issue is to minimize the cost of accessing vertically distributed objects. Toward this goal, we exploit prior skylines that have overlapped subspaces to the given subspace. In particular, we develop algorithms for three scenarios- when the subspace of prior skylines is superspace, subspace, or the rest. Our experimental results validate that our proposed algorithm shows significantly better performance than the state-of-the-art algorithms.

  17. Towards highly efficient water photoelectrolysis

    Science.gov (United States)

    Elavambedu Prakasam, Haripriya

    ethylene glycol resulted in remarkable growth characteristics of titania nanotube arrays, hexagonal closed packed up to 1 mm in length, with tube aspect ratios of approximately 10,000. For the first time, complete anodization of the starting titanium foil has been demonstrated resulting in back to back nanotube array membranes ranging from 360 mum--1 mm in length. The nanotubes exhibited growth rates of up to 15 mum/hr. A detailed study on the factors affecting the growth rate and nanotube dimensions is presented. It is suggested that faster high field ionic conduction through a thinner barrier layer is responsible for the higher growth rates observed in electrolytes containing ethylene glycol. Methods to fabricate free standing, titania nanotube array membranes ranging in thickness from 50 microm--1000 mum has also been an outcome of this dissertation. In an effort to combine the charge transport properties of titania with the light absorption properties of iron (III) oxide, films comprised of vertically oriented Ti-Fe-O nanotube arrays on FTO coated glass substrates have been successfully synthesized in ethylene glycol electrolytes. Depending upon the Fe content the bandgap of the resulting films varied from about 3.26 to 2.17 eV. The Ti-Fe oxide nanotube array films demonstrated a photocurrent of 2 mA/cm2 under global AM 1.5 illumination with a 1.2% (two-electrode) photoconversion efficiency, demonstrating a sustained, time-energy normalized hydrogen evolution rate by water splitting of 7.1 mL/W·hr in a 1 M KOH solution with a platinum counter electrode under an applied bias of 0.7 V. The Ti-Fe-O material architecture demonstrates properties useful for hydrogen generation by water photoelectrolysis and, more importantly, this dissertation demonstrates that the general nanotube-array synthesis technique can be extended to other ternary oxide compositions of interest for water photoelectrolysis.

  18. Efficiency and reliability assessments of retrofitted high-efficiency motors

    Energy Technology Data Exchange (ETDEWEB)

    Hsu, John S.; Otaduy, P.J.; Dueck, J.D.

    1994-12-31

    The majority of electric-motor applications are pumps, fans, blowers, and certain compressors that follow the load torque pattern described in this paper. It has been known for many years that simply replacing the old motor with a high-efficiency motor might not produce the expected efficiency gain. This paper suggests the calculations for the effective efficiency and temperature rise of the high-efficiency motor. The reliability in terms of temperature rise, downsizing, power factor, harmonics, mechanical structure, etc., are discussed.

  19. A computable expression of closure to efficient causation.

    Science.gov (United States)

    Mossio, Matteo; Longo, Giuseppe; Stewart, John

    2009-04-07

    In this paper, we propose a mathematical expression of closure to efficient causation in terms of lambda-calculus; we argue that this opens up the perspective of developing principled computer simulations of systems closed to efficient causation in an appropriate programming language. An important implication of our formulation is that, by exhibiting an expression in lambda-calculus, which is a paradigmatic formalism for computability and programming, we show that there are no conceptual or principled problems in realizing a computer simulation or model of closure to efficient causation. We conclude with a brief discussion of the question whether closure to efficient causation captures all relevant properties of living systems. We suggest that it might not be the case, and that more complex definitions could indeed create crucial some obstacles to computability.

  20. Efficient Computation of Optimal Trading Strategies

    CERN Document Server

    Boyarshinov, Victor

    2010-01-01

    Given the return series for a set of instruments, a \\emph{trading strategy} is a switching function that transfers wealth from one instrument to another at specified times. We present efficient algorithms for constructing (ex-post) trading strategies that are optimal with respect to the total return, the Sterling ratio and the Sharpe ratio. Such ex-post optimal strategies are useful analysis tools. They can be used to analyze the "profitability of a market" in terms of optimal trading; to develop benchmarks against which real trading can be compared; and, within an inductive framework, the optimal trades can be used to to teach learning systems (predictors) which are then used to identify future trading opportunities.

  1. HIGH-EFFICIENCY INFRARED RECEIVER

    Directory of Open Access Journals (Sweden)

    A. K. Esman

    2016-01-01

    Full Text Available Recent research and development show promising use of high-performance solid-state receivers of the electromagnetic radiation. These receivers are based on the low-barrier Schottky diodes. The approach to the design of the receivers on the basis of delta-doped low-barrier Schottky diodes with beam leads without bias is especially actively developing because for uncooled receivers of the microwave radiation these diodes have virtually no competition. The purpose of this work is to improve the main parameters and characteristics that determine the practical relevance of the receivers of mid-infrared electromagnetic radiation at the operating room temperature by modifying the electrodes configuration of the diode and optimizing the distance between them. Proposed original design solution of the integrated receiver of mid-infrared radiation on the basis of the low-barrier Schottky diodes with beam leads allows to effectively adjust its main parameters and characteristics. Simulation of the electromagnetic characteristics of the proposed receiver by using the software package HFSS with the basic algorithm of a finite element method which implemented to calculate the behavior of electromagnetic fields on an arbitrary geometry with a predetermined material properties have shown that when the inner parts of the electrodes of the low-barrier Schottky diode is performed in the concentric elliptical convex-concave shape, it can be reduce the reflection losses to -57.75 dB and the standing wave ratio to 1.003 while increasing the directivity up to 23 at a wavelength of 6.09 μm. At this time, the rounded radii of the inner parts of the anode and cathode electrodes are equal 212 nm and 318 nm respectively and the gap setting between them is 106 nm. These parameters will improve the efficiency of the developed infrared optical-promising and electronic equipment for various purposes intended for work in the mid-infrared wavelength range. 

  2. Study of Efficient Utilization of Power using green Computing

    Directory of Open Access Journals (Sweden)

    Ms .Dheera Jadhwani, Mr.Mayur Agrawal, Mr.Hemant Mande

    2012-12-01

    Full Text Available Green computing or green IT, basically concerns toenvironmentally sustainable computing or IT. Thefield of green computing is defined as "theknowledge and practice of designing,manufacturing, using, and disposing of computers,servers, and associated subsystems—which includeprinters, monitors, and networking, storage devicesand communications systems—efficiently andeffectively with minimal or no impact on theenvironment. this computing is similar to greenchemistry that is minimum utilization of hazardousmaterials and , maximizing energy efficiencyduring the product's lifetime, and also promote therecyclability or biodegradability of defunct productsand factory waste .

  3. High Efficiency Refrigeration Process Project

    Data.gov (United States)

    National Aeronautics and Space Administration — It has been proposed by NASA JSC studies, that the most mass efficient (non-nuclear) method of Lunar habitat cooling is via photovoltaic (PV) direct vapor...

  4. High efficiency turbine blade coatings.

    Energy Technology Data Exchange (ETDEWEB)

    Youchison, Dennis L.; Gallis, Michail A.

    2014-06-01

    The development of advanced thermal barrier coatings (TBCs) of yttria stabilized zirconia (YSZ) that exhibit lower thermal conductivity through better control of electron beam - physical vapor deposition (EB-PVD) processing is of prime interest to both the aerospace and power industries. This report summarizes the work performed under a two-year Lab-Directed Research and Development (LDRD) project (38664) to produce lower thermal conductivity, graded-layer thermal barrier coatings for turbine blades in an effort to increase the efficiency of high temperature gas turbines. This project was sponsored by the Nuclear Fuel Cycle Investment Area. Therefore, particular importance was given to the processing of the large blades required for industrial gas turbines proposed for use in the Brayton cycle of nuclear plants powered by high temperature gas-cooled reactors (HTGRs). During this modest (~1 full-time equivalent (FTE)) project, the processing technology was developed to create graded TBCs by coupling ion beam-assisted deposition (IBAD) with substrate pivoting in the alumina-YSZ system. The Electron Beam - 1200 kW (EB-1200) PVD system was used to deposit a variety of TBC coatings with micron layered microstructures and reduced thermal conductivity below 1.5 W/m.K. The use of IBAD produced fully stoichiometric coatings at a reduced substrate temperature of 600 oC and a reduced oxygen background pressure of 0.1 Pa. IBAD was also used to successfully demonstrate the transitioning of amorphous PVD-deposited alumina to the -phase alumina required as an oxygen diffusion barrier and for good adhesion to the substrate Ni2Al3 bondcoat. This process replaces the time consuming thermally grown oxide formation required before the YSZ deposition. In addition to the process technology, Direct Simulation Monte Carlo plume modeling and spectroscopic characterization of the PVD plumes were performed. The project consisted of five tasks. These included the production of layered

  5. High efficiency turbine blade coatings

    Energy Technology Data Exchange (ETDEWEB)

    Youchison, Dennis L. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Gallis, Michail A. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2014-06-01

    The development of advanced thermal barrier coatings (TBCs) of yttria stabilized zirconia (YSZ) that exhibit lower thermal conductivity through better control of electron beam - physical vapor deposition (EB-PVD) processing is of prime interest to both the aerospace and power industries. This report summarizes the work performed under a two-year Lab-Directed Research and Development (LDRD) project (38664) to produce lower thermal conductivity, graded-layer thermal barrier coatings for turbine blades in an effort to increase the efficiency of high temperature gas turbines. This project was sponsored by the Nuclear Fuel Cycle Investment Area. Therefore, particular importance was given to the processing of the large blades required for industrial gas turbines proposed for use in the Brayton cycle of nuclear plants powered by high temperature gas-cooled reactors (HTGRs). During this modest (~1 full-time equivalent (FTE)) project, the processing technology was developed to create graded TBCs by coupling ion beam-assisted deposition (IBAD) with substrate pivoting in the alumina-YSZ system. The Electron Beam - 1200 kW (EB-1200) PVD system was used to deposit a variety of TBC coatings with micron layered microstructures and reduced thermal conductivity below 1.5 W/m.K. The use of IBAD produced fully stoichiometric coatings at a reduced substrate temperature of 600°C and a reduced oxygen background pressure of 0.1 Pa. IBAD was also used to successfully demonstrate the transitioning of amorphous PVD-deposited alumina to the -phase alumina required as an oxygen diffusion barrier and for good adhesion to the substrate Ni2Al3 bondcoat. This process replaces the time consuming thermally grown oxide formation required before the YSZ deposition. In addition to the process technology, Direct Simulation Monte Carlo plume modeling and spectroscopic characterization of the PVD plumes were performed. The project consisted of five tasks. These included the

  6. Earthquake detection through computationally efficient similarity search

    Science.gov (United States)

    Yoon, Clara E.; O’Reilly, Ossian; Bergen, Karianne J.; Beroza, Gregory C.

    2015-01-01

    Seismology is experiencing rapid growth in the quantity of data, which has outpaced the development of processing algorithms. Earthquake detection—identification of seismic events in continuous data—is a fundamental operation for observational seismology. We developed an efficient method to detect earthquakes using waveform similarity that overcomes the disadvantages of existing detection methods. Our method, called Fingerprint And Similarity Thresholding (FAST), can analyze a week of continuous seismic waveform data in less than 2 hours, or 140 times faster than autocorrelation. FAST adapts a data mining algorithm, originally designed to identify similar audio clips within large databases; it first creates compact “fingerprints” of waveforms by extracting key discriminative features, then groups similar fingerprints together within a database to facilitate fast, scalable search for similar fingerprint pairs, and finally generates a list of earthquake detections. FAST detected most (21 of 24) cataloged earthquakes and 68 uncataloged earthquakes in 1 week of continuous data from a station located near the Calaveras Fault in central California, achieving detection performance comparable to that of autocorrelation, with some additional false detections. FAST is expected to realize its full potential when applied to extremely long duration data sets over a distributed network of seismic stations. The widespread application of FAST has the potential to aid in the discovery of unexpected seismic signals, improve seismic monitoring, and promote a greater understanding of a variety of earthquake processes. PMID:26665176

  7. Secure Computation, I/O-Efficient Algorithms and Distributed Signatures

    DEFF Research Database (Denmark)

    Damgård, Ivan Bjerre; Kölker, Jonas; Toft, Tomas

    2012-01-01

    adversary corrupting a constant fraction of the players and servers. Using packed secret sharing, the data can be stored in a compact way but will only be accessible in a block-wise fashion. We explore the possibility of using I/O-efficient algorithms to nevertheless compute on the data as efficiently...

  8. High efficiency stationary hydrogen storage

    Energy Technology Data Exchange (ETDEWEB)

    Hynek, S.; Fuller, W.; Truslow, S. [Arthur D. Little, Inc., Cambridge, MA (United States)

    1995-09-01

    Stationary storage of hydrogen permits one to make hydrogen now and use it later. With stationary hydrogen storage, one can use excess electrical generation capacity to power an electrolyzer, and store the resultant hydrogen for later use or transshipment. One can also use stationary hydrogen as a buffer at fueling stations to accommodate non-steady fueling demand, thus permitting the hydrogen supply system (e.g., methane reformer or electrolyzer) to be sized to meet the average, rather than the peak, demand. We at ADL designed, built, and tested a stationary hydrogen storage device that thermally couples a high-temperature metal hydride to a phase change material (PCM). The PCM captures and stores the heat of the hydriding reaction as its own heat of fusion (that is, it melts), and subsequently returns that heat of fusion (by freezing) to facilitate the dehydriding reaction. A key component of this stationary hydrogen storage device is the metal hydride itself. We used nickel-coated magnesium powder (NCMP) - magnesium particles coated with a thin layer of nickel by means of chemical vapor deposition (CVD). Magnesium hydride can store a higher weight fraction of hydrogen than any other practical metal hydride, and it is less expensive than any other metal hydride. We designed and constructed an experimental NCM/PCM reactor out of 310 stainless steel in the form of a shell-and-tube heat exchanger, with the tube side packed with NCMP and the shell side filled with a eutectic mixture of NaCL, KCl, and MgCl{sub 2}. Our experimental results indicate that with proper attention to limiting thermal losses, our overall efficiency will exceed 90% (DOE goal: >75%) and our overall system cost will be only 33% (DOE goal: <50%) of the value of the delivered hydrogen. It appears that NCMP can be used to purify hydrogen streams and store hydrogen at the same time. These prospects make the NCMP/PCM reactor an attractive component in a reformer-based hydrogen fueling station.

  9. Energy efficiency of computer power supply units - Final report

    Energy Technology Data Exchange (ETDEWEB)

    Aebischer, B. [cepe - Centre for Energy Policy and Economics, Swiss Federal Institute of Technology Zuerich, Zuerich (Switzerland); Huser, H. [Encontrol GmbH, Niederrohrdorf (Switzerland)

    2002-11-15

    This final report for the Swiss Federal Office of Energy (SFOE) takes a look at the efficiency of computer power supply units, which decreases rapidly during average computer use. The background and the purpose of the project are examined. The power supplies for personal computers are discussed and the testing arrangement used is described. Efficiency, power-factor and operating points of the units are examined. Potentials for improvement and measures to be taken are discussed. Also, action to be taken by those involved in the design and operation of such power units is proposed. Finally, recommendations for further work are made.

  10. The New and Computationally Efficient MIL-SOM Algorithm: Potential Benefits for Visualization and Analysis of a Large-Scale High-Dimensional Clinically Acquired Geographic Data

    Directory of Open Access Journals (Sweden)

    Tonny J. Oyana

    2012-01-01

    Full Text Available The objective of this paper is to introduce an efficient algorithm, namely, the mathematically improved learning-self organizing map (MIL-SOM algorithm, which speeds up the self-organizing map (SOM training process. In the proposed MIL-SOM algorithm, the weights of Kohonen’s SOM are based on the proportional-integral-derivative (PID controller. Thus, in a typical SOM learning setting, this improvement translates to faster convergence. The basic idea is primarily motivated by the urgent need to develop algorithms with the competence to converge faster and more efficiently than conventional techniques. The MIL-SOM algorithm is tested on four training geographic datasets representing biomedical and disease informatics application domains. Experimental results show that the MIL-SOM algorithm provides a competitive, better updating procedure and performance, good robustness, and it runs faster than Kohonen’s SOM.

  11. Overview of Ecological Agriculture with High Efficiency

    Institute of Scientific and Technical Information of China (English)

    HUANG Guo-qin; ZHAO Qi-guo; GONG Shao-lin; SHI Qing-hua

    2012-01-01

    From the presentation, connotation, characteristics, principles, pattern, and technologies of ecological agriculture with high efficiency, we conduct comprehensive and systematic analysis and discussion of the theoretical and practical progress of ecological agriculture with high efficiency. (i) Ecological agriculture with high efficiency was first advanced in China in 1991. (ii) Ecological agriculture with high efficiency highlights "high efficiency", "ecology", and "combination". (iii) Ecological agriculture with high efficiency is characterized by diverse organisms, good environment, good structure, powerful function, good quality, high benefit, low emission, sustainability. (iv) The yield increase and efficiency increase principle of ecological agriculture with high efficiency lies in full land use, three-dimensional light use, sufficient use of season, multi-layer water consumption, efficient fertilizer consumption, symbiosis and mutual supplement, ecological disaster reduction, recycling. (v) The typical pattern of ecological agriculture with high efficiency includes three-dimensional use pattern, biological symbiosis pattern, multi-industry combination pattern, industrial extension pattern, technology-driven pattern, environmental renovation pattern, resource recycling pattern, leisure and sight-seeing pattern. (vi) The key technologies of ecological agriculture with high efficiency include resource-saving technology, water and fertilizer regulation technology, biological technology for increasing soil fertility, disaster prevention and mitigation technology, comprehensive utilization technology, water conservation technology, structural adjustment technology, energy development technology, watershed control technology, and modern high-tech technology.

  12. High Efficiency Ka-Band Spatial Combiner

    Directory of Open Access Journals (Sweden)

    D. Passi

    2014-12-01

    Full Text Available A Ka-Band, High Efficiency, Small Size Spatial Combiner (SPC is proposed in this paper, which uses an innovatively matched quadruple Fin Lines to microstrip (FLuS transitions. At the date of this paper and at the Author's best knowledge no such FLuS innovative transitions have been reported in literature before. These transitions are inserted into a WR28 waveguide T-junction, in order to allow the integration of 16 Monolithic Microwave Integrated Circuit (MMIC Solid State Power Amplifiers (SSPA's. A computational electromagnetic model using the finite elements method has been implemented. A mean insertion loss of 2 dB is achieved with a return loss better the 10 dB in the 31-37 GHz bandwidth.

  13. High Efficiency, High Performance Clothes Dryer

    Energy Technology Data Exchange (ETDEWEB)

    Peter Pescatore; Phil Carbone

    2005-03-31

    This program covered the development of two separate products; an electric heat pump clothes dryer and a modulating gas dryer. These development efforts were independent of one another and are presented in this report in two separate volumes. Volume 1 details the Heat Pump Dryer Development while Volume 2 details the Modulating Gas Dryer Development. In both product development efforts, the intent was to develop high efficiency, high performance designs that would be attractive to US consumers. Working with Whirlpool Corporation as our commercial partner, TIAX applied this approach of satisfying consumer needs throughout the Product Development Process for both dryer designs. Heat pump clothes dryers have been in existence for years, especially in Europe, but have not been able to penetrate the market. This has been especially true in the US market where no volume production heat pump dryers are available. The issue has typically been around two key areas: cost and performance. Cost is a given in that a heat pump clothes dryer has numerous additional components associated with it. While heat pump dryers have been able to achieve significant energy savings compared to standard electric resistance dryers (over 50% in some cases), designs to date have been hampered by excessively long dry times, a major market driver in the US. The development work done on the heat pump dryer over the course of this program led to a demonstration dryer that delivered the following performance characteristics: (1) 40-50% energy savings on large loads with 35 F lower fabric temperatures and similar dry times; (2) 10-30 F reduction in fabric temperature for delicate loads with up to 50% energy savings and 30-40% time savings; (3) Improved fabric temperature uniformity; and (4) Robust performance across a range of vent restrictions. For the gas dryer development, the concept developed was one of modulating the gas flow to the dryer throughout the dry cycle. Through heat modulation in a

  14. High Efficiency, High Performance Clothes Dryer

    Energy Technology Data Exchange (ETDEWEB)

    Peter Pescatore; Phil Carbone

    2005-03-31

    This program covered the development of two separate products; an electric heat pump clothes dryer and a modulating gas dryer. These development efforts were independent of one another and are presented in this report in two separate volumes. Volume 1 details the Heat Pump Dryer Development while Volume 2 details the Modulating Gas Dryer Development. In both product development efforts, the intent was to develop high efficiency, high performance designs that would be attractive to US consumers. Working with Whirlpool Corporation as our commercial partner, TIAX applied this approach of satisfying consumer needs throughout the Product Development Process for both dryer designs. Heat pump clothes dryers have been in existence for years, especially in Europe, but have not been able to penetrate the market. This has been especially true in the US market where no volume production heat pump dryers are available. The issue has typically been around two key areas: cost and performance. Cost is a given in that a heat pump clothes dryer has numerous additional components associated with it. While heat pump dryers have been able to achieve significant energy savings compared to standard electric resistance dryers (over 50% in some cases), designs to date have been hampered by excessively long dry times, a major market driver in the US. The development work done on the heat pump dryer over the course of this program led to a demonstration dryer that delivered the following performance characteristics: (1) 40-50% energy savings on large loads with 35 F lower fabric temperatures and similar dry times; (2) 10-30 F reduction in fabric temperature for delicate loads with up to 50% energy savings and 30-40% time savings; (3) Improved fabric temperature uniformity; and (4) Robust performance across a range of vent restrictions. For the gas dryer development, the concept developed was one of modulating the gas flow to the dryer throughout the dry cycle. Through heat modulation in a

  15. An Efficient Speedup Strategy for Constant Sum Game Computations

    Directory of Open Access Journals (Sweden)

    Alexandru-Ioan STAN

    2014-01-01

    Full Text Available Large classes of game theoretic problems seem to defy attempts of finding polynomial-time al-gorithms while analyzing large amounts of data. This premise leads naturally to the possibility of using efficient parallel computing implementations when seeking exact solutions to some of these problems. Although alpha beta algorithms for more than one-player game-tree searches show moderate parallel performance, this paper sets forth an alpha beta strategy enhanced with transposition tables in order to offer satisfactory speedups on high performance servers. When the access to the transposition tables is done in low constant delay time, the achieved speedups should approach the theoretical upper bounds of the code parallelism. We tested the strategy on a well-known combinatorial game.

  16. Teuchos C++ memory management classes, idioms, and related topics, the complete reference : a comprehensive strategy for safe and efficient memory management in C++ for high performance computing.

    Energy Technology Data Exchange (ETDEWEB)

    Bartlett, Roscoe Ainsworth

    2010-05-01

    The ubiquitous use of raw pointers in higher-level code is the primary cause of all memory usage problems and memory leaks in C++ programs. This paper describes what might be considered a radical approach to the problem which is to encapsulate the use of all raw pointers and all raw calls to new and delete in higher-level C++ code. Instead, a set of cooperating template classes developed in the Trilinos package Teuchos are used to encapsulate every use of raw C++ pointers in every use case where it appears in high-level code. Included in the set of memory management classes is the typical reference-counted smart pointer class similar to boost::shared ptr (and therefore C++0x std::shared ptr). However, what is missing in boost and the new standard library are non-reference counted classes for remaining use cases where raw C++ pointers would need to be used. These classes have a debug build mode where nearly all programmer errors are caught and gracefully reported at runtime. The default optimized build mode strips all runtime checks and allows the code to perform as efficiently as raw C++ pointers with reasonable usage. Also included is a novel approach for dealing with the circular references problem that imparts little extra overhead and is almost completely invisible to most of the code (unlike the boost and therefore C++0x approach). Rather than being a radical approach, encapsulating all raw C++ pointers is simply the logical progression of a trend in the C++ development and standards community that started with std::auto ptr and is continued (but not finished) with std::shared ptr in C++0x. Using the Teuchos reference-counted memory management classes allows one to remove unnecessary constraints in the use of objects by removing arbitrary lifetime ordering constraints which are a type of unnecessary coupling [23]. The code one writes with these classes will be more likely to be correct on first writing, will be less likely to contain silent (but deadly) memory

  17. Teuchos C++ memory management classes, idioms, and related topics, the complete reference : a comprehensive strategy for safe and efficient memory management in C++ for high performance computing.

    Energy Technology Data Exchange (ETDEWEB)

    Bartlett, Roscoe Ainsworth

    2010-05-01

    The ubiquitous use of raw pointers in higher-level code is the primary cause of all memory usage problems and memory leaks in C++ programs. This paper describes what might be considered a radical approach to the problem which is to encapsulate the use of all raw pointers and all raw calls to new and delete in higher-level C++ code. Instead, a set of cooperating template classes developed in the Trilinos package Teuchos are used to encapsulate every use of raw C++ pointers in every use case where it appears in high-level code. Included in the set of memory management classes is the typical reference-counted smart pointer class similar to boost::shared ptr (and therefore C++0x std::shared ptr). However, what is missing in boost and the new standard library are non-reference counted classes for remaining use cases where raw C++ pointers would need to be used. These classes have a debug build mode where nearly all programmer errors are caught and gracefully reported at runtime. The default optimized build mode strips all runtime checks and allows the code to perform as efficiently as raw C++ pointers with reasonable usage. Also included is a novel approach for dealing with the circular references problem that imparts little extra overhead and is almost completely invisible to most of the code (unlike the boost and therefore C++0x approach). Rather than being a radical approach, encapsulating all raw C++ pointers is simply the logical progression of a trend in the C++ development and standards community that started with std::auto ptr and is continued (but not finished) with std::shared ptr in C++0x. Using the Teuchos reference-counted memory management classes allows one to remove unnecessary constraints in the use of objects by removing arbitrary lifetime ordering constraints which are a type of unnecessary coupling [23]. The code one writes with these classes will be more likely to be correct on first writing, will be less likely to contain silent (but deadly) memory

  18. Highly Efficient Freestyle Magnetic Nanoswimmer.

    Science.gov (United States)

    Li, Tianlong; Li, Jinxing; Morozov, Konstantin I; Wu, Zhiguang; Xu, Tailin; Rozen, Isaac; Leshansky, Alexander M; Li, Longqiu; Wang, Joseph

    2017-08-09

    The unique swimming strategies of natural microorganisms have inspired recent development of magnetic micro/nanorobots powered by artificial helical or flexible flagella. However, as artificial nanoswimmers with unique geometries are being developed, it is critical to explore new potential modes for kinetic optimization. For example, the freestyle stroke is the most efficient of the competitive swimming strokes for humans. Here we report a new type of magnetic nanorobot, a symmetric multilinked two-arm nanoswimmer, capable of efficient "freestyle" swimming at low Reynolds numbers. Excellent agreement between the experimental observations and theoretical predictions indicates that the powerful "freestyle" propulsion of the two-arm nanorobot is attributed to synchronized oscillatory deformations of the nanorobot under the combined action of magnetic field and viscous forces. It is demonstrated for the first time that the nonplanar propulsion gait due to the cooperative "freestyle" stroke of the two magnetic arms can be powered by a plane oscillatory magnetic field. These two-arm nanorobots are capable of a powerful propulsion up to 12 body lengths per second, along with on-demand speed regulation and remote navigation. Furthermore, the nonplanar propulsion gait powered by the consecutive swinging of the achiral magnetic arms is more efficient than that of common chiral nanohelical swimmers. This new swimming mechanism and its attractive performance opens new possibilities in designing remotely actuated nanorobots for biomedical operation at the nanoscale.

  19. FPGA Implementation of High Efficiency Parallel Computing of PDSCH in LTE Systems%LTE基站中PDSCH高效并行计算的FPGA实现

    Institute of Scientific and Technical Information of China (English)

    廖晓强; 钱俊伟; 朱宇霞

    2012-01-01

    A parallel computing method of bit-level signal processing of the PDSCH based on FPGA in 3GPP LTE eNodeB is presented. Because of less data traffic in PDCCH compared to PDSCH, the parallel computing is used in order to make sure the data timing of PDCCH and PDSCH are aligned and hardware resources are saved greatly to meet the requirement of LTE system test. The method is successfully proved and optimized on Virtex -6 FPGA of Xilinx by VHDL.%提出一种第三代合作伙伴(3GPP)长期演进(LTE)基站中下行共享信道(PDSCH)中比特级信号处理并行计算方案,其并行运算是基于现场可编程门阵列(FPGA)的.由于下行控制信道中数据流量相对下行共享信道偏少,为了保证控制信道与共享信道下行数据的时序对齐,并且最大程度上节省硬件资源,以满足LTE系统测试要求,必须采用并行计算的处理方式.采用VHDL语言在Xilinx公司的Virtex -6系列FPGA芯片内成功对该方案进行了验证,并对其进行优化.

  20. Efficient Unsteady Flow Visualization with High-Order Access Dependencies

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Jiang; Guo, Hanqi; Yuan, Xiaoru

    2016-04-19

    We present a novel high-order access dependencies based model for efficient pathline computation in unsteady flow visualization. By taking longer access sequences into account to model more sophisticated data access patterns in particle tracing, our method greatly improves the accuracy and reliability in data access prediction. In our work, high-order access dependencies are calculated by tracing uniformly-seeded pathlines in both forward and backward directions in a preprocessing stage. The effectiveness of our proposed approach is demonstrated through a parallel particle tracing framework with high-order data prefetching. Results show that our method achieves higher data locality and hence improves the efficiency of pathline computation.

  1. High-efficiency solar concentrator

    Science.gov (United States)

    Lansing, F. L.; Dorman, J.

    1980-01-01

    A new type of solar concentrator is presented using liquid lenses and simple translational tracking mechanism. The concentrator achieves a 100:1 nominal concentration ratio and is compared in performance with a flat-plate collector having two sheets of glazing and non-selective coating. The results of the thermal analysis show that higher temperatures can be obtained with the concentrator than is possible with the non-concentrator flat-plate type. Furthermore, the thermal efficiency far exceeds that of the comparative flat-plate type for all operating conditions.

  2. Monitoring SLAC High Performance UNIX Computing Systems

    Energy Technology Data Exchange (ETDEWEB)

    Lettsome, Annette K.; /Bethune-Cookman Coll. /SLAC

    2005-12-15

    Knowledge of the effectiveness and efficiency of computers is important when working with high performance systems. The monitoring of such systems is advantageous in order to foresee possible misfortunes or system failures. Ganglia is a software system designed for high performance computing systems to retrieve specific monitoring information. An alternative storage facility for Ganglia's collected data is needed since its default storage system, the round-robin database (RRD), struggles with data integrity. The creation of a script-driven MySQL database solves this dilemma. This paper describes the process took in the creation and implementation of the MySQL database for use by Ganglia. Comparisons between data storage by both databases are made using gnuplot and Ganglia's real-time graphical user interface.

  3. RIOT: I/O-Efficient Numerical Computing without SQL

    CERN Document Server

    Zhang, Yi; Yang, Jun

    2009-01-01

    R is a numerical computing environment that is widely popular for statistical data analysis. Like many such environments, R performs poorly for large datasets whose sizes exceed that of physical memory. We present our vision of RIOT (R with I/O Transparency), a system that makes R programs I/O-efficient in a way transparent to the users. We describe our experience with RIOT-DB, an initial prototype that uses a relational database system as a backend. Despite the overhead and inadequacy of generic database systems in handling array data and numerical computation, RIOT-DB significantly outperforms R in many large-data scenarios, thanks to a suite of high-level, inter-operation optimizations that integrate seamlessly into R. While many techniques in RIOT are inspired by databases (and, for RIOT-DB, realized by a database system), RIOT users are insulated from anything database related. Compared with previous approaches that require users to learn new languages and rewrite their programs to interface with a datab...

  4. The NumPy array: a structure for efficient numerical computation

    CERN Document Server

    Van Der Walt, Stefan; Varoquaux, Gaël

    2011-01-01

    In the Python world, NumPy arrays are the standard representation for numerical data. Here, we show how these arrays enable efficient implementation of numerical computations in a high-level language. Overall, three techniques are applied to improve performance: vectorizing calculations, avoiding copying data in memory, and minimizing operation counts. We first present the NumPy array structure, then show how to use it for efficient computation, and finally how to share array data with other libraries.

  5. The NumPy array: a structure for efficient numerical computation

    OpenAIRE

    Van der Walt, Stefan; Colbert, S. Chris; Varoquaux, Gaël

    2011-01-01

    International audience; In the Python world, NumPy arrays are the standard representation for numerical data. Here, we show how these arrays enable efficient implementation of numerical computations in a high-level language. Overall, three techniques are applied to improve performance: vectorizing calculations, avoiding copying data in memory, and minimizing operation counts. We first present the NumPy array structure, then show how to use it for efficient computation, and finally how to shar...

  6. Repeatable, accurate, and high speed multi-level programming of memristor 1T1R arrays for power efficient analog computing applications.

    Science.gov (United States)

    Merced-Grafals, Emmanuelle J; Dávila, Noraica; Ge, Ning; Williams, R Stanley; Strachan, John Paul

    2016-09-09

    Beyond use as high density non-volatile memories, memristors have potential as synaptic components of neuromorphic systems. We investigated the suitability of tantalum oxide (TaOx) transistor-memristor (1T1R) arrays for such applications, particularly the ability to accurately, repeatedly, and rapidly reach arbitrary conductance states. Programming is performed by applying an adaptive pulsed algorithm that utilizes the transistor gate voltage to control the SET switching operation and increase programming speed of the 1T1R cells. We show the capability of programming 64 conductance levels with programming speed and programming error. The algorithm is also utilized to program 16 conductance levels on a population of cells in the 1T1R array showing robustness to cell-to-cell variability. In general, the proposed algorithm results in approximately 10× improvement in programming speed over standard algorithms that do not use the transistor gate to control memristor switching. In addition, after only two programming pulses (an initialization pulse followed by a programming pulse), the resulting conductance values are within 12% of the target values in all cases. Finally, endurance of more than 10(6) cycles is shown through open-loop (single pulses) programming across multiple conductance levels using the optimized gate voltage of the transistor. These results are relevant for applications that require high speed, accurate, and repeatable programming of the cells such as in neural networks and analog data processing.

  7. Repeatable, accurate, and high speed multi-level programming of memristor 1T1R arrays for power efficient analog computing applications

    Science.gov (United States)

    Merced-Grafals, Emmanuelle J.; Dávila, Noraica; Ge, Ning; Williams, R. Stanley; Strachan, John Paul

    2016-09-01

    Beyond use as high density non-volatile memories, memristors have potential as synaptic components of neuromorphic systems. We investigated the suitability of tantalum oxide (TaOx) transistor-memristor (1T1R) arrays for such applications, particularly the ability to accurately, repeatedly, and rapidly reach arbitrary conductance states. Programming is performed by applying an adaptive pulsed algorithm that utilizes the transistor gate voltage to control the SET switching operation and increase programming speed of the 1T1R cells. We show the capability of programming 64 conductance levels with cells in the 1T1R array showing robustness to cell-to-cell variability. In general, the proposed algorithm results in approximately 10× improvement in programming speed over standard algorithms that do not use the transistor gate to control memristor switching. In addition, after only two programming pulses (an initialization pulse followed by a programming pulse), the resulting conductance values are within 12% of the target values in all cases. Finally, endurance of more than 106 cycles is shown through open-loop (single pulses) programming across multiple conductance levels using the optimized gate voltage of the transistor. These results are relevant for applications that require high speed, accurate, and repeatable programming of the cells such as in neural networks and analog data processing.

  8. EFFICIENT VM LOAD BALANCING ALGORITHM FOR A CLOUD COMPUTING ENVIRONMENT

    Directory of Open Access Journals (Sweden)

    Jasmin James

    2012-09-01

    Full Text Available Cloud computing is a fast growing area in computing research and industry today. With the advancement of the Cloud, there are new possibilities opening up on how applications can be built and how different services can be offered to the end user through Virtualization, on the internet. There are the cloud service providers who provide large scaled computing infrastructure defined on usage, and provide the infrastructure services in a very flexiblemanner which the users can scale up or down at will. The establishment of an effective load balancing algorithm and how to use Cloud computing resources efficiently for effective and efficient cloud computing is one of the Cloud computing service providers’ultimate goals. In this paper firstly analysis of different Virtual Machine (VM load balancing algorithms is done. Secondly, a new VM load balancing algorithm has been proposed and implemented for an IaaS framework in Simulated cloud computing environment; i.e. ‘Weighted Active Monitoring Load Balancing Algorithm’ using CloudSimtools, for the Datacenter to effectively load balance requests between the available virtual machines assigning a weight, in order to achieve better performance parameters such as response time and Data processing time.

  9. Popescu-Rohrlich correlations imply efficient instantaneous nonlocal quantum computation

    Science.gov (United States)

    Broadbent, Anne

    2016-08-01

    In instantaneous nonlocal quantum computation, two parties cooperate in order to perform a quantum computation on their joint inputs, while being restricted to a single round of simultaneous communication. Previous results showed that instantaneous nonlocal quantum computation is possible, at the cost of an exponential amount of prior shared entanglement (in the size of the input). Here, we show that a linear amount of entanglement suffices, (in the size of the computation), as long as the parties share nonlocal correlations as given by the Popescu-Rohrlich box. This means that communication is not required for efficient instantaneous nonlocal quantum computation. Exploiting the well-known relation to position-based cryptography, our result also implies the impossibility of secure position-based cryptography against adversaries with nonsignaling correlations. Furthermore, our construction establishes a quantum analog of the classical communication complexity collapse under nonsignaling correlations.

  10. High-Efficiency dc/dc Converter

    Science.gov (United States)

    Sturman, J.

    1982-01-01

    High-efficiency dc/dc converter has been developed that provides commonly used voltages of plus or minus 12 Volts from an unregulated dc source of from 14 to 40 Volts. Unique features of converter are its high efficiency at low power level and ability to provide output either larger or smaller than input voltage.

  11. High Efficiency Microwave Power Amplifier (HEMPA) Design

    Science.gov (United States)

    Sims, W. Herbert

    2004-01-01

    This paper will focus on developing an exotic switching technique that enhances the DC-to-RF conversion efficiency of microwave power amplifiers. For years, switching techniques implemented in the 10 kHz to 30 MHz region have resulted in DC-to-RF conversion efficiencies of 90-95-percent. Currently amplifier conversion efficiency, in the 2-3 GHz region approaches, 10-20-percent. Using a combination of analytical modeling and hardware testing, a High Efficiency Microwave Power Amplifier was built that demonstrated conversion efficiencies four to five times higher than current state of the art.

  12. Multicolor, High Efficiency, Nanotextured LEDs

    Energy Technology Data Exchange (ETDEWEB)

    Jung Han; Arto Nurmikko

    2011-09-30

    We report on research results in this project which synergize advanced material science approaches with fundamental optical physics concepts pertaining to light-matter interaction, with the goal of solving seminal problems for the development of very high performance light emitting diodes (LEDs) in the blue and green for Solid State Lighting applications. Accomplishments in the duration of the contract period include (i) heteroepitaxy of nitrogen-polar LEDs on sapphire, (ii) heteroepitaxy of semipolar (11{bar 2}2) green LEDs on sapphire, (iii) synthesis of quantum-dot loaded nanoporous GaN that emits white light without phosphor conversion, (iv) demonstration of the highest quality semipolar (11{bar 2}2) GaN on sapphire using orientation-controlled epitaxy, (v) synthesis of nanoscale GaN and InGaN medium, and (vi) development of a novel liftoff process for manufacturing GaN thin-film vertical LEDs. The body of results is presented in this report shows how a solid foundation has been laid, with several noticeable accomplishments, for innovative research, consistent with the stated milestones.

  13. Efficient Approach for Load Balancing in Virtual Cloud Computing Environment

    Directory of Open Access Journals (Sweden)

    Harvinder singh

    2014-10-01

    Full Text Available Cloud computing technology is changing the focus of IT world and it is becoming famous because of its great characteristics. Load balancing is one of the main challenges in cloud computing for distributing workloads across multiple computers or a computer cluster, network links, central processing units, disk drives, or other resources. Successful load balancing optimizes resource use, maximizes throughput, minimizes response time, and avoids overload. The objective of this paper to propose an approach for scheduling algorithms that can maintain the load balancing and provides better improved strategies through efficient job scheduling and modified resource allocation techniques. The results discussed in this paper, based on existing round robin, least connection, throttled load balance, fastest response time and a new proposed algorithm fastest with least connection scheduling algorithms. This new algorithm identifies the overall response time and data centre processing time is improved as well as cost is reduced in comparison to the existing scheduling parameters.

  14. EFFICIENT APPROACH FOR LOAD BALANCING IN VIRTUAL CLOUD COMPUTING ENVIRONMENT

    Directory of Open Access Journals (Sweden)

    Harvinder Singh

    2015-10-01

    Full Text Available Cloud computing technology is changing the focus of IT world and it is becoming famous because of its great characteristics. Load balancing is one of the main challenges in cloud computing for distributing workloads across multiple computers or a computer cluster, network links, central processing units, disk drives, or other resources. Successful load balancing optimizes resource use, maximizes throughput, minimizes response time, and avoids overload. The objective of this paper to propose an approach for scheduling algorithms that can maintain the load balancing and provides better improved strategies through efficient job scheduling and modified resource allocation techniques. The results discussed in this paper, based on existing round robin, least connection, throttled load balance, fastest response time and a new proposed algorithm fastest with least connection scheduling algorithms. This new algorithm identifies the overall response time and data centre processing time is improved as well as cost is reduced in comparison to the existing scheduling parameters.

  15. Efficient computation of clipped Voronoi diagram for mesh generation

    KAUST Repository

    Yan, Dongming

    2013-04-01

    The Voronoi diagram is a fundamental geometric structure widely used in various fields, especially in computer graphics and geometry computing. For a set of points in a compact domain (i.e. a bounded and closed 2D region or a 3D volume), some Voronoi cells of their Voronoi diagram are infinite or partially outside of the domain, but in practice only the parts of the cells inside the domain are needed, as when computing the centroidal Voronoi tessellation. Such a Voronoi diagram confined to a compact domain is called a clipped Voronoi diagram. We present an efficient algorithm to compute the clipped Voronoi diagram for a set of sites with respect to a compact 2D region or a 3D volume. We also apply the proposed method to optimal mesh generation based on the centroidal Voronoi tessellation. Crown Copyright © 2011 Published by Elsevier Ltd. All rights reserved.

  16. Highly-efficient high-power pumps for fiber lasers

    Science.gov (United States)

    Gapontsev, V.; Moshegov, N.; Berezin, I.; Komissarov, A.; Trubenko, P.; Miftakhutdinov, D.; Berishev, I.; Chuyanov, V.; Raisky, O.; Ovtchinnikov, A.

    2017-02-01

    We report on high efficiency multimode pumps that enable ultra-high efficiency high power ECO Fiber Lasers. We discuss chip and packaged pump design and performance. Peak out-of-fiber power efficiency of ECO Fiber Laser pumps was reported to be as high as 68% and was achieved with passive cooling. For applications that do not require Fiber Lasers with ultimate power efficiency, we have developed passively cooled pumps with out-of-fiber power efficiency greater than 50%, maintained at operating current up to 22A. We report on approaches to diode chip and packaged pump design that possess such performance.

  17. High Energy Computed Tomographic Inspection of Munitions

    Science.gov (United States)

    2016-11-01

    UNCLASSIFIED UNCLASSIFIED AD-E403 815 Technical Report AREIS-TR-16006 HIGH ENERGY COMPUTED TOMOGRAPHIC INSPECTION OF MUNITIONS...REPORT DATE (DD-MM-YYYY) November 2016 2. REPORT TYPE Final 3. DATES COVERED (From – To) 4. TITLE AND SUBTITLE HIGH ENERGY COMPUTED...otherwise be accomplished by other nondestructive testing methods. 15. SUBJECT TERMS Radiography High energy Computed tomography (CT

  18. Efficient quantum circuits for one-way quantum computing.

    Science.gov (United States)

    Tanamoto, Tetsufumi; Liu, Yu-Xi; Hu, Xuedong; Nori, Franco

    2009-03-13

    While Ising-type interactions are ideal for implementing controlled phase flip gates in one-way quantum computing, natural interactions between solid-state qubits are most often described by either the XY or the Heisenberg models. We show an efficient way of generating cluster states directly using either the imaginary SWAP (iSWAP) gate for the XY model, or the sqrt[SWAP] gate for the Heisenberg model. Our approach thus makes one-way quantum computing more feasible for solid-state devices.

  19. A Computationally Efficient Aggregation Optimization Strategy of Model Predictive Control

    Institute of Scientific and Technical Information of China (English)

    2002-01-01

    Model Predictive Control (MPC) is a popular technique and has been successfully used in various industrial applications. However, the big drawback of MPC involved in the formidable on-line computational effort limits its applicability to relatively slow and/or small processes with a moderate number of inputs. This paper develops an aggregation optimization strategy for MPC that can improve the computational efficiency of MPC. For the regulation problem, an input decaying aggregation optimization algorithm is presented by aggregating all the original optimized variables on control horizon with the decaying sequence in respect of the current control action.

  20. Universally Composable Efficient Multiparty Computation from Threshold Homomorphic Encryption

    DEFF Research Database (Denmark)

    Damgård, Ivan Bjerre; Nielsen, Jesper Buus

    2003-01-01

    We present a new general multiparty computation protocol for the cryptographic scenario which is universally composable — in particular, it is secure against an active and adaptive adversary, corrupting any minority of the parties. The protocol is as efficient as the best known statically secure...... solutions, in particular the number of bits broadcast (which dominates the complexity) is Ω (nk |C|), where n is the number of parties, k is a security parameter, and |C| is the size of a circuit doing the desired computation. Unlike previous adaptively secure protocols for the cryptographic model, our...

  1. Very High Efficiency Solar Cell Modules

    Energy Technology Data Exchange (ETDEWEB)

    Barnett, A.; Kirkpatrick, D.; Honsberg, C.; Moore, D.; Wanlass, M.; Emery, K.; Schwartz, R.; Carlson, D.; Bowden, S.; Aiken, D.; Gray, A.; Kurtz, S.; Kazmerski, L., et al

    2009-01-01

    The Very High Efficiency Solar Cell (VHESC) program is developing integrated optical system - PV modules for portable applications that operate at greater than 50% efficiency. We are integrating the optical design with the solar cell design, and have entered previously unoccupied design space. Our approach is driven by proven quantitative models for the solar cell design, the optical design, and the integration of these designs. Optical systems efficiency with an optical efficiency of 93% and solar cell device results under ideal dichroic splitting optics summing to 42.7 {+-} 2.5% are described.

  2. Energy-efficient computing and networking. Revised selected papers

    Energy Technology Data Exchange (ETDEWEB)

    Hatziargyriou, Nikos; Dimeas, Aris [Ethnikon Metsovion Polytechneion, Athens (Greece); Weidlich, Anke (eds.) [SAP Research Center, Karlsruhe (Germany); Tomtsi, Thomai

    2011-07-01

    This book constitutes the postproceedings of the First International Conference on Energy-Efficient Computing and Networking, E-Energy, held in Passau, Germany in April 2010. The 23 revised papers presented were carefully reviewed and selected for inclusion in the post-proceedings. The papers are organized in topical sections on energy market and algorithms, ICT technology for the energy market, implementation of smart grid and smart home technology, microgrids and energy management, and energy efficiency through distributed energy management and buildings. (orig.)

  3. High Efficiency Low Scatter Echelle Grating Project

    Data.gov (United States)

    National Aeronautics and Space Administration — A high efficiency low scatter echelle grating will be developed using a novel technique of multiple diamond shaving cuts. The grating will have mirror surfaces on...

  4. Multi Band Gap High Efficiency Converter (RAINBOW)

    Science.gov (United States)

    Bekey, I.; Lewis, C.; Phillips, W.; Shields, V.; Stella, P.

    1997-01-01

    The RAINBOW multi band gap system represents a unique combination of solar cells, concentrators and beam splitters. RAINBOW is a flexible system which can readily expand as new high efficiency components are developed.

  5. High Efficiency Solar Furnace Core Project

    Data.gov (United States)

    National Aeronautics and Space Administration — It is proposed to develop a high efficiency solar furnace core that greatly lessens the heat losses from the furnace core, either greatly reducing the amount of...

  6. Efficient Protocols for Principal Eigenvector Computation over Private Data

    Directory of Open Access Journals (Sweden)

    Manas A. Pathak

    2011-12-01

    Full Text Available In this paper we present a protocol for computing the principal eigenvector of a collection of data matrices belonging to multiple semi-honest parties with privacy constraints. Our proposed protocol is based on secure multi-party computation with a semi-honest arbitrator who deals with data encrypted by the other parties using an additive homomorphic cryptosystem. We augment the protocol with randomization and oblivious transfer to make it difficult for any party to estimate properties of the data belonging to other parties from the intermediate steps. The previous approaches towards this problem were based on expensive QR decomposition of correlation matrices, we present an efficient algorithm using the power iteration method. We present an analysis of the correctness, security, and efficiency of protocol along with experimental results using a prototype implementation over simulated data and USPS handwritten digits dataset.

  7. Compact High Efficiency Adsorption Heat Pump

    OpenAIRE

    TeGrotenhuis, Ward E; Humble, Paul H; Sweeney, Josh B

    2012-01-01

    An innovative adsorption cycle heat pump technology is presented that is compact and capable of achieving high energy efficiency for integrated space heating, air conditioning, and water heating. High energy efficiency is accomplished by effectively recuperating heat within the system to minimize energy consumption. This substantially reduces the thermodynamic losses that occur when the sorbent beds are thermally cycled without effective heat recuperation. Furthermore, equipment cost is reduc...

  8. Computationally Efficient and Noise Robust DOA and Pitch Estimation

    DEFF Research Database (Denmark)

    Karimian-Azari, Sam; Jensen, Jesper Rindom; Christensen, Mads Græsbøll

    2016-01-01

    a joint DOA and pitch estimator. In white Gaussian noise, we derive even more computationally efficient solutions which are designed using the narrowband power spectrum of the harmonics. Numerical results reveal the performance of the estimators in colored noise compared with the Cram\\'{e}r-Rao lower...... bound. Experiments on real-life signals indicate the applicability of the methods in practical low local signal-to-noise ratios....

  9. Universally Composable Efficient Multiparty Computation from Threshold Homomorphic Encryption

    DEFF Research Database (Denmark)

    Damgård, Ivan Bjerre; Nielsen, Jesper Buus

    2003-01-01

    We present a new general multiparty computation protocol for the cryptographic scenario which is universally composable — in particular, it is secure against an active and adaptive adversary, corrupting any minority of the parties. The protocol is as efficient as the best known statically secure ...... protocol does not use non-committing encryption, instead it is based on homomorphic threshold encryption, in particular the Paillier cryptosystem....

  10. High efficiency quantum cascade laser frequency comb

    Science.gov (United States)

    Lu, Quanyong; Wu, Donghai; Slivken, Steven; Razeghi, Manijeh

    2017-03-01

    An efficient mid-infrared frequency comb source is of great interest to high speed, high resolution spectroscopy and metrology. Here we demonstrate a mid-IR quantum cascade laser frequency comb with a high power output and narrow beatnote linewidth at room temperature. The active region was designed with a strong-coupling between the injector and the upper lasing level for high internal quantum efficiency and a broadband gain. The group velocity dispersion was engineered for efficient, broadband mode-locking via four wave mixing. The comb device exhibits a narrow intermode beatnote linewidth of 50.5 Hz and a maximum wall-plug efficiency of 6.5% covering a spectral coverage of 110 cm‑1 at λ ~ 8 μm. The efficiency is improved by a factor of 6 compared with previous demonstrations. The high power efficiency and narrow beatnote linewidth will greatly expand the applications of quantum cascade laser frequency combs including high-precision remote sensing and spectroscopy.

  11. High efficiency quantum cascade laser frequency comb

    Science.gov (United States)

    Lu, Quanyong; Wu, Donghai; Slivken, Steven; Razeghi, Manijeh

    2017-01-01

    An efficient mid-infrared frequency comb source is of great interest to high speed, high resolution spectroscopy and metrology. Here we demonstrate a mid-IR quantum cascade laser frequency comb with a high power output and narrow beatnote linewidth at room temperature. The active region was designed with a strong-coupling between the injector and the upper lasing level for high internal quantum efficiency and a broadband gain. The group velocity dispersion was engineered for efficient, broadband mode-locking via four wave mixing. The comb device exhibits a narrow intermode beatnote linewidth of 50.5 Hz and a maximum wall-plug efficiency of 6.5% covering a spectral coverage of 110 cm−1 at λ ~ 8 μm. The efficiency is improved by a factor of 6 compared with previous demonstrations. The high power efficiency and narrow beatnote linewidth will greatly expand the applications of quantum cascade laser frequency combs including high-precision remote sensing and spectroscopy. PMID:28262834

  12. High Efficiency Polymer Solar Cells Technologies

    Institute of Scientific and Technical Information of China (English)

    Abdrhman M G; LI Hang-quan; ZHANG Li-ye; ZHOU Bing

    2006-01-01

    The conjugated polymer-based solar cell is one of the most promising devices in search of sustainable, renewable energy sources in last decade. It is the youngest field in organic solar cell research and also is certainly the fastest growing one at the moment. In addition, the key factor for polymer-based solar cells with high-efficiency is to invent new materials. Organic solar cell has attracted significant researches and commercial interest due to its low cost in fabrication and flexibility in applications. However, they suffer from relatively low conversion efficiency. The summarization of the significance and concept of high efficiency polymer solar cell technologies are presented.

  13. Efficient MATLAB computations with sparse and factored tensors.

    Energy Technology Data Exchange (ETDEWEB)

    Bader, Brett William; Kolda, Tamara Gibson (Sandia National Lab, Livermore, CA)

    2006-12-01

    In this paper, the term tensor refers simply to a multidimensional or N-way array, and we consider how specially structured tensors allow for efficient storage and computation. First, we study sparse tensors, which have the property that the vast majority of the elements are zero. We propose storing sparse tensors using coordinate format and describe the computational efficiency of this scheme for various mathematical operations, including those typical to tensor decomposition algorithms. Second, we study factored tensors, which have the property that they can be assembled from more basic components. We consider two specific types: a Tucker tensor can be expressed as the product of a core tensor (which itself may be dense, sparse, or factored) and a matrix along each mode, and a Kruskal tensor can be expressed as the sum of rank-1 tensors. We are interested in the case where the storage of the components is less than the storage of the full tensor, and we demonstrate that many elementary operations can be computed using only the components. All of the efficiencies described in this paper are implemented in the Tensor Toolbox for MATLAB.

  14. A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems

    CERN Document Server

    Beloglazov, Anton; Lee, Young Choon; Zomaya, Albert

    2010-01-01

    Traditionally, the development of computing systems has been focused on performance improvements driven by the demand of applications from consumer, scientific and business domains. However, the ever increasing energy consumption of computing systems has started to limit further performance growth due to overwhelming electricity bills and carbon dioxide footprints. Therefore, the goal of the computer system design has been shifted to power and energy efficiency. To identify open challenges in the area and facilitate future advancements it is essential to synthesize and classify the research on power and energy-efficient design conducted to date. In this work we discuss causes and problems of high power / energy consumption, and present a taxonomy of energy-efficient design of computing systems covering the hardware, operating system, virtualization and data center levels. We survey various key works in the area and map them to our taxonomy to guide future design and development efforts. This chapter is conclu...

  15. Recovery Act - CAREER: Sustainable Silicon -- Energy-Efficient VLSI Interconnect for Extreme-Scale Computing

    Energy Technology Data Exchange (ETDEWEB)

    Chiang, Patrick [Oregon State Univ., Corvallis, OR (United States)

    2014-01-31

    The research goal of this CAREER proposal is to develop energy-efficient, VLSI interconnect circuits and systems that will facilitate future massively-parallel, high-performance computing. Extreme-scale computing will exhibit massive parallelism on multiple vertical levels, from thou­ sands of computational units on a single processor to thousands of processors in a single data center. Unfortunately, the energy required to communicate between these units at every level (on­ chip, off-chip, off-rack) will be the critical limitation to energy efficiency. Therefore, the PI's career goal is to become a leading researcher in the design of energy-efficient VLSI interconnect for future computing systems.

  16. Costs evaluation methodic of energy efficient computer network reengineering

    Directory of Open Access Journals (Sweden)

    S.A. Nesterenko

    2016-09-01

    Full Text Available A key direction of modern computer networks reengineering is their transfer to a new energy-saving technology IEEE 802.3az. To make a reasoned decision about the transition to the new technology is needed a technique that allows network engineers to answer the question about the economic feasibility of a network upgrade. Aim: The aim of this research is development of methodic for calculating the cost-effectiveness of energy-efficient computer network reengineering. Materials and Methods: The methodic uses analytical models for calculating power consumption of a computer network port operating in IEEE 802.3 standard and energy-efficient mode of IEEE 802.3az standard. For frame transmission time calculation in the communication channel used the queuing model. To determine the values of the network operation parameters proposed to use multiagent network monitoring method. Results: The methodic allows calculating the economic impact of a computer network transfer to energy-saving technology IEEE 802.3az. To determine the network performance parameters proposed to use network SNMP monitoring systems based on RMON MIB agents.

  17. A Simple and Resource-efficient Setup for the Computer-aided Drug Design Laboratory.

    Science.gov (United States)

    Moretti, Loris; Sartori, Luca

    2016-10-01

    Undertaking modelling investigations for Computer-Aided Drug Design (CADD) requires a proper environment. In principle, this could be done on a single computer, but the reality of a drug discovery program requires robustness and high-throughput computing (HTC) to efficiently support the research. Therefore, a more capable alternative is needed but its implementation has no widespread solution. Here, the realization of such a computing facility is discussed, from general layout to technical details all aspects are covered. © 2016 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  18. CRPC research into linear algebra software for high performance computers

    Energy Technology Data Exchange (ETDEWEB)

    Choi, J.; Walker, D.W. [Oak Ridge National Lab., TN (United States). Mathematical Sciences Section; Dongarra, J.J. [Univ. of Tennessee, Knoxville, TN (United States). Dept. of Computer Science]|[Oak Ridge National Lab., TN (United States). Mathematical Sciences Section; Pozo, R. [Univ. of Tennessee, Knoxville, TN (United States). Dept. of Computer Science; Sorensen, D.C. [Rice Univ., Houston, TX (United States). Dept. of Computational and Applied Mathematics

    1994-12-31

    In this paper the authors look at a number of approaches being investigated in the Center for Research on Parallel Computation (CRPC) to develop linear algebra software for high-performance computers. These approaches are exemplified by the LAPACK, templates, and ARPACK projects. LAPACK is a software library for performing dense and banded linear algebra computations, and was designed to run efficiently on high-performance computers. The authors focus on the design of the distributed-memory version of LAPACK, and on an object-oriented interface to LAPACK.

  19. Computing support for High Energy Physics

    Energy Technology Data Exchange (ETDEWEB)

    Avery, P.; Yelton, J. [Univ. of Florida, Gainesville, FL (United States)

    1996-12-01

    This computing proposal (Task S) is submitted separately but in support of the High Energy Experiment (CLEO, Fermilab, CMS) and Theory tasks. The authors have built a very strong computing base at Florida over the past 8 years. In fact, computing has been one of the main contributions to their experimental collaborations, involving not just computing capacity for running Monte Carlos and data reduction, but participation in many computing initiatives, industrial partnerships, computing committees and collaborations. These facts justify the submission of a separate computing proposal.

  20. Energy-Efficient Scheduling of HPC Applications in Cloud Computing Environments

    CERN Document Server

    Garg, Saurabh Kumar; Anandasivam, Arun; Buyya, Rajkumar

    2009-01-01

    The use of High Performance Computing (HPC) in commercial and consumer IT applications is becoming popular. They need the ability to gain rapid and scalable access to high-end computing capabilities. Cloud computing promises to deliver such a computing infrastructure using data centers so that HPC users can access applications and data from a Cloud anywhere in the world on demand and pay based on what they use. However, the growing demand drastically increases the energy consumption of data centers, which has become a critical issue. High energy consumption not only translates to high energy cost, which will reduce the profit margin of Cloud providers, but also high carbon emissions which is not environmentally sustainable. Hence, energy-efficient solutions are required that can address the high increase in the energy consumption from the perspective of not only Cloud provider but also from the environment. To address this issue we propose near-optimal scheduling policies that exploits heterogeneity across mu...

  1. Energy efficient hybrid computing systems using spin devices

    Science.gov (United States)

    Sharad, Mrigank

    Emerging spin-devices like magnetic tunnel junctions (MTJ's), spin-valves and domain wall magnets (DWM) have opened new avenues for spin-based logic design. This work explored potential computing applications which can exploit such devices for higher energy-efficiency and performance. The proposed applications involve hybrid design schemes, where charge-based devices supplement the spin-devices, to gain large benefits at the system level. As an example, lateral spin valves (LSV) involve switching of nanomagnets using spin-polarized current injection through a metallic channel such as Cu. Such spin-torque based devices possess several interesting properties that can be exploited for ultra-low power computation. Analog characteristic of spin current facilitate non-Boolean computation like majority evaluation that can be used to model a neuron. The magneto-metallic neurons can operate at ultra-low terminal voltage of ˜20mV, thereby resulting in small computation power. Moreover, since nano-magnets inherently act as memory elements, these devices can facilitate integration of logic and memory in interesting ways. The spin based neurons can be integrated with CMOS and other emerging devices leading to different classes of neuromorphic/non-Von-Neumann architectures. The spin-based designs involve `mixed-mode' processing and hence can provide very compact and ultra-low energy solutions for complex computation blocks, both digital as well as analog. Such low-power, hybrid designs can be suitable for various data processing applications like cognitive computing, associative memory, and currentmode on-chip global interconnects. Simulation results for these applications based on device-circuit co-simulation framework predict more than ˜100x improvement in computation energy as compared to state of the art CMOS design, for optimal spin-device parameters.

  2. The path toward HEP High Performance Computing

    Science.gov (United States)

    Apostolakis, John; Brun, René; Carminati, Federico; Gheata, Andrei; Wenzel, Sandro

    2014-06-01

    High Energy Physics code has been known for making poor use of high performance computing architectures. Efforts in optimising HEP code on vector and RISC architectures have yield limited results and recent studies have shown that, on modern architectures, it achieves a performance between 10% and 50% of the peak one. Although several successful attempts have been made to port selected codes on GPUs, no major HEP code suite has a "High Performance" implementation. With LHC undergoing a major upgrade and a number of challenging experiments on the drawing board, HEP cannot any longer neglect the less-than-optimal performance of its code and it has to try making the best usage of the hardware. This activity is one of the foci of the SFT group at CERN, which hosts, among others, the Root and Geant4 project. The activity of the experiments is shared and coordinated via a Concurrency Forum, where the experience in optimising HEP code is presented and discussed. Another activity is the Geant-V project, centred on the development of a highperformance prototype for particle transport. Achieving a good concurrency level on the emerging parallel architectures without a complete redesign of the framework can only be done by parallelizing at event level, or with a much larger effort at track level. Apart the shareable data structures, this typically implies a multiplication factor in terms of memory consumption compared to the single threaded version, together with sub-optimal handling of event processing tails. Besides this, the low level instruction pipelining of modern processors cannot be used efficiently to speedup the program. We have implemented a framework that allows scheduling vectors of particles to an arbitrary number of computing resources in a fine grain parallel approach. The talk will review the current optimisation activities within the SFT group with a particular emphasis on the development perspectives towards a simulation framework able to profit best from

  3. High Efficiency ELID Grinding of Garnet Ferrite

    Institute of Scientific and Technical Information of China (English)

    2002-01-01

    Hard and brittle materials such as ferrite, optical glass and ceramics have been widely used in many fields because of their good characteristics and still gain more attentions. However, it is difficult to machine and get good surface quality. Some parts made of these materials have large machining allowances and need to be produced with large batch, but the machining efficiency is very low with usual grinding method. So it is of great importance to research the high efficiency grinding technology of hard ...

  4. Improving robustness and computational efficiency using modern C++

    Energy Technology Data Exchange (ETDEWEB)

    Paterno, M. [Fermilab; Kowalkowski, J. [Fermilab; Green, C. [Fermilab

    2014-01-01

    For nearly two decades, the C++ programming language has been the dominant programming language for experimental HEP. The publication of ISO/IEC 14882:2011, the current version of the international standard for the C++ programming language, makes available a variety of language and library facilities for improving the robustness, expressiveness, and computational efficiency of C++ code. However, much of the C++ written by the experimental HEP community does not take advantage of the features of the language to obtain these benefits, either due to lack of familiarity with these features or concern that these features must somehow be computationally inefficient. In this paper, we address some of the features of modern C+-+, and show how they can be used to make programs that are both robust and computationally efficient. We compare and contrast simple yet realistic examples of some common implementation patterns in C, currently-typical C++, and modern C++, and show (when necessary, down to the level of generated assembly language code) the quality of the executable code produced by recent C++ compilers, with the aim of allowing the HEP community to make informed decisions on the costs and benefits of the use of modern C++.

  5. Progress and Challenges in High Performance Computer Technology

    Institute of Scientific and Technical Information of China (English)

    Xue-Jun Yang; Yong Dou; Qing-Feng Hu

    2006-01-01

    High performance computers provide strategic computing power in the construction of national economy and defense, and become one of symbols of the country's overall strength. Over 30 years, with the supports of governments, the technology of high performance computers is in the process of rapid development, during which the computing performance increases nearly 3 million times and the processors number expands over 10 hundred thousands times. To solve the critical issues related with parallel efficiency and scalability, scientific researchers pursued extensive theoretical studies and technical innovations. The paper briefly looks back the course of building high performance computer systems both at home and abroad,and summarizes the significant breakthroughs of international high performance computer technology. We also overview the technology progress of China in the area of parallel computer architecture, parallel operating system and resource management,parallel compiler and performance optimization, environment for parallel programming and network computing. Finally, we examine the challenging issues, "memory wall", system scalability and "power wall", and discuss the issues of high productivity computers, which is the trend in building next generation high performance computers.

  6. High-performance computing using FPGAs

    CERN Document Server

    Benkrid, Khaled

    2013-01-01

    This book is concerned with the emerging field of High Performance Reconfigurable Computing (HPRC), which aims to harness the high performance and relative low power of reconfigurable hardware–in the form Field Programmable Gate Arrays (FPGAs)–in High Performance Computing (HPC) applications. It presents the latest developments in this field from applications, architecture, and tools and methodologies points of view. We hope that this work will form a reference for existing researchers in the field, and entice new researchers and developers to join the HPRC community.  The book includes:  Thirteen application chapters which present the most important application areas tackled by high performance reconfigurable computers, namely: financial computing, bioinformatics and computational biology, data search and processing, stencil computation e.g. computational fluid dynamics and seismic modeling, cryptanalysis, astronomical N-body simulation, and circuit simulation.     Seven architecture chapters which...

  7. Technology Development for High Efficiency Optical Communications

    Science.gov (United States)

    Farr, William H.

    2012-01-01

    Deep space optical communications is a significantly more challenging operational domain than near Earth space optical communications, primarily due to effects resulting from the vastly increased range between transmitter and receiver. The NASA Game Changing Development Program Deep Space Optical Communications Project is developing four key technologies for the implementation of a high efficiency telecommunications system that will enable greater than 10X the data rate of a state-of-the-art deep space RF system (Ka-band) for similar transceiver mass and power burden on the spacecraft. These technologies are a low mass spacecraft disturbance isolation assembly, a flight qualified photon counting detector array, a high efficiency flight laser amplifier and a high efficiency photon counting detector array for the ground-based receiver.

  8. NaI(Tl) Detector Efficiency Computation Using Radioactive Parallelepiped Sources Based on Efficiency Transfer Principle

    OpenAIRE

    MOHAMED S. BADAWI; Mona M. Gouda; Ahmed M. El-Khatib; Thabet, Abouzeid A.; Salim, Ahmed A.; Mahmoud I. Abbas

    2015-01-01

    The efficiency transfer (ET) principle is considered as a simple numerical simulation method, which can be used to calculate the full-energy peak efficiency (FEPE) of 3″×3″ NaI(Tl) scintillation detector over a wide energy range. In this work, the calculations of FEPE are based on computing the effective solid angle ratio between a radioactive point and parallelepiped sources located at various distances from the detector surface. Besides, the attenuation of the photon by the source-to-detect...

  9. Measure Guideline: High Efficiency Natural Gas Furnaces

    Energy Technology Data Exchange (ETDEWEB)

    Brand, L.; Rose, W.

    2012-10-01

    This Measure Guideline covers installation of high-efficiency gas furnaces. Topics covered include when to install a high-efficiency gas furnace as a retrofit measure, how to identify and address risks, and the steps to be used in the selection and installation process. The guideline is written for Building America practitioners and HVAC contractors and installers. It includes a compilation of information provided by manufacturers, researchers, and the Department of Energy as well as recent research results from the Partnership for Advanced Residential Retrofit (PARR) Building America team.

  10. Highly efficient heralding of entangled single photons.

    Science.gov (United States)

    Ramelow, Sven; Mech, Alexandra; Giustina, Marissa; Gröblacher, Simon; Wieczorek, Witlef; Beyer, Jörn; Lita, Adriana; Calkins, Brice; Gerrits, Thomas; Nam, Sae Woo; Zeilinger, Anton; Ursin, Rupert

    2013-03-25

    Single photons are an important prerequisite for a broad spectrum of quantum optical applications. We experimentally demonstrate a heralded single-photon source based on spontaneous parametric down-conversion in collinear bulk optics, and fiber-coupled bolometric transition-edge sensors. Without correcting for background, losses, or detection inefficiencies, we measure an overall heralding efficiency of 83%. By violating a Bell inequality, we confirm the single-photon character and high-quality entanglement of our heralded single photons which, in combination with the high heralding efficiency, are a necessary ingredient for advanced quantum communication protocols such as one-sided device-independent quantum key distribution.

  11. Measure Guideline. High Efficiency Natural Gas Furnaces

    Energy Technology Data Exchange (ETDEWEB)

    Brand, L. [Partnership for Advanced Residential Retrofit (PARR), Des Plaines, IL (United States); Rose, W. [Partnership for Advanced Residential Retrofit (PARR), Des Plaines, IL (United States)

    2012-10-01

    This measure guideline covers installation of high-efficiency gas furnaces, including: when to install a high-efficiency gas furnace as a retrofit measure; how to identify and address risks; and the steps to be used in the selection and installation process. The guideline is written for Building America practitioners and HVAC contractors and installers. It includes a compilation of information provided by manufacturers, researchers, and the Department of Energy as well as recent research results from the Partnership for Advanced Residential Retrofit (PARR) Building America team.

  12. Efficient Capacity Computation and Power Optimization for Relay Networks

    CERN Document Server

    Parvaresh, Farzad

    2011-01-01

    The capacity or approximations to capacity of various single-source single-destination relay network models has been characterized in terms of the cut-set upper bound. In principle, a direct computation of this bound requires evaluating the cut capacity over exponentially many cuts. We show that the minimum cut capacity of a relay network under some special assumptions can be cast as a minimization of a submodular function, and as a result, can be computed efficiently. We use this result to show that the capacity, or an approximation to the capacity within a constant gap for the Gaussian, wireless erasure, and Avestimehr-Diggavi-Tse deterministic relay network models can be computed in polynomial time. We present some empirical results showing that computing constant-gap approximations to the capacity of Gaussian relay networks with around 300 nodes can be done in order of minutes. For Gaussian networks, cut-set capacities are also functions of the powers assigned to the nodes. We consider a family of power o...

  13. Efficient Computation of Distance Sketches in Distributed Networks

    CERN Document Server

    Sarma, Atish Das; Pandurangan, Gopal

    2011-01-01

    Distance computation is one of the most fundamental primitives used in communication networks. The cost of effectively and accurately computing pairwise network distances can become prohibitive in large-scale networks such as the Internet and Peer-to-Peer (P2P) networks. To negotiate the rising need for very efficient distance computation, approximation techniques for numerous variants of this question have recently received significant attention in the literature. The goal is to preprocess the graph and store a small amount of information such that whenever a query for any pairwise distance is issued, the distance can be well approximated (i.e., with small stretch) very quickly in an online fashion. Specifically, the pre-processing (usually) involves storing a small sketch with each node, such that at query time only the sketches of the concerned nodes need to be looked up to compute the approximate distance. In this paper, we present the first theoretical study of distance sketches derived from distance ora...

  14. Differential area profiles: decomposition properties and efficient computation.

    Science.gov (United States)

    Ouzounis, Georgios K; Pesaresi, Martino; Soille, Pierre

    2012-08-01

    Differential area profiles (DAPs) are point-based multiscale descriptors used in pattern analysis and image segmentation. They are defined through sets of size-based connected morphological filters that constitute a joint area opening top-hat and area closing bottom-hat scale-space of the input image. The work presented in this paper explores the properties of this image decomposition through sets of area zones. An area zone defines a single plane of the DAP vector field and contains all the peak components of the input image, whose size is between the zone's attribute extrema. Area zones can be computed efficiently from hierarchical image representation structures, in a way similar to regular attribute filters. Operations on the DAP vector field can then be computed without the need for exporting it first, and an example with the leveling-like convex/concave segmentation scheme is given. This is referred to as the one-pass method and it is demonstrated on the Max-Tree structure. Its computational performance is tested and compared against conventional means for computing differential profiles, relying on iterative application of area openings and closings. Applications making use of the area zone decomposition are demonstrated in problems related to remote sensing and medical image analysis.

  15. High-level language computer architecture

    CERN Document Server

    Chu, Yaohan

    1975-01-01

    High-Level Language Computer Architecture offers a tutorial on high-level language computer architecture, including von Neumann architecture and syntax-oriented architecture as well as direct and indirect execution architecture. Design concepts of Japanese-language data processing systems are discussed, along with the architecture of stack machines and the SYMBOL computer system. The conceptual design of a direct high-level language processor is also described.Comprised of seven chapters, this book first presents a classification of high-level language computer architecture according to the pr

  16. Increasing Computational Efficiency of Cochlear Models Using Boundary Layers

    Science.gov (United States)

    Alkhairy, Samiya A.; Shera, Christopher A.

    2016-01-01

    Our goal is to develop methods to improve the efficiency of computational models of the cochlea for applications that require the solution accurately only within a basal region of interest, specifically by decreasing the number of spatial sections needed for simulation of the problem with good accuracy. We design algebraic spatial and parametric transformations to computational models of the cochlea. These transformations are applied after the basal region of interest and allow for spatial preservation, driven by the natural characteristics of approximate spatial causality of cochlear models. The project is of foundational nature and hence the goal is to design, characterize and develop an understanding and framework rather than optimization and globalization. Our scope is as follows: designing the transformations; understanding the mechanisms by which computational load is decreased for each transformation; development of performance criteria; characterization of the results of applying each transformation to a specific physical model and discretization and solution schemes. In this manuscript, we introduce one of the proposed methods (complex spatial transformation) for a case study physical model that is a linear, passive, transmission line model in which the various abstraction layers (electric parameters, filter parameters, wave parameters) are clearer than other models. This is conducted in the frequency domain for multiple frequencies using a second order finite difference scheme for discretization and direct elimination for solving the discrete system of equations. The performance is evaluated using two developed simulative criteria for each of the transformations. In conclusion, the developed methods serve to increase efficiency of a computational traveling wave cochlear model when spatial preservation can hold, while maintaining good correspondence with the solution of interest and good accuracy, for applications in which the interest is in the solution

  17. Increasing computational efficiency of cochlear models using boundary layers

    Science.gov (United States)

    Alkhairy, Samiya A.; Shera, Christopher A.

    2015-12-01

    Our goal is to develop methods to improve the efficiency of computational models of the cochlea for applications that require the solution accurately only within a basal region of interest, specifically by decreasing the number of spatial sections needed for simulation of the problem with good accuracy. We design algebraic spatial and parametric transformations to computational models of the cochlea. These transformations are applied after the basal region of interest and allow for spatial preservation, driven by the natural characteristics of approximate spatial causality of cochlear models. The project is of foundational nature and hence the goal is to design, characterize and develop an understanding and framework rather than optimization and globalization. Our scope is as follows: designing the transformations; understanding the mechanisms by which computational load is decreased for each transformation; development of performance criteria; characterization of the results of applying each transformation to a specific physical model and discretization and solution schemes. In this manuscript, we introduce one of the proposed methods (complex spatial transformation) for a case study physical model that is a linear, passive, transmission line model in which the various abstraction layers (electric parameters, filter parameters, wave parameters) are clearer than other models. This is conducted in the frequency domain for multiple frequencies using a second order finite difference scheme for discretization and direct elimination for solving the discrete system of equations. The performance is evaluated using two developed simulative criteria for each of the transformations. In conclusion, the developed methods serve to increase efficiency of a computational traveling wave cochlear model when spatial preservation can hold, while maintaining good correspondence with the solution of interest and good accuracy, for applications in which the interest is in the solution

  18. Efficient Use of Preisach Hysteresis Model in Computer Aided Design

    Directory of Open Access Journals (Sweden)

    IONITA, V.

    2013-05-01

    Full Text Available The paper presents a practical detailed analysis regarding the use of the classical Preisach hysteresis model, covering all the steps, from measuring the necessary data for the model identification to the implementation in a software code for Computer Aided Design (CAD in Electrical Engineering. An efficient numerical method is proposed and the hysteresis modeling accuracy is tested on magnetic recording materials. The procedure includes the correction of the experimental data, which are used for the hysteresis model identification, taking into account the demagnetizing effect for the sample that is measured in an open-circuit device (a vibrating sample magnetometer.

  19. Efficient quantum algorithm for computing n-time correlation functions.

    Science.gov (United States)

    Pedernales, J S; Di Candia, R; Egusquiza, I L; Casanova, J; Solano, E

    2014-07-11

    We propose a method for computing n-time correlation functions of arbitrary spinorial, fermionic, and bosonic operators, consisting of an efficient quantum algorithm that encodes these correlations in an initially added ancillary qubit for probe and control tasks. For spinorial and fermionic systems, the reconstruction of arbitrary n-time correlation functions requires the measurement of two ancilla observables, while for bosonic variables time derivatives of the same observables are needed. Finally, we provide examples applicable to different quantum platforms in the frame of the linear response theory.

  20. Using Weighted Graphs for Computationally Efficient WLAN Location Determination

    DEFF Research Database (Denmark)

    Thomsen, Bent; Hansen, Rene

    2007-01-01

    Indoor location-based services hold promise for a multitude of valuable services, but require micro-detailed geo-referencing not achievable with "outdoor" technologies such as GPS and cellular networks. A widely used technique for accurate indoor positioning is location fingerprinting which makes...... burden for large buildings and is thus problematic for tracking users in real time on processor-constrained mobile devices. In this paper we present a technique for improving the computational efficiency of the fingerprinting technique such that location determination becomes tractable on a mobile device...

  1. Scientific and high-performance computing at FAIR

    Directory of Open Access Journals (Sweden)

    Kisel Ivan

    2015-01-01

    Full Text Available Future FAIR experiments have to deal with very high input rates, large track multiplicities, make full event reconstruction and selection on-line on a large dedicated computer farm equipped with heterogeneous many-core CPU/GPU compute nodes. To develop efficient and fast algorithms, which are optimized for parallel computations, is a challenge for the groups of experts dealing with the HPC computing. Here we present and discuss the status and perspectives of the data reconstruction and physics analysis software of one of the future FAIR experiments, namely, the CBM experiment.

  2. IMPROVING TACONITE PROCESSING PLANT EFFICIENCY BY COMPUTER SIMULATION, Final Report

    Energy Technology Data Exchange (ETDEWEB)

    William M. Bond; Salih Ersayin

    2007-03-30

    This project involved industrial scale testing of a mineral processing simulator to improve the efficiency of a taconite processing plant, namely the Minorca mine. The Concentrator Modeling Center at the Coleraine Minerals Research Laboratory, University of Minnesota Duluth, enhanced the capabilities of available software, Usim Pac, by developing mathematical models needed for accurate simulation of taconite plants. This project provided funding for this technology to prove itself in the industrial environment. As the first step, data representing existing plant conditions were collected by sampling and sample analysis. Data were then balanced and provided a basis for assessing the efficiency of individual devices and the plant, and also for performing simulations aimed at improving plant efficiency. Performance evaluation served as a guide in developing alternative process strategies for more efficient production. A large number of computer simulations were then performed to quantify the benefits and effects of implementing these alternative schemes. Modification of makeup ball size was selected as the most feasible option for the target performance improvement. This was combined with replacement of existing hydrocyclones with more efficient ones. After plant implementation of these modifications, plant sampling surveys were carried out to validate findings of the simulation-based study. Plant data showed very good agreement with the simulated data, confirming results of simulation. After the implementation of modifications in the plant, several upstream bottlenecks became visible. Despite these bottlenecks limiting full capacity, concentrator energy improvement of 7% was obtained. Further improvements in energy efficiency are expected in the near future. The success of this project demonstrated the feasibility of a simulation-based approach. Currently, the Center provides simulation-based service to all the iron ore mining companies operating in northern

  3. High efficient solar tracker based on a simple shutter structure

    Science.gov (United States)

    Chen, Jin-Jia; Liu, Te-Shu; Huang, Kuang-Lung; Lin, Po-Chih

    2013-09-01

    In many photovoltaic (PV) or sunlight-illumination systems, solar trackers are always essential to obtain high energy/flux concentration efficiency, and that would lead to increase cost and extra power consumption due to the complex structure and heavy weight of the trackers. To decrease the cost while without sacrificing efficiency, a Fresnellens concentrator incorporated with a simple and cheap shutter, which consists of high reflective mirrors instead of conventional trackers, is proposed in this paper to provide solar tracking during the daytime. Thus, the time-variant and slant-incident sunlight rays can be redirected to vertically incident upon the surface of the Fresnel lens by appropriately arranging mirrors and swinging them to the proper slant angles with respect to the orientation of sunlight. The computer simulation results show that power concentration efficiency over 90%, as compared with the efficiency of directly normal incident sunlight, can be achieved with the mirror reflectance of 0.97 and for any solar incident angle within +/-75 degrees to the normal of the Fresnel lens. To verify the feasibility and performance of the concentrator with the proposed shutter, a sunlight illumination system based on this novel structure is demonstrated. Both computer simulation and practical measurement results for the prototype of the sunlight illumination system are also given to compare with. The results prove the simple and high efficient shutter applicable to general PV or sunlight-illumination systems for solar tracking.

  4. Efficient Computation of Transition State Resonances and Reaction Rates from a Quantum Normal Form

    NARCIS (Netherlands)

    Schubert, Roman; Waalkens, Holger; Wiggins, Stephen

    2006-01-01

    A quantum version of a recent formulation of transition state theory in phase space is presented. The theory developed provides an algorithm to compute quantum reaction rates and the associated Gamov-Siegert resonances with very high accuracy. The algorithm is especially efficient for multi-degree-o

  5. A high performance scientific cloud computing environment for materials simulations

    CERN Document Server

    Jorissen, Kevin; Rehr, John J

    2011-01-01

    We describe the development of a scientific cloud computing (SCC) platform that offers high performance computation capability. The platform consists of a scientific virtual machine prototype containing a UNIX operating system and several materials science codes, together with essential interface tools (an SCC toolset) that offers functionality comparable to local compute clusters. In particular, our SCC toolset provides automatic creation of virtual clusters for parallel computing, including tools for execution and monitoring performance, as well as efficient I/O utilities that enable seamless connections to and from the cloud. Our SCC platform is optimized for the Amazon Elastic Compute Cloud (EC2). We present benchmarks for prototypical scientific applications and demonstrate performance comparable to local compute clusters. To facilitate code execution and provide user-friendly access, we have also integrated cloud computing capability in a JAVA-based GUI. Our SCC platform may be an alternative to traditi...

  6. A Computationally Efficient Parallel Levenberg-Marquardt Algorithm for Large-Scale Big-Data Inversion

    Science.gov (United States)

    Lin, Y.; O'Malley, D.; Vesselinov, V. V.

    2015-12-01

    Inverse modeling seeks model parameters given a set of observed state variables. However, for many practical problems due to the facts that the observed data sets are often large and model parameters are often numerous, conventional methods for solving the inverse modeling can be computationally expensive. We have developed a new, computationally-efficient Levenberg-Marquardt method for solving large-scale inverse modeling. Levenberg-Marquardt methods require the solution of a dense linear system of equations which can be prohibitively expensive to compute for large-scale inverse problems. Our novel method projects the original large-scale linear problem down to a Krylov subspace, such that the dimensionality of the measurements can be significantly reduced. Furthermore, instead of solving the linear system for every Levenberg-Marquardt damping parameter, we store the Krylov subspace computed when solving the first damping parameter and recycle it for all the following damping parameters. The efficiency of our new inverse modeling algorithm is significantly improved by using these computational techniques. We apply this new inverse modeling method to invert for a random transitivity field. Our algorithm is fast enough to solve for the distributed model parameters (transitivity) at each computational node in the model domain. The inversion is also aided by the use regularization techniques. The algorithm is coded in Julia and implemented in the MADS computational framework (http://mads.lanl.gov). Julia is an advanced high-level scientific programing language that allows for efficient memory management and utilization of high-performance computational resources. By comparing with a Levenberg-Marquardt method using standard linear inversion techniques, our Levenberg-Marquardt method yields speed-up ratio of 15 in a multi-core computational environment and a speed-up ratio of 45 in a single-core computational environment. Therefore, our new inverse modeling method is a

  7. High power, high efficiency millimeter wavelength traveling wave tubes for high rate communications from deep space

    Science.gov (United States)

    Dayton, James A., Jr.

    1991-01-01

    The high-power transmitters needed for high data rate communications from deep space will require a new class of compact, high efficiency traveling wave tubes (TWT's). Many of the recent TWT developments in the microwave frequency range are generically applicable to mm wave devices, in particular much of the technology of computer aided design, cathodes, and multistage depressed collectors. However, because TWT dimensions scale approximately with wavelength, mm wave devices will be physically much smaller with inherently more stringent fabrication tolerances and sensitivity to thermal dissipation.

  8. Co-operative Scheduled Energy Aware Load-Balancing technique for an Efficient Computational Cloud

    Directory of Open Access Journals (Sweden)

    T R V Anandharajan

    2011-03-01

    Full Text Available Cloud Computing in the recent years has been taking its evolution from the scientific to the non scientific and commercial applications. Power consumption and Load balancing are very important and complex problem in computational Cloud. A computational Cloud differs from traditional high-performance computing systems in the heterogeneity of the computing nodes, as well as the communication links that connect the different nodes together. Load Balancing is a very important component in the commodity services based cloud computing. There is a need to develop algorithms that can capture this complexity yet can be easily implemented and used to solve a wide range of load-balancing scenarios in a Data and Computing intensive applications. In this paper, we propose to find the best EFFICIENT cloud resource by Co-operative Power aware Scheduled Load Balancing solution to the Cloud load-balancing problem. The algorithm developed combines the inherent efficiency of the centralized approach, energy efficient and the fault-tolerant nature of the distributed environment like Cloud.

  9. Highly efficient, gearless drive; Hocheffizienter, getriebeloser Antrieb

    Energy Technology Data Exchange (ETDEWEB)

    Niederer, R.

    2004-07-01

    Highly efficient, gearless variable-speed drive systems for low-speed applications have been developed. These systems consist of an inverter with active switches (IGBTs, MOSFETs, resp.) and a synchronous machine excited with permanent magnets. Therefore, these systems can be used for drive as well as for generator applications. They operate very efficiently since a gearbox is obsolete, furthermore weight, dimensions, noise and maintenance can be reduced. The inverter controllers do not require any speed sensors, thus reliability is increased and costs are decreased. Application for low-speed variable-speed drive systems can be found in industrial applications, cable railways or wind turbines. Both systems have been optimized in several iterative loops, in what regards overall efficiency and material expenditure. For both systems, prototypes have been developed and tested. Both prototypes performed reliably and fulfilled the expectations. The high power system (1200 kW, 20 rpm) operated at rated load with an overall efficiency of 93.1%, the lower power system (3 kW, 60 rpm) with an overall efficiency of 85%. Thus the losses of these new systems are at rated load about 4% lower compared to conventional drive systems equipped with a mechanical gearbox. (author)

  10. Computationally efficient sub-band coding of ECG signals.

    Science.gov (United States)

    Husøy, J H; Gjerde, T

    1996-03-01

    A data compression technique is presented for the compression of discrete time electrocardiogram (ECG) signals. The compression system is based on sub-band coding, a technique traditionally used for compressing speech and images. The sub-band coder employs quadrature mirror filter banks (QMF) with up to 32 critically sampled sub-bands. Both finite impulse response (FIR) and the more computationally efficient infinite impulse response (IIR) filter banks are considered as candidates in a complete ECG coding system. The sub-bands are threshold, quantized using uniform quantizers and run-length coded. The output of the run-length coder is further compressed by a Huffman coder. Extensive simulations indicate that 16 sub-bands are a suitable choice for this application. Furthermore, IIR filter banks are preferable due to their superiority in terms of computational efficiency. We conclude that the present scheme, which is suitable for real time implementation on a PC, can provide compression ratios between 5 and 15 without loss of clinical information.

  11. Statistically and Computationally Efficient Estimating Equations for Large Spatial Datasets

    KAUST Repository

    Sun, Ying

    2014-11-07

    For Gaussian process models, likelihood based methods are often difficult to use with large irregularly spaced spatial datasets, because exact calculations of the likelihood for n observations require O(n3) operations and O(n2) memory. Various approximation methods have been developed to address the computational difficulties. In this paper, we propose new unbiased estimating equations based on score equation approximations that are both computationally and statistically efficient. We replace the inverse covariance matrix that appears in the score equations by a sparse matrix to approximate the quadratic forms, then set the resulting quadratic forms equal to their expected values to obtain unbiased estimating equations. The sparse matrix is constructed by a sparse inverse Cholesky approach to approximate the inverse covariance matrix. The statistical efficiency of the resulting unbiased estimating equations are evaluated both in theory and by numerical studies. Our methods are applied to nearly 90,000 satellite-based measurements of water vapor levels over a region in the Southeast Pacific Ocean.

  12. Introduction to High Performance Scientific Computing

    OpenAIRE

    2016-01-01

    The field of high performance scientific computing lies at the crossroads of a number of disciplines and skill sets, and correspondingly, for someone to be successful at using high performance computing in science requires at least elementary knowledge of and skills in all these areas. Computations stem from an application context, so some acquaintance with physics and engineering sciences is desirable. Then, problems in these application areas are typically translated into linear algebraic, ...

  13. Resource estimation in high performance medical image computing.

    Science.gov (United States)

    Banalagay, Rueben; Covington, Kelsie Jade; Wilkes, D M; Landman, Bennett A

    2014-10-01

    Medical imaging analysis processes often involve the concatenation of many steps (e.g., multi-stage scripts) to integrate and realize advancements from image acquisition, image processing, and computational analysis. With the dramatic increase in data size for medical imaging studies (e.g., improved resolution, higher throughput acquisition, shared databases), interesting study designs are becoming intractable or impractical on individual workstations and servers. Modern pipeline environments provide control structures to distribute computational load in high performance computing (HPC) environments. However, high performance computing environments are often shared resources, and scheduling computation across these resources necessitates higher level modeling of resource utilization. Submission of 'jobs' requires an estimate of the CPU runtime and memory usage. The resource requirements for medical image processing algorithms are difficult to predict since the requirements can vary greatly between different machines, different execution instances, and different data inputs. Poor resource estimates can lead to wasted resources in high performance environments due to incomplete executions and extended queue wait times. Hence, resource estimation is becoming a major hurdle for medical image processing algorithms to efficiently leverage high performance computing environments. Herein, we present our implementation of a resource estimation system to overcome these difficulties and ultimately provide users with the ability to more efficiently utilize high performance computing resources.

  14. Scalable resource management in high performance computers.

    Energy Technology Data Exchange (ETDEWEB)

    Frachtenberg, E. (Eitan); Petrini, F. (Fabrizio); Fernandez Peinador, J. (Juan); Coll, S. (Salvador)

    2002-01-01

    Clusters of workstations have emerged as an important platform for building cost-effective, scalable and highly-available computers. Although many hardware solutions are available today, the largest challenge in making large-scale clusters usable lies in the system software. In this paper we present STORM, a resource management tool designed to provide scalability, low overhead and the flexibility necessary to efficiently support and analyze a wide range of job scheduling algorithms. STORM achieves these feats by closely integrating the management daemons with the low-level features that are common in state-of-the-art high-performance system area networks. The architecture of STORM is based on three main technical innovations. First, a sizable part of the scheduler runs in the thread processor located on the network interface. Second, we use hardware collectives that are highly scalable both for implementing control heartbeats and to distribute the binary of a parallel job in near-constant time, irrespective of job and machine sizes. Third, we use an I/O bypass protocol that allows fast data movements from the file system to the communication buffers in the network interface and vice versa. The experimental results show that STORM can launch a job with a binary of 12MB on a 64 processor/32 node cluster in less than 0.25 sec on an empty network, in less than 0.45 sec when all the processors are busy computing other jobs, and in less than 0.65 sec when the network is flooded with a background traffic. This paper provides experimental and analytical evidence that these results scale to a much larger number of nodes. To the best of our knowledge, STORM is at least two orders of magnitude faster than existing production schedulers in launching jobs, performing resource management tasks and gang scheduling.

  15. High-efficiency design of a mixed-flow pump

    Institute of Scientific and Technical Information of China (English)

    KIM; Jin-Hyuk; AHN; Hyung-Jin; KIM; Kwang-Yong

    2010-01-01

    High-efficiency design of a mixed-flow pump has been carried out based on numerical analysis of a three-dimensional viscous flow.For analysis,the Reynolds-averaged Navier-Stokes equations with a shear stress transport turbulence model were discretized by finite-volume approximations.Structured grid system was constructed in the computational domain,which has O-type grids near the blade surfaces and H/J-type grids in other regions.The numerical results were validated with experimental data for the heads and hydraulic efficiencies at different flow coefficients.The hydraulic efficiency at the design flow coefficient was evaluated with variation of the geometric variables,i.e.,the area of the discharge and length of the vane in the diffuser.The result has shown that the hydraulic efficiency of a mixed-flow pump at the design condition is improved by the modification of the geometry.

  16. China's High Performance Computer Standard Commission Established

    Institute of Scientific and Technical Information of China (English)

    2007-01-01

    @@ China's High Performance Computer Standard Commission was established on March 28, 2007, under the guidance of the Science and Technology Bureau of the Ministry of Information Industry. It will prepare relevant professional standards on high performance computers to break through the monopoly in the field by foreign manufacturers and vendors.

  17. Highly efficient charged particle veto detector CUP

    Energy Technology Data Exchange (ETDEWEB)

    Palacz, M. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland)]. E-mail: palacz@slcj.uw.edu.pl; Nyberg, J. [Department of Radiation Sciences, Uppsala University, Uppsala (Sweden); Bednarczyk, P. [Institute de Recherches Subatomiques, Strasbourg (France); Henryk Niewodniczanski Institute of Nuclear Physics, Polish Academy of Sciences, Cracow (Poland); Dworski, J. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland); Gorska, M. [Gesellschaft fuer Schwerionenforschung, Darmstadt (Germany); Iwanicki, J. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland); Kapusta, M. [Soltan Institute for Nuclear Studies, Swierk (Poland); Kownacki, J. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland); Kulczycka, E. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland); Lagergren, K. [Royal Institute of Technology, Stockholm (Sweden); Moszynski, M. [Soltan Institute for Nuclear Studies, Swierk (Poland); Pienkowski, L. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland); Stolarz, A. [Heavy Ion Laboratory, Warsaw University, ul. Pasteura 5A, PL 02-093 Warsaw (Poland); Wolski, D. [Soltan Institute for Nuclear Studies, Swierk (Poland); Zieblinski, M. [Henryk Niewodniczanski Institute of Nuclear Physics, Polish Academy of Sciences, Cracow (Poland)

    2005-09-11

    A novel, highly efficient, plastic scintillator detector has been constructed. The primary application of the detector is to act as a veto device in heavy-ion-induced fusion-evaporation reactions, in which the structure of proton-rich nuclides is investigated by {gamma}-ray spectroscopy methods. The detector rejects events in which light charged particles, like protons and {alpha} particles, are emitted in the evaporation process, facilitating selection of reaction channels associated with emission of only neutrons. The detector was used in a EUROBALL experiment, with achieved efficiencies of 80% and 63% for protons and {alpha} particles, respectively. The design of the detector, its performance and limitations are discussed.

  18. High Efficiency Solar Integrated Roof Membrane Product

    Energy Technology Data Exchange (ETDEWEB)

    Partyka, Eric; Shenoy, Anil

    2013-05-15

    This project was designed to address the Solar Energy Technology Program objective, to develop new methods to integrate photovoltaic (PV) cells or modules within a building-integrated photovoltaic (BIPV) application that will result in lower installed cost as well as higher efficiencies of the encapsulated/embedded PV module. The technology assessment and development focused on the evaluation and identification of manufacturing technologies and equipment capable of producing such low-cost, high-efficiency, flexible BIPV solar cells on single-ply roofing membranes.

  19. High-efficiency electrical charger for nanoparticles

    Energy Technology Data Exchange (ETDEWEB)

    Alonso, M., E-mail: malonso@cenim.csic.es [National Centre for Metallurgical Research (CENIM-CSIC) (Spain); Huang, C. H. [Yuanpei University, Department of Environmental Engineering and Health (China)

    2015-08-15

    An electrical charger, based on a point-to-plate DC corona discharge, for the high-efficiency charging of aerosol particles with diameter of a few nanometers, has been designed, constructed, and evaluated. The discharge takes place between a needle and a perforated plate, and the results presented here have shown that this specific design allows reduction of electrostatic losses of charged particles within the charger in comparison with other typical designs. Besides, the small effective volume of the charger leads to a relatively small diffusion loss of particles. As a consequence of the reduced electrostatic and diffusion losses, the extrinsic charging efficiency attainable is higher than in similar devices.

  20. Efficient parameter sensitivity computation for spatially extended reaction networks

    Science.gov (United States)

    Lester, C.; Yates, C. A.; Baker, R. E.

    2017-01-01

    Reaction-diffusion models are widely used to study spatially extended chemical reaction systems. In order to understand how the dynamics of a reaction-diffusion model are affected by changes in its input parameters, efficient methods for computing parametric sensitivities are required. In this work, we focus on the stochastic models of spatially extended chemical reaction systems that involve partitioning the computational domain into voxels. Parametric sensitivities are often calculated using Monte Carlo techniques that are typically computationally expensive; however, variance reduction techniques can decrease the number of Monte Carlo simulations required. By exploiting the characteristic dynamics of spatially extended reaction networks, we are able to adapt existing finite difference schemes to robustly estimate parametric sensitivities in a spatially extended network. We show that algorithmic performance depends on the dynamics of the given network and the choice of summary statistics. We then describe a hybrid technique that dynamically chooses the most appropriate simulation method for the network of interest. Our method is tested for functionality and accuracy in a range of different scenarios.

  1. High Efficiency Reversible Fuel Cell Power Converter

    DEFF Research Database (Denmark)

    Pittini, Riccardo

    The large scale integration of renewable energy sources requires suitable energy storage systems to balance energy production and demand in the electrical grid. Bidirectional fuel cells are an attractive technology for energy storage systems due to the high energy density of fuel. Compared...... entitled "High Efficiency Reversible Fuel Cell Power Converter" and it presents the design of a high efficiency dc-dc converter developed and optimized for bidirectional fuel cell applications. First, a brief overview of fuel cell and energy storage technologies is presented. Different system topologies...... to traditional unidirectional fuel cell, bidirectional fuel cells have increased operating voltage and current ranges. These characteristics increase the stresses on dc-dc and dc-ac converters in the electrical system, which require proper design and advanced optimization. This work is part of the PhD project...

  2. A computationally efficient method for hand-eye calibration.

    Science.gov (United States)

    Zhang, Zhiqiang; Zhang, Lin; Yang, Guang-Zhong

    2017-07-19

    Surgical robots with cooperative control and semiautonomous features have shown increasing clinical potential, particularly for repetitive tasks under imaging and vision guidance. Effective performance of an autonomous task requires accurate hand-eye calibration so that the transformation between the robot coordinate frame and the camera coordinates is well defined. In practice, due to changes in surgical instruments, online hand-eye calibration must be performed regularly. In order to ensure seamless execution of the surgical procedure without affecting the normal surgical workflow, it is important to derive fast and efficient hand-eye calibration methods. We present a computationally efficient iterative method for hand-eye calibration. In this method, dual quaternion is introduced to represent the rigid transformation, and a two-step iterative method is proposed to recover the real and dual parts of the dual quaternion simultaneously, and thus the estimation of rotation and translation of the transformation. The proposed method was applied to determine the rigid transformation between the stereo laparoscope and the robot manipulator. Promising experimental and simulation results have shown significant convergence speed improvement to 3 iterations from larger than 30 with regard to standard optimization method, which illustrates the effectiveness and efficiency of the proposed method.

  3. Spin-neurons: A possible path to energy-efficient neuromorphic computers

    Energy Technology Data Exchange (ETDEWEB)

    Sharad, Mrigank; Fan, Deliang; Roy, Kaushik [School of Electrical and Computer Engineering, Purdue University, West Lafayette, Indiana 47907 (United States)

    2013-12-21

    Recent years have witnessed growing interest in the field of brain-inspired computing based on neural-network architectures. In order to translate the related algorithmic models into powerful, yet energy-efficient cognitive-computing hardware, computing-devices beyond CMOS may need to be explored. The suitability of such devices to this field of computing would strongly depend upon how closely their physical characteristics match with the essential computing primitives employed in such models. In this work, we discuss the rationale of applying emerging spin-torque devices for bio-inspired computing. Recent spin-torque experiments have shown the path to low-current, low-voltage, and high-speed magnetization switching in nano-scale magnetic devices. Such magneto-metallic, current-mode spin-torque switches can mimic the analog summing and “thresholding” operation of an artificial neuron with high energy-efficiency. Comparison with CMOS-based analog circuit-model of a neuron shows that “spin-neurons” (spin based circuit model of neurons) can achieve more than two orders of magnitude lower energy and beyond three orders of magnitude reduction in energy-delay product. The application of spin-neurons can therefore be an attractive option for neuromorphic computers of future.

  4. Spin-neurons: A possible path to energy-efficient neuromorphic computers

    Science.gov (United States)

    Sharad, Mrigank; Fan, Deliang; Roy, Kaushik

    2013-12-01

    Recent years have witnessed growing interest in the field of brain-inspired computing based on neural-network architectures. In order to translate the related algorithmic models into powerful, yet energy-efficient cognitive-computing hardware, computing-devices beyond CMOS may need to be explored. The suitability of such devices to this field of computing would strongly depend upon how closely their physical characteristics match with the essential computing primitives employed in such models. In this work, we discuss the rationale of applying emerging spin-torque devices for bio-inspired computing. Recent spin-torque experiments have shown the path to low-current, low-voltage, and high-speed magnetization switching in nano-scale magnetic devices. Such magneto-metallic, current-mode spin-torque switches can mimic the analog summing and "thresholding" operation of an artificial neuron with high energy-efficiency. Comparison with CMOS-based analog circuit-model of a neuron shows that "spin-neurons" (spin based circuit model of neurons) can achieve more than two orders of magnitude lower energy and beyond three orders of magnitude reduction in energy-delay product. The application of spin-neurons can therefore be an attractive option for neuromorphic computers of future.

  5. Methodologies for high efficiency perovskite solar cells.

    Science.gov (United States)

    Park, Nam-Gyu

    2016-01-01

    Since the report on long-term durable solid-state perovskite solar cell in 2012, perovskite solar cells based on lead halide perovskites having organic cations such as methylammonium CH3NH3PbI3 or formamidinium HC(NH2)2PbI3 have received great attention because of superb photovoltaic performance with power conversion efficiency exceeding 22 %. In this review, emergence of perovskite solar cell is briefly introduced. Since understanding fundamentals of light absorbers is directly related to their photovoltaic performance, opto-electronic properties of organo lead halide perovskites are investigated in order to provide insight into design of higher efficiency perovskite solar cells. Since the conversion efficiency of perovskite solar cell is found to depend significantly on perovskite film quality, methodologies for fabricating high quality perovskite films are particularly emphasized, including various solution-processes and vacuum deposition method.

  6. Methodologies for high efficiency perovskite solar cells

    Science.gov (United States)

    Park, Nam-Gyu

    2016-06-01

    Since the report on long-term durable solid-state perovskite solar cell in 2012, perovskite solar cells based on lead halide perovskites having organic cations such as methylammonium CH3NH3PbI3 or formamidinium HC(NH2)2PbI3 have received great attention because of superb photovoltaic performance with power conversion efficiency exceeding 22 %. In this review, emergence of perovskite solar cell is briefly introduced. Since understanding fundamentals of light absorbers is directly related to their photovoltaic performance, opto-electronic properties of organo lead halide perovskites are investigated in order to provide insight into design of higher efficiency perovskite solar cells. Since the conversion efficiency of perovskite solar cell is found to depend significantly on perovskite film quality, methodologies for fabricating high quality perovskite films are particularly emphasized, including various solution-processes and vacuum deposition method.

  7. High efficiency electrotransformation of Lactobacillus casei.

    Science.gov (United States)

    Welker, Dennis L; Hughes, Joanne E; Steele, James L; Broadbent, Jeff R

    2015-01-01

    We investigated whether protocols allowing high efficiency electrotransformation of other lactic acid bacteria were applicable to five strains of Lactobacillus casei (12A, 32G, A2-362, ATCC 334 and BL23). Addition of 1% glycine or 0.9 M NaCl during cell growth, limitation of the growth of the cell cultures to OD600 0.6-0.8, pre-electroporation treatment of cells with water or with a lithium acetate (100 mM)/dithiothreitol (10 mM) solution and optimization of electroporation conditions all improved transformation efficiencies. However, the five strains varied in their responses to these treatments. Transformation efficiencies of 10(6) colony forming units μg(-1) pTRKH2 DNA and higher were obtained with three strains which is sufficient for construction of chromosomal gene knock-outs and gene replacements.

  8. High Efficiency, Low Emission Refrigeration System

    Energy Technology Data Exchange (ETDEWEB)

    Fricke, Brian A [ORNL; Sharma, Vishaldeep [ORNL

    2016-08-01

    Supermarket refrigeration systems account for approximately 50% of supermarket energy use, placing this class of equipment among the highest energy consumers in the commercial building domain. In addition, the commonly used refrigeration system in supermarket applications is the multiplex direct expansion (DX) system, which is prone to refrigerant leaks due to its long lengths of refrigerant piping. This leakage reduces the efficiency of the system and increases the impact of the system on the environment. The high Global Warming Potential (GWP) of the hydrofluorocarbon (HFC) refrigerants commonly used in these systems, coupled with the large refrigerant charge and the high refrigerant leakage rates leads to significant direct emissions of greenhouse gases into the atmosphere. Methods for reducing refrigerant leakage and energy consumption are available, but underutilized. Further work needs to be done to reduce costs of advanced system designs to improve market utilization. In addition, refrigeration system retrofits that result in reduced energy consumption are needed since the majority of applications address retrofits rather than new stores. The retrofit market is also of most concern since it involves large-volume refrigerant systems with high leak rates. Finally, alternative refrigerants for new and retrofit applications are needed to reduce emissions and reduce the impact on the environment. The objective of this Collaborative Research and Development Agreement (CRADA) between the Oak Ridge National Laboratory and Hill Phoenix is to develop a supermarket refrigeration system that reduces greenhouse gas emissions and has 25 to 30 percent lower energy consumption than existing systems. The outcomes of this project will include the design of a low emission, high efficiency commercial refrigeration system suitable for use in current U.S. supermarkets. In addition, a prototype low emission, high efficiency supermarket refrigeration system will be produced for

  9. High-throughput computing in the sciences.

    Science.gov (United States)

    Morgan, Mark; Grimshaw, Andrew

    2009-01-01

    While it is true that the modern computer is many orders of magnitude faster than that of yesteryear; this tremendous growth in CPU clock rates is now over. Unfortunately, however, the growth in demand for computational power has not abated; whereas researchers a decade ago could simply wait for computers to get faster, today the only solution to the growing need for more powerful computational resource lies in the exploitation of parallelism. Software parallelization falls generally into two broad categories--"true parallel" and high-throughput computing. This chapter focuses on the latter of these two types of parallelism. With high-throughput computing, users can run many copies of their software at the same time across many different computers. This technique for achieving parallelism is powerful in its ability to provide high degrees of parallelism, yet simple in its conceptual implementation. This chapter covers various patterns of high-throughput computing usage and the skills and techniques necessary to take full advantage of them. By utilizing numerous examples and sample codes and scripts, we hope to provide the reader not only with a deeper understanding of the principles behind high-throughput computing, but also with a set of tools and references that will prove invaluable as she explores software parallelism with her own software applications and research.

  10. The Efficiency of Various Computers and Optimizations in Performing Finite Element Computations

    Science.gov (United States)

    Marcus, Martin H.; Broduer, Steve (Technical Monitor)

    2001-01-01

    With the advent of computers with many processors, it becomes unclear how to best exploit this advantage. For example, matrices can be inverted by applying several processors to each vector operation, or one processor can be applied to each matrix. The former approach has diminishing returns beyond a handful of processors, but how many processors depends on the computer architecture. Applying one processor to each matrix is feasible with enough ram memory and scratch disk space, but the speed at which this is done is found to vary by a factor of three depending on how it is done. The cost of the computer must also be taken into account. A computer with many processors and fast interprocessor communication is much more expensive than the same computer and processors with slow interprocessor communication. Consequently, for problems that require several matrices to be inverted, the best speed per dollar for computers is found to be several small workstations that are networked together, such as in a Beowulf cluster. Since these machines typically have two processors per node, each matrix is most efficiently inverted with no more than two processors assigned to it.

  11. Creation of High Efficient Firefly Luciferase

    Science.gov (United States)

    Nakatsu, Toru

    Firefly emits visible yellow-green light. The bioluminescence reaction is carried out by the enzyme luciferase. The bioluminescence of luciferase is widely used as an excellent tool for monitoring gene expression, the measurement of the amount of ATP and in vivo imaging. Recently a study of the cancer metastasis is carried out by in vivo luminescence imaging system, because luminescence imaging is less toxic and more useful for long-term assay than fluorescence imaging by GFP. However the luminescence is much dimmer than fluorescence. Then bioluminescence imaging in living organisms demands the high efficient luciferase which emits near infrared lights or enhances the emission intensity. Here I introduce an idea for creating the high efficient luciferase based on the crystal structure.

  12. High Efficiency Thermoelectric Materials and Devices

    Science.gov (United States)

    Kochergin, Vladimir (Inventor)

    2013-01-01

    Growth of thermoelectric materials in the form of quantum well super-lattices on three-dimensionally structured substrates provide the means to achieve high conversion efficiency of the thermoelectric module combined with inexpensiveness of fabrication and compatibility with large scale production. Thermoelectric devices utilizing thermoelectric materials in the form of quantum well semiconductor super-lattices grown on three-dimensionally structured substrates provide improved thermoelectric characteristics that can be used for power generation, cooling and other applications..

  13. Clean and Highly Efficient Utilization of Coal

    Institute of Scientific and Technical Information of China (English)

    WANG Jianguo; YANG Li

    2011-01-01

    @@ Clean and highly efficient utilization of coal is an important scientific and technological issue.As the petroleum resource decreases but its consumption increases, all of the countries in the world have to face the big issue of sustainable development of energy and economy and protection of environment.Therefore, study on clean coal technology (CCT) has attracted much attention and become one of important themes of energy research.

  14. High Efficiency Power Converter for Low Voltage High Power Applications

    DEFF Research Database (Denmark)

    Nymand, Morten

    , and remote power generation for light towers, camper vans, boats, beacons, and buoys etc. A review of current state-of-the-art is presented. The best performing converters achieve moderately high peak efficiencies at high input voltage and medium power level. However, system dimensioning and cost are often......The topic of this thesis is the design of high efficiency power electronic dc-to-dc converters for high-power, low-input-voltage to high-output-voltage applications. These converters are increasingly required for emerging sustainable energy systems such as fuel cell, battery or photo voltaic based...... determined by the performance at the system worst case operating point which is usually at minimum input voltage and maximum power. Except for the non-regulating V6 converters, all published solutions exhibit a very significant drop in conversion efficiency at minimum input voltage and maximum output power...

  15. NASA High-End Computing Program Website

    Science.gov (United States)

    Cohen, Jarrett S.

    2008-01-01

    If you are a NASA-sponsored scientist or engineer. computing time is available to you at the High-End Computing (HEC) Program's NASA Advanced Supercomputing (NAS) Facility and NASA Center for Computational Sciences (NCCS). The Science Mission Directorate will select from requests NCCS Portals submitted to the e-Books online system for awards beginning on May 1. Current projects set to explore on April 30 must have a request in e-Books to be considered for renewal

  16. NASA High-End Computing Program Website

    Science.gov (United States)

    Cohen, Jarrett S.

    2008-01-01

    If you are a NASA-sponsored scientist or engineer. computing time is available to you at the High-End Computing (HEC) Program's NASA Advanced Supercomputing (NAS) Facility and NASA Center for Computational Sciences (NCCS). The Science Mission Directorate will select from requests NCCS Portals submitted to the e-Books online system for awards beginning on May 1. Current projects set to explore on April 30 must have a request in e-Books to be considered for renewal

  17. Bioblendstocks that Enable High Efficiency Engine Designs

    Energy Technology Data Exchange (ETDEWEB)

    McCormick, Robert L.; Fioroni, Gina M.; Ratcliff, Matthew A.; Zigler, Bradley T.; Farrell, John

    2016-11-03

    The past decade has seen a high level of innovation in production of biofuels from sugar, lipid, and lignocellulose feedstocks. As discussed in several talks at this workshop, ethanol blends in the E25 to E50 range could enable more highly efficient spark-ignited (SI) engines. This is because of their knock resistance properties that include not only high research octane number (RON), but also charge cooling from high heat of vaporization, and high flame speed. Emerging alcohol fuels such as isobutanol or mixed alcohols have desirable properties such as reduced gasoline blend vapor pressure, but also have lower RON than ethanol. These fuels may be able to achieve the same knock resistance benefits, but likely will require higher blend levels or higher RON hydrocarbon blendstocks. A group of very high RON (>150) oxygenates such as dimethyl furan, methyl anisole, and related compounds are also produced from biomass. While providing no increase in charge cooling, their very high octane numbers may provide adequate knock resistance for future highly efficient SI engines. Given this range of options for highly knock resistant fuels there appears to be a critical need for a fuel knock resistance metric that includes effects of octane number, heat of vaporization, and potentially flame speed. Emerging diesel fuels include highly branched long-chain alkanes from hydroprocessing of fats and oils, as well as sugar-derived terpenoids. These have relatively high cetane number (CN), which may have some benefits in designing more efficient CI engines. Fast pyrolysis of biomass can produce diesel boiling range streams that are high in aromatic, oxygen and acid contents. Hydroprocessing can be applied to remove oxygen and consequently reduce acidity, however there are strong economic incentives to leave up to 2 wt% oxygen in the product. This oxygen will primarily be present as low CN alkyl phenols and aryl ethers. While these have high heating value, their presence in diesel fuel

  18. Developing a computationally efficient dynamic multilevel hybrid optimization scheme using multifidelity model interactions.

    Energy Technology Data Exchange (ETDEWEB)

    Hough, Patricia Diane (Sandia National Laboratories, Livermore, CA); Gray, Genetha Anne (Sandia National Laboratories, Livermore, CA); Castro, Joseph Pete Jr. (; .); Giunta, Anthony Andrew

    2006-01-01

    Many engineering application problems use optimization algorithms in conjunction with numerical simulators to search for solutions. The formulation of relevant objective functions and constraints dictate possible optimization algorithms. Often, a gradient based approach is not possible since objective functions and constraints can be nonlinear, nonconvex, non-differentiable, or even discontinuous and the simulations involved can be computationally expensive. Moreover, computational efficiency and accuracy are desirable and also influence the choice of solution method. With the advent and increasing availability of massively parallel computers, computational speed has increased tremendously. Unfortunately, the numerical and model complexities of many problems still demand significant computational resources. Moreover, in optimization, these expenses can be a limiting factor since obtaining solutions often requires the completion of numerous computationally intensive simulations. Therefore, we propose a multifidelity optimization algorithm (MFO) designed to improve the computational efficiency of an optimization method for a wide range of applications. In developing the MFO algorithm, we take advantage of the interactions between multi fidelity models to develop a dynamic and computational time saving optimization algorithm. First, a direct search method is applied to the high fidelity model over a reduced design space. In conjunction with this search, a specialized oracle is employed to map the design space of this high fidelity model to that of a computationally cheaper low fidelity model using space mapping techniques. Then, in the low fidelity space, an optimum is obtained using gradient or non-gradient based optimization, and it is mapped back to the high fidelity space. In this paper, we describe the theory and implementation details of our MFO algorithm. We also demonstrate our MFO method on some example problems and on two applications: earth penetrators and

  19. NaI(Tl Detector Efficiency Computation Using Radioactive Parallelepiped Sources Based on Efficiency Transfer Principle

    Directory of Open Access Journals (Sweden)

    Mohamed S. Badawi

    2015-01-01

    Full Text Available The efficiency transfer (ET principle is considered as a simple numerical simulation method, which can be used to calculate the full-energy peak efficiency (FEPE of 3″×3″ NaI(Tl scintillation detector over a wide energy range. In this work, the calculations of FEPE are based on computing the effective solid angle ratio between a radioactive point and parallelepiped sources located at various distances from the detector surface. Besides, the attenuation of the photon by the source-to-detector system (detector material, detector end cap, and holder material was considered and determined. This method is straightforwardly useful in setting up the efficiency calibration curve for NaI(Tl scintillation detector, when no calibration sources exist in volume shape. The values of the efficiency calculations using theoretical method are compared with the measured ones and the results show that the discrepancies in general for all the measurements are found to be less than 6%.

  20. Highly efficient fully transparent inverted OLEDs

    Science.gov (United States)

    Meyer, J.; Winkler, T.; Hamwi, S.; Schmale, S.; Kröger, M.; Görrn, P.; Johannes, H.-H.; Riedl, T.; Lang, E.; Becker, D.; Dobbertin, T.; Kowalsky, W.

    2007-09-01

    One of the unique selling propositions of OLEDs is their potential to realize highly transparent devices over the visible spectrum. This is because organic semiconductors provide a large Stokes-Shift and low intrinsic absorption losses. Hence, new areas of applications for displays and ambient lighting become accessible, for instance, the integration of OLEDs into the windshield or the ceiling of automobiles. The main challenge in the realization of fully transparent devices is the deposition of the top electrode. ITO is commonly used as transparent bottom anode in a conventional OLED. To obtain uniform light emission over the entire viewing angle and a low series resistance, a TCO such as ITO is desirable as top contact as well. However, sputter deposition of ITO on top of organic layers causes damage induced by high energetic particles and UV radiation. We have found an efficient process to protect the organic layers against the ITO rf magnetron deposition process of ITO for an inverted OLED (IOLED). The inverted structure allows the integration of OLEDs in more powerful n-channel transistors used in active matrix backplanes. Employing the green electrophosphorescent material Ir(ppy) 3 lead to IOLED with a current efficiency of 50 cd/A and power efficiency of 24 lm/W at 100 cd/m2. The average transmittance exceeds 80 % in the visible region. The on-set voltage for light emission is lower than 3 V. In addition, by vertical stacking we achieved a very high current efficiency of more than 70 cd/A for transparent IOLED.

  1. Debugging a high performance computing program

    Science.gov (United States)

    Gooding, Thomas M.

    2013-08-20

    Methods, apparatus, and computer program products are disclosed for debugging a high performance computing program by gathering lists of addresses of calling instructions for a plurality of threads of execution of the program, assigning the threads to groups in dependence upon the addresses, and displaying the groups to identify defective threads.

  2. A New Stochastic Computing Methodology for Efficient Neural Network Implementation.

    Science.gov (United States)

    Canals, Vincent; Morro, Antoni; Oliver, Antoni; Alomar, Miquel L; Rosselló, Josep L

    2016-03-01

    This paper presents a new methodology for the hardware implementation of neural networks (NNs) based on probabilistic laws. The proposed encoding scheme circumvents the limitations of classical stochastic computing (based on unipolar or bipolar encoding) extending the representation range to any real number using the ratio of two bipolar-encoded pulsed signals. Furthermore, the novel approach presents practically a total noise-immunity capability due to its specific codification. We introduce different designs for building the fundamental blocks needed to implement NNs. The validity of the present approach is demonstrated through a regression and a pattern recognition task. The low cost of the methodology in terms of hardware, along with its capacity to implement complex mathematical functions (such as the hyperbolic tangent), allows its use for building highly reliable systems and parallel computing.

  3. High Efficiency Colloidal Quantum Dot Phosphors

    Energy Technology Data Exchange (ETDEWEB)

    Kahen, Keith

    2013-12-31

    The project showed that non-Cd containing, InP-based nanocrystals (semiconductor materials with dimensions of ~6 nm) have high potential for enabling next-generation, nanocrystal-based, on chip phosphors for solid state lighting. Typical nanocrystals fall short of the requirements for on chip phosphors due to their loss of quantum efficiency under the operating conditions of LEDs, such as, high temperature (up to 150 °C) and high optical flux (up to 200 W/cm2). The InP-based nanocrystals invented during this project maintain high quantum efficiency (>80%) in polymer-based films under these operating conditions for emission wavelengths ranging from ~530 to 620 nm. These nanocrystals also show other desirable attributes, such as, lack of blinking (a common problem with nanocrystals which limits their performance) and no increase in the emission spectral width from room to 150 °C (emitters with narrower spectral widths enable higher efficiency LEDs). Prior to these nanocrystals, no nanocrystal system (regardless of nanocrystal type) showed this collection of properties; in fact, other nanocrystal systems are typically limited to showing only one desirable trait (such as high temperature stability) but being deficient in other properties (such as high flux stability). The project showed that one can reproducibly obtain these properties by generating a novel compositional structure inside of the nanomaterials; in addition, the project formulated an initial theoretical framework linking the compositional structure to the list of high performance optical properties. Over the course of the project, the synthetic methodology for producing the novel composition was evolved to enable the synthesis of these nanomaterials at a cost approximately equal to that required for forming typical conventional nanocrystals. Given the above results, the last major remaining step prior to scale up of the nanomaterials is to limit the oxidation of these materials during the tens of

  4. Computationally efficient measure of topological redundancy of biological and social networks

    Science.gov (United States)

    Albert, Réka; Dasgupta, Bhaskar; Hegde, Rashmi; Sivanathan, Gowri Sangeetha; Gitter, Anthony; Gürsoy, Gamze; Paul, Pradyut; Sontag, Eduardo

    2011-09-01

    It is well known that biological and social interaction networks have a varying degree of redundancy, though a consensus of the precise cause of this is so far lacking. In this paper, we introduce a topological redundancy measure for labeled directed networks that is formal, computationally efficient, and applicable to a variety of directed networks such as cellular signaling, and metabolic and social interaction networks. We demonstrate the computational efficiency of our measure by computing its value and statistical significance on a number of biological and social networks with up to several thousands of nodes and edges. Our results suggest a number of interesting observations: (1) Social networks are more redundant that their biological counterparts, (2) transcriptional networks are less redundant than signaling networks, (3) the topological redundancy of the C. elegans metabolic network is largely due to its inclusion of currency metabolites, and (4) the redundancy of signaling networks is highly (negatively) correlated with the monotonicity of their dynamics.

  5. A New Computationally Efficient Measure of Topological Redundancy of Biological and Social Networks

    CERN Document Server

    Albert, Reka; Gitter, Anthony; Gursoy, Gamze; Hegde, Rashmi; Paul, Pradyut; Sivanathan, Gowri Sangeetha; Sontag, Eduardo

    2011-01-01

    It is well-known that biological and social interaction networks have a varying degree of redundancy, though a consensus of the precise cause of this is so far lacking. In this paper, we introduce a topological redundancy measure for labeled directed networks that is formal, computationally efficient and applicable to a variety of directed networks such as cellular signaling, metabolic and social interaction networks. We demonstrate the computational efficiency of our measure by computing its value and statistical significance on a number of biological and social networks with up to several thousands of nodes and edges. Our results suggest a number of interesting observations: (1) social networks are more redundant that their biological counterparts, (2) transcriptional networks are less redundant than signaling networks, (3) the topological redundancy of the C. elegans metabolic network is largely due to its inclusion of currency metabolites, and (4) the redundancy of signaling networks is highly (negatively...

  6. High Efficiency Power Converter for Low Voltage High Power Applications

    DEFF Research Database (Denmark)

    Nymand, Morten

    , and remote power generation for light towers, camper vans, boats, beacons, and buoys etc. In chapter 2, a review of current state-of-the-art is presented. The best performing converters achieve moderately high peak efficiencies at high input voltage and medium power level. However, system dimensioning...

  7. Efficient scatter model for simulation of ultrasound images from computed tomography data

    Science.gov (United States)

    D'Amato, J. P.; Lo Vercio, L.; Rubi, P.; Fernandez Vera, E.; Barbuzza, R.; Del Fresno, M.; Larrabide, I.

    2015-12-01

    Background and motivation: Real-time ultrasound simulation refers to the process of computationally creating fully synthetic ultrasound images instantly. Due to the high value of specialized low cost training for healthcare professionals, there is a growing interest in the use of this technology and the development of high fidelity systems that simulate the acquisitions of echographic images. The objective is to create an efficient and reproducible simulator that can run either on notebooks or desktops using low cost devices. Materials and methods: We present an interactive ultrasound simulator based on CT data. This simulator is based on ray-casting and provides real-time interaction capabilities. The simulation of scattering that is coherent with the transducer position in real time is also introduced. Such noise is produced using a simplified model of multiplicative noise and convolution with point spread functions (PSF) tailored for this purpose. Results: The computational efficiency of scattering maps generation was revised with an improved performance. This allowed a more efficient simulation of coherent scattering in the synthetic echographic images while providing highly realistic result. We describe some quality and performance metrics to validate these results, where a performance of up to 55fps was achieved. Conclusion: The proposed technique for real-time scattering modeling provides realistic yet computationally efficient scatter distributions. The error between the original image and the simulated scattering image was compared for the proposed method and the state-of-the-art, showing negligible differences in its distribution.

  8. Quantum wells for high-efficiency photovoltaics

    Science.gov (United States)

    Alonso-Álvarez, Diego; Ekins-Daukes, Nicholas

    2016-03-01

    Over the last couple of decades, there has been an intense research on strain balanced semiconductor quantum wells (QW) to increase the efficiency of multi-junction solar (MJ) solar cells grown monolithically on germanium. So far, the most successful application of QWs have required just to tailor a few tens of nanometers the absorption edge of a given subcell in order to reach the optimum spectral position. However, the demand for higher efficiency devices requiring 3, 4 or more junctions, represents a major difference in the challenges QWs must face: tailoring the absorption edge of a host material is not enough, but a complete new device, absorbing light in a different spectral region, must be designed. Among the most important issues to solve is the need for an optically thick structure to absorb enough light while keeping excellent carrier extraction using highly strained materials. Improvement of the growth techniques, smarter device designs - involving superlattices and shifted QWs, for example - or the use of quantum wires rather than QWs, have proven to be very effective steps towards high efficient MJ solar cells based on nanostructures in the last couple of years. But more is to be done to reach the target performances. This work discusses all these challenges, the limitations they represent and the different approaches that are being used to overcome them.

  9. HIGH-EFFICIENCY AUTONOMOUS LASER ADAPTIVE OPTICS

    Energy Technology Data Exchange (ETDEWEB)

    Baranec, Christoph [Institute for Astronomy, University of Hawai' i at Mānoa, Hilo, HI, NZ 96720-2700 (United States); Riddle, Reed; Tendulkar, Shriharsh; Hogstrom, Kristina; Bui, Khanh; Dekany, Richard; Kulkarni, Shrinivas [Division of Physics, Mathematics, and Astronomy, California Institute of Technology, Pasadena, CA 91125 (United States); Law, Nicholas M. [Department of Physics and Astronomy, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599-3255 (United States); Ramaprakash, A. N.; Burse, Mahesh; Chordia, Pravin; Das, Hillol; Punnadi, Sujit, E-mail: baranec@hawaii.edu [Inter-University Centre for Astronomy and Astrophysics, Ganeshkhind, Pune 411007 (India)

    2014-07-20

    As new large-scale astronomical surveys greatly increase the number of objects targeted and discoveries made, the requirement for efficient follow-up observations is crucial. Adaptive optics imaging, which compensates for the image-blurring effects of Earth's turbulent atmosphere, is essential for these surveys, but the scarcity, complexity and high demand of current systems limit their availability for following up large numbers of targets. To address this need, we have engineered and implemented Robo-AO, a fully autonomous laser adaptive optics and imaging system that routinely images over 200 objects per night with an acuity 10 times sharper at visible wavelengths than typically possible from the ground. By greatly improving the angular resolution, sensitivity, and efficiency of 1-3 m class telescopes, we have eliminated a major obstacle in the follow-up of the discoveries from current and future large astronomical surveys.

  10. High-efficiency Autonomous Laser Adaptive Optics

    CERN Document Server

    Baranec, Christoph; Law, Nicholas M; Ramaprakash, A N; Tendulkar, Shriharsh; Hogstrom, Kristina; Bui, Khanh; Burse, Mahesh; Chordia, Pravin; Das, Hillol; Dekany, Richard; Kulkarni, Shrinivas; Punnadi, Sujit

    2014-01-01

    As new large-scale astronomical surveys greatly increase the number of objects targeted and discoveries made, the requirement for efficient follow-up observations is crucial. Adaptive optics imaging, which compensates for the image-blurring effects of Earth's turbulent atmosphere, is essential for these surveys, but the scarcity, complexity and high demand of current systems limits their availability for following up large numbers of targets. To address this need, we have engineered and implemented Robo-AO, a fully autonomous laser adaptive optics and imaging system that routinely images over 200 objects per night with an acuity 10 times sharper at visible wavelengths than typically possible from the ground. By greatly improving the angular resolution, sensitivity, and efficiency of 1-3 m class telescopes, we have eliminated a major obstacle in the follow-up of the discoveries from current and future large astronomical surveys.

  11. High efficiency motors; Motores de alta eficiencia

    Energy Technology Data Exchange (ETDEWEB)

    Uranga Favela, Ivan Jaime [Energia Controlada de Mexico, S. A. de C. V., Mexico, D. F. (Mexico)

    1992-12-31

    This paper is a technical-financial study of the high efficiency and super-premium motors. As it is widely known, more than 60% of the electrical energy generated in the country is used for the operation of motors, in industry as well as in commerce. Therefore the importance that the motors have in the efficient energy use. [Espanol] El presente trabajo es un estudio tecnico-financiero de los motores de alta eficiencia y los motores super premium. Como es ampliamente conocido, mas del 60% de la energia electrica generada en el pais, es utilizada para accionar motores, dentro de la industria y el comercio. De alli la importancia que los motores tienen en el uso eficiente de la energia.

  12. Federal High End Computing (HEC) Information Portal

    Data.gov (United States)

    Networking and Information Technology Research and Development, Executive Office of the President — This portal provides information about opportunities to engage in U.S. Federal government high performance computing activities, including supercomputer use,...

  13. High School Physics and the Affordable Computer.

    Science.gov (United States)

    Harvey, Norman L.

    1978-01-01

    Explains how the computer was used in a high school physics course; Project Physics program and individualized study PSSC physics program. Evaluates the capabilities and limitations of a $600 microcomputer system. (GA)

  14. NINJA: Java for High Performance Numerical Computing

    Directory of Open Access Journals (Sweden)

    José E. Moreira

    2002-01-01

    Full Text Available When Java was first introduced, there was a perception that its many benefits came at a significant performance cost. In the particularly performance-sensitive field of numerical computing, initial measurements indicated a hundred-fold performance disadvantage between Java and more established languages such as Fortran and C. Although much progress has been made, and Java now can be competitive with C/C++ in many important situations, significant performance challenges remain. Existing Java virtual machines are not yet capable of performing the advanced loop transformations and automatic parallelization that are now common in state-of-the-art Fortran compilers. Java also has difficulties in implementing complex arithmetic efficiently. These performance deficiencies can be attacked with a combination of class libraries (packages, in Java that implement truly multidimensional arrays and complex numbers, and new compiler techniques that exploit the properties of these class libraries to enable other, more conventional, optimizations. Two compiler techniques, versioning and semantic expansion, can be leveraged to allow fully automatic optimization and parallelization of Java code. Our measurements with the NINJA prototype Java environment show that Java can be competitive in performance with highly optimized and tuned Fortran code.

  15. High Quantum Efficiency OLED Lighting Systems

    Energy Technology Data Exchange (ETDEWEB)

    Shiang, Joseph [General Electric (GE) Global Research, Fairfield, CT (United States)

    2011-09-30

    The overall goal of the program was to apply improvements in light outcoupling technology to a practical large area plastic luminaire, and thus enable the product vision of an extremely thin form factor high efficiency large area light source. The target substrate was plastic and the baseline device was operating at 35 LPW at the start of the program. The target LPW of the program was a >2x improvement in the LPW efficacy and the overall amount of light to be delivered was relatively high 900 lumens. Despite the extremely difficult challenges associated with scaling up a wet solution process on plastic substrates, the program was able to make substantial progress. A small molecule wet solution process was successfully implemented on plastic substrates with almost no loss in efficiency in transitioning from the laboratory scale glass to large area plastic substrates. By transitioning to a small molecule based process, the LPW entitlement increased from 35 LPW to 60 LPW. A further 10% improvement in outcoupling efficiency was demonstrated via the use of a highly reflecting cathode, which reduced absorptive loss in the OLED device. The calculated potential improvement in some cases is even larger, ~30%, and thus there is considerable room for optimism in improving the net light coupling efficacy, provided absorptive loss mechanisms are eliminated. Further improvements are possible if scattering schemes such as the silver nanowire based hard coat structure are fully developed. The wet coating processes were successfully scaled to large area plastic substrate and resulted in the construction of a 900 lumens luminaire device.

  16. Nanooptics for high efficient photon managment

    Science.gov (United States)

    Wyrowski, Frank; Schimmel, Hagen

    2005-09-01

    Optical systems for photon management, that is the generation of tailored electromagnetic fields, constitute one of the keys for innovation through photonics. An important subfield of photon management deals with the transformation of an incident light field into a field of specified intensity distribution. In this paper we consider some basic aspects of the nature of systems for those light transformations. It turns out, that the transversal redistribution of energy (TRE) is of central concern to achieve systems with high transformation efficiency. Besides established techniques nanostructured optical elements (NOE) are demanded to implement transversal energy redistribution. That builds a bridge between the needs of photon management, optical engineering, and nanooptics.

  17. Analog models of computations \\& Effective Church Turing Thesis: Efficient simulation of Turing machines by the General Purpose Analog Computer

    CERN Document Server

    Pouly, Amaury; Graça, Daniel S

    2012-01-01

    \\emph{Are analog models of computations more powerful than classical models of computations?} From a series of recent papers, it is now clear that many realistic analog models of computations are provably equivalent to classical digital models of computations from a \\emph{computability} point of view. Take, for example, the probably most realistic model of analog computation, the General Purpose Analog Computer (GPAC) model from Claude Shannon, a model for Differential Analyzers, which are analog machines used from 1930s to early 1960s to solve various problems. It is now known that functions computable by Turing machines are provably exactly those that are computable by GPAC. This paper is about next step: understanding if this equivalence also holds at the \\emph{complexity} level. In this paper we show that the realistic models of analog computation -- namely the General Purpose Analog Computer (GPAC) -- can simulate Turing machines in a computationally efficient manner. More concretely we show that, modulo...

  18. Studying an Eulerian Computer Model on Different High-performance Computer Platforms and Some Applications

    Science.gov (United States)

    Georgiev, K.; Zlatev, Z.

    2010-11-01

    The Danish Eulerian Model (DEM) is an Eulerian model for studying the transport of air pollutants on large scale. Originally, the model was developed at the National Environmental Research Institute of Denmark. The model computational domain covers Europe and some neighbour parts belong to the Atlantic Ocean, Asia and Africa. If DEM model is to be applied by using fine grids, then its discretization leads to a huge computational problem. This implies that such a model as DEM must be run only on high-performance computer architectures. The implementation and tuning of such a complex large-scale model on each different computer is a non-trivial task. Here, some comparison results of running of this model on different kind of vector (CRAY C92A, Fujitsu, etc.), parallel computers with distributed memory (IBM SP, CRAY T3E, Beowulf clusters, Macintosh G4 clusters, etc.), parallel computers with shared memory (SGI Origin, SUN, etc.) and parallel computers with two levels of parallelism (IBM SMP, IBM BlueGene/P, clusters of multiprocessor nodes, etc.) will be presented. The main idea in the parallel version of DEM is domain partitioning approach. Discussions according to the effective use of the cache and hierarchical memories of the modern computers as well as the performance, speed-ups and efficiency achieved will be done. The parallel code of DEM, created by using MPI standard library, appears to be highly portable and shows good efficiency and scalability on different kind of vector and parallel computers. Some important applications of the computer model output are presented in short.

  19. MADLVF: An Energy Efficient Resource Utilization Approach for Cloud Computing

    Directory of Open Access Journals (Sweden)

    J.K. Verma

    2014-06-01

    Full Text Available Last few decades have remained the witness of steeper growth in demand for higher computational power. It is merely due to shift from the industrial age to Information and Communication Technology (ICT age which was marginally the result of digital revolution. Such trend in demand caused establishment of large-scale data centers situated at geographically apart locations. These large-scale data centers consume a large amount of electrical energy which results into very high operating cost and large amount of carbon dioxide (CO2 emission due to resource underutilization. We propose MADLVF algorithm to overcome the problems such as resource underutilization, high energy consumption, and large CO2 emissions. Further, we present a comparative study between the proposed algorithm and MADRS algorithms showing proposed methodology outperforms over the existing one in terms of energy consumption and the number of VM migrations.

  20. Computationally efficient finite element evaluation of natural patellofemoral mechanics.

    Science.gov (United States)

    Fitzpatrick, Clare K; Baldwin, Mark A; Rullkoetter, Paul J

    2010-12-01

    pressures averaged 8.3%, 11.2%, and 5.7% between rigid and deformable analyses in the tibiofemoral joint. As statistical, probabilistic, and optimization techniques can require hundreds to thousands of analyses, a viable platform is crucial to component evaluation or clinical applications. The computationally efficient rigid body platform described in this study may be integrated with statistical and probabilistic methods and has potential clinical application in understanding in vivo joint mechanics on a subject-specific or population basis.

  1. Vacuum MOCVD fabrication of high efficience cells

    Science.gov (United States)

    Partain, L. D.; Fraas, L. M.; Mcleod, P. S.; Cape, J. A.

    1985-01-01

    Vacuum metal-organic-chemical-vapor-deposition (MOCVD) is a new fabrication process with improved safety and easier scalability due to its metal rather than glass construction and its uniform multiport gas injection system. It uses source materials more efficiently than other methods because the vacuum molecular flow conditions allow the high sticking coefficient reactants to reach the substrates as undeflected molecular beams and the hot chamber walls cause the low sticking coefficient reactants to bounce off the walls and interact with the substrates many times. This high source utilization reduces the materials costs power device and substantially decreases the amounts of toxic materials that must be handled as process effluents. The molecular beams allow precise growth control. With improved source purifications, vacuum MOCVD has provided p GaAs layers with 10-micron minority carrier diffusion lengths and GaAs and GaAsSb solar cells with 20% AMO efficiencies at 59X and 99X sunlight concentration ratios. Mechanical stacking has been identified as the quickest, most direct and logical path to stacked multiple-junction solar cells that perform better than the best single-junction devices. The mechanical stack is configured for immediate use in solar arrays and allows interconnections that improve the system end-of-life performance in space.

  2. High efficiency dielectric metasurfaces at visible wavelengths

    CERN Document Server

    Devlin, Robert C; Chen, Wei-Ting; Oh, Jaewon; Capasso, Federico

    2016-01-01

    Metasurfaces are planar optical elements that hold promise for overcoming the limitations of refractive and conventional diffractive optics1-3. Dielectric metasurfaces demonstrated thus far4-10 are limited to transparency windows at infrared wavelengths because of significant optical absorption and loss at visible wavelengths. It is critical that new materials and fabrication techniques be developed for dielectric metasurfaces at visible wavelengths to enable applications such as three-dimensional displays, wearable optics and planar optical systems11. Here, we demonstrate high performance titanium dioxide dielectric metasurfaces in the form of holograms for red, green and blue wavelengths with record absolute efficiency (>78%). We use atomic layer deposition of amorphous titanium dioxide that exhibits low surface roughness of 0.738 nm and ideal optical properties. To fabricate the metasurfaces we use a lift-off-like process that allows us to produce highly anisotropic nanofins with shape birefringence. This ...

  3. Efficient Parallel Computation of Nearest Neighbor Interchange Distances

    CERN Document Server

    Gast, Mikael

    2012-01-01

    The nni-distance is a well-known distance measure for phylogenetic trees. We construct an efficient parallel approximation algorithm for the nni-distance in the CRCW-PRAM model running in O(log n) time on O(n) processors. Given two phylogenetic trees T1 and T2 on the same set of taxa and with the same multi-set of edge-weights, the algorithm constructs a sequence of nni-operations of weight at most O(log n) \\cdot opt, where opt denotes the minimum weight of a sequence of nni-operations transforming T1 into T2 . This algorithm is based on the sequential approximation algorithm for the nni-distance given by DasGupta et al. (2000). Furthermore, we show that the problem of identifying so called good edge-pairs between two weighted phylogenies can be computed in O(log n) time on O(n log n) processors.

  4. Efficient computation of bifurcation diagrams via adaptive ROMs

    Energy Technology Data Exchange (ETDEWEB)

    Terragni, F [Gregorio Millán Institute for Fluid Dynamics, Nanoscience and Industrial Mathematics, Universidad Carlos III de Madrid, E-28911 Leganés (Spain); Vega, J M, E-mail: fterragn@ing.uc3m.es [E.T.S.I. Aeronáuticos, Universidad Politécnica de Madrid, E-28040 Madrid (Spain)

    2014-08-01

    Various ideas concerning model reduction based on proper orthogonal decomposition are discussed, exploited, and suited to the approximation of complex bifurcations in some dissipative systems. The observation that the most energetic modes involved in these low dimensional descriptions depend only weakly on the actual values of the problem parameters is firstly highlighted and used to develop a simple strategy to capture the transitions occurring over a given bifurcation parameter span. Flexibility of the approach is stressed by means of some numerical experiments. A significant improvement is obtained by introducing a truncation error estimate to detect when the approximation fails. Thus, the considered modes are suitably updated on demand, as the bifurcation parameter is varied, in order to account for possible changes in the phase space of the system that might be missed. A further extension of the method to more complex (quasi-periodic and chaotic) attractors is finally outlined by implementing a control of truncation instabilities, which leads to a general, adaptive reduced order model for the construction of bifurcation diagrams. Illustration of the ideas and methods in the complex Ginzburg–Landau equation (a paradigm of laminar flows on a bounded domain) evidences a fairly good computational efficiency. (paper)

  5. Efficient computation of bifurcation diagrams via adaptive ROMs

    Science.gov (United States)

    Terragni, F.; Vega, J. M.

    2014-08-01

    Various ideas concerning model reduction based on proper orthogonal decomposition are discussed, exploited, and suited to the approximation of complex bifurcations in some dissipative systems. The observation that the most energetic modes involved in these low dimensional descriptions depend only weakly on the actual values of the problem parameters is firstly highlighted and used to develop a simple strategy to capture the transitions occurring over a given bifurcation parameter span. Flexibility of the approach is stressed by means of some numerical experiments. A significant improvement is obtained by introducing a truncation error estimate to detect when the approximation fails. Thus, the considered modes are suitably updated on demand, as the bifurcation parameter is varied, in order to account for possible changes in the phase space of the system that might be missed. A further extension of the method to more complex (quasi-periodic and chaotic) attractors is finally outlined by implementing a control of truncation instabilities, which leads to a general, adaptive reduced order model for the construction of bifurcation diagrams. Illustration of the ideas and methods in the complex Ginzburg-Landau equation (a paradigm of laminar flows on a bounded domain) evidences a fairly good computational efficiency.

  6. Development of a computationally efficient urban modeling approach

    DEFF Research Database (Denmark)

    Wolfs, Vincent; Murla, Damian; Ntegeka, Victor;

    2016-01-01

    This paper presents a parsimonious and data-driven modelling approach to simulate urban floods. Flood levels simulated by detailed 1D-2D hydrodynamic models can be emulated using the presented conceptual modelling approach with a very short calculation time. In addition, the model detail can...... be adjust-ed, allowing the modeller to focus on flood-prone locations. This results in efficiently parameterized models that can be tailored to applications. The simulated flood levels are transformed into flood extent maps using a high resolution (0.5-meter) digital terrain model in GIS. To illustrate...... the developed methodology, a case study for the city of Ghent in Belgium is elaborated. The configured conceptual model mimics the flood levels of a detailed 1D-2D hydrodynamic InfoWorks ICM model accurately, while the calculation time is an order of magnitude of 106 times shorter than the original highly...

  7. Simple Motor Control Concept Results High Efficiency at High Velocities

    Science.gov (United States)

    Starin, Scott; Engel, Chris

    2013-09-01

    The need for high velocity motors in space applications for reaction wheels and detectors has stressed the limits of Brushless Permanent Magnet Motors (BPMM). Due to inherent hysteresis core losses, conventional BPMMs try to balance the need for torque verses hysteresis losses. Cong-less motors have significantly less hysteresis losses but suffer from lower efficiencies. Additionally, the inherent low inductance in cog-less motors result in high ripple currents or high switching frequencies, which lowers overall efficiency and increases performance demands on the control electronics.However, using a somewhat forgotten but fully qualified technology of Isotropic Magnet Motors (IMM), extremely high velocities may be achieved at low power input using conventional drive electronics. This paper will discuss the trade study efforts and empirical test data on a 34,000 RPM IMM.

  8. Grid computing in high energy physics

    CERN Document Server

    Avery, P

    2004-01-01

    Over the next two decades, major high energy physics (HEP) experiments, particularly at the Large Hadron Collider, will face unprecedented challenges to achieving their scientific potential. These challenges arise primarily from the rapidly increasing size and complexity of HEP datasets that will be collected and the enormous computational, storage and networking resources that will be deployed by global collaborations in order to process, distribute and analyze them. Coupling such vast information technology resources to globally distributed collaborations of several thousand physicists requires extremely capable computing infrastructures supporting several key areas: (1) computing (providing sufficient computational and storage resources for all processing, simulation and analysis tasks undertaken by the collaborations); (2) networking (deploying high speed networks to transport data quickly between institutions around the world); (3) software (supporting simple and transparent access to data and software r...

  9. Dawning4000A high performance computer

    Institute of Scientific and Technical Information of China (English)

    SUN Ninghui; MENG Dan

    2007-01-01

    Dawning4000A is an AMD Opteron-based Linux Cluster with 11.2Tflops peak performance and 8.06Tflops Linpack performance.It was developed for the Shanghai Supercomputer Center (SSC)as one of the computing power stations of the China National Grid (CNGrid)project.The Massively Cluster Computer (MCC)architecture is proposed to put added-value on the industry standard system.Several grid-enabling components are developed to support the running environment of the CNGrid.It is an achievement for a high performance computer with the low-cost approach.

  10. The Effect of Computer Automation on Institutional Review Board (IRB) Office Efficiency

    Science.gov (United States)

    Oder, Karl; Pittman, Stephanie

    2015-01-01

    Companies purchase computer systems to make their processes more efficient through automation. Some academic medical centers (AMC) have purchased computer systems for their institutional review boards (IRB) to increase efficiency and compliance with regulations. IRB computer systems are expensive to purchase, deploy, and maintain. An AMC should…

  11. Federal Plan for High-End Computing. Report of the High-End Computing Revitalization Task Force (HECRTF)

    Science.gov (United States)

    2004-07-01

    and other energy feedstock more efficiently. Signal Transduction Pathways Develop atomic-level computational models and simulations of complex...biomolecules to explain and predict cell signal pathways and their disrupters. Yield understanding of initiation of cancer and other diseases and their...calculations also introduces a requirement for a high degree of internodal connectivity (high bisection bandwidth). These needs cannot be met simply by

  12. Multiscale approaches to high efficiency photovoltaics

    Directory of Open Access Journals (Sweden)

    Connolly James Patrick

    2016-01-01

    Full Text Available While renewable energies are achieving parity around the globe, efforts to reach higher solar cell efficiencies becomes ever more difficult as they approach the limiting efficiency. The so-called third generation concepts attempt to break this limit through a combination of novel physical processes and new materials and concepts in organic and inorganic systems. Some examples of semi-empirical modelling in the field are reviewed, in particular for multispectral solar cells on silicon (French ANR project MultiSolSi. Their achievements are outlined, and the limits of these approaches shown. This introduces the main topic of this contribution, which is the use of multiscale experimental and theoretical techniques to go beyond the semi-empirical understanding of these systems. This approach has already led to great advances at modelling which have led to modelling software, which is widely known. Yet, a survey of the topic reveals a fragmentation of efforts across disciplines, firstly, such as organic and inorganic fields, but also between the high efficiency concepts such as hot carrier cells and intermediate band concepts. We show how this obstacle to the resolution of practical research obstacles may be lifted by inter-disciplinary cooperation across length scales, and across experimental and theoretical fields, and finally across materials systems. We present a European COST Action “MultiscaleSolar” kicking off in early 2015, which brings together experimental and theoretical partners in order to develop multiscale research in organic and inorganic materials. The goal of this defragmentation and interdisciplinary collaboration is to develop understanding across length scales, which will enable the full potential of third generation concepts to be evaluated in practise, for societal and industrial applications.

  13. Novel Nanophosphors for High Efficiency Fluorescent Lamps

    Energy Technology Data Exchange (ETDEWEB)

    Alok Srivatava

    2007-03-31

    This is the Final Report of the Novel Nanophosphors for High Efficiency Fluorescent Lamps, Department of Energy (DOE). The overall goal of this three-year program is to develop novel hybrid phosphors by coating commercially available lamp phosphors with highly stable wide band-gap nanocrystalline phosphors (NCP). The prime technical approach is the development of NCP quantum-splitting phosphor (QSP) and ultra-violet (UV) emitting phosphors with quantum efficiencies exceeding that of the conventional phosphors at 185 nm. The novel hybrid phosphors will increase the efficiency of the fluorescent lamps by up to 32%, enabling total energy savings of 0.26 quads, the reduction in the U.S. energy bill by $6.5 billion and the reduction of the annual carbon emission by 4.1 billion kilogram. Our work started by investigating through modeling calculations the requirement for the particle size of the NCP. Our work to develop suitable nanocrystalline phosphors started with the known oxide quantum splitting and UV emitting phosphors. We demonstrated several synthesis techniques for the production of high quality nanocrystalline materials that crystallizes in the desired phase and with the desired particle size. In collaboration with our subcontractor we demonstrated the feasibility for the manufacture of NC phosphors. We also demonstrated novel techniques of coating the NCP on the surface of micron sized phosphors. Our chief achievement pertains to the successful testing of the coated hybrid phosphor systems in linear fluorescent lamps. In linear fluorescent lamp tests, we have demonstrated up to 7% increase in the efficacy of hybrid phosphors over the conventional (uncoated) phosphors. We have also demonstrated the improvement in the lumen maintenance of the coated phosphors. A hybrid phosphor system based on the commercial red emitting phosphor, Y{sub 2}O{sub 3}:Eu{sup 3+} did not show the anticipated improvement in lamp efficacy. We explored the reasons for this observation

  14. EMMP :a highly efficient membership management protocol

    Institute of Scientific and Technical Information of China (English)

    LI Renfa; XIE Yunlong; WEN Jigang; YUE Guangxue

    2007-01-01

    Gossip (or epidemic) algorithms have recently become popular solutions to multicast message dissemination in peer-to-peer systems.Nevertheless,it is not straightforward to apply gossip to on-demand streaming because it often fails to achieve a timely delivery.To solve this problem and taking into account the characteristic of peers randomly joining and leaving in peer-to-peer systems,an Efficient Membership Management Protocol (EMMP) has been presented.Every node only needs to keep contact with O (log(N)) nodes,and EMMP can support the reliable dissemination of messages.Considering the "distance" between peers,it causes the major data to be transmitted in a local area and reduces the backbone's traffic,and speeds up the dissemination of messages between peers.This paper has adopted the"goodfriend" mechanism to reduce the influence on the system when a peer fails or leaves.Simulation results show that EMMP is highly efficient,and both the redundancy and the delay of the system are well solved.

  15. Efficiency of High Order Spectral Element Methods on Petascale Architectures

    KAUST Repository

    Hutchinson, Maxwell

    2016-06-14

    High order methods for the solution of PDEs expose a tradeoff between computational cost and accuracy on a per degree of freedom basis. In many cases, the cost increases due to higher arithmetic intensity while affecting data movement minimally. As architectures tend towards wider vector instructions and expect higher arithmetic intensities, the best order for a particular simulation may change. This study highlights preferred orders by identifying the high order efficiency frontier of the spectral element method implemented in Nek5000 and NekBox: the set of orders and meshes that minimize computational cost at fixed accuracy. First, we extract Nek’s order-dependent computational kernels and demonstrate exceptional hardware utilization by hardware-aware implementations. Then, we perform productionscale calculations of the nonlinear single mode Rayleigh-Taylor instability on BlueGene/Q and Cray XC40-based supercomputers to highlight the influence of the architecture. Accuracy is defined with respect to physical observables, and computational costs are measured by the corehour charge of the entire application. The total number of grid points needed to achieve a given accuracy is reduced by increasing the polynomial order. On the XC40 and BlueGene/Q, polynomial orders as high as 31 and 15 come at no marginal cost per timestep, respectively. Taken together, these observations lead to a strong preference for high order discretizations that use fewer degrees of freedom. From a performance point of view, we demonstrate up to 60% full application bandwidth utilization at scale and achieve ≈1PFlop/s of compute performance in Nek’s most flop-intense methods.

  16. Design of High Efficient MPPT Solar Inverter

    Directory of Open Access Journals (Sweden)

    Sunitha K. A.

    2017-01-01

    Full Text Available This work aims to design a High Efficient Maximum Power Point Tracking (MPPT Solar Inverter. A boost converter is designed in the system to boost the power from the photovoltaic panel. By this experimental setup a room consisting of 500 Watts load (eight fluorescent tubes is completely controlled. It is aimed to decrease the maintenance cost. A microcontroller is introduced for tracking the P&O (Perturb and Observe algorithm used for tracking the maximum power point. The duty cycle for the operation of the boost convertor is optimally adjusted by using MPPT controller. There is a MPPT charge controller to charge the battery as well as fed to inverter which runs the load. Both the P&O scheme with the fixed variation for the reference current and the intelligent MPPT algorithm were able to identify the global Maximum power point, however the performance of the MPPT algorithm was better.

  17. COMPUTER APPLICATION SYSTEM FOR OPERATIONAL EFFICIENCY OF DIESEL RAILBUSES

    Directory of Open Access Journals (Sweden)

    Łukasz WOJCIECHOWSKI

    2016-09-01

    Full Text Available The article presents a computer algorithm to calculate the estimated operating cost analysis rail bus. This computer application system compares the cost of employment locomotive and wagon, the cost of using locomotives and cost of using rail bus. An intensive growth of passenger railway traffic increased a demand for modern computer systems to management means of transportation. Described computer application operates on the basis of selected operating parameters of rail buses.

  18. Popescu-Rohrlich correlations imply efficient instantaneous nonlocal quantum computation

    OpenAIRE

    Broadbent, Anne

    2015-01-01

    In instantaneous nonlocal quantum computation, two parties cooperate in order to perform a quantum computation on their joint inputs, while being restricted to a single round of simultaneous communication. Previous results showed that instantaneous nonlocal quantum computation is possible, at the cost of an exponential amount of prior shared entanglement (in the size of the input). Here, we show that a linear amount of entanglement suffices, (in the size of the computation), as long as the pa...

  19. High Efficiency Centrifugal Compressor for Rotorcraft Applications

    Science.gov (United States)

    Medic, Gorazd; Sharma, Om P.; Jongwook, Joo; Hardin, Larry W.; McCormick, Duane C.; Cousins, William T.; Lurie, Elizabeth A.; Shabbir, Aamir; Holley, Brian M.; Van Slooten, Paul R.

    2014-01-01

    The report "High Efficiency Centrifugal Compressor for Rotorcraft Applications" documents the work conducted at UTRC under the NRA Contract NNC08CB03C, with cost share 2/3 NASA, and 1/3 UTRC, that has been extended to 4.5 years. The purpose of this effort was to identify key technical barriers to advancing the state-of-the-art of small centrifugal compressor stages; to delineate the measurements required to provide insight into the flow physics of the technical barriers; to design, fabricate, install, and test a state-of-the-art research compressor that is representative of the rear stage of an axial-centrifugal aero-engine; and to acquire detailed aerodynamic performance and research quality data to clarify flow physics and to establish detailed data sets for future application. The design activity centered on meeting the goal set outlined in the NASA solicitation-the design target was to increase efficiency at higher work factor, while also reducing the maximum diameter of the stage. To fit within the existing Small Engine Components Test Facility at NASA Glenn Research Center (GRC) and to facilitate component re-use, certain key design parameters were fixed by UTRC, including impeller tip diameter, impeller rotational speed, and impeller inlet hub and shroud radii. This report describes the design effort of the High Efficiency Centrifugal Compressor stage (HECC) and delineation of measurements, fabrication of the compressor, and the initial tests that were performed. A new High-Efficiency Centrifugal Compressor stage with a very challenging reduction in radius ratio was successfully designed, fabricated and installed at GRC. The testing was successful, with no mechanical problems and the running clearances were achieved without impeller rubs. Overall, measured pressure ratio of 4.68, work factor of 0.81, and at design exit corrected flow rate of 3 lbm/s met the target requirements. Polytropic efficiency of 85.5 percent and stall margin of 7.5 percent were

  20. A high-efficiency superhydrophobic plasma separator.

    Science.gov (United States)

    Liu, Changchun; Liao, Shih-Chuan; Song, Jinzhao; Mauk, Michael G; Li, Xuanwen; Wu, Gaoxiang; Ge, Dengteng; Greenberg, Robert M; Yang, Shu; Bau, Haim H

    2016-02-01

    To meet stringent limit-of-detection specifications for low abundance target molecules, a relatively large volume of plasma is needed for many blood-based clinical diagnostics. Conventional centrifugation methods for plasma separation are not suitable for on-site testing or bedside diagnostics. Here, we report a simple, yet high-efficiency, clamshell-style, superhydrophobic plasma separator that is capable of separating a relatively large volume of plasma from several hundred microliters of whole blood (finger-prick blood volume). The plasma separator consists of a superhydrophobic top cover with a separation membrane and a superhydrophobic bottom substrate. Unlike previously reported membrane-based plasma separators, the separation membrane in our device is positioned at the top of the sandwiched whole blood film to increase the membrane separation capacity and plasma yield. In addition, the device's superhydrophobic characteristics (i) facilitates the formation of well-defined, contracted, thin blood film with a high contact angle; (ii) minimizes biomolecular adhesion to surfaces; (iii) increases blood clotting time; and (iv) reduces blood cell hemolysis. The device demonstrated a "blood in-plasma out" capability, consistently extracting 65 ± 21.5 μL of plasma from 200 μL of whole blood in less than 10 min without electrical power. The device was used to separate plasma from Schistosoma mansoni genomic DNA-spiked whole blood with a recovery efficiency of >84.5 ± 25.8%. The S. mansoni genomic DNA in the separated plasma was successfully tested on our custom-made microfluidic chip by using loop mediated isothermal amplification (LAMP) method.

  1. Building Efficient Wireless Infrastructures for Pervasive Computing Environments

    Science.gov (United States)

    Sheng, Bo

    2010-01-01

    Pervasive computing is an emerging concept that thoroughly brings computing devices and the consequent technology into people's daily life and activities. Most of these computing devices are very small, sometimes even "invisible", and often embedded into the objects surrounding people. In addition, these devices usually are not isolated, but…

  2. Advanced Klystrons for High Efficiency Accelerator Systems - Final Report

    Energy Technology Data Exchange (ETDEWEB)

    Read, Michael; Ives, Robert Lawrence

    2014-03-26

    This program explored tailoring of RF pulses used to drive accelerator cavities. Simulations indicated that properly shaping the pulse risetime to match accelerator cavity characteristics reduced reflected power and increased total efficiency. Tailoring the pulse requires a high power, gridded, klystron to shape the risetime while also controlling the beam current. The Phase I program generated a preliminary design of a gridded electron gun for a klystron producing 5-10 MW of RF power. This required design of a segmented cathode using Controlled Porosity Reservoir cathodes to limit power deposition on the grid. The program was successful in computationally designing a gun producing a high quality electron beam with grid control. Additional analysis of pulse tailoring indicated that technique would only be useful for cavity drive pulses that were less than approximately 2-3 times the risetime. Otherwise, the efficiency gained during the risetime of the pulse became insignificant when considering the efficiency over the entire pulse. Consequently, it was determined that a Phase II program would not provide sufficient return to justify the cost. Never the less, other applications for a high power gridded gun are currently being pursued. This klystron, for example, would facilitate development inverse Comptom x-ray sources by providing a high repetition rate (10 -100 kHz) RF source.

  3. High-performance computing for airborne applications

    Energy Technology Data Exchange (ETDEWEB)

    Quinn, Heather M [Los Alamos National Laboratory; Manuzzato, Andrea [Los Alamos National Laboratory; Fairbanks, Tom [Los Alamos National Laboratory; Dallmann, Nicholas [Los Alamos National Laboratory; Desgeorges, Rose [Los Alamos National Laboratory

    2010-06-28

    Recently, there has been attempts to move common satellite tasks to unmanned aerial vehicles (UAVs). UAVs are significantly cheaper to buy than satellites and easier to deploy on an as-needed basis. The more benign radiation environment also allows for an aggressive adoption of state-of-the-art commercial computational devices, which increases the amount of data that can be collected. There are a number of commercial computing devices currently available that are well-suited to high-performance computing. These devices range from specialized computational devices, such as field-programmable gate arrays (FPGAs) and digital signal processors (DSPs), to traditional computing platforms, such as microprocessors. Even though the radiation environment is relatively benign, these devices could be susceptible to single-event effects. In this paper, we will present radiation data for high-performance computing devices in a accelerated neutron environment. These devices include a multi-core digital signal processor, two field-programmable gate arrays, and a microprocessor. From these results, we found that all of these devices are suitable for many airplane environments without reliability problems.

  4. A Power Efficient Exaflop Computer Design for Global Cloud System Resolving Climate Models.

    Science.gov (United States)

    Wehner, M. F.; Oliker, L.; Shalf, J.

    2008-12-01

    Exascale computers would allow routine ensemble modeling of the global climate system at the cloud system resolving scale. Power and cost requirements of traditional architecture systems are likely to delay such capability for many years. We present an alternative route to the exascale using embedded processor technology to design a system optimized for ultra high resolution climate modeling. These power efficient processors, used in consumer electronic devices such as mobile phones, portable music players, cameras, etc., can be tailored to the specific needs of scientific computing. We project that a system capable of integrating a kilometer scale climate model a thousand times faster than real time could be designed and built in a five year time scale for US$75M with a power consumption of 3MW. This is cheaper, more power efficient and sooner than any other existing technology.

  5. Efficient acceleration of mutual information computation for nonrigid registration using CUDA.

    Science.gov (United States)

    Ikeda, Kei; Ino, Fumihiko; Hagihara, Kenichi

    2014-05-01

    In this paper, we propose an efficient acceleration method for the nonrigid registration of multimodal images that uses a graphics processing unit. The key contribution of our method is efficient utilization of on-chip memory for both normalized mutual information (NMI) computation and hierarchical B-spline deformation, which compose a well-known registration algorithm. We implement this registration algorithm as a compute unified device architecture program with an efficient parallel scheme and several optimization techniques such as hierarchical data organization, data reuse, and multiresolution representation. We experimentally evaluate our method with four clinical datasets consisting of up to 512 × 512 × 296 voxels. We find that exploitation of on-chip memory achieves a 12-fold increase in speed over an off-chip memory version and, therefore, it increases the efficiency of parallel execution from 4% to 46%. We also find that our method running on a GeForce GTX 580 card is approximately 14 times faster than a fully optimized CPU-based implementation running on four cores. Some multimodal registration results are also provided to understand the limitation of our method. We believe that our highly efficient method, which completes an alignment task within a few tens of seconds, will be useful to realize rapid nonrigid registration.

  6. Automated Development of Accurate Algorithms and Efficient Codes for Computational Aeroacoustics

    Science.gov (United States)

    Goodrich, John W.; Dyson, Rodger W.

    1999-01-01

    The simulation of sound generation and propagation in three space dimensions with realistic aircraft components is a very large time dependent computation with fine details. Simulations in open domains with embedded objects require accurate and robust algorithms for propagation, for artificial inflow and outflow boundaries, and for the definition of geometrically complex objects. The development, implementation, and validation of methods for solving these demanding problems is being done to support the NASA pillar goals for reducing aircraft noise levels. Our goal is to provide algorithms which are sufficiently accurate and efficient to produce usable results rapidly enough to allow design engineers to study the effects on sound levels of design changes in propulsion systems, and in the integration of propulsion systems with airframes. There is a lack of design tools for these purposes at this time. Our technical approach to this problem combines the development of new, algorithms with the use of Mathematica and Unix utilities to automate the algorithm development, code implementation, and validation. We use explicit methods to ensure effective implementation by domain decomposition for SPMD parallel computing. There are several orders of magnitude difference in the computational efficiencies of the algorithms which we have considered. We currently have new artificial inflow and outflow boundary conditions that are stable, accurate, and unobtrusive, with implementations that match the accuracy and efficiency of the propagation methods. The artificial numerical boundary treatments have been proven to have solutions which converge to the full open domain problems, so that the error from the boundary treatments can be driven as low as is required. The purpose of this paper is to briefly present a method for developing highly accurate algorithms for computational aeroacoustics, the use of computer automation in this process, and a brief survey of the algorithms that

  7. A Monomial Chaos Approach for Efficient Uncertainty Quantification in Computational Fluid Dynamics

    NARCIS (Netherlands)

    Witteveen, J.A.S.; Bijl, H.

    2006-01-01

    A monomial chaos approach is proposed for efficient uncertainty quantification in nonlinear computational problems. Propagating uncertainty through nonlinear equations can still be computationally intensive for existing uncertainty quantification methods. It usually results in a set of nonlinear equ

  8. High-performance computing in seismology

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1996-09-01

    The scientific, technical, and economic importance of the issues discussed here presents a clear agenda for future research in computational seismology. In this way these problems will drive advances in high-performance computing in the field of seismology. There is a broad community that will benefit from this work, including the petroleum industry, research geophysicists, engineers concerned with seismic hazard mitigation, and governments charged with enforcing a comprehensive test ban treaty. These advances may also lead to new applications for seismological research. The recent application of high-resolution seismic imaging of the shallow subsurface for the environmental remediation industry is an example of this activity. This report makes the following recommendations: (1) focused efforts to develop validated documented software for seismological computations should be supported, with special emphasis on scalable algorithms for parallel processors; (2) the education of seismologists in high-performance computing technologies and methodologies should be improved; (3) collaborations between seismologists and computational scientists and engineers should be increased; (4) the infrastructure for archiving, disseminating, and processing large volumes of seismological data should be improved.

  9. White LED with High Package Extraction Efficiency

    Energy Technology Data Exchange (ETDEWEB)

    Yi Zheng; Matthew Stough

    2008-09-30

    The goal of this project is to develop a high efficiency phosphor converting (white) Light Emitting Diode (pcLED) 1-Watt package through an increase in package extraction efficiency. A transparent/translucent monolithic phosphor is proposed to replace the powdered phosphor to reduce the scattering caused by phosphor particles. Additionally, a multi-layer thin film selectively reflecting filter is proposed between blue LED die and phosphor layer to recover inward yellow emission. At the end of the project we expect to recycle approximately 50% of the unrecovered backward light in current package construction, and develop a pcLED device with 80 lm/W{sub e} using our technology improvements and commercially available chip/package source. The success of the project will benefit luminous efficacy of white LEDs by increasing package extraction efficiency. In most phosphor-converting white LEDs, the white color is obtained by combining a blue LED die (or chip) with a powdered phosphor layer. The phosphor partially absorbs the blue light from the LED die and converts it into a broad green-yellow emission. The mixture of the transmitted blue light and green-yellow light emerging gives white light. There are two major drawbacks for current pcLEDs in terms of package extraction efficiency. The first is light scattering caused by phosphor particles. When the blue photons from the chip strike the phosphor particles, some blue light will be scattered by phosphor particles. Converted yellow emission photons are also scattered. A portion of scattered light is in the backward direction toward the die. The amount of this backward light varies and depends in part on the particle size of phosphors. The other drawback is that yellow emission from phosphor powders is isotropic. Although some backward light can be recovered by the reflector in current LED packages, there is still a portion of backward light that will be absorbed inside the package and further converted to heat. Heat

  10. Tailored Materials for High Efficiency CIDI Engines

    Energy Technology Data Exchange (ETDEWEB)

    Grant, G.J.; Jana, S.

    2012-03-30

    The overall goal of the project, Tailored Materials for High Efficiency Compression Ignition Direct Injection (CIDI) Engines, is to enable the implementation of new combustion strategies, such as homogeneous charge compression ignition (HCCI), that have the potential to significantly increase the energy efficiency of current diesel engines and decrease fuel consumption and environmental emissions. These strategies, however, are increasing the demands on conventional engine materials, either from increases in peak cylinder pressure (PCP) or from increases in the temperature of operation. The specific objective of this project is to investigate the application of a new material processing technology, friction stir processing (FSP), to improve the thermal and mechanical properties of engine components. The concept is to modify the surfaces of conventional, low-cost engine materials. The project focused primarily on FSP in aluminum materials that are compositional analogs to the typical piston and head alloys seen in small- to mid-sized CIDI engines. Investigations have been primarily of two types over the duration of this project: (1) FSP of a cast hypoeutectic Al-Si-Mg (A356/357) alloy with no introduction of any new components, and (2) FSP of Al-Cu-Ni alloys (Alloy 339) by physically stirring-in various quantities of carbon nanotubes/nanofibers or carbon fibers. Experimental work to date on aluminum systems has shown significant increases in fatigue lifetime and stress-level performance in aluminum-silicon alloys using friction processing alone, but work to demonstrate the addition of carbon nanotubes and fibers into aluminum substrates has shown mixed results due primarily to the difficulty in achieving porosity-free, homogeneous distributions of the particulate. A limited effort to understand the effects of FSP on steel materials was also undertaken during the course of this project. Processed regions were created in high-strength, low-alloyed steels up to 0.5 in

  11. High efficiency diffusion molecular retention tumor targeting.

    Directory of Open Access Journals (Sweden)

    Yanyan Guo

    Full Text Available Here we introduce diffusion molecular retention (DMR tumor targeting, a technique that employs PEG-fluorochrome shielded probes that, after a peritumoral (PT injection, undergo slow vascular uptake and extensive interstitial diffusion, with tumor retention only through integrin molecular recognition. To demonstrate DMR, RGD (integrin binding and RAD (control probes were synthesized bearing DOTA (for (111 In(3+, a NIR fluorochrome, and 5 kDa PEG that endows probes with a protein-like volume of 25 kDa and decreases non-specific interactions. With a GFP-BT-20 breast carcinoma model, tumor targeting by the DMR or i.v. methods was assessed by surface fluorescence, biodistribution of [(111In] RGD and [(111In] RAD probes, and whole animal SPECT. After a PT injection, both probes rapidly diffused through the normal and tumor interstitium, with retention of the RGD probe due to integrin interactions. With PT injection and the [(111In] RGD probe, SPECT indicated a highly tumor specific uptake at 24 h post injection, with 352%ID/g tumor obtained by DMR (vs 4.14%ID/g by i.v.. The high efficiency molecular targeting of DMR employed low probe doses (e.g. 25 ng as RGD peptide, which minimizes toxicity risks and facilitates clinical translation. DMR applications include the delivery of fluorochromes for intraoperative tumor margin delineation, the delivery of radioisotopes (e.g. toxic, short range alpha emitters for radiotherapy, or the delivery of photosensitizers to tumors accessible to light.

  12. High collection efficiency CVD diamond alpha detectors

    Energy Technology Data Exchange (ETDEWEB)

    Bergonzo, P.; Foulon, F.; Marshall, R.D.; Jany, C.; Brambilla, A. [CEA/Saclay, Gif-sur-Yvette (France); McKeag, R.D.; Jackman, R.B. [University College London (United Kingdom). Electronic and Electrical Engineering Dept.

    1998-06-01

    Advances in Chemical Vapor Deposited (CVD) diamond have enabled the routine use of this material for sensor device fabrication, allowing exploitation of its unique combination of physical properties (low temperature susceptibility (> 500 C), high resistance to radiation damage (> 100 Mrad) and to corrosive media). A consequence of CVD diamond growth on silicon is the formation of polycrystalline films which has a profound influence on the physical and electronic properties with respect to those measured on monocrystalline diamond. The authors report the optimization of physical and geometrical device parameters for radiation detection in the counting mode. Sandwich and co-planar electrode geometries are tested and their performances evaluated with regard to the nature of the field profile and drift distances inherent in such devices. The carrier drift length before trapping was measured under alpha particles and values as high as 40% of the overall film thickness are reported. Further, by optimizing the device geometry, they show that a gain in collection efficiency, defined as the induced charge divided by the deposited charge within the material, can be achieved even though lower bias values are used.

  13. The green computing book tackling energy efficiency at large scale

    CERN Document Server

    Feng, Wu-chun

    2014-01-01

    Low-Power, Massively Parallel, Energy-Efficient Supercomputers The Blue Gene TeamCompiler-Driven Energy Efficiency Mahmut Kandemir and Shekhar Srikantaiah An Adaptive Run-Time System for Improving Energy Efficiency Chung-Hsing Hsu, Wu-chun Feng, and Stephen W. PooleEnergy-Efficient Multithreading through Run-Time Adaptation Exploring Trade-Offs between Energy Savings and Reliability in Storage Systems Ali R. Butt, Puranjoy Bhattacharjee, Guanying Wang, and Chris GniadyCross-Layer Power Management Zhikui Wang and Parthasarathy Ranganathan Energy-Efficient Virtualized Systems Ripal Nathuji and K

  14. Linear algebra on high-performance computers

    Energy Technology Data Exchange (ETDEWEB)

    Dongarra, J.J.; Sorensen, D.C.

    1986-01-01

    This paper surveys work recently done at Argonne National Laboratory in an attempt to discover ways to construct numerical software for high-performance computers. The numerical algorithms are taken from several areas of numerical linear algebra. We discuss certain architectural features of advanced-computer architectures that will affect the design of algorithms. The technique of restructuring algorithms in terms of certain modules is reviewed. This technique has proved successful in obtaining a high level of transportability without severe loss of performance on a wide variety of both vector and parallel computers. The module technique is demonstrably effective for dense linear algebra problems. However, in the case of sparse and structured problems it may be difficult to identify general modules that will be as effective. New algorithms have been devised for certain problems in this category. We present examples in three important areas: banded systems, sparse QR factorization, and symmetric eigenvalue problems. 32 refs., 10 figs., 6 tabs.

  15. Efficient quantum computation in a network with probabilistic gates and logical encoding

    DEFF Research Database (Denmark)

    Borregaard, J.; Sørensen, A. S.; Cirac, J. I.

    2017-01-01

    An approach to efficient quantum computation with probabilistic gates is proposed and analyzed in both a local and nonlocal setting. It combines heralded gates previously studied for atom or atomlike qubits with logical encoding from linear optical quantum computation in order to perform high......-fidelity quantum gates across a quantum network. The error-detecting properties of the heralded operations ensure high fidelity while the encoding makes it possible to correct for failed attempts such that deterministic and high-quality gates can be achieved. Importantly, this is robust to photon loss, which...... is typically the main obstacle to photonic-based quantum information processing. Overall this approach opens a path toward quantum networks with atomic nodes and photonic links....

  16. The computational optimization of heat exchange efficiency in stack chimneys

    Energy Technology Data Exchange (ETDEWEB)

    Van Goch, T.A.J.

    2012-02-15

    For many industrial processes, the chimney is the final step before hot fumes, with high thermal energy content, are discharged into the atmosphere. Tapping into this energy and utilizing it for heating or cooling applications, could improve sustainability, efficiency and/or reduce operational costs. Alternatively, an unused chimney, like the monumental chimney at the Eindhoven University of Technology, could serve as an 'energy channeler' once more; it can enhance free cooling by exploiting the stack effect. This study aims to identify design parameters that influence annual heat exchange in such stack chimney applications and optimize these parameters for specific scenarios to maximize the performance. Performance is defined by annual heat exchange, system efficiency and costs. The energy required for the water pump as compared to the energy exchanged, defines the system efficiency, which is expressed in an efficiency coefficient (EC). This study is an example of applying building performance simulation (BPS) tools for decision support in the early phase of the design process. In this study, BPS tools are used to provide design guidance, performance evaluation and optimization. A general method for optimization of simulation models will be studied, and applied in two case studies with different applications (heating/cooling), namely; (1) CERES case: 'Eindhoven University of Technology monumental stack chimney equipped with a heat exchanger, rejects heat to load the cold source of the aquifer system on the campus of the university and/or provides free cooling to the CERES building'; and (2) Industrial case: 'Heat exchanger in an industrial stack chimney, which recoups heat for use in e.g. absorption cooling'. The main research question, addressing the concerns of both cases, is expressed as follows: 'what is the optimal set of design parameters so heat exchange in stack chimneys is optimized annually for the cases in which a

  17. Numerics of High Performance Computers and Benchmark Evaluation of Distributed Memory Computers

    Directory of Open Access Journals (Sweden)

    H. S. Krishna

    2004-07-01

    Full Text Available The internal representation of numerical data, their speed of manipulation to generate the desired result through efficient utilisation of central processing unit, memory, and communication links are essential steps of all high performance scientific computations. Machine parameters, in particular, reveal accuracy and error bounds of computation, required for performance tuning of codes. This paper reports diagnosis of machine parameters, measurement of computing power of several workstations, serial and parallel computers, and a component-wise test procedure for distributed memory computers. Hierarchical memory structure is illustrated by block copying and unrolling techniques. Locality of reference for cache reuse of data is amply demonstrated by fast Fourier transform codes. Cache and register-blocking technique results in their optimum utilisation with consequent gain in throughput during vector-matrix operations. Implementation of these memory management techniques reduces cache inefficiency loss, which is known to be proportional to the number of processors. Of the two Linux clusters-ANUP16, HPC22 and HPC64, it has been found from the measurement of intrinsic parameters and from application benchmark of multi-block Euler code test run that ANUP16 is suitable for problems that exhibit fine-grained parallelism. The delivered performance of ANUP16 is of immense utility for developing high-end PC clusters like HPC64 and customised parallel computers with added advantage of speed and high degree of parallelism.

  18. 高效使用电脑,提高工作效率%Using Computer Efficiently and Improving Working Efficiency

    Institute of Scientific and Technical Information of China (English)

    王伟

    2012-01-01

    文章结合实际工作,提出了高效使用电脑提高工作效率的方式方法.%The method that how to use computer efficiently to improve working efficiency is presented integrated with practice.

  19. High bandgap III-V alloys for high efficiency optoelectronics

    Science.gov (United States)

    Alberi, Kirstin; Mascarenhas, Angelo; Wanlass, Mark

    2017-01-10

    High bandgap alloys for high efficiency optoelectronics are disclosed. An exemplary optoelectronic device may include a substrate, at least one Al.sub.1-xIn.sub.xP layer, and a step-grade buffer between the substrate and at least one Al.sub.1-xIn.sub.xP layer. The buffer may begin with a layer that is substantially lattice matched to GaAs, and may then incrementally increase the lattice constant in each sequential layer until a predetermined lattice constant of Al.sub.1-xIn.sub.xP is reached.

  20. High bandgap III-V alloys for high efficiency optoelectronics

    Energy Technology Data Exchange (ETDEWEB)

    Alberi, Kirstin; Mascarenhas, Angelo; Wanlass, Mark

    2017-01-10

    High bandgap alloys for high efficiency optoelectronics are disclosed. An exemplary optoelectronic device may include a substrate, at least one Al.sub.1-xIn.sub.xP layer, and a step-grade buffer between the substrate and at least one Al.sub.1-xIn.sub.xP layer. The buffer may begin with a layer that is substantially lattice matched to GaAs, and may then incrementally increase the lattice constant in each sequential layer until a predetermined lattice constant of Al.sub.1-xIn.sub.xP is reached.

  1. High performance computing on vector systems

    CERN Document Server

    Roller, Sabine

    2008-01-01

    Presents the developments in high-performance computing and simulation on modern supercomputer architectures. This book covers trends in hardware and software development in general and specifically the vector-based systems and heterogeneous architectures. It presents innovative fields like coupled multi-physics or multi-scale simulations.

  2. Efficient Quantification of Uncertainties in Complex Computer Code Results Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Propagation of parameter uncertainties through large computer models can be very resource intensive. Frameworks and tools for uncertainty quantification are...

  3. High-performance computers for unmanned vehicles

    Science.gov (United States)

    Toms, David; Ettinger, Gil J.

    2005-10-01

    The present trend of increasing functionality onboard unmanned vehicles is made possible by rapid advances in high-performance computers (HPCs). An HPC is characterized by very high computational capability (100s of billions of operations per second) contained in lightweight, rugged, low-power packages. HPCs are critical to the processing of sensor data onboard these vehicles. Operations such as radar image formation, target tracking, target recognition, signal intelligence signature collection and analysis, electro-optic image compression, and onboard data exploitation are provided by these machines. The net effect of an HPC is to minimize communication bandwidth requirements and maximize mission flexibility. This paper focuses on new and emerging technologies in the HPC market. Emerging capabilities include new lightweight, low-power computing systems: multi-mission computing (using a common computer to support several sensors); onboard data exploitation; and large image data storage capacities. These new capabilities will enable an entirely new generation of deployed capabilities at reduced cost. New software tools and architectures available to unmanned vehicle developers will enable them to rapidly develop optimum solutions with maximum productivity and return on investment. These new technologies effectively open the trade space for unmanned vehicle designers.

  4. ABOUT THE SUITABILITY OF CLOUDS IN HIGH-PERFORMANCE COMPUTING

    Directory of Open Access Journals (Sweden)

    Harald Richter

    2016-01-01

    Full Text Available Cloud computing has become the ubiquitous computing and storage paradigm. It is also attractive for scientists, because they do not have to care any more for their own IT infrastructure, but can outsource it to a Cloud Service Provider of their choice. However, for the case of High-Performance Computing (HPC in a cloud, as it is needed in simulations or for Big Data analysis, things are getting more intricate, because HPC codes must stay highly efficient, even when executed by many virtual cores (vCPUs. Older clouds or new standard clouds can fulfil this only under special precautions, which are given in this article. The results can be extrapolated to other cloud OSes than OpenStack and to other codes than OpenFOAM, which were used as examples.

  5. Series-Tuned High Efficiency RF-Power Amplifiers

    DEFF Research Database (Denmark)

    Vidkjær, Jens

    2008-01-01

    An approach to high efficiency RF-power amplifier design is presented. It addresses simultaneously efficiency optimization and peak voltage limitations when transistors are pushed towards their power limits.......An approach to high efficiency RF-power amplifier design is presented. It addresses simultaneously efficiency optimization and peak voltage limitations when transistors are pushed towards their power limits....

  6. High-Degree Neurons Feed Cortical Computations.

    Directory of Open Access Journals (Sweden)

    Nicholas M Timme

    2016-05-01

    Full Text Available Recent work has shown that functional connectivity among cortical neurons is highly varied, with a small percentage of neurons having many more connections than others. Also, recent theoretical developments now make it possible to quantify how neurons modify information from the connections they receive. Therefore, it is now possible to investigate how information modification, or computation, depends on the number of connections a neuron receives (in-degree or sends out (out-degree. To do this, we recorded the simultaneous spiking activity of hundreds of neurons in cortico-hippocampal slice cultures using a high-density 512-electrode array. This preparation and recording method combination produced large numbers of neurons recorded at temporal and spatial resolutions that are not currently available in any in vivo recording system. We utilized transfer entropy (a well-established method for detecting linear and nonlinear interactions in time series and the partial information decomposition (a powerful, recently developed tool for dissecting multivariate information processing into distinct parts to quantify computation between neurons where information flows converged. We found that computations did not occur equally in all neurons throughout the networks. Surprisingly, neurons that computed large amounts of information tended to receive connections from high out-degree neurons. However, the in-degree of a neuron was not related to the amount of information it computed. To gain insight into these findings, we developed a simple feedforward network model. We found that a degree-modified Hebbian wiring rule best reproduced the pattern of computation and degree correlation results seen in the real data. Interestingly, this rule also maximized signal propagation in the presence of network-wide correlations, suggesting a mechanism by which cortex could deal with common random background input. These are the first results to show that the extent to

  7. Efficient Quantification of Uncertainties in Complex Computer Code Results Project

    Data.gov (United States)

    National Aeronautics and Space Administration — This proposal addresses methods for efficient quantification of margins and uncertainties (QMU) for models that couple multiple, large-scale commercial or...

  8. Experiences with Efficient Methodologies for Teaching Computer Programming to Geoscientists

    Science.gov (United States)

    Jacobs, Christian T.; Gorman, Gerard J.; Rees, Huw E.; Craig, Lorraine E.

    2016-01-01

    Computer programming was once thought of as a skill required only by professional software developers. But today, given the ubiquitous nature of computation and data science it is quickly becoming necessary for all scientists and engineers to have at least a basic knowledge of how to program. Teaching how to program, particularly to those students…

  9. An efficient algorithm for computing the H-infinity norm

    NARCIS (Netherlands)

    Belur, Madhu N.; Praagman, C.

    2011-01-01

    This technical note addresses the computation of the H-infinity norm by directly computing the isolated common zeros of two bivariate polynomials, unlike the iteration algorithm that is currently used to find the H-infinity norm. The proposed method to H-infinity norm calculation is compared with th

  10. Efficient computation of steady, 3D water-wave patterns

    NARCIS (Netherlands)

    Lewis, M.R.; Koren, B.

    2003-01-01

    Numerical methods for the computation of stationary free surfaces is the subject of much current research in computational engineering. The present report is directed towards free surfaces in maritime engineering. Of interest here are the long steady waves generated by ships, the gravity waves. In t

  11. High efficient wind-hydrogen facility

    Energy Technology Data Exchange (ETDEWEB)

    Bolcich, J.C. [Centro Atomico Bariloche, San Carlos de Bariloche, Rio Negro (Argentina); Bolcich, A.; Bolcich, D.; Wandyk, N. [ZVALT S.R.L., San Carlos de Bariloche, Rio Negro (Argentina)

    2001-06-01

    Considered a high efficiency and economical option for the conversion of wind energy, a new concept was developed in Patagonia, Argentina. It is called the Ducted Multirotor-Wind Energy Converter (DM-WEC) and was designed to be used in remote areas. Under study is the first prototype (10-30 Kilowatts), while the second prototype (3-5 Kilowatt) is being installed and is undergoing testing. This second prototype is of the multirotor type with the duct for air canalization as an added alternative. In this presentation, the authors described the technical attributes of the two prototypes and provided an evaluation of the potential for hydrogen production using small electrolysis units. They also included a comparison between the ducted and non-ducted prototypes with regard to maximum power attainable. The results obtained so far indicated that it represents an affordable energy source for remote areas with a higher power per unit cross section swept area. The rotating parts are encapsulated, reducing the noise level and vibrations. figs.

  12. A high-efficiency aerothermoelastic analysis method

    Science.gov (United States)

    Wan, ZhiQiang; Wang, YaoKun; Liu, YunZhen; Yang, Chao

    2014-06-01

    In this paper, a high-efficiency aerothermoelastic analysis method based on unified hypersonic lifting surface theory is established. The method adopts a two-way coupling form that couples the structure, aerodynamic force, and aerodynamic thermo and heat conduction. The aerodynamic force is first calculated based on unified hypersonic lifting surface theory, and then the Eckert reference temperature method is used to solve the temperature field, where the transient heat conduction is solved using Fourier's law, and the modal method is used for the aeroelastic correction. Finally, flutter is analyzed based on the p-k method. The aerothermoelastic behavior of a typical hypersonic low-aspect ratio wing is then analyzed, and the results indicate the following: (1) the combined effects of the aerodynamic load and thermal load both deform the wing, which would increase if the flexibility, size, and flight time of the hypersonic aircraft increase; (2) the effect of heat accumulation should be noted, and therefore, the trajectory parameters should be considered in the design of hypersonic flight vehicles to avoid hazardous conditions, such as flutter.

  13. Highly Efficient Contactless Electrical Energy Transmission System

    Science.gov (United States)

    Ayano, Hideki; Nagase, Hiroshi; Inaba, Hiromi

    This paper proposes a new concept for a contactless electrical energy transmission system for an elevator and an automated guided vehicle. The system has rechargeable batteries on the car and electrical energy is supplied at a specific place. When electric power is supplied to the car, it runs automatically and approaches the battery charger. Therefore, a comparatively large gap is needed between the primary transformer at the battery charger and the secondary transformer on the car in order to prevent damage which would be caused by a collision. In this case, a drop of the transformer coupling rate due to the large gap must be prevented. In conventional contactless electrical energy transmission technology, since electric power is received by a pick-up coil from a power line, a large-sized transformer is required. And when the distance over which the car runs is long, the copper loss of the line also increases. The developed system adopts a high frequency inverter using a soft switching method to miniaturize the transformer. The system has a coupling rate of 0.88 for a transformer gap length of 10mm and can operate at 91% efficiency.

  14. High performance computing and communications panel report

    Energy Technology Data Exchange (ETDEWEB)

    1992-12-01

    In FY92, a presidential initiative entitled High Performance Computing and Communications (HPCC) was launched, aimed at securing U.S. preeminence in high performance computing and related communication technologies. The stated goal of the initiative is threefold: extend U.S. technological leadership in high performance computing and computer communications; provide wide dissemination and application of the technologies; and spur gains in U.S. productivity and industrial competitiveness, all within the context of the mission needs of federal agencies. Because of the importance of the HPCC program to the national well-being, especially its potential implication for industrial competitiveness, the Assistant to the President for Science and Technology has asked that the President's Council of Advisors in Science and Technology (PCAST) establish a panel to advise PCAST on the strengths and weaknesses of the HPCC program. The report presents a program analysis based on strategy, balance, management, and vision. Both constructive recommendations for program improvement and positive reinforcement of successful program elements are contained within the report.

  15. Efficient Backprojection-Based Synthetic Aperture Radar Computation with Many-Core Processors

    Directory of Open Access Journals (Sweden)

    Jongsoo Park

    2013-01-01

    Full Text Available Tackling computationally challenging problems with high efficiency often requires the combination of algorithmic innovation, advanced architecture, and thorough exploitation of parallelism. We demonstrate this synergy through synthetic aperture radar (SAR via backprojection, an image reconstruction method that can require hundreds of TFLOPS. Computation cost is significantly reduced by our new algorithm of approximate strength reduction; data movement cost is economized by software locality optimizations facilitated by advanced architecture support; parallelism is fully harnessed in various patterns and granularities. We deliver over 35 billion backprojections per second throughput per compute node on an Intel® Xeon® processor E5-2670-based cluster, equipped with Intel® Xeon Phi™ coprocessors. This corresponds to processing a 3K×3K image within a second using a single node. Our study can be extended to other settings: backprojection is applicable elsewhere including medical imaging, approximate strength reduction is a general code transformation technique, and many-core processors are emerging as a solution to energy-efficient computing.

  16. Efficient computation method for two-dimensional nonlinear waves

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    The theory and simulation of fully-nonlinear waves in a truncated two-dimensional wave tank in time domain are presented. A piston-type wave-maker is used to generate gravity waves into the tank field in finite water depth. A damping zone is added in front of the wave-maker which makes it become one kind of absorbing wave-maker and ensures the prescribed Neumann condition. The efficiency of nmerical tank is further enhanced by installation of a sponge layer beach (SLB) in front of downtank to absorb longer weak waves that leak through the entire wave train front. Assume potential flow, the space- periodic irrotational surface waves can be represented by mixed Euler- Lagrange particles. Solving the integral equation at each time step for new normal velocities, the instantaneous free surface is integrated following time history by use of fourth-order Runge- Kutta method. The double node technique is used to deal with geometric discontinuity at the wave- body intersections. Several precise smoothing methods have been introduced to treat surface point with high curvature. No saw-tooth like instability is observed during the total simulation.The advantage of proposed wave tank has been verified by comparing with linear theoretical solution and other nonlinear results, excellent agreement in the whole range of frequencies of interest has been obtained.

  17. High-Precision Computation and Mathematical Physics

    Energy Technology Data Exchange (ETDEWEB)

    Bailey, David H.; Borwein, Jonathan M.

    2008-11-03

    At the present time, IEEE 64-bit floating-point arithmetic is sufficiently accurate for most scientific applications. However, for a rapidly growing body of important scientific computing applications, a higher level of numeric precision is required. Such calculations are facilitated by high-precision software packages that include high-level language translation modules to minimize the conversion effort. This paper presents a survey of recent applications of these techniques and provides some analysis of their numerical requirements. These applications include supernova simulations, climate modeling, planetary orbit calculations, Coulomb n-body atomic systems, scattering amplitudes of quarks, gluons and bosons, nonlinear oscillator theory, Ising theory, quantum field theory and experimental mathematics. We conclude that high-precision arithmetic facilities are now an indispensable component of a modern large-scale scientific computing environment.

  18. Energy efficient computing exploiting the properties of light

    Science.gov (United States)

    Shamir, Joseph

    2013-09-01

    Significant reduction of energy dissipation in computing can be achieved by addressing the theoretical lower limit of energy consumption and replacing arrays of traditional Boolean logic gates by other methods of implementing logic operations. In particular, a slight modification of the concept of computing allows the incorporation of fundamentally lossless optical processes as part of the computing operation. While the introduced new concepts can be implemented electronically or by other means, using optics eliminates also energy dissipation involved in the translation of electric charges. A possible realization of the indicated concepts is based on directed logic networks composed of reversible optical logic gate arrays.

  19. Efficient Data-parallel Computations on Distributed Systems

    Institute of Scientific and Technical Information of China (English)

    2002-01-01

    Task scheduling determines the performance of NOW computing to a large extent.However,the computer system architecture, computing capability and sys tem load are rarely proposed together.In this paper,a biggest-heterogeneous scheduling algorithm is presented.It fully considers the system characterist ics (from application view), structure and state.So it always can utilize all processing resource under a reasonable premise.The results of experiment show the algorithm can significantly shorten the response time of jobs.

  20. Highly Efficient Boundary Element Analysis of Whispering Gallery Microcavities

    CERN Document Server

    Pan, Leyuan

    2014-01-01

    We demonstrate that the efficiency of the boundary element whispering gallery microcavity analysis can be improved by orders of magnitude with the inclusion of Fresnel approximation. Using this formulation, simulation of a microdisk with wave-number-radius product as large as $kR\\approx8,000$ was demonstrated in contrast to a previous record of $kR\\approx100$. In addition to its high accuracy on computing the modal field distribution and resonance wavelength, this method yields a relative error of $10%$ in calculating the quality factor as high as $10^{11}$ through a direct root searching method where the conventional boundary element method failed to achieve. Finally, quadrupole shaped cavities and double disks as large as $100 {\\mu}m$ in diameter were modeled by employing as few as $512$ boundary elements whilst the simulation of such large cavities using conventional boundary element method were not reported previously.

  1. Towards high efficiency segmented thermoelectric unicouples

    DEFF Research Database (Denmark)

    Pham, Hoang Ngan; Christensen, Dennis Valbjørn; Snyder, Gerald Jeffrey

    2014-01-01

    Segmentation of thermoelectric (TE) materials is a widely used solution to improve the efficiency of thermoelectric generators over a wide working temperature range. However, the improvement can only be obtained with appropriate material selections. In this work, we provide an overview...... of the theoretical efficiency of the best performing unicouples designed from segmenting the state-of-the-art TE materials. The efficiencies are evaluated using a 1D numerical model which includes all thermoelectric effects, heat conduction, Joule effects and temperature dependent material properties, but neglects...

  2. High efficiency quasi-monochromatic infrared emitter

    Science.gov (United States)

    Brucoli, Giovanni; Bouchon, Patrick; Haïdar, Riad; Besbes, Mondher; Benisty, Henri; Greffet, Jean-Jacques

    2014-02-01

    Incandescent radiation sources are widely used as mid-infrared emitters owing to the lack of alternative for compact and low cost sources. A drawback of miniature hot systems such as membranes is their low efficiency, e.g., for battery powered systems. For targeted narrow-band applications such as gas spectroscopy, the efficiency is even lower. In this paper, we introduce design rules valid for very generic membranes demonstrating that their energy efficiency for use as incandescent infrared sources can be increased by two orders of magnitude.

  3. PREFACE: High Performance Computing Symposium 2011

    Science.gov (United States)

    Talon, Suzanne; Mousseau, Normand; Peslherbe, Gilles; Bertrand, François; Gauthier, Pierre; Kadem, Lyes; Moitessier, Nicolas; Rouleau, Guy; Wittig, Rod

    2012-02-01

    HPCS (High Performance Computing Symposium) is a multidisciplinary conference that focuses on research involving High Performance Computing and its application. Attended by Canadian and international experts and renowned researchers in the sciences, all areas of engineering, the applied sciences, medicine and life sciences, mathematics, the humanities and social sciences, it is Canada's pre-eminent forum for HPC. The 25th edition was held in Montréal, at the Université du Québec à Montréal, from 15-17 June and focused on HPC in Medical Science. The conference was preceded by tutorials held at Concordia University, where 56 participants learned about HPC best practices, GPU computing, parallel computing, debugging and a number of high-level languages. 274 participants from six countries attended the main conference, which involved 11 invited and 37 contributed oral presentations, 33 posters, and an exhibit hall with 16 booths from our sponsors. The work that follows is a collection of papers presented at the conference covering HPC topics ranging from computer science to bioinformatics. They are divided here into four sections: HPC in Engineering, Physics and Materials Science, HPC in Medical Science, HPC Enabling to Explore our World and New Algorithms for HPC. We would once more like to thank the participants and invited speakers, the members of the Scientific Committee, the referees who spent time reviewing the papers and our invaluable sponsors. To hear the invited talks and learn about 25 years of HPC development in Canada visit the Symposium website: http://2011.hpcs.ca/lang/en/conference/keynote-speakers/ Enjoy the excellent papers that follow, and we look forward to seeing you in Vancouver for HPCS 2012! Gilles Peslherbe Chair of the Scientific Committee Normand Mousseau Co-Chair of HPCS 2011 Suzanne Talon Chair of the Organizing Committee UQAM Sponsors The PDF also contains photographs from the conference banquet.

  4. An efficient algorithm for nucleolus and prekernel computation in some classes of TU-games

    NARCIS (Netherlands)

    Faigle, U.; Kern, W.; Kuipers, J.

    1998-01-01

    We consider classes of TU-games. We show that we can efficiently compute an allocation in the intersection of the prekernel and the least core of the game if we can efficiently compute the minimum excess for any given allocation. In the case where the prekernel of the game contains exactly one core

  5. High Performance Computing in Science and Engineering '15 : Transactions of the High Performance Computing Center

    CERN Document Server

    Kröner, Dietmar; Resch, Michael

    2016-01-01

    This book presents the state-of-the-art in supercomputer simulation. It includes the latest findings from leading researchers using systems from the High Performance Computing Center Stuttgart (HLRS) in 2015. The reports cover all fields of computational science and engineering ranging from CFD to computational physics and from chemistry to computer science with a special emphasis on industrially relevant applications. Presenting findings of one of Europe’s leading systems, this volume covers a wide variety of applications that deliver a high level of sustained performance. The book covers the main methods in high-performance computing. Its outstanding results in achieving the best performance for production codes are of particular interest for both scientists and engineers. The book comes with a wealth of color illustrations and tables of results.

  6. High-efficiency Transformerless PV Inverter Circuits

    OpenAIRE

    Chen, Baifeng

    2015-01-01

    With worldwide growing demand for electric energy, there has been a great interest in exploring photovoltaic (PV) sources. For the PV generation system, the power converter is the most essential part for the efficiency and function performance. In recent years, there have been quite a few new transformerless PV inverters topologies, which eliminate the traditional line frequency transformers to achieve lower cost and higher efficiency, and maintain lower leakage current as well. With an ov...

  7. High efficiency silicon solar cell review

    Science.gov (United States)

    Godlewski, M. P. (Editor)

    1975-01-01

    An overview is presented of the current research and development efforts to improve the performance of the silicon solar cell. The 24 papers presented reviewed experimental and analytic modeling work which emphasizes the improvment of conversion efficiency and the reduction of manufacturing costs. A summary is given of the round-table discussion, in which the near- and far-term directions of future efficiency improvements were discussed.

  8. An efficient computational strategy for composite laminates assemblies including variability

    National Research Council Canada - National Science Library

    Roulet, V; Boucard, P.-A; Champaney, L

    2013-01-01

    The aim of this work is to present an efficient numerical strategy for studying the influence of the material parameters on problems involving 3D assemblies of composite parts with contact and friction...

  9. Efficient CUDA Polynomial Preconditioned Conjugate Gradient Solver for Finite Element Computation of Elasticity Problems

    Directory of Open Access Journals (Sweden)

    Jianfei Zhang

    2013-01-01

    Full Text Available Graphics processing unit (GPU has obtained great success in scientific computations for its tremendous computational horsepower and very high memory bandwidth. This paper discusses the efficient way to implement polynomial preconditioned conjugate gradient solver for the finite element computation of elasticity on NVIDIA GPUs using compute unified device architecture (CUDA. Sliced block ELLPACK (SBELL format is introduced to store sparse matrix arising from finite element discretization of elasticity with fewer padding zeros than traditional ELLPACK-based formats. Polynomial preconditioning methods have been investigated both in convergence and running time. From the overall performance, the least-squares (L-S polynomial method is chosen as a preconditioner in PCG solver to finite element equations derived from elasticity for its best results on different example meshes. In the PCG solver, mixed precision algorithm is used not only to reduce the overall computational, storage requirements and bandwidth but to make full use of the capacity of the GPU devices. With SBELL format and mixed precision algorithm, the GPU-based L-S preconditioned CG can get a speedup of about 7–9 to CPU-implementation.

  10. Efficient high-order analysis of bowtie nanoantennas using the locally corrected Nystrom method

    CERN Document Server

    Chorsi, Hamid T

    2016-01-01

    It is demonstrated that the Locally Corrected Nystrom (LCN) method is a versatile and numerically efficient computational method for the modeling of scattering from plasmonic bowtie nanoantennas. The LCN method is a high-order analysis method that can provide exponential convergence. It is straightforward to implement, accurate and computationally efficient. To the best of the authors knowledge, the high-order LCN is here applied for the first time to 3D nanostructures. Numerical results show the accuracy and efficiency of the LCN applied to the electromagnetic analysis of nanostructures.

  11. Parallel computation of seismic analysis of high arch dam

    Institute of Scientific and Technical Information of China (English)

    Chen Houqun; Ma Huaifa; Tu Jin; Cheng Guangqing; Tang Juzhen

    2008-01-01

    Parallel computation programs are developed for three-dimensional meso-mechanics analysis of fully-graded dam concrete and seismic response analysis of high arch dams (ADs), based on the Parallel Finite Element Program Generator (PFEPG). The computational algorithms of the numerical simulation of the meso-structure of concrete specimens were studied. Taking into account damage evolution, static preload, strain rate effect, and the heterogeneity of the meso-structure of dam concrete, the fracture processes of damage evolution and configuration of the cracks can be directly simulated. In the seismic response analysis of ADs, all the following factors are involved, such as the nonlinear contact due to the opening and slipping of the contraction joints, energy dispersion of the far-field foundation, dynamic interactions of the dam-foundation-reservoir system, and the combining effects of seismic action with all static loads. The correctness, reliability and efficiency of the two parallel computational programs are verified with practical illustrations.

  12. Limits on efficient computation in the physical world

    Science.gov (United States)

    Aaronson, Scott Joel

    More than a speculative technology, quantum computing seems to challenge our most basic intuitions about how the physical world should behave. In this thesis I show that, while some intuitions from classical computer science must be jettisoned in the light of modern physics, many others emerge nearly unscathed; and I use powerful tools from computational complexity theory to help determine which are which. In the first part of the thesis, I attack the common belief that quantum computing resembles classical exponential parallelism, by showing that quantum computers would face serious limitations on a wider range of problems than was previously known. In particular, any quantum algorithm that solves the collision problem---that of deciding whether a sequence of n integers is one-to-one or two-to-one---must query the sequence O (n1/5) times. This resolves a question that was open for years; previously no lower bound better than constant was known. A corollary is that there is no "black-box" quantum algorithm to break cryptographic hash functions or solve the Graph Isomorphism problem in polynomial time. I also show that relative to an oracle, quantum computers could not solve NP-complete problems in polynomial time, even with the help of nonuniform "quantum advice states"; and that any quantum algorithm needs O (2n/4/n) queries to find a local minimum of a black-box function on the n-dimensional hypercube. Surprisingly, the latter result also leads to new classical lower bounds for the local search problem. Finally, I give new lower bounds on quantum one-way communication complexity, and on the quantum query complexity of total Boolean functions and recursive Fourier sampling. The second part of the thesis studies the relationship of the quantum computing model to physical reality. I first examine the arguments of Leonid Levin, Stephen Wolfram, and others who believe quantum computing to be fundamentally impossible. I find their arguments unconvincing without a "Sure

  13. Efficient algorithm and computing tool for shading calculation

    Directory of Open Access Journals (Sweden)

    Chanadda Pongpattana

    2006-03-01

    Full Text Available The window is always part of a building envelope. It earns its respect in creating architectural elegance of a building. Despite a major advantage of daylight utilization, a window would inevitably allow heat from solar radiation to penetrate into a building. Hence, a window design must be performed under a careful consideration in order to achieve an energy-conscious design for which the daylight utilization and heat gain are optimized. This paper presents the validation of the vectorial formulation of shading calculation by comparing the computational results with experimental ones, overhang, fin, and eggcrate. A computational algorithm and interactive computer software for computing the shadow were developed. The software was designed in order to be user-friendly and capable of presenting profiles of the shadow graphically and computing corresponding shaded areas for a given window system. It was found that software simulation results were in excellent agreement with experimental results. The average percentage of error is approximately 0.25%, 0.52%, and 0.21% for overhang, fin, and eggcrate, respectively.

  14. An efficient numerical integral in three-dimensional electromagnetic field computations

    Science.gov (United States)

    Whetten, Frank L.; Liu, Kefeng; Balanis, Constantine A.

    1990-01-01

    An improved algorithm for efficiently computing a sinusoid and an exponential integral commonly encountered in method-of-moments solutions is presented. The new algorithm has been tested for accuracy and computer execution time against both numerical integration and other existing numerical algorithms, and has outperformed them. Typical execution time comparisons on several computers are given.

  15. Improving the Hydraulic Efficiency of Centrifugal Pumps through Computational Fluid Dynamics Based Design optimization

    Directory of Open Access Journals (Sweden)

    Abdellah Ait moussa

    2014-08-01

    Full Text Available The design and optimization of turbo machine impellers such as those in pumps and turbines is a highly complicated task due to the complex three-dimensional shape of the impeller blades and surrounding devices. Small differences in geometry can lead to significant changes in the performance of these machines. We report here an efficient numerical technique that automatically optimizes the geometry of these blades for maximum performance. The technique combines, mathematical modeling of the impeller blades using non-uniform rational B-spline (NURBS, Computational fluid dynamics (CFD with Geometry Parameterizations in turbulent flow simulation and the Globalized and bounded Nelder-Mead (GBNM algorithm in geometry optimization.

  16. An Efficient Role Specification Management Model for Highly Distributed Environments

    Directory of Open Access Journals (Sweden)

    Soomi Yang

    2006-07-01

    Full Text Available Highly distributed environments such as pervasive computing environments not having global or broad control, need another attribute certificate management technique. For an efficient role based access control using attribute certificate, we use a technique of structuring role specification certificates. It can provide more flexible and secure collaborating environments. The roles are grouped and made them into the relation tree. It can reduce management cost and overhead incurred when changing the specification of the role. Further we use caching of frequently used role specification certificate for better performance in case applying the role. Tree structured role specification results secure and efficient role renewing and distribution. Caching of role specification helps an application of role. In order to be scalable distribution of the role specification certificate, we use multicasting packets. Also, performance enhancement of structuring role specification certificates is quantified in the sense of taking into account of the packet loss. In the experimental section, it is shown that role updating and distribution are secured and efficient.

  17. Work-Efficient Parallel Skyline Computation for the GPU

    DEFF Research Database (Denmark)

    Bøgh, Kenneth Sejdenfaden; Chester, Sean; Assent, Ira

    2015-01-01

    The skyline operator returns records in a dataset that provide optimal trade-offs of multiple dimensions. State-of-the-art skyline computation involves complex tree traversals, data-ordering, and conditional branching to minimize the number of point-to-point comparisons. Meanwhile, GPGPU computing...... a global, static partitioning scheme. With the partitioning, we can permit controlled branching to exploit transitive relationships and avoid most point-to-point comparisons. The result is a non-traditional GPU algorithm, SkyAlign, that prioritizes work-effciency and respectable throughput, rather than...... maximal throughput, to achieve orders of magnitude faster performance....

  18. The Analysis of Distillation Tray Column Efficiency by Fluid Dynamics and Mass Transfer Computation

    Institute of Scientific and Technical Information of China (English)

    刘春江; 袁希钢; 余国琮

    2005-01-01

    It has long been found that the flow pattern of the liquid phase on distillation tray is of great importance on distillation process performance. But until now, there was very few published work on quantitative investigation of this subject. By combining the computational fluid dynamics (CFD) with the mass transfer equation, a theoretical model is proposed for predicting the details of velocity and concentration distributions as well as the tray efficiency of distillation tray column. Using the proposed model, four different cases corresponding to different assumptions of liquid and vapor flowing condition for a distillation tray column were investigated. In Case I, the distributions of velocity and concentration of the incoming liquid from the downcomer and the uprising vapor from the underneath tray spacing are uniform. In Case Ⅱ, the distribution of the incoming liquid is non-uniform but the uprising vapor is uniform. In Case Ⅲ, the distribution of the incoming liquid is uniform but the uprising vapor is non-uniform.In Case IV, the distributions of both the incoming liquid and the uprising vapor are non-uniform. The details of velocity and concentration distributions on a multiple sieve tray distillation column in four different cases were simulated using the proposed model. It is found that the shape of the simulated concentration profiles of vapor and the liquid is quite different from case to case. The computed results also show that the tray efficiency is highly reduced by the maldistribution of velocity and concentration of the incoming liquid and uprising vapor. The tray efficiency for Case Ⅰ is higher than Case Ⅱ or Case Ⅲ, and that for Case Ⅳis the lowest. It also reveals that the accumulated effect of maldistribution becomes more pronounced when the number of column trays increased. The present study demonstrates that the use of computational method to predict the mass transfer efficiency for the tray column, especially for the large one, is

  19. Designing high efficient solar powered lighting systems

    DEFF Research Database (Denmark)

    Poulsen, Peter Behrensdorff; Thorsteinsson, Sune; Lindén, Johannes;

    2016-01-01

    Some major challenges in the development of L2L products is the lack of efficient converter electronics, modelling tools for dimensioning and furthermore, characterization facilities to support the successful development of the products. We report the development of 2 Three-Port-Converters respec......Some major challenges in the development of L2L products is the lack of efficient converter electronics, modelling tools for dimensioning and furthermore, characterization facilities to support the successful development of the products. We report the development of 2 Three......-Port-Converters respectively for 1-10Wp and 10-50 Wp with a peak efficiency of 97% at 1.8 W of PV power for the 10 Wp version. Furthermore, a modelling tool for L2L products has been developed and a laboratory for feeding in component data not available in the datasheets to the model is described....

  20. Domain Decomposition Based High Performance Parallel Computing

    CERN Document Server

    Raju, Mandhapati P

    2009-01-01

    The study deals with the parallelization of finite element based Navier-Stokes codes using domain decomposition and state-ofart sparse direct solvers. There has been significant improvement in the performance of sparse direct solvers. Parallel sparse direct solvers are not found to exhibit good scalability. Hence, the parallelization of sparse direct solvers is done using domain decomposition techniques. A highly efficient sparse direct solver PARDISO is used in this study. The scalability of both Newton and modified Newton algorithms are tested.

  1. Probabilistic structural analysis algorithm development for computational efficiency

    Science.gov (United States)

    Wu, Y.-T.

    1991-01-01

    The PSAM (Probabilistic Structural Analysis Methods) program is developing a probabilistic structural risk assessment capability for the SSME components. An advanced probabilistic structural analysis software system, NESSUS (Numerical Evaluation of Stochastic Structures Under Stress), is being developed as part of the PSAM effort to accurately simulate stochastic structures operating under severe random loading conditions. One of the challenges in developing the NESSUS system is the development of the probabilistic algorithms that provide both efficiency and accuracy. The main probability algorithms developed and implemented in the NESSUS system are efficient, but approximate in nature. In the last six years, the algorithms have improved very significantly.

  2. An Efficient Virtual Machine Consolidation Scheme for Multimedia Cloud Computing.

    Science.gov (United States)

    Han, Guangjie; Que, Wenhui; Jia, Gangyong; Shu, Lei

    2016-02-18

    Cloud computing has innovated the IT industry in recent years, as it can delivery subscription-based services to users in the pay-as-you-go model. Meanwhile, multimedia cloud computing is emerging based on cloud computing to provide a variety of media services on the Internet. However, with the growing popularity of multimedia cloud computing, its large energy consumption cannot only contribute to greenhouse gas emissions, but also result in the rising of cloud users' costs. Therefore, the multimedia cloud providers should try to minimize its energy consumption as much as possible while satisfying the consumers' resource requirements and guaranteeing quality of service (QoS). In this paper, we have proposed a remaining utilization-aware (RUA) algorithm for virtual machine (VM) placement, and a power-aware algorithm (PA) is proposed to find proper hosts to shut down for energy saving. These two algorithms have been combined and applied to cloud data centers for completing the process of VM consolidation. Simulation results have shown that there exists a trade-off between the cloud data center's energy consumption and service-level agreement (SLA) violations. Besides, the RUA algorithm is able to deal with variable workload to prevent hosts from overloading after VM placement and to reduce the SLA violations dramatically.

  3. Efficient computation of exposure profiles for counterparty credit risk

    NARCIS (Netherlands)

    Graaf, C.S.L. de; Feng, Q.; Kandhai, B.D.; Oosterlee, C.W.

    2014-01-01

    Three computational techniques for approximation of counterparty exposure for financial derivatives are presented. The exposure can be used to quantify so-called Credit Valuation Adjustment (CVA) and Potential Future Exposure (PFE), which are of utmost importance for modern risk management in the fi

  4. Efficient Computation of Exposure Profiles for Counterparty Credit Risk

    NARCIS (Netherlands)

    de Graaf, C.S.L.; Feng, Q.; Kandhai, D.; Oosterlee, C.W.

    2014-01-01

    Three computational techniques for approximation of counterparty exposure for financial derivatives are presented. The exposure can be used to quantify so-called Credit Valuation Adjustment (CVA) and Potential Future Exposure (PFE), which are of utmost importance for modern risk management in the fi

  5. Efficient Computation of Exposure Profiles for Counterparty Credit Risk

    NARCIS (Netherlands)

    de Graaf, C.S.L.; Feng, Q.; Kandhai, D.; Oosterlee, C.W.

    2014-01-01

    Three computational techniques for approximation of counterparty exposure for financial derivatives are presented. The exposure can be used to quantify so-called Credit Valuation Adjustment (CVA) and Potential Future Exposure (PFE), which are of utmost importance for modern risk management in the

  6. Efficient computation of exposure profiles for counterparty credit risk

    NARCIS (Netherlands)

    C.S.L. de Graaf (Kees); Q. Feng (Qian); B.D. Kandhai; C.W. Oosterlee (Cornelis)

    2014-01-01

    htmlabstractThree computational techniques for approximation of counterparty exposure for financial derivatives are presented. The exposure can be used to quantify so-called Credit Valuation Adjustment (CVA) and Potential Future Exposure (PFE), which are of utmost importance for modern risk

  7. Efficient Adjoint Computation of Hybrid Systems of Differential Algebraic Equations with Applications in Power Systems

    Energy Technology Data Exchange (ETDEWEB)

    Abhyankar, Shrirang [Argonne National Lab. (ANL), Argonne, IL (United States); Anitescu, Mihai [Argonne National Lab. (ANL), Argonne, IL (United States); Constantinescu, Emil [Argonne National Lab. (ANL), Argonne, IL (United States); Zhang, Hong [Argonne National Lab. (ANL), Argonne, IL (United States)

    2016-03-31

    Sensitivity analysis is an important tool to describe power system dynamic behavior in response to parameter variations. It is a central component in preventive and corrective control applications. The existing approaches for sensitivity calculations, namely, finite-difference and forward sensitivity analysis, require a computational effort that increases linearly with the number of sensitivity parameters. In this work, we investigate, implement, and test a discrete adjoint sensitivity approach whose computational effort is effectively independent of the number of sensitivity parameters. The proposed approach is highly efficient for calculating trajectory sensitivities of larger systems and is consistent, within machine precision, with the function whose sensitivity we are seeking. This is an essential feature for use in optimization applications. Moreover, our approach includes a consistent treatment of systems with switching, such as DC exciters, by deriving and implementing the adjoint jump conditions that arise from state and time-dependent discontinuities. The accuracy and the computational efficiency of the proposed approach are demonstrated in comparison with the forward sensitivity analysis approach.

  8. Designing high efficient solar powered lighting systems

    DEFF Research Database (Denmark)

    Poulsen, Peter Behrensdorff; Thorsteinsson, Sune; Lindén, Johannes

    2016-01-01

    Some major challenges in the development of L2L products is the lack of efficient converter electronics, modelling tools for dimensioning and furthermore, characterization facilities to support the successful development of the products. We report the development of 2 Three-Port-Converters respec...

  9. Designing high efficient solar powered lighting systems

    DEFF Research Database (Denmark)

    Poulsen, Peter Behrensdorff; Thorsteinsson, Sune; Lindén, Johannes;

    2016-01-01

    Some major challenges in the development of L2L products is the lack of efficient converter electronics, modelling tools for dimensioning and furthermore, characterization facilities to support the successful development of the products. We report the development of 2 Three-Port-Converters respec...

  10. Designing high efficient solar powered lighting systems

    DEFF Research Database (Denmark)

    Poulsen, Peter Behrensdorff; Thorsteinsson, Sune; Lindén, Johannes;

    Some major challenges in the development of L2L products is the lack of efficient converter electronics, modelling tools for dimensioning and furthermore, characterization facilities to support the successful development of the products. We report the development of 2 Three-Port-Converters respec...

  11. Computationally efficient locally-recurrent neural networks for online signal processing

    CERN Document Server

    Hussain, A; Shim, I

    1999-01-01

    A general class of computationally efficient locally recurrent networks (CERN) is described for real-time adaptive signal processing. The structure of the CERN is based on linear-in-the- parameters single-hidden-layered feedforward neural networks such as the radial basis function (RBF) network, the Volterra neural network (VNN) and the functionally expanded neural network (FENN), adapted to employ local output feedback. The corresponding learning algorithms are derived and key structural and computational complexity comparisons are made between the CERN and conventional recurrent neural networks. Two case studies are performed involving the real- time adaptive nonlinear prediction of real-world chaotic, highly non- stationary laser time series and an actual speech signal, which show that a recurrent FENN based adaptive CERN predictor can significantly outperform the corresponding feedforward FENN and conventionally employed linear adaptive filtering models. (13 refs).

  12. Computationally efficient DOD and DOA estimation for bistatic MIMO radar with propagator method

    Science.gov (United States)

    Zhang, Xiaofei; Wu, Hailang; Li, Jianfeng; Xu, Dazhuan

    2012-09-01

    In this article, we consider a computationally efficient direction of departure and direction of arrival estimation problem for a bistatic multiple-input multiple-output (MIMO) radar. The computational loads of the propagator method (PM) can be significantly smaller since the PM does not require any eigenvalue decomposition of the cross correlation matrix and singular value decomposition of the received data. An improved PM algorithm is proposed to obtain automatically paired transmit and receive angle estimations in the MIMO radar. The proposed algorithm has very close angle estimation performance to conventional PM, which has a much higher complexity than our algorithm. For high signal-to-noise ratio, the proposed algorithm has very close angle estimation to estimation of signal parameters via rotational invariance technique algorithm. The variance of the estimation error and Cramér-Rao bound of angle estimation are derived. Simulation results verify the usefulness of our algorithm.

  13. AN EFFICIENT AND STABLE STRUCTURE PRESERVING ALGORITHM FOR COMPUTING THE EIGENVALUES OF A HAMILTONIAN MATRIX

    Institute of Scientific and Technical Information of China (English)

    闫庆友; 熊西文

    2002-01-01

    An efficient and stable structure preserving algorithm, which is a variant of the QR like (SR) algorithm due to Bunse-Gerstner and Mehrmann, is presented for computing the eigenvalues and stable invariant subspaces of a Hamiltonian matrix. In the algorithm two strategies are employed, one of which is called dis- unstabilization technique and the other is preprocessing technique. Together with them, a so-called ratio-reduction equation and a backtrack technique are introduced to avoid the instability and breakdown in the original algorithm. It is shown that the new algorithm can overcome the instability and breakdown at low cost. Numerical results have demonstrated that the algorithm is stable and can compute the eigenvalues to very high accuracy.

  14. Efficiency using computer simulation of Reverse Threshold Model Theory on assessing a “One Laptop Per Child” computer versus desktop computer

    Directory of Open Access Journals (Sweden)

    Supat Faarungsang

    2017-04-01

    Full Text Available The Reverse Threshold Model Theory (RTMT model was introduced based on limiting factor concepts, but its efficiency compared to the Conventional Model (CM has not been published. This investigation assessed the efficiency of RTMT compared to CM using computer simulation on the “One Laptop Per Child” computer and a desktop computer. Based on probability values, it was found that RTMT was more efficient than CM among eight treatment combinations and an earlier study verified that RTMT gives complete elimination of random error. Furthermore, RTMT has several advantages over CM and is therefore proposed to be applied to most research data.

  15. A Memory and Computation Efficient Sparse Level-Set Method

    NARCIS (Netherlands)

    Laan, Wladimir J. van der; Jalba, Andrei C.; Roerdink, Jos B.T.M.

    2011-01-01

    Since its introduction, the level set method has become the favorite technique for capturing and tracking moving interfaces, and found applications in a wide variety of scientific fields. In this paper we present efficient data structures and algorithms for tracking dynamic interfaces through the le

  16. A Memory and Computation Efficient Sparse Level-Set Method

    NARCIS (Netherlands)

    Laan, Wladimir J. van der; Jalba, Andrei C.; Roerdink, Jos B.T.M.

    Since its introduction, the level set method has become the favorite technique for capturing and tracking moving interfaces, and found applications in a wide variety of scientific fields. In this paper we present efficient data structures and algorithms for tracking dynamic interfaces through the

  17. Efficient protein engineering by combining computational design and directed evolution

    NARCIS (Netherlands)

    Floor, Robert

    2015-01-01

    Het gebruik van enzymatische katalyse kan een belangrijke verduurzaming van de chemie opleveren. Enzymen hebben vaak een hoge selectiviteit en activiteit, waardoor een grondstof efficiënt wordt omgezet in een product. Echter, een belangrijke beperking is dat enzymen vaak niet stabiel genoeg zijn om

  18. An Accurate and Computationally Efficient Model for Membrane-Type Circular-Symmetric Micro-Hotplates

    Directory of Open Access Journals (Sweden)

    Usman Khan

    2014-04-01

    Full Text Available Ideally, the design of high-performance micro-hotplates would require a large number of simulations because of the existence of many important design parameters as well as the possibly crucial effects of both spread and drift. However, the computational cost of FEM simulations, which are the only available tool for accurately predicting the temperature in micro-hotplates, is very high. As a result, micro-hotplate designers generally have no effective simulation-tools for the optimization. In order to circumvent these issues, here, we propose a model for practical circular-symmetric micro-hot-plates which takes advantage of modified Bessel functions, computationally efficient matrix-approach for considering the relevant boundary conditions, Taylor linearization for modeling the Joule heating and radiation losses, and external-region-segmentation strategy in order to accurately take into account radiation losses in the entire micro-hotplate. The proposed model is almost as accurate as FEM simulations and two to three orders of magnitude more computationally efficient (e.g., 45 s versus more than 8 h. The residual errors, which are mainly associated to the undesired heating in the electrical contacts, are small (e.g., few degrees Celsius for an 800 °C operating temperature and, for important analyses, almost constant. Therefore, we also introduce a computationally-easy single-FEM-compensation strategy in order to reduce the residual errors to about 1 °C. As illustrative examples of the power of our approach, we report the systematic investigation of a spread in the membrane thermal conductivity and of combined variations of both ambient and bulk temperatures. Our model enables a much faster characterization of micro-hotplates and, thus, a much more effective optimization prior to fabrication.

  19. A highly efficient 3D micromixer using soft PDMS bonding

    Science.gov (United States)

    Cha, Junghun; Kim, Jinseok; Ryu, Suk-Kyu; Park, Jungyul; Jeong, Yongwon; Park, Sewan; Park, Sukho; Kim, Hyeon Cheol; Chun, Kukjin

    2006-09-01

    This paper presents a novel highly efficient passive micromixer that employs diffusion for micromixing. Since conventional fabrication methods cannot form precise aligned microchannels, the realization of a complex 3D micromixer has been difficult. Here, we report a novel micromixer, named a chessboard mixer. In addition, a new polydimethylsiloxane (PDMS) bonding method was developed to produce the proposed mixer. The new PDMS bonding technique requires PDMS-to-PDMS bonding and the moldable flexibility of partially cured PDMS to form the structure. Accordingly, a two-step curing process was used to solve these problems. Adhesion control was also considered when forming the PDMS membranes. Complex 3D microchannels in the micromixer were aligned within 1 m using the proposed new bonding method. The presented micromixer could increase the mixing effect by expanding interfaces between mixing fluids. Thus, this mixer makes it possible to mix within a shorter distance than other pre-existing micromixers do. A simulation using computational fluid dynamics (CFD)-ACE software showed a highly efficient performance, and an experiment involving the mixing of NaOH and phenolphthalein confirmed the rapid mixing performance (<1400 µm).

  20. Computationally Efficient Implementation of Convolution-based Locally Adaptive Binarization Techniques

    OpenAIRE

    Mollah, Ayatullah Faruk; Basu, Subhadip; Nasipuri, Mita

    2012-01-01

    One of the most important steps of document image processing is binarization. The computational requirements of locally adaptive binarization techniques make them unsuitable for devices with limited computing facilities. In this paper, we have presented a computationally efficient implementation of convolution based locally adaptive binarization techniques keeping the performance comparable to the original implementation. The computational complexity has been reduced from O(W2N2) to O(WN2) wh...

  1. Efficient Model for Distributed Computing based on Smart Embedded Agent

    Directory of Open Access Journals (Sweden)

    Hassna Bensag

    2017-02-01

    Full Text Available Technological advances of embedded computing exposed humans to an increasing intrusion of computing in their day-to-day life (e.g. smart devices. Cooperation, autonomy, and mobility made the agent a promising mechanism for embedded devices. The work aims to present a new model of an embedded agent designed to be implemented in smart devices in order to achieve parallel tasks in a distribute environment. To validate the proposed model, a case study was developed for medical image segmentation using Cardiac Magnetic Resonance Image (MRI. In the first part of this paper, we focus on implementing the parallel algorithm of classification using C-means method in embedded systems. We propose then a new concept of distributed classification using multi-agent systems based on JADE and Raspberry PI 2 devices.

  2. Efficient Flow Control Scheme in Multimedia Cloud Computing

    Directory of Open Access Journals (Sweden)

    Jinsheng Tan

    2013-10-01

    Full Text Available As multimedia cloud computing involving a great deal of calculations about graphics, images, audio and video, which consume a lot of resources and is a key issue toward traffic control. The characteristics of traditional HTB serial determine the bottleneck of its processing speed. The author provides a kind of mechanism based on multi-core processors pipeline style and parallelization of HTB flow control, makes the improvement of the analysis and algorithm toward flow control, and finally carries out experimental testing. The results show that: compared to traditional flow control, the multi-core processors pipeline style and parallelization of HTB flow control not only has greatly improved on the processing power, but still maintained a good stability, so as to meet the multimedia cloud computing users and data scale

  3. Application of Green computing in Framing Energy Efficient Software Engineering

    Directory of Open Access Journals (Sweden)

    Aritra Mitra, Riya Basu, Avik Guha, Shalabh Agarwal, Asoke Nath

    2013-03-01

    Full Text Available Green computing and energy saving is now a veryimportant issue in Computer science andinformation technology. Dueto tremendousgrowth in information technology now the bigchallenge is how to minimize the power usage andhow to reduce the carbon foot print. Greencomputing is now a prime research area where thepeople are trying to minimize the carbon footprintand minimum usage of energy. To minimize theusage of energy there are two independentapproaches one is designing suitable hardware andthe second one is to redesign the softwaremethodology. In the present paper the authorshave tried to explore the software methodologiesand designs that can be used today to save energy.The authors have also tried to extend mobileplatform battery time as well as the various toolsthat support the development of energy-efficientsoftware.

  4. Computationally efficient statistical differential equation modeling using homogenization

    Science.gov (United States)

    Hooten, Mevin B.; Garlick, Martha J.; Powell, James A.

    2013-01-01

    Statistical models using partial differential equations (PDEs) to describe dynamically evolving natural systems are appearing in the scientific literature with some regularity in recent years. Often such studies seek to characterize the dynamics of temporal or spatio-temporal phenomena such as invasive species, consumer-resource interactions, community evolution, and resource selection. Specifically, in the spatial setting, data are often available at varying spatial and temporal scales. Additionally, the necessary numerical integration of a PDE may be computationally infeasible over the spatial support of interest. We present an approach to impose computationally advantageous changes of support in statistical implementations of PDE models and demonstrate its utility through simulation using a form of PDE known as “ecological diffusion.” We also apply a statistical ecological diffusion model to a data set involving the spread of mountain pine beetle (Dendroctonus ponderosae) in Idaho, USA.

  5. Computer Controlled High Precise,High Voltage Pules Generator

    Institute of Scientific and Technical Information of China (English)

    但果; 邹积岩; 丛吉远; 董恩源

    2003-01-01

    High precise, high voltage pulse generator made up of high-power IGBT and pulse transformers controlled by a computer are described. A simple main circuit topology employed in this pulse generator can reduce the cost meanwhile it still meets special requirements for pulsed electric fields (PEFs) in food process. The pulse generator utilizes a complex programmable logic device (CPLD) to generate trigger signals. Pulse-frequency, pulse-width and pulse-number are controlled via RS232 bus by a computer. The high voltage pulse generator well suits to the application for fluid food non-thermal effect in pulsed electric fields, for it can increase and decrease by the step length 1.

  6. Efficient computations of wave loads on offshore structures

    DEFF Research Database (Denmark)

    Paulsen, Bo Terp

    The present thesis considers numerical computations of fully nonlinear wave impacts on bottom mounted surface piercing circular cylinders at intermediate water depths. The aim of the thesis is to provide new knowledge regarding wave loads on foundations for offshore wind turbines. Hence, the dime......The present thesis considers numerical computations of fully nonlinear wave impacts on bottom mounted surface piercing circular cylinders at intermediate water depths. The aim of the thesis is to provide new knowledge regarding wave loads on foundations for offshore wind turbines. Hence...... is carefully validated against experimental measurements of regular-, irregular- and multi-directional irregular waves. The ability of the numerical model to accurately reproduce experiments is also investigated. Wave impacts on a bottom mounted circular cylinder from steep regular waves are presented. Here......, the inline forces and the motion of the free surface is described as a function of the non-dimensional wave steepness, the relative water depth, the relative cylinder diameter and a co-existing current. From the computations, higher harmonic forces are determined and compared against the Morison equation...

  7. Achieving High Performance Distributed System: Using Grid, Cluster and Cloud Computing

    Directory of Open Access Journals (Sweden)

    Sunil Kr Singh

    2015-02-01

    Full Text Available To increase the efficiency of any task, we require a system that would provide high performance along with flexibilities and cost efficiencies for user. Distributed computing, as we are all aware, has become very popular over the past decade. Distributed computing has three major types, namely, cluster, grid and cloud. In order to develop a high performance distributed system, we need to utilize all the above mentioned three types of computing. In this paper, we shall first have an introduction of all the three types of distributed computing. Subsequently examining them we shall explore trends in computing and green sustainable computing to enhance the performance of a distributed system. Finally presenting the future scope, we conclude the paper suggesting a path to achieve a Green high performance distributed system using cluster, grid and cloud computing

  8. Heterogeneous High Throughput Scientific Computing with APM X-Gene and Intel Xeon Phi

    CERN Document Server

    Abdurachmanov, David; Elmer, Peter; Eulisse, Giulio; Knight, Robert; Muzaffar, Shahzad

    2014-01-01

    Electrical power requirements will be a constraint on the future growth of Distributed High Throughput Computing (DHTC) as used by High Energy Physics. Performance-per-watt is a critical metric for the evaluation of computer architectures for cost- efficient computing. Additionally, future performance growth will come from heterogeneous, many-core, and high computing density platforms with specialized processors. In this paper, we examine the Intel Xeon Phi Many Integrated Cores (MIC) co-processor and Applied Micro X-Gene ARMv8 64-bit low-power server system-on-a-chip (SoC) solutions for scientific computing applications. We report our experience on software porting, performance and energy efficiency and evaluate the potential for use of such technologies in the context of distributed computing systems such as the Worldwide LHC Computing Grid (WLCG).

  9. Heterogeneous High Throughput Scientific Computing with APM X-Gene and Intel Xeon Phi

    Science.gov (United States)

    Abdurachmanov, David; Bockelman, Brian; Elmer, Peter; Eulisse, Giulio; Knight, Robert; Muzaffar, Shahzad

    2015-05-01

    Electrical power requirements will be a constraint on the future growth of Distributed High Throughput Computing (DHTC) as used by High Energy Physics. Performance-per-watt is a critical metric for the evaluation of computer architectures for cost- efficient computing. Additionally, future performance growth will come from heterogeneous, many-core, and high computing density platforms with specialized processors. In this paper, we examine the Intel Xeon Phi Many Integrated Cores (MIC) co-processor and Applied Micro X-Gene ARMv8 64-bit low-power server system-on-a-chip (SoC) solutions for scientific computing applications. We report our experience on software porting, performance and energy efficiency and evaluate the potential for use of such technologies in the context of distributed computing systems such as the Worldwide LHC Computing Grid (WLCG).

  10. Efficient and Highly Aldehyde Selective Wacker Oxidation

    KAUST Repository

    Teo, Peili

    2012-07-06

    A method for efficient and aldehyde-selective Wacker oxidation of aryl-substituted olefins using PdCl 2(MeCN) 2, 1,4-benzoquinone, and t-BuOH in air is described. Up to a 96% yield of aldehyde can be obtained, and up to 99% selectivity can be achieved with styrene-related substrates. © 2012 American Chemical Society.

  11. Computer vision for high content screening.

    Science.gov (United States)

    Kraus, Oren Z; Frey, Brendan J

    2016-01-01

    High Content Screening (HCS) technologies that combine automated fluorescence microscopy with high throughput biotechnology have become powerful systems for studying cell biology and drug screening. These systems can produce more than 100 000 images per day, making their success dependent on automated image analysis. In this review, we describe the steps involved in quantifying microscopy images and different approaches for each step. Typically, individual cells are segmented from the background using a segmentation algorithm. Each cell is then quantified by extracting numerical features, such as area and intensity measurements. As these feature representations are typically high dimensional (>500), modern machine learning algorithms are used to classify, cluster and visualize cells in HCS experiments. Machine learning algorithms that learn feature representations, in addition to the classification or clustering task, have recently advanced the state of the art on several benchmarking tasks in the computer vision community. These techniques have also recently been applied to HCS image analysis.

  12. Computing High Accuracy Power Spectra with Pico

    CERN Document Server

    Fendt, William A

    2007-01-01

    This paper presents the second release of Pico (Parameters for the Impatient COsmologist). Pico is a general purpose machine learning code which we have applied to computing the CMB power spectra and the WMAP likelihood. For this release, we have made improvements to the algorithm as well as the data sets used to train Pico, leading to a significant improvement in accuracy. For the 9 parameter nonflat case presented here Pico can on average compute the TT, TE and EE spectra to better than 1% of cosmic standard deviation for nearly all $\\ell$ values over a large region of parameter space. Performing a cosmological parameter analysis of current CMB and large scale structure data, we show that these power spectra give very accurate 1 and 2 dimensional parameter posteriors. We have extended Pico to allow computation of the tensor power spectrum and the matter transfer function. Pico runs about 1500 times faster than CAMB at the default accuracy and about 250,000 times faster at high accuracy. Training Pico can be...

  13. Algorithms for Efficient Computation of Transfer Functions for Large Order Flexible Systems

    Science.gov (United States)

    Maghami, Peiman G.; Giesy, Daniel P.

    1998-01-01

    An efficient and robust computational scheme is given for the calculation of the frequency response function of a large order, flexible system implemented with a linear, time invariant control system. Advantage is taken of the highly structured sparsity of the system matrix of the plant based on a model of the structure using normal mode coordinates. The computational time per frequency point of the new computational scheme is a linear function of system size, a significant improvement over traditional, still-matrix techniques whose computational times per frequency point range from quadratic to cubic functions of system size. This permits the practical frequency domain analysis of systems of much larger order than by traditional, full-matrix techniques. Formulations are given for both open- and closed-loop systems. Numerical examples are presented showing the advantages of the present formulation over traditional approaches, both in speed and in accuracy. Using a model with 703 structural modes, the present method was up to two orders of magnitude faster than a traditional method. The present method generally showed good to excellent accuracy throughout the range of test frequencies, while traditional methods gave adequate accuracy for lower frequencies, but generally deteriorated in performance at higher frequencies with worst case errors being many orders of magnitude times the correct values.

  14. High Thrust Efficiency MPD Thruster Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Magnetoplasmadynamic (MPD) thrusters can provide the high-specific impulse, high-power propulsion required to support human and robotic exploration missions to the...

  15. Computationally Efficient Power Allocation Algorithm in Multicarrier-Based Cognitive Radio Networks: OFDM and FBMC Systems

    Science.gov (United States)

    Shaat, Musbah; Bader, Faouzi

    2010-12-01

    Cognitive Radio (CR) systems have been proposed to increase the spectrum utilization by opportunistically access the unused spectrum. Multicarrier communication systems are promising candidates for CR systems. Due to its high spectral efficiency, filter bank multicarrier (FBMC) can be considered as an alternative to conventional orthogonal frequency division multiplexing (OFDM) for transmission over the CR networks. This paper addresses the problem of resource allocation in multicarrier-based CR networks. The objective is to maximize the downlink capacity of the network under both total power and interference introduced to the primary users (PUs) constraints. The optimal solution has high computational complexity which makes it unsuitable for practical applications and hence a low complexity suboptimal solution is proposed. The proposed algorithm utilizes the spectrum holes in PUs bands as well as active PU bands. The performance of the proposed algorithm is investigated for OFDM and FBMC based CR systems. Simulation results illustrate that the proposed resource allocation algorithm with low computational complexity achieves near optimal performance and proves the efficiency of using FBMC in CR context.

  16. Increasing the computational efficient of digital cross correlation by a vectorization method

    Science.gov (United States)

    Chang, Ching-Yuan; Ma, Chien-Ching

    2017-08-01

    This study presents a vectorization method for use in MATLAB programming aimed at increasing the computational efficiency of digital cross correlation in sound and images, resulting in a speedup of 6.387 and 36.044 times compared with performance values obtained from looped expression. This work bridges the gap between matrix operations and loop iteration, preserving flexibility and efficiency in program testing. This paper uses numerical simulation to verify the speedup of the proposed vectorization method as well as experiments to measure the quantitative transient displacement response subjected to dynamic impact loading. The experiment involved the use of a high speed camera as well as a fiber optic system to measure the transient displacement in a cantilever beam under impact from a steel ball. Experimental measurement data obtained from the two methods are in excellent agreement in both the time and frequency domain, with discrepancies of only 0.68%. Numerical and experiment results demonstrate the efficacy of the proposed vectorization method with regard to computational speed in signal processing and high precision in the correlation algorithm. We also present the source code with which to build MATLAB-executable functions on Windows as well as Linux platforms, and provide a series of examples to demonstrate the application of the proposed vectorization method.

  17. Computationally Efficient Power Allocation Algorithm in Multicarrier-Based Cognitive Radio Networks: OFDM and FBMC Systems

    Directory of Open Access Journals (Sweden)

    Shaat Musbah

    2010-01-01

    Full Text Available Cognitive Radio (CR systems have been proposed to increase the spectrum utilization by opportunistically access the unused spectrum. Multicarrier communication systems are promising candidates for CR systems. Due to its high spectral efficiency, filter bank multicarrier (FBMC can be considered as an alternative to conventional orthogonal frequency division multiplexing (OFDM for transmission over the CR networks. This paper addresses the problem of resource allocation in multicarrier-based CR networks. The objective is to maximize the downlink capacity of the network under both total power and interference introduced to the primary users (PUs constraints. The optimal solution has high computational complexity which makes it unsuitable for practical applications and hence a low complexity suboptimal solution is proposed. The proposed algorithm utilizes the spectrum holes in PUs bands as well as active PU bands. The performance of the proposed algorithm is investigated for OFDM and FBMC based CR systems. Simulation results illustrate that the proposed resource allocation algorithm with low computational complexity achieves near optimal performance and proves the efficiency of using FBMC in CR context.

  18. Design Strategies for Ultra-high Efficiency Photovoltaics

    Science.gov (United States)

    Warmann, Emily Cathryn

    While concentrator photovoltaic cells have shown significant improvements in efficiency in the past ten years, once these cells are integrated into concentrating optics, connected to a power conditioning system and deployed in the field, the overall module efficiency drops to only 34 to 36%. This efficiency is impressive compared to conventional flat plate modules, but it is far short of the theoretical limits for solar energy conversion. Designing a system capable of achieving ultra high efficiency of 50% or greater cannot be achieved by refinement and iteration of current design approaches. This thesis takes a systems approach to designing a photovoltaic system capable of 50% efficient performance using conventional diode-based solar cells. The effort began with an exploration of the limiting efficiency of spectrum splitting ensembles with 2 to 20 sub cells in different electrical configurations. Incorporating realistic non-ideal performance with the computationally simple detailed balance approach resulted in practical limits that are useful to identify specific cell performance requirements. This effort quantified the relative benefit of additional cells and concentration for system efficiency, which will help in designing practical optical systems. Efforts to improve the quality of the solar cells themselves focused on the development of tunable lattice constant epitaxial templates. Initially intended to enable lattice matched multijunction solar cells, these templates would enable increased flexibility in band gap selection for spectrum splitting ensembles and enhanced radiative quality relative to metamorphic growth. The III-V material family is commonly used for multijunction solar cells both for its high radiative quality and for the ease of integrating multiple band gaps into one monolithic growth. The band gap flexibility is limited by the lattice constant of available growth templates. The virtual substrate consists of a thin III-V film with the desired

  19. Automatic Generation of Very Efficient Programs by Generalized Partial Computation

    Institute of Scientific and Technical Information of China (English)

    2001-01-01

    Generalized Partial Computation (GPC) is a program transformationmethod utilizi ng partial information about input data, properties of auxiliary functions and t he logical structure of a source program. GPC uses both an inference engine such as a theorem prover and a classical partial evaluator to optimize programs. The refore, GPC is more powerful than classical partial evaluators but harder to imp lement and control. We have implemented an experimental GPC system called WSDFU (Waseda Simplify-Distribute-Fold-Unfold). This paper discusses the power of t he program transformation system, its theorem prover and future works.

  20. Work-Efficient Parallel Skyline Computation for the GPU

    DEFF Research Database (Denmark)

    Bøgh, Kenneth Sejdenfaden; Chester, Sean; Assent, Ira

    2015-01-01

    offers the potential for parallelizing skyline computation across thousands of cores. However, attempts to port skyline algorithms to the GPU have prioritized throughput and failed to outperform sequential algorithms. In this paper, we introduce a new skyline algorithm, designed for the GPU, that uses...... a global, static partitioning scheme. With the partitioning, we can permit controlled branching to exploit transitive relationships and avoid most point-to-point comparisons. The result is a non-traditional GPU algorithm, SkyAlign, that prioritizes work-effciency and respectable throughput, rather than...

  1. Techniques for Efficiently Ensuring Data Storage Security in Cloud Computing

    DEFF Research Database (Denmark)

    Banoth, Rajkumar

    2011-01-01

    The Cloud Computing is the next generation architecture of IT Enterprise. It moves the application software and databases to the large data centers, where the management of the data and services may not be fully trustworthy. Here, focus is on cloud data storage security, an important aspect...... of quality of service. To ensure the correctness of users’ data in the cloud, we propose an effective and flexible distributed scheme with two salient features. By utilizing the homomorphic token with distributed verification of erasure-coded data, the scheme achieves the integration of storage correctness...

  2. PMMA lens with high efficiency and reliability

    Science.gov (United States)

    Matsuzaki, Ichiro; Abe, Koji; Fujita, Katsuhiro

    2013-09-01

    Polymethyl Methacrylate (PMMA) Fresnel lenses are increasingly being used in concentrated photovoltaic (CPV) systems installed outdoors and, accordingly, emphasis is being placed on the durability of such lenses with regard to light transmittance when subject to ultraviolet (UV) light and dust exposure. Accelerated testing methods for evaluating durability under UV exposure were established, allowing development of a lens material with improved UV resistance. Simultaneously, through a proprietary molding method, a Fresnel lens that boasts favorable light concentration efficiency with little deformation even after prolonged outdoor use was developed. Moreover, the lens incorporates a new hard-coat finish that possesses sand durability and UV resistance comparable to that of tempered glass.

  3. Energy-Efficient Management of Data Center Resources for Cloud Computing: A Vision, Architectural Elements, and Open Challenges

    CERN Document Server

    Buyya, Rajkumar; Abawajy, Jemal

    2010-01-01

    Cloud computing is offering utility-oriented IT services to users worldwide. Based on a pay-as-you-go model, it enables hosting of pervasive applications from consumer, scientific, and business domains. However, data centers hosting Cloud applications consume huge amounts of energy, contributing to high operational costs and carbon footprints to the environment. Therefore, we need Green Cloud computing solutions that can not only save energy for the environment but also reduce operational costs. This paper presents vision, challenges, and architectural elements for energy-efficient management of Cloud computing environments. We focus on the development of dynamic resource provisioning and allocation algorithms that consider the synergy between various data center infrastructures (i.e., the hardware, power units, cooling and software), and holistically work to boost data center energy efficiency and performance. In particular, this paper proposes (a) architectural principles for energy-efficient management of ...

  4. Computationally Efficient Nonlinearity Compensation for Coherent Fiber-Optic Systems

    Institute of Scientific and Technical Information of China (English)

    Likai Zhu; Guifang Li

    2012-01-01

    Split-step digital backward propagation (DBP) can be combined with coherent detection to compensate for fiber nonlinear impairments. A large number of DBP steps is usually needed for a long-haul fiber system, and this creates a heavy computational load. In a trade-off between complexity and performance, interchannel nonlinearity can be disregarded in order to simplify the DBP algorithm. The number of steps can also be reduced at the expense of performance. In periodic dispersion-managed long-haul transmission systems, optical waveform distortion is dominated by chromatic dispersion. As a result, the nonlinearity of the optical signal repeats in every dispersion period. Because of this periodic behavior, DBP of many fiber spans can be folded into one span. Using this distance-folded DBP method, the required computation for a transoceanic transmission system with full inline dispersion compensation can be reduced by up to two orders of magnitude with negligible penalty. The folded DBP method can be modified to compensate for nonlinearity in fiber links with non-zero residua dispersion per span.

  5. Probabilistic Damage Characterization Using the Computationally-Efficient Bayesian Approach

    Science.gov (United States)

    Warner, James E.; Hochhalter, Jacob D.

    2016-01-01

    This work presents a computationally-ecient approach for damage determination that quanti es uncertainty in the provided diagnosis. Given strain sensor data that are polluted with measurement errors, Bayesian inference is used to estimate the location, size, and orientation of damage. This approach uses Bayes' Theorem to combine any prior knowledge an analyst may have about the nature of the damage with information provided implicitly by the strain sensor data to form a posterior probability distribution over possible damage states. The unknown damage parameters are then estimated based on samples drawn numerically from this distribution using a Markov Chain Monte Carlo (MCMC) sampling algorithm. Several modi cations are made to the traditional Bayesian inference approach to provide signi cant computational speedup. First, an ecient surrogate model is constructed using sparse grid interpolation to replace a costly nite element model that must otherwise be evaluated for each sample drawn with MCMC. Next, the standard Bayesian posterior distribution is modi ed using a weighted likelihood formulation, which is shown to improve the convergence of the sampling process. Finally, a robust MCMC algorithm, Delayed Rejection Adaptive Metropolis (DRAM), is adopted to sample the probability distribution more eciently. Numerical examples demonstrate that the proposed framework e ectively provides damage estimates with uncertainty quanti cation and can yield orders of magnitude speedup over standard Bayesian approaches.

  6. Quantum Confined Semiconductors for High Efficiency Photovoltaics

    Science.gov (United States)

    Beard, Matthew

    2014-03-01

    Semiconductor nanostructures, where at least one dimension is small enough to produce quantum confinement effects, provide new pathways for controlling energy flow and therefore have the potential to increase the efficiency of the primary photon-to-free energy conversion step. In this discussion, I will present the current status of research efforts towards utilizing the unique properties of colloidal quantum dots (NCs confined in three dimensions) in prototype solar cells and demonstrate that these unique systems have the potential to bypass the Shockley-Queisser single-junction limit for solar photon conversion. The solar cells are constructed using a low temperature solution based deposition of PbS or PbSe QDs as the absorber layer. Different chemical treatments of the QD layer are employed in order to obtain good electrical communication while maintaining the quantum-confined properties of the QDs. We have characterized the transport and carrier dynamics using a transient absorption, time-resolved THz, and temperature-dependent photoluminescence. I will discuss the interplay between carrier generation, recombination, and mobility within the QD layers. A unique aspect of our devices is that the QDs exhibit multiple exciton generation with an efficiency that is ~ 2 to 3 times greater than the parental bulk semiconductor.

  7. High performance computing for beam physics applications

    Science.gov (United States)

    Ryne, R. D.; Habib, S.

    Several countries are now involved in efforts aimed at utilizing accelerator-driven technologies to solve problems of national and international importance. These technologies have both economic and environmental implications. The technologies include waste transmutation, plutonium conversion, neutron production for materials science and biological science research, neutron production for fusion materials testing, fission energy production systems, and tritium production. All of these projects require a high-intensity linear accelerator that operates with extremely low beam loss. This presents a formidable computational challenge: One must design and optimize over a kilometer of complex accelerating structures while taking into account beam loss to an accuracy of 10 parts per billion per meter. Such modeling is essential if one is to have confidence that the accelerator will meet its beam loss requirement, which ultimately affects system reliability, safety and cost. At Los Alamos, the authors are developing a capability to model ultra-low loss accelerators using the CM-5 at the Advanced Computing Laboratory. They are developing PIC, Vlasov/Poisson, and Langevin/Fokker-Planck codes for this purpose. With slight modification, they have also applied their codes to modeling mesoscopic systems and astrophysical systems. In this paper, they will first describe HPC activities in the accelerator community. Then they will discuss the tools they have developed to model classical and quantum evolution equations. Lastly they will describe how these tools have been used to study beam halo in high current, mismatched charged particle beams.

  8. Recent development of high efficiency arc welding systems in Japan

    Institute of Scientific and Technical Information of China (English)

    USHIO Masao; SUGITANI Yuji; TANAKA Manabu

    2005-01-01

    The paper describes recent trends to carry forward the automation in welding process in Japanese industries.The present situation on the introduction of computer integrated manufacturing is surveyed and its concept and relation among the constituent subsystems are shown.Also recent developments of arc welding processes and related tools are reviewed.They are as follows.1) Schematic image of the structure of computer integrated manufacturing system (CIM).2) Short descriptions for the trend of introducing CIM system.An example of multi-robot welding system in shipyard,virtual factory,network monitoring of welding quality,clarification of welder's skill.3) Development of high efficiency welding processes.Copper-less wire electrode,tandem large fillet welding,tandem pulse GMA welding,digital wire feeder with new push-pull system,overhead horizontal tandem GMA welding for ship bottom shell,one-sided fillet groove GMA welding for construction machinery,ultra-NGW hot wire GTA process,Al-alloy sheet welding with laser assisted AC pulsed MIG process.

  9. Evaluation of high-performance computing software

    Energy Technology Data Exchange (ETDEWEB)

    Browne, S.; Dongarra, J. [Univ. of Tennessee, Knoxville, TN (United States); Rowan, T. [Oak Ridge National Lab., TN (United States)

    1996-12-31

    The absence of unbiased and up to date comparative evaluations of high-performance computing software complicates a user`s search for the appropriate software package. The National HPCC Software Exchange (NHSE) is attacking this problem using an approach that includes independent evaluations of software, incorporation of author and user feedback into the evaluations, and Web access to the evaluations. We are applying this approach to the Parallel Tools Library (PTLIB), a new software repository for parallel systems software and tools, and HPC-Netlib, a high performance branch of the Netlib mathematical software repository. Updating the evaluations with feed-back and making it available via the Web helps ensure accuracy and timeliness, and using independent reviewers produces unbiased comparative evaluations difficult to find elsewhere.

  10. Evaluation of high-performance computing software

    Energy Technology Data Exchange (ETDEWEB)

    Browne, S.; Dongarra, J. [Univ. of Tennessee, Knoxville, TN (United States); Rowan, T. [Oak Ridge National Lab., TN (United States)

    1996-12-31

    The absence of unbiased and up to date comparative evaluations of high-performance computing software complicates a user`s search for the appropriate software package. The National HPCC Software Exchange (NHSE) is attacking this problem using an approach that includes independent evaluations of software, incorporation of author and user feedback into the evaluations, and Web access to the evaluations. We are applying this approach to the Parallel Tools Library (PTLIB), a new software repository for parallel systems software and tools, and HPC-Netlib, a high performance branch of the Netlib mathematical software repository. Updating the evaluations with feed-back and making it available via the Web helps ensure accuracy and timeliness, and using independent reviewers produces unbiased comparative evaluations difficult to find elsewhere.

  11. Ultra-high resolution computed tomography imaging

    Energy Technology Data Exchange (ETDEWEB)

    Paulus, Michael J. (Knoxville, TN); Sari-Sarraf, Hamed (Knoxville, TN); Tobin, Jr., Kenneth William (Harriman, TN); Gleason, Shaun S. (Knoxville, TN); Thomas, Jr., Clarence E. (Knoxville, TN)

    2002-01-01

    A method for ultra-high resolution computed tomography imaging, comprising the steps of: focusing a high energy particle beam, for example x-rays or gamma-rays, onto a target object; acquiring a 2-dimensional projection data set representative of the target object; generating a corrected projection data set by applying a deconvolution algorithm, having an experimentally determined a transfer function, to the 2-dimensional data set; storing the corrected projection data set; incrementally rotating the target object through an angle of approximately 180.degree., and after each the incremental rotation, repeating the radiating, acquiring, generating and storing steps; and, after the rotating step, applying a cone-beam algorithm, for example a modified tomographic reconstruction algorithm, to the corrected projection data sets to generate a 3-dimensional image. The size of the spot focus of the beam is reduced to not greater than approximately 1 micron, and even to not greater than approximately 0.5 microns.

  12. High-Efficient Circuits for Ternary Addition

    Directory of Open Access Journals (Sweden)

    Reza Faghih Mirzaee

    2014-01-01

    Full Text Available New ternary adders, which are fundamental components of ternary addition, are presented in this paper. They are on the basis of a logic style which mostly generates binary signals. Therefore, static power dissipation reaches its minimum extent. Extensive different analyses are carried out to examine how efficient the new designs are. For instance, the ternary ripple adder constructed by the proposed ternary half and full adders consumes 2.33 μW less power than the one implemented by the previous adder cells. It is almost twice faster as well. Due to their unique superior characteristics for ternary circuitry, carbon nanotube field-effect transistors are used to form the novel circuits, which are entirely suitable for practical applications.

  13. Efficient relaxed-Jacobi smoothers for multigrid on parallel computers

    Science.gov (United States)

    Yang, Xiang; Mittal, Rajat

    2017-03-01

    In this Technical Note, we present a family of Jacobi-based multigrid smoothers suitable for the solution of discretized elliptic equations. These smoothers are based on the idea of scheduled-relaxation Jacobi proposed recently by Yang & Mittal (2014) [18] and employ two or three successive relaxed Jacobi iterations with relaxation factors derived so as to maximize the smoothing property of these iterations. The performance of these new smoothers measured in terms of convergence acceleration and computational workload, is assessed for multi-domain implementations typical of parallelized solvers, and compared to the lexicographic point Gauss-Seidel smoother. The tests include the geometric multigrid method on structured grids as well as the algebraic grid method on unstructured grids. The tests demonstrate that unlike Gauss-Seidel, the convergence of these Jacobi-based smoothers is unaffected by domain decomposition, and furthermore, they outperform the lexicographic Gauss-Seidel by factors that increase with domain partition count.

  14. Communication efficient basic linear algebra computations on hypercube architectures

    Energy Technology Data Exchange (ETDEWEB)

    Johnsson, S.L.

    1987-04-01

    This paper presents a few algorithms for embedding loops and multidimensional arrays in hypercubes with emphasis on proximity preserving embeddings. A proximity preserving embedding minimizes the need for communication bandwidth in computations requiring nearest neighbor communication. Two storage schemes for ''large'' problems on ''small'' machines are suggested and analyzed, and algorithms for matrix transpose, multiplying matrices, factoring matrices, and solving triangular linear systems are presented. A few complete binary tree embeddings are described and analyzed. The data movement in the matrix algorithms is analyzed and it is shown that in the majority of cases the directed routing paths intersect only at nodes of the hypercube allowing for a maximum degree of pipelining.

  15. Computational approaches for efficiently modelling of small atmospheric clusters

    DEFF Research Database (Denmark)

    Elm, Jonas; Mikkelsen, Kurt Valentin

    2014-01-01

    Utilizing a comprehensive test set of 205 clusters of atmospheric relevance, we investigate how different DFT functionals (M06-2X, PW91, ωB97X-D) and basis sets (6-311++G(3df,3pd), 6-31++G(d,p), 6-31+G(d)) affect the thermal contribution to the Gibbs free energy and single point energy. Reducing...... the basis set used in the geometry and frequency calculation from 6-311++G(3df,3pd) → 6-31++G(d,p) implies a significant speed-up in computational time and only leads to small errors in the thermal contribution to the Gibbs free energy and subsequent coupled cluster single point energy calculation....

  16. Novel Polymers for High Efficiency Renewable and Portable Power Applications

    Science.gov (United States)

    2015-07-30

    systems with frontier orbital levels (HOMOs and LUMOs) and morphologies systematically investigated and optimized for high efficiency photoelectric and...of polymer-dye covalently linked systems that could efficiently convert light/heat into electrical power. Therefore, frontier orbital matching...force between the polymer and dye would result in weaker PL quenching and optoelectronic device power conversion efficiency , this experimentally

  17. A review of high-efficiency silicon solar cells

    Science.gov (United States)

    Rohatgi, A.

    1986-01-01

    Various parameters that affect solar cell efficiency were discussed. It is not understood why solar cells produced from less expensive Czochralski (Cz) silicon are less efficient than cells fabricated from more expensive float-zone (Fz) silicon. Performance characteristics were presented for recently produced, high-efficient solar cells fabricated by Westinghouse Electric Corp., Spire Corp., University of New South Wales, and Stanford University.

  18. Efficient Sustainable Operation Mechanism of Distributed Desktop Integration Storage Based on Virtualization with Ubiquitous Computing

    Directory of Open Access Journals (Sweden)

    Hyun-Woo Kim

    2015-06-01

    Full Text Available Following the rapid growth of ubiquitous computing, many jobs that were previously manual have now been automated. This automation has increased the amount of time available for leisure; diverse services are now being developed for this leisure time. In addition, the development of small and portable devices like smartphones, diverse Internet services can be used regardless of time and place. Studies regarding diverse virtualization are currently in progress. These studies aim to determine ways to efficiently store and process the big data generated by the multitude of devices and services in use. One topic of such studies is desktop storage virtualization, which integrates distributed desktop resources and provides these resources to users to integrate into distributed legacy desktops via virtualization. In the case of desktop storage virtualization, high availability of virtualization is necessary and important for providing reliability to users. Studies regarding hierarchical structures and resource integration are currently in progress. These studies aim to create efficient data distribution and storage for distributed desktops based on resource integration environments. However, studies regarding efficient responses to server faults occurring in desktop-based resource integration environments have been insufficient. This paper proposes a mechanism for the sustainable operation of desktop storage (SODS for high operational availability. It allows for the easy addition and removal of desktops in desktop-based integration environments. It also activates alternative servers when a fault occurs within a system.

  19. Performance Comparison of Hybrid Signed Digit Arithmetic in Efficient Computing

    Directory of Open Access Journals (Sweden)

    VISHAL AWASTHI

    2011-10-01

    Full Text Available In redundant representations, addition can be carried out in a constant time independent of the word length of the operands. Adder forms a fundamental building block in almost majority of VLSI designs. A hybrid adder can add an unsigned number to a signed-digit number and hence their efficient performance greatly determinesthe quality of the final output of the concerned circuit. In this paper we designed and compared the speed of adders by reducing the carry propagation time with the help of combined effect of improved architectures of adders and signed digit representation of number systems. The key idea is to draw out a compromise between execution time of fast adding process and area available which is often very limited. In this paper we also tried to verify the various algorithms of signed digit and hybrid signed digit adders.

  20. Efficient Multidimensional Data Redistribution for Resizable Parallel Computations

    CERN Document Server

    Sudarsan, Rajesh

    2007-01-01

    Traditional parallel schedulers running on cluster supercomputers support only static scheduling, where the number of processors allocated to an application remains fixed throughout the execution of the job. This results in under-utilization of idle system resources thereby decreasing overall system throughput. In our research, we have developed a prototype framework called ReSHAPE, which supports dynamic resizing of parallel MPI applications executing on distributed memory platforms. The resizing library in ReSHAPE includes support for releasing and acquiring processors and efficiently redistributing application state to a new set of processors. In this paper, we derive an algorithm for redistributing two-dimensional block-cyclic arrays from $P$ to $Q$ processors, organized as 2-D processor grids. The algorithm ensures a contention-free communication schedule for data redistribution if $P_r \\leq Q_r$ and $P_c \\leq Q_c$. In other cases, the algorithm implements circular row and column shifts on the communicat...

  1. High Efficiency Regenerative Helium Compressor Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Helium plays several critical rolls in spacecraft propulsion. High pressure helium is commonly used to pressurize propellant fuel tanks. Helium cryocoolers can be...

  2. Advanced High Efficiency Durable DACS Thruster Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Systima is developing a high performance 25 lbf DACS thruster that operates with a novel non-toxic monopropellant. The monopropellant has a 30% higher...

  3. Molecular Dynamics Simulations on High-Performance Reconfigurable Computing Systems.

    Science.gov (United States)

    Chiu, Matt; Herbordt, Martin C

    2010-11-01

    The acceleration of molecular dynamics (MD) simulations using high-performance reconfigurable computing (HPRC) has been much studied. Given the intense competition from multicore and GPUs, there is now a question whether MD on HPRC can be competitive. We concentrate here on the MD kernel computation: determining the short-range force between particle pairs. In one part of the study, we systematically explore the design space of the force pipeline with respect to arithmetic algorithm, arithmetic mode, precision, and various other optimizations. We examine simplifications and find that some have little effect on simulation quality. In the other part, we present the first FPGA study of the filtering of particle pairs with nearly zero mutual force, a standard optimization in MD codes. There are several innovations, including a novel partitioning of the particle space, and new methods for filtering and mapping work onto the pipelines. As a consequence, highly efficient filtering can be implemented with only a small fraction of the FPGA's resources. Overall, we find that, for an Altera Stratix-III EP3ES260, 8 force pipelines running at nearly 200 MHz can fit on the FPGA, and that they can perform at 95% efficiency. This results in an 80-fold per core speed-up for the short-range force, which is likely to make FPGAs highly competitive for MD.

  4. Benchmarking: More Aspects of High Performance Computing

    Energy Technology Data Exchange (ETDEWEB)

    Ravindrudu, Rahul [Iowa State Univ., Ames, IA (United States)

    2004-01-01

    The original HPL algorithm makes the assumption that all data can be fit entirely in the main memory. This assumption will obviously give a good performance due to the absence of disk I/O. However, not all applications can fit their entire data in memory. These applications which require a fair amount of I/O to move data to and from main memory and secondary storage, are more indicative of usage of an Massively Parallel Processor (MPP) System. Given this scenario a well designed I/O architecture will play a significant part in the performance of the MPP System on regular jobs. And, this is not represented in the current Benchmark. The modified HPL algorithm is hoped to be a step in filling this void. The most important factor in the performance of out-of-core algorithms is the actual I/O operations performed and their efficiency in transferring data to/from main memory and disk, Various methods were introduced in the report for performing I/O operations. The I/O method to use depends on the design of the out-of-core algorithm. Conversely, the performance of the out-of-core algorithm is affected by the choice of I/O operations. This implies, good performance is achieved when I/O efficiency is closely tied with the out-of-core algorithms. The out-of-core algorithms must be designed from the start. It is easily observed in the timings for various plots, that I/O plays a significant part in the overall execution time. This leads to an important conclusion, retro-fitting an existing code may not be the best choice. The right-looking algorithm selected for the LU factorization is a recursive algorithm and performs well when the entire dataset is in memory. At each stage of the loop the entire trailing submatrix is read into memory panel by panel. This gives a polynomial number of I/O reads and writes. If the left-looking algorithm was selected for the main loop, the number of I/O operations involved will be linear on the number of columns. This is due to the data access

  5. Implementing an Affordable High-Performance Computing for Teaching-Oriented Computer Science Curriculum

    Science.gov (United States)

    Abuzaghleh, Omar; Goldschmidt, Kathleen; Elleithy, Yasser; Lee, Jeongkyu

    2013-01-01

    With the advances in computing power, high-performance computing (HPC) platforms have had an impact on not only scientific research in advanced organizations but also computer science curriculum in the educational community. For example, multicore programming and parallel systems are highly desired courses in the computer science major. However,…

  6. Implementing an Affordable High-Performance Computing for Teaching-Oriented Computer Science Curriculum

    Science.gov (United States)

    Abuzaghleh, Omar; Goldschmidt, Kathleen; Elleithy, Yasser; Lee, Jeongkyu

    2013-01-01

    With the advances in computing power, high-performance computing (HPC) platforms have had an impact on not only scientific research in advanced organizations but also computer science curriculum in the educational community. For example, multicore programming and parallel systems are highly desired courses in the computer science major. However,…

  7. Efficient Computing of some Vector Operations over GF(3) and GF(4)

    OpenAIRE

    Bouyukliev, Iliya; Bakoev, Valentin

    2008-01-01

    The problem of efficient computing of the affine vector operations (addition of two vectors and multiplication of a vector by a scalar over GF (q)), and also the weight of a given vector, is important for many problems in coding theory, cryptography, VLSI technology etc. In this paper we propose a new way of representing vectors over GF (3) and GF (4) and we describe an efficient performance of these affine operations. Computing weights of binary vectors is also discussed.

  8. Iterative coupling reservoir simulation on high performance computers

    Institute of Scientific and Technical Information of China (English)

    Lu Bo; Wheeler Mary F

    2009-01-01

    In this paper, the iterative coupling approach is proposed for applications to solving multiphase flow equation systems in reservoir simulation, as it provides a more flexible time-stepping strategy than existing approaches. The iterative method decouples the whole equation systems into pressure and saturation/concentration equations, and then solves them in sequence, implicitly and semi-implicitly. At each time step, a series of iterations are computed, which involve solving linearized equations using specific tolerances that are iteration dependent. Following convergence of subproblems, material balance is checked. Convergence of time steps is based on material balance errors. Key components of the iterative method include phase scaling for deriving a pressure equation and use of several advanced numerical techniques. The iterative model is implemented for parallel computing platforms and shows high parallel efficiency and scalability.

  9. An accurate and efficient computation method of the hydration free energy of a large, complex molecule.

    Science.gov (United States)

    Yoshidome, Takashi; Ekimoto, Toru; Matubayasi, Nobuyuki; Harano, Yuichi; Kinoshita, Masahiro; Ikeguchi, Mitsunori

    2015-05-07

    The hydration free energy (HFE) is a crucially important physical quantity to discuss various chemical processes in aqueous solutions. Although an explicit-solvent computation with molecular dynamics (MD) simulations is a preferable treatment of the HFE, huge computational load has been inevitable for large, complex solutes like proteins. In the present paper, we propose an efficient computation method for the HFE. In our method, the HFE is computed as a sum of 〈UUV〉/2 (〈UUV〉 is the ensemble average of the sum of pair interaction energy between solute and water molecule) and the water reorganization term mainly reflecting the excluded volume effect. Since 〈UUV〉 can readily be computed through a MD of the system composed of solute and water, an efficient computation of the latter term leads to a reduction of computational load. We demonstrate that the water reorganization term can quantitatively be calculated using the morphometric approach (MA) which expresses the term as the linear combinations of the four geometric measures of a solute and the corresponding coefficients determined with the energy representation (ER) method. Since the MA enables us to finish the computation of the solvent reorganization term in less than 0.1 s once the coefficients are determined, the use of the MA enables us to provide an efficient computation of the HFE even for large, complex solutes. Through the applications, we find that our method has almost the same quantitative performance as the ER method with substantial reduction of the computational load.

  10. Preparation of highly efficient manganese catalase mimics.

    Science.gov (United States)

    Triller, Michael U; Hsieh, Wen-Yuan; Pecoraro, Vincent L; Rompel, Annette; Krebs, Bernt

    2002-10-21

    The series of compounds [Mn(bpia)(mu-OAc)](2)(ClO(4))(2) (1), [Mn(2)(bpia)(2)(muO)(mu-OAc)](ClO(4))(3).CH(3)CN (2), [Mn(bpia)(mu-O)](2)(ClO(4))(2)(PF(6)).2CH(3)CN (3), [Mn(bpia)(Cl)(2)](ClO)(4) (4), and [(Mn(bpia)(Cl))(2)(mu-O)](ClO(4))(2).2CH(3)CN (5) (bpia = bis(picolyl)(N-methylimidazol-2-yl)amine) represents a structural, spectroscopic, and functional model system for manganese catalases. Compounds 3 and 5 have been synthesized from 2 via bulk electrolysis and ligand exchange, respectively. All complexes have been structurally characterized by X-ray crystallography and by UV-vis and EPR spectroscopies. The different bridging ligands including the rare mono-mu-oxo and mono-mu-oxo-mono-mu-carboxylato motifs lead to a variation of the Mn-Mn separation across the four binuclear compounds of 1.50 A (Mn(2)(II,II) = 4.128 A, Mn(2)(III,III) = 3.5326 and 3.2533 A, Mn(2)(III,IV) = 2.624 A). Complexes 1, 2, and 3 are mimics for the Mn(2)(II,II), the Mn(2)(III,III), and the Mn(2)(III,IV) oxidation states of the native enzyme. UV-vis spectra of these compounds show similarities to those of the corresponding oxidation states of manganese catalase from Thermus thermophilus and Lactobacillus plantarum. Compound 2 exhibits a rare example of a Jahn-Teller compression. While complexes 1 and 3 are efficient catalysts for the disproportionation of hydrogen peroxide and contain an N(4)O(2) donor set, 4 and 5 show no catalase activity. These complexes have an N(4)Cl(2) and N(4)OCl donor set, respectively, and serve as mimics for halide inhibited manganese catalases. Cyclovoltammetric data show that the substitution of oxygen donor atoms with chloride causes a shift of redox potentials to more positive values. To our knowledge, complex 1 is the most efficient binuclear functional manganese catalase mimic exhibiting saturation kinetics to date.

  11. Energy-Efficient Office Buildings at High Latitudes

    Energy Technology Data Exchange (ETDEWEB)

    Lerum, V.

    1996-12-31

    This doctoral thesis describes a method for energy efficient office building design at high latitudes and cold climates. The method combines daylighting, passive solar heating, solar protection, and ventilative cooling. The thesis focuses on optimal design of an equatorial-facing fenestration system. A spreadsheet framework linking existing simplified methods is used. The daylight analysis uses location specific data on frequency distribution of diffuse daylight on vertical surfaces to estimate energy savings from optimal window and room configurations in combination with a daylight-responsive electric lighting system. The passive solar heating analysis is a generalization of a solar load ratio method adapted to cold climates by combining it with the Norwegian standard NS3031 for winter months when the solar savings fraction is negative. The emphasis is on very high computational efficiency to permit rapid and comprehensive examination of a large number of options early in design. The procedure is illustrated for a location in Trondheim, Norway, testing the relative significance of various design improvement options relative to a base case. The method is also tested for two other locations in Norway, at latitudes 58 and 70 degrees North. The band of latitudes between these limits covers cities in Alaska, Canada, Greenland, Iceland, Scandinavia, Finland, Russia, and Northern Japan. A comprehensive study of the ``whole building approach`` shows the impact of integrated daylighting and low-energy design strategies. In general, consumption of lighting electricity may be reduced by 50-80%, even at extremely high latitudes. The reduced internal heat from electric lights is replaced by passive solar heating. 113 refs., 85 figs., 25 tabs.

  12. Photosynthetic Diurnal Variation of Soybean Cultivars with High Photosynthetic Efficiency

    Institute of Scientific and Technical Information of China (English)

    MAN Wei-qun; DU Wei-guang; ZHANG Gui-ru; LUAN Xiao-yan; GE Qiao-ying; HAO Nai-bin; CHEN Yi

    2002-01-01

    The photosynthetic characters were investigated among soybean cultivars with high photosynthetic efficiency and high yield. The results indicated that: 1) There were significant differences in photosynthetic rate (Ph) and dark respiration rate (DR) under saturation light intensity and appropriate temperature.2) There were a little difference in light compensation point among them. Photo flux density (PFD) were mong the cultivars. Diurnal variation of Pn was shown a curve with two peaks. 4) The cultivars with high photosynthetic efficiency were subjected less to photoinhibition than that with high yield. Critical temperatures of photoinhibition in high photosynthetic efficiency cultivars were higher than that of high yield.

  13. Efficient computation of coherent synchrotron radiation in a rectangular chamber

    Science.gov (United States)

    Warnock, Robert L.; Bizzozero, David A.

    2016-09-01

    We study coherent synchrotron radiation (CSR) in a perfectly conducting vacuum chamber of rectangular cross section, in a formalism allowing an arbitrary sequence of bends and straight sections. We apply the paraxial method in the frequency domain, with a Fourier development in the vertical coordinate but with no other mode expansions. A line charge source is handled numerically by a new method that rids the equations of singularities through a change of dependent variable. The resulting algorithm is fast compared to earlier methods, works for short bunches with complicated structure, and yields all six field components at any space-time point. As an example we compute the tangential magnetic field at the walls. From that one can make a perturbative treatment of the Poynting flux to estimate the energy deposited in resistive walls. The calculation was motivated by a design issue for LCLS-II, the question of how much wall heating from CSR occurs in the last bend of a bunch compressor and the following straight section. Working with a realistic longitudinal bunch form of r.m.s. length 10.4 μ m and a charge of 100 pC we conclude that the radiated power is quite small (28 W at a 1 MHz repetition rate), and all radiated energy is absorbed in the walls within 7 m along the straight section.

  14. Analog computation through high-dimensional physical chaotic neuro-dynamics

    Science.gov (United States)

    Horio, Yoshihiko; Aihara, Kazuyuki

    2008-07-01

    Conventional von Neumann computers have difficulty in solving complex and ill-posed real-world problems. However, living organisms often face such problems in real life, and must quickly obtain suitable solutions through physical, dynamical, and collective computations involving vast assemblies of neurons. These highly parallel computations through high-dimensional dynamics (computation through dynamics) are completely different from the numerical computations on von Neumann computers (computation through algorithms). In this paper, we explore a novel computational mechanism with high-dimensional physical chaotic neuro-dynamics. We physically constructed two hardware prototypes using analog chaotic-neuron integrated circuits. These systems combine analog computations with chaotic neuro-dynamics and digital computation through algorithms. We used quadratic assignment problems (QAPs) as benchmarks. The first prototype utilizes an analog chaotic neural network with 800-dimensional dynamics. An external algorithm constructs a solution for a QAP using the internal dynamics of the network. In the second system, 300-dimensional analog chaotic neuro-dynamics drive a tabu-search algorithm. We demonstrate experimentally that both systems efficiently solve QAPs through physical chaotic dynamics. We also qualitatively analyze the underlying mechanism of the highly parallel and collective analog computations by observing global and local dynamics. Furthermore, we introduce spatial and temporal mutual information to quantitatively evaluate the system dynamics. The experimental results confirm the validity and efficiency of the proposed computational paradigm with the physical analog chaotic neuro-dynamics.

  15. Robust fault detection of linear systems using a computationally efficient set-membership method

    DEFF Research Database (Denmark)

    Tabatabaeipour, Mojtaba; Bak, Thomas

    2014-01-01

    In this paper, a computationally efficient set-membership method for robust fault detection of linear systems is proposed. The method computes an interval outer-approximation of the output of the system that is consistent with the model, the bounds on noise and disturbance, and the past...

  16. Efficient Minimum-Phase Prefilter Computation Using Fast QL-Factorization

    DEFF Research Database (Denmark)

    Hansen, Morten; Christensen, Lars P.B.

    2009-01-01

    This paper presents a novel approach for computing both the minimum-phase filter and the associated all-pass filter in a computationally efficient way using the fast QL-factorization. A desirable property of this approach is that the complexity is independent on the size of the matrix which is QL...

  17. Computational Efficiency through Visual Argument: Do Graphic Organizers Communicate Relations in Text Too Effectively?

    Science.gov (United States)

    Robinson, Daniel H.; Schraw, Gregory

    1994-01-01

    Three experiments involving 138 college students investigated why one type of graphic organizer (a matrix) may communicate interconcept relations better than an outline or text. Results suggest that a matrix is more computationally efficient than either outline or text, allowing the easier computation of relationships. (SLD)

  18. High Performance Computing in Science and Engineering '02 : Transactions of the High Performance Computing Center

    CERN Document Server

    Jäger, Willi

    2003-01-01

    This book presents the state-of-the-art in modeling and simulation on supercomputers. Leading German research groups present their results achieved on high-end systems of the High Performance Computing Center Stuttgart (HLRS) for the year 2002. Reports cover all fields of supercomputing simulation ranging from computational fluid dynamics to computer science. Special emphasis is given to industrially relevant applications. Moreover, by presenting results for both vector sytems and micro-processor based systems the book allows to compare performance levels and usability of a variety of supercomputer architectures. It therefore becomes an indispensable guidebook to assess the impact of the Japanese Earth Simulator project on supercomputing in the years to come.

  19. Development of a computationally efficient urban flood modelling approach

    DEFF Research Database (Denmark)

    Wolfs, Vincent; Ntegeka, Victor; Murla, Damian

    the developed methodology, a case study for the city of Ghent in Belgium is elaborated. The configured conceptual model mimics the flood levels of a detailed 1D-2D hydrodynamic InfoWorks ICM model accurately, while the calculation time is an order of magnitude of 106 times shorter than the original highly...

  20. Efficient Skyline Computation in Structured Peer-to-Peer Systems

    DEFF Research Database (Denmark)

    Cui, Bin; Chen, Lijiang; Xu, Linhao;

    2009-01-01

    An increasing number of large-scale applications exploit peer-to-peer network architecture to provide highly scalable and flexible services. Among these applications, data management in peer-to-peer systems is one of the interesting domains. In this paper, we investigate the multidimensional skyl...

  1. Time efficient 3-D electromagnetic modeling on massively parallel computers

    Energy Technology Data Exchange (ETDEWEB)

    Alumbaugh, D.L.; Newman, G.A.

    1995-08-01

    A numerical modeling algorithm has been developed to simulate the electromagnetic response of a three dimensional earth to a dipole source for frequencies ranging from 100 to 100MHz. The numerical problem is formulated in terms of a frequency domain--modified vector Helmholtz equation for the scattered electric fields. The resulting differential equation is approximated using a staggered finite difference grid which results in a linear system of equations for which the matrix is sparse and complex symmetric. The system of equations is solved using a preconditioned quasi-minimum-residual method. Dirichlet boundary conditions are employed at the edges of the mesh by setting the tangential electric fields equal to zero. At frequencies less than 1MHz, normal grid stretching is employed to mitigate unwanted reflections off the grid boundaries. For frequencies greater than this, absorbing boundary conditions must be employed by making the stretching parameters of the modified vector Helmholtz equation complex which introduces loss at the boundaries. To allow for faster calculation of realistic models, the original serial version of the code has been modified to run on a massively parallel architecture. This modification involves three distinct tasks; (1) mapping the finite difference stencil to a processor stencil which allows for the necessary information to be exchanged between processors that contain adjacent nodes in the model, (2) determining the most efficient method to input the model which is accomplished by dividing the input into ``global`` and ``local`` data and then reading the two sets in differently, and (3) deciding how to output the data which is an inherently nonparallel process.

  2. High Efficiency Micromachining System Applied in Nanolithography

    Science.gov (United States)

    Chen, Xing; Lee, Dong Weon; Choi, Young Soo

    Scanning probe lithography such as direct-writing lithographic processes and nanoscratching techniques based on scanning probe microscopy have presented new micromachining methods for microelectromechanical system (MEMS). In this paper, a micromachining system for thermal scanning probe lithography is introduced, which consists of the cantilever arrays and a big stroke micro XY-stage. A large machining area and high machining speed can be realized by combining arrays of cantilevers possessing sharp tips at their top with the novel micro XY-stage which can obtain big displacements under relatively low driving voltage and in a small size. According to the above configuration, this micromachining system is provided with high throughputs and suitable for industrialization due to its MEMS-based simple fabrication process. The novel micro XY-stage applied in this system is presented in detail including the unique structure and principles, which shows an obvious improvement and distinct advantages in comparison with traditional structures. It is analyzed by mathematical model and then simulated using finite element method (FEM), it is proved to be able to practically serve the micromachining system with high capability.

  3. Cross-scale Efficient Tensor Contractions for Coupled Cluster Computations Through Multiple Programming Model Backends

    Energy Technology Data Exchange (ETDEWEB)

    Ibrahim, Khaled Z. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States). Computational Research Division; Epifanovsky, Evgeny [Q-Chem, Inc., Pleasanton, CA (United States); Williams, Samuel W. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States). Computational Research Division; Krylov, Anna I. [Univ. of Southern California, Los Angeles, CA (United States). Dept. of Chemistry

    2016-07-26

    Coupled-cluster methods provide highly accurate models of molecular structure by explicit numerical calculation of tensors representing the correlation between electrons. These calculations are dominated by a sequence of tensor contractions, motivating the development of numerical libraries for such operations. While based on matrix-matrix multiplication, these libraries are specialized to exploit symmetries in the molecular structure and in electronic interactions, and thus reduce the size of the tensor representation and the complexity of contractions. The resulting algorithms are irregular and their parallelization has been previously achieved via the use of dynamic scheduling or specialized data decompositions. We introduce our efforts to extend the Libtensor framework to work in the distributed memory environment in a scalable and energy efficient manner. We achieve up to 240 speedup compared with the best optimized shared memory implementation. We attain scalability to hundreds of thousands of compute cores on three distributed-memory architectures, (Cray XC30&XC40, BlueGene/Q), and on a heterogeneous GPU-CPU system (Cray XK7). As the bottlenecks shift from being compute-bound DGEMM's to communication-bound collectives as the size of the molecular system scales, we adopt two radically different parallelization approaches for handling load-imbalance. Nevertheless, we preserve a uni ed interface to both programming models to maintain the productivity of computational quantum chemists.

  4. Methods for Efficiently and Accurately Computing Quantum Mechanical Free Energies for Enzyme Catalysis.

    Science.gov (United States)

    Kearns, F L; Hudson, P S; Boresch, S; Woodcock, H L

    2016-01-01

    Enzyme activity is inherently linked to free energies of transition states, ligand binding, protonation/deprotonation, etc.; these free energies, and thus enzyme function, can be affected by residue mutations, allosterically induced conformational changes, and much more. Therefore, being able to predict free energies associated with enzymatic processes is critical to understanding and predicting their function. Free energy simulation (FES) has historically been a computational challenge as it requires both the accurate description of inter- and intramolecular interactions and adequate sampling of all relevant conformational degrees of freedom. The hybrid quantum mechanical molecular mechanical (QM/MM) framework is the current tool of choice when accurate computations of macromolecular systems are essential. Unfortunately, robust and efficient approaches that employ the high levels of computational theory needed to accurately describe many reactive processes (ie, ab initio, DFT), while also including explicit solvation effects and accounting for extensive conformational sampling are essentially nonexistent. In this chapter, we will give a brief overview of two recently developed methods that mitigate several major challenges associated with QM/MM FES: the QM non-Boltzmann Bennett's acceptance ratio method and the QM nonequilibrium work method. We will also describe usage of these methods to calculate free energies associated with (1) relative properties and (2) along reaction paths, using simple test cases with relevance to enzymes examples. © 2016 Elsevier Inc. All rights reserved.

  5. A computationally-efficient secondary organic aerosol module for three-dimensional air quality models

    Science.gov (United States)

    Liu, P.; Zhang, Y.

    2008-07-01

    Accurately simulating secondary organic aerosols (SOA) in three-dimensional (3-D) air quality models is challenging due to the complexity of the physics and chemistry involved and the high computational demand required. A computationally-efficient yet accurate SOA module is necessary in 3-D applications for long-term simulations and real-time air quality forecasting. A coupled gas and aerosol box model (i.e., 0-D CMAQ-MADRID 2) is used to optimize relevant processes in order to develop such a SOA module. Solving the partitioning equations for condensable volatile organic compounds (VOCs) and calculating their activity coefficients in the multicomponent mixtures are identified to be the most computationally-expensive processes. The two processes can be speeded up by relaxing the error tolerance levels and reducing the maximum number of iterations of the numerical solver for the partitioning equations for organic species; conditionally activating organic-inorganic interactions; and parameterizing the calculation of activity coefficients for organic mixtures in the hydrophilic module. The optimal speed-up method can reduce the total CPU cost by up to a factor of 31.4 from benchmark under the rural conditions with 2 ppb isoprene and by factors of 10 71 under various test conditions with 2 10 ppb isoprene and >40% relative humidity while maintaining ±15% deviation. These speed-up methods are applicable to other SOA modules that are based on partitioning theories.

  6. The path toward HEP High Performance Computing

    CERN Document Server

    Apostolakis, John; Carminati, Federico; Gheata, Andrei; Wenzel, Sandro

    2014-01-01

    High Energy Physics code has been known for making poor use of high performance computing architectures. Efforts in optimising HEP code on vector and RISC architectures have yield limited results and recent studies have shown that, on modern architectures, it achieves a performance between 10% and 50% of the peak one. Although several successful attempts have been made to port selected codes on GPUs, no major HEP code suite has a 'High Performance' implementation. With LHC undergoing a major upgrade and a number of challenging experiments on the drawing board, HEP cannot any longer neglect the less-than-optimal performance of its code and it has to try making the best usage of the hardware. This activity is one of the foci of the SFT group at CERN, which hosts, among others, the Root and Geant4 project. The activity of the experiments is shared and coordinated via a Concurrency Forum, where the experience in optimising HEP code is presented and discussed. Another activity is the Geant-V project, centred on th...

  7. Biologically inspired highly efficient buoyancy engine

    Science.gov (United States)

    Akle, Barbar; Habchi, Wassim; Abdelnour, Rita; Blottman, John, III; Leo, Donald

    2012-04-01

    Undersea distributed networked sensor systems require a miniaturization of platforms and a means of both spatial and temporal persistence. One aspect of this system is the necessity to modulate sensor depth for optimal positioning and station-keeping. Current approaches involve pneumatic bladders or electrolysis; both require mechanical subsystems and consume significant power. These are not suitable for the miniaturization of sensor platforms. Presented in this study is a novel biologically inspired method that relies on ionic motion and osmotic pressures to displace a volume of water from the ocean into and out of the proposed buoyancy engine. At a constant device volume, the displaced water will alter buoyancy leading to either sinking or floating. The engine is composed of an enclosure sided on the ocean's end by a Nafion ionomer and by a flexible membrane separating the water from a gas enclosure. Two electrodes are placed one inside the enclosure and the other attached to the engine on the outside. The semi-permeable membrane Nafion allows water motion in and out of the enclosure while blocking anions from being transferred. The two electrodes generate local concentration changes of ions upon the application of an electrical field; these changes lead to osmotic pressures and hence the transfer of water through the semi-permeable membrane. Some aquatic organisms such as pelagic crustacean perform this buoyancy control using an exchange of ions through their tissue to modulate its density relative to the ambient sea water. In this paper, the authors provide an experimental proof of concept of this buoyancy engine. The efficiency of changing the engine's buoyancy is calculated and optimized as a function of electrode surface area. For example electrodes made of a 3mm diameter Ag/AgCl proved to transfer approximately 4mm3 of water consuming 4 Joules of electrical energy. The speed of displacement is optimized as a function of the surface area of the Nafion

  8. Computational Properties of the Hippocampus Increase the Efficiency of Goal-Directed Foraging through Hierarchical Reinforcement Learning.

    Science.gov (United States)

    Chalmers, Eric; Luczak, Artur; Gruber, Aaron J

    2016-01-01

    The mammalian brain is thought to use a version of Model-based Reinforcement Learning (MBRL) to guide "goal-directed" behavior, wherein animals consider goals and make plans to acquire desired outcomes. However, conventional MBRL algorithms do not fully explain animals' ability to rapidly adapt to environmental changes, or learn multiple complex tasks. They also require extensive computation, suggesting that goal-directed behavior is cognitively expensive. We propose here that key features of processing in the hippocampus support a flexible MBRL mechanism for spatial navigation that is computationally efficient and can adapt quickly to change. We investigate this idea by implementing a computational MBRL framework that incorporates features inspired by computational properties of the hippocampus: a hierarchical representation of space, "forward sweeps" through future spatial trajectories, and context-driven remapping of place cells. We find that a hierarchical abstraction of space greatly reduces the computational load (mental effort) required for adaptation to changing environmental conditions, and allows efficient scaling to large problems. It also allows abstract knowledge gained at high levels to guide adaptation to new obstacles. Moreover, a context-driven remapping mechanism allows learning and memory of multiple tasks. Simulating dorsal or ventral hippocampal lesions in our computational framework qualitatively reproduces behavioral deficits observed in rodents with analogous lesions. The framework may thus embody key features of how the brain organizes model-based RL to efficiently solve navigation and other difficult tasks.

  9. Computational Properties of the Hippocampus Increase the Efficiency of Goal-Directed Foraging through Hierarchical Reinforcement Learning

    Directory of Open Access Journals (Sweden)

    Eric Chalmers

    2016-12-01

    Full Text Available The mammalian brain is thought to use a version of Model-based Reinforcement Learning (MBRL to guide goal-directed behavior, wherein animals consider goals and make plans to acquire desired outcomes. However, conventional MBRL algorithms do not fully explain animals’ ability to rapidly adapt to environmental changes, or learn multiple complex tasks. They also require extensive computation, suggesting that goal-directed behavior is cognitively expensive. We propose here that key features of processing in the hippocampus support a flexible MBRL mechanism for spatial navigation that is computationally efficient and can adapt quickly to change. We investigate this idea by implementing a computational MBRL framework that incorporates features inspired by computational properties of the hippocampus: a hierarchical representation of space, forward sweeps through future spatial trajectories, and context-driven remapping of place cells. We find that a hierarchical abstraction of space greatly reduces the computational load (mental effort required for adaptation to changing environmental conditions, and allows efficient scaling to large problems. It also allows abstract knowledge gained at high levels to guide adaptation to new obstacles. Moreover, a context-driven remapping mechanism allows learning and memory of multiple tasks. Simulating dorsal or ventral hippocampal lesions in our computational framework qualitatively reproduces behavioral deficits observed in rodents with analogous lesions. The framework may thus embody key features of how the brain organizes model-based RL to efficiently solve navigation and other difficult tasks.

  10. Slovak High School Students' Attitudes toward Computers

    Science.gov (United States)

    Kubiatko, Milan; Halakova, Zuzana; Nagyova, Sona; Nagy, Tibor

    2011-01-01

    The pervasive involvement of information and communication technologies and computers in our daily lives influences changes of attitude toward computers. We focused on finding these ecological effects in the differences in computer attitudes as a function of gender and age. A questionnaire with 34 Likert-type items was used in our research. The…

  11. High speed and large scale scientific computing

    CERN Document Server

    Gentzsch, W; Joubert, GR

    2010-01-01

    Over the years parallel technologies have completely transformed main stream computing. This book deals with the issues related to the area of cloud computing and discusses developments in grids, applications and information processing, as well as e-science. It is suitable for computer scientists, IT engineers and IT managers.

  12. High Speed Travelling Wave Single-Photon Detectors With Near-Unity Quantum Efficiency

    CERN Document Server

    Pernice, W; Minaeva, O; Li, M; Goltsman, G N; Sergienko, A V; Tang, H X

    2011-01-01

    Ultrafast, high quantum efficiency single photon detectors are among the most sought-after elements in modern quantum optics and quantum communication. Close-to-unity photon detection efficiency is essential for scalable measurement-based quantum computation, quantum key distribution, and loophole-free Bell experiments. However, imperfect modal matching and finite photon absorption rates have usually limited the maximum attainable detection efficiency of single photon detectors. Here we demonstrate a superconducting nanowire detector atop nanophotonic waveguides and achieve single photon detection efficiency up to 94% at telecom wavelengths. Our detectors are fully embedded in a scalable, low loss silicon photonic circuit and provide ultrashort timing jitter of 18ps at multi-GHz detection rates. Exploiting this high temporal resolution we demonstrate ballistic photon transport in silicon ring resonators. The direct implementation of such a detector with high quantum efficiency, high detection speed and low ji...

  13. 40 CFR 761.71 - High efficiency boilers.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 30 2010-07-01 2010-07-01 false High efficiency boilers. 761.71... PROHIBITIONS Storage and Disposal § 761.71 High efficiency boilers. (a) To burn mineral oil dielectric fluid containing a PCB concentration of ≥50 ppm, but boiler shall comply with the...

  14. High efficiency low cost GaAs/Ge cell technology

    Science.gov (United States)

    Ho, Frank

    1990-01-01

    Viewgraphs on high efficiency low cost GaAs/Ge cell technology are presented. Topics covered include: high efficiency, low cost GaAs/Ge solar cells; advantages of Ge; comparison of typical production cells for space applications; panel level comparisons; and solar cell technology trends.

  15. Energy Efficient and Compact RF High-Power Amplifiers

    NARCIS (Netherlands)

    Calvillo Cortés, D.A.

    2014-01-01

    The main objectives of this thesis are to improve the energy efficiency and physical form-factor of high-power amplifiers in base station applications. As such, the focus of this dissertation is placed on the outphasing amplifier concept, which can offer high-efficiency, good linearity and excellent

  16. Highly efficient carrier multiplication in PbS nanosheets

    NARCIS (Netherlands)

    Aerts, M.; Bielewicz, T.; Klinke, C.; Grozema, F.C.; Houtepen, A.J.; Schins, J.M.; Siebbeles, L.D.A.

    2014-01-01

    Semiconductor nanocrystals are promising for use in cheap and highly efficient solar cells. A high efficiency can be achieved by carrier multiplication (CM), which yields multiple electron-hole pairs for a single absorbed photon. Lead chalcogenide nanocrystals are of specific interest, since their b

  17. Next-generation sequencing: big data meets high performance computing.

    Science.gov (United States)

    Schmidt, Bertil; Hildebrandt, Andreas

    2017-02-02

    The progress of next-generation sequencing has a major impact on medical and genomic research. This high-throughput technology can now produce billions of short DNA or RNA fragments in excess of a few terabytes of data in a single run. This leads to massive datasets used by a wide range of applications including personalized cancer treatment and precision medicine. In addition to the hugely increased throughput, the cost of using high-throughput technologies has been dramatically decreasing. A low sequencing cost of around US$1000 per genome has now rendered large population-scale projects feasible. However, to make effective use of the produced data, the design of big data algorithms and their efficient implementation on modern high performance computing systems is required.

  18. Highly efficient sources of single indistinguishable photons

    DEFF Research Database (Denmark)

    Gregersen, Niels

    2013-01-01

    Solid-state sources capable of emitting single photons on demand are of great interest in quantum information applications. Ideally, such a source should emit exactly one photon into the collection optics per trigger, the emitted photons should be indistinguishable and the source should...... be electrically driven. Several design strategies addressing these requirements have been proposed. In the cavity-based source, light emission is controlled using resonant cavity quantum electrodynamics effects, whereas in the waveguide-based source, broadband electric field screening effects are employed...... to direct the light emission into the optical mode of interest. For all the strategies, accurate modeling and careful optical engineering is required to achieve high performance....

  19. Efficient and configurable transmission protocol based on UDP in grid computing

    Institute of Scientific and Technical Information of China (English)

    Jigang WANG; Guochang GU; Chunguang MA; Weidong ZHONG

    2009-01-01

    At present,mainstream data transfer protocols are not always a good match for the diverse demands of grid computing.Considering this situation,this article proposes an efficient and configurable data transfer protocol (ECUDP) for grid computing.The ECUDP is based on the standard user datagram protocol (UDP),but with a collection of optimizations that meet the challenge of providing configurability and reliability while main-taining performance that meets the communication requirements of demanding applications.Experimental results show that the ECUDP performs efficiently in various grid computing scenarios and the performance analysis model can provide a good estimation of its performance.

  20. X-Ray Line Measurements with High Efficiency Bragg Crystals

    Energy Technology Data Exchange (ETDEWEB)

    Pak, A; Gregori, G; Knight, J; Campbell, K; Landen, O; Glenzer, S

    2004-04-01

    We have studied the focusing properties of two highly oriented pyrolitic graphite (HOPG) spectrometers, which differ in the degree of the mosaic spread: ZYA with a low mosaic spread ({gamma}=0.4 degrees) and ZYH with a large mosaic spread ({gamma}=3.5 degrees). In order to assess the crystal performance for a variety of different experiments, various K{alpha} and K{beta} x-ray lines have been produced using a high-intensity ({approx}>10{sup 17} W/cm{sup 2}) short-pulse ({approx} 100 fs) laser beam focused onto Ti, V, Zn, and Cu foils. The measured spectral resolution of the HOPG crystals in both first and second order diffraction has been compared with theoretical predictions. Using known values for the peak reflectivity of HOPG crystals, we have also computed K{alpha} x-ray conversion efficiencies of Ti, V, Zn, and Cu. These results are important to estimate the optimal conditions under which different types of HOPG monochromators can be used for the detection of weak x-ray signals as the one encountered in x-ray Thomson/Compton scattering experiments.

  1. Improving Computational Efficiency of Model Predictive Control Genetic Algorithms for Real-Time Decision Support

    Science.gov (United States)

    Minsker, B. S.; Zimmer, A. L.; Ostfeld, A.; Schmidt, A.

    2014-12-01

    Enabling real-time decision support, particularly under conditions of uncertainty, requires computationally efficient algorithms that can rapidly generate recommendations. In this paper, a suite of model predictive control (MPC) genetic algorithms are developed and tested offline to explore their value for reducing CSOs during real-time use in a deep-tunnel sewer system. MPC approaches include the micro-GA, the probability-based compact GA, and domain-specific GA methods that reduce the number of decision variable values analyzed within the sewer hydraulic model, thus reducing algorithm search space. Minimum fitness and constraint values achieved by all GA approaches, as well as computational times required to reach the minimum values, are compared to large population sizes with long convergence times. Optimization results for a subset of the Chicago combined sewer system indicate that genetic algorithm variations with coarse decision variable representation, eventually transitioning to the entire range of decision variable values, are most efficient at addressing the CSO control problem. Although diversity-enhancing micro-GAs evaluate a larger search space and exhibit shorter convergence times, these representations do not reach minimum fitness and constraint values. The domain-specific GAs prove to be the most efficient and are used to test CSO sensitivity to energy costs, CSO penalties, and pressurization constraint values. The results show that CSO volumes are highly dependent on the tunnel pressurization constraint, with reductions of 13% to 77% possible with less conservative operational strategies. Because current management practices may not account for varying costs at CSO locations and electricity rate changes in the summer and winter, the sensitivity of the results is evaluated for variable seasonal and diurnal CSO penalty costs and electricity-related system maintenance costs, as well as different sluice gate constraint levels. These findings indicate

  2. Compact and highly efficient laser pump cavity

    Energy Technology Data Exchange (ETDEWEB)

    Chang, Jim J. (Dublin, CA); Bass, Isaac L. (Castro Valley, CA); Zapata, Luis E. (Livermore, CA)

    1999-01-01

    A new, compact, side-pumped laser pump cavity design which uses non-conventional optics for injection of laser-diode light into a laser pump chamber includes a plurality of elongated light concentration channels. In one embodiment, the light concentration channels are compound parabolic concentrators (CPC) which have very small exit apertures so that light will not escape from the pumping chamber and will be multiply reflected through the laser rod. This new design effectively traps the pump radiation inside the pump chamber that encloses the laser rod. It enables more uniform laser pumping and highly effective recycle of pump radiation, leading to significantly improved laser performance. This new design also effectively widens the acceptable radiation wavelength of the diodes, resulting in a more reliable laser performance with lower cost.

  3. 2250-MHz High Efficiency Microwave Power Amplifier (HEMPA)

    Science.gov (United States)

    Sims, W. Herbert; Bell, Joseph L. (Technical Monitor)

    2001-01-01

    Tnis paper will focus on developing an exotic switching technique that enhances the DC-to-RF conversion efficiency of microwave power amplifiers. For years, switching techniques implemented in the 10 kHz to 30 MHz region have resulted in DC-to-RF conversion efficiencies of 90-95-percent. Currently amplifier conversion efficiency, in the 2-3 GHz region approaches, 10-20-percent. Using a combination of analytical modeling and hardware testing, a High Efficiency Microwave Power Amplifier was built that demonstrated conversion efficiencies four to five times higher than current state of the art.

  4. Computationally generated velocity taper for efficiency enhancement in a coupled-cavity traveling-wave tube

    Science.gov (United States)

    Wilson, Jeffrey D.

    1989-01-01

    A computational routine has been created to generate velocity tapers for efficiency enhancement in coupled-cavity TWTs. Programmed into the NASA multidimensional large-signal coupled-cavity TWT computer code, the routine generates the gradually decreasing cavity periods required to maintain a prescribed relationship between the circuit phase velocity and the electron-bunch velocity. Computational results for several computer-generated tapers are compared to those for an existing coupled-cavity TWT with a three-step taper. Guidelines are developed for prescribing the bunch-phase profile to produce a taper for efficiency. The resulting taper provides a calculated RF efficiency 45 percent higher than the step taper at center frequency and at least 37 percent higher over the bandwidth.

  5. Toward High-Power Klystrons With RF Power Conversion Efficiency on the Order of 90%

    CERN Document Server

    Baikov, Andrey Yu; Syratchev, Igor

    2015-01-01

    The increase in efficiency of RF power generation for future large accelerators is considered a high priority issue. The vast majority of the existing commercial high-power RF klystrons operates in the electronic efficiency range between 40% and 55%. Only a few klystrons available on the market are capable of operating with 65% efficiency or above. In this paper, a new method to achieve 90% RF power conversion efficiency in a klystron amplifier is presented. The essential part of this method is a new bunching technique - bunching with bunch core oscillations. Computer simulations confirm that the RF production efficiency above 90% can be reached with this new bunching method. The results of a preliminary study of an L-band, 20-MW peak RF power multibeam klystron for Compact Linear Collider with the efficiency above 85% are presented.

  6. High Performance Computing in Science and Engineering '98 : Transactions of the High Performance Computing Center

    CERN Document Server

    Jäger, Willi

    1999-01-01

    The book contains reports about the most significant projects from science and industry that are using the supercomputers of the Federal High Performance Computing Center Stuttgart (HLRS). These projects are from different scientific disciplines, with a focus on engineering, physics and chemistry. They were carefully selected in a peer-review process and are showcases for an innovative combination of state-of-the-art physical modeling, novel algorithms and the use of leading-edge parallel computer technology. As HLRS is in close cooperation with industrial companies, special emphasis has been put on the industrial relevance of results and methods.

  7. High Performance Computing in Science and Engineering '99 : Transactions of the High Performance Computing Center

    CERN Document Server

    Jäger, Willi

    2000-01-01

    The book contains reports about the most significant projects from science and engineering of the Federal High Performance Computing Center Stuttgart (HLRS). They were carefully selected in a peer-review process and are showcases of an innovative combination of state-of-the-art modeling, novel algorithms and the use of leading-edge parallel computer technology. The projects of HLRS are using supercomputer systems operated jointly by university and industry and therefore a special emphasis has been put on the industrial relevance of results and methods.

  8. Resilient and Robust High Performance Computing Platforms for Scientific Computing Integrity

    Energy Technology Data Exchange (ETDEWEB)

    Jin, Yier [Univ. of Central Florida, Orlando, FL (United States)

    2017-07-14

    As technology advances, computer systems are subject to increasingly sophisticated cyber-attacks that compromise both their security and integrity. High performance computing platforms used in commercial and scientific applications involving sensitive, or even classified data, are frequently targeted by powerful adversaries. This situation is made worse by a lack of fundamental security solutions that both perform efficiently and are effective at preventing threats. Current security solutions fail to address the threat landscape and ensure the integrity of sensitive data. As challenges rise, both private and public sectors will require robust technologies to protect its computing infrastructure. The research outcomes from this project try to address all these challenges. For example, we present LAZARUS, a novel technique to harden kernel Address Space Layout Randomization (KASLR) against paging-based side-channel attacks. In particular, our scheme allows for fine-grained protection of the virtual memory mappings that implement the randomization. We demonstrate the effectiveness of our approach by hardening a recent Linux kernel with LAZARUS, mitigating all of the previously presented side-channel attacks on KASLR. Our extensive evaluation shows that LAZARUS incurs only 0.943% overhead for standard benchmarks, and is therefore highly practical. We also introduced HA2lloc, a hardware-assisted allocator that is capable of leveraging an extended memory management unit to detect memory errors in the heap. We also perform testing using HA2lloc in a simulation environment and find that the approach is capable of preventing common memory vulnerabilities.

  9. Energy efficient engine high-pressure turbine detailed design report

    Science.gov (United States)

    Thulin, R. D.; Howe, D. C.; Singer, I. D.

    1982-01-01

    The energy efficient engine high-pressure turbine is a single stage system based on technology advancements in the areas of aerodynamics, structures and materials to achieve high performance, low operating economics and durability commensurate with commercial service requirements. Low loss performance features combined with a low through-flow velocity approach results in a predicted efficiency of 88.8 for a flight propulsion system. Turbine airfoil durability goals are achieved through the use of advanced high-strength and high-temperature capability single crystal materials and effective cooling management. Overall, this design reflects a considerable extension in turbine technology that is applicable to future, energy efficient gas-turbine engines.

  10. Computationally efficient SVM multi-class image recognition with confidence measures

    Energy Technology Data Exchange (ETDEWEB)

    Makili, Lazaro [Dpto. Informatica y Automatica - UNED, Madrid (Spain); Vega, Jesus, E-mail: jesus.vega@ciemat.es [Asociacion EURATOM/CIEMAT para Fusion, Madrid (Spain); Dormido-Canto, Sebastian [Dpto. Informatica y Automatica - UNED, Madrid (Spain); Pastor, Ignacio [Asociacion EURATOM/CIEMAT para Fusion, Madrid (Spain); Murari, Andrea [Associazione EURATOM-CIEMAT per la Fusione, Consorzio RFX, Padova (Italy)

    2011-10-15

    Typically, machine learning methods produce non-qualified estimates, i.e. the accuracy and reliability of the predictions are not provided. Transductive predictors are very recent classifiers able to provide, simultaneously with the prediction, a couple of values (confidence and credibility) to reflect the quality of the prediction. Usually, a drawback of the transductive techniques for huge datasets and large dimensionality is the high computational time. To overcome this issue, a more efficient classifier has been used in a multi-class image classification problem in the TJ-II stellarator database. It is based on the creation of a hash function to generate several 'one versus the rest' classifiers for every class. By using Support Vector Machines as the underlying classifier, a comparison between the pure transductive approach and the new method has been performed. In both cases, the success rates are high and the computation time with the new method is up to 0.4 times the old one.

  11. ZrOCl2.8H2O on montmorillonite K10: an efficient green catalyst for One-pot synthesis of highly substituted imidazoles under Solvent-Free Conditions: DFT Computational investigation

    Directory of Open Access Journals (Sweden)

    Ahmad Movahedpour

    2017-01-01

    Full Text Available ZrOCl2.8H2O/K10 has been found to be an extremely efficient catalyst for the preparation of synthesis of 2,4,5- trisubstituted and 1,2,4,5-tetra substituted imidazoles in excellent yields. This catalystexhibited remarkable catalytic activity with respect to the reaction time (2-3min, amount of catalyst. Present protocol using ZrOCl2.8H2O/K10 as catalyst is convincingly superior to the recently reported methodologies. Density functional theory (DFT method and B3LYP/6-31G(d,p level of theory were used to determine some parameters such as vibrational modes, dipole moment (μ, EHOMO, ELUMO and gap energy (∆E of compounds.

  12. High efficiency in human muscle: an anomaly and an opportunity?

    Science.gov (United States)

    Nelson, Frank E; Ortega, Justus D; Jubrias, Sharon A; Conley, Kevin E; Kushmerick, Martin J

    2011-08-15

    Can human muscle be highly efficient in vivo? Animal muscles typically show contraction-coupling efficiencies FDI) muscle of the hand has an efficiency value in vivo of 68%. We examine two key factors that could account for this apparently high efficiency value: (1) transfer of cross-bridge work into mechanical work and (2) the use of elastic energy to do external work. Our analysis supports a high contractile efficiency reflective of nearly complete transfer of muscular to mechanical work with no contribution by recycling of elastic energy to mechanical work. Our survey of reported contraction-coupling efficiency values puts the FDI value higher than typical values found in small animals in vitro but within the range of values for human muscle in vivo. These high efficiency values support recent studies that suggest lower Ca(2+) cycling costs in working contractions and a decline in cost during repeated contractions. In the end, our analysis indicates that the FDI muscle may be exceptional in having an efficiency value on the higher end of that reported for human muscle. Thus, the FDI muscle may be an exception both in contraction-coupling efficiency and in Ca(2+) cycling costs, which makes it an ideal muscle model system offering prime conditions for studying the energetics of muscle contraction in vivo.

  13. Efficient Computation of Power, Force, and Torque in BEM Scattering Calculations

    CERN Document Server

    Reid, M T Homer

    2013-01-01

    We present concise, computationally efficient formulas for several quantities of interest -- including absorbed and scattered power, optical force (radiation pressure), and torque -- in scattering calculations performed using the boundary-element method (BEM) [also known as the method of moments (MOM)]. Our formulas compute the quantities of interest \\textit{directly} from the BEM surface currents with no need ever to compute the scattered electromagnetic fields. We derive our new formulas and demonstrate their effectiveness by computing power, force, and torque in a number of example geometries. Free, open-source software implementations of our formulas are available for download online.

  14. Energy- and cost-efficient lattice-QCD computations using graphics processing units

    Energy Technology Data Exchange (ETDEWEB)

    Bach, Matthias

    2014-07-01

    Quarks and gluons are the building blocks of all hadronic matter, like protons and neutrons. Their interaction is described by Quantum Chromodynamics (QCD), a theory under test by large scale experiments like the Large Hadron Collider (LHC) at CERN and in the future at the Facility for Antiproton and Ion Research (FAIR) at GSI. However, perturbative methods can only be applied to QCD for high energies. Studies from first principles are possible via a discretization onto an Euclidean space-time grid. This discretization of QCD is called Lattice QCD (LQCD) and is the only ab-initio option outside of the high-energy regime. LQCD is extremely compute and memory intensive. In particular, it is by definition always bandwidth limited. Thus - despite the complexity of LQCD applications - it led to the development of several specialized compute platforms and influenced the development of others. However, in recent years General-Purpose computation on Graphics Processing Units (GPGPU) came up as a new means for parallel computing. Contrary to machines traditionally used for LQCD, graphics processing units (GPUs) are a massmarket product. This promises advantages in both the pace at which higher-performing hardware becomes available and its price. CL2QCD is an OpenCL based implementation of LQCD using Wilson fermions that was developed within this thesis. It operates on GPUs by all major vendors as well as on central processing units (CPUs). On the AMD Radeon HD 7970 it provides the fastest double-precision D kernel for a single GPU, achieving 120GFLOPS. D - the most compute intensive kernel in LQCD simulations - is commonly used to compare LQCD platforms. This performance is enabled by an in-depth analysis of optimization techniques for bandwidth-limited codes on GPUs. Further, analysis of the communication between GPU and CPU, as well as between multiple GPUs, enables high-performance Krylov space solvers and linear scaling to multiple GPUs within a single system. LQCD

  15. Application of a computationally efficient method to approximate gap model results with a probabilistic approach

    Science.gov (United States)

    Scherstjanoi, M.; Kaplan, J. O.; Lischke, H.

    2014-07-01

    To be able to simulate climate change effects on forest dynamics over the whole of Switzerland, we adapted the second-generation DGVM (dynamic global vegetation model) LPJ-GUESS (Lund-Potsdam-Jena General Ecosystem Simulator) to the Alpine environment. We modified model functions, tuned model parameters, and implemented new tree species to represent the potential natural vegetation of Alpine landscapes. Furthermore, we increased the computational efficiency of the model to enable area-covering simulations in a fine resolution (1 km) sufficient for the complex topography of the Alps, which resulted in more than 32 000 simulation grid cells. To this aim, we applied the recently developed method GAPPARD (approximating GAP model results with a Probabilistic Approach to account for stand Replacing Disturbances) (Scherstjanoi et al., 2013) to LPJ-GUESS. GAPPARD derives mean output values from a combination of simulation runs without disturbances and a patch age distribution defined by the disturbance frequency. With this computationally efficient method, which increased the model's speed by approximately the factor 8, we were able to faster detect the shortcomings of LPJ-GUESS functions and parameters. We used the adapted LPJ-GUESS together with GAPPARD to assess the influence of one climate change scenario on dynamics of tree species composition and biomass throughout the 21st century in Switzerland. To allow for comparison with the original model, we additionally simulated forest dynamics along a north-south transect through Switzerland. The results from this transect confirmed the high value of the GAPPARD method despite some limitations towards extreme climatic events. It allowed for the first time to obtain area-wide, detailed high-resolution LPJ-GUESS simulation results for a large part of the Alpine region.

  16. High brilliance and high efficiency: optimized high power diode laser bars

    Science.gov (United States)

    Hülsewede, R.; Schulze, H.; Sebastian, J.; Schröder, D.; Meusel, J.; Hennig, P.

    2008-02-01

    The strong increasing laser market has ongoing demands to reduce the costs of diode laser pumped systems. For that reason JENOPTIK Diode Lab GmbH (JDL) optimized the bar brilliance (small vertical far field divergence) and bar efficiency (higher optical power operation) with respect to the pump applications. High efficiency reduces the costs for mounting and cooling and high brilliance increases the coupling efficiency. Both are carefully adjusted in the 9xx nm - high power diode laser bars for pump applications in disc- and fiber lasers. Based on low loss waveguide structures high brilliance bars with 19° fast axis beam divergence (FWHM) with 58 % maximum efficiency and 27° fast axis beam divergence (FWHM) with 62 % maximum efficiency are developed. Mounted on conductive cooled heat sinks high power operation with lifetime > 20.000 hours at 120 W output power level (50 % filling factor bars) and 80W (20 % filling factor bars) is demonstrated. 808nm bars used as pump sources for Nd:YAG solid state lasers are still dominating in the market. With respect to the demands on high reliability at high power operation current results of a 100 W high power life time test are showing more than 9000 hour operation time for passively cooled packaged high efficiency 50 % filling factor bars. Measurement of the COMD-level after this hard pulse life time test demonstrates very high power levels with no significant droop in COMD-power level. This confirms the high facet stability of JDL's facet technology. New high power diode laser bars with wavelength of 825 nm and 885 nm are still under development and first results are presented.

  17. Dense, Efficient Chip-to-Chip Communication at the Extremes of Computing

    Science.gov (United States)

    Loh, Matthew

    2013-01-01

    The scalability of CMOS technology has driven computation into a diverse range of applications across the power consumption, performance and size spectra. Communication is a necessary adjunct to computation, and whether this is to push data from node-to-node in a high-performance computing cluster or from the receiver of wireless link to a neural…

  18. Software Synthesis for High Productivity Exascale Computing

    Energy Technology Data Exchange (ETDEWEB)

    Bodik, Rastislav [Univ. of Washington, Seattle, WA (United States)

    2010-09-01

    Over the three years of our project, we accomplished three key milestones: We demonstrated how ideas from generative programming and software synthesis can help support the development of bulk-synchronous distributed memory kernels. These ideas are realized in a new language called MSL, a C-like language that combines synthesis features with high level notations for array manipulation and bulk-synchronous parallelism to simplify the semantic analysis required for synthesis. We also demonstrated that these high level notations map easily to low level C code and show that the performance of this generated code matches that of handwritten Fortran. Second, we introduced the idea of solver-aided domain-specific languages (SDSLs), which are an emerging class of computer-aided programming systems. SDSLs ease the construction of programs by automating tasks such as verification, debugging, synthesis, and non-deterministic execution. SDSLs are implemented by translating the DSL program into logical constraints. Next, we developed a symbolic virtual machine called Rosette, which simplifies the construction of such SDSLs and their compilers. We have used Rosette to build SynthCL, a subset of OpenCL that supports synthesis. Third, we developed novel numeric algorithms that move as little data as possible, either between levels of a memory hierarchy or between parallel processors over a network. We achieved progress in three aspects of this problem. First we determined lower bounds on communication. Second, we compared these lower bounds to widely used versions of these algorithms, and noted that these widely used algorithms usually communicate asymptotically more than is necessary. Third, we identified or invented new algorithms for most linear algebra problems that do attain these lower bounds, and demonstrated large speed-ups in theory and practice.

  19. Highly-Efficient and Modular Medium-Voltage Converters

    Science.gov (United States)

    2015-09-28

    4. TITLE AND SUBTITLE Highly-Efficient and Modula Medium -Voltage Converters 6. AUTHOR(S) Maryam Saeedifard 7. PERFORMING ORGANIZATIC i NAME(S...realization of highly efficient, modular medium - voltage dc-ac and dc-dc energy conversion systems by development of new control strategies that improve the...Z39.18 a 01^ 100(0^5 Final Report for Grant N00014-14-1-0615 Highly-Efficient and Modular Medium -Voltage Converters Lead Organization: Georgia Tech

  20. High-performance Scientific Computing using Parallel Computing to Improve Performance Optimization Problems

    Directory of Open Access Journals (Sweden)

    Florica Novăcescu

    2011-10-01

    Full Text Available HPC (High Performance Computing has become essential for the acceleration of innovation and the companies’ assistance in creating new inventions, better models and more reliable products as well as obtaining processes and services at low costs. The information in this paper focuses particularly on: description the field of high performance scientific computing, parallel computing, scientific computing, parallel computers, and trends in the HPC field, presented here reveal important new directions toward the realization of a high performance computational society. The practical part of the work is an example of use of the HPC tool to accelerate solving an electrostatic optimization problem using the Parallel Computing Toolbox that allows solving computational and data-intensive problems using MATLAB and Simulink on multicore and multiprocessor computers.

  1. Computational Thinking and Practice - A Generic Approach to Computing in Danish High Schools

    DEFF Research Database (Denmark)

    Caspersen, Michael E.; Nowack, Palle

    2014-01-01

    Internationally, there is a growing awareness on the necessity of providing relevant computing education in schools, particularly high schools. We present a new and generic approach to Computing in Danish High Schools based on a conceptual framework derived from ideas related to computational thi...

  2. High Fidelity Adiabatic Quantum Computation via Dynamical Decoupling

    CERN Document Server

    Quiroz, Gregory

    2012-01-01

    We introduce high-order dynamical decoupling strategies for open system adiabatic quantum computation. Our numerical results demonstrate that a judicious choice of high-order dynamical decoupling method, in conjunction with an encoding which allows computation to proceed alongside decoupling, can dramatically enhance the fidelity of adiabatic quantum computation in spite of decoherence.

  3. High-reliability computing for the smarter planet

    Energy Technology Data Exchange (ETDEWEB)

    Quinn, Heather M [Los Alamos National Laboratory; Graham, Paul [Los Alamos National Laboratory; Manuzzato, Andrea [UNIV OF PADOVA; Dehon, Andre [UNIV OF PENN; Carter, Nicholas [INTEL CORPORATION

    2010-01-01

    The geometric rate of improvement of transistor size and integrated circuit performance, known as Moore's Law, has been an engine of growth for our economy, enabling new products and services, creating new value and wealth, increasing safety, and removing menial tasks from our daily lives. Affordable, highly integrated components have enabled both life-saving technologies and rich entertainment applications. Anti-lock brakes, insulin monitors, and GPS-enabled emergency response systems save lives. Cell phones, internet appliances, virtual worlds, realistic video games, and mp3 players enrich our lives and connect us together. Over the past 40 years of silicon scaling, the increasing capabilities of inexpensive computation have transformed our society through automation and ubiquitous communications. In this paper, we will present the concept of the smarter planet, how reliability failures affect current systems, and methods that can be used to increase the reliable adoption of new automation in the future. We will illustrate these issues using a number of different electronic devices in a couple of different scenarios. Recently IBM has been presenting the idea of a 'smarter planet.' In smarter planet documents, IBM discusses increased computer automation of roadways, banking, healthcare, and infrastructure, as automation could create more efficient systems. A necessary component of the smarter planet concept is to ensure that these new systems have very high reliability. Even extremely rare reliability problems can easily escalate to problematic scenarios when implemented at very large scales. For life-critical systems, such as automobiles, infrastructure, medical implantables, and avionic systems, unmitigated failures could be dangerous. As more automation moves into these types of critical systems, reliability failures will need to be managed. As computer automation continues to increase in our society, the need for greater radiation reliability is

  4. Study of application technology of ultra-high speed computer to the elucidation of complex phenomena

    Energy Technology Data Exchange (ETDEWEB)

    Sekiguchi, Tomotsugu [Electrotechnical Lab., Tsukuba, Ibaraki (Japan)

    1996-06-01

    The basic design of numerical information library in the decentralized computer network was explained at the first step of constructing the application technology of ultra-high speed computer to the elucidation of complex phenomena. Establishment of the system makes possible to construct the efficient application environment of ultra-high speed computer system to be scalable with the different computing systems. We named the system Ninf (Network Information Library for High Performance Computing). The summary of application technology of library was described as follows: the application technology of library under the distributed environment, numeric constants, retrieval of value, library of special functions, computing library, Ninf library interface, Ninf remote library and registration. By the system, user is able to use the program concentrating the analyzing technology of numerical value with high precision, reliability and speed. (S.Y.)

  5. High efficiency interdigitated back contact silicon solar cells

    Science.gov (United States)

    Verlinden, P.; van de Wiele, F.; Stehelin, G.; Floret, F.; David, J. P.

    Interdigitated back contact (IBC) silicon solar cells with 25.6 percent efficiency at 10 W/sq cm and 24.4 percent at 30 W/ sq cm were fabricated. The authors report on the technological process, which produces a high effective carrier lifetime in the bulk (780 microsec), and on the characterization of the cells. The front side of these cells is textured and has an efficient polka-dot floating tandem junction. IBC and point-contact (PC) cells are fabricated on the same substrate and their efficiencies are compared. The possiblity of reaching 29 percent efficiency at 300X is shown.

  6. Multi Objective Optimization for Calibration and Efficient Uncertainty Analysis of Computationally Expensive Watershed Models

    Science.gov (United States)

    Akhtar, T.; Shoemaker, C. A.

    2011-12-01

    Assessing the sensitivity of calibration results to different calibration criteria can be done through multi objective optimization that considers multiple calibration criteria. This analysis can be extended to uncertainty analysis by comparing the results of simulation of the model with parameter sets from many points along a Pareto Front. In this study we employ multi-objective optimization in order to understand which parameter values should be used for flow parameters of a SWAT model, (Soil and Water Assessment Tool) designed to simulate flow in the Cannonsville Reservoir in upstate New York. The comprehensive analysis procedure encapsulates identification of suitable objectives, analysis of trade-offs obtained through multi-objective optimization, and the impact of the trade-offs uncertainty. Examples of multiple criteria can include a) quality of the fit in different seasons, b) quality of the fit for high flow events and for low flow events, c) quality of the fit for different constituents (e.g. water versus nutrients). Many distributed watershed models are computationally expensive and include a large number of parameters that are to be calibrated. Efficient optimization algorithms are hence needed to find good solutions to multi-criteria calibration problems in a feasible amount of time. We apply a new algorithm called Gap Optimized Multi-Objective Optimization using Response Surfaces (GOMORS), for efficient multi-criteria optimization of the Cannonsville SWAT watershed calibration problem. GOMORS is a stochastic optimization method, which makes use of Radial Basis Functions for approximation of the computationally expensive objectives. GOMORS performance is also compared against other multi-objective algorithms ParEGO and NSGA-II. ParEGO is a kriging based efficient multi-objective optimization algorithm, whereas NSGA-II is a well-known multi-objective evolutionary optimization algorithm. GOMORS is more efficient than both ParEGO and NSGA-II in providing

  7. Efficient high-precision matrix algebra on parallel architectures for nonlinear combinatorial optimization

    KAUST Repository

    Gunnels, John

    2010-06-01

    We provide a first demonstration of the idea that matrix-based algorithms for nonlinear combinatorial optimization problems can be efficiently implemented. Such algorithms were mainly conceived by theoretical computer scientists for proving efficiency. We are able to demonstrate the practicality of our approach by developing an implementation on a massively parallel architecture, and exploiting scalable and efficient parallel implementations of algorithms for ultra high-precision linear algebra. Additionally, we have delineated and implemented the necessary algorithmic and coding changes required in order to address problems several orders of magnitude larger, dealing with the limits of scalability from memory footprint, computational efficiency, reliability, and interconnect perspectives. © Springer and Mathematical Programming Society 2010.

  8. An Efficient Adaptive Load Balancing Algorithm for Cloud Computing Under Bursty Workloads

    Directory of Open Access Journals (Sweden)

    S. F. Issawi

    2015-06-01

    Full Text Available Cloud computing is a recent, emerging technology in the IT industry. It is an evolution of previous models such as grid computing. It enables a wide range of users to access a large sharing pool of resources over the internet. In such complex system, there is a tremendous need for an efficient load balancing scheme in order to satisfy peak user demands and provide high quality of services. One of the challenging problems that degrade the performance of a load balancing process is bursty workloads. Although there are a lot of researches proposing different load balancing algorithms, most of them neglect the problem of bursty workloads. Motivated by this problem, this paper proposes a new burstness-aware load balancing algorithm which can adapt to the variation in the request rate by adopting two load balancing algorithms: RR in burst and Random in non-burst state. Fuzzy logic is used in order to assign the received request to a balanced VM. The algorithm has been evaluated and compared with other algorithms using Cloud Analyst simulator. Results show that the proposed algorithm improves the average response time and average processing time in comparison with other algorithms.

  9. Highly Efficient Solid Oxide Electrolyzer & Sabatier System Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Paragon Space Development Corporation (Paragon) and ENrG Incorporated (ENrG) are teaming to provide a highly efficient reactor for carbon monoxide/carbon dioxide...

  10. Highly Efficient Solid Oxide Electrolyzer & Sabatier System Project

    Data.gov (United States)

    National Aeronautics and Space Administration — Paragon Space Development Corporation® (Paragon) and ENrG Incorporated (ENrG) are teaming to provide a highly efficient reactor for carbon monoxide/carbon...

  11. High Efficiency Lighting with Integrated Adaptive Control (HELIAC) Project

    Data.gov (United States)

    National Aeronautics and Space Administration — The proposed project is the continued development of the High Efficiency Lighting with Integrated Adaptive Control (HELIAC) system. Solar radiation is not a viable...

  12. Compact Tunable High-Efficiency Entangled Photon Source Project

    Data.gov (United States)

    National Aeronautics and Space Administration — MagiQ proposes to develop a compact tunable high-efficiency low-power-consumption entangled photon source. The source, based on inter-Fabry-Perot-cavity Spontaneous...

  13. High Efficiency Lighting with Integrated Adaptive Control (HELIAC) Project

    Data.gov (United States)

    National Aeronautics and Space Administration — The innovation of the proposed project is the development of High Efficiency Lighting with Integrated Adaptive Control (HELIAC) systems to drive plant growth. Solar...

  14. High Efficiency Direct Methane Solid Oxide Fuel Cell System Project

    Data.gov (United States)

    National Aeronautics and Space Administration — NASA has a defined need for energy dense and highly efficient energy storage and power delivery systems for future space missions. Compared to other fuel cell...

  15. An Improved, Highly Efficient Method for the Synthesis of Bisphenols

    Directory of Open Access Journals (Sweden)

    L. S. Patil

    2011-01-01

    Full Text Available An efficient synthesis of bisphenols is described by condensation of substituted phenols with corresponding cyclic ketones in presence of cetyltrimethylammonium chloride and 3-mercaptopropionic acid as a catalyst in extremely high purity and yields.

  16. Computational Intelligence and Wavelet Transform Based Metamodel for Efficient Generation of Not-Yet Simulated Waveforms.

    Science.gov (United States)

    Oltean, Gabriel; Ivanciu, Laura-Nicoleta

    2016-01-01

    The design and verification of complex electronic systems, especially the analog and mixed-signal ones, prove to be extremely time consuming tasks, if only circuit-level simulations are involved. A significant amount of time can be saved if a cost effective solution is used for the extensive analysis of the system, under all conceivable conditions. This paper proposes a data-driven method to build fast to evaluate, but also accurate metamodels capable of generating not-yet simulated waveforms as a function of different combinations of the parameters of the system. The necessary data are obtained by early-stage simulation of an electronic control system from the automotive industry. The metamodel development is based on three key elements: a wavelet transform for waveform characterization, a genetic algorithm optimization to detect the optimal wavelet transform and to identify the most relevant decomposition coefficients, and an artificial neuronal network to derive the relevant coefficients of the wavelet transform for any new parameters combination. The resulted metamodels for three different waveform families are fully reliable. They satisfy the required key points: high accuracy (a maximum mean squared error of 7.1x10-5 for the unity-based normalized waveforms), efficiency (fully affordable computational effort for metamodel build-up: maximum 18 minutes on a general purpose computer), and simplicity (less than 1 second for running the metamodel, the user only provides the parameters combination). The metamodels can be used for very efficient generation of new waveforms, for any possible combination of dependent parameters, offering the possibility to explore the entire design space. A wide range of possibilities becomes achievable for the user, such as: all design corners can be analyzed, possible worst-case situations can be investigated, extreme values of waveforms can be discovered, sensitivity analyses can be performed (the influence of each parameter on the

  17. Computational Intelligence and Wavelet Transform Based Metamodel for Efficient Generation of Not-Yet Simulated Waveforms

    Science.gov (United States)

    Oltean, Gabriel; Ivanciu, Laura-Nicoleta

    2016-01-01

    The design and verification of complex electronic systems, especially the analog and mixed-signal ones, prove to be extremely time consuming tasks, if only circuit-level simulations are involved. A significant amount of time can be saved if a cost effective solution is used for the extensive analysis of the system, under all conceivable conditions. This paper proposes a data-driven method to build fast to evaluate, but also accurate metamodels capable of generating not-yet simulated waveforms as a function of different combinations of the parameters of the system. The necessary data are obtained by early-stage simulation of an electronic control system from the automotive industry. The metamodel development is based on three key elements: a wavelet transform for waveform characterization, a genetic algorithm optimization to detect the optimal wavelet transform and to identify the most relevant decomposition coefficients, and an artificial neuronal network to derive the relevant coefficients of the wavelet transform for any new parameters combination. The resulted metamodels for three different waveform families are fully reliable. They satisfy the required key points: high accuracy (a maximum mean squared error of 7.1x10-5 for the unity-based normalized waveforms), efficiency (fully affordable computational effort for metamodel build-up: maximum 18 minutes on a general purpose computer), and simplicity (less than 1 second for running the metamodel, the user only provides the parameters combination). The metamodels can be used for very efficient generation of new waveforms, for any possible combination of dependent parameters, offering the possibility to explore the entire design space. A wide range of possibilities becomes achievable for the user, such as: all design corners can be analyzed, possible worst-case situations can be investigated, extreme values of waveforms can be discovered, sensitivity analyses can be performed (the influence of each parameter on the

  18. Computational Intelligence and Wavelet Transform Based Metamodel for Efficient Generation of Not-Yet Simulated Waveforms.

    Directory of Open Access Journals (Sweden)

    Gabriel Oltean

    Full Text Available The design and verification of complex electronic systems, especially the analog and mixed-signal ones, prove to be extremely time consuming tasks, if only circuit-level simulations are involved. A significant amount of time can be saved if a cost effective solution is used for the extensive analysis of the system, under all conceivable conditions. This paper proposes a data-driven method to build fast to evaluate, but also accurate metamodels capable of generating not-yet simulated waveforms as a function of different combinations of the parameters of the system. The necessary data are obtained by early-stage simulation of an electronic control system from the automotive industry. The metamodel development is based on three key elements: a wavelet transform for waveform characterization, a genetic algorithm optimization to detect the optimal wavelet transform and to identify the most relevant decomposition coefficients, and an artificial neuronal network to derive the relevant coefficients of the wavelet transform for any new parameters combination. The resulted metamodels for three different waveform families are fully reliable. They satisfy the required key points: high accuracy (a maximum mean squared error of 7.1x10-5 for the unity-based normalized waveforms, efficiency (fully affordable computational effort for metamodel build-up: maximum 18 minutes on a general purpose computer, and simplicity (less than 1 second for running the metamodel, the user only provides the parameters combination. The metamodels can be used for very efficient generation of new waveforms, for any possible combination of dependent parameters, offering the possibility to explore the entire design space. A wide range of possibilities becomes achievable for the user, such as: all design corners can be analyzed, possible worst-case situations can be investigated, extreme values of waveforms can be discovered, sensitivity analyses can be performed (the influence of each

  19. Contributions to Desktop Grid Computing : From High Throughput Computing to Data-Intensive Sciences on Hybrid Distributed Computing Infrastructures

    OpenAIRE

    Fedak, Gilles

    2015-01-01

    Since the mid 90’s, Desktop Grid Computing - i.e the idea of using a large number of remote PCs distributed on the Internet to execute large parallel applications - has proved to be an efficient paradigm to provide a large computational power at the fraction of the cost of a dedicated computing infrastructure.This document presents my contributions over the last decade to broaden the scope of Desktop Grid Computing. My research has followed three different directions. The first direction has ...

  20. An efficient numerical algorithm for computing densely distributed positive interior transmission eigenvalues

    Science.gov (United States)

    Li, Tiexiang; Huang, Tsung-Ming; Lin, Wen-Wei; Wang, Jenn-Nan

    2017-03-01

    We propose an efficient eigensolver for computing densely distributed spectra of the two-dimensional transmission eigenvalue problem (TEP), which is derived from Maxwell’s equations with Tellegen media and the transverse magnetic mode. The governing equations, when discretized by the standard piecewise linear finite element method, give rise to a large-scale quadratic eigenvalue problem (QEP). Our numerical simulation shows that half of the positive eigenvalues of the QEP are densely distributed in some interval near the origin. The quadratic Jacobi-Davidson method with a so-called non-equivalence deflation technique is proposed to compute the dense spectrum of the QEP. Extensive numerical simulations show that our proposed method processes the convergence efficiently, even when it needs to compute more than 5000 desired eigenpairs. Numerical results also illustrate that the computed eigenvalue curves can be approximated by nonlinear functions, which can be applied to estimate the denseness of the eigenvalues for the TEP.

  1. Efficient computation of GW energy level corrections for molecules described in a plane wave basis

    Science.gov (United States)

    Rousseau, Bruno; Laflamme Janssen, Jonathan; Côté, Michel

    2013-03-01

    An efficient computational approach is presented to compute the ionisation energy and quasiparticle band gap at the level of the GW approximation when the Hilbert space is described in terms of plane waves. The method relies on ab initio calculations as a starting point. Then, the use of the Sternheimer equation eliminates slowly convergent sums on conduction states. Further, the Lanczos method is used to efficiently extract the most important eigenstates of the dielectric operator. This approach avoids the explicit computation of matrix elements of the dielectric operator in the plane wave basis, a crippling bottleneck of the brute force approach. The method is initially applied to organic molecules of current interest in the field of organic photovoltaics. Given the completeness of the plane wave basis, systematic convergence studies can be conducted. Furthermore, the method can readily be extended to describe polymers, which are also of interest for photovoltaic applications, but remain a significant computational challenge for methods based on localized basis sets.

  2. Accurate and efficient computation of nonlocal potentials based on Gaussian-sum approximation

    Science.gov (United States)

    Exl, Lukas; Mauser, Norbert J.; Zhang, Yong

    2016-12-01

    We introduce an accurate and efficient method for the numerical evaluation of nonlocal potentials, including the 3D/2D Coulomb, 2D Poisson and 3D dipole-dipole potentials. Our method is based on a Gaussian-sum approximation of the singular convolution kernel combined with a Taylor expansion of the density. Starting from the convolution formulation of the nonlocal potential, for smooth and fast decaying densities, we make a full use of the Fourier pseudospectral (plane wave) approximation of the density and a separable Gaussian-sum approximation of the kernel in an interval where the singularity (the origin) is excluded. The potential is separated into a regular integral and a near-field singular correction integral. The first is computed with the Fourier pseudospectral method, while the latter is well resolved utilizing a low-order Taylor expansion of the density. Both parts are accelerated by fast Fourier transforms (FFT). The method is accurate (14-16 digits), efficient (O (Nlog ⁡ N) complexity), low in storage, easily adaptable to other different kernels, applicable for anisotropic densities and highly parallelizable.

  3. One high-efficiency analysis method for high-speed circuit networks containing distributed parameter elements

    Institute of Scientific and Technical Information of China (English)

    Lei DOU; Zhiquan WANG

    2005-01-01

    In the field of high-speed circuits,the analysis of mixed circuit networks containing both distributed parameter elements and lumped parameter elements becomes ever important.This paper presents a new method for analyzing mixed circuit networks.It adds transmission line end currents to the circuit variables of the classical modified nodal approach and can be applied directly to the mixed circuit networks.We also introduce a frequency-domain technique without requiring decoupling for multiconductor transmission lines.The two methods are combined together to efficiently analyze high-speed circuit networks containing uniform,nonuniform,and frequency-dependent transmission lines.Numerical experiment is presented and the results are compared with that computed by PSPICE.

  4. Global climate change: Mitigation opportunities high efficiency large chiller technology

    Energy Technology Data Exchange (ETDEWEB)

    Stanga, M.V.

    1997-12-31

    This paper, comprised of presentation viewgraphs, examines the impact of high efficiency large chiller technology on world electricity consumption and carbon dioxide emissions. Background data are summarized, and sample calculations are presented. Calculations show that presently available high energy efficiency chiller technology has the ability to substantially reduce energy consumption from large chillers. If this technology is widely implemented on a global basis, it could reduce carbon dioxide emissions by 65 million tons by 2010.

  5. Highly efficient procedure for the transesterification of vegetable oil

    Energy Technology Data Exchange (ETDEWEB)

    Liang, Xuezheng; Gao, Shan; He, Mingyuan [Shanghai Key Laboratory of Green Chemistry and Chemical Process, Department of Chemistry, East China Normal University, Shanghai 200062 (China); Yang, Jianguo [Shanghai Key Laboratory of Green Chemistry and Chemical Process, Department of Chemistry, East China Normal University, Shanghai 200062 (China); Energy Institute, Department of Materials Science and Engineering, Pennsylvania State University, University Park, PA 16802 (United States)

    2009-10-15

    The highly efficient procedure has been developed for the synthesis of biodiesel from vegetable oil and methanol. The KF/MgO has been selected as the most efficient catalyst for the reactions with the yield of 99.3%. Operational simplicity, without need of the purification of raw vegetable oil, low cost of the catalyst used, high activities, no saponification and reusability are the key features of this methodology. (author)

  6. The photonic nanowire: A highly efficient single-photon source

    DEFF Research Database (Denmark)

    Gregersen, Niels

    2014-01-01

    The photonic nanowire represents an attractive platform for a quantum light emitter. However, careful optical engineering using the modal method, which elegantly allows access to all relevant physical parameters, is crucial to ensure high efficiency.......The photonic nanowire represents an attractive platform for a quantum light emitter. However, careful optical engineering using the modal method, which elegantly allows access to all relevant physical parameters, is crucial to ensure high efficiency....

  7. Microresonator Kerr frequency combs with high conversion efficiency

    CERN Document Server

    Xue, Xiaoxiao; Xuan, Yi; Qi, Minghao; Weiner, Andrew M

    2016-01-01

    Microresonator-based Kerr frequency comb (microcomb) generation can potentially revolutionize a variety of applications ranging from telecommunications to optical frequency synthesis. However, phase-locked microcombs have generally had low conversion efficiency limited to a few percent. Here we report experimental results that achieve ~30% conversion efficiency (~200 mW on-chip comb power excluding the pump) in the fiber telecommunication band with broadband mode-locked dark-pulse combs. We present a general analysis on the efficiency which is applicable to any phase-locked microcomb state. The effective coupling condition for the pump as well as the duty cycle of localized time-domain structures play a key role in determining the conversion efficiency. Our observation of high efficiency comb states is relevant for applications such as optical communications which require high power per comb line.

  8. Bayesian uncertainty quantification and propagation in molecular dynamics simulations: A high performance computing framework

    Science.gov (United States)

    Angelikopoulos, Panagiotis; Papadimitriou, Costas; Koumoutsakos, Petros

    2012-10-01

    We present a Bayesian probabilistic framework for quantifying and propagating the uncertainties in the parameters of force fields employed in molecular dynamics (MD) simulations. We propose a highly parallel implementation of the transitional Markov chain Monte Carlo for populating the posterior probability distribution of the MD force-field parameters. Efficient scheduling algorithms are proposed to handle the MD model runs and to distribute the computations in clusters with heterogeneous architectures. Furthermore, adaptive surrogate models are proposed in order to reduce the computational cost associated with the large number of MD model runs. The effectiveness and computational efficiency of the proposed Bayesian framework is demonstrated in MD simulations of liquid and gaseous argon.

  9. An Efficient Method for Solving Spread Option Pricing Problem: Numerical Analysis and Computing

    Directory of Open Access Journals (Sweden)

    R. Company

    2016-01-01

    Full Text Available This paper deals with numerical analysis and computing of spread option pricing problem described by a two-spatial variables partial differential equation. Both European and American cases are treated. Taking advantage of a cross derivative removing technique, an explicit difference scheme is developed retaining the benefits of the one-dimensional finite difference method, preserving positivity, accuracy, and computational time efficiency. Numerical results illustrate the interest of the approach.

  10. A computationally efficient denoising and hole-filling method for depth image enhancement

    Science.gov (United States)

    Liu, Soulan; Chen, Chen; Kehtarnavaz, Nasser

    2016-04-01

    Depth maps captured by Kinect depth cameras are being widely used for 3D action recognition. However, such images often appear noisy and contain missing pixels or black holes. This paper presents a computationally efficient method for both denoising and hole-filling in depth images. The denoising is achieved by utilizing a combination of Gaussian kernel filtering and anisotropic filtering. The hole-filling is achieved by utilizing a combination of morphological filtering and zero block filtering. Experimental results using the publicly available datasets are provided indicating the superiority of the developed method in terms of both depth error and computational efficiency compared to three existing methods.

  11. High efficiency USC power plant - present status and future potential

    Energy Technology Data Exchange (ETDEWEB)

    Blum, R. [Faelleskemikerne I/S Fynsvaerket (Denmark); Hald, J. [Elsam/Elkraft/TU Denmark (Denmark)

    1998-12-31

    Increasing demand for energy production with low impact on the environment and minimised fuel consumption can be met with high efficient coal fired power plants with advanced steam parameters. An important key to this improvement is the development of high temperature materials with optimised mechanical strength. Based on the results of more than ten years of development a coal fired power plant with an efficiency above 50 % can now be realised. Future developments focus on materials which enable an efficiency of 52-55 %. (orig.) 25 refs.

  12. DEVELOPMENT OF A COMPUTER PROGRAM TO SUPPORT AN EFFICIENT NON-REGRESSION TEST OF A THERMAL-HYDRAULIC SYSTEM CODE

    Directory of Open Access Journals (Sweden)

    JUN YEOB LEE

    2014-10-01

    Full Text Available During the development process of a thermal-hydraulic system code, a non-regression test (NRT must be performed repeatedly in order to prevent software regression. The NRT process, however, is time-consuming and labor-intensive. Thus, automation of this process is an ideal solution. In this study, we have developed a program to support an efficient NRT for the SPACE code and demonstrated its usability. This results in a high degree of efficiency for code development. The program was developed using the Visual Basic for Applications and designed so that it can be easily customized for the NRT of other computer codes.

  13. Computational study of power conversion and luminous efficiency performance for semiconductor quantum dot nanophosphors on light-emitting diodes.

    Science.gov (United States)

    Erdem, Talha; Nizamoglu, Sedat; Demir, Hilmi Volkan

    2012-01-30

    We present power conversion efficiency (PCE) and luminous efficiency (LE) performance levels of high photometric quality white LEDs integrated with quantum dots (QDs) achieving an averaged color rendering index of ≥90 (with R9 at least 70), a luminous efficacy of optical radiation of ≥380 lm/W(opt) a correlated color temperature of ≤4000 K, and a chromaticity difference dC quantum efficiency of 43%, 61%, and 80% in film, respectively, using state-of-the-art blue LED chips (81.3% PCE). Furthermore, our computational analyses suggest that QD-LEDs can be both photometrically and electrically more efficient than phosphor based LEDs when state-of-the-art QDs are used.

  14. Energy-Efficient Computational Chemistry: Comparison of x86 and ARM Systems.

    Science.gov (United States)

    Keipert, Kristopher; Mitra, Gaurav; Sunriyal, Vaibhav; Leang, Sarom S; Sosonkina, Masha; Rendell, Alistair P; Gordon, Mark S

    2015-11-10

    The computational efficiency and energy-to-solution of several applications using the GAMESS quantum chemistry suite of codes is evaluated for 32-bit and 64-bit ARM-based computers, and compared to an x86 machine. The x86 system completes all benchmark computations more quickly than either ARM system and is the best choice to minimize time to solution. The ARM64 and ARM32 computational performances are similar to each other for Hartree-Fock and density functional theory energy calculations. However, for memory-intensive second-order perturbation theory energy and gradient computations the lower ARM32 read/write memory bandwidth results in computation times as much as 86% longer than on the ARM64 system. The ARM32 system is more energy efficient than the x86 and ARM64 CPUs for all benchmarked methods, while the ARM64 CPU is more energy efficient than the x86 CPU for some core counts and molecular sizes.

  15. Towards robust dynamical decoupling and high fidelity adiabatic quantum computation

    Science.gov (United States)

    Quiroz, Gregory

    Quantum computation (QC) relies on the ability to implement high-fidelity quantum gate operations and successfully preserve quantum state coherence. One of the most challenging obstacles for reliable QC is overcoming the inevitable interaction between a quantum system and its environment. Unwanted interactions result in decoherence processes that cause quantum states to deviate from a desired evolution, consequently leading to computational errors and loss of coherence. Dynamical decoupling (DD) is one such method, which seeks to attenuate the effects of decoherence by applying strong and expeditious control pulses solely to the system. Provided the pulses are applied over a time duration sufficiently shorter than the correlation time associated with the environment dynamics, DD effectively averages out undesirable interactions and preserves quantum states with a low probability of error, or fidelity loss. In this study various aspects of this approach are studied from sequence construction to applications of DD to protecting QC. First, a comprehensive examination of the error suppression properties of a near-optimal DD approach is given to understand the relationship between error suppression capabilities and the number of required DD control pulses in the case of ideal, instantaneous pulses. While such considerations are instructive for examining DD efficiency, i.e., performance vs the number of control pulses, high-fidelity DD in realizable systems is difficult to achieve due to intrinsic pulse imperfections which further contribute to decoherence. As a second consideration, it is shown how one can overcome this hurdle and achieve robustness and recover high-fidelity DD in the presence of faulty control pulses using Genetic Algorithm optimization and sequence symmetrization. Thirdly, to illustrate the implementation of DD in conjunction with QC, the utilization of DD and quantum error correction codes (QECCs) as a protection method for adiabatic quantum

  16. A highly efficient multi-core algorithm for clustering extremely large datasets

    Directory of Open Access Journals (Sweden)

    Kraus Johann M

    2010-04-01

    Full Text Available Abstract Background In recent years, the demand for computational power in computational biology has increased due to rapidly growing data sets from microarray and other high-throughput technologies. This demand is likely to increase. Standard algorithms for analyzing data, such as cluster algorithms, need to be parallelized for fast processing. Unfortunately, most approaches for parallelizing algorithms largely rely on network communication protocols connecting and requiring multiple computers. One answer to this problem is to utilize the intrinsic capabilities in current multi-core hardware to distribute the tasks among the different cores of one computer. Results We introduce a multi-core parallelization of the k-means and k-modes cluster algorithms based on the design principles of transactional memory for clustering gene expression microarray type data and categorial SNP data. Our new shared memory parallel algorithms show to be highly efficient. We demonstrate their computational power and show their utility in cluster stability and sensitivity analysis employing repeated runs with slightly changed parameters. Computation speed of our Java based algorithm was increased by a factor of 10 for large data sets while preserving computational accuracy compared to single-core implementations and a recently published network based parallelization. Conclusions Most desktop computers and even notebooks provide at least dual-core processors. Our multi-core algorithms show that using modern algorithmic concepts, parallelization makes it possible to perform even such laborious tasks as cluster sensitivity and cluster number estimation on the laboratory computer.

  17. High Efficiency of Two Efficient QSDC with Authentication Is at the Cost of 1Their Security

    Institute of Scientific and Technical Information of China (English)

    QIN Su-Juan; WEN Qiao-Yan; MENG Luo-Ming; ZHU Fu-Chen

    2009-01-01

    Two efficient protocols of quantum secure direct communication with authentication [Chin. Phys. Lett. 25 (2008)2354] were recently proposed by Liu et al. to improve the efficiency of two protocols presented in [Phys. Rev. A 75 (2007) 026301] by four Pauli operations. We show that the high efficiency of the two protocols is at the expense of their security. The authenticator Trent can reach half the secret by a particular attack strategy in the first protocol. In the second protocol, not only Trent but also an eavesdropper outside can elicit half-information about the secret from the public declaration.

  18. Highly efficient Nd: YAG ceramic CW laser with 59.8% slope-efficiency

    Institute of Scientific and Technical Information of China (English)

    Yunfeng Qi; Qihong Lou; Haixia Ma; Jingxing Dong

    2005-01-01

    @@ In this paper, a highly efficient Ti:sapphire end-pumped 1 at.-% Nd:YAG ceramic laser that is comparable in efficiency with Nd:YAG single crystal lasers has been developed. Optical absorption and emission spectra for Nd:YAG ceramics have been measured. With 673-mW pumping, 295-mW laser output at 1064 nm has been obtained. The laser threshold is only 13 mW. Deducted the transmitted light, the corresponding optical-to-optical conversion efficiency is 58.4%. The lasing characteristics of Nd:YAG ceramic are nearly equal to those of Nd:YAG single crystal.

  19. Very-High Efficiency, High Power Laser Diodes Project

    Data.gov (United States)

    National Aeronautics and Space Administration — AdTech Photonics, in collaboration with the Center for Advanced Studies in Photonics Research (CASPR) at UMBC, is pleased to submit this proposal entitled ?Very-High...

  20. Research on stable, high-efficiency, amorphous silicon multijunction modules

    Energy Technology Data Exchange (ETDEWEB)

    Catalano, A.; Bennett, M.; Chen, L.; D' Aiello, R.; Fieselmann, B.; Li, Y.; Newton, J.; Podlesny, R.; Yang, L. (Solarex Corp., Newtown, PA (United States). Thin Film Div.)

    1992-08-01

    This report describes work to demonstrate a multijunction module with a stabilized'' efficiency (600 h, 50{degrees}C, AM1.5) of 10.5%. Triple-junction devices and modules using a-Si:H alloys with carbon and germanium were developed to meet program goals. ZnO was used to provide a high optical transmission front contact. Proof of concept was obtained for several important advances deemed to be important for obtaining high (12.5%) stabilized efficiency. They were (1) stable, high-quality a-SiC:H devices and (2) high-transmission, textured ZnO. Although these developments were not scaled up and included in modules, triple-junction module efficiencies as high as 10.85% were demonstrated. NREL measured 9.62% and 9.00% indoors and outdoors, respectively. The modules are expected to lose no more than 20% of their initial performance. 28 refs.